Next Article in Journal
Spatio-Temporal Variation of Critical Relative Humidity Based on Multiple Datasets
Previous Article in Journal
Bitemporal Remote Sensing Image Change Detection Network Based on Siamese-Attention Feedback Architecture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Attention Multi-Image Super-Resolution Transformer (MAST) for Remote Sensing

1
Aerospace Information Research Institute, Chinese Academy of Sciences, No.9 Dengzhuang South Road, Haidian District, Beijing 100094, China
2
School of Optoelectronics, University of Chinese Academy of Sciences, No.19(A) Yuquan Road, Shijingshan District, Beijing 100049, China
3
Department of Key Laboratory of Computational Optical Imagine Technology, CAS, No.9 Dengzhuang South Road, Haidian District, Beijing 100094, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2023, 15(17), 4183; https://doi.org/10.3390/rs15174183
Submission received: 16 July 2023 / Revised: 17 August 2023 / Accepted: 24 August 2023 / Published: 25 August 2023

Abstract

:
Deep-learning-driven multi-image super-resolution (MISR) reconstruction techniques have significant application value in the field of aerospace remote sensing. In particular, Transformer-based models have shown outstanding performance in super-resolution tasks. However, current MISR models have some deficiencies in the application of multi-scale information and the modeling of the attention mechanism, leading to an insufficient utilization of complementary information in multiple images. In this context, we innovatively propose a Multi-Attention Multi-Image Super-Resolution Transformer (MAST), which involves improvements in two main aspects. Firstly, we present a Multi-Scale and Mixed Attention Block (MMAB). With its multi-scale structure, the network is able to extract image features from different scales to obtain more contextual information. Additionally, the introduction of mixed attention allows the network to fully explore high-frequency features of the images in both channel and spatial dimensions. Secondly, we propose a Collaborative Attention Fusion Block (CAFB). By incorporating channel attention into the self-attention layer of the Transformer, we aim to better establish global correlations between multiple images. To improve the network’s perception ability of local detailed features, we introduce a Residual Local Attention Block (RLAB). With the aforementioned improvements, our model can better extract and utilize non-redundant information, achieving a superior restoration effect that balances the global structure and local details of the image. The results from the comparative experiments reveal that our approach demonstrated a notable enhancement in cPSNR, with improvements of 0.91 dB and 0.81 dB observed in the NIR and RED bands of the PROBA-V dataset, respectively, in comparison to the existing state-of-the-art methods. Extensive experiments demonstrate that the method proposed in this paper can provide a valuable reference for solving multi-image super-resolution tasks for remote sensing.

1. Introduction

Space optical remote sensing utilizes optical cameras in space to observe the Earth or space targets, capturing remote sensing images that reflect their reflection or the radiation information of the targets. It enables applications such as classification, identification, and comprehensive detection, serving as an important means of obtaining information for global change monitoring, regional surveillance, resource exploration, urban planning, and other fields [1]. The higher the geometric resolution of a space optical camera, the greater the precision in differentiating, identifying, and classifying targets.
Although increasing the optical aperture and focal length can directly improve the resolution, it significantly increases the cost, weight, volume, and development complexity of the camera. Additionally, regardless of the size, aperture, or focal length of the imaging system, the imaging resolution is inevitably limited by physical factors in the imaging chain. This leads to a decrease in the resolution index and is primarily caused by atmospheric blurring, optical aberrations and diffraction-induced blurring, aliasing due to an undersampling of CCD/CMOS detectors, and electronic noise [2].
Therefore, overcoming the physical limitations mentioned above and improving the resolution without changing the optical hardware is an important issue that needs to be addressed in space-based optical remote sensing. Image super-resolution reconstruction technology can overcome the limitations of traditional optical imaging and is an effective approach to solve the associated problems [3].
Image super-resolution reconstruction is the reconstruction of a high-resolution (HR) image from one or multiple low-resolution (LR) images. It possesses ill-posedness, falls within the realm of inverse problems, and represents a challenging research direction [4]. Since the problem was proposed by Harris [5] in 1964, scholars have proposed various super-resolution methods around the inverse convolution algorithm [6]. With the rapid development of deep learning technology, the use of neural networks to accomplish super-resolution reconstruction tasks has become a frontier approach in academic research, and a series of excellent super-resolution algorithms have emerged, often achieving state-of-the-art performance in various benchmarks of super-resolution [7]. In terms of network models, they can be divided into convolutional neural network algorithms [8,9,10], generative adversarial neural network algorithms [11,12,13], Transformer network algorithms [14,15,16,17], etc.
Single-image super-resolution aims to reconstruct an HR image from a noisy and low-quality LR image [18]. Multi-image super-resolution utilizes complementary information between multiple LR images of the same scene [19] to reconstruct an HR image through sub-pixel level information fusion [20,21]. Compared to single-image super-resolution algorithms, multi-image super-resolution (MISR) algorithms leverage complementary sub-pixel information [22], which can yield better reconstruction results [23].
The principle of the MISR algorithm is to extract and fuse sub-pixel features from multiple images to reconstruct the high-frequency information of the image and improve the resolution. It primarily focuses on three key aspects: (1) establishing a mechanism for extracting and fusing complementary sub-pixel features; (2) reducing the impact of registration errors between multiple images on the reconstruction; and (3) suppressing the influence of blurring and noise contamination on the reconstruction. In recent years, researchers in the field of remote sensing have conducted innovative research in high-frequency information feature extraction [24,25], registration error suppression [26,27,28], noise error suppression [29,30], and other aspects related to these three key aspects. However, there are still some issues that need to be addressed.
In terms of feature extraction, the existing algorithms are still inadequate in extracting the depth features that contribute to the super-resolution task in multiple images; the complementarity of features between each image is not sufficient, and features that do not contribute to the reconstruction task are also passed to the feature fusion step, which affects the reconstruction effect. In terms of feature fusion, the existing algorithms are still inadequate in terms of the local detail fusion mechanism, which leads to the fusion of sub-pixel features of multiple images while also fusing noise and blurred features, affecting the reconstruction effect. Finally, no optimization algorithm has been proposed that can comprehensively address the three abovementioned key aspects.
Based on the analysis of the limitations of the existing MISR methods for remote sensing, this paper innovatively proposes a synthetically improved Multi-Attention Multi-Image Super-Resolution Transformer (MAST). This network mitigates the issue of effective information loss that is present in existing methods by employing a multi-scale and multi-attention approach. Comparative experiments showed that the proposed algorithm achieved state-of-the-art super-resolution results in the NIR and RED bands of the PROBA-V MISR dataset, which can provide valuable references and insights for future studies in this field. In brief, the main contributions of this work are as follows:
  • We innovatively proposed a Multi-Scale and Mixed Attention Block (MMAB) that enabled a model to fully extract complementary information from multiple images. Specifically, we introduced a Multi-Scale Extraction Block (MSEB) to obtain more reference information and effectively suppress registration errors. We also presented a Channel Attention Guided Block (CAGB) and a Spatial Attention Guided Block (SAGB) to enhance the model’s perception of high-frequency features with high contributions to the super-resolution task in each image.
  • We also presented a Collaborative Attention Fusion Block (CAFB) based on the Transformer, which constructed a more comprehensive feature fusion mechanism incorporating self-attention, channel attention, and local attention. The embedding of channel attention enhanced the advantage of the Transformer in establishing global correlations between multiple frames. The introduction of the Residual Local Attention Block (RLAB) largely broke through the limitations of the Transformer in capturing local high-frequency information, enabling the model to simultaneously restore the overall structure and local details of the image.

2. Related Work

2.1. Multi-Image Super-Resolution

The multi-image degradation model for the same scene imaged k times can be expressed as follows:
y k = [ A     H k     C k ]     x + n k
where y k   represents the k-th LR image, x represents the HR image, the symbol ⊗ denotes convolution, n k   represents additive system noise, A represents the detector sampling coefficients, H k   denotes the blur coefficients, and C k   represents the inter-frame displacement relationship function. Multi-image super-resolution reconstruction involves solving for the HR image from this equation.
Tsai and Huang [31] first proposed the concept of multi-frame image super-resolution in 1984. They presented a frequency domain method that improves the spatial resolution of an image by utilizing the complementary information brought by the relative displacement between multiple LR images. However, the limitation of frequency domain methods is that they are restricted to global translational motion and linear space-invariant degradation models. Spatial domain methods, on the other hand, possess strong capabilities of incorporating prior constraints. These methods include iterative back-projection methods, convex set projection methods, maximum a posteriori methods, and more [32].
Although the deconvolution algorithm provides a solution for MISR, using it to address the multi-image super-resolution reconstruction problem involves solving an ill-posed problem that is inherently unstable. Registration errors and modeling errors are gradually amplified during the iterative process, making it difficult to effectively suppress them. With the continuous development of artificial intelligence technology, deep learning has provided a new approach to solve the multi-image super-resolution problem using deep neural networks. This approach avoids the instability caused by the ill-posed nature of deconvolution and provides a more effective solution.

2.2. Deep Learning for Multi-Image Super-Resolution

Deep learning methods have shown great potential in addressing multi-frame super-resolution problems [33,34,35,36,37]. Deudon et al. [22] proposed HighRes-net, which first utilizes deep learning to solve the multi-frame super-resolution task. They achieved implicit registration by setting a reference frame channel and recursively fused LR image pairs using a global fusion operator. Molini et al. [38] introduced the registration filter RegNet with feature learning capabilities in DeepSUM, integrating the image registration task into a CNN. They further incorporated non-local characteristics into the network using graph convolution methods in DeepSUM++. Arefin et al. [39] presented MISR-GRU, which employs the ConvGRU module to compute the hidden state of the feature representation generated by the encoder and utilizes global average pooling to obtain fusion features. Dorr et al. [40] introduced 3DWDSRNet, a 3D wide activation residual network that replaces 3D convolutions with wide activation blocks to enhance the spatio-temporal correlation of multiple images. They also employed frame permutation techniques for data augmentation.
By introducing attention mechanisms, the model’s perception and utilization of high-frequency features can be enhanced, leading to a better reconstruction of image details and further improving the performance of super-resolution tasks [41,42,43,44]. Salvetti et al. [45] introduced a 3D convolutional feature attention mechanism with embedded residual connections in the RAMS network, enabling the network to focus on extracting high-frequency information in both the temporal and spatial dimensions. Valsesia et al. [46] utilized self-attention mechanisms in the temporal dimension and studied the relationship between the uncertainty of super-resolution images and temporal variations. They constructed an MISR model, PIUnet, which possesses permutation invariance for temporal images. In the burst super-resolution task, Bhat et al. [47] aligned multiple frames using optical flow and employed attention-based adaptive weighting and fusion methods to generate denoised super-resolution RGB images.

2.3. Transformers for Multi-Image Super-Resolution

The Transformer model is a deep learning model based on the encoder–decoder architecture with a self-attention mechanism [48]. Compared to RNNs and CNNs, the Transformer model has two key advantages. Firstly, it enables fully parallel computation, which improves the efficiency of training and inference. Secondly, the Transformer has a global receptive field, which allows it to capture global information more effectively. Recently, significant progress has been made in image super-resolution research using the Transformer model [14,15,17,36,49].
An et al. [50] were the first to apply a Transformer to the remote sensing image MISR task, proposing TR-MISR, and through extensive experiments, they demonstrated the effectiveness of the Transformer model in multi-frame fusion and in improving the model’s robustness to noise. Luo et al. [36] reduced alignment errors by combining optical flow estimation and deformable convolution and proposed BSRT, utilizing blocks and groups based on the Swin Transformer as the backbone structure to more effectively extract global contextual information. Dudhane et al. [37] introduced Burstormer, a Transformer-based hierarchical model, for burst image restoration and enhancement tasks. This model improved feature alignment and fusion by introducing enhanced deformable alignment modules and reference-free feature enrichment modules, utilizing multi-scale local and non-local features.

3. Methodology

3.1. Network Architecture

The overall network of the Multi-Attention Multi-Image Super-Resolution Transformer (MAST) is illustrated in Figure 1, which consists of four parts: shallow feature extraction, deep feature extraction, collaborative attention feature fusion, and HR image reconstruction.
In particular, the MAST network takes ILR as its input, which signifies a compilation of LR images captured from the identical scene. The shape of ILR is H × W × k × C0, with H, W, k, and C0 denoting the image’s height, width, the number of LR images, and the channels per image, respectively. It is noteworthy that the value of C0 is 1 in this context. To streamline our notation for simplicity, we abbreviate the shape of ILR as H × W × k.
Initially, we determined the median image of the input image collection using the median operation. This step allowed us to establish a reference image, denoted as Ref, for each individual LR image. The process is illustrated below. In this context, x and y correspond to the pixel positions within the image, while i represents the index of the i-th among the initial 9 LR images.
Ref [ x , y ] = Median ( I LR [ x , y , i ] ) ,   i = 1 , 2 , , 9
Following this, we proceeded to concatenate each LR image with its corresponding reference image along the channel dimension, leading to the generation of LR-Ref image pairs referred to as IPR. These pairs possess a shape of H × W × 2k and are represented below. Here, j indicates the j-th LR image within the image set ILR, while z signifies the channel of IPR.
I PR [ x , y , z ] = LR j [ x , y , z ] ,           z [ 0 , k     1 ] Ref x , y , z k ,   z [ k , 2 k     1 ] ,   j = 1 , 2 , , k
This form of image pairing can be seen as an implicit initial image registration condition, and it exhibits a degree of resilience against challenges such as noise, distortion, or localized occlusion that are present in the images. It does not depend on explicit feature point matching or correspondence, thereby effectively mitigating some issues related to image misalignment.
The shallow feature extraction layer comprises a 3 × 3 convolutional layer, followed by a PReLU activation function layer. Through the process of feature extraction applied to the input LR images, we obtained the shallow feature FS, which has the shape of H × W × C, where C (C > k) represents the channels of FS.
F S = PReLU ( H 3 × 3   Conv ( I PR ) )
Introducing convolutional layers into the Transformer model not only enhances the network’s training process, but also provides a more comprehensive feature representation for subsequent deep feature extraction and multi-frame information fusion [17]. Convolutional layers excel at capturing local patterns and structures in images, which are crucial for many visual tasks. Incorporating the convolutional layer into the Transformer model enhances the model’s ability to perceive these local features, resulting in a richer feature representation [51]. The important characteristics of convolutional layers are parameter sharing and local connectivity, which, to some extent, reduce the model’s parameter count, making it more amenable to training and optimization. Furthermore, convolutional layers can provide more stable gradient information during the early stages of training, aiding in accelerating the network’s convergence process [52].
Next, we proceeded to conduct deep feature extraction on the input feature, facilitating the network’s capacity to draw out additional complementary information through the integration of multi-scale and mixed attention guidance. This procedure resulted in a more intricate feature representation FD, depicted as follows:
F i = MMAB i ( F i     1 ) ,   i = 1 , 2 , , N , F D = BN ( H 3 × 3   Conv ( F N ) )
where Fi−1 and Fi indicate the input feature and output feature of i-th MMAB, respectively. BN represents the batch normalization operation.
By incorporating collaborative attention, the network in the fusion stage gains the ability to simultaneously grasp both global and local information within the image. This fosters a more comprehensive feature fusion, achieving a harmonious restoration of both the overall image structure and its intricate local details. FM represents the fusion feature derived from the deep features, as demonstrated below. In this context, Fj−1 and Fj indicate the input and output features of the CAFB module, respectively.
F j = CAFB j ( F j 1 ) ,   j = 1 , 2 , , M ,
The HR image reconstruction segment comprises a 3 × 3 convolutional layer, a PReLU layer (the two are denoted together as H3✕3 Conv(·) below for brevity), and a pixel shuffle layer. Through this reconstruction process, we derived the super-resolution image ISR, which can be represented as
I SR = PixelShuffle ( H 3 × 3   Conv ( F M ) )

3.2. Deep Feature Extraction Module

The Multi-Scale and Mixed Attention Block (MMAB) serves as the pivotal module within the network’s deep feature extraction segment, with its structure illustrated in Figure 2. In our pursuit of effectively acquiring complementary high-frequency feature information, we harnessed three distinct mechanism-driven feature extraction modules: MESB, CAGB, and SAGB. Among these, MESB operates on the foundation of multi-scale feature extraction principles, CAGB amalgamates channel dimension compression and expansion techniques with channel attention mechanisms, and SAGB prioritizes the utilization of spatial attention for feature extraction.
By harmonizing the outputs of these three modules in a parallel fashion, our model can apprehend an expansive spectrum of complementary information across various frequencies during the feature extraction phase. This, in turn, enriches the contextual information available for image reconstruction. Additionally, existing research indicates that parallel fusion modules contribute to alleviating the adverse effects of motion blur on the outcomes, thereby enhancing the super-resolution capabilities of the model [53].
Moreover, we incorporated two learnable adaptive weight scaling parameters to dynamically fine-tune the weights of both the residual path and the backbone path [54]. λB signifies the weight assigned to the backbone path, whereas λR represents the weight attributed to the residual path. Throughout the entire training phase of the neural network, as the adaptive weights undergo continuous updates, they not only enhance the network’s focus on high-frequency features, but also improve the propagation of gradient information. This leads to an acceleration in the convergence of the network’s training process. The complete computational process of an MMAB can be succinctly described as follows:
G 1 = MSEB Y ,   G 2   = CAGB Y ,   G 3 = SAGB Y F C [ x , y , z ] = G 1 [ x , y , z ] ,           z [ 0 , C 1 ] G 2 [ x , y , z C ] ,   z [ C , 2 C 1 ] G 3 [ x , y , z 2 C ] ,   z [ 2 C , 3 C 1 ]
Z = λ R · Y + λ B · H 3 × 3 Conv ( H 1 × 1 Conv ( F C ) )
where MSEB(·), CAGB(·), and SAGB(·) correspond to the three sub-blocks present within the MMAB module. G1, G2, and G3 denote the output features of the three modules. Y symbolizes the input feature. FC indicates the concatenated output feature derived from the three branches, whereas Z signifies the resulting feature following adaptive weighting. The shapes of G1, G2, G3, Y, and Z all maintain dimensions of H × W × C, while the shape of FC adopts H × W × 3C. Additionally, H3✕3 Conv(·) and H1✕1 Conv(·) refer to the 3 × 3 and 1 × 1 convolution layers, respectively.

3.2.1. Multi-Scale Extraction Block (MSEB)

Objects exhibit different detailed characteristics at various scales. Multi-scale structures can capture contextual information at different scales through receptive fields of different sizes, thereby enhancing feature representation. Moreover, it was shown that the use of multi-scale structures in networks can help to suppress alignment errors [26,27,28]. We borrowed the idea of a Pyramid Pooling Module (PPM) [55] in our design and introduced adaptive weight scaling parameters to create a residual-structured multiscale feature extraction block (MSEB), the structure of which is shown in Figure 3. By incorporating the adaptive weight scaling parameter [54], the MSEB can dynamically fine-tune the feature contents within both the residual and backbone paths, enabling the model to extract complementary information more efficiently and optimize the network’s training process.
To elaborate, F0 represents the input feature, and FH signifies the features extracted via H3✕3 Conv(·). MSLi(·) denotes the feature extraction layer with varying pooling kernels. The specific procedure entails initiating average pooling to generate feature maps with sizes of 1 × 1, 2 × 2, 3 × 3, and 6 × 6. Among these, the 1 × 1 feature map encapsulates global information, while the remaining three encompass local details. Local features are employed to capture subtle local variations, while global features are utilized to maintain overall coherence. Subsequently, these feature maps are restored to their original sizes using bilinear interpolation. Concat(·) indicates the concatenation of feature maps of identical shape, following the same process as outlined in Equation (8). By fusing multi-scale contextual features, the adverse impacts of local noise and inaccuracies on the overall registration outcome can be effectively alleviated. FA represents the fused multi-scale feature, while FB signifies the output feature after the addition of residual connection. The overall calculation process is described as follows:
F A = H 1 × 1   Conv ( Concat ( MSL 1 ( F H ) , ,   MSL 4 ( F H ) ) )
F B = λ B · H 3 × 3   Conv ( F A ) + λ R · F 0

3.2.2. Channel Attention Guided Block (CAGB)

The Channel Attention Guided Block (CAGB) is a sub-module of our proposed MMAB, the structure of which is shown in Figure 4, designed to direct the network to thoroughly harness high-frequency feature information within the channel dimension. By increasing the dimensionality, we can introduce a greater variety of distinct feature channels, enabling the model to comprehensively capture richer feature information [56]. Additionally, by reducing the dimensionality of the feature map in the channel dimension, the network can focus on more meaningful high-frequency features while reducing data redundancy [57].
Through the application of various dimension transformations, we obtained multi-level contextual information and performed information fusion to extract effective complementary information from the fused features. Furthermore, we introduced the channel attention mechanism, enabling the network to adaptively learn relationships between different feature channels, thus enhancing the perception of crucial high-frequency features.
In the depicted diagram, CTLi denotes the channel attention layer. Among these, CTL1(·) and CTL2(·) indicate the channel expansion of input feature F0, with expansion factors of 2 and 4, respectively, followed by a reduction to match the input channels. On the other hand, CTL3(·) and CTL4(·) signify the dimension reduction in F0, with reduction factors of 1/2 and 1/4, respectively, followed by an expansion. FK and FQ represent the context-fused feature and the output feature after the addition of residual connection, respectively. HCA(·) stands for the channel attention layer. The entire process unfolds as follows:
F K = H 1 × 1   Conv ( Concat ( CTL 1 ( F 0 ) , ,   CTL 4 ( F 0 ) ) )
F Q = λ B · ( H 3 × 3   Conv ( H CA ( F K ) ) ) + λ R · F 0

3.2.3. Spatial Attention Guided Block (SAGB)

The Spatial Attention Guided Block (SAGB) is another constituent of our proposed MMAB, the structure of which is shown in Figure 5, dedicated to delving more deeply into high-frequency feature information from the spatial dimension. By subtracting smoothed filtered feature maps from the original ones, we derived high-frequency filtered features. This process of enhancing high-frequency information can accentuate image details and edges. Subsequently, we applied computed spatial attention weights to the feature map. These weights prioritize different locations, thereby enhancing feature representation in significant image regions and subduing less important areas. This strategy effectively extracts high-frequency information. The computation process of the SAGB can be presented as follows:
F L = H UP ( H Avg ( F G ) )
F E = H 3 × 3   Conv ( H SA ( F G     F L ) )
where FG represents the feature obtained after H3×3 Conv (·) with a size of H × W × C. FL signifies the smoothed feature with a size of H/2 × W/2 × C. FE denotes the output features. HAvg(·) indicates the average pooling operation, while HUP(·) represents bilinear interpolation. HSA(·) represents the spatial attention layer, which adjusts the attention distribution in the spatial dimension, enabling the network to focus on high-frequency information regions.

3.3. Collaborative Attention Feature Fusion Module

The Collaborative Attention Fusion Block (CAFB) is the feature fusion module proposed by us based on the Transformer architecture, and its structure is illustrated in Figure 6. The reshaping process depicted in the figure below involves transforming the input feature map into a suitable format for the Transformer to process tokens. The Transformer model is suitable for multi-frame image processing as it is range-independent in computing attention weights and can capture distant dependencies throughout the input sequence. It demonstrates significant advantages in capturing correlation information between multiple images and enhancing robustness to noise. Our enhancement stems from the incorporation of both channel attention and residual local attention within the Transformer model. Compared to traditional attention mechanisms like CBAM, we adopted a parallel combination approach to simultaneously incorporate both global and local information.
By incorporating the channel attention mechanism to extract importance and correlation information from diverse channels and embedding this information into the Transformer, we further amplified the Transformer’s ability to capture comprehensive correlations among multi-frame information [58,59,60].
Meanwhile, the Residual Local Attention Block (RLAB) enhances the network’s ability to capture local high-frequency features. It does so by employing SimAM attention to extract the importance of neurons through the analytical solution of the energy function, directly inferring the 3D attention weights of the feature map [61]. Ultimately, through the application of weighted processing to the outputs of the two branches, an effective fusion of global and local information is accomplished. This empowers the super-resolution network to effectively restore both the overarching image structure and intricate details with proficiency during the reconstruction process. The complete computation process of CAFB unfolds as follows:
X M = MHA ( LN ( X ) ) + α · CA ( LN ( X ) ) + X
X N = MLP ( LN ( X M ) ) + X M
X U = X N + α · ( H 3 × 3   Conv ( RLAB ( X ) ) + X )
where X signifies the input of the CAFB module, constituting the feature acquired through deep extraction. XU represents the resulting feature, incorporating insights from both the Transformer and RLAB. XM and XN correspond to the interim features derived during the processing stages. MHA corresponds to the standard multi-head attention, LN stands for the LayerNorm layer, and MLP represents the multi-layer perceptron.
To attain optimal fusion performance, we introduced the hyperparameter α. Through judicious weight allocation, we can harmonize the contributions of global structural information and local detail information. This allows for the extraction of comprehensive global structural insights while accentuating crucial local details, culminating in an enhanced feature representation. In the experimental context, we set α to a value of 0.01.

4. Experiments

In this section, we introduce the PROBA-V Kelvin Dataset, which is the MISR dataset used in the satellite imagery experiments. Furthermore, we delve into the exploration of evaluation metrics employed to gauge the quality of super-resolution images. Additionally, we offer a comprehensive introduction to both the data preprocessing process and the configurations of the model parameter settings.

4.1. PROBA-V Kelvin Dataset

The PROBA-V dataset is a real remote sensing image dataset provided by the PROBA-V satellite and serves as a benchmark dataset for multi-frame super-resolution tasks [62]. The dataset consists of satellite data collected from 74 manually selected regions worldwide. It includes imagery captured in two bands: NIR and RED. These images were acquired at different times. The dataset provides images with two different spatial resolutions. The LR images have a resolution of 300 m and a size of 128 × 128 pixels. The HR images, on the other hand, have a resolution of 100 m and a size of 384 × 384 pixels.
The dataset consists of a total of 1450 scenes, with 1160 scenes used for training and 290 scenes used for testing. Each scene includes an HR image and multiple LR images. The number of LR images in each scene ranges from 9 to 35. Since these images were acquired at different times and under different weather conditions, the occlusion of variable features such as clouds and snow may affect the results of the image alignment. The PROBA-V dataset provides a manually annotated quality map for each image in the scene. The quality map is encoded in binary format, using 0 to represent occluded pixels and using 1 to represent non-obscured pixels. Only non-obscured pixels are considered clear pixels, and we performed super-resolution reconstruction only for the portion of the image that is clear.
The PROBA-V dataset is a significant publicly available benchmark dataset that is currently used for remote sensing image tasks and multi-image super-resolution tasks. Several MISR algorithms have been trained on this dataset, providing reliable solutions to the tasks [22,45,46,50]. This dataset serves as compelling evidence to validate the superiority of our algorithms. In the future, we will actively strive to acquire more datasets to verify the generalizability of our model across various data distributions.

4.2. Evaluation Metric

In super-resolution tasks, PSNR and SSIM are commonly used to evaluate objective image quality metrics, but they may not be effective enough for evaluating the quality of super-resolution remote sensing images, especially when there are problems such as pixel shift and brightness bias in the images, and they have less reference value for assessing the quality of reconstructed images. To assess the reconstruction quality of remote sensing images more accurately, we introduced cPSNR and cSSIM, which are metrics that are used in the PROBA-V challenge. These metrics have been improved to address pixel offset, brightness deviation, and cloud-related factors that are present in remote sensing images.
The size of the LR in the validation set is 128 × 128 pixels, and the size of the super-resolution image SR is 384 × 384. As the acquisition of LR and HR images involves separate processes, there might be a slight misalignment at the pixel level between them. It was assumed that the maximum potential offset between the two images is 3 pixels. To account for the potential errors due to misalignment, a 3-pixel-wide strip along the edges of the SR was removed. This resulted in a cropped image denoted as I 3 , 3 SR , which has a size of 378 × 378 pixels. Subsequently, we proceeded to crop the HR image from the upper-left coordinate (u, v), resulting in a cropped image denoted as I u , v HR that is the same size as I 3 , 3 SR . The procedure for computing the brightness deviation between I 3 , 3 SR and I u , v HR is outlined in Equation (19). In this equation, Mu,v represents a binary mask image. ‖·‖denotes the L1 norm of the matrix. When calculating various metrics, we use the mask to restrict the range of statistics and only consider the results from clear pixels.
b u , v = 1 M u , v ( I u , v HR     I 3 , 3 SR ) · M u , v
In the experiments of this paper, we adopted negative cPSNR as the loss function for network training. The cPSNR and cSSIM are calculated as follows:
cMSE = max u , v 0 , 6 1 M u , v [ ( I u , v HR   I 3 , 3 SR ) · M u , v + b u , v ] 2
cPSNR = 10   lg ( cMSE )
cSSIM = max u , v 0 , 6 SSIM ( I u , v HR · M u , v ,   I u , v SR · M u , v )

4.3. Experimental Settings

4.3.1. Data Pre-Processing

We used 70% of the training set for model training, while the remaining 30% was used to validate the performance of the model. To obtain more image scenes, we used image cropping for data enhancement. By cropping, we could obtain more LR and HR image patches with sizes of 64 × 64 and 192 × 192 pixels, respectively. To ensure the quality of the reconstructed images and to maintain a balance of computational effort during the training process, we chose to use only the scenes with an average reliable pixel count higher than 85% to participate in the model training.
We sorted the images in the dataset from high to low based on the number of clear pixels in the LR images. For the experimental setup, the network took k LR images as the input. If there were more images in the scene than k, we selected the top k LR images with the highest number of clear pixels as the input for the network. If the number of LR images in the scene was less than k, we filled the missing images with zero matrices to meet the input requirements.

4.3.2. Parameter Settings

In the experiment, the input number of LR images k accepted by our network was 32. The LR images are grayscale images with a channel number C0 of 1. The channels C for shallow feature FS, deep feature FD, and fusion feature FM were all set to 64. In the MAST network, we set the number of the MMAB and CAFB to 2 and 6, respectively. The number of heads in the multi-head attention of the Transformer model was set to 8, and the dimension of the hidden layer in the MLP was set to 128.
We conducted experiments on the Ubuntu 18.04 system. The hardware setup included four Nvidia RTX A5000 GPUs (with a total memory of 96 GB) and an Intel Xeon (R) Gold 5320 CPU @ 2.20 GHz. The entire code framework was implemented using Python 3.7 and PyTorch 1.7.1. In the experiments, we set the batch size to 8 based on the computational capacity of the devices. For the optimizer, we used Adam with an initial learning rate of 5 × 10−4. The learning rate decay strategy employed cosine decay, with a minimum learning rate of 5 × 10−6. During the training process, we trained the model for a total of 400 epochs, which took approximately 76 h. Subsequently, we fine-tuned the model with a learning rate of 1 × 10−5 and performed 80 epochs of fine-tuning training. The model had a parameter count of 3.3 million.

5. Discussion

5.1. Comparison with State-of-the-Art Methods

We compared MAST with other existing MISR methods of remote sensing. Since the PROBA-V dataset did not provide ground-truth for the test set, we evaluated the reconstruction performance of different methods on the validation set using cPSNR and cSSIM as performance metrics to assess the quality of the reconstructed images. Table 1 presents the quantitative results obtained using various methods on the NIR and RED bands.
In Figure 7, we present a visual comparison of the test results between MAST and the baseline method using Bicubic interpolation in two bands: NIR and RED. The results show that MAST exhibited excellent reconstruction performance in almost all scenarios. Specifically, in 98.23% of the NIR band, the reconstruction performance of the MAST method surpassed that of the baseline method. Similarly, in the RED band, our MAST method outperformed the baseline method in as many as 98.86% of the scenarios in terms of reconstruction performance.
From the image regions shown in Figure 8 and Figure 9, it is visually apparent that our method significantly improved the preservation of textures and fine details, resulting in better visual results. These results further validate the excellent performance of the MAST method in the super-resolution task of multiple remote sensing images.
We used Bicubic interpolation as the baseline method for super-resolution reconstruction. For each scene, we applied Bicubic interpolation to the clearest image to obtain an HR image. The red boxes in the figures indicate the reconstructed regions we are comparing.
  • Classical MISR Methods:
IBP obtains an initial image through Bicubic interpolation and performs registration using phase correlation. It gradually improves the estimation results by leveraging the image’s own information and reconstruction errors. BTV combines the characteristics of total variation regularization and bilateral filtering to solve the bilateral total variation minimization problem. This method can remove noise while preserving the edge information in the image.
2.
SISR Method:
RCAN learns the high-frequency details of images through residual learning and enhances the high-frequency features using the channel attention mechanism. The RCAN framework consists of five residual groups, each containing five residual channel attention blocks.
3.
Video Super-Resolution Method:
VSR-DUF utilizes dynamic upsampling filters to reconstruct HR images and enhances the image details by calculating residuals. We chose a base network consisting of 16 layers and used nine LR images as the input for the network.
4.
Deep Learning Super-Resolution Methods:
HighRes-net utilizes the reference frame for implicit alignment and recursively fuses multiple images. The experiments were conducted using the default framework, and training and testing were conducted with 16 LR images.
The 3DWDSRNet method utilizes 3D WDSR blocks instead of 3D convolutions to fuse multiple images, which helps to maintain lower GPU memory requirements. The experiments were conducted using seven LR images as the input.
DeepSUM introduces a dynamic registration subnetwork and utilizes 3D convolutions to fuse multiple images. The experiments were conducted using the clearest nine LR images to train the network. DeepSUM++ improves the non-local characteristics of the network by introducing graph convolutions.
MISR-GRU utilizes 24 LR images for implicit co-registration. It employs a fusion strategy based on ConvGRU and performs registration on super-resolution images using the ShiftNet network.
RAMS introduces an attention mechanism with 3D convolutions to better extract high-frequency features. The input consists of the clearest nine LR images. RAMS+20 improves upon the inference process by adopting a temporal self-ensemble strategy. It randomly permutes the input images 20 times and takes the average of the super-resolution results as the output.
PIUnet exhibits permutation invariance in the temporal dimension. However, it only performs well when the number of test images is close to the number of input images. Based on the parameter settings in the paper, we set the input to be nine LR images.
TR-MISR is currently the top-ranked algorithm in the leaderboard, and its network’s core is a fusion module based on a Transformer. In the experiments, we set the number of LR images to 32.
To address the issue of the insufficient utilization of complementary information from multiple frames in current MISR models, caused by a single scale and a single attention mechanism, this paper proposes an improved super-resolution network that combines multi-scale and multi-attention approaches. In terms of feature extraction, we introduce the Multi-Scale and Mixed Attention Block (MMAB) to enhance the extraction and utilization of complementary information from multiple images. Additionally, in the aspect of feature fusion, we propose the Collaborative Attention Fusion Block (CAFB) that simultaneously utilizes global and local information to achieve a balance between the restoration of the overall structure and the preservation of local details in the image.
Both the qualitative and quantitative results from the comparative experiments demonstrate that our method achieved state-of-the-art performance in the super-resolution reconstruction task on the PROBA-V dataset, further validating the effectiveness of our approach.

5.2. Analysis of Ablation Experiments

5.2.1. Effectiveness of Each Module

To validate the effectiveness of the two key modules, MMAB and CAFB, in the MAST network structure, we conducted a series of ablation experiments. We evaluated the contribution of each submodule to the overall performance, and the experimental results are recorded in Table 2 and Table 3. The results of the ablation experiments intuitively demonstrate the importance of each submodule and provide guidance for further improvement and optimization of the network.
  • Importance of MMAB
To evaluate the effectiveness of the MMAB in feature extraction, we conducted a series of ablation experiments with TR-MISR as the baseline. The feature extraction part of TR-MISR consists of two residual blocks. Apart from the differences in the feature extraction part, we maintained the same network structure as TR-MISR. In Table 2, we compare the cPSNR values of the models using three different versions of the feature extraction block on the validation set: (1) MSEB, (2) MSEB + SAGB, and (3) MSEB + SAGB + CAGB.
The multi-scale structure has a stronger ability to adapt to different scale feature information, allowing for more comprehensive feature extraction. The results in Table 2 show that when using only the MSEB for feature extraction, there was an average improvement of 0.2 dB on the validation set for both NIR and RED. The inclusion of attention mechanisms can guide the network to explore high-frequency information in the images and extract more complementary information from multiple images. The complete MMAB for feature extraction achieved cPSNR improvements of 0.52 dB and 0.55 dB on NIR and RED, respectively.
2.
Importance of CAFB
To evaluate the effectiveness of the CAFB, we conducted ablation experiments with TR-MISR as the baseline. Apart from the difference in the fusion module, we maintained the same network structure as TR-MISR. In Table 3, we compare the cPSNR values of the models using two different versions of the fusion module on the validation set: one with the Transformer + CA network and the other with the complete network (Transformer + CA + RLAB). Here, CA represents channel attention. RLAB represents the Residual Local Attention Block that we designed.
By incorporating channel attention, the ability of the Transformer model to capture more contextual features is effectively enhanced. Compared to the TR-MISR model that solely relies on the Transformer-based fusion block, our network achieved a surpassing improvement of 0.17 dB on both the NIR and RED validation sets when channel attention was added. Additionally, the inclusion of the local attention mechanism enables the model to focus more on high-frequency information. The complete fusion network showed a significant improvement in the model’s performance, with an average improvement of over 0.5 dB.

5.2.2. Analysis of the Effect of the Number of Input Images

In theory, as the number of input images increases, the MISR model can utilize more complementary information. However, it also introduces more noise and redundant information. The increase in noise can potentially degrade the quality of the reconstructed images. MISR utilizes complementary information between multiple images. If there is excessive redundant information in the network, it will increase the computational burden of the model and reduce the convergence speed of the network. The robustness of the model to noise has a significant impact on the quality of image reconstruction.
Table 4 presents the cPSNR comparison results of our proposed method on the validation set using different numbers of input images (without fine-tuning). When the number of input images k was 16, our method achieved higher cPSNR in both the NIR and RED bands compared to the other comparative methods. As the number of input images increased to 32, the reconstruction performance of the model reached its best, significantly outperforming the current state-of-the-art MISR method. Specifically, in the NIR band, our method improved cPSNR by 0.71 dB, while in the RED band, it showed a relative improvement of 0.69 dB in cPSNR.
Figure 10 illustrate specific regions of the super-resolution images under different input conditions. It is visually evident that as the number of input images increased, the reconstructed images exhibited richer details and texture information. The experimental results strongly demonstrate that our proposed method can better utilize the complementary information from multiple images and exhibit improved robustness in handling noise. Our method achieved superior reconstruction performance on the PROBA-V dataset.

5.2.3. Impact of the Number of Blocks

To investigate the impact of the number of the proposed two main blocks on the model performance, we conducted a series of comparative experiments by varying the number of MMABs and CAFBs (denoted by N and M, respectively) in the network structure based on our current hardware configuration. The quantitative results obtained on the validation set are recorded in Table 5.
The experimental results indicate that within a certain range, there was an improvement in the reconstruction performance as the number of blocks increased. It is noteworthy that the model achieved optimal reconstruction performance when N and M were set to 2 and 6, respectively. This implies that, in a fixed dataset size, appropriately increasing the number of modules can enhance the model’s capacity, leading to a better fitting of the input features and an improved performance in super-resolution reconstruction. The results further validate the effectiveness of the MMAB and CAFB modules.

5.2.4. Impact of the Size of the Training Images

The size of the input images often has an impact on the overall performance and speed of the model. To investigate the influence of image size in the training data on the super-resolution results, we conducted three sets of comparative experiments. Within the constraints of the computational conditions, we cropped the LR images in the training set to 64 × 64, 68 × 68, and 72 × 72, respectively. Correspondingly, the paired HR images were also cropped to three times the size of the corresponding LR images. The quantitative results of the validation set are reported in Table 6.
Larger images usually contain more details and textures, which can provide more realistic information for the super-resolution algorithm and help to generate more accurate high-resolution details. The computational complexity of the Vision Transformer is proportional to the square of the image size, and it requires more computational resources and time to deal with larger images. Therefore, it is particularly important to select the appropriate image size in the model training.

6. Conclusions

In this article, we proposed a Multi-Attention Multi-Image Super-Resolution Transformer (MAST), a novel multi-image super-resolution network of remote sensing, which is innovative in two aspects. Firstly, we introduced the MMAB for deep feature extraction. Through the design of a multi-scale architecture, we enhanced the network’s ability to capture features at various scales. Additionally, we employed a method of mixed attention to guide the network in exploring high-frequency features across both channel and spatial dimensions. Through these improvements, the network’s capability to extract complementary high-frequency information from multiple images was elevated. Secondly, we presented the CAFB for feature fusion. By incorporating channel attention within the Transformer architecture, the MISR model gained the ability to effectively harness correlation information among multiple images. Furthermore, the introduction of RLAB enhanced the network’s capacity to discern local intricate features. These combined enhancements empowered the network to achieve a harmonious equilibrium between the utilization of local and global information during fusion. On the PROBA-V dataset, MAST achieved cPSNR values of 49.63 dB and 51.48 dB for the NIR and RED bands, respectively, along with cSSIM values of 0.9884 and 0.9924, demonstrating state-of-the-art super-resolution performance. The extensive experiments validated the superiority of MAST compared to other advanced methods. In the future, our work will focus on establishing more accurate registration mechanisms to further improve the super-resolution results of multiple images by reducing registration errors.

Author Contributions

Conceptualization, J.L. and Z.T.; methodology, J.L.; software, J.L.; investigation, J.L., W.Z., B.Z., and G.Z.; writing—original draft preparation, J.L.; writing—review and editing, J.L. and Z.T.; project administration, Q.L.; funding acquisition, Z.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Program Project of Science and Technology Innovation of the Chinese Academy of Sciences (no. KGFZD-135-20-03-02) and by the Innovation Foundation of the Key Laboratory of Computational Optical Imaging Technology, CAS (no. CXJJ-23S016).

Data Availability Statement

The data presented in this study are available upon request from the corresponding authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hussain, S.; Lu, L.; Mubeen, M.; Nasim, W.; Karuppannan, S.; Fahad, S.; Tariq, A.; Mousa, B.; Mumtaz, F.; Aslam, M. Spatiotemporal variation in land use land cover in the response to local climate change using multispectral remote sensing data. Land 2022, 11, 595. [Google Scholar] [CrossRef]
  2. Ngo, T.D.; Bui, T.T.; Pham, T.M.; Thai, H.T.; Nguyen, G.L.; Nguyen, T.N. Image deconvolution for optical small satellite with deep learning and real-time GPU acceleration. J. Real-Time Image Process. 2021, 18, 1697–1710. [Google Scholar] [CrossRef]
  3. Wang, X.; Yi, J.; Guo, J.; Song, Y.; Lyu, J.; Xu, J.; Yan, W.; Zhao, J.; Cai, Q.; Min, H. A review of image super-resolution approaches based on deep learning and applications in remote sensing. Remote Sens. 2022, 14, 5423. [Google Scholar] [CrossRef]
  4. Jo, Y.; Oh, S.W.; Vajda, P.; Kim, S.J. Tackling the ill-posedness of super-resolution through adaptive target generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16236–16245. [Google Scholar]
  5. Harris, J. Diffraction and Resolving Power. J. Opt. Soc. Am. 1964, 54, 931. [Google Scholar] [CrossRef]
  6. Milanfar, P. Super-Resolution Imaging; CRC Press: Boca Raton, FL, USA, 2011. [Google Scholar]
  7. Wang, Z.; Chen, J.; Hoi, S.C. Deep learning for image super-resolution: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3365–3387. [Google Scholar] [CrossRef] [PubMed]
  8. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [PubMed]
  9. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  10. Mei, Y.; Fan, Y.; Zhou, Y. Image super-resolution with non-local sparse attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3517–3526. [Google Scholar]
  11. Chan, K.C.; Wang, X.; Xu, X.; Gu, J.; Loy, C.C. Glean: Generative latent bank for large-factor image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14245–14254. [Google Scholar]
  12. Wang, X.; Xie, L.; Dong, C.; Shan, Y. Real-esrgan: Training real-world blind super-resolution with pure synthetic data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 1905–1914. [Google Scholar]
  13. Shi, Y.; Han, L.; Han, L.; Chang, S.; Hu, T.; Dancey, D. A latent encoder coupled generative adversarial network (le-gan) for efficient hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–19. [Google Scholar] [CrossRef]
  14. Yang, F.; Yang, H.; Fu, J.; Lu, H.; Guo, B. Learning texture transformer network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5791–5800. [Google Scholar]
  15. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12299–12310. [Google Scholar]
  16. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  17. Chen, X.; Wang, X.; Zhou, J.; Qiao, Y.; Dong, C. Activating more pixels in image super-resolution transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 22367–22377. [Google Scholar]
  18. Zhang, J.; Xu, T.; Li, J.; Jiang, S.; Zhang, Y. Single-Image Super Resolution of Remote Sensing Images with Real-World Degradation Modeling. Remote Sens. 2022, 14, 2895. [Google Scholar] [CrossRef]
  19. Yue, L.; Shen, H.; Li, J.; Yuan, Q.; Zhang, H.; Zhang, L. Image super-resolution: The techniques, applications, and future. Signal Process. 2016, 128, 389–408. [Google Scholar] [CrossRef]
  20. Wronski, B.; Garcia-Dorado, I.; Ernst, M.; Kelly, D.; Krainin, M.; Liang, C.-K.; Levoy, M.; Milanfar, P. Handheld multi-frame super-resolution. ACM Trans. Graph. (ToG) 2019, 38, 1–18. [Google Scholar] [CrossRef]
  21. Tarasiewicz, T.; Nalepa, J.; Kawulok, M. A graph neural network for multiple-image super-resolution. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Virtual, 19–22 September 2021; pp. 1824–1828. [Google Scholar]
  22. Deudon, M.; Kalaitzis, A.; Goytom, I.; Arefin, M.R.; Lin, Z.; Sankaran, K.; Michalski, V.; Kahou, S.E.; Cornebise, J.; Bengio, Y. Highres-net: Recursive fusion for multi-frame super-resolution of satellite imagery. arXiv 2020, arXiv:2002.06460. [Google Scholar]
  23. Bhat, G.; Danelljan, M.; Timofte, R.; Cao, Y.; Cao, Y.; Chen, M.; Chen, X.; Cheng, S.; Dudhane, A.; Fan, H. NTIRE 2022 burst super-resolution challenge. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 1041–1061. [Google Scholar]
  24. Huang, B.; He, B.; Wu, L.; Guo, Z. Deep residual dual-attention network for super-resolution reconstruction of remote sensing images. Remote Sens. 2021, 13, 2784. [Google Scholar] [CrossRef]
  25. Jia, S.; Wang, Z.; Li, Q.; Jia, X.; Xu, M. Multiattention generative adversarial network for remote sensing image super-resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  26. Yang, Z.; Dan, T.; Yang, Y. Multi-temporal remote sensing image registration using deep convolutional features. IEEE Access 2018, 6, 38544–38555. [Google Scholar] [CrossRef]
  27. Qin, P.; Huang, H.; Tang, H.; Wang, J.; Liu, C. MUSTFN: A spatiotemporal fusion method for multi-scale and multi-sensor remote sensing images based on a convolutional neural network. Int. J. Appl. Earth Obs. Geoinf. 2022, 115, 103113. [Google Scholar] [CrossRef]
  28. Ye, Y.; Tang, T.; Zhu, B.; Yang, C.; Li, B.; Hao, S. A multiscale framework with unsupervised learning for remote sensing image registration. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  29. Qiao, B.; Xu, B.; Xie, Y.; Lin, Y.; Liu, Y.; Zuo, X. HMFT: Hyperspectral and Multispectral Image Fusion Super-Resolution Method Based on Efficient Transformer and Spatial-Spectral Attention Mechanism. Comput. Intell. Neurosci. 2023, 2023, 4725986. [Google Scholar] [CrossRef]
  30. Qiu, Z.; Shen, H.; Yue, L.; Zheng, G. Cross-sensor remote sensing imagery super-resolution via an edge-guided attention-based network. ISPRS J. Photogramm. Remote Sens. 2023, 199, 226–241. [Google Scholar] [CrossRef]
  31. TSAI, R. Multiframe Image Restoraition and Registration. Adv. Comput. Vis. Image Process. 1984, 1, 317–339. [Google Scholar]
  32. Guo, M.; Zhang, Z.; Liu, H.; Huang, Y. Ndsrgan: A novel dense generative adversarial network for real aerial imagery super-resolution reconstruction. Remote Sens. 2022, 14, 1574. [Google Scholar] [CrossRef]
  33. Bhat, G.; Danelljan, M.; Yu, F.; Van Gool, L.; Timofte, R. Deep reparametrization of multi-frame super-resolution and denoising. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 2460–2470. [Google Scholar]
  34. Luo, Z.; Yu, L.; Mo, X.; Li, Y.; Jia, L.; Fan, H.; Sun, J.; Liu, S. Ebsr: Feature enhanced burst super-resolution with deformable alignment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 471–478. [Google Scholar]
  35. Dudhane, A.; Zamir, S.W.; Khan, S.; Khan, F.S.; Yang, M.-H. Burst image restoration and enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 5759–5768. [Google Scholar]
  36. Luo, Z.; Li, Y.; Cheng, S.; Yu, L.; Wu, Q.; Wen, Z.; Fan, H.; Sun, J.; Liu, S. BSRT: Improving burst super-resolution with swin transformer and flow-guided deformable alignment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 998–1008. [Google Scholar]
  37. Dudhane, A.; Zamir, S.W.; Khan, S.; Khan, F.S.; Yang, M.-H. Burstormer: Burst Image Restoration and Enhancement Transformer. arXiv 2023, arXiv:2304.01194. [Google Scholar]
  38. Molini, A.B.; Valsesia, D.; Fracastoro, G.; Magli, E. Deepsum: Deep neural network for super-resolution of unregistered multitemporal images. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3644–3656. [Google Scholar] [CrossRef]
  39. Arefin, M.R.; Michalski, V.; St-Charles, P.-L.; Kalaitzis, A.; Kim, S.; Kahou, S.E.; Bengio, Y. Multi-image super-resolution for remote sensing using deep recurrent networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 206–207. [Google Scholar]
  40. Dorr, F. Satellite image multi-frame super resolution using 3D wide-activation neural networks. Remote Sens. 2020, 12, 3812. [Google Scholar] [CrossRef]
  41. Ghaffarian, S.; Valente, J.; Van Der Voort, M.; Tekinerdogan, B. Effect of attention mechanism in deep learning-based remote sensing image processing: A systematic literature review. Remote Sens. 2021, 13, 2965. [Google Scholar] [CrossRef]
  42. Lu, E.; Hu, X. Image super-resolution via channel attention and spatial attention. Appl. Intell. 2022, 52, 2260–2268. [Google Scholar] [CrossRef]
  43. Xia, B.; Hang, Y.; Tian, Y.; Yang, W.; Liao, Q.; Zhou, J. Efficient non-local contrastive attention for image super-resolution. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; pp. 2759–2767. [Google Scholar]
  44. Zhang, X.; Zeng, H.; Guo, S.; Zhang, L. Efficient long-range attention network for image super-resolution. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Part XVII. pp. 649–667. [Google Scholar]
  45. Salvetti, F.; Mazzia, V.; Khaliq, A.; Chiaberge, M. Multi-image super resolution of remotely sensed images using residual attention deep neural networks. Remote Sens. 2020, 12, 2207. [Google Scholar] [CrossRef]
  46. Valsesia, D.; Magli, E. Permutation invariance and uncertainty in multitemporal image super-resolution. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  47. Bhat, G.; Danelljan, M.; Van Gool, L.; Timofte, R. Deep burst super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9209–9218. [Google Scholar]
  48. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  49. Zhang, D.; Huang, F.; Liu, S.; Wang, X.; Jin, Z. SwinFIR: Revisiting the SWINIR with fast Fourier convolution and improved training for image super-resolution. arXiv 2022, arXiv:2208.11247. [Google Scholar]
  50. An, T.; Zhang, X.; Huo, C.; Xue, B.; Wang, L.; Pan, C. TR-MISR: Multiimage super-resolution based on feature fusion with transformers. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 1373–1388. [Google Scholar] [CrossRef]
  51. Li, K.; Wang, Y.; Zhang, J.; Gao, P.; Song, G.; Liu, Y.; Li, H.; Qiao, Y. Uniformer: Unifying convolution and self-attention for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 1–18. [Google Scholar] [CrossRef] [PubMed]
  52. Xiao, T.; Singh, M.; Mintun, E.; Darrell, T.; Dollár, P.; Girshick, R. Early convolutions help transformers see better. Adv. Neural Inf. Process. Syst. 2021, 34, 30392–30400. [Google Scholar]
  53. Fang, N.; Zhan, Z. High-resolution optical flow and frame-recurrent network for video super-resolution and deblurring. Neurocomputing 2022, 489, 128–138. [Google Scholar] [CrossRef]
  54. Lu, Z.; Li, J.; Liu, H.; Huang, C.; Zhang, L.; Zeng, T. Transformer for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 457–466. [Google Scholar]
  55. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  56. Recanatesi, S.; Farrell, M.; Advani, M.; Moore, T.; Lajoie, G.; Shea-Brown, E. Dimensionality compression and expansion in deep neural networks. arXiv 2019, arXiv:1906.00443. [Google Scholar]
  57. Zebari, R.; Abdulazeez, A.; Zeebaree, D.; Zebari, D.; Saeed, J. A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction. J. Appl. Sci. Technol. Trends 2020, 1, 56–70. [Google Scholar] [CrossRef]
  58. d’Ascoli, S.; Touvron, H.; Leavitt, M.L.; Morcos, A.S.; Biroli, G.; Sagun, L. Convit: Improving vision transformers with soft convolutional inductive biases. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 2286–2296. [Google Scholar]
  59. Han, K.; Xiao, A.; Wu, E.; Guo, J.; Xu, C.; Wang, Y. Transformer in transformer. Adv. Neural Inf. Process. Syst. 2021, 34, 15908–15919. [Google Scholar]
  60. Patel, K.; Bur, A.M.; Li, F.; Wang, G. Aggregating global features into local vision transformer. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 1141–1147. [Google Scholar]
  61. Yang, L.; Zhang, R.-Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 11863–11874. [Google Scholar]
  62. Märtens, M.; Izzo, D.; Krzic, A.; Cox, D. Super-resolution of PROBA-V images using convolutional neural networks. Astrodynamics 2019, 3, 387–402. [Google Scholar] [CrossRef]
  63. Farsiu, S.; Robinson, M.D.; Elad, M.; Milanfar, P. Fast and robust multiframe super resolution. IEEE Trans. Image Process. 2004, 13, 1327–1344. [Google Scholar] [CrossRef]
  64. Irani, M.; Peleg, S. Improving resolution by image registration. CVGIP Graph. Models Image Process. 1991, 53, 231–239. [Google Scholar] [CrossRef]
  65. Jo, Y.; Oh, S.W.; Kang, J.; Kim, S.J. Deep video super-resolution network using dynamic upsampling filters without explicit motion compensation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3224–3232. [Google Scholar]
  66. Molini, A.B.; Valsesia, D.; Fracastoro, G.; Magli, E. Deepsum++: Non-local deep neural network for super-resolution of unregistered multitemporal images. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 609–612. [Google Scholar]
Figure 1. The overall architecture of MAST. In this context, k indicates the number of LR images. N and M represent the number of proposed modules of MMAB and CAFB, respectively.
Figure 1. The overall architecture of MAST. In this context, k indicates the number of LR images. N and M represent the number of proposed modules of MMAB and CAFB, respectively.
Remotesensing 15 04183 g001
Figure 2. The structure of an MMAB.
Figure 2. The structure of an MMAB.
Remotesensing 15 04183 g002
Figure 3. The structure of the MSEB.
Figure 3. The structure of the MSEB.
Remotesensing 15 04183 g003
Figure 4. The structure of the CAGB.
Figure 4. The structure of the CAGB.
Remotesensing 15 04183 g004
Figure 5. The structure of the SAGB.
Figure 5. The structure of the SAGB.
Remotesensing 15 04183 g005
Figure 6. The structure of the CAFB.
Figure 6. The structure of the CAFB.
Remotesensing 15 04183 g006
Figure 7. Comparison of cPSNR for MAST and Bicubic on the validation set. Each point in the figure represents a scene. The point above the line y = x represents the scene in which the MAST reconstruction outperforms the baseline method.
Figure 7. Comparison of cPSNR for MAST and Bicubic on the validation set. Each point in the figure represents a scene. The point above the line y = x represents the scene in which the MAST reconstruction outperforms the baseline method.
Remotesensing 15 04183 g007
Figure 8. Comparison between different MISR methods on the imgset0198 scene of the RED band.
Figure 8. Comparison between different MISR methods on the imgset0198 scene of the RED band.
Remotesensing 15 04183 g008
Figure 9. Comparison between different MISR methods on the imgset1125 scene of the NIR band.
Figure 9. Comparison between different MISR methods on the imgset1125 scene of the NIR band.
Remotesensing 15 04183 g009
Figure 10. (a) The effect of different input images k on the imgset0454 scene of the RED band. The reconstruction outcome achieves the best result when k = 32. (b) The effect of different input images k on the imgset0870 scene of the NIR band. The reconstruction outcome achieves the best result when k = 32.
Figure 10. (a) The effect of different input images k on the imgset0454 scene of the RED band. The reconstruction outcome achieves the best result when k = 32. (b) The effect of different input images k on the imgset0870 scene of the NIR band. The reconstruction outcome achieves the best result when k = 32.
Remotesensing 15 04183 g010
Table 1. Quantitative comparison with different methods on the validation set.
Table 1. Quantitative comparison with different methods on the validation set.
BandNIRRED
MetriccPSNRcSSIMcPSNRcSSIM
Bicubic45.440.977047.330.9840
BTV [63]45.930.979448.120.9861
IBP [64]45.960.979648.210.9965
RCAN [9]45.660.979849.220.9870
VSR-DUF [65]47.200.985049.590.9902
HighRes-net [22]47.550.985549.750.9904
3DWDSRNet [40]47.580.985649.900.9908
DeepSUM [38]47.840.985850.000.9908
MISR-GRU [39]47.880.986150.110.9910
DeepSUM++ [66]47.930.986250.080.9912
RAMS [45]48.230.987550.170.9913
RAMS+20 [45]48.510.988050.440.9917
PIUnet [46]48.720.988350.620.9921
TR-MISR [50]48.540.988250.670.9921
MAST (this study)49.630.988451.480.9924
Table 2. Ablation results of individual components of MMAB on the validation set.
Table 2. Ablation results of individual components of MMAB on the validation set.
MMABNIRRED
MSEBSAGBCAGBcPSNRcSSIMcPSNRcSSIM
××48.930.986750.960.9916
×48.980.986950.990.9919
50.990.987951.220.9921
Table 3. Ablation results of individual components of CAFB on the validation set.
Table 3. Ablation results of individual components of CAFB on the validation set.
CAFBNIRRED
TransformerCARLABcPSNRcSSIMcPSNRcSSIM
×48.890.987450.840.9914
49.230.988151.190.9920
Table 4. Comparison of evaluation metrics on the validation dataset for different numbers of input LR images.
Table 4. Comparison of evaluation metrics on the validation dataset for different numbers of input LR images.
BandNIRRED
kcPSNRcSSIMcPSNRcSSIM
3249.430.988351.360.9922
2449.260.988451.270.9922
1648.870.987951.020.9920
848.410.986550.640.9908
Table 5. Comparison of evaluation metrics on the validation dataset for different numbers of modules.
Table 5. Comparison of evaluation metrics on the validation dataset for different numbers of modules.
BandNIRRED
NMMABNCAFBcPSNRcSSIMcPSNRcSSIM
2649.430.988351.360.9922
2449.210.987351.150.9918
2248.960.986850.820.9912
3449.300.987751.250.9923
1648.500.986250.350.9908
Table 6. Comparison of evaluation metrics on the validation dataset for different input image sizes.
Table 6. Comparison of evaluation metrics on the validation dataset for different input image sizes.
BandNIRRED
SizecPSNRcSSIMcPSNRcSSIM
6449.630.988451.480.9924
6849.680.989251.520.9925
7249.770.988951.540.9925
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Lv, Q.; Zhang, W.; Zhu, B.; Zhang, G.; Tan, Z. Multi-Attention Multi-Image Super-Resolution Transformer (MAST) for Remote Sensing. Remote Sens. 2023, 15, 4183. https://doi.org/10.3390/rs15174183

AMA Style

Li J, Lv Q, Zhang W, Zhu B, Zhang G, Tan Z. Multi-Attention Multi-Image Super-Resolution Transformer (MAST) for Remote Sensing. Remote Sensing. 2023; 15(17):4183. https://doi.org/10.3390/rs15174183

Chicago/Turabian Style

Li, Jiaao, Qunbo Lv, Wenjian Zhang, Baoyu Zhu, Guiyu Zhang, and Zheng Tan. 2023. "Multi-Attention Multi-Image Super-Resolution Transformer (MAST) for Remote Sensing" Remote Sensing 15, no. 17: 4183. https://doi.org/10.3390/rs15174183

APA Style

Li, J., Lv, Q., Zhang, W., Zhu, B., Zhang, G., & Tan, Z. (2023). Multi-Attention Multi-Image Super-Resolution Transformer (MAST) for Remote Sensing. Remote Sensing, 15(17), 4183. https://doi.org/10.3390/rs15174183

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop