Next Article in Journal
Complementary Local–Global Optimization for Few-Shot Object Detection in Remote Sensing
Previous Article in Journal
Image Characteristic-Guided Learning Method for Remote-Sensing Image Inpainting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mamba-STFM: A Mamba-Based Spatiotemporal Fusion Method for Remote Sensing Images

1
School of Computer Technology and Application, Qinghai University, Xining 810016, China
2
Qinghai Provincial Laboratory for Intelligent Computing and Application, Xining 810016, China
3
Qinghai Provincial Institute of Meteorological Sciences, Xining 810001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(13), 2135; https://doi.org/10.3390/rs17132135
Submission received: 12 May 2025 / Revised: 16 June 2025 / Accepted: 19 June 2025 / Published: 21 June 2025

Abstract

Spatiotemporal fusion techniques can generate remote sensing imagery with high spatial and temporal resolutions, thereby facilitating Earth observation. However, traditional methods are constrained by linear assumptions; generative adversarial networks suffer from mode collapse; convolutional neural networks struggle to capture global context; and Transformers are hard to scale due to quadratic computational complexity and high memory consumption. To address these challenges, this study introduces an end-to-end remote sensing image spatiotemporal fusion approach based on the Mamba architecture (Mamba-spatiotemporal fusion model, Mamba-STFM), marking the first application of Mamba in this domain and presenting a novel paradigm for spatiotemporal fusion model design. Mamba-STFM consists of a feature extraction encoder and a feature fusion decoder. At the core of the encoder is the visual state space-FuseCore-AttNet block (VSS-FCAN block), which deeply integrates linear complexity cross-scan global perception with a channel attention mechanism, significantly reducing quadratic-level computation and memory overhead while improving inference throughput through parallel scanning and kernel fusion techniques. The decoder’s core is the spatiotemporal mixture-of-experts fusion module (STF-MoE block), composed of our novel spatial expert and temporal expert modules. The spatial expert adaptively adjusts channel weights to optimize spatial feature representation, enabling precise alignment and fusion of multi-resolution images, while the temporal expert incorporates a temporal squeeze-and-excitation mechanism and selective state space model (SSM) techniques to efficiently capture short-range temporal dependencies, maintain linear sequence modeling complexity, and further enhance overall spatiotemporal fusion throughput. Extensive experiments on public datasets demonstrate that Mamba-STFM outperforms existing methods in fusion quality; ablation studies validate the effectiveness of each core module; and efficiency analyses and application comparisons further confirm the model’s superior performance.

1. Introduction

As a core technology for monitoring the Earth’s surface systems, remote sensing plays a fundamental role in a wide range of applications due to its broad coverage, rapid acquisition, and frequent observation capabilities. From precision agriculture [1] and dynamic environmental assessment [2,3,4] to rapid disaster response [5] and scientific urban planning [6], remote sensing data serve as a critical driving force. In these fields, the simultaneous acquisition of high-spatial resolution (HSR) and high-temporal resolution (HTR) remote sensing image time series is of paramount importance. However, existing single-platform remote sensing satellites generally face multiple limitations [7,8,9]. Constrained by cloud cover, orbital cycles, and inherent trade-offs in sensor technology, these platforms struggle to provide data with both high spatial and high temporal resolution simultaneously—an often unattainable compromise. For instance, the Landsat satellites offer spatial resolutions of approximately 25 m [10], which are sufficient for capturing surface details, yet their intrinsic 16-day revisit cycle limits their ability to meet high temporal resolution demands. In contrast, the MODIS sensor [11,12] provides near-daily or even more frequent global observations, but at the cost of spatial precision, with resolutions ranging from 250 to 1000 m.
In response to the challenge of acquiring high-spatial and high-temporal resolution data simultaneously, spatiotemporal fusion (STF) of remote sensing images has emerged [13] and evolved into a key research direction in the field of remote sensing data processing. The value of STF lies in its ability to overcome the limitations of single-sensor systems by integrating complementary data from multiple sensors [14,15]. Specifically, the technique typically uses reference pairs composed of low-spatial resolution/high-temporal resolution (LSR/HTR) images and high-spatial resolution/low-temporal resolution (HSR/LTR) images from known dates, along with a coarse image from the target prediction date, to generate a high-quality image sequence at the target date with both HSR and HTR.
After decades of exploration and development [13], spatiotemporal fusion (STF) algorithms have evolved into a diverse array of technical systems, each rooted in different forms of prior knowledge and model design philosophies. From the perspective of core principles and implementation strategies, existing algorithms can be broadly categorized into five main types: (1) weight-based methods [15,16]; (2) methods based on mixed-pixel decomposition [17,18]; (3) hybrid approaches [19]; (4) shallow learning-based methods; and (5) deep learning-based methods [20]. Among these, weight function-based methods represent some of the earliest and most widely applied techniques in the STF domain. Their key mechanism involves estimating the pixel values of the fine image on the target date by performing a weighted average of reference image pixels surrounding the target pixel. The weights are typically computed based on spatial proximity, spectral similarity, and temporal consistency. For example, the STARFM model [15], a representative algorithm in this category, assigns higher weights to spectrally and structurally similar pure pixels within a defined neighborhood. Although subsequent studies have proposed improvements in weighting strategies [21], cross-sensor adaptation [22], and method integration [23], weight-based methods still face significant challenges in improving fusion accuracy when applied to scenes with abrupt land cover changes or high spatial complexity due to their reliance on assumptions of local windows and linear relationships.
Mixed-pixel decomposition-based methods [24,25] are grounded in the understanding that low-resolution pixels are, in fact, mixtures of spectral signals from multiple high-resolution land surface components. These techniques employ spectral unmixing processes [17,18] to extract the fractional abundances of various land cover types from coarse-resolution images and then synthesize high-resolution images for the target date by incorporating spatial distributions or endmember spectral information derived from reference high-resolution imagery. While these methods offer advantages in handling sub-pixel-level land cover heterogeneity, their final accuracy is highly dependent on the adaptability of the spectral mixing model, the quality of the endmember library, and the ability to effectively capture and model temporal variations in land surface spectral properties.
To address the limitations of single-principle approaches, researchers have developed hybrid fusion strategies. These methods [19] aim to organically integrate the strengths of different techniques, including weight functions, mixed-pixel decomposition, and even change detection. In practice, FSDAF proposed by Zhu et al. [26] is a representative method that cleverly combines weight functions with mixed-pixel decomposition. Another example is CSAFM, employed by Wang et al. for evapotranspiration estimation [27], which also merges these two strategies. Furthermore, FSDAF 2.0, an improved version developed by Guo et al. [28], incorporates change detection and optimization techniques. These hybrid methods are typically more flexible in design and capable of delivering improved fusion performance. However, this often comes at the cost of increased model complexity, requiring more parameter tuning and reliance on additional assumptions.
To overcome the limitations of traditional spatiotemporal fusion algorithms—such as weight-based and mixed-pixel decomposition methods—which rely on fixed assumptions, shallow learning techniques from the domain of machine learning have been introduced as data-driven alternatives or enhancements. These approaches encompass several technical paths: Bayesian-based methods [29,30], which formulate fusion as a maximum a posteriori (MAP) inference problem by modeling temporal evolution and scale transformation; sparse representation–based methods [31], which exploit the sparsity of image patches by learning dictionaries and sparse coding relationships; and dictionary learning-based methods [32], which focus on learning paired dictionaries capable of capturing structural characteristics of images. Compared with traditional physics- or empirical model–based approaches, these shallow learning models demonstrate greater adaptability and offer moderate performance improvements. However, their relatively low model complexity and limited capacity for deep feature extraction often hinder their ability to manage complex spatiotemporal dynamics and highly heterogeneous regions. Nevertheless, their exploration of data-driven paradigms has laid a critical foundation for the subsequent application of more powerful deep learning methods in the STF domain.
Benefiting from their powerful nonlinear modeling capabilities and end-to-end learning frameworks, deep learning methods have achieved significant breakthroughs in the field of spatiotemporal fusion (STF) of remote sensing imagery, substantially raising performance ceilings and surpassing previous techniques. Deep learning models such as convolutional neural networks (CNNs), generative adversarial networks (GANs), and Transformers can autonomously learn complex nonlinear correspondences between images across different spatial and temporal scales. In CNN-based approaches, researchers have leveraged the spatial feature extraction strength of CNNs to design fusion models. Representative works include the early STFCNN [33], as well as more sophisticated architectures such as DCSTFN [34], EDCSTFN [20], the dual-stream CNN model STFNet [35], attention-augmented models [36], and BiaSTF [37], which addresses sensor biases. In parallel, GAN-based approaches [38,39,40,41] have also demonstrated outstanding performance on STF tasks. Notably, GAN-STFM [42] innovated the input paradigm by requiring only a coarse image from the target date and a fine-resolution image from any reference date to efficiently generate high-accuracy fused outputs. The Transformer architecture [43] has attracted increasing attention due to its unique self-attention mechanism, which enables direct modeling of dependencies among input elements, thereby facilitating effective global information representation. This capability is particularly critical for addressing large-scale spatial heterogeneity and complex temporal variations. Pioneering work by Liu et al. [44] introduced MSNet, which combines CNN and Transformer architectures for STF, effectively mitigating the local receptive field limitations of CNNs. This work laid the foundation for subsequent hybrid or Transformer-based fusion models such as SwinSTFM [45], CTSTFM [46], and EMSNet [47], offering crucial insights and architectural inspiration.
Despite the substantial progress achieved by deep learning–based spatiotemporal fusion (STF) methods, their practical application still faces several challenges. For CNN-based methods, a key limitation lies in their inherently local receptive field, which restricts the effective modeling of global and long-range spatial dependencies—often leading to performance bottlenecks in scenarios with large-scale surface heterogeneity or abrupt changes. Additionally, the shared weight design of convolutional layers may lack the flexibility to capture fine-grained, highly variant local patterns. GAN-based methods, on the other hand, suffer from issues related to computational efficiency and training stability. Performing generative operations directly in high-resolution space [42] is beneficial for preserving detail but comes with significant computational cost; dimensionality reduction strategies may alleviate this, but often at the expense of critical spatial information, requiring post-processing compensation. Furthermore, GAN training is notoriously unstable and prone to mode collapse [48,49], which poses significant challenges when fitting and generalizing to the complex and variable characteristics of remote sensing data. Transformer-based architectures also face distinct hurdles in STF tasks: the quadratic computational complexity of their self-attention mechanism leads to prohibitive computational and memory demands when processing high-resolution imagery; the absence of local inductive bias, as found in CNNs, hampers the accurate capture of fine textures and edges; and the need for large-scale labeled datasets conflicts with the scarcity of annotated samples in the remote sensing domain, limiting their training effectiveness. Moreover, in long-sequence modeling, self-attention mechanisms can introduce abrupt shifts and noise, necessitating the integration of temporal consistency modules to ensure stability.
To address the aforementioned limitations of existing architectures, this study introduces the Mamba architecture [50] into the domain of remote sensing spatiotemporal fusion and proposes an end-to-end method termed the Mamba-spatiotemporal fusion model (Mamba-STFM). The Mamba-STFM framework comprises a multi-branch feature extraction encoder and a multi-scale feature fusion decoder. Its core components include the visual state space–FuseCore–AttNet block (VSS–FCAN block), a spatial expert, and a temporal expert. The VSS-FCAN block is an enhanced version of the conventional visual state space (VSS) block, in which the standard Feed-Forward Network (FFN) is replaced with a FuseCore-AttNet block incorporating channel attention. The spatial expert and temporal expert are novel modules originally designed in this work. The key contributions of these core modules are as follows.
(1)
The VSS-FCAN block combines cross-scan global perception with channel attention mechanisms, significantly reducing computational and memory overhead at quadratic complexity. Additionally, it enhances model inference throughput through parallel scanning and kernel fusion techniques.
(2)
The spatial expert utilizes an enhanced 2D residual convolutional module and incorporates a channel attention mechanism, adaptively adjusting channel weights to optimize spatial feature representation, thereby achieving precise alignment and fusion of coarse and fine-resolution images.
(3)
The temporal expert introduces a 3D residual convolutional module for adjacent time steps, combining a temporal squeeze-and-excitation mechanism with selective state space model (SSM) techniques. This design efficiently captures short-range temporal dependencies while maintaining linear complexity in temporal sequence modeling, further improving the overall throughput of spatiotemporal fusion.
To comprehensively evaluate the performance of our proposed method, we compare it with several representative spatiotemporal fusion models, including both traditional and deep learning-based approaches. Among traditional methods, STARFM is the first model based on a weighting function, which predicts pixel values using spectral differences and spatial similarities, and ESTARFM improves upon it by introducing transformation coefficients that enhance the prediction of small and linear features, particularly in heterogeneous and seasonally dynamic landscapes. For deep learning-based baselines, GANSTFM is a typical GAN-based model that breaks the time constraint of reference image selection and uses switchable normalization to better capture temporal characteristics. EDCSTFN represents CNN-based models with a dual-branch structure and feature-level fusion, showing robustness to low-quality data and limited cloud coverage. Swin-STFM is the first model to apply the Swin Transformer to spatiotemporal fusion, combining spectral unmixing theory with attention-based feature extraction to enhance contextual understanding and generation quality. ECPW-STFM is the first to introduce wavelet transform into spatiotemporal image fusion, enabling separate learning of high- and low-frequency components and achieving a better balance between global structure and local detail. STFDiff is the first to employ a diffusion model for spatiotemporal fusion, progressively refining Gaussian noise into the target image through a dual-stream UNet and enhanced noise feature extraction. These baselines span classical traditional models, CNNs, GANs, Transformers, wavelet-enhanced methods, and diffusion models, providing a comprehensive benchmark to validate the effectiveness of our approach.
This Mamba-based method, equipped with a spatiotemporal expert mixture fusion module, not only offers linear computational complexity and high inference throughput but also achieves a significant performance breakthrough in spatiotemporal fusion tasks. Comparative experiments conducted on two widely used public remote sensing spatiotemporal fusion datasets demonstrate that the Mamba-STFM model outperforms existing state-of-the-art (SOTA) methods, achieving superior fusion performance.
The remainder of this paper is organized as follows: Section 2 introduces the overall architecture and core modules of Mamba-STFM; Section 3 describes the experimental design and process; Section 4 presents a series of comparisons with multiple methods on public datasets, ablation studies of Mamba-STFM, efficiency evaluations, and performance comparisons in practical applications; Section 5 discusses the results; Section 6 concludes the paper and discusses future directions.

2. Methodology

2.1. Overall Structure

The Mamba-STFM model proposed in this study is an end-to-end remote sensing spatiotemporal fusion model based on Mamba, designed to effectively integrate multi-source heterogeneous remote sensing data to generate synthesized images with high spatiotemporal resolution. The overall model adopts an encoder–decoder architecture, as shown in Figure 1, and primarily consists of a multi-branch feature extraction encoder and a multi-scale feature fusion decoder. The multi-branch feature extraction encoder is an improvement based on the Vision Mamba backbone network, while the multi-scale feature fusion decoder is composed of multiple STF-MoE blocks and VSS-FCAN blocks. The input to Mamba-STFM consists of a pair of coarse and fine images at the T0 time step and a coarse image at the T1 time step, while the output is a fine-resolution image at the T1 time step.
The multi-branch feature extraction encoder is used to process the input images and obtain multi-scale feature representations. Specifically, the input coarse image C 0 R B × C × H × W at the T0 time step, fine image F 0 R B × C × H × W at the T0 time step, and coarse image C 1 R B × C × H × W at the T1 time step are fed in parallel into three independent feature extraction branches, with each branch being constructed based on a hierarchical Vision Mamba structure. The input images are first processed through a patch embedding layer, which transforms them into feature tokens and expands their dimensionality. These feature tokens then pass through a series of VSS-FCAN blocks, which are adept at capturing both local and global dependencies, to perform transformation and information aggregation. During this process, spatial downsampling is applied to reduce the spatial resolution of feature maps while increasing the number of channels, thereby enabling multi-scale feature extraction. As a result, multi-scale feature lists E s o u r c e i are generated for C 0 , F 0 , and C 1 , respectively, where “source” denotes the image type and “i” indicates the i-th scale. Additionally, the differences between the coarse-resolution features at T1 and T0 are computed.
Δ E C 1 , C 0 i = E C 0 i E C 1 i , i 1 , 2 , 3 , 4 .
Finally, the encoder outputs multi-scale feature lists E C 0 i , Δ E C 1 , C 0 i , E F 0 i , and E C 1 i .
The multi-scale feature fusion decoder processes the multi-scale feature lists generated by the encoder. This component is responsible for integrating multi-source and multitemporal feature information and progressively reconstructing high-resolution imagery. The process can be conceptually divided into two stages: intra-scale feature fusion and hierarchical feature reconstruction. During the intra-scale feature fusion stage, for any given scale i, the input features E C 0 i , Δ E C 1 , C 0 i , E F 0 i , and E C 1 i are fed into a dedicated spatiotemporal mixture-of-experts fusion module (STF-MoE block). This module consists of a temporal expert pathway and a spatial expert pathway. The temporal expert handles the coarse-resolution features from T0 and T1, while the spatial expert focuses on the fine-resolution features from T0. The outputs of these experts are weighted using a dynamic gating mechanism generated based on feature E F 0 i . The model then computes the correspondence between the enhanced features and the original features and predicts the changes in fine resolution. Multiple prediction results are aggregated and refined through channel attention. Finally, residual connections are used to integrate the original information, generating the fused feature Y i at the corresponding scale. This process is applied in parallel across all scales, resulting in a multi-scale fused feature list Y 1 , Y 2 , Y 3 , Y 4 , where C represents the fused output at scale i.
Y i = ScaleFusionUnit i E C 0 i , Δ E C 1 , C 0 i , E F 0 i , E C 1 i , i { 1 , 2 , 3 , 4 } .
These fused features are then passed into the hierarchical feature reconstruction stage, which is responsible for restoring the feature resolution to match that of the original fine-resolution image. A bottom-up strategy is employed to integrate low-resolution contextual information with fine-grained details from the current scale.
X combined i = Y i + BilinearUpsample X u p i + 1 , i { 1 , 2 , 3 } .
Here, X u p i + 1 refers to the upsampled feature from scale i + 1. The combined feature X c o m b i n e d i is further enhanced and its spatial resolution expanded through a series of upsampling and feature refinement operations. This iterative process is performed progressively across scales until a feature map matching the spatial dimensions of the original fine-resolution image is obtained. The final feature map is normalized to generate the predicted fine-resolution image at time T1, denoted as F 1 . X u p 1 represents the highest-resolution feature prior to the final layer normalization step.
F 1 = OutputLayer FinalUpsample LayerNorm X u p 1

2.2. Visual State Space-FuseCore-AttNet Block

Because the Vision Transformer (ViT) successfully introduced the Transformer architecture into the field of computer vision, self-attention-based models have achieved remarkable progress in tasks such as image classification, object detection, and semantic segmentation, demonstrating strong global feature modeling capabilities. However, the standard ViT requires images to be partitioned into fixed-size patches and flattened into sequences, leading to quadratic computational complexity with respect to image size and limited ability to capture fine-grained local structures. To address these limitations, the Swin Transformer was proposed, which incorporates a shifted window mechanism that confines self-attention computations to local windows while enabling cross-window interactions through a window-shifting strategy. The Swin Transformer effectively reduces computational complexity, making it suitable for high-resolution image processing, and its hierarchical design allows for the extraction of multi-scale features, compensating for the lack of local inductive bias in ViT. Nonetheless, the Swin Transformer still relies on explicit attention matrix computation, which remains a bottleneck when handling very long sequences or scenarios that demand higher computational efficiency.
In recent years, state space models (SSMs) have demonstrated outstanding performance and computational efficiency in sequence modeling. In particular, the introduction of the Mamba architecture has enabled efficient linear-time sequence processing through a hardware-aware selective scan mechanism. Inspired by this advancement, Vision Mamba (VMamba) extends the advantages of SSMs to visual tasks, constructing a highly efficient visual backbone network. By leveraging the selective scan mechanism of SSMs, VMamba performs efficient information aggregation over flattened image sequences, effectively capturing long-range dependencies while avoiding the quadratic computational overhead typically associated with explicit attention. This architecture exhibits superior speed and performance potential compared with existing Transformer and Transformer-like models.
Leveraging the strengths of VMamba, this study designs a multi-branch feature extraction encoder that efficiently captures multi-scale features with both global and local contextual information from the input images. The encoder consists of three independent VMamba backbone branches, each responsible for processing inputs C 0 , F 0 , and C 1 , respectively. Each branch adopts a hierarchical design consisting of a patch embedding layer followed by a series of downsampling stages. The patch embedding layer transforms the input image into an initial feature sequence via convolution, simultaneously enhancing feature dimensionality and adjusting spatial structure. Each subsequent downsampling stage comprises multiple VSS-FCAN blocks and concludes with a downsampling layer (patch merging) to reduce spatial resolution while increasing channel dimensions. The VSS-FCAN block serves as the core computational unit of the encoder, as illustrated in Figure 1. It is an enhanced version of the standard VSS block, where the multilayer perceptron (MLP) feed-forward network is replaced with a convolutional feed-forward network integrated with channel attention (FuseCore-AttNet), better suited for 2D image processing. Internally, the VSS-FCAN block consists of a normalization layer, a feature transformation module based on selective scan (SS2D), and the FuseCore-AttNet. Specifically, the input feature X i n R B × C × H × W is first normalized through a preprocessing layer.
X n o r m 1 = N o r m a l i z e 1 X i n
Subsequently, the normalized feature X n o r m 1 is passed into the SS2D module. This module applies content-aware selective state space modeling (SSM) scanning mechanisms across multiple directions—such as horizontal, vertical, and their reverse counterparts—effectively capturing spatial dependencies within the two-dimensional domain and producing output feature Y S S M R B × C × H × W .
Y S S M = S S 2 D X n o r m 1
The VSS-FCAN block combines the output of the SS2D module with the input feature through the first residual connection and applies DropPath for stochastic regularization to enhance the model’s generalization capability.
Y i n t = X i n + D r o p P a t h Y S S M
Next, to introduce nonlinear transformations and spatial fusion, the intermediate features are preprocessed by a second normalization layer and then fed into the FuseCore-AttNet module. This module receives input X n o r m 2 , performs feature transformation, and finally adds the output Y F u s e C o r e A t t N e t to X n o r m 1 via a residual connection to produce the final output X o u t of the VSS-FCAN block.
X n o r m 2 = N o r m a l i z e 2 X i n t
X o u t = X n o r m 1 + D r o p P a t h Y F u s e C o r e A t t N e t

2.3. FuseCore-AttNet

In a standard VSS block, the feed-forward network (FFN) typically consists of two linear layers separated by a nonlinear activation function. This structure is highly efficient for processing flattened sequential data. However, for image features with two-dimensional spatial structures, purely linear layers fail to effectively capture the spatial correlations between neighboring pixels. To better adapt to the processing of image features and to introduce stronger local spatial awareness into the VSS block, we propose a convolutional feed-forward network, termed FuseCore-AttNet, to replace the standard FFN, as illustrated in Figure 2. The FuseCore-AttNet module takes a two-dimensional image feature tensor as the input. Its core idea is to replace the dimensionality expansion and reduction functions of linear layers with a series of convolutional operations. Moreover, it incorporates depth-wise separable convolutions and channel attention mechanisms to efficiently mix spatial features. Specifically, the input feature map X n o r m 2 is first passed through a 1 × 1 convolutional layer to increase the channel dimension. This operation expands the feature dimension from D to a hidden dimension D h i d d e n = D × m u l t without altering the spatial resolution. Then, a 3 × 3 depth-wise separable convolution followed by batch normalization is applied to the up-projected feature map X e x p . A GELU activation function is subsequently employed to introduce nonlinearity. This sequence forms the main pathway of the module, which is responsible for initial spatial feature mixing and nonlinear transformation.
X main _ act = GELU BN DWConv 3 × 3 X exp
A key improvement of FuseCore-AttNet lies in the introduction of two parallel branches designed to enhance the extraction of spatial features. In Branch 1, a standard 3 × 3 depth-wise separable convolution is employed to focus on capturing spatial information within local neighborhoods.
B 1 = GELU BN DWConv 3 × 3 std X main _ act
For Branch 2, a 3 × 3 depth-wise separable convolution with dilation is used. By employing dilated convolutions, this branch can expand the receptive field and capture a larger spatial context without increasing the number of parameters or computational cost.
B 2 = GELU BN DWConv 3 × 3 dilated X main _ act
The outputs of these two parallel branches are then concatenated along the channel dimension and subsequently passed through a 1 × 1 convolutional layer for channel dimension reduction to obtain X r e d u c e d . To further enhance the model’s channel-wise feature adaptivity, a squeeze-excitation (SE) attention module is applied to the reduced features. The SE module first compresses the spatial information of each channel into a scalar using global average pooling, then models the channel dependencies through two 1 × 1 convolutional layers (with ReLU activation and channel reduction in between), and finally generates channel-wise attention weights W s e R B × D × 1 × 1 through a Sigmoid activation. These weights are element-wise multiplied with X r e d u c e d , allowing for adaptive adjustment of the features along the channel dimension.
w s e = Sigmoid Conv 1 × 1 s e 2 ReLU Conv 1 × 1 s e 1 AdaptiveAvgPool2d X reduced
X main _ act = X reduced w se
Finally, Dropout is applied to the SE-weighted features X s e s c a l e d for regularization, and a residual connection is established between X s e s c a l e d and the original input features X n o r m 2 , yielding the final output of the FuseCore-AttNet module.
Y FuseCore - AttNet = X norm 2 + DropPath Dropout X main _ act

2.4. Spatiotemporal Mixture-of-Experts Fusion Block

The core functionality of the spatiotemporal mixture-of-experts fusion block (STF-MoE block) lies in the fine-grained interaction and integration of multi-source and multitemporal features at each independent scale, as shown in Figure 3A. For each scale i, a dedicated fusion unit (stf-union) receives four corresponding input feature sets: E C 0 i , Δ E C 1 , C 0 i , E F 0 i , and E C 1 i . To fully exploit different types of information, we design both spatial expert and temporal expert modules. The spatial expert operates on the fine-resolution feature E F 0 i at time T0, aiming to enhance and refine the fine-grained spatial details contained in F 0 . It achieves this through a 2D residual convolutional block integrated with a channel attention mechanism, which refines the spatial features of E F 0 i . By enhancing discriminative spatial representations, the spatial expert provides a high-quality spatial foundation for predicting fine-resolution information at time T1.
S P i = S p a t i a l E x p e r t E F 0 i
The temporal expert operates on the coarse-resolution features E C 0 i and E C 1 i at time points T0 and T1, respectively, aiming to capture and model the change patterns and dynamic information across time. Features E C 0 i and E C 1 i are stacked along a new temporal dimension to form a 5 D tensor, which is then fed into the temporal expert module. By analyzing this temporal cube, the temporal expert extracts change-sensitive features, providing crucial temporal context for predicting fine-resolution variations at T1.
T E i = TemporalExpert E C 0 i , E C 1 i stacked
To adaptively integrate the outputs of the spatial and temporal experts with the original fine-resolution information at T0, the model generates gating weights based on E F 0 i . Specifically, E F 0 i is passed through a 1 × 1 convolutional layer followed by a Sigmoid activation to produce the spatial gate G S i and temporal gate G T i . These two gates dynamically modulate the contributions of the expert outputs. The enhanced feature E is obtained by combining the original feature F with the gated expert outputs, where ⊙ denotes element-wise multiplication.
G S i , G T i = σ Conv1x 1 E F 0 i ( split into two halves )
X enhanced i = E F 0 i + G S i S P i + G T i T E i
Subsequently, the normalized output X n o r m i , together with E C 0 i , Δ E C 1 , C 0 i , and E C 1 i , is fed into the stf-union. The stf-union first computes the correlation feature c o r i between X n o r m i and E C 0 i , modeling the correspondence between the coarse- and fine-resolution features at time T0.
cor i = SiLU Conv X norm i E C 0 i
Then, c o r i is used to perform both change-based prediction and T1 coarse-resolution-based prediction. For change-based prediction, c o r i is used to modulate the coarse-resolution change feature Δ E C 1 , C 0 i to predict the fine-resolution change increment Δ p r e d i , resulting in an initial prediction P 1 i . For the T1 coarse-resolution-based prediction, c o r i is used to map the T1 coarse-resolution feature E C 1 i into the fine-resolution space, yielding the final prediction P 2 i .
Δ pred i = ConvDiff Δ E C 1 , C 0 i cor i
P 1 i = X norm i + Δ pred . i
P 2 i = E C 1 i cor i
Finally, the two prediction results are combined and further refined using channel attention to obtain P i . The refined output is then fused with the original features E F 0 i and E C 1 i through a residual connection, resulting in the final fused feature Y i .
Y i = ChannelAttentionBlock P i + E F 0 i + E C 1 i

2.5. Temporal Expert

One of the key challenges in spatiotemporal fusion lies in effectively capturing and leveraging changes across different time points, particularly under resolution mismatch. To specifically model the dynamic changes from time T0 to T1 at the coarse resolution, we design a temporal expert, as illustrated in Figure 3B. This expert operates on the T0 and T1 coarse-resolution features E C 0 i and E C 1 i at the same scale i, aiming to extract feature representations that are sensitive to temporal variations. The temporal expert incorporates 3D convolutional layers, batch normalization, and a nonlinear activation function (SiLU), with information propagated through a residual connection structure. The 3D convolutions enable simultaneous feature extraction and fusion across both temporal and spatial dimensions, thereby capturing interactions between temporal changes and spatial patterns. To further enhance the model’s sensitivity to the importance of features at different time steps, we embed a temporal SE (squeeze-and-excitation) attention mechanism within the temporal expert. Unlike the standard SE attention that performs compression and excitation along the channel dimension, the temporal SE attention focuses specifically on the temporal dimension. It first applies a pooling operation across the spatial dimensions of ( H i × W i ) to compress the spatial information of each channel at each time step into a scalar, forming a tensor X temporal i with shape R B × D i × 2 × 1 × 1 . This tensor is then fed into a lightweight 3 D convolutional network to learn importance weights for different combinations of time steps and channels. Finally, a Sigmoid activation is applied to constrain the weights between 0 and 1, resulting in the temporal attention weight tensor W t e m p o r a l .
W temporal = Sigmoid Conv 1 × 1 × 1 s e 2 ReLU Conv 1 × 1 × 1 s e 1 Adaptive AvgPool 3 d X temporal i
These temporal attention weights are applied element-wise to the feature maps produced by the main 3D convolutional layers, adaptively modulating the contributions of features at different time steps.
X temporal = 3 DConvBlock X temporal i
X temporal = X temporal W temporal
Finally, the temporal expert outputs the time-aware feature T E i for the given scale through a residual connection.
T E i = ProcessTimeDimension ResidualConnection X temporal i , X temporal

2.6. Spatial Expert

In multi-scale spatiotemporal fusion tasks, accurately capturing and preserving spatial details in fine-resolution imagery is critical. To enhance the model’s ability to process spatial information from the T0 fine-resolution image F 0 , we design a spatial expert for each scale, as illustrated in Figure 3C. This expert takes the fine-resolution feature E F 0 i at the corresponding scale of T0 as input, aiming to enhance discriminative spatial features and texture details through refined spatial processing. Internally, it consists of two 3 × 3 convolutional layers, batch normalization, a nonlinear activation function (SiLU), and a channel attention module for adaptive weighting of channel-wise features. The input feature C E F 0 i first passes through the first 3 × 3 convolutional layer, followed by batch normalization and SiLU activation.
X 1 = SiLU BN 1 Conv 3 × 3 1 E F 0 i
The activated feature X 1 is passed through a second 3 × 3 convolutional layer followed by batch normalization to produce X 2 . To enable the model to adaptively focus on the importance of different channels, we incorporate a channel attention mechanism into the main pathway. The channel attention module takes the feature map as input and first applies global average pooling and global max pooling to compress the spatial information of each channel into two scalars, which are then summed. This channel descriptor vector is then passed through a lightweight network—composed of two 1 × 1 convolutional layers with an intermediate SiLU nonlinear activation—to capture complex inter-channel dependencies. Finally, a Sigmoid activation function is applied to generate the channel-wise weights W C A , each ranging from 0 to 1. Here, C o n v 1 × 1 s e 1 performs channel reduction, while C o n v 1 × 1 s e 2 performs channel expansion.
W C A = σ Conv 1 × 1 s e 2 SiLU Conv 1 × 1 s e 1 AvgPool X 2 + MaxPool X 2
The learned channel weights W C A are applied to the feature map X 2 via element-wise multiplication to enhance the features of important channels, resulting in feature map X C A _ s c a l e d . Finally, to maintain smooth information flow and facilitate training, the attention-adjusted feature map X C A _ s c a l e d is combined with the original input feature E F 0 i through a residual connection. The resulting tensor is then passed through a final SiLU activation layer to produce the output S P i of the spatial expert at this scale.
X C A _ scaled = X 2 W C A
X residual = E F 0 i + X C A _ scaled
S P i = SiLU X residual

2.7. Loss Function

To effectively train the proposed Mamba-STFM model and enable it to generate synthetic images F 1 ^ that closely match the real high-resolution images in both pixel values and spatial structure, we employ a composite loss function L that combines pixel-level error and structural similarity. The primary optimization objective is to ensure that the pixel values of the predicted image F 1 ^ closely approximate those of the ground truth image F 1 . We adopt the mean squared error (MSE) as the pixel-level loss component L M S E . The M S E loss measures the average squared differences between the predicted output and the ground truth, effectively guiding model convergence and penalizing large pixel-wise deviations. Here, P and G T denote the predicted and ground truth images, respectively; C, H, and W represent the number of channels, height, and width; and P c , i , j and G T c , i , j indicate their pixel values at channel c, row i, and column j.
L M S E ( P , G T ) = 1 C · H · W c = 1 C i = 1 H j = 1 W P c , i , j G T c , i , j 2
Relying solely on pixel-level loss often fails to adequately capture the spatial structural information of images, which may result in blurred edges or distorted textures in the synthesized images. To guide the model in generating outputs that are structurally closer to real images, we introduce a loss term L SSIM based on the structural similarity index measure ( S S I M ). S S I M is a widely used metric for evaluating image quality and similarity, which assesses luminance, contrast, and structural information. Its values range from 0 to 1, with higher values indicating greater similarity between two images. We convert S S I M into a loss function by minimizing its complement, thereby maximizing structural similarity during training.
L S S I M P , G T = 1 S S I M P , G T
The final loss function L used to train the Mamba-STFM model is formulated as a linear combination of the M S E loss and the S S I M -based structural loss.
L = λ M S E L M S E Y ^ t 2 fine , Y t 2 fine + λ S S I M L S S I M Y ^ t 2 fine , Y t 2 fine
λ M S E and L M S E are non-negative weighting coefficients used to balance the contributions of the two loss components to the overall optimization objective. The model parameters are updated using standard gradient descent and backpropagation algorithms to minimize the combined loss function L .

3. Experiments

3.1. Study Area and Datasets

In the field of spatiotemporal fusion of remote sensing imagery, data availability is of critical importance. Among available datasets, the widely recognized Coleambally Irrigation Area (CIA) and Lower Gwydir Catchment (LGC) serve as foundational benchmarks for evaluating the performance of fusion algorithms. These datasets are valuable because they offer multiband remote sensing observation sequences across multiple time points, providing a standardized platform for researchers to assess the accuracy and robustness of spatiotemporal fusion algorithms in capturing and reconstructing various surface dynamics—ranging from gradual phenological changes to abrupt environmental events. Within the framework of this study, we employ these two datasets to conduct a detailed comparative evaluation of the proposed Mamba-STFM model against representative existing methods in terms of fusion performance.
Specifically, the CIA dataset [51] was designed to capture seasonal phenological cycles in agricultural regions. The dataset contains 17 carefully selected pairs of cloud-free Landsat and MODIS images, each covering an area of 2040 × 1720 pixels. The time series spans from 7 October 2001 to 3 May 2002, capturing surface conditions during that period. Geographically, the dataset corresponds to the Coleambally Irrigation Area in southern New South Wales, Australia.
In contrast to the CIA dataset, the LGC dataset [52] focuses on detecting and modeling rapid and drastic surface changes, with flood events serving as a typical example. This dataset comprises 14 pairs of cloud-free Landsat and MODIS images, with a larger spatial coverage of 3200 × 2720 pixels. It captures surface changes that occurred between 16 April 2004 and 3 April 2005. The study area is located in the Lower Gwydir Catchment in Northern New South Wales, Australia. Notably, during the time span of this dataset, a major flood event occurred in mid-December 2004, inundating approximately 44% of the area. This provides a realistic and challenging test scenario for evaluating algorithm performance under sudden environmental changes.

3.2. Experimental Details

To evaluate the performance of the proposed Mamba-STFM model, we conducted comparative experiments on the CIA and LGC datasets against seven representative spatiotemporal fusion methods. These methods include STARFM [15] and ESTARFM [16] (two classical traditional approaches), as well as GANSTFM [42], EDCSTFN [20], Swin-STFM [45], ECPW-STFN [53], and STFDiff [54] (all deep learning-based models). Each of these deep learning models has distinct characteristics: GANSTFM employs a conditional generative adversarial network, EDCSTFN is a deep model based on convolutional neural networks, Swin-STFM is built on the Transformer architecture, STFDiff performs fusion using a diffusion model, and ECPW-STFN is a wavelet-based pairing network.
Our experimental design focuses on four core components, all conducted on the publicly available CIA and LGC datasets. Prior to formal evaluation, the datasets were partitioned into training and testing sets. For the CIA dataset, the training period spans from 8 October 2001 to 11 April 2002 and the testing period from April 18 to 4 May 2002, with the prediction target date set as 27 April 2002. For the LGC dataset, the training set includes data from 16 April to 25 October 2004 and 13 January to 3 April 2005. The testing set consists of data from 26 November to 28 December 2004, with the prediction target date being 12 December 2004. Once the data preparation was complete, we carried out the following experiments: (1) quantitative performance comparisons of various fusion methods on the test sets; (2) qualitative analysis of the fused image details produced by each method; (3) ablation studies to evaluate the contribution of key components within the Mamba-STFM model; and (4) comparative evaluation of the classification performance of different models using the K-means clustering algorithm.
The proposed Mamba-STFM fusion method was implemented using Python 3.9.21 and PyTorch 1.13.1. The encoder comprises four stages, with the number of VSS-FCAN blocks set to two, two, nine, and two for each respective stage. The dimensionality parameters for downsampling are set to 96, 192, 384, and 768, respectively. The spatiotemporal fusion module operates across four scale levels, each containing one STF-Union unit. The MambaDecoder in the decoder stage has depths configured as one, one, one, and one.
The model is trained using the Adaptive Moment Estimation (ADAM) optimizer with an initial learning rate of 1 × 10 4 , a batch size of 4, a patch size of 256, and for 50 training epochs. A dynamic learning rate adjustment strategy is employed: if the training loss does not decrease for four consecutive epochs, the learning rate is automatically reduced by a factor of 0.5. Sensitivity analysis shows that the model is relatively robust to batch size variations, and its performance remains stable across batch sizes ranging from 2 to 16. The training process does not exhibit any oscillation or divergence, and the loss curves demonstrate smooth and consistent convergence.
The parameter settings for other spatiotemporal fusion methods follow the original authors’ configurations to ensure fair comparison. All baseline models were retrained using the same training, validation, and test splits as Mamba-STFM. All experiments were conducted on a server equipped with an NVIDIA RTX 3090 GPU (24 GB), an Intel Platinum 8362 CPU, and 45 GB of RAM.

3.3. Assessment of Metrics

To comprehensively evaluate the performance of the fused images, five key metrics are employed in this study: root mean square error (RMSE), structural similarity index (SSIM), universal image quality index (UIQI), correlation coefficient (CC), and Erreur Relative Globale Adimensionnelle de Synthèse (ERGAS). These metrics assess fusion quality from multiple perspectives: RMSE quantifies pixel-wise deviations, where lower values indicate higher accuracy; SSIM evaluates structural fidelity, with values ranging from 0 to 1, where values closer to 1 denote greater structural similarity; UIQI assesses image quality by analyzing distortion in correlation, luminance, and contrast, ranging from −1 to 1, with 1 indicating perfect statistical consistency; CC reflects the strength of linear correlation between images, also ranging from −1 to 1, where 1 indicates perfect positive correlation; ERGAS offers a band-wise global relative error assessment, where lower values signify better overall fusion performance. Details of each metric are provided in Table 1.

4. Results

4.1. Comparison of Various Fusion Methods on the CIA Dataset

4.1.1. Quantitative Comparison

Based on the CIA dataset, we conducted a comprehensive quantitative evaluation of eight spatiotemporal fusion models—STARFM, ESTARFM, GANSTFM, EDCSTFN, SwinSTFM, ECPW-STFN, STFDiff, and Mamba-STFM—using key metrics widely adopted in the remote sensing image fusion field, including RMSE, SSIM, UIQI, CC, and ERGAS. The evaluation results in Table 2 indicate that the Mamba-STFM model achieves state-of-the-art performance. As a global indicator of overall spectral and spatial quality (where lower values indicate better performance), Mamba-STFM yielded the lowest ERGAS value of 0.8615. This represents a reduction of 0.0534 compared with the second best model, STFDiff (0.9149), and a substantial decrease of 0.8471 compared with the weakest traditional model, STARFM (1.7086), highlighting a significant overall improvement. Furthermore, Mamba-STFM consistently demonstrated superior performance across individual spectral bands, achieving the lowest RMSE and the highest SSIM, UIQI, and CC values in the vast majority of bands. Notably, in terms of the UIQI metric, which evaluates overall image quality, Mamba-STFM attained the highest values across all six spectral bands, underscoring its strong capability in preserving image fidelity. These comprehensive quantitative results strongly support the superior performance and leading position of the Mamba-STFM model in spatiotemporal image fusion on this dataset.

4.1.2. Detail Comparison

To provide a more intuitive assessment of the fusion performance of different methods, we compared the details of various fusion products on the CIA dataset. As shown in Figure 4 (27 April 2002, CIA dataset agricultural area), significant differences are observed among the methods in terms of micro-texture and field boundary restoration: The STARFM fusion results are overly smoothed, with blurred small-scale field boundaries and discontinuous spectral transitions; ESTARFM insufficiently reconstructs fragmented structures, with noticeable edge diffusion. While GANSTFM introduces some texture details, it is accompanied by artifact stripes and spectral distortions; EDCSTFN and ECPW-STFN show limited improvement in overall texture, with residual noise interference and block effects; SwinSTFM results exhibit a noticeable grid-like block structure, with unclear distribution of multi-scale objects; STFDiff has limited high-frequency detail restoration, with significant boundary noise. In contrast, Mamba-STFM performs best in maintaining field geometry and color contrast, accurately reconstructing small-scale vegetation block distributions, preserving sharp boundaries, minimizing spectral bias, and demonstrating the highest structural similarity and optimal fusion details.
As shown in Figure 5 (27 April 2002, CIA dataset), there are significant differences in the pixel-level RMSE distributions of the various methods: GANSTFM produces large “high-error areas” at field boundaries and mixed crop interfaces, with pixel RMSE frequently exceeding 0.06; STARFM and ESTARFM reduce errors in homogeneous areas but are accompanied by scattered high-error patches (RMSE > 0.05), corresponding to pseudo-textures; EDCSTFN and ECPW-STFN suppress overall errors compared with STARFM but still leave relatively high RMSEs in localized regions due to blocky artifacts. While SwinSTFM and STFDiff better capture fine structures and reduce large-scale hotspots, they show increased errors at boundaries, with RMSE distribution tails extending; in contrast, Mamba-STFM exhibits the most uniform error field, with the majority of pixel RMSE values below 0.05 and minimal high errors, providing strong evidence of its ability to achieve optimal fusion with minimal spectral distortion when reconstructing both homogeneous and heterogeneous agricultural features.

4.2. Comparison of Various Fusion Methods on the LGC Dataset

4.2.1. Quantitative Comparison

A quantitative evaluation on the LGC dataset also compared eight spatiotemporal fusion models, including Mamba-STFM. The analysis results, as shown in Table 3, provide a clear performance overview: the Mamba-STFM model leads comprehensively among all evaluated models, ranking first across all performance metrics. Specifically, Mamba-STFM achieved the lowest RMSE values across all six spectral bands in the reconstruction error dimension; simultaneously, in terms of image quality, structural similarity, and correlation, as reflected by SSIM, UIQI, and CC, it also set new records for the highest values in each band. Particularly crucial is the ERGAS metric, which provides a global performance evaluation (where lower values indicate better performance). Mamba-STFM achieved the lowest score of 1.2472, securing its outstanding position. This result not only outperforms the second best STFDiff model (1.4260) by 0.1788 points but also shows a significant improvement of 0.8845 compared with the lowest-performing GANSTFM model (2.1317). These overwhelming quantitative results collectively establish Mamba-STFM’s state-of-the-art performance on the LGC dataset.

4.2.2. Detail Comparison

Figure 6 presents a comparison of the details after spatiotemporal fusion of the flood area on the LGC dataset from December 12, 2004, using various fusion methods. Significant differences are observed in the sharpness of water boundaries and the reproduction of details at object conflicts. The STARFM results generally exhibit over-smoothing, with blurred flood edges and unnatural spectral jumps at the land–water interface; ESTARFM underestimates the spatial expansion of newly formed water channels and flooded patches, resulting in a contracted water body range. While GANSTFM enhances the texture of localized water bodies, it introduces strip artifacts and spectral shifts, which hinder accurate flood area identification; EDCSTFN and ECPW-STFN show slight improvements in detail recovery over STARFM, but residual noise and blocky artifacts still impede the resolution of small-scale tributary networks; SwinSTFM exhibits noticeable grid-like textures, with discontinuous breaks in the flood distribution; STFDiff, while introducing some high-frequency information, still shows a blurred land–water boundary and spectral deviations. In contrast, Mamba-STFM best preserves the flood geometry and spatial details, clearly depicting the floodplain and small-scale tributary systems, while achieving smooth transitions and spectral consistency at the land–water interface. It also exhibits the highest overall structural similarity index.
Figure 7 presents the pixel-level RMSE error distribution of fusion results from various spatiotemporal models on the LGC dataset from 12 December 2004 (representing a land cover change scenario caused by flooding). The color intensity in the error map visually reflects the difference between the fusion results and the true image. As seen in the figure, models such as STARFM, ESTARFM, GANSTFM, EDCSTFN, SwinSTFM, and ECPW-STFN generally exhibit high fusion errors in regions with significant land cover changes (such as the darker areas in the lower left and center of the image). In contrast, the STFDiff model shows reduced errors in these change areas, while the error map generated by the Mamba-STFM model has the lightest overall color and the most uniform distribution, particularly with significantly lower errors in the flood area compared with all other models. This indicates that the Mamba-STFM model demonstrates the best performance in handling the complex land cover changes represented by the LGC dataset (such as flooding). It more effectively captures and predicts surface changes, producing more precise fusion results and significantly reducing pixel-level errors.

4.3. Comparison of the Efficiency of Various Spatiotemporal Fusion Methods

To evaluate the performance of various deep learning-based spatiotemporal fusion models for remote sensing images, this study focuses on comparing the computational efficiency and fusion accuracy of six models based on different network architectures. The evaluation metrics include computational complexity (G FLOPs), number of parameters (M parameters), average inference time (ms), max memory usage (MB), and the average correlation coefficient (CC) for each channel across two different datasets. All model evaluations were conducted on a server equipped with an RTX 3090 GPU, with the input image size fixed at 256 × 256 and the number of channels set to six.
As shown in Table 4, the six models exhibit notable differences in performance. The GAN-based GANSTFM model (evaluated only with its generator for fairness) shows the highest computational complexity (37.7586 G FLOPs) and relatively fewer parameters (0.5771 M parameters), but delivers the worst fusion accuracy. The deep CNN-based EDCSTFN model is the most lightweight, with only 18.5838 G FLOPs and 0.2825 M parameters. However, its spatiotemporal fusion performance remains unsatisfactory, with average CC values of 0.8793 and 0.7531 on the two datasets, only slightly better than GANSTFM. The Transformer-based SwinSTFM achieves moderate accuracy (0.8931 and 0.8621) with 28.1822 G FLOPs and 37.4656 M parameters. ECPW-STFN (wavelet-based) and STFDiff (diffusion-based) both achieve competitive fusion performance, ranking just below Mamba-STFM. Notably, ECPW-STFN has the second highest computational complexity (30.7110 G FLOPs), and STFDiff has the second largest parameter count (42.8226 M parameters).
In contrast, the Mamba-STFM model achieves the best fusion accuracy, with average CC values of 0.9277 and 0.9046, while maintaining relatively low computational complexity. Importantly, it also demonstrates excellent inference efficiency—with an average inference time of only 44.8670 ms and maximum memory usage of 288.7332 MB—substantially outperforming STFDiff (236.1877 ms, 733.5072 MB) and ECPW-STFN (120.3042 ms, 605.1192 MB) and even showing better memory efficiency than SwinSTFM (311.1792 MB). These results confirm that Mamba-STFM provides an outstanding balance between fusion accuracy and computational efficiency, making it highly suitable for real-time and resource-constrained remote sensing applications.

4.4. Ablation Study

To investigate the importance of the three modules—FuseCore-AttNet, temporal expert, and spatial expert—in the Mamba-STFM model, this study conducted an ablation experiment. The ablation experiment used the LGC dataset, with evaluation metrics including SSIM, UIQI, and CC (higher values typically indicate better quality), as well as RMSE and ERGAS (lower values indicate smaller errors and better quality), to quantify the model’s fusion performance under different configurations. The row marked with ∖ represents the complete Mamba-STFM model configuration, which includes all the designed core modules. The remaining rows show the performance changes after removing or replacing specific core modules. Specifically, the row marked as “ALL” represents a control configuration where all custom core modules of the model are replaced with the original Mamba structure. As shown in the experimental results in Table 5, the complete Mamba-STFM model (∖ row) achieved the best performance across all evaluation metrics. Removing or replacing any individual core module led to a decline in model performance, which proves the positive contribution of the custom modules—temporal expert, spatial expert, and FuseCore-AttNet—in enhancing spatiotemporal fusion performance. Particularly notable is the case when all core modules were replaced with the original Mamba structure (the “ALL” row), where the model performance showed the most severe degradation, performing the worst on all metrics. This strongly indicates the importance and effectiveness of the custom-designed Mamba-STFM core modules for spatiotemporal fusion tasks, as compared with the generic Mamba structure.

4.5. Application Comparison

4.5.1. Comparison of Clustering Results in the CIA Study Area

Figure 8 presents the land cover classification results derived from unsupervised K-means clustering applied to the fused images generated by various spatiotemporal fusion models (including STARFM, GANSTFM, EDCSTFN, SwinSTFM, ECPW-STFN, STFDiff, and the proposed Mamba-STFM), as well as the ground truth image, within the CIA study area focused on cropland phenology. By comparing the classification results with the ground truth, we can evaluate each model’s ability to capture surface complexity and phenological variation. As shown in the figure, the classification results of the traditional method STARFM exhibit significant discrepancies from the ground truth in terms of spatial details and class boundaries, manifesting as blurred edges, fragmented features, and misclassifications. Among deep learning methods, GANSTFM and EDCSTFN demonstrate noticeable improvements over STARFM, yet still suffer from classification errors, blurring, and spectral distortions. SwinSTFM, ECPW-STFN, and STFDiff further improve classification performance to a certain extent, capturing more spatial detail. Notably, the classification results based on the fused image generated by the proposed Mamba-STFM method show the highest consistency with the ground truth. Mamba-STFM not only accurately identifies and distinguishes between various land cover types (e.g., cropland, vegetation, bare soil), but also preserves parcel boundaries and internal texture structures with high fidelity. In particular, within the marked regions of interest (highlighted in red, yellow, and blue boxes), the classification results of Mamba-STFM are nearly identical to the ground truth in terms of detail, shape, and spatial layout, significantly outperforming all other methods. These results strongly demonstrate that Mamba-STFM can generate high-quality spatiotemporal fused images, effectively capturing complex cropland phenological dynamics, thereby laying a solid foundation for accurate land cover classification and change detection. The classification results provide intuitive and compelling evidence of the superior performance of Mamba-STFM in spatiotemporal fusion tasks.
Figure 9 illustrates the pixel-level scatter distributions of RED (red dots) and NIR (blue dots) bands between the fused images generated by different spatiotemporal fusion methods and the corresponding ground truth images in the CIA study area. The x-axis represents pixel values from the fused images, and the y-axis shows values from the ground truth. The black dashed line indicates perfect agreement (1:1 correspondence). These scatter plots provide a visual assessment of each method’s ability to reproduce actual surface reflectance at the pixel level, which is especially critical for vegetation index calculations such as NDVI that rely on RED and NIR bands. As shown in the figure, the scatter distributions from STARFM, ESTARFM, GANSTFM, and EDCSTFN are relatively dispersed, with point clouds deviating from the 1:1 line, particularly in high-value regions. This indicates that these methods exhibit substantial errors and uncertainty in predicting surface reflectance and fail to fully capture complex phenological dynamics and surface details. Some deep learning-based methods, such as SwinSTFM, ECPW-STFN, and STFDiff, partially improve the clustering of scatter points, yet still display noticeable outliers and deviations from the 1:1 line. In contrast, the scatter plots generated by the proposed Mamba-STFM model demonstrate superior performance. The scatter points for both RED and NIR bands cluster closely around the 1:1 diagonal line with the lowest degree of dispersion, although some outliers remain. These results strongly indicate that the Mamba-STFM model can predict and reconstruct surface reflectance in the RED and NIR bands with high accuracy, achieving excellent pixel-level consistency with the ground truth. This high-precision spectral reconstruction capability provides a reliable data foundation for remote sensing applications that depend on accurate spectral information, such as precision agriculture monitoring and land cover classification.

4.5.2. Comparison of Clustering Results in the LGC Study Area

Figure 10 presents the land cover classification results obtained by applying unsupervised K-means clustering to the fused images generated by various spatiotemporal fusion models, alongside the ground truth in the LGC study area, which focuses on flood-induced land cover conflicts. Compared with the ground truth, STARFM exhibits pronounced block effects along complex land cover boundaries, with poor identification of small patches and significant boundary misalignment. Although GANSTFM improves global structure reconstruction to some extent, its classification results are overly smoothed, significantly weakening the micro-scale spatial heterogeneity in flood-affected conflict zones. EDCSTFN demonstrates good alignment with the large-scale distribution of vegetation and bare land but fails to represent small-scale features in conflict zones, leading to frequent local misclassification. SwinSTFM enhances the depiction of patch edge details through window-based self-attention mechanisms, while ECPW-STFN leverages hierarchical convolutions and a progressively weighted multi-scale context aggregation strategy to adaptively assign weights to features of different resolutions. Nevertheless, both methods struggle to fully reconstruct the fragmented spatial patterns induced by subtle floodplain topographies. STFDiff improves texture retention in high-contrast regions, but the overall category distribution still deviates from the actual conditions. In contrast, Mamba-STFM achieves an optimal balance between multi-scale feature fusion and spatiotemporal interaction modeling, accurately reproducing patch shapes and spatial distribution, but also retaining the highest level of micro-scale heterogeneity in flood-induced conflict zones, significantly improving classification accuracy and boundary resolution.
Figure 11 illustrates the scatter distributions of pixel values in the RED (red points) and NIR (blue points) bands between the fused images generated by various spatiotemporal fusion methods and the corresponding ground truth in the LGC study area. The overall trends reveal significant differences in the ability of each method to reproduce pixel-level surface reflectance. Specifically, STARFM and ESTARFM tend to underestimate reflectance in the medium-to-high range, with scattered points widely deviating from the 1:1 reference line. GANSTFM shows some improvement in the high reflectance range but exhibits substantial dispersion in the low reflectance region. EDCSTFN and ECPW-STFN display signs of dynamic range compression; their scatter points are more aligned along the diagonal but generally shift toward the mid-to-low value region. In contrast, SwinSTFM, STFDiff, and Mamba-STFM produce scatter distributions that closely align with the 1:1 reference line, especially at the high reflectance end, where deviations are minimal. This indicates that these three methods exhibit superior performance in reproducing pixel-level surface reflectance. Such accuracy contributes directly to improving the precision of vegetation indices (e.g., NDVI) derived from the RED and NIR bands.

5. Discussion

In this study, we introduce for the first time the Mamba architecture into the spatiotemporal fusion of remote sensing imagery, proposing the Mamba-STFM model. By integrating a linear complexity visual state space (VSS) module and a FuseCore-AttNet attention block within the encoder, alongside a dual “spatial expert” and “temporal expert” mixture-of-experts structure in the decoder, our design efficiently captures both global and local features. This approach overcomes the quadratic computational bottleneck of conventional Transformers while preserving fine-grained spatial details and modeling short-range temporal dependencies, thereby establishing a novel paradigm for spatiotemporal fusion.
Extensive experiments on two public benchmarks (CIA and LGC) demonstrate that Mamba-STFM outperforms seven state-of-the-art methods— including STARFM, GAN-STFM, Swin-STFM, and the recent diffusion-based STFDiff—across all evaluation metrics. For instance, on the CIA dataset, Mamba-STFM achieves the lowest ERGAS (0.8615) alongside the highest SSIM, UIQI, and CC scores; on the LGC dataset, it records an ERGAS of 1.2472, substantially better than the runner-up’s 1.4260. These results confirm the model’s superiority in maintaining spatial fidelity and spectral consistency.
Compared with alternative deep-learning approaches, our VSS-FCAN module leverages parallel scanning and kernel fusion to deliver linear-time global receptive fields, significantly reducing both computational load and memory usage. On an RTX 3090 GPU, Mamba-STFM achieves a 20–30% increase in inference throughput versus traditional GAN- and Transformer-based models while keeping parameter counts and FLOPs at practical levels, underscoring its promise for industrial deployment.
The proposed Mamba-STFM model not only achieves high-quality spatiotemporal fusion but also demonstrates strong potential for real-world applications such as change detection, crop monitoring, and land cover classification. By effectively capturing both spatial and temporal dependencies, it produces semantically meaningful features that support accurate identification of land use changes, vegetation dynamics, and thematic mapping. Preliminary results show that the fused outputs align well with known land cover types, indicating that Mamba-STFM is a scalable and practical tool for downstream Earth observation tasks.
Despite its strong performance, Mamba-STFM generalization under challenging conditions—such as cloud cover, high-latitude polar regions, or abrupt weather changes—remains to be thoroughly evaluated. Furthermore, the efficiency of distributed training across multiple GPUs with very high-resolution inputs warrants further optimization.

6. Conclusions

This study is the first to introduce the Mamba architecture into the domain of remote sensing image spatiotemporal fusion, proposing an end-to-end method named Mamba-STFM. This novel approach expands the design paradigm and application scope of fusion models, providing a new framework for the efficient processing of high-resolution and multitemporal remote sensing data.
The proposed Mamba-STFM consists of three core modules: the VSS-FCAN block, spatial expert, and temporal expert. The VSS-FCAN block significantly reduces quadratic computational complexity and memory overhead while achieving higher inference throughput. The spatial expert adopts an enhanced 2D residual convolution module with channel attention to adaptively recalibrate feature channel weights, thus optimizing spatial feature representation. The temporal expert employs a 3D residual convolution module across adjacent time steps to efficiently capture short-range temporal dependencies with linear sequential modeling complexity.
Comprehensive experiments conducted on the publicly available CIA and LGC datasets demonstrate that Mamba-STFM consistently outperforms existing fusion methods. Quantitatively, it achieves a 1.50%–20.09% improvement in the average correlation coefficient (CC) over all baselines. In terms of qualitative evaluation, Mamba-STFM produces fusion results that are most visually consistent with ground truth. Efficiency-wise, it shows the best balance between computational cost and fusion accuracy. Furthermore, in downstream applications such as land cover classification and change detection, Mamba-STFM exhibits strong practical potential. In summary, Mamba-STFM offers a powerful and scalable solution for remote sensing image fusion tasks and holds significant promise for future development and deployment in real-world Earth observation applications.
Beyond pixel-level fusion, clustering of spatiotemporal features is critical for downstream applications such as land cover classification and change detection. In the context of Mamba-STFM, the rich joint embeddings produced by the VSS-FCAN and the mixture-of-experts decoder naturally lend themselves to unsupervised grouping: the global context captured by the visual state space supports the separation of broad thematic classes, while the fine-grained attention in FuseCore-AttNet preserves local texture cues essential for delineating subtle boundaries. Preliminary experiments have shown that clusters correspond well to known land cover types. This indicates that Mamba-STFM not only excels at image reconstruction but also produces semantically meaningful features suited for clustering-based analyses.
Although Mamba-STFM demonstrates outstanding performance in spatiotemporal fusion tasks, its multi-branch and multi-module architecture inevitably introduces considerable parameter redundancy and high demand for GPU resources during both training and inference. Therefore, our future work will focus on optimizing the network architecture to reduce parameter redundancy and improve fusion efficiency.

Author Contributions

Q.Z.: Conceptualization, methodology, software, validation, investigation formal analysis, data curation, and writing—original draft. X.Z.: Funding acquisition and writing—review and editing. C.Q., T.Z., W.H. and Y.H.: Data curation. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the major science and technology project of Qinghai Province (Grant No. 2024-SF-A6).

Data Availability Statement

The data that support the findings of this study are openly available at https://doi.org/10.4225/08/514B7FD1C798C (accessed on 2 March 2025).

Acknowledgments

The authors sincerely appreciate the hard work and valuable comments from the editor and reviewers. The authors also gratefully acknowledge the support from the Qinghai Provincial Laboratory for Intelligent Computing, the Qinghai Province Kunlun Talent High level Education Teaching Talent Project in 2023 year and Application Platform, and the Qinghai Provincial Institute of Meteorological Sciences.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Defourny, P.; Bontemps, S.; Bellemans, N.; Cara, C.; Dedieu, G.; Guzzonato, E.; Hagolle, O.; Inglada, J.; Nicola, L.; Rabaute, T.; et al. Near real-time agriculture monitoring at national scale at parcel resolution: Performance assessment of the Sen2-Agri automated system in various cropping systems around the world. Remote Sens. Environ. 2019, 221, 551–568. [Google Scholar] [CrossRef]
  2. Yuan, Q.; Shen, H.; Li, T.; Li, Z.; Li, S.; Jiang, Y.; Xu, H.; Tan, W.; Yang, Q.; Wang, J.; et al. Deep learning in environmental remote sensing: Achievements and challenges. Remote Sens. Environ. 2020, 241, 111716. [Google Scholar] [CrossRef]
  3. Du, Z.; Li, X.; Miao, J.; Huang, Y.; Shen, H.; Zhang, L. Concatenated deep-learning framework for multitask change detection of optical and SAR images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 17, 719–731. [Google Scholar] [CrossRef]
  4. Doña, C.; Chang, N.B.; Caselles, V.; Sánchez, J.M.; Camacho, A.; Delegido, J.; Vannah, B.W. Integrated satellite data fusion and mining for monitoring lake water quality status of the Albufera de Valencia in Spain. J. Environ. Manag. 2015, 151, 416–426. [Google Scholar] [CrossRef]
  5. Zhong, L.; Dzurisin, D.; Jung, H.; Zhang, J.; Zhang, Y. Radar image and data fusion for natural hazards characterisation. Int. J. Image Data Fusion 2010, 1, 217–242. [Google Scholar]
  6. Shao, Z.; Wu, W.; Li, D. Spatio-temporal-spectral observation model for urban remote sensing. Geo-Spat. Inf. Sci. 2021, 24, 372–386. [Google Scholar] [CrossRef]
  7. Xiao, J.; Aggarwal, A.K.; Duc, N.H.; Arya, A.; Rage, U.K.; Avtar, R. A review of remote sensing image spatiotemporal fusion: Challenges, applications and recent trends. Remote Sens. Appl. Soc. Environ. 2023, 32, 101005. [Google Scholar] [CrossRef]
  8. Li, J.; Li, Y.; He, L.; Chen, J.; Plaza, A. Spatio-temporal fusion for remote sensing data: An overview and new benchmark. Sci. China Inf. Sci. 2020, 63, 140301. [Google Scholar] [CrossRef]
  9. Shen, H.; Meng, X.; Zhang, L. An integrated framework for the spatio–temporal–spectral fusion of remote sensing images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7135–7148. [Google Scholar] [CrossRef]
  10. Roy, D.P.; Wulder, M.A.; Loveland, T.R.; Woodcock, C.E.; Allen, R.G.; Anderson, M.C.; Helder, D.; Irons, J.R.; Johnson, D.M.; Kennedy, R.; et al. Landsat-8: Science and product vision for terrestrial global change research. Remote Sens. Environ. 2014, 145, 154–172. [Google Scholar] [CrossRef]
  11. Justice, C.O.; Vermote, E.; Townshend, J.R.; Defries, R.; Roy, D.P.; Hall, D.K.; Salomonson, V.V.; Privette, J.L.; Riggs, G.; Strahler, A.; et al. The Moderate Resolution Imaging Spectroradiometer (MODIS): Land remote sensing for global change research. IEEE Trans. Geosci. Remote Sens. 1998, 36, 1228–1249. [Google Scholar] [CrossRef]
  12. Barnes, W.L.; Salomonson, V.V. MODIS: A global imaging spectroradiometer for the Earth Observing System. In Optical Technologies for Aerospace Sensing: A Critical Review; SPIE: St. Bellingham, WA, USA, 1992; Volume 10269, pp. 280–302. [Google Scholar]
  13. Zhu, X.; Cai, F.; Tian, J.; Williams, T.K.A. Spatiotemporal fusion of multisource remote sensing data: Literature survey, taxonomy, principles, applications, and future directions. Remote Sens. 2018, 10, 527. [Google Scholar] [CrossRef]
  14. Belgiu, M.; Stein, A. Spatiotemporal image fusion in remote sensing. Remote Sens. 2019, 11, 818. [Google Scholar] [CrossRef]
  15. Gao, F.; Masek, J.; Schwaller, M.; Hall, F. On the blending of the Landsat and MODIS surface reflectance: Predicting daily Landsat surface reflectance. IEEE Trans. Geosci. Remote Sens. 2006, 44, 2207–2218. [Google Scholar]
  16. Zhu, X.; Chen, J.; Gao, F.; Chen, X.; Masek, J.G. An enhanced spatial and temporal adaptive reflectance fusion model for complex heterogeneous regions. Remote Sens. Environ. 2010, 114, 2610–2623. [Google Scholar] [CrossRef]
  17. Zhukov, B.; Oertel, D.; Lanzl, F.; Reinhackel, G. Unmixing-based multisensor multiresolution image fusion. IEEE Trans. Geosci. Remote Sens. 1999, 37, 1212–1226. [Google Scholar] [CrossRef]
  18. Lu, M.; Chen, J.; Tang, H.; Rao, Y.; Yang, P.; Wu, W. Land cover change detection by integrating object-based data blending model of Landsat and MODIS. Remote Sens. Environ. 2016, 184, 374–386. [Google Scholar] [CrossRef]
  19. Wang, Q.; Atkinson, P.M. Spatio-temporal fusion for daily Sentinel-2 images. Remote Sens. Environ. 2018, 204, 31–42. [Google Scholar] [CrossRef]
  20. Tan, Z.; Di, L.; Zhang, M.; Guo, L.; Gao, M. An enhanced deep convolutional model for spatiotemporal image fusion. Remote Sens. 2019, 11, 2898. [Google Scholar] [CrossRef]
  21. Liao, C.; Wang, J.; Pritchard, I.; Liu, J.; Shang, J. A spatio-temporal data fusion model for generating NDVI time series in heterogeneous regions. Remote Sens. 2017, 9, 1125. [Google Scholar] [CrossRef]
  22. Ghosh, R.; Gupta, P.K.; Tolpekin, V.; Srivastav, S. An enhanced spatiotemporal fusion method–Implications for coal fire monitoring using satellite imagery. Int. J. Appl. Earth Obs. Geoinf. 2020, 88, 102056. [Google Scholar] [CrossRef]
  23. Dong, R.; Zhang, L.; Fu, H. RRSGAN: Reference-based super-resolution for remote sensing image. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5601117. [Google Scholar] [CrossRef]
  24. Wu, M.; Niu, Z.; Wang, C.; Wu, C.; Wang, L. Use of MODIS and Landsat time series data to generate high-resolution temporal synthetic Landsat data using a spatial and temporal reflectance fusion model. J. Appl. Remote Sens. 2012, 6, 063507. [Google Scholar]
  25. Lu, L.; Huang, Y.; Di, L.; Hang, D. A new spatial attraction model for improving subpixel land cover classification. Remote Sens. 2017, 9, 360. [Google Scholar] [CrossRef]
  26. Zhu, X.; Helmer, E.H.; Gao, F.; Liu, D.; Chen, J.; Lefsky, M.A. A flexible spatiotemporal method for fusing satellite images with different resolutions. Remote Sens. Environ. 2016, 172, 165–177. [Google Scholar] [CrossRef]
  27. Wang, S.; Wang, C.; Zhang, C.; Xue, J.; Wang, P.; Wang, X.; Wang, W.; Zhang, X.; Li, W.; Huang, G.; et al. A classification-based spatiotemporal adaptive fusion model for the evaluation of remotely sensed evapotranspiration in heterogeneous irrigated agricultural area. Remote Sens. Environ. 2022, 273, 112962. [Google Scholar] [CrossRef]
  28. Guo, D.; Shi, W.; Hao, M.; Zhu, X. FSDAF 2.0: Improving the performance of retrieving land cover changes and preserving spatial details. Remote Sens. Environ. 2020, 248, 111973. [Google Scholar] [CrossRef]
  29. Xue, J.; Leung, Y.; Fung, T. A Bayesian data fusion approach to spatio-temporal fusion of remotely sensed images. Remote Sens. 2017, 9, 1310. [Google Scholar] [CrossRef]
  30. Li, A.; Bo, Y.; Zhu, Y.; Guo, P.; Bi, J.; He, Y. Blending multi-resolution satellite sea surface temperature (SST) products using Bayesian maximum entropy method. Remote Sens. Environ. 2013, 135, 52–63. [Google Scholar] [CrossRef]
  31. Huang, B.; Song, H. Spatiotemporal reflectance fusion via sparse representation. IEEE Trans. Geosci. Remote Sens. 2012, 50, 3707–3716. [Google Scholar] [CrossRef]
  32. Song, H.; Huang, B. Spatiotemporal satellite image fusion through one-pair image learning. IEEE Trans. Geosci. Remote Sens. 2012, 51, 1883–1896. [Google Scholar] [CrossRef]
  33. Song, H.; Liu, Q.; Wang, G.; Hang, R.; Huang, B. Spatiotemporal satellite image fusion using deep convolutional neural networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 821–829. [Google Scholar] [CrossRef]
  34. Tan, Z.; Yue, P.; Di, L.; Tang, J. Deriving high spatiotemporal remote sensing images using deep convolutional network. Remote Sens. 2018, 10, 1066. [Google Scholar] [CrossRef]
  35. Liu, X.; Deng, C.; Chanussot, J.; Hong, D.; Zhao, B. StfNet: A two-stream convolutional neural network for spatiotemporal image fusion. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6552–6564. [Google Scholar] [CrossRef]
  36. Cheng, F.; Fu, Z.; Tang, B.; Huang, L.; Huang, K.; Ji, X. Stf-egfa: A remote sensing spatiotemporal fusion network with edge-guided feature attention. Remote Sens. 2022, 14, 3057. [Google Scholar] [CrossRef]
  37. Li, Y.; Li, J.; He, L.; Chen, J.; Plaza, A. A new sensor bias-driven spatio-temporal fusion model based on convolutional neural networks. Sci. China Inf. Sci. 2020, 63, 140302. [Google Scholar] [CrossRef]
  38. Zhang, H.; Song, Y.; Han, C.; Zhang, L. Remote sensing image spatiotemporal fusion using a generative adversarial network. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4273–4286. [Google Scholar] [CrossRef]
  39. Song, B.; Liu, P.; Li, J.; Wang, L.; Zhang, L.; He, G.; Chen, L.; Liu, J. MLFF-GAN: A multilevel feature fusion with GAN for spatiotemporal remote sensing images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4410816. [Google Scholar] [CrossRef]
  40. Liu, H.; Yang, G.; Deng, F.; Qian, Y.; Fan, Y. MCBAM-GAN: The GAN spatiotemporal fusion model based on multiscale and CBAM for remote sensing images. Remote Sens. 2023, 15, 1583. [Google Scholar] [CrossRef]
  41. Chen, J.; Wang, L.; Feng, R.; Liu, P.; Han, W.; Chen, X. CycleGAN-STF: Spatiotemporal fusion via CycleGAN-based image generation. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5851–5865. [Google Scholar] [CrossRef]
  42. Tan, Z.; Gao, M.; Li, X.; Jiang, L. A flexible reference-insensitive spatiotemporal fusion model for remote sensing images using conditional generative adversarial network. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5601413. [Google Scholar] [CrossRef]
  43. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  44. Li, W.; Cao, D.; Peng, Y.; Yang, C. MSNet: A multi-stream fusion network for remote sensing spatiotemporal fusion based on transformer and convolution. Remote Sens. 2021, 13, 3724. [Google Scholar] [CrossRef]
  45. Chen, G.; Jiao, P.; Hu, Q.; Xiao, L.; Ye, Z. SwinSTFM: Remote sensing spatiotemporal fusion using Swin transformer. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5410618. [Google Scholar] [CrossRef]
  46. Jiang, M.; Shao, H. A cnn-transformer combined remote sensing imagery spatiotemporal fusion model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 13995–14009. [Google Scholar] [CrossRef]
  47. Hu, M.; Wu, C.; Du, B. EMS-Net: Efficient multi-temporal self-attention for hyperspectral change detection. In Proceedings of the IGARSS 2023–2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; pp. 6664–6667. [Google Scholar]
  48. Metz, L.; Poole, B.; Pfau, D.; Sohl-Dickstein, J. Unrolled generative adversarial networks. arXiv 2016, arXiv:1611.02163. [Google Scholar]
  49. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of wasserstein gans. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 5769–5779. [Google Scholar]
  50. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Jiao, J.; Liu, Y. Vmamba: Visual state space model. In Proceedings of the Advances in Neural Information Processing Systems 37 (NIPS 2024), Vancouver, BC, Canada, 10–15 December 2024; pp. 103031–103063. [Google Scholar]
  51. Emelyanova, I.; McVicar, T.; Van Niel, T.; Li, L.; Van Dijk, A. Landsat and MODIS data for the Coleambally Irrigation Area; v3. Data Collection; CSIRO: Canberra, Australia, 2013. [Google Scholar] [CrossRef]
  52. Emelyanova, I.; McVicar, T.; Van Niel, T.; Li, L.; Van Dijk, A. Landsat and MODIS Data for the Lower Gwydir Catchment Study Site; v3. Data Collection; CSIRO: Canberra, Australia, 2013. [Google Scholar] [CrossRef]
  53. Zhang, X.; Li, S.; Tan, Z.; Li, X. Enhanced wavelet based spatiotemporal fusion networks using cross-paired remote sensing images. ISPRS J. Photogramm. Remote Sens. 2024, 211, 281–297. [Google Scholar] [CrossRef]
  54. Huang, H.; He, W.; Zhang, H.; Xia, Y.; Zhang, L. STFDiff: Remote sensing image spatiotemporal fusion with diffusion models. Inf. Fusion 2024, 111, 102505. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the Mamba-STFM model and the structure of the VSS-FCAN block.
Figure 1. The overall architecture of the Mamba-STFM model and the structure of the VSS-FCAN block.
Remotesensing 17 02135 g001
Figure 2. Network architecture diagram of FuseCore-AttNet.
Figure 2. Network architecture diagram of FuseCore-AttNet.
Remotesensing 17 02135 g002
Figure 3. Network structure diagram of the STF-MoE block.
Figure 3. Network structure diagram of the STF-MoE block.
Remotesensing 17 02135 g003
Figure 4. Detailed comparison of different fusion products with ground truth values on the CIA dataset.
Figure 4. Detailed comparison of different fusion products with ground truth values on the CIA dataset.
Remotesensing 17 02135 g004
Figure 5. RMSE error map of different fusion products on the CIA dataset.
Figure 5. RMSE error map of different fusion products on the CIA dataset.
Remotesensing 17 02135 g005
Figure 6. Detailed comparison of different fusion products with ground truth values on the LGC dataset.
Figure 6. Detailed comparison of different fusion products with ground truth values on the LGC dataset.
Remotesensing 17 02135 g006
Figure 7. RMSE error map of different fusion products on the LGC dataset.
Figure 7. RMSE error map of different fusion products on the LGC dataset.
Remotesensing 17 02135 g007
Figure 8. K-means clustering results of fused images by various methods in the CIA region, where green represents Class 1 and purple represents Class 2.
Figure 8. K-means clustering results of fused images by various methods in the CIA region, where green represents Class 1 and purple represents Class 2.
Remotesensing 17 02135 g008
Figure 9. Scatter distribution of pixel values between different fused images and the corresponding real images in the RED and NIR bands within the CIA study area.
Figure 9. Scatter distribution of pixel values between different fused images and the corresponding real images in the RED and NIR bands within the CIA study area.
Remotesensing 17 02135 g009
Figure 10. K-means clustering results of fused images by various methods in the LGC region, where green represents Class 1 and purple represents Class 2.
Figure 10. K-means clustering results of fused images by various methods in the LGC region, where green represents Class 1 and purple represents Class 2.
Remotesensing 17 02135 g010
Figure 11. Scatter distribution of pixel values between different fused images and the corresponding real images in the RED and NIR bands within the LGC study area.
Figure 11. Scatter distribution of pixel values between different fused images and the corresponding real images in the RED and NIR bands within the LGC study area.
Remotesensing 17 02135 g011
Table 1. Image quality evaluation metrics and formula explanation.
Table 1. Image quality evaluation metrics and formula explanation.
MetricFormulaSymbol Explanation
RMSE RMSE = 1 N i = 1 N ( a i b i ) 2 N: Total number of pixels
a i : The i-th pixel in the reference image
b i : The i-th pixel in the image to be evaluated
SSIM SSIM = ( 2 m a m b + C 1 ) ( 2 c a b + C 2 ) ( m a 2 + m b 2 + C 1 ) ( v a + v b + C 2 ) m a , m b : Local mean of the two images
v a , v b : Local variance of the two images
c a b : Local covariance of the two images
C 1 , C 2 : Stability constants
UIQI Q = 4 c a b m a m b ( v a + v b ) ( m a 2 + m b 2 ) Same symbols as SSIM:
m a , m b : Local mean
v a , v b : Local variance
c a b : Local covariance
CC CC = i = 1 N ( a i m a ) ( b i m b ) i = 1 N ( a i m a ) 2 i = 1 N ( b i m b ) 2 N: Number of pixels
a i , b i : Pixel values in the two images
m a = 1 N i = 1 N a i , m b = 1 N i = 1 N b i : Global mean of the images
ERGAS ERGAS = 100 · h L h H 1 B k = 1 B R k M k 2 B: Number of bands
R k : RMSE of the k-th band
M k : Mean value of the reference image for the k-th band
h L , h H : Pixel size of low/high resolution images
Table 2. Quantitative comparison of various fusion methods on the CIA dataset. The black bold font represents the best value.
Table 2. Quantitative comparison of various fusion methods on the CIA dataset. The black bold font represents the best value.
MetricsBandSTARFMESTARFMGANSTFMEDCSTFNSwinSTFMECPW-STFNSTFDiffMamba-STFM
RMSE10.01740.01360.01390.01260.01120.01250.00930.0086
20.02360.01950.02020.01610.01480.01370.01280.0115
30.03330.03060.03130.02550.02170.01520.01970.0169
40.05380.05150.05470.04520.02990.02610.02630.0256
50.05600.05370.05630.04840.03630.03490.03350.0311
60.05440.04820.05520.04330.03060.02850.02880.0277
SSIM10.87560.90790.90370.91510.93870.93260.94240.9515
20.86740.88690.90930.92930.92540.93110.93850.9363
30.82200.85380.84050.86340.88450.89040.89010.8976
40.75160.78250.73500.84190.82160.84550.84680.8640
50.79360.78470.78910.78530.77250.78930.79140.8092
60.77170.78060.78220.79110.79170.79720.81160.8184
UIQI10.72270.77440.79830.82160.83620.83610.83550.8863
20.71570.78300.73710.85080.86560.88810.88160.9162
30.80560.82310.80620.86190.87280.89080.90430.9206
40.82280.83600.84190.90260.91680.92490.92220.9435
50.85310.85860.85040.92120.92070.93110.93650.9430
60.86040.87280.88490.92550.92150.93570.93310.9369
CC10.73050.78870.76360.82510.84630.86560.87390.8952
20.75540.79310.77600.84390.87230.90810.90580.9195
30.77910.82960.83940.85100.87850.89070.90320.9226
40.83360.83870.85320.91570.91880.92580.92170.9440
50.85290.85950.86480.92150.92090.93440.93310.9471
60.85140.87320.89170.91840.92150.93580.93860.9378
ERGAS-1.70861.59951.60571.25901.06580.96330.91490.8615
Table 3. Quantitative comparison of various fusion methods on the LGC dataset, with the best values in bold.
Table 3. Quantitative comparison of various fusion methods on the LGC dataset, with the best values in bold.
MetricsBandSTARFMESTARFMGANSTFMEDCSTFNSwinSTFMECPW-STFNSTFDiffMamba-STFM
RMSE10.01460.01390.01550.01340.01220.01170.01050.0098
20.01890.01850.02120.01960.01780.01450.01490.0127
30.02690.02680.02640.02550.02290.02140.01970.0165
40.03990.03950.04100.03810.02940.02880.02810.0234
50.05410.05520.05380.05450.04350.04140.04270.0335
60.04270.04240.03990.04060.03190.02990.03040.0257
SSIM10.93050.93810.92610.93270.93970.94720.94500.9494
20.89760.90240.90420.91870.90600.91140.91610.9246
30.85770.85820.87180.87120.87360.88190.88120.8958
40.73430.76720.79270.81290.81590.81420.80630.8395
50.61140.63140.63530.65170.70390.71970.72600.7569
60.66760.67920.69230.71600.76040.76460.77280.8026
UIQI10.60130.67920.65170.69830.77650.78550.79620.8775
20.67440.71530.67160.75790.77270.77500.78940.8826
30.69480.72980.69700.75440.76930.77840.77510.8832
40.74390.78010.71750.79730.87710.89130.88060.9244
50.74220.80860.76520.80650.86830.88400.87520.9230
60.75190.80600.76330.80730.86980.86990.88140.9148
CC10.68310.70180.61920.72360.78250.81440.80390.8845
20.71880.75110.63170.69640.77570.78290.78750.8873
30.70620.74550.65100.68770.77190.77530.78030.8868
40.76520.79420.74780.79730.88270.89660.90280.9266
50.79130.80560.78920.80650.87150.88020.88960.9247
60.80440.80320.78310.80730.87200.89110.89520.9174
ERGAS-1.90421.88722.13171.96881.61791.44131.42601.2472
Table 4. Computational complexity, number of parameters, average inference time, max memory usage, and average correlation coefficients (CCs) for each channel on the datasets for 6 deep learning spatiotemporal fusion models.
Table 4. Computational complexity, number of parameters, average inference time, max memory usage, and average correlation coefficients (CCs) for each channel on the datasets for 6 deep learning spatiotemporal fusion models.
ModelGANSTFMEDCSTFNSwinSTFMECPW-STFNSTFDiffMamba-STFM
FLOPs (G)37.758618.583828.182230.71104.593716.5276
Parms (M)0.57710.282537.46560.471942.822647.1078
Average inference time (ms)6.90532.518848.2391120.3042236.187744.8670
Max memory usage (MB)187.42991983.3047311.1792605.1192733.5072288.7332
CIA0.83150.87930.89310.91010.91270.9277
LGC0.70370.75310.82610.84010.84320.9046
Table 5. Ablation study of the core modules in the Mamba-STFM model. The symbol "∖" indicates the full model, while ↑ and ↓ show increasing and decreasing values in the column.
Table 5. Ablation study of the core modules in the Mamba-STFM model. The symbol "∖" indicates the full model, while ↑ and ↓ show increasing and decreasing values in the column.
Eliminate or ReplaceSSIM ↓UIQI ↓CC ↓RMSE ↑ERGAS ↑
0.86150.90090.90460.02031.2472
FuseCore-AttNet0.85530.89740.88250.02381.2528
Temporal Expert0.84720.87010.86830.02641.3575
Spatial Expert0.82580.85620.85190.03151.4091
All0.82110.84940.84470.03381.4216
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Q.; Zhang, X.; Quan, C.; Zhao, T.; Huo, W.; Huang, Y. Mamba-STFM: A Mamba-Based Spatiotemporal Fusion Method for Remote Sensing Images. Remote Sens. 2025, 17, 2135. https://doi.org/10.3390/rs17132135

AMA Style

Zhang Q, Zhang X, Quan C, Zhao T, Huo W, Huang Y. Mamba-STFM: A Mamba-Based Spatiotemporal Fusion Method for Remote Sensing Images. Remote Sensing. 2025; 17(13):2135. https://doi.org/10.3390/rs17132135

Chicago/Turabian Style

Zhang, Qiyuan, Xiaodan Zhang, Chen Quan, Tong Zhao, Wei Huo, and Yuanchen Huang. 2025. "Mamba-STFM: A Mamba-Based Spatiotemporal Fusion Method for Remote Sensing Images" Remote Sensing 17, no. 13: 2135. https://doi.org/10.3390/rs17132135

APA Style

Zhang, Q., Zhang, X., Quan, C., Zhao, T., Huo, W., & Huang, Y. (2025). Mamba-STFM: A Mamba-Based Spatiotemporal Fusion Method for Remote Sensing Images. Remote Sensing, 17(13), 2135. https://doi.org/10.3390/rs17132135

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop