Next Article in Journal
Effectiveness of Therapeutic Exercise in Physiotherapy with Blood Flow Restriction in Patients with Knee Pathologies: A Systematic Review
Previous Article in Journal
From Misinformation to Resilient Communication: Strategic Simulation of Social Network Dynamics in the Pharmaceutical Industry
Previous Article in Special Issue
LightMFF: A Simple and Efficient Ultra-Lightweight Multi-Focus Image Fusion Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ESSTformer: A CNN-Transformer Hybrid with Decoupled Spatial Spectral Transformers for Hyperspectral Image Super-Resolution

School of Communication and Information Engineering, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(21), 11738; https://doi.org/10.3390/app152111738
Submission received: 9 October 2025 / Revised: 28 October 2025 / Accepted: 31 October 2025 / Published: 4 November 2025
(This article belongs to the Special Issue Advances in Optical Imaging and Deep Learning)

Abstract

Hyperspectral images (HSIs) are crucial for ground object classification, target detection, and related applications due to their rich spatial spectral information. However, hardware limitations in imaging systems make it challenging to directly acquire HSIs with a high spatial resolution. While deep learning-based single hyperspectral image super-resolution (SHSR) methods have made significant progress, existing approaches primarily rely on convolutional neural networks (CNNs) with fixed geometric kernels, which struggle to model global spatial spectral dependencies effectively. To address this, we propose ESSTformer, a novel SHSR framework that synergistically integrates CNNs’ local feature extraction and Transformers’ global modeling capabilities. Specifically, we design a multi-scale spectral attention module (MSAM) based on dilated convolutions to capture local multi-scale spatial spectral features. Considering the inherent differences between spatial and spectral information, we adopt a decoupled processing strategy by constructing separate spatial and Spectral Transformers. The Spatial Transformer employs window attention mechanisms and an improved convolutional multi-layer perceptron (CMLP) to model long-range spatial dependencies, while the Spectral Transformer utilizes self-attention mechanisms combined with a spectral enhancement module to focus on discriminative spectral features. Extensive experiments on three hyperspectral datasets demonstrate that the proposed ESSTformer achieves a superior performance in super-resolution reconstruction compared to state-of-the-art methods.

1. Introduction

Hyperspectral images (HSIs) capture the reflectance or transmittance information of objects across dozens to hundreds of continuous spectral bands. Compared with standard natural images, HSIs possess more spectral bands, which enables them to acquire richer spectral detail information that more accurately characterizes the fine features and intrinsic properties of objects. Consequently, hyperspectral imaging has been widely applied in numerous fields including remote sensing [1,2,3], medical imaging [4,5,6], and industrial inspection [7,8,9]. However, constrained by imaging mechanisms and equipment limitations, acquired hyperspectral images require tradeoffs among spatial resolution, spectral resolution, swath width, and signal-to-noise ratio, making it difficult to directly obtain high-spatial-resolution hyperspectral images. Yet, in application scenarios such as mineral exploration [10], urban fine-scale mapping [11], and small target detection [12], a higher spatial resolution is essential. This limited spatial resolution has become a bottleneck restricting hyperspectral imaging applications in remote sensing technology. Although hardware improvements can enhance resolution, such approaches often involve high costs and lengthy cycles. In comparison, image processing techniques for resolution enhancement offer greater practical value.
Super-resolution (SR) [13] technology enables the reconstruction of high-resolution images consistent with the original scene from either a single or a sequence of low-resolution images. For hyperspectral image super-resolution reconstruction, a key classification hinges on whether auxiliary information (e.g., panchromatic images [14,15], RGB images [16,17], or multispectral images [18]) is employed; this divides the methods into two main categories, namely fusion-based hyperspectral image super-resolution and single image hyperspectral super-resolution (SHSR). The fusion-based approach enhances the spatial resolution of the target hyperspectral image by fusing low-resolution hyperspectral images with high-resolution auxiliary images. However, this method relies on a critical assumption involving a perfect registration between the low-resolution hyperspectral images and the high-resolution auxiliary images. In practical scenarios, acquiring high-quality auxiliary images is often challenging, if not entirely unfeasible. In contrast, SHSR requires no auxiliary information throughout the reconstruction process. It can yield satisfactory performance in most cases, thereby demonstrating a broad application potential.
SHSR methods can be primarily categorized into traditional approaches and deep learning-based methods. Most traditional methods, such as those based on sparse representation [19,20] and low-rank matrices [21,22], typically rely on manually designed prior knowledge (e.g., self-similarity, sparsity, and low-rank properties) as regularization terms to guide the reconstruction process. While these methods can deliver relatively favorable results in specific scenarios, their reliance on handcrafted priors gives rise to two key limitations, manifesting in computationally intensive processes and constrained representation capability. These drawbacks further restrict their ability to fully capture the inherent characteristics of hyperspectral data. With the rapid advancement of deep learning techniques, especially the widespread adoption of convolutional neural networks (CNNs), deep learning-based SHSR methods have achieved remarkable performance breakthroughs [23,24]. Nevertheless, due to the fixed size of their convolutional kernels, CNNs inherently possess limited receptive fields, which renders them inefficient in modeling long-range dependencies. Additionally, most of these CNN-based methods only focus on single-scale spatial features of HSIs, while overlooking the rich multi-scale mapping relationships across multi-scale spaces. This oversight ultimately limits the further improvement of their overall reconstruction performance [25].
In recent years, Transformer models [26] have been successfully applied to HSI classification tasks, owing to their exceptional ability to capture global information. However, hyperspectral datasets typically have a limited number of training samples, which poses a particular challenge for training effective Transformer models for SR tasks. Furthermore, research on Transformer architectures specifically tailored to SR remains relatively scarce. To address the limitation of Transformers struggling with small-scale datasets, several studies have attempted to integrate CNNs with Transformers [27,28,29]. For example, Interactformer [28] adopts a hybrid architecture that combines 3D convolutions with Transformer modules, enabling the simultaneous extraction of local and global spatial spectral features. Nevertheless, the extensive use of 3D convolutions and parallel structures leads to a high computational complexity, as well as increased demands on hardware memory resources.
The Enhanced Spatial Spectral Transformer (ESSTformer) is proposed for SHSR to address these challenges. This novel framework innovatively integrates the strengths of CNNs and Transformer architectures, achieving significant improvements in reconstruction performance by synergistically extracting local and global spatial spectral features. For local feature extraction, we design a multi scale spectral attention module (MSAM) based on dilated convolutions. This module captures multi-scale spatial features via its dilated convolution design, while leveraging a spectral attention mechanism to dynamically adjust weights across different spectral bands, enabling the more effective extraction of local spatial spectral information. For global feature modeling, we adopt a decoupled processing strategy that separately constructs spectral and Spatial Transformers to better handle the distinct characteristics of these modalities. In the Spectral Transformer, inter-band correlations are preserved through self-attention mechanisms, and a spectral enhancement module is introduced to dynamically adjust inter-band weights, which emphasizes critical spectral information and enables the more accurate selection and fusion of important spectral features, thus improving the spectral fidelity of reconstructed images. In the Spatial Transformer, windowed attention is employed to model long-range spatial dependencies, while a locally enhanced feed-forward network maintains essential local neighborhood information. This design simultaneously preserves spatial details and mitigates the excessive computational costs associated with standard self-attention. Comprehensive qualitative and quantitative experiments across three hyperspectral datasets demonstrate the effectiveness of the ESSTformer. Our main contributions are summarized as follows:
  • We propose ESSTformer, a novel CNN-Transformer hybrid framework for SHSR that effectively exploits both local and global spatial spectral information, significantly boosting the super-resolution performance.
  • We design the MSAM to learn multi-scale interactions between local spatial and spectral features, substantially enhancing the local feature representation.
  • Considering the inherent differences between spatial and spectral characteristics, we develop a decoupled processing strategy with dedicated Transformer modules; the Spatial Transformer captures global spatial dependencies while the Spectral Transformer models long-range spectral relationships, working synergistically to improve feature extraction precision.
  • We replace standard MLPs with convolutional multi-layer perceptrons (CMLPs) to better leverage neighborhood spatial context, thereby enhancing the model’s representational capacity and adaptability.

2. Materials and Methods

In this section, we present a detailed description of the proposed SHSR method, ESSTformer, which is structured around five essential components: the overall framework, MSAM, the Spectral Transformer, the Spatial Transformer and the loss function.

2.1. Overall Framework

As illustrated in Figure 1, the proposed ESSTformer model in this paper combines CNNs and Transformer architectures to simultaneously process both local and global spatial spectral information of hyperspectral images. The model takes a low-resolution hyperspectral image I L R R H × W × B with dimensions H × W × B (where H , W , and B denote the height, width, and number of spectral as input, and outputs a high-resolution hyperspectral image I S R R s H × s W × B with a super-resolution scale factor of s. The objective of the ESSTformer network is to predict the super-resolved image ISR from the input low-resolution image ILR, making it as close as possible to the original high-resolution hyperspectral image I H R R s H × s W × B , which can be expressed as follows:
I S R = H E S S T f o r m e r ( I L R ) ,
where H E S S T f o r m e r ( ) represents the function corresponding to the proposed ESSTformer methodology.
For the purpose of extracting local spatial spectral information, we designed the MSAM to capture features across different scales. The module can be formally expressed as follows:
F loc = H M S A M ( I L R ) ,
where F l o c represents the extracted local spatial spectral features, and H M S A M ( ) denotes the function corresponding to the operation of the MSAM applied to the input I L R .
For the full exploitation of global spatial spectral information, we first project the extracted local features F l o c into the hyperspectral dimension, which can be represented as follows:
F 0 = H 0 ( F l o c ) ,
where F 0 R H × W × B represents the hyperspectral spatial spectral features with C spectral bands, H × W denotes the spatial resolution, and H 0 ( ) refers to the 1 × 1 convolutional function.
The extracted features F 0 are then fed into an Enhanced Spectral Transformer Block (EETB) to model inter-band correlations across different spectral wavelengths, thereby capturing global spectral characteristics. Subsequently, the feature maps output from the EETB module are processed by an Enhanced Spatial Transformer Block(EATB) to establish long-range dependencies in the spatial dimension. We construct multiple cascaded EETB and EATB modules to comprehensively extract both spatial and spectral features. The output feature map F from the final EATB module is fused with feature map F 0 obtained through long skip connections, generating a new feature representation F K . To prevent the loss of low-frequency information, we employ residual connections by concatenating each module’s output with F K , followed by a 1 × 1 convolution for dimensionality reduction to ensure feature map F f maintains the same dimensions as F. Finally, an upsampling module expands the spatial resolution of the fused deep spatial spectral features F f , which can be formally expressed as follows:
F u p = H u p ( F f ) ,
where H u p ( ) denotes the upsampling function based on the Pixel Shuffle technique [30], and F up has C spectral bands with spatial dimensions of s H × s W . Through this series of processing steps, the feature map is significantly enhanced in both spectral representation and spatial resolution, leading to a richer, multi-dimensional feature map that effectively captures the global spatial and spectral context of the input data.
Finally, bicubic interpolation is applied to upsample the input features, mitigating training complexity and helping to preserve original information through subsequent residual connections to the network’s output. After bicubic interpolation, a 1 × 1 convolutional layer is used to adjust the spatial dimensions of the interpolated feature map, ensuring it aligns with the output feature dimensions F f , thus generating the residual feature F r e s . To maintain the spatial resolution of the final reconstructed image, consistent with the original high-resolution hyperspectral image, an additional 1 × 1 convolutional layer is applied to the feature map F r e s , which has already been adjusted by both interpolation and convolution. The upsampling process can be formalized as follows:
F r e s = F u p + H 1 ( I L R ) ,
I S R = H 2 ( F r e s ) ,
where F r e s R s H × s W × C represents the residual feature, H 1 ( ) and H 2 ( ) denote the 1 × 1 convolutional layers, I L R is the bicubic upsampled version of the input low-resolution hyperspectral image, and I S R is the reconstructed high-resolution hyperspectral image.

2.2. MSAM

Currently, most SHSR methods primarily rely on 2D or 3D convolutions to extract spatial spectral features at a single scale, which inherently limits the model’s ability to explore rich scale-wise mapping relationships in multi-scale spaces. The MSAM, illustrated in Figure 2, is designed to address this limitation by capturing local multi-scale spatial spectral features of the image.
While existing methods often rely on convolutional kernels of fixed receptive fields, their capacity to represent multi-scale spatial spectral contexts remains limited. Accordingly, we design the MSAM, aimed at explicitly capturing local features at multiple scales. The module takes I L R as input. Specifically, we employ dilated convolutions with three distinct dilation rates (1, 3, 5), combined with ReLU activation functions, to capture multi-scale spatial information. The extracted features F i at each scale are expressed as follows:
F 1 = H D ( I L R , 1 ) ,
F 2 = H D ( I L R , 2 ) ,
F 3 = H D ( I L R , 3 ) ,
where H D ( , ) represents the dilated convolution function. Subsequently, the multi-scale features F i extracted at different scales are fused and further refined through a Residual Spatial Module to obtain residual multi-scale spatial features F s p a , which can be formally expressed as follows:
F d = F 1 + F 2 + F 3 ,
F s p a = F d + H 3 ( F d ) ,
where H 3 ( ) represents the composite function of two 3 × 3 convolutional layers (with intermediate RELU activation) in the Residual Spatial Module. A Residual Channel Attention Block (RCAB) [31] is incorporated to enhance the representation by modeling inter-band dependencies. The process is formally expressed as follows:
F l o c = H R C A B ( F s p a ) ,
where F l o c represents the local multi-scale spatial spectral features and H R C A B ( ) denotes the RCAB operation function.
Through MSAM processing, we successfully extract local multi-scale spatial spectral features from HSIs, while establishing a robust foundation for subsequent global spatial spectral feature extraction.

2.3. EETB

The strong inter-band spectral correlations in HSIs are particularly crucial for SHSR analysis. Therefore, to effectively capture these long-range dependencies between different spectral bands, we have designed the EETB module as shown in Figure 1b. The EETB module consists of two residual blocks, the first containing a Layer Normalization (LN) layer followed by a deformable convolution-based self-attention layer integrated with a spectral enhancement (SE) module to enhance feature representation capability, and the second comprising a LN layer with a subsequent MLP layer designed to deeply explore non-linear feature relationships. Let F e R H × W × C represent the input to a single EETB module; the processing flow can be mathematically expressed as follows:
F ^ e = M H S A ( L N ( F e ) ) + S E ( L N ( F e ) ) + F e ,
F e = M L P ( L N ( F ^ e ) ) + F ^ e ,
where F ^ e denotes the enhanced spectral features, F e represents the global spectral features output by the EETB module, LN(·) refers to the function implemented by the Layer Normalization (LN) layer, and MLP(·) denotes the function implemented by the MLP layer.

2.3.1. Long Self-Attention

An adaptive multi-head self-attention mechanism is proposed to better capture correlations and variations among spectral bands. Its query ( Q ) is generated using deformable convolution, which allows the module to adaptively focus on more relevant spatial contexts based on input features, thereby achieving a more flexible receptive field than standard convolutions.
As illustrated in Figure 3, the normalized feature F e R H × W × C is projected into query ( Q R H × W × C ), key ( K R H × W × C ), and value ( V R H × W × C ) tensors. This projection process can be mathematically represented as follows:
Q = W d e f o r m a b l e ( F e ) , K = W K ( F e ) , V = W V ( F e ) ,
where W d e f o r m a b l e ( ) denotes the deformable convolution function, and W K ( ) and W V ( ) represent the 1 × 1 point-wise convolution functions.
Subsequently, the projected tensors are reshaped to Q R H W × C , K R C × H W , and V R H W × C , respectively, then split into N heads along the channel dimension. The attention mechanism for each head is computed as follows:
A tt e n t i o n i ( Q i , K i , V i ) = V i S o f t M a x ( Q i K i T d k ) ,
where d k = C / N is the feature dimension per head. The outputs from all heads are concatenated and projected:
M u l t i H e a d = C o n c a t ( A t t e n t i o n 1 , A t t e n t i o n 2 , , A t t e n t i o n N ) ,
Finally, the output is reshaped back to H × W × C and processed by a 1 × 1 convolution to produce the enhanced spectral attentional features F e . Our deformable attention mechanism effectively captures long-range spectral spatial dependencies through content-aware spatial sampling, significantly improving feature representation for SHSR tasks.

2.3.2. Spectral Enhancement

The SE module is incorporated to complement the spatial adaptive attention mechanism and further optimize the modeling of inter-channel dependencies. Designed as a computational unit for adaptive recalibration, it emphasizes informative spectral bands while suppressing less useful ones.
As illustrated in Figure 4, the input feature F e R H × W × C is first squeezed into a channel descriptor vector F c R 1 × 1 × C via global average pooling. Subsequently, an excitation operation is applied through a simple gating mechanism with a sigmoid activation:
F z = σ ( W h δ ( W l F c ) ) ,
where δ denotes the ReLU function, W l R K × C and W h R C × K are the weights of two linear layers forming a bottleneck structure for dimensionality reduction (K = C/r) and restoration, and σ is the sigmoid function. The final output is obtained by scaling the original input with the computed channel weights:
F e = F e F z ,
By explicitly modeling spectral channel relationships, this lightweight module ensures that subsequent processing stages focus on the most discriminative spectral features, thereby enhancing the overall representational capacity of our framework for hyperspectral image reconstruction.

2.4. EATB

While the EETB effectively models global inter-band correlations, its attention mechanism, which operates on flattened spectral vectors, is inherently limited in capturing long-range spatial dependencies. This is because the spectral self-attention prioritizes relationships between bands across all spatial locations but does not explicitly model the contextual relationships between different pixels or regions within the same band. Consequently, a standalone Spectral Transformer may struggle to reconstruct fine spatial structures and edges that require integrating information from distant parts of the image.
We advocate for a decoupled spatial spectral modeling strategy to address this fundamental limitation and achieve a more comprehensive representation. This design explicitly separates the learning of spectral and spatial features into dedicated Transformer blocks. Following the spectral feature extraction by the EETB, the EATB is introduced to explicitly and efficiently model long-range spatial contexts.
As shown in Figure 1c, the EATB module consists of two residual blocks: the first contains a LN followed by a window-based attention layer, while the second comprises an LN layer followed by a CMLP layer. Let F a R H × W × C be the input to a single EATB module; its processing can be expressed as follows:
F ^ a = W M S A ( L N ( F a ) ) + F a ,
F a = C M L P ( L N ( F ^ a ) ) + F ^ a ,
where F ^ a R H × W × C represents the spatial features captured through the window attention mechanism, F a R H × W × C denotes the global spatial features output by the EATB module, LN(·) refers to the operation function of the Layer Normalization layer, and CMLP(·) represents the operation function of the CMLP layer.

2.4.1. WMSA

The Window-based Multi-head Self-Attention (WMSA) module is adopted to efficiently model long-range spatial dependencies while avoiding the quadratic complexity of standard self-attention. The input feature map of size H × W × C is first partitioned into non-overlapping M × M windows, reshaping the input to H W M 2 × M 2 × C .
Within each window, the feature F a R M × M × C is projected into queries ( Q R M × M × C ), keys ( K R M × M × C ), and values ( V R M × M × C ) via 1 × 1 convolutions:
Q = W Q ( F a ) , K = W K ( F a ) , V = W V ( F a ) ,
Subsequently, the projecte Q , K , and V are reshaped into Q R M 2 × C , K R M 2 × C and V R M 2 × C , respectively. The attention matrix is then computed by applying the self-attention mechanism within local windows, which is formulated as follows:
W M S A i ( Q , K , V ) = S o f t M a x ( Q i K i T d k ) V i ,
where W M S A i ( Q , K , V ) represents the output of the i-th window, and d k = C / N is the feature dimension per head, with N being the total number of attention heads.
Subsequently, the outputs from multiple windows are concatenated, which can be expressed as follows:
W M S A = C o n c a t ( W M S A 1 , W M S A 2 , , W M S A N ) ,
The concatenated features are then reshaped and restored to their original dimensions through a 1 × 1 convolutional operation, yielding the final window attention features F a .

2.4.2. CMLP

In standard Transformer architectures, the feed-forward network (FFN) employs position-wise layers to transform features. While effective for modeling channel-wise interactions, this design possesses a fundamental limitation for visual tasks: it operates independently on each pixel location, thereby failing to capture the local spatial context that is crucial for understanding hyperspectral imagery [32].
In response to this gap, we introduce the convolutional MLP (CMLP), a powerful alternative to the standard FFN. The core innovation of our CMLP lies in its gated convolution mechanism, which explicitly incorporates local spatial feature extraction and adaptive modulation into the feed-forward process.
As illustrated in Figure 5, the CMLP first projects the input feature F ^ a using two parallel 3 × 3 convolutional layers. Unlike the standard FFN, this setup processes each pixel by considering its neighboring context. The outputs of these two paths are then fused through a gating mechanism formulated as follows:
F ˜ a = f c o n v 3 ( F ^ a ) f G E L U ( f c o n v 3 ( F ^ a ) ) ,
where f c o n v 3 ( ) corresponds to 1 × 1 and 3×3 convolution operations, and ‘·’ is the element-wise multiplication. This design, inspired by gated linear units, allows one path to non-linearly transform the features while the other acts as a gate, dynamically modulating which spatial features should be emphasized or suppressed.
The gated output F ˜ a is subsequently refined by a spatial attention module(SA) to prioritize globally salient regions; the spatial attention map is generated by aggregating channel information via both global average pooling and global max pooling, followed by a convolution and a sigmoid activation function.
Finally, a 1 × 1 convolutional layer f c o n v 1 ( ) projects the refined features back to the original channel dimension, ensuring compatibility with the subsequent Transformer blocks:
F ^ a = f c o n v 1 ( F ˜ a ) ,
By replacing the channel-wise layers with local convolutional processing, a gated feature modulation mechanism, and global spatial attention, our CMLP effectively captures the intricate spatial spectral patterns inherent in hyperspectral data, thereby significantly enhancing the representational capacity of the Transformer backbone.

2.5. Loss Function

In evaluating the performance of SR reconstruction, selecting appropriate loss functions is critical for quantifying the discrepancy between reconstructed images and their corresponding ground-truth counterparts. Extensive research has demonstrated that both l 1 and l 2 can effectively facilitate SR tasks [33]. However, the l 2 typically optimizes for pixel-level mean values, which often leads to overly smoothed reconstructed results and a compromised preservation of fine-grained details. In contrast, the l 1 yields a more balanced error distribution across image pixels, thereby guiding the model to learn more accurate and detail-rich representations. Therefore, this study employs l 1 to measure the similarity between reconstructed images and ground-truth images, and its mathematical formulation is expressed as follows:
l 1 ( Θ ) = 1 N n = 1 N I S R n I H R n 1 ,
where N represents the total number of images in a training batch, Θ denotes the set of parameters of the network, I S R n refers to the n-th reconstructed high-resolution hyperspectral image, and I H R n corresponds to the n-th original high-resolution hyperspectral image. In designing the loss function for super-resolution reconstruction tasks, special attention must be paid to the correlation between spectral features in hyperspectral images to prevent spectral distortion. To address this, we incorporate the Spectral Angle Mapper (SAM) loss to enforce spectral consistency, which is formulated as follows:
l s p e ( Θ ) = 1 N n = 1 N 1 Π arc cos ( I S R n I H R n I S R n 2 I H R n 2 ) ,
Additionally, to further enhance structural details and edge information while preventing blurring effects, we introduce a gradient loss inspired by Wang et al. [34]. Gradient information plays a crucial role in improving image structural details, as it provides supplementary high-frequency information that enables the model to more effectively restore image sharpness. The gradient loss is formulated as follows:
l g r a ( Θ ) = 1 N n = 1 N M ( I S R n ) M ( I H R n ) 1 ,
where M ( ) denotes the operator for computing the image gradient map, which is obtained by calculating M ( H ) = ( h H , w H , l H ) 2 . The operators h , w , and l represent the gradient calculations along the horizontal, vertical, and spectral dimensions, respectively.
Finally, the total loss of the network can be expressed as follows:
l t o t a l ( Θ ) = l 1 + λ 1 l s p c + λ 2 l g r a ,
where λ 1 and λ 2 are the balancing parameters for the different loss terms, and the values of λ 1 and λ 2 are set to 0.5 and 0.1. This choice is informed by prior work [34] and is corroborated by our empirical analysis in Figure 10.

3. Results

3.1. Datasets

In this section, we evaluate the performance of the proposed method on three publicly available hyperspectral image datasets: CAVE [35], Harvard [36], and Chikusei [37].
CAVE dataset: This dataset comprises hyperspectral images of 32 distinct scenes, each accompanied by corresponding RGB images, capturing a wide range of real-world materials and objects. It consists of 31 spectral bands, spanning from 400 nm to 700 nm with 10 nm intervals, and a spatial resolution of 512 × 512 pixels.
Harvard dataset: The Harvard dataset contains hyperspectral images from 77 different scenes, including 50 indoor and outdoor environments and 27 indoor scenes under artificial or mixed lighting conditions.The dataset, it includes 31 spectral bands, covering the range from 420 nm to 720 nm with 10 nm intervals, and a spatial resolution of 1040 × 1392 pixels.
Chikusei dataset: Captured using the Hyperspec VNIR-C imaging sensor, this dataset includes hyperspectral imagery of agricultural and urban areas in Chikusei and Ibaraki, Japan. It features 128 spectral bands, spanning from 363 nm to 1018 nm, with a spatial resolution of 2517 × 2335 pixels and a ground sampling distance of 2.5 m.
The super-resolution (SR) performance was thoroughly assessed using a set of standard evaluation metrics: Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), Spectral Angle Mapper (SAM), Cross-Correlation (CC), Root Mean Square Error (RMSE), and the Relative Global Dimensionality Reduction Synthesis (ERGAS).

3.2. Implementation Rules

In this study, for both the CAVE and Harvard datasets, 80% of the samples were allocated for training, with the remaining 20% reserved for testing. To enhance the diversity of the training set, 24 image patches were randomly cropped from each image. The sample set was further augmented through three types of transformations: scale adjustments (applying scaling factors of 1, 0.75, and 0.5), rotational operations (rotating at 90°, 180°, and 270° angles), and horizontal/vertical flips. During preprocessing, bicubic downsampling was performed on images at scale factors of 2, 3, and 4 to generate low- resolution hyperspectral images with dimensions B × 32 × 32, where B denotes the number of spectral bands. For testing, to improve computational efficiency, only the top left 512 × 512 region of each test image was selected for evaluation. For the Chikusei dataset, due to the presence of invalid data in the edge regions, a central area of 2304 × 2048 × 128 pixels was first cropped for processing. Four non-overlapping images of size 512 × 512 × 128 were then extracted from the upper portion of the cropped region for testing, while the remaining area was used for training. When processing images at a scale factor of 4, image patches of 64 × 64 × 128 with a 32-pixel overlap were extracted; for a scale factor of 8, patches of 128 × 128 × 128 with a 128-pixel overlap were used. These patches were subsequently downsampled via bicubic interpolation to produce corresponding low resolution images at the specified scales.
In the proposed ESSTFormer network, for shallow spatial spectral feature extraction within the MSAM, dilated convolutions are employed with dilation rates set to 1, 3, and 5, respectively. Subsequently, spatial features are extracted using a 3 × 3 convolution kernel, while all other convolutions for channel expansion or shrinkage utilize 1 × 1 kernels. For deep spatial spectral feature extraction, the number of feature maps C in EETB and EATB is set to 240, and there are four consecutive EETB and EATB modules (see ablation study for details). Finally, a progressive upsampling strategy based on PixelShuffle is adopted to enlarge the spatial size of the input low-resolution hyperspectral image.
The network was trained using the Adam optimizer for 60 epochs, with a mini-batch size of 16 and an initial learning rate of 1 × 10−5. The model was implemented using the PyTorch 2.1.0, Meta Platforms, Inc., Menlo Park, CA, USA framework and trained on an NVIDIA RTX 4070 GPU (NVIDIA Corporation, Santa Clara, CA, USA).

3.3. Experimental Results and Analysis

3.3.1. Experimental Results on Chikusei

As shown in Table 1, we compared the proposed ESSTformer with several advanced methods on the Chikusei dataset and evaluated its performance under scale factors of ×4 and ×8 using four objective quantitative metrics. The comparative results demonstrate that ESSTformer outperforms other algorithms across all evaluation metrics. This superiority can be attributed to the fact that ESSTformer extracts local spatial spectral features via CNNs and further models long-range spatial spectral features separately by leveraging the heterogeneity of spatial and spectral features, thereby ultimately achieving an outstanding performance.
Specifically, although 3DFCNN [38] extracts spectral and spatial information through 3D convolutions, it fails to effectively capture critical spatial spectral features and suffers from a high computational complexity. MCNet [39] extracts the spatial spectral features of images by combining 2D and 3D convolutions, but still cannot capture global spatial spectral features due to the limitations of convolution kernels. SSPSR [40] adopts a grouping strategy, achieving a favorable performance in terms of spatial and spectral similarity. MSDFormer [41] also integrates CNN and Transformer architectures; despite its overall excellent performance, it is slightly inferior to ESSTformer in terms of PSNR and SAM metrics because it does not account for the heterogeneity between spectral and spatial features.
Qualitative results on the Chikusei dataset (Figure 6) demonstrate the advantage of our ESSTformer. CNN-based methods (3D-FCNN, MCNet, SSPSR) suffer from severe blurring due to limited receptive fields. A key differentiator is MSDFormer: while it uses a Spectral Transformer, it still relies on local convolutions for spatial detail recovery, restricting its global spatial modeling ability and leading to inconsistent spatial structure reconstruction (e.g., blurred field boundaries in Figure 6). Our ESSTformer addresses this with a decoupled dual-Transformer architecture: a dedicated Spatial Transformer explicitly captures long-range spatial contexts, working synergistically with the Spectral Transformer. This design enables ESSTformer to closely match the ground truth in both spatial sharpness (e.g., crisp river edges) and spectral consistency, outperforming MSDFormer and other methods.

3.3.2. Experimental Results on CAVE

As shown in Table 2, we compared the proposed ESSTformer with several advanced methods on the CAVE dataset and evaluated its performance under scale factors of ×2, ×3, and ×4 using three objective quantitative metrics. The results demonstrate that ESSTformer achieves an excellent performance across all evaluation metrics, significantly outperforming other comparative methods. Due to its adoption of progressive upsampling, SSPSR [40] exhibits limited performance when the scale factor is small.
For evaluation, we randomly selected one test image from the CAVE dataset, displayed its spectral band, and analyzed the corresponding absolute error maps (Figure 7), with the GT as the reference. It is evident that 3DFCNN, MCNet, and SSPSR exhibit prominent error regions (brighter areas), indicating significant deviations in edge and texture reconstruction. MSDFormer reduces errors to some extent but still has noticeable residual errors. In contrast, ESSTFormer shows the darkest error map, meaning it has the smallest deviation from GT, directly demonstrating its superiority in spatial detail restoration and reconstruction fidelity. Such performance gain originates from our decoupled spatial spectral Transformer architecture and gated convolution module, which enable a more precise modeling of spatial textures in hyperspectral images.

3.3.3. Experimental Results on Harvard

As shown in Table 3, the ESSTformer proposed in this paper outperforms other methods on the Harvard dataset.
We selected a test image and displayed its spectral band for evaluation. As shown in Figure 8, the absolute error maps of multiple methods are presented to compare the spatial reconstruction performance. It can be seen from the figure that the reconstructed images exhibit clearer edges and more realistic visual information.

3.4. Ablation Experiments

(1)
Effectiveness of the multi-scale feature extraction module: In the MSAM, we employed dilated convolutions with three different dilation rates to extract multi-scale features of the image. To verify the effectiveness of multi-scale features, we replaced the dilated convolutions with standard convolutions to extract single-scale spatial spectral features and named this variant “Ours w/o DConv”. As indicated in Table 4, all performance metrics exhibit a noticeable degradation, further highlighting the critical role of multi-scale structures in super-resolution image reconstruction.
(2)
Effectiveness of the EETB module: In the proposed ESSTformer, we use the EETB module to capture the global spectral features of the image, thus enhancing the model’s representational capability. To evaluate the effectiveness of the EETB module, we replace it with a network made up of CNN modules. Specifically, we substitute the EETB module with the CA module, which is commonly used in SR, and name this modified model “Ours w/o EETB” as presented in Table 4. From the results, all metrics show a significant decrease.
(3)
Effectiveness of the SE Module: In the EETB, we introduce the SE module. It compresses squeeze and excitation in the features of each spectral channel, enabling the network to adaptively focus on more important spectral channels and thus enhance the spectral features. To verify the effectiveness of the SE module, we design a control network with the SE module removed, named “Ours w/o SE”. As can be seen from the results in Table 4, after removing the SE module, the network performance degrades, which further proves the crucial role of the SE module in improving network performance.
(4)
Effectiveness of the EATB module: In the proposed ESSTformer, considering the heterogeneity between spatial and spectral information in images, we process long-range spatial and spectral features separately. Specifically, we designed the EATB module to capture the global spatial features of images. To evaluate the effectiveness of the EATB module, we replaced it with a network composed of 2D convolution modules and named this variant “Ours w/o EATB” in Table 4. The experimental results demonstrate that, after removing the EATB module, all evaluation metrics exhibit a significant decline.
(5)
Effectiveness of the CMLP module: In conventional Transformers, the feed-forward network (FFN) module fails to fully capture the local spatial information in hyperspectral images. To address this limitation, we propose the CMLP module by integrating convolutional layers. To verify its effectiveness, we replaced the CMLP module with a standard MLP and labeled this variant “Ours w/o CMLP” in Table 4. The experimental results indicate that, after removing the CMLP module, the PSNR decreases significantly, which further confirms the critical role of the CMLP module in capturing the local spatial information of images.
(6)
To extract global spatial spectral information, we incorporate N Transformer modules into the network. The experimental results are presented in Table 5. When fewer modules are used (N = 2), all quantitative metrics achieve the worst performance. When the number of Transformer modules increases to 4 (N = 4), metrics such as PSNR and SSIM improve significantly. However, with a further increase in the number of modules, the model complexity rises accordingly, leading to overfitting and subsequently a gradual decline in reconstruction performance.
(7)
To determine the optimal position of the SE module within the EETB framework, we tested two variants, named EETB1 and EETB2, respectively. EETB1 places the SE module outside the EETB, while EETB2 embeds the SE module into the self-attention computation, as shown in Figure 9. The comparative experimental results, presented in Table 6, demonstrate that EETB2 achieves a superior reconstruction quality. The reason for this is that EETB2 can more effectively dynamically enhance the spectral information of images during feature computation, facilitating information interaction and fusion across different bands. Through this embedded approach, the model can more accurately capture the complex relationships between spatial and spectral information, thereby achieving better performance in super-resolution tasks.
(8)
Following prior work [34], we set the initial loss weights, λ 1 = 0.5 and λ 2 = 0.1 . To confirm these hyperparameters for our task, we compared the PSNR values of different weight combinations on the Chikusei dataset. As shown in Figure 10, the PSNR indeed reaches its maximum when λ 1 = 0.5 and λ 2 = 0.1 , which validates this setting and led us to its adoption.

4. Conclusions

In this paper, we propose a model named ESSTformer, which integrates the MSAM, EETB, and EATB modules to fully leverage the advantages of CNN structures and Transformer architectures in extracting local and global spatial spectral features. Specifically, the MSAM, composed of a multi-scale convolution module, a Residual Spatial Module, and a Residual Channel Attention Module, is designed to extract the shallow spatial spectral features of images. The EETB module extracts long-range spectral features based on the self-attention mechanism and is supplemented with a spectral enhancement module to help the network focus on important spectral features. Considering the heterogeneity of spatial and spectral information in images, the EATB module adopts a window-based attention mechanism to effectively capture global spatial features while reducing the computational complexity. Unlike traditional MLPs, the CMLP component in EATB pays more attention to the spatial neighborhood information of images, which is beneficial for the restoration of image details. Finally, the effectiveness of each module is verified through extensive ablation experiments. Both qualitative and quantitative experimental results demonstrate that the proposed method outperforms the existing methods on three hyperspectral image datasets, especially under different scale factors.

Author Contributions

Conceptualization, H.L. and C.Y.; methodology, H.L.; software, H.L.; validation, H.L. and C.Y.; formal analysis, H.L.; investigation, Y.D. and Z.Z.; resources, C.Y.; data curation, J.L., Y.D. and Z.Z.; writing—original draft preparation, H.L.; writing—review and editing, C.Y.; visualization, H.L.; supervision, C.Y.; project administration, C.Y.; funding acquisition, C.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 62201457).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cao, R.; Fang, L.; Lu, T.; He, N. Self-Attention-Based Deep Feature Fusion for Remote Sensing Scene Classification. IEEE Geosci. Remote Sens. Lett. 2021, 18, 43–47. [Google Scholar] [CrossRef]
  2. Alhichri, H.; Alswayed, A.S.; Bazi, Y.; Ammour, N.; Alajlan, N.A. Classification of Remote Sensing Images Using EfficientNet-B3 CNN Model with Attention. IEEE Access 2021, 9, 14078–14094. [Google Scholar] [CrossRef]
  3. Chen, W.; Gao, Y.; Chen, A.; Zhou, G.; Wang, J.; Yang, X.; Jiang, R. Remote Sensing Scene Classification with Multi-Spatial Scale Frequency Covariance Pooling. Multimed. Tools Appl. 2022, 81, 30413–30435. [Google Scholar] [CrossRef]
  4. Wisotzky, E.L.; Hilsmann, A.; Eisert, P. 3D Hyperspectral Light-Field Imaging: A First Intraoperative Implementation. Curr. Dir. Biomed. Eng. 2023, 9, 611–614. [Google Scholar] [CrossRef]
  5. Wu, I.-C.; Chen, Y.-C.; Karmakar, R.; Mukundan, A.; Gabriel, G.; Wang, C.-C.; Wang, H.-C. Advancements in Hyperspectral Imaging and Computer-Aided Diagnostic Methods for the Enhanced Detection and Diagnosis of Head and Neck Cancer. Biomedicines 2024, 12, 2315. [Google Scholar] [CrossRef]
  6. Menon, S.; Trudgill, N. How Commonly Is Upper Gastrointestinal Cancer Missed at Endoscopy? A Meta-Analysis. Endosc. Int. Open 2014, 2, E46–E50. [Google Scholar] [CrossRef]
  7. Mishra, G.; Panda, B.K.; Ramirez, W.A.; Jung, H.; Singh, C.B.; Lee, S.-H.; Lee, I. Application of SWIR Hyperspectral Imaging Coupled with Chemometrics for Rapid and Non-Destructive Prediction of Aflatoxin B1 in Single Kernel Almonds. LWT-Food Sci. Technol. 2022, 155, 112954. [Google Scholar] [CrossRef]
  8. Sun, D.-W.; Pu, H.; Yu, J. Applications of Hyperspectral Imaging Technology in the Food Industry. Nat. Rev. Electr. Eng. 2024, 1, 251–263. [Google Scholar] [CrossRef]
  9. Gruber, F.; Wollmann, P.; Grählert, W.; Kaskel, S. Hyperspectral Imaging Using Laser Excitation for Fast Raman and Fluorescence Hyperspectral Imaging for Sorting and Quality Control Applications. J. Imaging 2018, 4, 110. [Google Scholar] [CrossRef]
  10. Barton, I.F.; Gabriel, M.J.; Lyons-Baral, J. Extending Geometallurgy to the Mine Scale with Hyperspectral Imaging: A Pilot Study Using Drone- and Ground-Based Scanning. Mining Metall. Explor. 2021, 38, 799–818. [Google Scholar] [CrossRef]
  11. Yuan, J.; Wang, S.; Wu, C.; Xu, Y. Fine-Grained Classification of Urban Functional Zones and Landscape Pattern Analysis Using Hyperspectral Satellite Imagery: A Case Study of Wuhan. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3972–3991. [Google Scholar] [CrossRef]
  12. Zhao, X.; Huang, J.; Gao, Y.; Wang, Q. Hyperspectral Target Detection Based on Prior Spectral Perception and Local Graph Fusion. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 13936–13948. [Google Scholar] [CrossRef]
  13. Hu, J.-F.; Huang, T.-Z.; Deng, L.-J.; Jiang, T.-X.; Vivone, G.; Chanussot, J. Hyperspectral Image Super-Resolution via Deep Spatiospectral Attention Convolutional Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 7251–7265. [Google Scholar] [CrossRef] [PubMed]
  14. Zheng, Y.; Li, J.; Li, Y.; Guo, J.; Wu, X.; Chanussot, J. Hyperspectral Pansharpening Using Deep Prior and Dual Attention Residual Network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8059–8076. [Google Scholar] [CrossRef]
  15. Dong, W.; Qu, J.; Zhang, T.; Li, Y.; Du, Q. Context-Aware Guided Attention Based Cross-Feedback Dense Network for Hyperspectral Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5530814. [Google Scholar] [CrossRef]
  16. Zhao, C.; Liu, H.; Su, N.; Yan, Y. TFTN: A Transformer-Based Fusion Tracking Framework of Hyperspectral and RGB. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5542515. [Google Scholar] [CrossRef]
  17. Wan, W.; Zhang, B.; Vella, M.; Mota, J.F.C.; Chen, W. Robust RGB-Guided Super-Resolution of Hyperspectral Images via TV3 Minimization. IEEE Signal Process. Lett. 2022, 29, 957–961. [Google Scholar] [CrossRef]
  18. Xu, Y.; Wu, Z.; Chanussot, J.; Wei, Z. Nonlocal Patch Tensor Sparse Representation for Hyperspectral Image Super-Resolution. IEEE Trans. Image Process. 2019, 28, 3034–3047. [Google Scholar] [CrossRef] [PubMed]
  19. Duan, Y.; Wang, N.; Zhang, Y.; Song, C. Tensor-Based Sparse Representation for Hyperspectral Image Reconstruction Using RGB Inputs. Mathematics 2024, 12, 708. [Google Scholar] [CrossRef]
  20. Gao, L.; Hong, D.; Yao, J.; Zhang, B.; Gamba, P.; Chanussot, J. Spectral Superresolution of Multispectral Imagery With Joint Sparse and Low-Rank Learning. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2269–2280. [Google Scholar] [CrossRef]
  21. Cao, M.; Bao, W.; Qu, K. Hyperspectral Super-Resolution Via Joint Regularization of Low-Rank Tensor Decomposition. Remote Sens. 2021, 13, 4116. [Google Scholar] [CrossRef]
  22. Xue, J.; Zhao, Y.-Q.; Bu, Y.; Liao, W.; Chan, J.C.-W.; Philips, W. Spatial-Spectral Structured Sparse Low-Rank Representation for Hyperspectral Image Super-Resolution. IEEE Trans. Image Process. 2021, 30, 3084–3097. [Google Scholar] [CrossRef] [PubMed]
  23. Hu, J.; Jia, X.; Li, Y.; He, G.; Zhao, M. Hyperspectral Image Super-Resolution via Intrafusion Network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7459–7471. [Google Scholar] [CrossRef]
  24. Zhang, J.; Liu, J.; Yang, J.; Wu, Z. Crossed Dual-Branch U-Net for Hyperspectral Image Super-Resolution. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2024, 17, 2296–2307. [Google Scholar] [CrossRef]
  25. Qi, W.; Huang, C.; Wang, Y.; Zhang, X.; Sun, W.; Zhang, L. Global–Local 3-D Convolutional Transformer Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5510820. [Google Scholar] [CrossRef]
  26. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A. SpectralFormer: Rethinking Hyperspectral Image Classification With Transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  27. Peng, Z.; Hou, Q.; Hu, J.; Xiao, B.; Torr, P.H.S.; Feng, J. Conformer: Local Features Coupling Global Representations for Recognition and Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 9454–9468. [Google Scholar] [CrossRef]
  28. Liu, Y.; Hu, J.; Kang, X.; Luo, J.; Fan, S. Interactformer: Interactive transformer and CNN for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531715. [Google Scholar] [CrossRef]
  29. Ma, Q.J.; Jiang, J.; Liu, X.M.; Ma, J.Y. Learning a 3D-CNN and Transformer Prior for Hyperspectral Image Super-Resolution. Inf. Fusion 2023, 100, 101907. [Google Scholar] [CrossRef]
  30. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar] [CrossRef]
  31. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. arXiv 2018, arXiv:1807.02758. [Google Scholar] [CrossRef]
  32. Sun, H.; Xu, J.; Meng, F.; Cheng, M.; Cao, Q. Spectral-Spatial Convolutional Hybrid Transformer for Hyperspectral Image Classification. IEEE Access 2025, 13, 59102–59117. [Google Scholar] [CrossRef]
  33. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. arXiv 2017, arXiv:1707.02921. [Google Scholar] [CrossRef]
  34. Wang, X.; Ma, J.; Jiang, J. Hyperspectral Image Super-Resolution via Recurrent Feedback Embedding and Spatial–Spectral Consistency Regularization. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5503113. [Google Scholar] [CrossRef]
  35. Yasuma, F.; Mitsunaga, T.; Iso, D.; Nayar, S.K. Generalized Assorted Pixel Camera: Postcapture Control of Resolution, Dynamic Range, and Spectrum. IEEE Trans. Image Process. 2010, 19, 2241–2253. [Google Scholar] [CrossRef]
  36. Chakrabarti, A.; Zickler, T. Statistics of Real-World Hyperspectral Images. In Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 193–200. [Google Scholar] [CrossRef]
  37. Yokoya, N.; Iwasaki, A. Airborne Hyperspectral Data over Chikusei; Technical Report SAL-2016-05-27; Space Application Laboratory, University of Tokyo: Tokyo, Japan, 2016. [Google Scholar]
  38. Mei, S.; Yuan, X.; Ji, J.; Zhang, Y.; Wan, S.; Du, Q. Hyperspectral Image Spatial Super-Resolution via 3D Full Convolutional Neural Network. Remote Sens. 2017, 9, 1139. [Google Scholar] [CrossRef]
  39. Li, Q.; Wang, Q.; Li, X. Mixed 2D/3D Convolutional Network for Hyperspectral Image Super-Resolution. Remote Sens. 2020, 12, 1660. [Google Scholar] [CrossRef]
  40. Jiang, J.; Sun, H.; Liu, X.; Ma, J. Learning Spatial-Spectral Prior for Super-Resolution of Hyperspectral Imagery. IEEE Trans. Comput. Imaging 2020, 6, 1082–1096. [Google Scholar] [CrossRef]
  41. Chen, S.; Zhang, L.; Zhang, L. MSDformer: Multiscale deformable transformer for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5525614. [Google Scholar] [CrossRef]
Figure 1. Overall architecture of the proposed ESSTformer method. (a) Enhanced Spatial Spectral Transformer (ESST), (b) Enhanced Spectral Transformer Block (EETB), (c) Enhanced Spatial Transformer Block (EATB).
Figure 1. Overall architecture of the proposed ESSTformer method. (a) Enhanced Spatial Spectral Transformer (ESST), (b) Enhanced Spectral Transformer Block (EETB), (c) Enhanced Spatial Transformer Block (EATB).
Applsci 15 11738 g001
Figure 2. The structure of the designed MSAM, which captures the shallow spatial spectral features of the image.
Figure 2. The structure of the designed MSAM, which captures the shallow spatial spectral features of the image.
Applsci 15 11738 g002
Figure 3. The structure of the designed MHSA, which is a key component of the EETB.
Figure 3. The structure of the designed MHSA, which is a key component of the EETB.
Applsci 15 11738 g003
Figure 4. The structure of the designed SE module, which emphasizes critical spectral information, facilitating superior reconstruction of image details.
Figure 4. The structure of the designed SE module, which emphasizes critical spectral information, facilitating superior reconstruction of image details.
Applsci 15 11738 g004
Figure 5. The structure of the designed CMLP, which is primarily designed to capture neighboring spatial information.
Figure 5. The structure of the designed CMLP, which is primarily designed to capture neighboring spatial information.
Applsci 15 11738 g005
Figure 6. Reconstruction images of different comparison methods on Chikusei dataset.
Figure 6. Reconstruction images of different comparison methods on Chikusei dataset.
Applsci 15 11738 g006
Figure 7. Absolute error maps of different comparison methods on CAVE dataset.
Figure 7. Absolute error maps of different comparison methods on CAVE dataset.
Applsci 15 11738 g007
Figure 8. Absolute error maps of different comparison methods on Harvard dataset.
Figure 8. Absolute error maps of different comparison methods on Harvard dataset.
Applsci 15 11738 g008aApplsci 15 11738 g008b
Figure 9. The structure of the EETB1 and EETB2.
Figure 9. The structure of the EETB1 and EETB2.
Applsci 15 11738 g009
Figure 10. PSNR of different combinations of loss weights on Chikusei dataset.
Figure 10. PSNR of different combinations of loss weights on Chikusei dataset.
Applsci 15 11738 g010
Table 1. Quantitative comparison of different methods on the Chikusei dataset.
Table 1. Quantitative comparison of different methods on the Chikusei dataset.
dPSNRSAMRMSECC
3DFCNN438.60913.11740.01400.9355
MCNet439.89502.46560.01120.9507
SSPSR439.99552.38640.01190.9530
MSDFormer440.09022.39810.01180.9539
Ours440.11382.36350.01170.9542
3DFCNN834.83754.84320.02150.8428
MCNet835.50494.27850.01190.8661
SSPSR835.16434.69110.02060.8560
MSDFormer835.59144.13810.01970.8693
Ours835.63144.07680.01950.8706
Table 2. Quantitative comparison of different methods on the CAVE dataset.
Table 2. Quantitative comparison of different methods on the CAVE dataset.
Metricsd3DFCNNMCNetSSPSRMSDFormerOurs
PSNR244.15445.092-45.20445.484
SAM22.4052.301-2.3332.359
SSIM20.97860.9838-0.98560.9876
PSNR340.21942.031-42.34542.589
SAM32.9302.809-2.6722.548
SSIM30.96530.9726-0.97540.9798
PSNR437.32639.20738.36639.627840.058
SAM43.3603.2923.4843.1773.019
SSIM40.95950.96540.96190.96720.9689
Table 3. Quantitative comparison of different methods on the Harvard dataset.
Table 3. Quantitative comparison of different methods on the Harvard dataset.
Metricsd3DFCNNMCNetSSPSRMSDFormerOurs
PSNR245.26446.213-46.64246.985
SAM21.7941.703-1.6131.644
SSIM20.98080.9853-0.98650.9878
PSNR342.38543.681-43.91244.151
SAM32.9301.987-1.9031.817
SSIM30.95800.9627-0.96530.9674
PSNR439.03740.54440.47440.63540.866
SAM42.1852.1082.2012.0142.001
SSIM40.92190.92670.92410.92710.9297
Table 4. Ablation studies on the Chikusei dataset.
Table 4. Ablation studies on the Chikusei dataset.
MethodParams.
10 6 )
FLOPs
10 9 )
PSNRSAMRMSECC
Ours w/o DConv15.872151.224340.03632.38890.01180.9533
Ours w/o EETB14.407348.043639.96692.39790.01180.9526
Ours w/o SE15.926552.104740.08092.39130.01170.9533
Ours w/o EATB14.856249.671439.99732.39260.01180.9528
Ours w/o CMLP15.764850.712939.84162.41960.01210.9512
Ours16.008153.381640.11382.36350.01170.9542
Table 5. Number of Transformer modules.
Table 5. Number of Transformer modules.
NumberPSNRSAMRMSECC
N = 239.91102.42340.01210.9520
N = 440.11382.36350.01170.9542
N = 640.05282.38010.01180.9534
N = 839.96182.38910.01190.9401
Table 6. Reconstruction results of SE at different positions.
Table 6. Reconstruction results of SE at different positions.
NumberPSNRSAMRMSECC
EETB140.10132.36830.01170.9539
EETB240.11382.36350.01170.9542
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, H.; Yi, C.; Liu, J.; Zhang, Z.; Dong, Y. ESSTformer: A CNN-Transformer Hybrid with Decoupled Spatial Spectral Transformers for Hyperspectral Image Super-Resolution. Appl. Sci. 2025, 15, 11738. https://doi.org/10.3390/app152111738

AMA Style

Li H, Yi C, Liu J, Zhang Z, Dong Y. ESSTformer: A CNN-Transformer Hybrid with Decoupled Spatial Spectral Transformers for Hyperspectral Image Super-Resolution. Applied Sciences. 2025; 15(21):11738. https://doi.org/10.3390/app152111738

Chicago/Turabian Style

Li, Hehuan, Chen Yi, Jiming Liu, Zhen Zhang, and Yu Dong. 2025. "ESSTformer: A CNN-Transformer Hybrid with Decoupled Spatial Spectral Transformers for Hyperspectral Image Super-Resolution" Applied Sciences 15, no. 21: 11738. https://doi.org/10.3390/app152111738

APA Style

Li, H., Yi, C., Liu, J., Zhang, Z., & Dong, Y. (2025). ESSTformer: A CNN-Transformer Hybrid with Decoupled Spatial Spectral Transformers for Hyperspectral Image Super-Resolution. Applied Sciences, 15(21), 11738. https://doi.org/10.3390/app152111738

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop