Next Article in Journal
Preliminary Results on Hydrogen Concentration Time Series in Spring Gases from the Pamir–Western Himalayan Syntaxis: Variability and Tectonic Instability
Previous Article in Journal
Unconventional Yeast in the Bakery Industry: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid-Domain Synergistic Transformer for Hyperspectral Image Denoising

by
Haoyue Li
and
Di Wu
*
School of Optoelectronic Science and Engineering, Soochow University, Suzhou 215000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(17), 9735; https://doi.org/10.3390/app15179735
Submission received: 29 July 2025 / Revised: 28 August 2025 / Accepted: 3 September 2025 / Published: 4 September 2025

Abstract

Featured Application

Hyperspectral denoising technology can improve the quality of remote sensing images and is applicable to fields such as environmental monitoring and precision agriculture.

Abstract

Hyperspectral image (HSI) denoising is challenged by complex spatial-spectral noise coupling. Existing deep learning methods, primarily designed for RGB images, fail to address HSI-specific noise distributions and spectral correlations. This paper proposes a Hybrid-Domain Synergistic Transformer (HDST) integrating frequency-domain enhancement and multiscale modeling. Key contributions include (1) a Fourier-based preprocessing module decoupling spectral noise; (2) a dynamic cross-domain attention mechanism adaptively fusing spatial-frequency features; and (3) a hierarchical architecture combining global noise modeling and detail recovery. Experiments on realistic and synthetic datasets show HDST outperforms state-of-the-art methods in PSNR, with fewer parameters. Visual results confirm effective noise suppression without spectral distortion. The framework provides a robust solution for HSI denoising, demonstrating potential for high-dimensional visual data processing.

1. Introduction

Image denoising, a traditional image processing technique, aims to recover clean information from observed signals and finds extensive applications in tasks such as remote sensing [1,2], target recognition [3], and 3D reconstruction [4]. Traditional methods such as wavelet transform [5], nonlocal means [6] rely on hand-crafted priors and are inadequate for handling complex noise distributions. In recent years, deep learning-based denoising models such as DnCNN [7] and SwinIR [8] have significantly improved performance through powerful image feature extraction capabilities.
These image denoising algorithms primarily focus on RGB images, which typically have only 3 channels and a simple structure. However, hyperspectral images (HSI), widely used in numerous scientific and applied fields, capture the spectral-spatial information of scenes through hundreds of consecutive narrow spectral bands, significantly expanding the dimension and capability of information expression. Since the late 1960s, hyperspectral imaging technology has been mainly applied in areas such as agriculture, mineral exploration, and satellite remote sensing monitoring [9]. Typical applications include crop health assessment, mineral identification, and land use classification based on satellite and airborne systems [9]. In recent years, with the continuous advancement of sensor technology and analytical methods, the applications of hyperspectral images have expanded to multiple cutting-edge fields, including agriculture, military [10], art and cultural heritage protection [11], food safety, and medical diagnosis [12]. For instance, in geophysical exploration, HSIs are used to invert the composition of surface materials [13]; in heritage science, they assist in the non-destructive identification of art materials and the restoration of historical documents [11]; moreover, in security and industrial testing, HSIs are also widely applied in chemical identification, material classification [14], and pollutant diffusion tracking. These applications demonstrate the unique advantages of hyperspectral images in acquiring information from complex scenes, while also placing higher-dimensional and more refined requirements on the research of HSI-related algorithms. Traditional algorithms cannot be directly applied to HSI denoising due to their neglect of channel correlations [15]. Compared to RGB images, HSI denoising presents the following unique challenges:
  • Coupled multidimensional noise and cross-band correlations: Noise is non-uniformly distributed across spatial dimensions and complexly coupled with spectral dimensions, exhibiting inter-band dependencies [16].
  • Sensitivity to high-frequency information: Inconsistent imaging equipment causes HSIs to frequently have bit depths different from RGB. Weak spectral features are easily drowned out by noise, requiring fine-grained frequency-domain guidance [1].
  • Detail preservation and computational efficiency: Hundreds of channels substantially increase computational complexity. Traditional methods lack efficiency and capability in jointly modeling local spatial details and global spectral continuity [17].
Traditional HSI denoising methods based on frequency domain processing primarily utilize Fourier transform and wavelet transform techniques [18,19]. The Fourier transform distinguishes noise and signal distributions through frequency domain conversion, where low-pass filters can suppress high-frequency noise but tend to lose image details while showing limited effectiveness for complex noise. Similarly, wavelet transform decomposes HSI into sub-bands of different frequencies and achieves denoising by processing coefficients within these sub-bands [5,20]. When noise energy dominates certain sub-bands, thresholding can be applied to suppress noise components. However, this approach faces critical challenges in selecting appropriate wavelet basis functions and determining optimal thresholds, where improper threshold selection may cause over-denoising (resulting in image blurring) or under-denoising (leaving significant noise) [20]. These traditional methods primarily consider frequency domain perspectives without fully integrating spatial and spectral characteristics of HSI, making them increasingly unsuitable for modern high-quality HSI processing requirements.
Early attempts to apply deep learning to HSI denoising encountered significant challenges. For instance, Xie and Li [21] pioneered the use of deep CNNs with trainable nonlinear functions for this task, but extending 2D models to 3D HSI data faced obstacles including under-utilization of spatial-spectral correlations, high computational complexity, and insufficient training data [22]. To address these limitations, Dong et al. developed a 3D denoising framework using an improved U-net architecture [22]. Their approach employed separable filtering that decomposes 3D operations into 2D spatial and 1D spectral components to reduce complexity, supplemented by transfer learning to mitigate data scarcity, ultimately outperforming model-based methods. Recent innovations have further advanced deep learning approaches: Liang et al.’s HSSD [23] combines CNN and Transformer architectures with a decoupling strategy to capture both local/nonlocal details and global spectral correlations, achieving superior spatial-spectral reconstruction, while Fu et al.’s GMSC-Net [24] adopts a model-based approach using sparse representation and iterative optimization, addressing the lack of clean-noisy pairs through clustering techniques.
In recent years, frequency domain analysis, when integrated with deep learning, has demonstrated unique advantages in image restoration tasks [25,26], leveraging frequency transformations to decouple noise and signal spectra while utilizing the nonlinear modeling capabilities of deep networks. Successful applications in RGB denoising include Fourier convolutions in FADformer [27] and cross-domain filtering in FCENet [28], while other exemplary approaches span tasks like deraining—such as Gao et al.’s contrastive regularization-based frequency domain method for improved feature distinguishability [27]—deblurring via Kong et al.’s Frequency-domain Self-Attention Solver (FSAS) with Discriminative Frequency-domain Feed-forward Network (DFFN) that applies convolution theorem principles [29], and adaptive fusion of NIR and RGB images through Wang et al.’s FCENet to exploit frequency domain complementarity [30]. Cross-domain fusion frameworks have also gained traction, as seen in Zhang et al.’s FS-Net, which separates frequency components of underwater images combined with spatial filtering for contrast enhancement [25], alongside HSI-specific implementations like Wang et al.’s UOANet [31] and Sheng et al.’s SANet [32] that optimize denoising through spectrum-spatial fusion and frequency-guided processing—though the adaptation of earlier RGB-focused methods to HSI remains largely unexplored.
Complementary to frequency approaches, Atrous Spatial Pyramid Pooling (ASPP) [33] captures multi-frequency noise components through multiscale atrous convolutions, with its feature extraction capabilities validated in diverse tasks like semantic segmentation [34]. This module finds wide application in image processing; for example, ASPP-DF-PVNet [35] enhances model generalizability by capturing multi-level details. When combined with frequency analysis, ASPP’s multi-dilation design can be mapped to multi-band frequency processing, enabling specific separation of periodic stripe noise spectra (characterized by low-frequency directional energy [33]) from high-frequency random sensor noise in HSI. This integration provides physically meaningful frequency priors for cross-domain feature fusion, though effectively combining frequency analysis with HSI’s spatial-spectral duality requires further investigation.
To this end, this paper proposes an HSI denoising framework based on frequency domain enhancement and multiscale modeling, achieving spatial-frequency-channel three-dimensional collaborative processing. Specifically, building upon Spectral Enhanced Rectangle Transformer (SERT) [36], it innovatively deeply integrates frequency domain noise decoupling with spatial multiscale modeling: (1) Introducing a fast Fourier transform (FFT) preprocessing module to transform HSI into the frequency domain for noise separation, combined with ASPP multiscale atrous convolution [33] to suppress noise components of different frequencies; (2) designing a dynamic cross-domain attention module to adaptively fuse spatial-frequency features through a learnable gating mechanism, such as enhancing noise suppression in low-frequency regions dominated by the frequency domain and preserving ground object textures in high-frequency regions dominated by the spatial domain; and (3) constructing a shallow global-deep local hierarchical architecture, utilizing the global average pooling branch of ASPP to capture noise statistical characteristics and guide detail recovery in deep networks. Experiments show that this framework improves the quality of HSI restoration in mixed noise scenarios while maintaining the efficiency of SERT [36] to a certain extent. This framework has verified its effectiveness in HSI denoising tasks, and it may be extended to a universal post-processing framework in the future. In general, our contributions can be summarized as follows:
  • Propose a spatial-frequency-channel three-dimensional collaborative dynamic cross-domain feature fusion framework, combining spatial domain Transformer with FFT-based frequency domain feature extraction in a post-processing manner. It realizes adaptive weight allocation of dual-domain features through a learnable gating mechanism, enabling cross-domain attention for dual-domain collaborative processing to address the coupling of HSI noise in spatial, spectral, and frequency dimensions.
  • Firstly, introduce spatial pyramid pooling into the joint processing of both spatial and frequency domains; while constructing a hierarchical denoising chain of global statistical guidance-local detail recovery, it forms cross-denoising capabilities of spatial multi-resolution + frequency multi-band.
  • Compared to the baseline, our model achieves a PSNR improvement of 0.94 dB on Realistic dataset and a maximum PSNR improvement of 0.52 dB on ICVL synthetic dataset.

2. Methods

Spatial domain methods are widely used in image restoration and can generally achieve relatively good denoising levels [37]. To enable our proposed frequency domain theoretical framework to be transplanted and applied to a large number of spatial domain denoising models, we integrate the idea of post-processing into the frequency domain theoretical framework while ensuring that the overall input and output of this framework maintain the same dimensional size, allowing information captured in the frequency domain to assist in improving the overall performance of the network.
The overall framework of the proposed method is shown in Figure 1, which is based on the Spectral Enhanced Rectangle Transformer (SERT) [36]. To further reduce high-frequency signals caused by noise, this paper proposes a frequency post-processing (FPP) module based on both frequency and spatial domains, along with a learnable gating mechanism to achieve adaptive weight allocation of dual-domain features.

2.1. Review of SERT

Spectral Enhanced Rectangle Transformer (SERT) [36] employs 3 Residual Transformer Layers (RTL), each cascading 6 transformer blocks. Its core modules include (1) the Rectangular Attention (RA) module, which realizes multidirectional nonlocal interaction through horizontal and vertical rectangular window partitioning and spectral shuffling; and (2) the Spectral Enhancement (SE) module, which utilizes a global memory unit to store spectral priors and enhances spectral-spatial correlations while suppressing noise through a dynamic low-rank selection mechanism. These two modules work synergistically to effectively address the issues of multidimensional noise coupling and cross-band correlations in HSI, integrating the spatial and channel information of the image.
However, effective information in HSI is mostly concentrated in low- and medium-frequency regions, while random noise manifests itself as random fluctuations in pixel values in the spatial domain, and its energy is often widely distributed in high-frequency parts when transformed to the frequency domain. Therefore, SERT cannot effectively distinguish between stripe noise and edge textures in HSI space, leading to loss of image details after denoising.

2.2. Hybrid-Domain Synergistic Transformer Network (HDST)

This framework takes RTL as the basic skeleton and innovatively introduces a frequency domain post-processing unit and a hybrid domain multiscale module, constructing a closed-loop optimization process of “spatial feature extraction–frequency domain noise decoupling–cross-domain information calibration”. The loop is shown in Figure 2.
Specifically, hierarchical domain adaptation is performed for the cascaded Transformer structure within RTL. Shallow Transformers capture long-range spatial dependencies and low-rank spectral characteristics through the RA and SE modules, laying the foundation for feature representation; FPP units are embedded in the deepest two layers of Transformers to decouple noise spectral components through frequency domain transformation, avoiding blurring of shallow details due to premature frequency domain processing.
Within the FPP unit, through a three-stage chain of frequency domain multiscale filtering–cross-domain feature calibration–dynamic residual fusion, frequency domain noise patterns are converted into guiding signals for spatial domain denoising. Through the frequency-spatial collaborative attention mechanism, frequency domain noise priors such as specific frequency noise patterns are fed back to spatial domain feature calibration, and then adaptive control of feature enhancement is achieved through learnable residual connections, ultimately forming a complete optimization chain of “noise decoupling–feature calibration–error compensation”.
Meanwhile, multiscale cross-enhancement is widely used to improve feature extraction capability. After each Transformer block, the HDMS module is used for multiscale sampling, covering multilevel information from edge details to global structures, providing a richer feature base for frequency domain processing; ASPP-FFT located within FSGF realizes refined separation of wide-band noise in the frequency domain through multi-dilation rate atrous convolution, effectively achieving two-dimensional multiscale analysis.

2.2.1. Atrous Spatial Pyramid Pooling (ASPP)

Atrous Spatial Pyramid Pooling (ASPP) is a module proposed in the DeepLab system [33]. It contains the following 5 parallel branches:
1 × 1 convolution ( preserve resolution ) Dilated conv ( dilation = [ 6 , 12 , 18 ] ) ( multiscale context ) Global Avg Pool 1 × 1 Conv Upsample ( global statistics ) .
The features are concatenated and fused by 1 × 1 convolution. The structure is also shown in the ASPP Reference section of Figure 1, with dilation rates [6,12,18].
Innovatively, we adapt this spatial domain module for frequency domain processing by employing multiscale frequency domain convolutions with dilation rates [2,4,8], and we name this module ASPP-FFT. This enables the extraction of both local frequency-band correlations and cross-band global contextual information, which subsequently drives the generation of frequency-guided spatial gating signals. Since the frequency domain inherently represents global information, the redundant global statistics branch is removed to avoid disturbance.
1 × 1 convolution ( preserve resolution ) Dilated conv ( dilation = [ 2 , 4 , 8 ] ) ( multiscale context ) .

2.2.2. FFT-Scale Gated Fusion (FSGF)

To achieve precise decoupling and adaptive fusion of noise in the frequency domain and spatial domain, we perform FFT on the input spatial feature S, separate the real and imaginary parts, and concatenate them into F c . Multiscale feature extraction in the frequency domain is achieved through the ASPP-FFT module. This enables wide-band noise filtering in the frequency domain, forming a spatial-frequency dual-dimensional multiscale analysis. Its forward propagation can be expressed as
F = F F T ( S ) , F c = C o n c a t ( F . r e a l , F . i m a g ) ,
F p r o c = ASPP-FFT ( F c , d i l a t i o n = [ 2 , 4 , 8 ] ) ,
where S represents the input spatial domain information, F is the frequency domain result obtained after a fast Fourier transform, and the feature F c , formed by concatenating the real and imaginary parts of F, is used for multiscale ASPP-based analysis. The combination of F c fully preserves complex phase information and adapts to real-valued convolution, enabling noise frequency band separation and alignment in the domain.
Subsequently, a dynamic gating mechanism generates a spectral mask G a t e , which adaptively selects the fusion path based on noise intensity. The hyperparameter α is used to control the injection intensity of the information from the frequency domain. This process is expressed by the following formulas:
S = C o n v 3 × 3 ( I F F T ( F p r o c ) ) ,
G a t e = σ ( C o n v 1 × 1 ( F p r o c ) ) ,
F = S G a t e + α · S ( 1 G a t e ) ,
where S is the spatial feature reconstructed from the frequency domain processing result, and 3 × 3 convolution denotes the reconstruction convolution operation; ⊙ represents element-wise multiplication; σ represents the sigmoid activation function; Gate retains target areas ( Gate 1 ) and replaces noisy regions ( Gate 0 ).

2.2.3. Frequency-Spatial Collaborative Attention (FSCA)

To establish cross-domain correlations between frequency-domain noise priors and spatial-domain features, the frequency-domain fusion features output by FSGF and the original input features are divided into non-overlapping windows, and cross-domain associations are established through a multi-head attention mechanism: using spatial features as queries (Q) and frequency-domain features as keys (K) and values (V), realizing the interaction logic of spatial positions querying their corresponding frequency-domain noise patterns. The spatial map S and the map F generated by combining original and reconstructed features with the mask are divided into M × M non-overlapping windows and flattened to reduce computational complexity. Then, unbiased multi-head attention calculation is performed on the flattened sequences, expressed as
Q = S f l a t W Q T , K = F f l a t W K T , V = F f l a t W V T ,
X i = X [ : , : , i C h : ( i + 1 ) C h ] , X { Q , K , V } ,
A t t n H e a d i = S o f t m a x Q i · K i T C h · V i ,
H = C o n c a t ( A t t n H e a d i ) · W O T ,
where W Q , W K , W V , and W O are learnable weight matrices, projecting input features to query, key, value, and output spaces; C h represents the dimension of a single attention head, and the multi-head attention mechanism is implemented by dividing features by dimension; Q i , K i , and V i represent the query, key, and value vectors of the i-th head.

2.2.4. Dynamic Fusion

Residual calculation is performed on the calibrated feature H output by FSCA through a learnable parameter β and a lightweight convolution fusion network, achieving dynamic fusion of “original features + frequency domain enhanced residuals”. Its forward propagation can be expressed as
H f u s e d = C o n v 3 × 3 ( G E L U ( C o n v 1 × 1 ( H ) ) ) ,
O u t p u t = S + β · H f u s e d ,
where H f u s e d is the residual output enhanced by the fusion network. The 3 × 3 convolution restores dimensions and introduces spatial context information and continuity constraints, avoiding block effects that may arise from frequency domain reconstruction; β is a learnable fusion coefficient initialized to 0.1, ensuring stability in the early training stage and intensity of enhancement after convergence, preventing overshoot of features or loss of information.

2.2.5. Hybrid-Domain Multiscale Module (HDMS)

As a basic component throughout each RTL layer, the HDMS module performs multiscale atrous convolution with dilation rates [2,4,6] on spatial features after each Transformer block. Small dilation rates capture edge and texture details, and larger dilation rates aggregate target structures and globally associated features. Through multi-branch parallel sampling, it compensates for the deficiency of single-scale convolution in handling complex noise, providing more stable feature input for subsequent frequency domain processing and significantly enhancing feature robustness.
Shallow HDMS strengthens the noise resistance of local details, while deep HDMS provides spatial feature bases containing multiscale semantics for FPP units, enabling frequency domain noise decoupling to more accurately adapt to content structures at different levels, achieving cross-layer information paving.

3. Results

To evaluate the proposed HDST framework, experiments are conducted against model-based and deep-learning hyperspectral denoising methods. Further ablation studies and computational efficiency analyses are performed to validate the contribution of key components and the practical viability of HDST.
Today, deep learning denoising combined with the frequency domain has been involved in RGB image processing [38]. However, HSI denoising methods that integrate deep learning with frequency domain processing are relatively rare. When applying related RGB image processing methods to HSI image denoising tasks, the increase in the number of channels will lead to potential performance degradation. Therefore, the method of comparative experiments is still based on comparison with commonly used hyperspectral denoising methods.
The comparison methods include several traditional model-based hyperspectral image denoising methods, such as the filter-based method (Block-Matching 4D algorithm (BM4D [39])), the orthogonal basis-based method (Non-local Graph Matching (NGMeet [40])), and the wavelet-based method (3D wavelets [41]); also included are three deep learning-based methods, namely Hyperspectral Image Denoising Convolutional Neural Network (HSID-CNN [42]), Multi-Attention Fusion Network (MAC-Net [43]), and a relatively new spatial-spectral recurrent transformer U-Net (SSRT-UNet [44]).
The proposed model is implemented using Pytorch 2.4.1, and experiments are carried out on a server equipped with an Intel Xeon Processor CPU and a GeForce RTX 3090 GPU running Ubuntu 22.04. This server is located in Inner Mongolia, China, and was assembled by AutoDL.
We selected a widely used metric from each three perspectives: image restoration, visual perception, and spectral similarity to fully characterize the performance of the model.
  • PSNR [45] (Peak Signal-to-Noise Ratio) focuses on pixel-level errors, calculating the mean square error between the original and distorted images, comparing it with the square of the maximum signal value, and expressing quality in decibels (larger values mean less distortion).
    PSNR = 10 log 10 L 2 MSE , MSE = 1 N i = 1 N ( x i y i ) 2 .
  • SSIM [45] (Structural Similarity) emphasizes image structure and perceptual quality, evaluating similarity via structure, luminance, and contrast. Its evaluation results are more consistent with human visual perception. Ranging 0–1, values closer to 1 indicate higher quality.
    SSIM ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 ) .
  • SAM [17] (Spectral Angle Mapping) focuses on spectral similarity, measuring it via the angle between spectral vectors; smaller angles mean more similar spectra and better quality.
    SAM = arccos x s p e c , y s p e c x s p e c 2 · y s p e c 2 .
This paper introduces the L2 loss function [46] for network training. Let the noisy data be x and the corresponding ground truth be y; the loss function can be expressed as:
arg min f θ f θ ( x ) y 2 2 .

3.1. Performance on Realistic Dataset

For the Realistic dataset [47], our experimental method is generally consistent with the training of the original SERT [36] network. The data set contains 59 noisy hyperspectral images and their corresponding clean, noise-free hyperspectral images. We followed the SERT division method, using the same 44 images and 15 images for training and testing. Each hyperspectral image channel contains 34 bands, ranging from 400 to 700 nm, and the image size within a single channel is 696 × 520. Overlapping 128 × 128 spatial regions are cropped, and data augmentation is performed for training [47]. Sufficient training can be achieved by setting the number of training epochs to 500. The learning rate is set to 1 e 4 for the first 200 epochs, changed to 5 e 5 at 200 epochs, and set to 1 e 5 at 400 epochs.
Table 1 shows the average results of different methods on the Realistic dataset. Our proposed HDST method significantly outperforms SERT [36] as the baseline and several other hyperspectral image denoising methods, with a maximum PSNR improvement of 0.94 dB, indicating the effectiveness of our method in handling real noise. In terms of inference time, the average processing time of our proposed HDST method on the Realistic dataset is only 0.607 s, demonstrating excellent computational efficiency. This time is significantly superior to that of traditional methods and also competitive among deep learning methods, being faster than MAC-Net and SSRT-UNet. Although the inference time of HDST is slightly longer than that of its baseline SERT by 0.056 seconds, this minimal cost is exchanged for a significant improvement in image quality. This reflects the good balance achieved by the model between performance and efficiency and inherits the lightweight architectural design of SERT.
We also show the denoising results of the above real noisy hyperspectral images in Figure 3. In terms of noise removal and detail preservation, the results of our improved method are superior to traditional denoising methods and deep learning methods. Note that the noisy images in these datasets generally feature lower brightness than the GroundTruth images. Therefore, during the model training process, to reduce L2 loss, the algorithm may appropriately increase the saturation of the reconstructed images to make them closer to the GroundTruth. In addition, since all HSI images used in these datasets are in 16-bit format, their high numerical values appear as white, barely distinguishable to the naked eye on a grayscale display, and this exceeds the color rendering capability of most monitors. Thus, the white areas in the visualization results do not indicate local overexposure or reconstruction failure.

3.2. Performance on ICVL Synthetic Dataset

For the ICVL synthetic dataset [48,49], which contains 201 HSIs with a size of 1392 × 1300, each hyperspectral image channel contains 31 bands, ranging from 400 to 700 nm. We followed the SERT [36] division method, using the same 100 images, 5 images, and 50 images for training, validation, and testing. Training images are cropped to a size of 64 × 64 at different ratios. In the testing phase, HSI is cropped to 512 × 512 × 31 to obtain affordable computational costs. The number of training epochs is set at 100, with the learning rate set to 1 e 4 for the first 50 epochs and changed to 1 e 5 for the last 50 epochs.
The noise setting method partially follows the baseline settings, using the noise patterns in the literature [48] for simulation. The noise patterns are independently and identically distributed Gaussian noise with random settings between levels 10 and 70, non-independently and identically distributed Gaussian noise (non-i.i.d Gaussian noise), Gaussian + Stripe noise, Gaussian + Deadline noise, Gaussian + Impulse noise, and Mixture noise, which contains stripe, deadline, and impulse noise.
Table 2 presents the comparison results. The proposed HDST method also outperforms SERT [36] as the baseline among the noise methods selected. In the case of small amounts of structured noise, some of the latest comparative methods perform excellently, even surpassing our proposed model; however, HDST demonstrates significant comprehensive advantages in mixed noise scenarios, enabling it to better adapt to denoising complex noise patterns such as those in real-world scenes.
Figure 4 presents the visualized comparison results. Our improved method has achieved relatively ideal visual effects, outperforming denoising methods with fewer noise points in some areas.

3.3. Ablation Study and Analysis

3.3.1. Module Interaction Analysis

To verify the theoretical analysis and understand the actual role of modules, in addition to the baseline and the proposed model, we additionally set up four groups of comparative models for experiments. Comparative Network 1 only adds frequency domain processing (FSGF + FSCA) to test the independent role of the frequency domain module. Comparative Network 2 adds dynamic fusion based on Network 1 to test the tuning ability of residual gating. Comparative Network 3 only adds multiscale pyramids (HDMS) to test the collaboration between multiscale and frequency domain ASPP. Comparative Network 4 includes frequency domain processing (without dynamic fusion) + multiscale pyramids (ASPP-FFT and HDMS) to analyze the bottleneck of cross-domain feature fusion and assist in verifying the collaborative role of modules. The performance results of each comparative network on the Realistic dataset and its corresponding training environment are shown in Table 3.
The experimental results of Net1 and Net2 show that the processing of the independent frequency domain module may destroy spatial continuity, leading to a slight increase in parameters such as PSNR; while under the correction of the dynamic fusion module, the frequency domain module can achieve effective improvement; the experimental results of Net3 show that the pure multiscale module brings a large single-module gain, verifying that multiscale modeling is a key pillar for performance improvement; the experimental results of Net4 show that the combination of frequency domain and multiscale methods achieves great improvement, proving that multiscale feature extraction provides a frequency spectrum decomposition basis with clear physical meaning for frequency domain processing.
The incremental growth in parameters of the entire ablation network (from Net1 to Net4) also confirms that frequency domain processing (0.34 M), dynamic fusion (0.28 M), and ASPP (0.55 M) contribute to the parameter count in a relatively balanced manner. This benefit stems from the low-parameter design approach adopted for each module, which effectively controls the complexity of every individual module.
Both theoretical analysis and the above experiments show that multiscale methods can effectively collaborate with frequency domain methods, and the design of each module in the improved model is reasonable.

3.3.2. Multiscale Dilation Rate Comparison

To investigate the differences in the impact of dilation rate configurations in the multiscale module on spatial domain and frequency domain processing, we designed six groups of comparative experiments with different dilation rates for HDMS and FSGF, respectively. Specifically, we tested classic spatial pyramid configurations such as [2,4,6], [4,8,12], and [6,12,18], as well as unconventional configurations such as [1,3,5] and [2,4,8]. Meanwhile, an extremely large dilation rate [16,32,32] was established as a boundary condition. In all experiments, the parameters of other modules were kept consistent on Realistic dataset. As shown in Table 4, we recorded the deviation values of each configuration relative to the optimal PSNR. This presentation method intuitively reflects the sensitivity of different scale combinations to the two heterogeneous modules, providing data support for subsequent domain-specific optimization strategies.
As can be seen from the data in Table 4, HDMS and FSGF exhibit differences in their sensitivity to dilation rates. For HDMS, its optimal dilation rate configuration is [2,4,6], with a PSNR difference of 0.04 dB compared to the suboptimal configuration [2,4,8]. This slight gap indicates that spatial domain modeling needs to align closely with local structural features. Specifically, when using small dilation rates [1,3,5], the PSNR decreases by 0.09 dB due to insufficient receptive fields; while excessively large dilation rates [16,32,32] disrupt local continuity, leading to a 0.2 dB reduction in PSNR.
In contrast, the optimal dilation rate configuration for FSGF is [2,4,8], suggesting that the frequency domain module requires a relatively wider receptive field to achieve effective separation of broadband noise. When dilation rates reach [16,32,32], the loss of high-frequency details caused by over-smoothing also verifies the global characteristics of the frequency domain module.
The technical reason for this difference lies in the fact that ASPP-FFT in the frequency domain needs to balance frequency resolution, where small dilation rates are advantageous, and noise separation capability, where large dilation rates are more beneficial. However, HDMS emphasizes texture preservation, dominated by small dilation rates. This phenomenon echoes the design innovation of “domain-specific multiscale design”, confirming the rationality and effectiveness of differentiated designs tailored to the characteristics of different domains.

3.4. Model Computational Efficiency Analysis

3.4.1. Complexity-Parameter Trade-Off

We conducted statistics on the number of parameters and the operation speed on the Realistic dataset and its corresponding training environment, and the statistical results are shown in Table 5. During the upgrade of the baseline structure, we observed that the number of parameters increased from 1.91 M to 3.08 M, an increase of 61%, while the amount of computation (GFLOPS) increased only approximately 14%.
The core reason is that the newly added modules adopt a parameter-intensive but computationally efficient design strategy, achieving decoupling of the two. The multiscale method uses multi-branch 1 × 1 convolution to build a spatial pyramid, whose parameters are concentrated in the channel projection layer, but the amount of computation is compressed through depth-wise separable convolution; the frequency domain processing module introduces fully connected layer parameters through real and imaginary part separation and dynamic gating but controls the computational increment using the zero-parameter characteristics of the FFT transform and dimensionality reduction strategy. Finally, the cross-domain attention mechanism limits the attention range through window partitioning.
The asymmetric growth of parameters and computation reflects the idea of “trading parameters for efficiency”, i.e., enhancing the model’s expressive ability by adding lightweight parameter modules and suppressing computational overhead through optimization, improving model capacity while maintaining computational efficiency.

3.4.2. Cross-Model Parameter-Performance Comparison

To demonstrate the advantages of the HDST model in balancing parameter efficiency and performance, we selected three of the most representative models for comparative analysis: the lightweight convolutional network MAC-Net [43], the Transformer-based SSRT-UNet [44], and the Transformer model SERT [36]. We statistically analyzed the parameter counts and performance metrics of the above models, compiled the data into Table 6, and plotted them in the model complexity-performance relationship graph in Figure 5.
In terms of efficiency advantages, as demonstrated in Table 6 and Figure 5, HDST achieves a notable PSNR with a relatively small number of parameters, demonstrating remarkable superiority over existing models. Compared to the lightweight MAC-Net [43], which with 0.43 M parameters attains a PSNR of 29.2 dB, HDST delivers a significant performance gain of 1.42 dB, clearly proving the effective conversion of its increased parameters into tangible performance improvements. Compared with the more parameter-heavy SSRT-UNet [44], HDST also achieves a better restoration quality while using 30 % of SSRT-UNet’s parameter count.
Figure 5 uses a scatter plot to compare the parameter counts and denoising performance (PSNR) of different models, clearly showing that HDST has achieved an optimal balance in terms of parameter efficiency (Pareto optimum), i.e., it obtains leading performance with relatively low model complexity.
This advantage is an intuitive confirmation of the complexity analysis in Section 3.4.1 and is fundamentally derived from the inherent efficiency of the hybrid-domain collaborative design paradigm proposed in this paper. The FPP module leverages the parameter-free nature of FFT/IFFT to achieve powerful noise decoupling capabilities, and subsequent lightweight dynamic gating and convolutional layers further bring significant performance gains. HDMS and ASPP-FFT adopt parameter-efficient depthwise separable convolutions and a multi-branch design, which greatly enrich the feature expression capability while moderately controlling the total number of parameters. In addition, FSCA constrains computational complexity by window partition. This collaborative design strategy of “parameter dense yet computationally efficient” collectively enables HDST to achieve a disproportionately large improvement in PSNR with a 61 % increase in parameter count, thus occupying the advantageous position at the lower right of Figure 5.
Hyperspectral imaging often relies on platforms with limited on-device computing resources, such as satellites and unmanned aerial vehicles (UAVs). The framework, characterized by high performance, a moderate number of parameters, and high inference efficiency, exhibits excellent potential for practical application and engineering deployment. Through port-based packaging and model pre-training, this model can be applied to numerous scenarios, including remote sensing-based precision agriculture, medical image enhancement, and industrial vision. For embedded or real-time systems with constrained computing resources, such as satellite and airborne platforms, model pruning and quantization can be employed to further reduce their memory and computational overhead, thereby meeting the stringent requirements of practical applications.

4. Discussion

Analysis of experimental results across different datasets reveals that the performance improvement of the HDST model on the Realistic dataset (real-world data) is significantly superior to that on the ICVL synthetic dataset. This discrepancy underscores a profound connection between the frequency-domain processing mechanism and the physical characteristics of noise. From the perspective of hyperspectral imaging physics, real HSI noise exhibits typical spatial non-uniformity and frequency-domain energy aggregation, which precisely align with the core innovation of HDST: multiscale spatial-frequency collaboration. This alignment between physical properties and algorithmic design amplifies the model’s advantages in real-world scenarios.
Further analysis indicates that the excellent performance in complex noise scenarios stems from the cascaded enhancement effect of multiscale spatial-frequency features. Specifically, the shallow HDMS module employs a combination of small dilation rates [2,4,6], simulating the local perception mechanism of the human visual system to preserve critical edge information for subsequent processing. In contrast, the deep FSGF captures global frequency-domain patterns through larger dilation rates. Their synergistic operation enhances image restoration performance. This closed-loop optimization process, "spatial detail guidance-frequency domain noise decoupling”, maintains texture continuity when handling complex scenes with mixed stripe and impulse noise, validating the necessity of cross-domain modeling.
In particular, the limited performance improvement on synthetic data exposes the limitations of existing simulation methods. The uniform noise distribution in synthetic data creates a performance bottleneck, suggesting that future research should focus more on realistic modeling of noise physical models rather than solely relying on the generalization ability of synthetic data. Although the sample size of the Realistic dataset is relatively small, which may raise concerns about statistical robustness, the HDST model still maintains certain advantages compared with all the methods included in the comparison. In addition, the baseline improvement achieved by HDST under complex noise conditions in the synthetic dataset is superior to that under other noise conditions. This indirectly demonstrates the performance of HDST in handling real-world noise, which echoes the results obtained from the real-world dataset. In the future, we will seek more public real-world hyperspectral image denoising datasets and evaluate HDST on these datasets to comprehensively verify the robustness and generalization of the model under different imaging conditions and noise distributions.
From an engineering application perspective, HDST demonstrates a favorable balance of computation performance. Despite a 61% increase in parameters, the inference latency increases only by 10.1% due to low-computation design strategies, making it well suited for scenarios requiring both speed and high accuracy. If further validated, the dimension-preserving post-processing model proposed herein may be widely transplanted to various image denoising frameworks.
The limitations of this work and future improvement directions merit discussion: First, the gate generation mechanism should be further optimized to capture frequency-domain characteristics. Second, since frequency band analysis still relies on preset dilation rates, developing adaptive spectral perception modules is necessary. Finally, efforts should be made to further encapsulate the model and attempt its transplantation to more existing methods to enhance its universality. These challenges provide clear technical pathways for future research.
In summary, the success of HDST essentially stems from the precise matching between physical properties and algorithmic design. This establishes an important design paradigm for frequency-domain enhanced denoising models: under the premise of maintaining computational efficiency, achieve collaborative optimization of noise frequency-domain attributes and network architecture through mechanisms such as multiscale perception, cross-domain attention, and dynamic gating. This approach can be extended to other imaging modalities, offering new methodological references for various fields.

5. Conclusions

Hyperspectral image denoising needs to simultaneously address spatially non-uniform noise and spectral correlation interference. Traditional methods ignore the inherent multidimensional noise coupling and cross-band correlations of HSI images, leading to the limitation of single-domain processing being unable to balance detail preservation and noise suppression. This framework takes residual Transformer layers as the basic skeleton and constructs a closed-loop optimization process of “spatial feature extraction–frequency domain noise decoupling–cross-domain information calibration” through cross-layer collaboration between frequency domain post-processing units and hybrid domain multiscale modules. This work demonstrates that hybrid domain learning is not merely algorithmic fusion but creates emergent properties: the closed-loop spatial-frequency interaction enables noise component transferability that single-domain methods cannot achieve. A large number of experimental results also prove the effectiveness of the proposed method.
This framework is particularly suitable for denoising images with complex noise in real-world environments. In most cases, it can achieve state-of-the-art quality in the field and enables inference with a moderate number of parameters and high inference speed, endowing it with excellent potential for engineering deployment in tasks such as agricultural remote sensing, medical image enhancement, archaeological document restoration, and industrial non-destructive testing. Optimization strategies like model pruning and quantization can meet the requirements of resource-constrained platforms, while further porting and model pre-training can enhance the model’s usability. Future research will focus on testing the generalization ability of the modules and the model to promote the application of this method in a wider range of scenarios.

Author Contributions

Conceptualization, H.L.; methodology, H.L.; software, H.L.; validation, H.L.; formal analysis, H.L.; investigation, H.L.; resources, D.W.; data curation, H.L.; writing—original draft preparation, H.L.; writing—review and editing, D.W.; visualization, H.L.; supervision, D.W.; project administration, D.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the Realistic dataset at https://github.com/ColinTaoZhang/HSIDwRD (accessed on 29 July 2025) and the ICVL dataset at https://huggingface.co/datasets/danaroth/icvl (accessed on 29 July 2025). The code implementing the proposed HDST is openly available on GitHub at https://github.com/lhy-cn/HDST-HSIDenoise (accessed on 25 July 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GELUGaussian Error Linear Unit
CONVConvolution
CNNConvolutional Neural Network
HSIHyperspectral Image
RGBImage Using Red Green Blue Color Model
FFTFast Fourier Transform
IFFTInverse Fast Fourier Transform
ASPPAtrous Spatial Pyramid Pooling
HDSTHybrid-Domain Synergistic Transformer
HDMSHybrid-Domain Multiscale Module
FSCAFrequency-Spatial Collaborative Attention
FSGFFFT-Scale Gated Fusion
PSNRPeak Signal-to-Noise Ratio
SAMSpectral Angle Mapper
SSIMStructural Similarity Index
ParamsNumber of Parameters
GFLOPSGiga Floating Point Operations Per Second

References

  1. Yang, C.; Zhang, C.; Shen, H.; Peng, T.; Wang, C.; Deng, L.; Chen, H.; He, L. HFAN: High-Frequency Attention Network for hyperspectral image denoising. Int. J. Mach. Learn. Cybern. 2024, 15, 837–851. [Google Scholar] [CrossRef]
  2. Lin, Y.; Cai, Z.; Li, J.; Zhang, J. Lightweight Remote Sensing Image Denoising via Knowledge Distillation. In Proceedings of the 2022 IEEE 24th International Workshop on Multimedia Signal Processing (MMSP), Shanghai, China, 26–28 September 2022; pp. 1–7. [Google Scholar] [CrossRef]
  3. Tang, J.; Zhang, F.; Ma, F.; Gao, F.; Yin, Q.; Zhou, Y. How SAR Image Denoise Affects the Performance of DCNN-Based Target Recognition Method. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 3609–3612. [Google Scholar] [CrossRef]
  4. Li, Y.; Zhou, F. Research on denoising optimization algorithm of point cloud in 3d reconstruction. J. Phys. Conf. Ser. 2019, 1345, 052073. [Google Scholar] [CrossRef]
  5. Rasti, B.; Sveinsson, J.R.; Ulfarsson, M.O.; Benediktsson, J.A. Hyperspectral image denoising using 3D wavelets. In Proceedings of the 2012 IEEE International Geoscience and Remote Sensing Symposium, Munich, Germany, 22–27 July 2012; pp. 1349–1352. [Google Scholar] [CrossRef]
  6. Kaur, N.; Neetu. Comparison of Non Local Means Filtering with Wavelet Transform Technique for Denoising ECG signal. Int. J. Innov. Res. Sci. Eng. Technol. 2015, 4, 7052–7057. [Google Scholar] [CrossRef]
  7. Zhao, Y.; Li, Y.; Dong, X.; Yang, B. Low-frequency noise suppression method based on improved DnCNN in desert seismic data. IEEE Geosci. Remote Sens. Lett. 2018, 16, 811–815. [Google Scholar] [CrossRef]
  8. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 10–17 October 2021; pp. 1833–1844. [Google Scholar]
  9. Khan, M.J.; Khan, H.S.; Yousaf, A.; Khurshid, K.; Abbas, A. Modern Trends in Hyperspectral Image Analysis: A Review. IEEE Access 2018, 6, 14118–14129. [Google Scholar] [CrossRef]
  10. Govender, M.; Chetty, K.; Bulcock, H. A review of hyperspectral remote sensing and its application in vegetation and water resource studies. Water Sa 2007, 33, 145–151. [Google Scholar] [CrossRef]
  11. Fischer, C.; Kakoulli, I. Multispectral and hyperspectral imaging technologies in conservation: Current research and potential applications. Stud. Conserv. 2006, 51, 3–16. [Google Scholar] [CrossRef]
  12. Carrasco, O.; Gomez, R.B.; Chainani, A.; Roper, W.E. Hyperspectral imaging applied to medical diagnoses and food safety. In Proceedings of the Geo-Spatial and Temporal Image and Data Exploitation III, SPIE, Orlando, FL, USA, 21–25 April 2003; Volume 5097, pp. 215–221. [Google Scholar]
  13. Clark, R.N.; Swayze, G.A. Mapping minerals, amorphous materials, environmental materials, vegetation, water, ice and snow, and other materials: The USGS Tricorder algorithm. In Proceedings of the JPL, Summaries of the Fifth Annual JPL Airborne Earth Science Workshop, AVIRIS Workshop, Pasadena, CA, USA, 23–26 January 1995; Volume 1. [Google Scholar]
  14. Kuula, J.; Pölönen, I.; Puupponen, H.H.; Selander, T.; Reinikainen, T.; Kalenius, T.; Saari, H. Using VIS/NIR and IR spectral cameras for detecting and separating crime scene details. In Proceedings of the Sensors, and Command, Control, Communications, and Intelligence (C3I) Technologies for Homeland Security and Homeland Defense XI, SPIE, Baltimore, MD, USA, 23–27 April 2012; Volume 8359, pp. 150–160. [Google Scholar]
  15. Maffei, A.; Haut, J.M.; Paoletti, M.E.; Plaza, J.; Bruzzone, L.; Plaza, A. A Single Model CNN for Hyperspectral Image Denoising. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2516–2529. [Google Scholar] [CrossRef]
  16. Xiong, F.; Zhou, J.; Tao, S.; Lu, J.; Zhou, J.; Qian, Y. SMDS-Net: Model Guided Spectral-Spatial Network for Hyperspectral Image Denoising. IEEE Trans. Image Process. 2022, 31, 5469–5483. [Google Scholar] [CrossRef]
  17. Ozdemir, A.; Polat, K. Deep Learning Applications for Hyperspectral Imaging: A Systematic Review. J. Inst. Electron. Comput. 2020, 2, 39–56. [Google Scholar] [CrossRef]
  18. Palsson, F.; Ulfarsson, M.O.; Sveinsson, J.R. Hyperspectral image denoising using a sparse low rank model and dual-tree complex wavelet transform. In Proceedings of the 2014 IEEE Geoscience and Remote Sensing Symposium, Quebec City, QC, Canada, 13–18 July 2014; pp. 3670–3673. [Google Scholar]
  19. Rasti, B.; Scheunders, P.; Ghamisi, P.; Licciardi, G.; Chanussot, J. Noise reduction in hyperspectral imagery: Overview and application. Remote Sens. 2018, 10, 482. [Google Scholar] [CrossRef]
  20. Zhou, D.; Wang, Q.J.; Tian, Q.J.; Lin, Q.Z.; Fu, W.X. Wavelet analysis and its application in denoising the spectrum of hyperspectral image. Spectrosc. Spectr. Anal. 2009, 29, 1941–1945. [Google Scholar]
  21. Xie, W.; Li, Y. Hyperspectral Imagery Denoising by Deep Learning With Trainable Nonlinearity Function. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1963–1967. [Google Scholar] [CrossRef]
  22. Dong, W.; Wang, H.; Wu, F.; Shi, G.; Li, X. Deep Spatial–Spectral Representation Learning for Hyperspectral Image Denoising. IEEE Trans. Comput. Imaging 2019, 5, 635–648. [Google Scholar] [CrossRef]
  23. Liang, H.; Ke, C.; Li, K. Hybrid Spatial-Spectral Neural Network for Hyperspectral Image Denoising. In Proceedings of the Computer Vision—ECCV 2024 Workshops, Milan, Italy, 29 September–4 October 2024; Del Bue, A., Canton, C., Pont-Tuset, J., Tommasi, T., Eds.; Springer: Cham, Switzerland, 2025; pp. 278–294. [Google Scholar]
  24. Fu, G.; Xiong, F.; Tao, S.; Lu, J.; Zhou, J.; Qian, Y. Learning a Model-Based Deep Hyperspectral Denoiser from a Single Noisy Hyperspectral Image. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 4131–4134. [Google Scholar] [CrossRef]
  25. Zhang, H.; Xu, H.; Yu, X.; Zhang, X.; Wu, C. Leveraging Frequency and Spatial Domain Information for Underwater Image Restoration. J. Phys. Conf. Ser. 2024, 2832, 012001. [Google Scholar] [CrossRef]
  26. Zheng, P.; Chen, K.; Huang, J.; Chen, B.; Liu, J.; Ren, Y.; Pu, X. Lightweight Medical Image Restoration via Integrating Reliable Lesion-Semantic Driven Prior. arXiv 2025, arXiv:2504.11286. [Google Scholar] [CrossRef]
  27. Gao, N.; Jiang, X.; Zhang, X.; Deng, Y. Efficient Frequency-Domain Image Deraining with Contrastive Regularization. In Proceedings of the 18th European Conference on Computer Vision—ECCV 2024, Milan, Italy, 29 September–4 October 2024; Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G., Eds.; Springer: Cham, Switzerland, 2025; pp. 240–257. [Google Scholar]
  28. Zhu, Y.; Chen, J.; Liang, L.; Kuang, Z.; Jin, L.; Zhang, W. Fourier Contour Embedding for Arbitrary-Shaped Text Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 3123–3131. [Google Scholar]
  29. Kong, L.; Dong, J.; Ge, J.; Li, M.; Pan, J. Efficient frequency domain-based transformers for high-quality image deblurring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5886–5895. [Google Scholar]
  30. Wang, Y.; Wang, H.; Wang, L.; Wang, X.; Zhu, L.; Lu, W.; Huang, H. Complementary advantages: Exploiting cross-field frequency correlation for nir-assisted image denoising. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 11–15 June 2025; pp. 12679–12689. [Google Scholar]
  31. Wang, S.; Li, L.; Li, X.; Zhang, J.; Zhao, L.; Su, X.; Chen, F. A Denoising Network Based on Frequency-Spectral- Spatial-Feature for Hyperspectral Image. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 6693–6710. [Google Scholar] [CrossRef]
  32. Sheng, Z.; Yu, Z.; Liu, X.; Cao, S.Y.; Liu, Y.; Shen, H.L.; Zhang, H. Structure Aggregation for Cross-Spectral Stereo Image Guided Denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 13997–14006. [Google Scholar]
  33. Chen, L.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. arXiv 2016, arXiv:1606.00915. [Google Scholar] [CrossRef]
  34. Lian, X.; Pang, Y.; Han, J.; Pan, J. Cascaded hierarchical atrous spatial pyramid pooling module for semantic segmentation. Pattern Recognit. 2021, 110, 107622. [Google Scholar] [CrossRef]
  35. Zhu, Y.; Wan, L.; Xu, W.; Wang, S. ASPP-DF-PVNet: Atrous Spatial Pyramid Pooling and Distance-Filtered PVNet for occlusion resistant 6D object pose estimation. Signal Process. Image Commun. 2021, 95, 116268. [Google Scholar] [CrossRef]
  36. Li, M.; Liu, J.; Fu, Y.; Zhang, Y.; Dou, D. Spectral Enhanced Rectangle Transformer for Hyperspectral Image Denoising. arXiv 2023, arXiv:2304.00844. [Google Scholar] [CrossRef]
  37. Zhang, Q.; Zheng, Y.; Yuan, Q.; Song, M.; Yu, H.; Xiao, Y. Hyperspectral Image Denoising: From Model-Driven, Data-Driven, to Model-Data-Driven. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 13143–13163. [Google Scholar] [CrossRef]
  38. Archana, R.; Jeevaraj, P.S.E. Deep learning models for digital image processing: A review. Artif. Intell. Rev. 2024, 57, 11. [Google Scholar] [CrossRef]
  39. Maggioni, M.; Katkovnik, V.; Egiazarian, K.; Foi, A. Nonlocal transform-domain filter for volumetric data denoising and reconstruction. IEEE Trans. Image Process. 2012, 22, 119–133. [Google Scholar] [CrossRef]
  40. He, W.; Yao, Q.; Li, C.; Yokoya, N.; Zhao, Q. Non-local meets global: An integrated paradigm for hyperspectral denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6868–6877. [Google Scholar]
  41. Brook, A. 3-D wavelets-based denoising and enhancement of hyperspectral imagery. In Proceedings of the 2014 6th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Lausanne, Switzerland, 24–27 June 2014; pp. 1–4. [Google Scholar] [CrossRef]
  42. Yuan, Q.; Zhang, Q.; Li, J.; Shen, H.; Zhang, L. Hyperspectral Image Denoising Employing a Spatial–Spectral Deep Residual Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1205–1218. [Google Scholar] [CrossRef]
  43. Xiong, F.; Zhou, J.; Zhao, Q.; Lu, J.; Qian, Y. MAC-Net: Model-aided nonlocal neural network for hyperspectral image denoising. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–14. [Google Scholar] [CrossRef]
  44. Fu, G.; Xiong, F.; Lu, J.; Zhou, J.; Zhou, J.; Qian, Y. Hyperspectral Image Denoising via Spatial–Spectral Recurrent Transformer. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–14. [Google Scholar] [CrossRef]
  45. Sara, U.; Akter, M.; Uddin, M.S. Image quality assessment through FSIM, SSIM, MSE and PSNR—A comparative study. J. Comput. Commun. 2019, 7, 8–18. [Google Scholar] [CrossRef]
  46. Terven, J.; Cordova-Esparza, D.M.; Romero-González, J.A.; Ramírez-Pedraza, A.; Chávez-Urbiola, E. A comprehensive survey of loss functions and metrics in deep learning. Artif. Intell. Rev. 2025, 58, 195. [Google Scholar] [CrossRef]
  47. Zhang, H.; Liu, L.; He, W.; Zhang, L. Hyperspectral Image Denoising With Total Variation Regularization and Nonlocal Low-Rank Tensor Decomposition. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3071–3084. [Google Scholar] [CrossRef]
  48. Wei, K.; Fu, Y.; Huang, H. 3-D Quasi-Recurrent Neural Network for Hyperspectral Image Denoising. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 363–375. [Google Scholar] [CrossRef] [PubMed]
  49. Bodrito, T.; Zouaoui, A.; Chanussot, J.; Mairal, J. A Trainable Spectral-Spatial Sparse Coding Model for Hyperspectral Image Restoration. arXiv 2021, arXiv:2111.09708. [Google Scholar] [CrossRef]
Figure 1. Overall framework of HDST. (a) The overall diagram of the HDST network formed by fusing the network framework of the residual attention layer and the frequency-domain post-processing module. (b) ASPP Reference: The module structure diagram of the original Atrous Spatial Pyramid Pooling. (c) The diagram of spatial-domain-frequency-domain-hybrid-domain feature extraction process based on ASPP. (d) The diagram of cross-domain attention and residual calibration process.
Figure 1. Overall framework of HDST. (a) The overall diagram of the HDST network formed by fusing the network framework of the residual attention layer and the frequency-domain post-processing module. (b) ASPP Reference: The module structure diagram of the original Atrous Spatial Pyramid Pooling. (c) The diagram of spatial-domain-frequency-domain-hybrid-domain feature extraction process based on ASPP. (d) The diagram of cross-domain attention and residual calibration process.
Applsci 15 09735 g001
Figure 2. Cross-domain feedback loop.
Figure 2. Cross-domain feedback loop.
Applsci 15 09735 g002
Figure 3. Visualization of denoising results on the Realistic dataset. The area for visual comparison of different models in the upper right part of this figure corresponds to the region enclosed by the yellow box in the thumbnail at the upper left. The regions enclosed by these three red boxes at the bottom of the figure are the detail areas where HDST demonstrates better reconstruction performance compared to SSRT-UNet.
Figure 3. Visualization of denoising results on the Realistic dataset. The area for visual comparison of different models in the upper right part of this figure corresponds to the region enclosed by the yellow box in the thumbnail at the upper left. The regions enclosed by these three red boxes at the bottom of the figure are the detail areas where HDST demonstrates better reconstruction performance compared to SSRT-UNet.
Applsci 15 09735 g003
Figure 4. Visualization of denoising results on ICVL synthetic dataset. The area for visual comparison of different models in the upper right part of this figure corresponds to the region enclosed by the yellow box in the thumbnail at the upper left. The regions enclosed by the red boxes at the bottom of the figure are the detail areas where HDST demonstrates better reconstruction performance compared to SSRT-UNet.
Figure 4. Visualization of denoising results on ICVL synthetic dataset. The area for visual comparison of different models in the upper right part of this figure corresponds to the region enclosed by the yellow box in the thumbnail at the upper left. The regions enclosed by the red boxes at the bottom of the figure are the detail areas where HDST demonstrates better reconstruction performance compared to SSRT-UNet.
Applsci 15 09735 g004
Figure 5. Scatter plot illustrating parameter efficiency versus PSNR performance.
Figure 5. Scatter plot illustrating parameter efficiency versus PSNR performance.
Applsci 15 09735 g005
Table 1. Average results of different denoising methods on the Realistic dataset.
Table 1. Average results of different denoising methods on the Realistic dataset.
CategoryMethodPSNRSSIMSAMTime
-Noisy23.260.760917.329-
TraditionalBM4D [39]29.040.94713.08711.83 m
Wavelet [41]28.050.90465.853.25 m
NGMeet [40]28.720.95112.73517.88 m
Deep LearningHSID-CNN [42]26.440.89925.2421.345 s
MAC-Net [43]29.20.94894.0992.157 s
SSRT-UNet [44]30.170.95492.50620.566 s
SERT [36]29.680.95332.5360.551 s
HDST (ours)30.620.95552.4170.607 s
All values are averaged over 15 test images. Higher PSNR/SSIM and lower SAM indicate better performance. Bold indicates best performance. (Based on PSNR.).
Table 2. Average results of different denoising methods on various noise types in the synthetic dataset.
Table 2. Average results of different denoising methods on various noise types in the synthetic dataset.
MethodGaussian+MixtureGaussian+DeadlineGaussian+Impulse
PSNRSSIMSAMPSNRSSIMSAMPSNRSSIMSAM
Noisy13.910.339651.5317.50.477047.5514.930.375846.98
BM4D28.010.841923.5933.770.96156.8529.790.861321.59
Wavelet26.090.794532.8332.570.94768.8728.870.885617.59
NGMeet26.130.779631.8933.410.96656.5527.020.788431.20
HSID-CNN35.300.95886.2938.330.97833.9936.210.96635.48
MAC-Net30.590.930014.5136.680.98605.6334.540.955310.20
SSRT-UNet40.170.99022.9644.010.99601.6642.590.99322.17
SERT40.000.99372.8443.660.99691.9942.670.99592.30
HDST (ours)40.520.99472.5143.810.99661.8342.890.99631.96
MethodGaussian+StripeNon-i.i.d GaussianGaussian 10–70
PSNRSSIMSAMPSNRSSIMSAMPSNRSSIMSAM
Noisy17.510.486746.9818.290.511646.217.240.478241.94
BM4D35.630.97306.2636.180.97675.7836.620.97705.51
Wavelet34.170.96486.7334.290.96586.6236.510.98743.52
NGMeet34.880.96655.4234.900.97455.3741.670.99372.19
HSID-CNN38.090.97654.5939.280.98193.8039.040.97763.71
MAC-Net39.030.99104.0339.980.96624.5541.310.99412.52
SSRT-UNet43.750.99531.7944.050.99621.6943.270.99471.89
SERT43.680.99691.9744.200.99711.6942.820.99571.88
HDST (ours)43.880.99631.7644.370.99701.6243.010.99671.83
All values are averaged over 50 test images. Bold text (except for the headers in the two columns at the top and middle of the table) indicates the best performance. Underlined indicates worst performance. (All based on PSNR.).
Table 3. Ablation study on the Realistic dataset.
Table 3. Ablation study on the Realistic dataset.
ModelFrequencyResidualASPPPSNRSSIMSAMPara
Baseline [36] 29.680.95332.5361.91M
Net1 29.730.95292.5282.25M
Net2 29.890.95412.5082.53M
Net3 30.270.95392.4752.46M
Net4 30.460.95502.4332.80M
HDST30.620.95552.4173.08M
✓ indicates the component is included. Frequency: frequency post-processing module. Residual: dynamic fusion. ASPP/HDMS: ASPP-FFT & Hybrid-Domain multiscale module.
Table 4. ASPP dilation rate comparison on the Realistic dataset.
Table 4. ASPP dilation rate comparison on the Realistic dataset.
ASPP Rates[6,12,18][1,3,5][2,4,6][2,4,8][4,8,12][16,32,32]
HDMS−0.17−0.09(30.62)−0.04−0.11−0.2
FSGF−0.13−0.07−0.02(30.62)−0.05−0.27
The values in the table represent the differences between the PSNR at the current rate and the PSNR at the optimal rate (marked as (30.62)).
Table 5. Model complexity analysis on the Realistic dataset.
Table 5. Model complexity analysis on the Realistic dataset.
MetricBaseline [36]HDST Δ (%)
GFLOPS4086.24686.8+14.7
Params (M)1.913.08+61.2
Inference Time (ms)551607+10.1
Tested on GeForce RTX 3090 GPU.
Table 6. PSNR and parameter count comparison.
Table 6. PSNR and parameter count comparison.
MetricMAC-Net [43]SSRT-UNet [44]SERT [36]HDST
PSNR (dB)29.230.1729.6830.62
Params (M)0.4310.21.913.08
Tested on the Realistic dataset.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, H.; Wu, D. Hybrid-Domain Synergistic Transformer for Hyperspectral Image Denoising. Appl. Sci. 2025, 15, 9735. https://doi.org/10.3390/app15179735

AMA Style

Li H, Wu D. Hybrid-Domain Synergistic Transformer for Hyperspectral Image Denoising. Applied Sciences. 2025; 15(17):9735. https://doi.org/10.3390/app15179735

Chicago/Turabian Style

Li, Haoyue, and Di Wu. 2025. "Hybrid-Domain Synergistic Transformer for Hyperspectral Image Denoising" Applied Sciences 15, no. 17: 9735. https://doi.org/10.3390/app15179735

APA Style

Li, H., & Wu, D. (2025). Hybrid-Domain Synergistic Transformer for Hyperspectral Image Denoising. Applied Sciences, 15(17), 9735. https://doi.org/10.3390/app15179735

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop