Next Article in Journal
Precision and Reliability of a Dynamometer for Trunk Extension Strength and Steadiness Assessment
Previous Article in Journal
Optimal Recovery Time for Post-Activation Performance Enhancement After an Acute Bout of Plyometric Exercise on Unilateral Countermovement Jump and Postural Sway in National-Level Female Volleyball Players
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectral Information Divergence-Driven Diffusion Networks for Hyperspectral Target Detection

1
Shanghai Institute of Technical Physics, Chinese Academy of Sciences, Shanghai 200083, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
School of Instrument Science and Optp-Electronics Engineering, Hefei University of Technology, Hefei 230009, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(8), 4076; https://doi.org/10.3390/app15084076
Submission received: 10 March 2025 / Revised: 26 March 2025 / Accepted: 4 April 2025 / Published: 8 April 2025

Abstract

:
Hyperspectral Imagery (HSI) plays a crucial role in military and civilian target detection. However, HSI target detection remains highly challenging due to the interference caused by complex and diverse real-world scenarios. This paper proposes a Spectral Information Divergence-driven Diffusion Network model (SID-DN) for hyperspectral target detection, which significantly enhances detection robustness in complex scenes by decoupling background distribution modeling from target detection. The proposed method focuses on learning the background distribution in hyperspectral data and achieves target detection by accurately reconstructing background samples to identify differences between background and target samples. This method introduces an adaptive coarse detection module, which optimizes the coarse detection process in generative hyperspectral target detection, effectively reducing the background-target misclassification. Additionally, a SID-based Diffusion model is designed to optimize the loss of Diffusion, effectively reducing the interference of suspected target samples during the background learning process. Experiments on three real-world datasets demonstrate that the method is highly competitive, with detection results significantly outperforming current state-of-the-art methods.

1. Introduction

The growing number of remote sensing satellites and the miniaturization of remote sensing equipment in recent years have enhanced the flexibility of remote sensing data acquisition, enabling quicker and more targeted responses to specific areas [1]. Among various remote sensing technologies, Hyperspectral Imagery (HSI) has attracted significant attention due to its capacity for capturing the distinct spectral properties of different objects in a scene. Hyperspectral sensors can provide hundreds of continuous spectral data across the electromagnetic spectrum, from the visible to the infrared (IR) bands. This enables the accurate observation of the reflection or emission characteristics of the object being tested [2]. Therefore, hyperspectral is capable of a wide range of vision tasks, including image classification [3], anomalous target detection [4], target detection [5], and change detection [6]. Furthermore, it plays a pivotal role in numerous practical applications, including land cover classification [7,8], sea and land military target detection [9], mineral resources exploration [10], and crop pathology detection [11], among others.
Hyperspectral Target Detection (HTD) identifies targets in complex backgrounds by exploiting subtle spectral differences between the targets and backgrounds in HSI. Because of impediments such as cloud cover and the poor spatial resolution of HSI, detected targets are often represented at a sub-pixel level, making the extraction process challenging. HTD methods are primarily divided into classical detection techniques and deep learning-based methods.
The classical HTD methods primarily include methods based on spectral angle detection [12], hypothesis testing [13], Constrained Energy Minimization (CEM) [14], signal decomposition [15] and sparse representation [16]. Spectral angle detection methods measure spectral similarity through geometric or probabilistic metrics, such as Spectral Angle Mapper (SAM) [12], Adaptive Coherence Estimator(ACE) [17] and Spectral Information Divergence (SID) [18].These methods are sensitive to spectral variations and often serve as pre- or post-processing steps due to their computational simplicity [19].Hypothesis testing-based detectors, including Adaptive Spatial Detector (ASD) [20] and Kernel-based Adaptive Subspace Detector (KASD) [21], rely on statistical signal theory and likelihood ratio tests but face limitations from Gaussian distribution assumptions. The CEM detection algorithm [14], minimizes background energy while enhancing target responses through linear filtering. Though simple, its linear nature struggles with nonlinear HSI features, prompting nonlinear extensions like hCEM [22] and machine learning variants [23,24].The Signal Decomposition Detection method [25,26] separate targets from backgrounds using models like Orthogonal Subspace Projection (OSP) [27] and Low-Rank and Sparse Matrix Decomposition (LRaSMD) [28], yet require precise spectral modeling and suffer from high complexity.The sparse representation methods [29,30,31] leverage sparse coefficients and dictionary learning (e.g., a dual sparsity constrained(DSC) [32]) to enhance target-background discrimination. While effective, their performance depends heavily on dictionary completeness, limiting generalization.Most traditional detection algorithms rely on manually predefined assumptions (e.g., spectral matching criteria, energy minimization, or distribution models) and struggle to effectively handle the nonlinear characteristics inherent in complex real-world scenarios. This limitation has driven researchers to explore and adopt end-to-end deep learning models, leveraging data-driven approaches to reduce reliance on handcrafted feature engineering.
Recently, the rapid advancements in deep learning have revitalized HTD. Deep learning can capture complex and diverse nonlinear characteristics of HSI using multi-layered network structures. Deep learning-based HTD methods can be categorized into two main approaches. The first approach constructed an end-to-end network for HSI to enable target detection across all images. Freitas et al. [33] developed a detection model using a 3D Convolutional Neural Network (CNN) for maritime HTD. Zhang et al. [34] introduced HTD-Net, a method combining an Auto Encoder (AE) with CNN for HTD in scenarios with missing labels. Qin et al. [35] introduced the HTD-ViT algorithm, which integrated spectral and spatial information for HTD. Jiao et al. [36] combined Fourier transform with Transformer to develop a dual-branch Fourier hybrid transform network. Wang et al. [37] innovatively enhanced spectral feature learning by integrating the momentum encoder with Transformer. Jiao et al. [38] introduced a triplet spectral transformer method for HTD, excelling at capturing long-term dependencies in spectral features.
The second approach involved reconstructing hyperspectral data using generative deep learning, followed by target detection based on differences between the reconstructed target and background. These methods typically relied on learning from background samples generated after coarse detection. With advancements in generative networks, Generative Adversarial Networks (GAN) [39], Variational Auto Encoders (VAE) [40], and AEs have been widely applied in HTD. Xie et al. [41] pioneered the field by proposing a target-constrained background GAN model, which relies on background learning based on target suppression constraint (BLTSC) for HTD. Xie et al. [42] introduced an unsupervised algorithm leveraging a deep latent spectral representation model to adaptively map latent representations to optimal subset relationships. Ali et al. [43] proposed a self-supervised background learning method by combining principal component analysis (PCA) with an adversarial autoencoder (AAE) network. Xie et al. [44] proposed a Spectral Regularized Unsupervised Network (SRUN) to enhance the spectral representation capabilities of Autoencoder and Variational Autoencoder (VAE) models. The framework integrated spectral angular difference-driven feature selection and adaptive weighting mechanisms to generate discriminative detection maps. Tian et al. [45] combined OSP theory with VAE and proposed the OS-VAE network framework, exploring new methods that fuse traditional methods with deep learning.In order to improve non-linear feature expression and physical interpretability, Shen et al. [46] designed a subspace representation network for HTD. Shi et al. successively integrated AE with HTD to develop the DCSSAE network [47] and 3DMMRAE network [48].
Despite the significant success of deep learning-based HTD, several challenges remain:
  • Supervised deep learning often demands extensive labeled training data, which is challenging to obtain in practical scenarios, potentially undermining its robustness. Recently, self-supervised generative HTD has advanced significantly, but coarse detection often suffers from background-target misclassification. Improving coarse detection remains a critical challenge in generative HTD.
  • Effective background learning is crucial for enhancing algorithm performance through improved background representation. However, HSI is often underestimated in practical applications, with its relationship to prior spectra poorly understood, leading to inadequate suppression of specific targets.
Based on the above considerations, this paper proposes a hyperspectral detection deep learning model based on Diffusion and target suppression constraints, with the intention of enhancing coarse detection’s efficiency and the ability to represent the background. The model is named by SID-DN. The main contributions of this paper are as follows:
  • This paper proposes a two-stream adaptive coarse detection algorithm that enhances the coarse detection process and better distinguishes background samples from target samples.
  • In order to enhance the learning of hyperspectral information by Diffusion, a SID-based Diffusion model is designed to optimize the loss of Diffusion. Compared with the traditional Diffusion, the performance of the algorithm is improved while using prior spectral information.
  • To verify the effectiveness of the method in this paper, the results are compared with the most advanced existing algorithms, which show that our model is superior to existing HTD algorithms.
The rest of this paper is organized as follows. Section 2 reviews the necessary knowledge and related works. Section 3 gives a detailed description of the proposed method. Section 4 presents the experimental setup and results on the data set. Section 5 analyses the experiments. Section 6 concludes and summarizes the paper.

2. Related Work

2.1. Notations

In this paper, tensors, matrices, vectors and scalars are denoted by H , H , h , h. For instance, the HSI tensor is represented by H R h × w × L ,where h , w , L correspond to height, width and the number of spectral band respectively. HSI tensor reshapes as two-dimensional matrix H R n × L , where n represents h × w . The prior spectral vector represents d R L × 1 .

2.2. Diffusion Models

A generative model generates samples resembling the training data by learning its probability distribution. The model is robust because it continuously learns the characteristics of the data distribution. In general, the generative models can be divided into naive Bayes, hidden Markov models, GANs, VAEs and diffusion models. Diffusion models have advantages over GANs and VAEs, including training stability, strong interpretability, and the ability to generate high-quality results. Therefore, they have become the mainstream approach for generative models. Denoising diffusion probabilistic models (DDPMs) are the most frequently utilized models [49]. The core concept involves generating data through forward and reverse data denoising processes. Both processes are modeled as parametric Markov chains. The forward process, which lacks learnable parameters, converts samples into standard Gaussian noise through iterative noise addition. The reverse process contains learnable parameters, and the image is denoised and gradually restored during the process, usually using a neural network approach. The model is generally a unet-based model. DDPMs have strong feature generation capabilities and are currently widely used in HSI denoising [50], HSI restoration [51], HSI classification [52], anomaly detection [53] and other tasks, but are rarely involved in the field of HTD. In HTD applications, DDPMs enhance spectral differences between the target and background, facilitating their separation.

2.3. Spectral Information Divergence

In this section, we introduce SID that is used in Section 3. SID is a critical parameter for measuring spectral correlation. It quantifies the spectral similarity of the target to other pixels using Kullback-Leibler (KL) divergence. Supposed x = ( x 1 , x 2 , , x L ) is target pixel, y = ( y 1 , y 2 , , y L ) represent the other pixel, and L is the number of bands. The probability vector of pixel x is given by p = ( p 1 , p 2 , , p L ) , and the probability vector y of pixel is given by q = ( q 1 , q 2 , , q L ) , where p i = x i i = 1 L x i , q i = y i i = 1 L y i . According to the definition of KL divergence, the relative entropy of the target pixel x with respect to pixel y is expressed as:
S ( x y ) = i = 1 L q i log q i p i
where S ( · ) represents KL divergence.The SID is composed of two directions of relative entropy, given by:
S I D ( x y ) = S ( x y ) + S ( y x )
Using this method, SID accounts for the bi-directional spectral differences between two pixels, effectively capturing the spectral information between them.

3. Proposed Method

This paper presents a novel HTD framework based on Diffusion joint spectral information divergence constraints. In this paper, the SID idea is integrated with Diffusion to propose a model for better representation of the background. As illustrated in Figure 1, the proposed deep learning method, based on diffusion constraints based on SID, comprises two primary modules: background learning and target comparison detection.

3.1. Double-Branch Adaptive Coarse Detection (DBACD)

The limited availability of target spectra presents a significant challenge for deep network training. Conversely, the abundance of background samples in HSI offers new opportunities for target detection by leveraging the differences between background samples and targets. By reconstructing diverse background samples, the model effectively learns the background distribution characteristics, while the reconstructed target remains significantly different from the background samples. However, previous studies often use a fixed threshold to initially select background samples, which lacks adaptability to dataset distributions, performs poorly across different datasets, and demonstrates limited robustness. In this regard, this paper proposes a Double-Branch Adaptive Coarse Detection method, which can better divide the background samples and target samples. As shown in Figure 2, the comparison between the proposed method and the traditional coarse detection is shown. Inspired by integrated learning, this paper combines two coarse detection algorithms CEM and MF adaptively determines the threshold for separating background and target samples based on the relationship between the mean and variance of the coarse detection results.
Given the input hyperspectral matrix as H R h × w × L and a prior spectral information as d R L × 1 , the output value of the CEM is
D CEM = w T H = d T R 1 d R 1 d T H
where R = ( 1 / n H H T ) is the inter-correlation tensor and D C E M R h × w × 1 . And the tensor D C E M R h × w × 1 unfolds as matrix D C E M R h × w . CEM minimizes the filter capacity under predefined conditions by employing a FIR linear filter design.
The Match Filter (MF) [54] method mainly models the background variation by using the mean vector μ MF and covariance matrix Σ MF . The output value of the MF, D MF R h × w  is:
D M F = d μ M F T Σ M F 1 H μ M F d μ M F T Σ M F 1 d μ M F
These two methods have their own advantages and disadvantages, and we adopt the idea of integrated learning here, so that D C E M and D M F are jointly coarse detected in the second order. The values of pixels that are similar to the a prior spectra are highlighted by the coarse detection. However, traditional methods typically use a fixed threshold for binarization, often leading to the misclassification of numerous background samples and insufficient learning of background characteristics. This paper adopts a method to adaptively determine the threshold value based on the mean and standard deviation of the image. The threshold t h is defined as t h = μ 1 + 3 × σ 1 , where μ 1 represents the mean of coarse detection map and σ 1 denotes the standard deviation of the coarse detection map. The coarse detection maps processed through CEM and MF are fused using the coarse detection fusion factor ι , The fusion result is expressed as:
D ^ = ι × D C E M + ( 1 ι ) × D M F
where D ^ represents the fused coarse detection map, serving as a crucial component in subsequent target detection processes.
After binarizing the images generated by the two coarse detections, the target and background values converge to 1 and 0, respectively. The threshold, derived from the statistical properties of the image, significantly reduces the likelihood of background misclassification. Following binarization, the intersection of D C E M and D M F produces the binarized map D . By binarizing the map D , the set of background samples B is identified as shown in Equation (6):
B ( i , j , : ) = 0 , D ( i , j ) = 1 H ( i , j , : ) , otherwise
where H ( i , j ) represents the spectral vector at position ( i , j ) . This approach enables the effective construction of training background samples using the a prior spectra d . We use 75% of the samples for training. Comparative experiments will be conducted in Section 5 to validate the effectiveness of the proposed method.

3.2. Background Learning Based on Diffusion Model

This paper builds on the diffusion model in the BSDM proposed by Ma et al. [53], using a Gaussian distribution N N μ , σ 2 , with the same mean and variance as the input HSI as the pseudo background noise, where μ is the mean of the background samples and σ is the standard deviation of the background samples. Figure 3 illustrates the schematic of the forward noise addition process and the denoising network in the diffusion model.
The zero-moment data b 0 is sampled from the background data distribution q ( x ) , i.e., b 0 q ( x ) . The forward diffusion process involves the gradual addition of T noises to b 0 , where ϵ 0 , , ϵ t 2 , ϵ t 1 all follow a Gaussian distribution N N μ , σ 2 . This entire process is represented mathematically in Equation (7):
q b t b t 1 = N b t ; 1 β t b t 1 + μ β t , σ 2 β t I s . t . , β t = λ t T
where b t represents the background sample after t-step diffusion, β t represents the noise intensity, λ is a hyperparameter controlling β t , and and I is the identity matrix. The forward process can be regarded as a Markov process since each moment t is only related to the t 1 moment, so the relationship between the t moment and the initial moment can be recursively introduced as shown in Equation (8):
q b 1 : T b 0 = t = 1 T q b t b t 1
To simplify the calculation, this paper uses the reparameterization technique in DDPM and defines α t = 1 β t , α ¯ t = t = 1 T α i , so that b t can be expressed as Equation (9):
b t = α t b t 1 + 1 α t ϵ t 1 = α t α t 1 b t 2 + 1 α t α t 1 ϵ ¯ t 2
Since ϵ t 1 and ϵ t 2 both follow a Gaussian distribution, they can be combined based on Gaussian properties, with  ϵ ¯ representing the resulting distribution. Equation (10) is derived through recursion.
b t = α ¯ t b 0 + θ t ϵ ¯ s . t . , θ t ϵ N μ θ t , 1 α ¯ t σ 2
ϵ ¯ represents a Gaussian function comprising multiple moments, and  θ t = k = 0 t 1 α ¯ t α t k α ¯ t k . Substituting this into Equation (8) results in Equation (11):
q b t b t 1 = N b t ; α ¯ t b t 1 + μ θ t , 1 α ¯ t σ 2 I
The above describes the forward diffusion process of the diffusion model, which obtains the distribution of background samples at T moments. This process is expressed in Equation (12):
B t = B t N = α ¯ t B + θ t N
The inverse process of the diffusion model is typically denoised using a deep neural network with the structure of the denoising network illustrated in Figure 3. The spectral vector in B t serves as the input, while the estimated noise n i at the t-th step is the output, as expressed in Equation (13):
n i = f B t ; γ
where f represents the deep learning network and γ denotes its parameters. The network f primarily comprises a temporal embedding layer and a differential Transformer. This paper initially adopts the same temporal embedding method as DDPM.
t ^ = Emb t t ; γ te , s . t . , γ te γ
where Emb t . ; γ t e denotes the temporal embedding function network and γ t e represents its parameters.
B t ˜ = g B t , γ linear + t ^
where g . ; γ l i n e a r is the linear transformation layer, with  γ l i n e a r as its parameter.
This paper employs a differential transformer-based network for the denoising model [55]. Transformer models often allocate excessive attention to task-irrelevant contextual information, leading to an overemphasis on non-critical parts. The differential attention mechanism mitigates this issue by removing attentional noise and directing the model’s focus toward crucial information. This approach is conceptually analogous to differential amplifiers in electrical engineering. For a given input background sample B ˜ R ( h × w ) × L , it is first projected onto queries, keys, and values: Q 1 , Q 2 , K 1 , K 2 R ( h × w ) × L and V R ( 2 × h × w ) × L . The differential attention D i f f ( · ) is then calculated as follows:
Q 1 ; Q 2 = B ˜ W Q , K 1 ; K 2 = B ˜ W K , V = B ˜ W V D i f f ( B ˜ ) = ( softmax ( Q 1 K 1 T d ) λ softmax ( Q 2 K 2 T d ) ) V
where W Q , W K , W V R ( 2 × h × w ) × L , and λ is learnable shared parameter, designed as
λ = exp λ q 1 · λ k 1 exp λ q 2 · λ k 2 + λ init
where λ q 1 , λ k 1 , λ q 2 , λ k 2 are learnable parameters, while λ init serves as a hyperparameter. Additionally, the multi-head attention mechanism employs, W i Q , W i K , W i V , where i 1 , h . W O is a learnable projection matrix.
head i = D B ˜ ; W i Q , W i K , W i V , λ head i ¯ = 1 λ init · head i ¯ MultiHead ( B ˜ ) = Concat head i ¯ , head h ¯ W O
The output n ˜ is obtained after a linear layer transformation. The mean squared error loss function L 1 used for the diffusion model is
L 1 = min θ , t N f B t ; γ 2
The background samples classified during coarse detection are often misclassified. To improve the reliability of background reconstruction, this paper proposes a SID-based loss function, which serves as a criterion for target suppression during training. It is introduced in Section 2. The SID is calculated as shown in Equation (2), and the average SID value for each sample is represented as L s i d , as given in Equation (20):
L s i d = 1 m i m S I D [ i ]
where m is the number of background samples. Consequently, the total loss function is expressed in Equation (21):
L t o t a l = L 1 δ × L s i d

3.3. Target Detection

The target detection process corresponds to the reverse denoising process in DDPM. Starting from a predefined pseudo-background noise, the trained denoising network iteratively restores all spectral samples in reverse order, from step t = T to t = 0 . The final output spectral samples closely approximate the true data distribution, denoted as H ^ .
Effective target detection is achieved by analyzing the differences in reconstructed spectra between H ^ and H . In this study, the spectral angle mapper (SAM) function is utilized, as proposed by previous scholars, to evaluate the degree of change in samples before and after reconstruction. For each h ( i , j ) and h ^ ( i , j ) , the SAD is calculated using the Equation (22):
d M i , j = cos 1 ( h ( i , j ) ) T h ^ ( i , j ) h ( i , j ) h ^ ( i , j ) = cos 1 h ( i , j ) h ^ ( i , j ) h ^ ( i , j ) T h ^ ( i , j ) h ( i , j ) T h ( i , j )
where h ( i , j ) and h ^ ( i , j ) represent the spectral profiles of the original and reconstructed spectra at various ( i , j ) , respectively, and  d M i , j represents the distance difference map at different ( i , j ) . To obtain detection targets, d M i , j at each position is multiplied by the detection map D ^ , which is weighted by the two coarse detections. However, hyperspectral data often exhibit special cases like spectral variability, where poor reconstruction accuracy can cause localized degradation in background suppression. To address this issue, a non-linear function is introduced to suppress pixels with smaller response values while preserving those with higher response values. The nonlinear function f ( x ) is defined as:
f ( x ) = 1 e β x , x 0 0 , x < 0
where β is a parameter that determines the background suppression capability. Each detected pixel d M i , j is multiplied pixel-wise with the weighted detection map D ^ after being processed through the non-linear variation, as defined by the following Equation (24):
d ^ M i , j = d M i , j f D ^ ( i , j )
By this method, the final result is the detection map R R h × w . The overall performance of the HSI is significantly improved, demonstrating substantial enhancement across different datasets. Algorithm 1 summarizes the pseudocode flow of the SID-DN model that we proposed.
Algorithm 1 SID-DN for HSI Target Detection
Require: HSI tensor X , prior target spectrum d
1:
Compute a coarse detection result D ^ from input B by (5)
2:
Construct the background training set X B from binarized map D and input X by (6)
3:
for each epoch do
4:
    Minimize the loss function L t o t a l in (21)
5:
end for
6:
Reconstruct H ^ with the trained model.
7:
Compute the spectral distance map d M by (22)
8:
Compute R from D ^ and d M by (23) and (24)
Ensure: Target map R

4. Results

This section begins with basic information about the test dataset, followed by the parameter and environment settings for both the proposed and state-of-the-art algorithms. Finally, the two algorithms are compared using objective and subjective metrics.

4.1. Description of Experimental Datasets

In this experiment, three real HTD datasets are selected to validate the proposed algorithm: the Gulfport dataset (part of the ABU dataset), the Los Angeles dataset, and the San Diego dataset. Their basic information is summarized in Table 1. The Gulfport and Los Angeles datasets are derived from the ABU dataset collected by Kang et al. [56], with aircraft as the detection objects. Low signal-to-noise spectra and spectra impacted by water vapor interference, including bands 1–6, 33–35, 97, 107–113, 153–166 and 221–224, are eliminated for the San Diego dataset. The remaining 189-dimensional spectra are used for experiments, with aircraft as the detection objects. The image sizes of the three datasets shown in Table 1 represent cropped dimensions.

4.2. Evaluation Metrics

To comprehensively evaluate HTD performance, this paper uses the 2-D and 3-D Receiver Operating Characteristic Curves (ROC) and the target-background separation map as objective evaluation metrics.
The ROC curve illustrates the relationship between the detection rate P D ( τ ) and the false alarm rate P F ( τ ) as the classification threshold τ varies. The corresponding formula is provided in Equations (25) and (26).
P D ( τ ) = T P T P + F N
P F ( τ ) = F P F P + T N
The detection rate P D ( τ ) is defined as the proportion of positive cases correctly detected at a given threshold τ , as described in Equation (25). The false alarm rate P F ( τ ) is the proportion of incorrectly detected positive cases relative to all negative cases at threshold τ , as shown in Equation (26). An ideal ROC curve approaches the upper left corner (0,1), indicating a high detection rate and a low false alarm rate.
By varying the threshold τ , 2D ROC curves ( τ , P D ( τ ) ) and ( τ , P F ( τ ) ) can be plotted. A faster rise in the 2D ROC curves ( τ , P D ( τ ) ) indicates a higher detection rate, and a faster decline in the 2D ROC curves ( τ , P D ( τ ) ) indicates a lower false alarm rate.
The Area Under the Curve (AUC) represents the area enclosed by the ROC curve and P D ( τ ) , as defined in Equation (27). A larger AUC value indicates a higher detection rate for the model.
AUC = 0 1 P D ( τ ) d P F ( τ )
To further evaluate the target detection performance of the model, Chang et al. [57] proposed a 3D ROC curve detection method. The 3D ROC curve includes five key metrics: AUC P F , P D , AUC τ , P D ( τ ) AUC τ , P F ( τ ) , AUC B S , AUC S N P R . Calculated using Equations (28) and (29):
AUC B S = AUC P F , P D AUC τ , P F ( τ )
AUC S N P R = AUC τ , P D AUC τ , P F ( τ )
The value of AUC P D , P F is interpreted the same way as AUC: a larger value indicates better detection algorithm performance, with an optimal value of 1. AUC P D , P F represents the overall performance of the algorithm and is the most important metric. However, AUC τ , P F ( τ ) is interpreted differently: a smaller value signifies better performance, with the optimal value being 0. a smaller AUC τ , P F ( τ ) value reflects a lower false alarm rate and better background suppression. AUC B S is a composite indicator calculated by weighting and summing the two AUC metrics. Consequently, a larger AUC B S value indicates better detection algorithm performance, with the optimal value being 1. AUC S N P R is inspired by the concept of Signal-to-Noise Ratio (SNR), where AUC P F , P D represents the signal and AUC τ , P F ( τ ) represents the noise. A larger AUC S N P R value indicates better performance of the detection algorithm, with the optimal value being + . The 3D ROC curve provides a three-dimensional visualization, allowing for intuitive assessment of the impact of threshold τ on the detection rate P D ( τ ) and the false alarm rate P F ( τ ) .
The target-background separation map, using the Box-Whisker method, analyzes the separation between target and background. There are two main criteria for evaluation. The first criterion is whether there is any overlap between the target box and the background box. If overlap exists, it indicates that the target and background are mixed during the detection process, making it difficult to distinguish them. The second criterion is based on the area of the background box. The upper and lower bounds of the box represent the 75th and 25th percentiles of the data, respectively. A smaller box indicates that the data are more concentrated, and background suppression is more effective.

4.3. Experimental Environment and Contrasting Models

The hardware used in this experiment includes an i7-12700K CPU, NVIDIA GTX 3090Ti 24 GB graphics card, and 128 GB RAM. The software environment comprises Windows Subsystem for Linux (WSL), PyTorch 1.11.0, and Python 3.8.
To evaluate the efficiency of the algorithms proposed in this paper, traditional target detection methods such as CEM [14], SID [18], MF [54] and sparse representation method as DSE [32], as well as deep learning models like BLTSC [41] and OS-VAE [45], were selected for comparison. All comparison algorithms were configured with the optimal parameters specified in the respective papers.

4.4. Comparative Results on HSI Target Detection

Figure 4 visualizes the detection maps generated by all comparison methods and the proposed method on three real datasets. The results demonstrate that the SID-DN method outperforms the other methods in both target detection and background suppression. Classical detection techniques like MF, CEM and SID exhibit high accuracy in target detection but struggle to filter extensive background information, resulting in insufficient background suppression. Likewise, the DSC method performs poorly in handling complex spectral scenarios, leading to increased false alarm targets. The SID-DN method excels in background suppression, effectively suppressing target-independent background information compared to recently proposed methods such as BLTSC and OS-VAE. BLTSC and OS-VAE, which are generative deep learning models that incorporate background generation representation, slightly lag behind the SID-DN algorithm in background suppression. This confirms the effectiveness of incorporating SID concepts into network design. Furthermore, using SID-DN, particularly on the Los Angeles dataset, the distinction between the target and background is highly evident, demonstrating that the reconstructed background closely resembles the real situation. Additionally, this paper enhances the coarse detection algorithm, significantly reducing target interference in the learning of background samples. The incorporation of SID concepts effectively eliminates potential bad samples, enabling accurate reconstruction of background samples. In summary, the SID-DN method proposed in this paper demonstrates competitive advantages in both target detection and background suppression.
Figure 5 presents the 3D ROC curves corresponding to the above detection results. Compared to the other algorithms, SID-DN achieves results closer to the upper left corner of the test dataset in terms of the AUC ( P D , P F ) metric. However, SID-DN performs slightly worse than DSE and MF in AUC P F , P D . This is because SID-DN focuses on learning background features through reconstruction, but the target spectral reconstruction remains suboptimal, resulting in weaker target intensity in detection and impacting this metric. Conversely, SID-DN is closer to the bottom right corner in the AUC τ , P F ( τ ) curve, indicating its superior background suppression capabilities compared to other methods. This observation highlights the importance of considering multiple metrics in evaluating algorithm performance. In summary, SID-DN achieves a highly advantageous average optimal detection rate across all three datasets.
While ROC curves visually represent the results, accurately assessing them can be challenging due to overlapping curves among the compared methods. Thus, the five AUC metrics across the three datasets are analyzed in detail in Table 2, Table 3 and Table 4.
In Table 2, Table 3 and Table 4, the best results are highlighted in bold. The proposed method achieves the best performance on the most critical metric, AUC ( P D , P F ) demonstrating the superiority of the algorithm. AUC τ , P D ( τ ) typically shows an inverse trend with AUC τ , P F ( τ ) , where a large value of AUC τ , P D ( τ ) corresponds to a small value of AUC τ , P F ( τ ) , or vice versa. Due to limitations in background reconstruction, SID-DN does not achieve the best results in AUC τ , P D ( τ ) . However, it achieves the highest average performance in AUC τ , P F ( τ ) , AUC B S and AUC S N P R , strongly indicating its effectiveness in suppressing the complex backgrounds of HSI. In conclusion, the AUC scores from the comprehensive performance analysis highlight the sophistication and robustness of the proposed SID-DN algorithm.
This paper further evaluates the target-background separation performance of SID-DN compared to other methods across three datasets. As shown in Figure 6, SID-DN achieves superior target-to-background separation ratios, better separation between target and background, and smaller overlapping regions across all datasets. Additionally, SID-DN effectively suppresses background interference, further demonstrating the algorithm’s effectiveness.
In conclusion, the proposed SID-DN method delivers outstanding performance in both subjective and objective metrics, including detection maps and 3D-ROC curves, while demonstrating superior target-background separation capability across multiple datasets.

5. Discussion

5.1. Ablation Comparison Experiment

The SID-DN proposed in this paper introduces two key innovations: the Double-Branch Adaptive Coarse Detection and the integration of the SID concept into Diffusion networks. To validate the effectiveness of these two innovations, ablation experiments were conducted on each dataset, with results summarized in Table 5. In Table 5, the first row of each dataset shows the detection results with a fixed threshold. The second row presents results after applying the Double-Branch Adaptive Coarse Detection, while the third row includes results after applying both the Double-Branch Adaptive Coarse Detection and the incorporation of SID. The results show that the quantitative detection results considering the SID loss outperform the quantitative detection results obtained without considering the SID. This finding highlights the robustness of the Double-Branch Adaptive Coarse Detection and SID loss in enhancing the Diffusion network for reconstructing HSI. By thresholding the adaptive coarse detection and integrating the SID concept into the network, the influence of potential targets on background learning is minimized. This reduces interference from target spectra on background spectra, ensuring a smaller discrepancy between the reconstructed and original background spectra, which enables accurate target-background separation. This work not only introduces novel approaches to the generative coarse detection process but also significantly improves the effectiveness of Diffusion-based generative spectral learning.

5.2. Parametric Analysis

This section discusses key hyperparameters in the SID-DN method, including coarse detection fusion parameters ι , the temporal embedding layer in Diffusion γ te , regularization parameters for SID δ , and background suppression β .
Firstly, selecting appropriate coarse detection fusion parameters is crucial for obtaining satisfactory detection result ι , as these parameters significantly influence the distribution of D ^ in Equation (5). Table 6 presents the detection results obtained with different coarse detection ratios across all datasets. As observed in Table 6, the Los datasets achieve maximum performance at a coarse detection ratio of 0.8, while the Gulfport reach their maximum at 0.6 and San Diego datasets reach their maximum at 0.4. This indicates that varying coarse detection fusion ratios can combine the strengths of both methods, thereby enhancing detection results.
Secondly, this paper examines the impact of the number of temporal embedding nodes γ te on the results. Generally, increasing the number of temporal embedding nodes improves performance; however, an excessive number of nodes can lead to parameter redundancy, ultimately degrading network performance. Therefore, it is important to set a reasonable number of temporal embedding nodes. As shown in Table 7, all datasets achieve optimal performance with 128 nodes. Accordingly, we set the number of temporal embedding nodes to 128 in our experiments.
Thirdly, the regularization parameter δ determines the impact of L s i d on the performance of the network algorithm. An appropriate L s i d suppresses potential targets, enabling effective background learning and ensuring that the reconstruction results successfully separate the target from the background. However, an excessively large L s i d can hinder target reconstruction, reducing performance. Table 8 demonstrates that an optimal balance between the two loss functions is achieved at δ = 0.01 . Consequently, we set δ to 0.01 in our experiments.
During the background suppression phase, selecting an appropriate parameter β effectively suppresses the background while highlighting the target. This parameter primarily reflects the strength of the non-linear suppression function in reducing background interference. As shown in Table 9, the background suppression capability increases with a higher beta initially, but then starts to decline. A smaller β leads to insufficient background suppression, while an excessively high beta causes the non-linear function to lose its ability to suppress, failing to preserve potential targets. Based on comprehensive experimental analysis, we selected β values of 8 for Los Angeles and Gulfport, 12 for San Diego. These values were chosen to achieve optimal performance across the datasets.
To further validate the impact of the diffusion denoising model on hyperspectral target detection, we selected CNN-based Unet network [53] and Vision Transformer [58] as denoising diffusion models, respectively, for comparison with our DiffTransformer. A U C ( P D , P F ) is primarily used to measure detection performance accuracy, while Params is used to measure the number of parameters in deep learning models, and FLOPs are used to assess the model’s computational efficiency and speed. As shown in the Table 10, our method achieves the best results in terms of A U C ( P D , P F ) , Params, and FLOPs, demonstrating the effectiveness of the method we employed.

6. Conclusions

This paper studies the challenges in generative HTD. From a theoretical perspective, we analyze the key aspects of generative HTD and address the shortcomings of the original algorithm. From a technical perspective, we propose a novel deep learning model that integrates SID with Diffusion to effectively guide the reconstruction of the generative model, incorporating SID information constraints. The SID loss and Diffusion process are co-optimized to enhance the reconstruction capability of the generative model. Experiments on multiple real HSI datasets demonstrate the effectiveness and robustness of the SID-DN algorithm. The results indicate that the SID-DN algorithm achieves state-of-the-art performance in both background suppression and detection accuracy when compared to existing methods. Specifically, our method provides an improvement of at least 50% in A U C S N P R . Future work will focus on optimizing the coarse detection process to minimize the influence of hyperparameters and further enhance detection accuracy, improving overall network performance.Furthermore, we plan to explore the potential applications of our proposed algorithm in real-world scenarios, such as military detection and environmental monitoring, as well as attempt multi-modal data fusion with hyperspectral and LiDAR data. Additionally, we aim to investigate target detection in complex scenes using SID-DN in combination with UAV-based hyperspectral imaging systems.

Author Contributions

Conceptualization, J.G.; data curation, X.D. and F.L.; formal analysis, J.G.; methodology, J.G.; software, J.G. and Z.H.; validation, Z.H. and Z.Y.; writing—original draft, J.G.; writing—review and editing, Z.H., X.D., Z.Y. and F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hong, D.; Li, C.; Zhang, B.; Yokoya, N.; Benediktsson, J.A.; Chanussot, J. Multimodal artificial intelligence foundation models: Unleashing the power of remote sensing big data in earth observation. Innovation 2024, 2, 100055. [Google Scholar] [CrossRef]
  2. Nasrabadi, N.M. Hyperspectral target detection: An overview of current and future challenges. IEEE Signal Process. Mag. 2013, 31, 34–44. [Google Scholar] [CrossRef]
  3. Ullah, F.; Ullah, I.; Khan, R.U.; Khan, S.; Khan, K.; Pau, G. Conventional to deep ensemble methods for hyperspectral image classification: A comprehensive survey. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 3878–3916. [Google Scholar]
  4. Lin, S.; Cheng, X.; Zeng, Y.; Huo, Y.; Zhang, M.; Wang, H. Low-rank and sparse representation inspired interpretable network for hyperspectral anomaly detection. IEEE Trans. Instrum. Meas. 2024, 73, 5033116. [Google Scholar]
  5. Chen, B.; Liu, L.; Zou, Z.; Shi, Z. Target detection in hyperspectral remote sensing image: Current status and challenges. Remote Sens. 2023, 15, 3223. [Google Scholar] [CrossRef]
  6. Cheng, G.; Huang, Y.; Li, X.; Lyu, S.; Xu, Z.; Zhao, H.; Zhao, Q.; Xiang, S. Change detection methods for remote sensing in the last decade: A comprehensive review. Remote Sens. 2024, 16, 2355. [Google Scholar] [CrossRef]
  7. Reddy, K.K.; Daduvy, A.; Mohana, R.M.; Assiri, B.; Shuaib, M.; Alam, S.; Sheneamer, A. Enhancing precision agriculture and land cover classification: A self-attention 3D convolutional neural network approach for hyperspectral image analysis. IEEE Access 2024, 12, 125592–125608. [Google Scholar] [CrossRef]
  8. Shen, M.; Zhao, W.; Jiang, N.; Liu, L.; Cao, R.; Yang, W.; Zhu, X.; Wang, C.; Chen, X.; Chen, J.; et al. Challenges in remote sensing of vegetation phenology. Innov. Geosci. 2024, 2, 100070. [Google Scholar]
  9. Li, Q.; Li, J.; Li, T.; Feng, Y. A Joint Framework for Underwater Hyperspectral Image Restoration and Target Detection with Conditional Diffusion Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 17263–17277. [Google Scholar]
  10. Bedini, E. The use of hyperspectral remote sensing for mineral exploration: A review. J. Hyperspectr. Remote Sens. 2017, 7, 189–211. [Google Scholar]
  11. Adetutu, A.E.; Bayo, Y.F.; Emmanuel, A.A.; Opeyemi, A.A.A. A Review of Hyperspectral Imaging Analysis Techniques for Onset Crop Disease Detection, Identification and Classification. J. For. Environ. Sci. 2024, 40, 80. [Google Scholar]
  12. Kruse, F.A.; Lefkoff, A.B.; Boardman, J.W.; Heidebrecht, K.B.; Shapiro, A.; Barloon, P.; Goetz, A.F. The spectral image processing system (SIPS)—interactive visualization and analysis of imaging spectrometer data. Remote Sens. Environ. 1993, 44, 145–163. [Google Scholar]
  13. Fuhrmann, D.R.; Kelly, E.J.; Nitzberg, R. A CFAR adaptive matched filter detector. IEEE Trans. Aerosp. Electron. Syst 1992, 28, 208–216. [Google Scholar]
  14. Farrand, W.H.; Harsanyi, J.C. Mapping the distribution of mine tailings in the Coeur d’Alene River Valley, Idaho, through the use of a constrained energy minimization technique. Remote Sens. Environ. 1997, 59, 64–76. [Google Scholar]
  15. Du, Q.; Chang, C.I. A signal-decomposed and interference-annihilated approach to hyperspectral target detection. IEEE Trans. Geosci. Remote Sens. 2004, 42, 892–906. [Google Scholar]
  16. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Sparse representation for target detection in hyperspectral imagery. IEEE J. Sel. Top. Signal Process. 2011, 5, 629–640. [Google Scholar]
  17. Kraut, S.; Scharf, L.L. The CFAR adaptive subspace detector is a scale-invariant GLRT. IEEE Trans. Signal Process. 1999, 47, 2538–2541. [Google Scholar]
  18. Chang, C.I. An information-theoretic approach to spectral variability, similarity, and discrimination for hyperspectral image analysis. IEEE Trans. Inf. Theory 2000, 46, 1927–1932. [Google Scholar]
  19. Wang, T.; Du, B.; Zhang, L. An automatic robust iteratively reweighted unstructured detector for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2367–2382. [Google Scholar]
  20. Kraut, S.; Scharf, L.L.; McWhorter, L.T. Adaptive subspace detectors. IEEE Trans. Signal Process. 2001, 49, 1–16. [Google Scholar]
  21. Kwon, H.; Nasrabadi, N.M. Kernel adaptive subspace detector for hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2006, 3, 271–275. [Google Scholar] [CrossRef]
  22. Zou, Z.; Shi, Z. Hierarchical suppression method for hyperspectral target detection. IEEE Trans. Geosci. Remote Sens. 2015, 54, 330–342. [Google Scholar] [CrossRef]
  23. Zhao, R.; Shi, Z.; Zou, Z.; Zhang, Z. Ensemble-based cascaded constrained energy minimization for hyperspectral target detection. Remote Sens. 2019, 11, 1310. [Google Scholar] [CrossRef]
  24. Yang, X.; Zhao, M.; Shi, S.; Chen, J. Deep constrained energy minimization for hyperspectral target detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 8049–8063. [Google Scholar] [CrossRef]
  25. Thai, B.; Healey, G. Invariant subpixel material detection in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2002, 40, 599–608. [Google Scholar] [CrossRef]
  26. Cheng, T.; Wang, B. Decomposition model with background dictionary learning for hyperspectral target detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1872–1884. [Google Scholar] [CrossRef]
  27. Harsanyi, J.C.; Chang, C.I. Hyperspectral image classification and dimensionality reduction: An orthogonal subspace projection approach. IEEE Trans. Geosci. Remote Sens. 1994, 32, 779–785. [Google Scholar] [CrossRef]
  28. Chang, C.I.; Chen, J. Orthogonal subspace projection using data sphering and low-rank and sparse matrix decomposition for hyperspectral target detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8704–8722. [Google Scholar] [CrossRef]
  29. Li, W.; Du, Q.; Zhang, B. Combined sparse and collaborative representation for hyperspectral target detection. Pattern Recognit. 2015, 48, 3904–3916. [Google Scholar] [CrossRef]
  30. Liu, R.; Wu, J.; Zhu, D.; Du, B. Weighted Discriminative Collaborative Competitive Representation with Global Dictionary for Hyperspectral Target Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5524613. [Google Scholar] [CrossRef]
  31. Li, Z.; Mu, T.; Wang, B.; Yang, Q.; Dai, H. Background covariance discriminative dictionary learning for hyperspectral target detection. Int. J. Appl. Earth Obs. Geoinf. 2024, 128, 103751. [Google Scholar]
  32. Shen, D.; Ma, X.; Wang, H.; Liu, J. A dual sparsity constrained approach for hyperspectral target detection. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 1963–1966. [Google Scholar]
  33. Freitas, S.; Silva, H.; Almeida, J.M.; Silva, E. Convolutional neural network target detection in hyperspectral imaging for maritime surveillance. Int. J. Adv. Robot. Syst. 2019, 16, 1729881419842991. [Google Scholar] [CrossRef]
  34. Zhang, G.; Zhao, S.; Li, W.; Du, Q.; Ran, Q.; Tao, R. HTD-Net: A deep convolutional neural network for target detection in hyperspectral imagery. Remote Sens. 2020, 12, 1489. [Google Scholar] [CrossRef]
  35. Qin, H.; Xie, W.; Li, Y.; Du, Q. HTD-VIT: Spectral-spatial joint hyperspectral target detection with vision transformer. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 1967–1970. [Google Scholar]
  36. Jiao, J.; Gong, Z.; Zhong, P. Dual-branch Fourier-mixing transformer network for hyperspectral target detection. Remote Sens. 2023, 15, 4675. [Google Scholar] [CrossRef]
  37. Wang, Y.; Chen, X.; Zhao, E.; Zhao, C.; Song, M.; Yu, C. An unsupervised momentum contrastive learning based transformer network for hyperspectral target detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 9053–9068. [Google Scholar]
  38. Jiao, J.; Gong, Z.; Zhong, P. Triplet spectralwise transformer network for hyperspectral target detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5519817. [Google Scholar]
  39. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  40. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114v11. [Google Scholar]
  41. Xie, W.; Zhang, X.; Li, Y.; Wang, K.; Du, Q. Background learning based on target suppression constraint for hyperspectral target detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5887–5897. [Google Scholar]
  42. Xie, W.; Lei, J.; Yang, J.; Li, Y.; Du, Q.; Li, Z. Deep latent spectral representation learning-based hyperspectral band selection for target detection. IEEE Trans. Geosci. Remote Sens. 2019, 58, 2015–2026. [Google Scholar]
  43. Ali, M.K.; Amin, B.; Maud, A.R.; Bhatti, F.A.; Sukhia, K.N.; Khurshid, K. Hyperspectral target detection using self-supervised background learning. Adv. Space Res. 2024, 74, 628–646. [Google Scholar]
  44. Xie, W.; Yang, J.; Lei, J.; Li, Y.; Du, Q.; He, G. SRUN: Spectral Regularized Unsupervised Networks for Hyperspectral Target Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1463–1474. [Google Scholar]
  45. Tian, Q.; He, C.; Xu, Y.; Wu, Z.; Wei, Z. Hyperspectral target detection: Learning faithful background representations via orthogonal subspace-guided variational autoencoder. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5516714. [Google Scholar]
  46. Shen, D.; Ma, X.; Kong, W.; Liu, J.; Wang, J.; Wang, H. Hyperspectral target detection based on interpretable representation network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5519416. [Google Scholar]
  47. Shi, Y.; Lei, J.; Yin, Y.; Cao, K.; Li, Y.; Chang, C.I. Discriminative feature learning with distance constrained stacked sparse autoencoder for hyperspectral target detection. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1462–1466. [Google Scholar]
  48. Shi, Y.; Li, J.; Yin, Y.; Xi, B.; Li, Y. Hyperspectral target detection with macro-micro feature extracted by 3-D residual autoencoder. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4907–4919. [Google Scholar]
  49. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  50. Zeng, H.; Cao, J.; Zhang, K.; Chen, Y.; Luong, H.; Philips, W. Unmixing diffusion for self-supervised hyperspectral image denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 27820–27830. [Google Scholar]
  51. Pang, L.; Rui, X.; Cui, L.; Wang, H.; Meng, D.; Cao, X. HIR-Diff: Unsupervised hyperspectral image restoration via improved diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 3005–3014. [Google Scholar]
  52. Chen, N.; Yue, J.; Fang, L.; Xia, S. SpectralDiff: A generative framework for hyperspectral image classification with diffusion models. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5522416. [Google Scholar]
  53. Ma, J.; Xie, W.; Li, Y.; Fang, L. Bsdm: Background suppression diffusion model for hyperspectral anomaly detection. arXiv 2023, arXiv:2307.09861. [Google Scholar]
  54. Manolakis, D.; Lockwood, R.; Cooley, T.; Jacobson, J. Is there a best hyperspectral detection algorithm? Algorithms Technol. Multispectr. Hyperspectr. Ultraspectr. Imag. XV 2009, 7334, 13–28. [Google Scholar]
  55. Ye, T.; Dong, L.; Xia, Y.; Sun, Y.; Zhu, Y.; Huang, G.; Wei, F. Differential transformer. arXiv 2024, arXiv:2410.05258. [Google Scholar]
  56. Kang, X.; Zhang, X.; Li, S.; Li, K.; Li, J.; Benediktsson, J.A. Hyperspectral anomaly detection with attribute and edge-preserving filters. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5600–5611. [Google Scholar]
  57. Chang, C.I. An effective evaluation tool for hyperspectral target detection: 3D receiver operating characteristic curve analysis. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5131–5153. [Google Scholar]
  58. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
Figure 1. An illustration of the proposed HSI target detection method.
Figure 1. An illustration of the proposed HSI target detection method.
Applsci 15 04076 g001
Figure 2. Schematic of Double-Branch Adaptive Coarse Detection.
Figure 2. Schematic of Double-Branch Adaptive Coarse Detection.
Applsci 15 04076 g002
Figure 3. Schematic of the forward noise addition process and the denoising network in the diffusion model.
Figure 3. Schematic of the forward noise addition process and the denoising network in the diffusion model.
Applsci 15 04076 g003
Figure 4. Pseudo-color representations of the three datasets are shown together with their corresponding ground-truth images and detection results obtained using various methods.
Figure 4. Pseudo-color representations of the three datasets are shown together with their corresponding ground-truth images and detection results obtained using various methods.
Applsci 15 04076 g004
Figure 5. ROC curves of different methods across three datasets: (a) Los Angeles, (b) Gulport, and (c) San Diego. From left to right: 3-D ROC curve, 2-D ROC curve for P F , P D , 2-D ROC curve for τ , P D ( τ ) , and 2-D ROC curve for τ , P F ( τ ) .
Figure 5. ROC curves of different methods across three datasets: (a) Los Angeles, (b) Gulport, and (c) San Diego. From left to right: 3-D ROC curve, 2-D ROC curve for P F , P D , 2-D ROC curve for τ , P D ( τ ) , and 2-D ROC curve for τ , P F ( τ ) .
Applsci 15 04076 g005
Figure 6. Background-target separation plots for various methods across three datasets: (a) Los Angeles, (b) Gulport, and (c) San Diego.
Figure 6. Background-target separation plots for various methods across three datasets: (a) Los Angeles, (b) Gulport, and (c) San Diego.
Applsci 15 04076 g006
Table 1. Basic information on Los Angeles, Gulport and San Diego.
Table 1. Basic information on Los Angeles, Gulport and San Diego.
DatasetLos AngelesGulportSan Diego
SensorAVRISAVRISAVRIS
Spatial resolution (m)7.13.47.5
Spatial Size100 × 100100 × 100100 × 100
Number of bands205191189
Spectral range ( μ m)0.4∼2.50.4∼2.50.37∼2.51
Spectral resolution (nm)101010
Number of Target (Pixels)876057
Table 2. Performance comparison of different methods across various AUC metrics on the Los Angeles dataset.
Table 2. Performance comparison of different methods across various AUC metrics on the Los Angeles dataset.
Method AUC ( P F , P D ) AUC ( τ , P D ) AUC ( τ , P F ) AUC BS AUC SNPR Running Time
CEM [14]0.99290.36180.01760.975320.490.14
SID [18]0.84990.20150.03610.81385.5720.09
MF [54]0.99580.43060.10150.89434.240.08
DSE [32]0.94700.47250.07620.87086.1921.92
BLTSC [41]0.99440.33530.00110.9933297.952.83
OS-VAE [45]0.99430.33110.00100.9933318.382.69
Proposed0.99670.31460.00070.9960410.530.37
Note: The best results are highlighted in bold.
Table 3. Performance comparison of different methods across various AUC metrics on the Gulfport dataset.
Table 3. Performance comparison of different methods across various AUC metrics on the Gulfport dataset.
Method AUC ( P F , P D ) AUC ( τ , P D ) AUC ( τ , P F ) AUC BS AUC SNPR Running Time
CEM [14]0.99930.50050.02250.976822.180.007
SID [18]0.94890.08000.00500.943913.430.09
MF [54]0.99920.54350.12860.87064.220.023
DSE [32]0.99000.56700.05250.937510.7818.78
BLTSC [41]0.99900.36710.00080.9982445.182.57
OS-VAE [45]0.99920.33810.00050.9987645.462.7
Proposed0.99940.35470.00020.99921742.730.47
Note: The best results are highlighted in bold.
Table 4. Performance comparison of different methods across various AUC metrics on San Diego dataset.
Table 4. Performance comparison of different methods across various AUC metrics on San Diego dataset.
Method AUC ( P F , P D ) AUC ( τ , P D ) AUC ( τ , P F ) AUC BS AUC SNPR Running Time
CEM [14]0.99550.62650.03950.956015.820.007
SID [18]0.99160.15880.00350.988144.270.083
MF [54]0.99620.70930.21400.78223.300.024
DSE [32]0.99410.64470.04550.948614.15271.08
BLTSC [41]0.99610.41670.00190.9942215.572.54
OS-VAE [45]0.99510.34540.00130.9938253.72.7
Proposed0.99690.37650.00020.99671409.550.35
Note: The best results are highlighted in bold.
Table 5. An ablation study was conducted on three datasets to evaluate the effectiveness of the proposed method.
Table 5. An ablation study was conducted on three datasets to evaluate the effectiveness of the proposed method.
DatasetDBACD L _ sid AUC ( P D , P F ) AUC _ ( τ , P D ) AUC ( τ , P F ) AUC BS AUC SNPR
Los Angeles××0.99420.36680.00280.9914126.77
×0.99600.32950.00070.9953430.21
0.99650.32430.00060.9959519.54
Gulfport××0.99920.36680.0050.994263.50
×0.99930.35810.00060.9987559.50
0.99940.32820.00010.99931997.28
San Diego××0.99470.33840.00490.989867.89
×0.99610.33250.00220.9939147.81
0.99690.37650.00020.99671409.55
Note: The best results are highlighted in bold.
Table 6. Sensitivity analysis of ι was conducted on three datasets.
Table 6. Sensitivity analysis of ι was conducted on three datasets.
ι 0.10.20.30.40.50.60.70.80.9
Los Angeles0.99490.99540.99610.99630.99650.99660.99660.99670.9967
Gulport0.99910.99930.99920.99930.99920.99940.99920.99870.9985
San Deigo0.99560.99590.99610.99690.99580.99560.99520.99450.9933
Note: The best results are highlighted in bold.
Table 7. Sensitivity analysis of γ te was conducted on three datasets.
Table 7. Sensitivity analysis of γ te was conducted on three datasets.
Dataset3264128256512
Los Angeles0.99640.99650.99660.99670.9964
Gulport0.99930.99920.99940.99880.9984
San Deigo0.95840.96770.99690.99590.9861
Note: The best results are highlighted in bold.
Table 8. Sensitivity analysis of δ was conducted on three datasets.
Table 8. Sensitivity analysis of δ was conducted on three datasets.
Dataset 1 × 10 4 1 × 10 3 1 × 10 2 1 × 10 1 1
Los Angeles0.99600.99600.99620.99650.9964
Gulport0.99940.99940.99940.99940.9993
San Deigo0.99640.99640.99630.99690.9966
Note: The best results are highlighted in bold.
Table 9. Sensitivity analysis of β was conducted on three datasets.
Table 9. Sensitivity analysis of β was conducted on three datasets.
Dataset48121620
Los Angeles0.99640.99670.99620.99530.9938
Gulport0.99930.99940.99930.99930.9992
San Deigo0.8420.98220.99070.99010.9969
Note: The best results are highlighted in bold.
Table 10. The performance comparison of different denoising diffusion models three datasets.
Table 10. The performance comparison of different denoising diffusion models three datasets.
Dataset U-NetVITSID-DN
Los AngelesParams1.15 M0.73 M0.46 M
FLOPs1.1 G0.75 G0.47 G
A U C ( P D , P F ) 0.99600.99610.9965
GulportParams1.13 M0.71 M0.44 M
FLOPs1.1G0.73 G0.45 G
A U C ( P D , P F ) 0.99920.99920.9994
SandiegoParams1.13 M0.83 M0.43 M
FLOPs1.1 G0.85 G0.44 G
A U C ( P D , P F ) 0.99580.99540.9969
Note: The best results are highlighted in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gong, J.; Huang, Z.; Yang, Z.; Ding, X.; Li, F. Spectral Information Divergence-Driven Diffusion Networks for Hyperspectral Target Detection. Appl. Sci. 2025, 15, 4076. https://doi.org/10.3390/app15084076

AMA Style

Gong J, Huang Z, Yang Z, Ding X, Li F. Spectral Information Divergence-Driven Diffusion Networks for Hyperspectral Target Detection. Applied Sciences. 2025; 15(8):4076. https://doi.org/10.3390/app15084076

Chicago/Turabian Style

Gong, Jinfu, Zhen Huang, Zhengye Yang, Xuezhuan Ding, and Fanming Li. 2025. "Spectral Information Divergence-Driven Diffusion Networks for Hyperspectral Target Detection" Applied Sciences 15, no. 8: 4076. https://doi.org/10.3390/app15084076

APA Style

Gong, J., Huang, Z., Yang, Z., Ding, X., & Li, F. (2025). Spectral Information Divergence-Driven Diffusion Networks for Hyperspectral Target Detection. Applied Sciences, 15(8), 4076. https://doi.org/10.3390/app15084076

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop