Highlights
What are the main findings?
- Developed a hyperspectral tracker with band-importance modeling and contextual feature refinement.
- Uses importance-guided spectral grouping to enhance tracking accuracy.
What is the implication of the main finding?
- Its spectral and contextual design helps track objects more accurately in complex scenes.
- Improves robustness of target localization in challenging hyperspectral videos.
Abstract
The scarcity of labeled hyperspectral video samples has motivated existing methods to leverage RGB-pretrained networks; however, many existing methods of hyperspectral object tracking (HOT) select only three representative spectral bands from hyperspectral images, leading to spectral information loss and weakened target discrimination. To address this issue, we propose the Band and Context Refinement Network (BCR-Net) for HOT. Firstly, we design a band importance learning module to partition hyperspectral images into multiple false-color images for pre-trained backbone network. Specifically, each hyperspectral band is expressed as a non-negative linear combination of other bands to form a correlation matrix. This correlation matrix is used to guide an importance ranking of the bands, enabling the grouping of bands into false-color images that supply informative spectral features for the multi-branch tracking framework. Furthermore, to exploit spectral–spatial relationships and contextual information, we design a Contextual Feature Refinement Module, which integrates multi-scale fusion and context-aware optimization to improve feature discrimination. Finally, to adaptively fuse multi-branch features according to band importance, we employ a correlation matrix-guided fusion strategy. Extensive experiments on two public hyperspectral video datasets show that BCR-Net achieves competitive performance compared with existing classical tracking methods.
1. Introduction
Hyperspectral object tracking (HOT) constitutes a key problem in computer vision, with broad applications in intelligent surveillance [], traffic monitoring [], and medical imaging analysis []. Given an initial target annotation in the first frame, a tracker must localize the target throughout subsequent frames despite appearance variation, motion dynamics, partial occlusion, illumination changes, and background clutter. While deep RGB-based trackers have achieved notable advances [,,], their reliance on only three broad spectral channels inherently limits discriminability, especially in low-contrast scenes, camouflage settings, and spectrally ambiguous backgrounds frequently encountered in remote sensing and geoscience monitoring.
Hyperspectral imaging (HSI) acquires contiguous narrow-band reflectance measurements, often spanning on the order of tens to over one hundred bands, providing rich joint spatial–spectral information capable of distinguishing targets with similar spatial texture but differing spectral signatures []. This fine-grained spectral resolution is particularly advantageous in scenarios involving small or partially occluded objects embedded in spectrally heterogeneous or cluttered environments. Consequently, HOT has emerged as a promising extension of traditional RGB tracking for improved robustness and semantic discrimination in complex environments.
A central challenge in hyperspectral object tracking is obtaining features that reliably distinguish the target from its background under varying conditions. Early HOT methods (such as references [,,]) mainly rely on manually designed features to construct the spectral or spatial expression of the target, such as material abundance, spectral gradient histogram or specific convolution kernel. While these feature representation strategies achieved successes, they lack adaptability, and it is difficult for them to handle rapid target appearance changes or severe occlusions. In contrast, deep learning has transformed RGB tracking by leveraging large-scale annotated datasets [,] to learn hierarchical, discriminative representations.
Due to the scarcity of large labeled hyperspectral video datasets, directly training deep HOT architectures end-to-end often leads to overfitting and unstable optimization. To mitigate this data limitation, a prevalent line of work transfers RGB-pretrained backbones (e.g., GOT-10K [], TrackingNet [], LaSOT []) to hyperspectral scenarios by first converting the high-dimensional hyperspectral data into one or more three-channel false-color composites []. Single-composite methods usually follow two paradigms: (i) selecting three key bands using conventional techniques for dimensionality reduction, such as PCA, or via learning-based selection networks [], and (ii) constructing a three-channel surrogate image by replicating existing bands. Specifically, band selection is sensitive and may exclude narrow yet discriminative spectral responses critical for small or low-contrast targets. Replication yields zero inter-channel diversity, preventing the pretrained filters from exploiting cross-channel interactions. Adjacent or sequential grouping preserves high local spectral correlation, causing redundant channel responses and underutilizing potential complementary cues. PCA emphasizes directions of maximal global variance; thus, minority-band features with high target–background discriminability but low energy risk being attenuated or entangled in principal components dominated by background structure. However, the above methods collectively reduce spectral separability, distort the statistical assumptions underlying RGB pretraining, and constrain downstream feature discrimination.
Alternatively, hyperspectral bands can be randomly grouped to form false-color composites, with the intention of reducing inter-channel correlation. However, this strategy treats all bands as equally important. In reality, different wavelengths vary significantly in their contribution to hyperspectral image formation and downstream tasks, as demonstrated by band selection studies []. Consequently, random grouping often mixes weakly informative or noisy bands with discriminative ones, which negatively impacts hyperspectral tracking and increases training difficulty. To mitigate this problem, Li et al. introduced the Band Attention Ensemble Network (BAE-Net) [], which employs a band attention module to estimate spectral importance and groups bands accordingly before constructing false-color images. This line of work highlights that incorporating band importance into grouping leads to more reliable representations than purely random partitioning.
Building on the idea of leveraging multiple spectral views, recent multi-branch trackers such as SST-Net [] and BRRF-Net [] extend this paradigm by generating several composites and processing them in parallel. Although methods like BAE-Net already consider band importance during grouping, most of these trackers still fuse the outputs of different branches through uniform averaging, effectively assuming equal reliability across composites. In practice, a branch dominated by weak or interference-prone bands can produce an outlier response map; equal weighting then propagates this error and increases the risk of drift. Compounding the issue, adverse imaging factors (e.g., motion blur, defocus, low illumination) erode high-level spatial structure and channel discriminability. Existing refinement modules rarely perform coupled spatial–channel enhancement, limiting their ability to recover degraded details. Although adaptive update mechanisms alleviate gradual appearance evolution, they do not explicitly quantify band importance or implement reliability-aware suppression of low-quality branches.
Despite these advances, hyperspectral tracking still faces two key challenges. First, severe spectral redundancy can cause the tracker to overemphasize noisy or weakly discriminative bands, thereby diluting target–background contrast and increasing false responses. Second, spatial degradation from factors such as blur or low resolution weakens high-level structural integrity, raising the risk of localization drift. We tackle these issues with a Band and Context Refinement Network (BCR-Net), which unifies explicit spectral importance estimation, contextual feature enhancement, and reliability-aware fusion. Specifically, the Band Importance Modeling Module (BIMM) leverages the linear self-representation property of hyperspectral imagery: a U-Net-based unrolled gradient descent solver produces a sparse band coefficient matrix that captures complementary relations while suppressing redundancy. This matrix is then used to (i) quantify per-band importance and (ii) construct multiple complementary false-color composites, allowing an RGB-pretrained backbone to harvest transferable semantic and spatial priors. Building upon the resulting high-level features, the Contextual Feature Refinement Module (CFRM) applies cross-dimensional attention to jointly enhance spatial detail fidelity and channel-wise discriminability. Finally, we perform dynamic reliability-aware fusion: branch-level tracking outputs are adaptively weighted by importance scores derived from the coefficient matrix, effectively attenuating unstable predictions from low-importance or interference-prone composites.
The main contributions of this study can be summarized as follows:
- We propose a unified hyperspectral tracking framework that integrates spectral band importance modeling with RGB-pretrained backbones. By selectively emphasizing informative bands and guiding feature construction, the framework achieves stable representations, efficient adaptation, and enhanced discriminative power during tracking.
- We reformulate the spectral self-expression prior into a parameterized module that learns band-wise importance and guides feature construction, explicitly addressing spectral redundancy and improving the discriminative power of the features.
- We employ a Contextual Feature Refinement Module to enhance multi-scale feature representations, improving discriminative power for target–background separation.
- Extensive experiments on two public hyperspectral video datasets demonstrate that BCR-Net is competitive with existing methods, achieving strong performance in both tracking accuracy and robustness.
2. Related Works
Current hyperspectral object tracking (HOT) methods have primarily evolved along two directions: correlation filter-based (CF) approaches and deep learning-based (DL) approaches. In the following, we review each category, highlighting their key techniques, advantages, and limitations.
2.1. Correlation Filter-Based HOT
CF methods constitute one of the earliest and most widely adopted strategies for HOT. These methods track targets by learning correlation filters that model the relationship between the target’s spectral–spatial characteristics and its surrounding context. Qian et al. [] proposed using normalized 3D spectral–spatial cubes as convolution filters to capture local spectral–spatial patterns for hyperspectral object tracking. Zhang et al. [] introduced a method that combines convolutional features from VGGNet with HOG descriptors and fuses their responses through correlation filters to improve tracking robustness under complex backgrounds. To address occlusions and illumination variations, Guo et al. [] developed a method that incorporates adaptive weighting and background-aware correlation filters to suppress anomalies in response maps.
To enhance feature extraction efficiency, Chen et al. [] proposed using fast spatial–spectral convolution kernels with closed-form solutions in the Fourier domain for real-time spectral–spatial feature learning. A tensor-based sparse correlation filter incorporating spatial–spectral weighted regularization was proposed by Hou et al. [] to reduce background interference and enhance spectral discriminability. Building on these approaches, Lou et al. [] proposed combining deep and spatial–spectral features by extracting 3D gradient-based spatial features through tensor singular spectrum analysis and convolutional features from a VGGNet pretrained on RGB images, thereby improving tracking accuracy by exploiting the complementarity of deep and enhanced spectral–spatial representations.
Overall, CF methods for HOT provide advantages such as low computational overhead, ease of implementation, and effective use of spectral–spatial priors. Nevertheless, their heavy reliance on handcrafted or predefined features limits adaptability in challenging scenarios, including occlusions, rapid appearance variations, and cluttered environments.
2.2. Deep Learning-Based HOT
DL methods have recently emerged as a major direction in HOT, offering the ability to learn adaptive spectral–spatial representations. The critical challenge, however, lies in the scarcity of labeled hyperspectral video data, which makes it impractical to train large-scale backbone networks tailored specifically for HOT. To mitigate this limitation, many studies transfer knowledge from RGB-pretrained models by dividing hyperspectral bands into groups and generating false-color images for training. Uzkent et al. [] proposed reducing hyperspectral data to three channels and applying an RGB-pretrained VGGNet to extract robust features for target representation. Wang et al. [] proposed a strategy that selects representative bands using joint entropy and utilizes RGB-pretrained Siamese networks to achieve effective semantic feature extraction under constrained training conditions.
Building upon RGB-pretrained backbones, many existing methods have specifically utilized spectral–spatial correlations to enhance feature discrimination. Specifically, Wu et al. [] reorganized hyperspectral bands into groups based on learned correlations and fused them for robust representation, but the final fusion relied on uniform averaging without fully exploiting learned band importance. Liu et al. [] designed a dual Siamese framework with a hyperspectral target-aware module and spatial–spectral cross-attention for adaptive fusion, yet lacked reliability-aware weighting and coupled spatial–channel refinement. Zhao et al. [] developed a transformer-based fusion network to capture intra- and inter-band interactions across modalities, effectively integrating hyperspectral and RGB information. Chen et al. [] proposed a spectral awareness module with feature interaction to model band dependencies and strengthen fusion, but did not incorporate importance learning into a unified fusion strategy. Jiang et al. [] introduced a channel-adaptive dual Siamese architecture to flexibly handle sequences with varied spectral channels. Xiong et al. [] introduced SSTCF, which employs spatial–spectral features and a low-rank constraint to enhance global spectral structure, while incorporating temporal consistency to maintain filter stability across consecutive frames. Gao et al. [] proposed SP-HST, a parameter-efficient method that freezes a pre-trained RGB network and trains only a lightweight branch to generate spectral prompts, which provide complementary HSI information fused with the frozen RGB features for the final tracking decision. Wang et al. [] presented SSF-Net, which employs spectral angle awareness to capture region-level spectral relationships, achieving more robust spatial–spectral representations for hyperspectral tracking.
Overall, DL-based HOT methods have demonstrated strong effectiveness in leveraging spectral–spatial fusion and knowledge transfer from pre-trained RGB models. Although the above-mentioned methods adopt band regrouping to improve hyperspectral tracking, they often fail to fully exploit the varying importance of spectral bands and suffer from degraded feature discrimination under complex conditions, which limits their efficiency and robustness. Therefore, in this paper, we propose BCR-Net to address these limitations through BIMM and CFRM, enabling adaptive band grouping and contextual refinement.
3. Proposed Method
Here, we describe the Band and Context Refinement Network (BCR-Net), which combines the Band Importance Modeling Module (BIMM) with the Contextual Feature Refinement Module (CFRM).
3.1. Overall Architecture
Since hyperspectral images contain high-dimensional spectral information that cannot be directly processed by conventional RGB-based object trackers, we propose the BCR-Net for hyperspectral tracking. It first learns the importance of each spectral channel and groups the hyperspectral data into multiple false-color images accordingly. Semantic features at a high level are extracted from the grouped images using a pretrained RGB backbone. Finally, the features from different groups are adaptively fused using importance-guided weighting, enabling robust spectral–spatial representation and accurate target localization.
Figure 1 illustrates the overall framework, which adopts a Siamese network architecture and includes three main components: (i) a dynamic spectral feature learning module, (ii) a backbone network for feature extraction, and (iii) an integrated prediction module. Input data is fed into two parallel branches: the template branch representing the initial target patch, and the search branch containing candidate regions from later frames.
Figure 1.
Overall framework of the proposed BCR-Net method. We adopt a Siamese network architecture and include three main components: (i) a dynamic spectral feature learning module, for adaptively modeling inter-band relationships and enhancing discriminative spectral representations; (ii) a backbone network for feature extraction, to capture robust multi-scale spectral–spatial features of both template and search regions; (iii) an integrated prediction module, for fusing multi-level features and generating precise target localization and scale estimation.
The dynamic spectral feature learning module first analyzes inter-band relationships and channel importance through an auxiliary reconstruction task. This step projects the original C-dimensional spectral cube into multiple three-channel false-color images that preserve discriminative information while reducing redundancy. These false-color groups are then fed into a pretrained RGB backbone (e.g., ResNet-50), shared by both template and search branches, to obtain deep semantic features. On top of the backbone output, a context-aware feature refinement module further processes the features to capture fine-grained spatial details and inter-channel correlations, ensuring the extracted representations remain robust for subsequent tracking.
Finally, the integrated prediction module aggregates the similarity maps produced by different false-color branches. To balance their contributions, each branch is adaptively weighted according to the spectral importance scores estimated in the dynamic spectral feature learning stage. The response map guides target localization, complemented by a dynamic template update that adjusts for temporal appearance changes.
3.2. Band Importance Modeling Module
Hyperspectral data are usually converted into three-channel composites to fit RGB-pretrained CNNs, but sequential grouping produces highly correlated channels, and independent processing overlooks spectral structure, both of which weaken discriminability. To overcome these issues, we adopt an importance-guided grouping strategy that ranks bands by significance and assembles them into complementary composites, reducing redundancy while preserving informative cues for robust multi-branch tracking.
For a hyperspectral image , we model intrinsic dependencies between spectral bands by leveraging the prior that each band can be roughly represented as a linear combination of the others. Accordingly, we learn a band correlation matrix via a spectral self-expression model, expressed as a non-negative reconstruction coefficient matrix . Each row of indicates how a band is reconstructed from the others, capturing inter-band dependencies, and summing the coefficients provides a quantitative measure of band importance, with larger values highlighting more discriminative spectral channels. Mathematically, the spectral self-expression model is formulated as
where denotes the Frobenius norm. The non-negativity constraint ensures physically meaningful importance values, while the zero-diagonal constraint prevents self-reconstruction and stabilizes the estimated band importance.
To implement the spectral self-expression problem in Equation (1), we employ an encoder–decoder architecture based on U-Net. Figure 2 depicts the connection between the last encoder block and the first decoder block, as well as the internal structure of each submodule. The reconstruction coefficient matrix is approximated by aggregating learned spectral–spatial features in a differentiable, end-to-end manner. The encoder captures multi-scale spectral–spatial representations, encoding both local patterns and global inter-band dependencies, while channel and spatial enhancement modules emphasize informative features. The decoder restores spatial resolution and fuses multi-scale features via a Semantic-Guided Refinement Module, where high-level features guide adaptive weighting of low-level features. From the decoder output, a global pooling operation aggregates feature correlations to produce an initial channel weight matrix , providing a coarse approximation of the reconstruction coefficients . Non-negativity and zero-diagonal constraints are enforced via simple activation and masking, and a lightweight Channel Weight Refinement module further optimizes to emphasize informative bands while suppressing redundancies.
Figure 2.
Illustration of the U-Net-based Band Importance Modeling Module. The top-left subfigure shows the connection between the last encoder block and the first decoder block, while the other subfigures depict detailed designs of its submodules.
3.2.1. Encoder of BIMM
In hyperspectral image reconstruction, the encoder must model both spectral and spatial correlations to extract discriminative features that distinguish the object from its background. We integrate two complementary modules: the Channel Enhancement Module (CEM) for modeling spectral dependencies and alleviating inter-band redundancy, and the spatial enhancement module (SEM) for identifying important spatial regions and enhancing structural details. This design adaptively enhances informative features while suppressing redundancy, improving discrimination and robustness.
Given an input feature map , CEM first applies adaptive average pooling and max pooling to obtain global channel descriptors and . Each descriptor is then passed through two fully connected layers: the first reduces the channel dimension to (with ), and the second restores it, with a ReLU activation in between. The outputs are summed and passed through a Softplus activation to generate the channel weight matrix , which is broadcasted to match the original feature dimensions and multiplied element-wise (⊙) with :
While CEM captures spectral relationships, it does not directly model spatial structures. To address this, the SEM module enhances spatial saliency by highlighting important positions and suppressing irrelevant regions. For an input , SEM computes channel-wise max and average maps, and , concatenates them, and applies a convolution with padding, followed by a Sigmoid activation () to generate the spatial weight map . The output is then obtained via element-wise multiplication (⊙) with :
where denotes concatenation along the channel dimension.
By stacking multiple 3D convolution layers, each followed by SEM and CEM, the encoder progressively captures and enhances spectral–spatial features. Specifically, SEM strengthens important spatial structures, while CEM emphasizes discriminative spectral channels. The output of the l-th encoder layer is expressed as
This design combines convolution with spectral–spatial enhancement in a complementary way, leading to more discriminative feature representations and laying a strong foundation for decoding and reconstruction.
3.2.2. Decoder of BIMM
A high-quality decoder is essential for restoring spatial resolution, reconstructing boundary details, and maintaining semantic consistency. In this work, the decoder addresses both feature reconstruction and semantic refinement by integrating two complementary modules. The Shuffle-Driven Upsampling Block (SDUB) progressively upsamples features while preserving fine-grained spatial–spectral information, ensuring accurate reconstruction of target details. The Semantic-Guided Refinement Module (SGRM) captures global spectral–spatial dependencies and enhances salient semantic features, thereby improving the overall discriminability of the reconstructed output. This design effectively recovers feature scales, fuses cross-level information, and enhances salient regions.
The SDUB is designed to efficiently enhance spatial resolution while preserving feature quality. It takes either the last encoder output (for the first decoder layer) or the previous decoder feature map (for subsequent layers) and first performs bilinear interpolation to double the spatial size. Local channel features are then extracted via a depthwise separable convolution (), followed by a channel rearrangement to promote cross-channel fusion, and finally a convolution adjusts the output to the target channel dimension . Formally, the operation on is expressed as
While SDUB efficiently upsamples feature maps, effectively fusing multi-level features remains challenging. Low-level decoder features contain fine-grained details but are noisy and semantically weak, whereas high-level encoder features provide rich semantic context but suffer from reduced spatial resolution and blurred boundaries. Simple concatenation or addition often preserves noise and dilutes semantics, resulting in suboptimal reconstruction.
In response, we introduce the Semantic-Guided Refinement Module (SGRM) to refine low-level decoder features using high-level encoder information. Specifically, for the l-th decoding layer, let and denote the encoder and previous decoder features after SDUB upsampling. Both are processed by grouped convolutions with batch normalization, spatially aligned via bilinear interpolation if needed, and fused through element-wise addition followed by ReLU to form the intermediate attention tensor :
where / are grouped convolutions.
The intermediate tensor passes through a convolution with batch normalization and Sigmoid activation, and the resulting spatial attention map modulates via element-wise multiplication to produce the fused decoder feature:
The decoder reconstructs multi-level semantic information by applying SDUB and SGRM in a nested manner:
In this structure, SDUB upsamples the previous decoder output, while SGRM fuses it with the corresponding encoder features, enabling efficient multi-scale feature reconstruction with enhanced spatial details and semantic consistency.
3.2.3. Channel Weight Refinement of BIMM
To capture inter-channel dependencies in the decoder features, we apply global average pooling () to the decoder output to obtain a channel-wise summary. The correlation matrix of these pooled features is then mapped through a Softplus function to produce the initial channel weight matrix :
where captures the similarity between different feature channels.
However, is a heuristic estimate that may contain noisy or suboptimal correlations, limiting its ability to highlight discriminative channels. To address this, we propose a single-step structural optimization module that refines while preserving its non-negativity and zero-diagonal constraints. A lightweight MLP () predicts a gradient-like update for each element, simulating conventional gradient descent in a learnable, end-to-end manner. The refined weight matrix is computed as
where is a diagonal suppression function that explicitly zeros out the diagonal elements of the matrix.
This single-step optimization captures the main structural patterns in channel correlations, promoting sparsity and highlighting important channels. By refining into a task-aligned, discriminative , the decoder emphasizes informative channels and suppresses irrelevant ones, enhancing both feature discrimination and robustness while remaining consistent with the initial estimation.
3.3. Contextual Feature Refinement Module
Previous studies by Chen et al. [] show that different backbone layers capture complementary information: shallow layers (e.g., conv3 in -50) retain high spatial resolution and encode local structures, aiding precise target localization, while deeper layers (e.g., conv4 and conv5) provide strong semantic abstraction, capturing high-level concepts and appearance variations. Accordingly, we select conv3, conv4, and conv5 for multi-scale feature fusion, forming a hierarchically complementary framework. To enhance context perception without increasing parameters, dilated convolutions with rates 2 and 4 are applied in conv4 and conv5, respectively. Spatial downsampling is removed in these stages to maintain consistent spatial sizes, ensuring alignment for multi-layer fusion and preserving both fine-grained details and semantic richness.
We further introduce a context-aware feature optimization module to enhance separability between template and search features. Channel-wise semantics are captured via global average and max pooling, while spatial dependencies are modeled by compressing the feature map into query and key matrices to compute a correlation score. A learnable convolution propagates contextual information, and the resulting channel and spatial cues are combined via element-wise Hadamard product to highlight salient regions and suppress background noise. Formally, this process is expressed as
where and are the input and output at the j-th pixel, N denotes the total number of pixels, and are learnable convolution layers, and compute global average and maximum pooling, is applied element-wise, and indicates concatenation along the channel dimension.
By jointly modeling channel semantics and spatial dependencies in this integrated manner, the Contextual Feature Refinement Module strengthens salient target representations and suppresses redundant background responses, providing robust, contextually enriched features for subsequent multi-branch tracking and prediction.
3.4. Training and Tracking
3.4.1. Offline Training
During training, the backbone -50 remains frozen, while the upstream modules—including the dynamic spectral feature learning module and multiple Box-Adaptive prediction heads—are optimized jointly in an end-to-end manner. Training pairs are constructed by sampling template and search regions from hyperspectral videos. The template and search regions are cropped from the original images and resized to and , respectively, before being fed into the network.
To exploit multi-scale information from different backbone stages, a hierarchical fusion strategy is applied. Feature maps from conv3, conv4, and conv5 are extracted for classification and regression, and their outputs are combined using learnable weights and to produce the final multi-level response maps:
where and denote the classification and regression outputs from the k-th convolutional block, and the weights and (initialized to 1) control each layer’s contribution.
For label assignment, we follow the ellipse sampling strategy []. In the search frame, the target is represented by a bounding box with top-left , center , and bottom-right coordinates, having width and height . Two ellipses are generated at the box center: with semi-axes and , and with semi-axes and :
where represents a candidate position. Points within are labeled positive, points outside negative, and those in between are ignored to prevent ambiguous supervision.
For each positive point, the regression target encodes its horizontal and vertical offsets relative to the left, top, right, and bottom boundaries of the ground-truth box:
Here, and indicate the horizontal distances to the left and right edges of the ground-truth box, while and correspond to the vertical distances to the top and bottom edges, respectively.
To jointly capture hyperspectral characteristics and dynamic target behavior, a multi-task loss function is adopted, combining spectral reconstruction, classification, and regression objectives:
where Z and X correspond to the features extracted from the template and search regions, respectively. and denote the band correlation matrices for the template and search regions, respectively. represents the importance weight of the n-th sample, and balance the classification and regression losses, and controls the influence of spectral reconstruction. The first term in the loss guides the dynamic spectral feature learning module to capture physically meaningful band relationships, while the second and third terms supervise classification and precise bounding box regression.
3.4.2. Online Tracking
The initial target template is extracted from the first frame to initialize the tracking process. Using the weights provided by the dynamic spectral feature learning module, the spectral bands are ordered and grouped to generate multiple false-color images. Subsequently, these template images are fed into the contextual feature module to extract multi-scale target features, which serve as references for tracking in the following frames.
For each new frame, the search region is extracted based on the previous target state. Its spectral bands are similarly ranked and grouped into false-color search images . The feature extraction network generates feature maps for each search image, which are paired with the corresponding template feature maps . Each pair is fed into the Box-Adaptive Head to produce classification scores and regression offsets :
where denotes a convolution, and ⋆ is cross-correlation operation. Each response map provides an independent target estimate, and the final position is determined by fusing these estimates based on the relative importance of each spectral band.
Target localization is achieved by finding the location with the highest response in the classification map , and reconstructing the bounding box using the corresponding regression values . To avoid unstable predictions, a cosine window is applied to discourage large positional shifts, and a scale regularization is introduced to prevent abrupt size fluctuations. These strategies collectively improve robustness and maintain tracking accuracy across frames.
4. Experiments
This section evaluates the proposed hyperspectral tracker against existing mainstream methods to demonstrate its effectiveness. Furthermore, we perform a detailed ablation study to assess the impact of each proposed module on overall tracking performance.
4.1. Experimental Setup
4.1.1. Dataset Description
To comprehensively assess performance across diverse scenarios, we employ two benchmark datasets.
The hyperspectral object tracking Competition (HOTC) benchmark [] comprises 75 video sequences, split into 40 for training and 35 for testing. Each sequence averages 390 frames at 25 FPS, providing three data modalities: hyperspectral (16 bands), false-color (3 bands), and RGB (3 bands). The dataset includes attribute annotations for 11 tracking challenges: fast motion, background clutter, illumination variation, out-of-plane rotation, occlusion, motion blur, low resolution, scale variation, in-plane rotation, out-of-view, and deformation.
The Multispectral UAV-based Spatio-temporal Tracking Dataset (MUST) [] features eight spectral channels (390–950 nm) captured at 1280 × 960 resolution and 5 FPS. This UAV-focused dataset balances spectral richness with temporal consistency, incorporating 12 challenging attributes: partial occlusion, background clutter, low resolution, similar object interference, scale variation, motion blur, fast motion, similar color, out-of-view, illumination variation, full occlusion, and camera motion.
4.1.2. Network Implementation
The tracking framework is implemented in Python 3.7.16 and evaluated on a machine with an Intel® Xeon® Silver 4310 CPU and NVIDIA RTX A6000 GPU. We adopt the optimizer configuration from SiamBAN [], setting the implicit momentum coefficient to 0.028. Training proceeds for 60 epochs using batch size 56, with learning rate initialized at 0.003 and decayed via standard scheduling for stable convergence.
4.1.3. Evaluation Indexes
We evaluate tracking performance using the One Pass Evaluation (OPE) protocol with two key metrics: area under the success rate curve (AUC), which measures average bounding box overlap across all frames, and distance precision (DP) at 20-pixel threshold, representing the percentage of frames where predicted centers fall within 20 pixels of ground truth. Additionally, we include frames per second (FPS), floating point operations (FLOPs), and model size as extra indicators to analyze the computational complexity of the algorithms. FPS reflects the computational efficiency of the tracker and indicates its practical suitability for real-time applications; FLOPs quantify the amount of computation required by the model; model size measures the memory footprint of the model, which is important for deployment efficiency.
4.2. Experimental Results Comparison
4.2.1. Experiments with the HOTC Dataset
To comprehensively assess the effectiveness of BCR-Net, we conduct evaluations on the HOTC dataset and compare its performance against a set of representative hyperspectral and RGB-based tracking methods. From the perspective of network architecture, these methods can be broadly categorized into three groups. The first group comprises hand-crafted feature-based trackers, including CNHT [] and MHT [], which rely on manually designed features for object representation. The second group consists of CNN-based methods, including DeepHKCF [], SiamRPN++ [], TSCFW [], and PHTrack []. Specifically, DeepHKCF leverages deep features while compressing spectral bands, SiamRPN++ uses a Siamese region proposal network to correlate template and search features, TSCFW utilizes tensor representations with context modeling to capture spectral–spatial dependencies, and PHTrack leverages prompt learning to adapt foundation models for hyperspectral tracking. The third group consists of transformer-based frameworks, including HOTMoE [], DASSP-Net [], and HyperTrack []. Specifically, HOTMoE models cross-spectral interactions through a modality mixture of experts, DASSP-Net employs dual attention to jointly learn spatial–spectral dependencies, and HyperTrack incorporates spectral reconstruction and a multi-branch framework to strengthen hyperspectral video representation.
Table 1 reports the tracking results of BCR-Net and the compared approaches. As shown in Table 1, correlation filter-based trackers demonstrate limited capability in modeling the complex spatial–spectral representation of hyperspectral data, leading to poor adaptability in cluttered or dynamic scenes. CNN-based methods achieve improved accuracy by learning hierarchical feature representations, thereby enhancing spatial discrimination and robustness. Specifically, PHTrack attains the highest AUC value of 0.586 owing to its foundation model adaptation via prompt learning, which strengthens overlap accuracy. Transformer-based frameworks further enhance the representation of spectral–spatial dependencies through attention mechanisms. In particular, HyperTrack integrates band embedding and multi-branch attention modules to capture both inter-band and spatial correlations, achieving an AUC of 0.845 and a DP of 0.563. Although the band embedding and multi-branch attention improve overall spectral–spatial feature representation compared with most CNN-based trackers, the limited interaction across branches and the additional structural complexity reduce localization stability. Compared with ViT-based frameworks, BCR-Net attains a higher DP score of 0.862, demonstrating that its explicit spectral–spatial interaction combined with adaptive channel refinement provides more stable and accurate target tracking under challenging conditions.
Table 1.
Comparison of tracking results across various algorithms on the HOTC dataset.
To evaluate the Inference speed of the compared methods and the proposed BCR-Net, we report FPS, FLOPs, and model size on the HOTC dataset (Table 1). Among this, “N/A” indicates that the method does not involve a complete end-to-end deep learning framework. STARK-ST50 and SiamRPN++ achieve high FPS by directly operating on false-color images, thereby avoiding the computational overhead. PHTrack and HotMoE exhibit moderate FPS, benefiting from prompt learning and selective expert activation, respectively, which reduce unnecessary spectral computations. DaSSP-Net and HyperTrack show slightly lower FPS due to the inclusion of spatial–spectral attention and multi-branch ViT modules. BCR-Net processes frames at 8.27 FPS, reflecting the additional computations introduced by the Dual Enhancement Module and adaptive band grouping, which contribute to improved tracking accuracy. Regarding computational complexity, PHTrack requires moderate FLOPs and model size since only its lightweight prompt branch is trained, whereas HotMoE’s multiple expert modules increase both FLOPs and memory consumption. DaSSP-Net and HyperTrack demand larger FLOPs and parameters owing to the ViT backbone and attention-based fusion. In conlusion, BCR-Net incurs slightly higher computational cost but achieves a deliberate balance between tracking precision and efficiency.
From Figure 3, our method achieves the highest DP score of 0.862, while PHTrack attains the highest AUC of 0.586. This indicates that our approach excels in localization stability, whereas PHTrack is more effective in overlap accuracy.
Figure 3.
Evaluation plots of competing hyperspectral trackers on the HOTC dataset. (a) Precision plots of OPE. (b) Success plots of OPE.
To evaluate model robustness, we selected four representative sequences: (fast motion), (background clutter), (target rotation and scale variation), and (occlusion and low resolution). As shown in Figure 4, our approach accurately tracks targets under these conditions, substantially outperforming other methods.
Figure 4.
Tracking results on four selected video sequences (coke, kangaroo, card, and playground), where bounding boxes of different colors represent the outputs of various trackers.
4.2.2. Experiments with the MUST Dataset
To further evaluate the practical performance of our proposed tracker in UAV video scenarios, we conducted systematic comparative experiments on the MUST dataset. A set of representative RGB object tracking algorithms was included for comparison, including SiamRPN [], SiamMask [], TaMOs [], ToMP [], SGDViT [], and OSTrack256 []. All methods were run using their officially published models and parameter settings on a unified experimental platform to ensure fairness and to minimize the influence of implementation or environmental differences.
Performance evaluation on the MUST dataset is presented in Table 2, where trackers are assessed using AUC and DP metrics. Our approach delivers superior results with AUC of 0.595 and DP of 0.388, highlighting its effectiveness and robustness in UAV aerial object tracking scenarios. Among the compared methods, TaMOs and ToMP perform relatively poorly, with AUCs of 0.316 and 0.461 and DPs of 0.177 and 0.318, respectively. Their limited modeling mechanisms reduce the ability to fully exploit spectral information, resulting in lower discrimination and tracking accuracy under complex conditions. In contrast, SiamRPN, SiamMask, SGDViT, and OSTrack256 maintain relatively stable performance, achieving AUCs around 0.539–0.552 and DPs around 0.369–0.385. These RGB-based deep models benefit from effective spatial and temporal feature extraction, yet they still lack comprehensive spectral modeling, which constrains their precision in scenarios with complex spectral variations. Our proposed method effectively leverages full spectral information through channel refinement and spectral–spatial correlation modules. By explicitly modeling inter-band relationships and enhancing deep feature representation, our tracker achieves superior accuracy and stability, surpassing both traditional and RGB-based deep models, and demonstrating enhanced cross-modal adaptability for UAV tracking.
Table 2.
Comparison of tracking results across various algorithms on the MUST dataset.
To evaluate model robustness, we selected four representative sequences: (fast motion), (scale variation), (partial occlusion), and (Similar Object). As shown in Figure 5, our approach accurately tracks targets under these conditions, substantially outperforming other methods.
Figure 5.
Tracking results on four selected video sequences (FM-Bicycle-001, SV-Bus-001, POC-Person-006, and SOB-Car-004), where bounding boxes of different colors represent the outputs of various trackers.
In summary, while existing RGB tracking algorithms demonstrate strong performance under standard visible light conditions, their performance generally degrades to varying degrees when faced with UAV videos, which present greater spectral complexity and perceptual challenges. By contrast, our method leverages spectral information and structural optimization, achieving a favorable balance between tracking accuracy and stability. This demonstrates its strong cross-modal adaptability and practical application potential in aerial tracking scenarios.
4.3. Ablation Study
The following presents a systematic ablation study on the HOTC dataset to evaluate the impact of each module and design choice on tracking performance.
4.3.1. Comparison of Tracking Performance with Proposed Modules
In the following, a systematic ablation evaluation is performed to quantify the effect of each key component. Specifically, a baseline tracker is first established, in which hyperspectral images are converted into false-color representations before being fed into the model for prediction. Building on this baseline, the Band Importance Modeling Module (BIMM) and the Contextual Feature Refinement Module (CFRM) are subsequently incorporated, and their effects are assessed on the HOTC dataset.
Table 3 illustrates the AUC and DP achieved by different module configurations. When BIMM is incorporated into the baseline, the AUC increases from 0.534 to 0.549 and the DP rises from 0.831 to 0.852, indicating that this module improves the selection of key spectral channels. Adding CFRM further increases the AUC to 0.572 and the DP to 0.862, suggesting that contextual information modeling contributes positively to hyperspectral target representation. Taken together, the inclusion of these two modules enhances both tracking accuracy and robustness while maintaining a simple model structure.
Table 3.
Comparison of tracking performance under different module combinations.
Secondly, we investigated how different initialization strategies affect the performance of the BIMM module. Specifically, we initialized the weight matrix to a mean matrix with all elements equal to , and compared this with a learned attention-based initialization. As shown in Table 4, using the learned attention initialization yields tracking results of 0.572 and 0.862 in terms of AUC and distance precision (DP), respectively, outperforming the mean initialization results of 0.544 and 0.837. These findings indicate that learning-based initialization provides a more effective starting point for band importance modeling, facilitating better feature selection and improving overall tracking accuracy.
Table 4.
Comparison of tracking performance under different BIMM initialization strategies.
4.3.2. Comparison of BIMM Component Configurations
To evaluate the effectiveness of each component in BIMM, we conduct ablation experiments by selectively removing one module at a time while keeping the others unchanged. The complete BIMM serves as the baseline. Table 5 reports the performance under different configurations.
Table 5.
Comparison of tracking performance under different BIMM component configurations.
Table 5 reports the performance of BIMM under different component configurations. The results reveal that each component—SEM, CEM, SDUB, and SGRM—contributes distinctly to the overall tracking capability. Removing any single component consistently degrades both AUC and DP, demonstrating the necessity of each module. SEM enhances semantic feature representation, improving discrimination and robustness. CEM provides crucial contextual information that strengthens feature correlation. SDUB facilitates effective spatial-depth feature integration, contributing to improved spatial representation. SGRM captures global spectral relationships, which is vital for handling hyperspectral data. Configurations combining multiple components achieve higher performance, confirming their complementary nature. The full BIMM configuration delivers the highest performance, validating that the integrated design of these modules effectively synergizes semantic, contextual, spatial, and spectral information to maximize tracking accuracy and robustness.
4.3.3. Comparison of Ensemble Prediction Methods
To evaluate the ensemble strategy’s effectiveness in hyperspectral tracking, we compared different prediction fusion methods and weighting mechanisms, analyzing their performance impact from two perspectives.
First, to investigate the differences between fusion mechanisms, the weighted ensemble was replaced by a simple average ensemble. As shown in Table 6, while average ensemble improves the ability to fuse multi-channel information to some extent, it slightly underperforms our proposed weighted ensemble scheme (AUC 0.572, DP 0.862) in both AUC (0.551) and DP (0.859). These findings suggest that weighting the influence of each false-color image during prediction improves feature integration and enhances overall tracking performance.
Table 6.
Comparison of tracking performance under different ensemble prediction methods.
Further, to investigate the effect of different weight generation strategies within the Band Importance Modeling Module, we compare our Integrated Prediction Module, which dynamically adjusts fusion weights, with a baseline that uses learnable fixed weights. In the baseline, all weights are learned during training but remain constant during testing, while all other settings are kept identical. As shown in Table 7, the fixed-weight strategy produces an AUC of 0.549 and a DP of 0.846, corresponding to decreases of 0.023 and 0.016 relative to the Integrated Prediction Module. This performance drop indicates that static weights cannot adapt to distribution shifts caused by changes in the target’s appearance or spectral properties across frames, thereby limiting effective feature fusion. In contrast, the Integrated Prediction Module leverages the BIMM to dynamically adjust fusion weights based on the input content of each frame, improving both generalization and robustness.
Table 7.
Comparison of tracking performance under different weighted settings.
4.3.4. Comparison of Band Grouping and Branch Number Strategies
In order to evaluate the band grouping design, four baseline strategies were considered in addition to our approach: sequential partitioning, which groups bands into equal intervals; single-band replication, which replicates each band three times to form a false-color image; PCA-based grouping, which projects hyperspectral bands into three principal components; and random grouping, which assigns bands arbitrarily to the RGB channels. All methods use the same tracking framework to ensure fairness.
As shown in Table 8, our strategy reaches 0.572 in AUC and 0.862 in DP, outperforming all other grouping approaches. Sequential partitioning fails to fully exploit complementary spectral relationships, while single-band replication introduces redundancy that limits feature discriminability. PCA-based grouping captures global variance but may overlook task-specific band importance, and random grouping arbitrarily assigns bands, disrupting natural spectral correlations. In contrast, our method selects bands based on learned importance, forming more informative false-color images that enhance feature representation and improve overall tracking performance.
Table 8.
Comparison of tracking performance under different band grouping strategies.
4.4. Parameter Analysis
To further investigate the impact of key hyper-parameters in the experiments, we conduct experiments on two factors: the depth of the MLP layers and the number of false-color images.
Firstly, as shown in Table 9, varying the MLP depth produces notable differences in tracking performance. The single-layer configuration achieves the highest performance. Increasing the depth to three or five layers leads to performance degradation, with AUC values of 0.528 and 0.549, respectively. This suggests that deeper MLP structures may lead to overfitting and reduced generalization ability.
Table 9.
Comparison of tracking performance under different MLP layer configurations.
Secondly, the results in Table 10 clearly demonstrate the effect of the number of false-color images on tracking performance. In our experiments, the input consists of 16-band hyperspectral images, and each false-color image is formed by grouping three bands, which allows a maximum of five images. The configuration with five images achieves the highest performance, with an AUC of 0.572 and a DP of 0.862. Reducing the number of images to three or one results in notable performance degradation, with AUC values of 0.541 and 0.505, respectively, accompanied by corresponding decreases in DP. This indicates that increasing the number of false-color images provides more diverse feature information, thereby enhancing tracking robustness.
Table 10.
Comparison of tracking performance under different numbers of false-color images.
5. Discussion
While the comprehensive experiments and analyses have validated the effectiveness of the proposed BCR-Net in hyperspectral video tracking, several important challenges remain that open avenues for future research.
- The shortage of high-quality hyperspectral video datasets with precise annotations presents a challenge for robust feature learning. Future research could focus on weakly supervised or semi-supervised learning, leveraging partial or noisy annotations. Additionally, knowledge transfer from well-labeled datasets in related domains could be explored to bootstrap hyperspectral trackers with minimal hyperspectral-specific annotations.
- Although the current fusion design improves feature representation, the interaction between spectral and spatial information remains insufficiently explored. In our earlier attempts, spectral attention mechanisms were employed to capture inter-band dependencies, but the resulting improvements were marginal. By contrast, the U-Net-based design adopted in this work yielded more substantial performance gains, highlighting the importance of structural modeling. Looking ahead, we plan to investigate adaptive fusion strategies that can adaptively integrate spectral and spatial features under varying scene conditions, with the goal of further improving both accuracy and computational efficiency in hyperspectral tracking.
- Our experiments demonstrate that explicit modeling of spectral information substantially improves target discrimination in challenging scenarios. Building upon this insight, future work will explore end-to-end network architectures that directly process raw multi-dimensional hyperspectral inputs, enabling more comprehensive utilization of spectral cues and potentially improving robustness in complex tracking environments.
6. Conclusions
A tracking algorithm for hyperspectral videos is proposed in this work, which benefits from spectral–spatial feature fusion and an improved network design. Based on the above experiments and analyses, the main conclusions are as follows:
- The band selection module ranks spectral bands by importance and groups them into complementary composites, improving feature discrimination and computational efficiency without compromising accuracy.
- By adaptively fusing results based on feature importance, our method achieves superior performance compared to uniform averaging, demonstrating more effective utilization of hyperspectral information for target discrimination.
- The proposed spectral–spatial fusion design, based on a U-Net architecture, demonstrates notable gains compared to earlier attempts with spectral attention mechanisms, validating its effectiveness in hyperspectral tracking.
- The method achieves a good balance between accuracy and efficiency, showing competitive performance on the HOTC dataset (with a DP of 0.862 and an AUC of 0.572), demonstrating its potential for real-world applications.
- Future work will further explore adaptive fusion strategies that dynamically balance spectral and spatial information to further improve robustness and generalization in challenging tracking scenarios.
Author Contributions
Conceptualization, J.Z. and Z.Z.; Methodology, Z.Z., K.N. and N.H.; Software, J.Z.; Validation, Z.Z., K.N. and J.Z.; Formal analysis, J.Z.; Investigation, J.Z.; Resources, Z.Z., K.N., N.H., Q.L. and P.L.; Data curation, J.Z.; Writing—original draft preparation, J.Z.; Writing—review and editing, Z.Z., K.N., N.H., Q.L. and P.L.; Visualization, J.Z.; Supervision, Z.Z., K.N. and N.H.; Project administration, Z.Z.; Funding acquisition, Z.Z. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded in part by the Frontier Technologies R&D Program of Jiangsu under Grant BF2024070, in part by the Nanjing University of Posts and Telecommunications Science Foundation (NUPTSF) under Grant NY222107, in part by the Natural Science Foundation of Jiangsu Higher Education Institutions of China under Grant 23KJB510022, in part by the Jiangsu Geological Bureau Research Project under Grant 2023KY11, and in part by the Natural Science Research Start-up Foundation of Recruiting Talents of Nanjing University of Posts and Telecommunications under Grant NY222017. The Article Processing Charge (APC) was funded by the authors.
Data Availability Statement
The hyperspectral video dataset used in this study is available in publicly accessible repositories. No new data were created in this study.
Conflicts of Interest
The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.
Abbreviations
The following abbreviations are used in this manuscript:
| AUC | Area Under Curve |
| DP | Distance Precision |
| OPE | One Pass Evaluation |
| CNN | Convolutional Neural Network |
| HOT | Hyperspectral Object Tracking |
| CF | Correlation Filter |
| DL | Deep Learning |
| BCR-Net | Band and Context Refinement Network |
References
- Joshi, K.A.; Thakore, D.G. A survey on moving object detection and tracking in video surveillance system. Int. J. Soft Comput. Eng. 2012, 2, 44–48. [Google Scholar]
- Liu, N.; Shapira, P.; Yue, X. Tracking developments in artificial intelligence research: Constructing and applying a new search strategy. Scientometrics 2021, 126, 3153–3192. [Google Scholar] [CrossRef]
- Li, X.; Hu, W.; Shen, C.; Zhang, Z.; Dick, A.; Van Den Hengel, A. A survey of appearance models in visual object tracking. ACM Trans. Intell. Syst. Technol. 2013, 4, 1–48. [Google Scholar] [CrossRef]
- Xiao, Y.; Su, X.; Yuan, Q.; Liu, D.; Shen, H.; Zhang, L. Satellite video super-resolution via multiscale deformable convolution alignment and temporal grouping projection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5610819. [Google Scholar] [CrossRef]
- Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. Siamrpn++: Evolution of siamese visual tracking with very deep networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; IEEE: New York, NY, USA; pp. 4282–4291. [Google Scholar]
- Xiao, Y.; Yuan, Q.; Jiang, K.; He, J.; Wang, Y.; Zhang, L. From degrade to upgrade: Learning a self-supervised degradation guided adaptive network for blind remote sensing image super-resolution. Inf. Fusion 2023, 96, 297–311. [Google Scholar] [CrossRef]
- Li, Z.; Xiong, F.; Zhou, J.; Lu, J.; Qian, Y. Learning a deep ensemble network with band importance for hyperspectral object tracking. IEEE Trans. Image Process. 2023, 32, 2901–2914. [Google Scholar] [CrossRef]
- Zhang, Z.; Qian, K.; Du, J.; Zhou, H. Multi-features integration based hyperspectral videos tracker. In Proceedings of the 2021 11th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 March 2021; IEEE: New York, NY, USA; pp. 1–5. [Google Scholar]
- Tang, Y.; Liu, Y.; Huang, H. Target-aware and spatial-spectral discriminant feature joint correlation filters for hyperspectral video object tracking. Comput. Vis. Image Underst. 2022, 223, 103535. [Google Scholar] [CrossRef]
- Hou, Z.; Li, W.; Zhou, J.; Tao, R. Spatial–spectral weighted and regularized tensor sparse correlation filter for object tracking in hyperspectral videos. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5541012. [Google Scholar] [CrossRef]
- Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. Atom: Accurate tracking by overlap maximization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; IEEE: New York, NY, USA; pp. 4660–4669. [Google Scholar]
- Xiao, Y.; Yuan, Q.; Jiang, K.; Jin, X.; He, J.; Zhang, L.; Lin, C.W. Local-global temporal difference learning for satellite video super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 2789–2802. [Google Scholar] [CrossRef]
- Huang, L.; Zhao, X.; Huang, K. Got-10k: A large high-diversity benchmark for generic object tracking in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1562–1577. [Google Scholar] [CrossRef] [PubMed]
- Muller, M.; Bibi, A.; Giancola, S.; Alsubaihi, S.; Ghanem, B. Trackingnet: A large-scale dataset and benchmark for object tracking in the wild. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 300–317. [Google Scholar]
- Fan, H.; Lin, L.; Yang, F.; Chu, P.; Deng, G.; Yu, S.; Bai, H.; Xu, Y.; Liao, C.; Ling, H. Lasot: A high-quality benchmark for large-scale single object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5374–5383. [Google Scholar]
- Liu, Z.; Wang, X.; Zhong, Y.; Shu, M.; Sun, C. SiamHYPER: Learning a hyperspectral object tracker from an RGB-based tracker. IEEE Trans. Image Process. 2022, 31, 7116–7129. [Google Scholar] [CrossRef] [PubMed]
- Tang, Y.; Liu, Y.; Ji, L.; Huang, H. Robust hyperspectral object tracking by exploiting background-aware spectral information with band selection network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6013405. [Google Scholar] [CrossRef]
- Wei, X.; Zhu, W.; Liao, B.; Cai, L. Scalable one-pass self-representation learning for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4360–4374. [Google Scholar] [CrossRef]
- Li, Z.; Xiong, F.; Zhou, J.; Wang, J.; Lu, J.; Qian, Y. BAE-Net: A band attention aware ensemble network for hyperspectral object tracking. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; IEEE: New York, NY, USA; pp. 2106–2110. [Google Scholar]
- Li, Z.; Ye, X.; Xiong, F.; Lu, J.; Zhou, J.; Qian, Y. Spectral-spatial-temporal attention network for hyperspectral tracking. In Proceedings of the 2021 11th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 March 2021; IEEE: New York, NY, USA; pp. 1–5. [Google Scholar]
- Ouyang, E.; Wu, J.; Li, B.; Zhao, L.; Hu, W. Band regrouping and response-level fusion for end-to-end hyperspectral object tracking. IEEE Geosci. Remote Sens. Lett. 2021, 19, 6005805. [Google Scholar] [CrossRef]
- Qian, K.; Zhou, J.; Xiong, F.; Zhou, H.; Du, J. Object tracking in hyperspectral videos with convolutional features and kernelized correlation filter. In Proceedings of the International Conference on Smart Multimedia, Toulon, France, 24–26 August 2018; Springer International Publishing: Cham, Switzerland; pp. 308–319. [Google Scholar]
- Guo, H.; Xu, Y.; Wu, Z.; Wei, Z. Hyperspectral video tracker based on anomaly suppression and multi-feature integration. In Proceedings of the 2023 13th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Athens, Greece, 31 October–2 November 2023; IEEE: New York, NY, USA; pp. 1–5. [Google Scholar]
- Chen, L.; Zhao, Y.; Yao, J.; Chen, J.; Li, N.; Chan, J.C.W.; Kong, S.G. Object tracking in hyperspectral-oriented video with fast spatial-spectral features. Remote Sens. 2021, 13, 1922. [Google Scholar] [CrossRef]
- Lou, H.; Rong, S.; Guo, H.; Ma, J. Hyperspectral video target tracking based on TensorSSA and deep features. In Proceedings of the 2023 IEEE 6th International Conference on Electronic Information and Communication Technology (ICEICT), Qingdao, China, 21–24 July 2023; IEEE: New York, NY, USA; pp. 1023–1028. [Google Scholar]
- Uzkent, B.; Rangnekar, A.; Hoffman, M.J. Tracking in aerial hyperspectral videos using deep kernelized correlation filters. IEEE Trans. Geosci. Remote Sens. 2018, 57, 449–461. [Google Scholar] [CrossRef]
- Wang, S.; Qian, K.; Chen, P. BS-SiamRPN: Hyperspectral video tracking based on band selection and the Siamese region proposal network. In Proceedings of the 2022 12th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Rome, Italy, 13–16 September 2022; IEEE: New York, NY, USA; pp. 1–8. [Google Scholar]
- Zhao, C.; Liu, H.; Su, N.; Yan, Y. TFTN: A transformer-based fusion tracking framework of hyperspectral and RGB. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5542515. [Google Scholar] [CrossRef]
- Chen, Y.; Yuan, Q.; Tang, Y.; Xiao, Y.; He, J.; Zhang, L. SPIRIT: Spectral awareness interaction network with dynamic template for hyperspectral object tracking. IEEE Trans. Geosci. Remote Sens. 2023, 62, 5503116. [Google Scholar] [CrossRef]
- Jiang, X.; Wang, X.; Sun, C.; Zhu, Z.; Zhong, Y. A channel adaptive dual siamese network for hyperspectral object tracking. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5403912. [Google Scholar] [CrossRef]
- Xiong, F.; Sun, Y.; Zhou, J.; Lu, J.; Qian, Y. Spatial-Spectral-Temporal Correlation Filter for Hyperspectral Object Tracking. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5506613. [Google Scholar] [CrossRef]
- He, G.; Gao, L.; Chen, L.; Jiang, Y.; Xie, W.; Li, Y. Hyperspectral Object Tracking with Spectral Information Prompt. IEEE Trans. Circuits Syst. Video Technol. 2025. [Google Scholar] [CrossRef]
- Wang, H.; Li, W.; Xia, X.G.; Du, Q.; Tian, J. SSF-Net: Spatial-spectral fusion network with spectral angle awareness for hyperspectral object tracking. IEEE Trans. Image Process. 2025, 34, 3518–3532. [Google Scholar] [CrossRef] [PubMed]
- Chen, Z.; Zhong, B.; Li, G.; Zhang, S.; Ji, R. Siamese box adaptive network for visual tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; IEEE: Piscataway, NJ, USA; pp. 6668–6677. [Google Scholar]
- Li, Z.; Xiong, F.; Lu, J.; Zhou, J.; Qian, Y. Material-guided siamese fusion network for hyperspectral object tracking. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; IEEE: New York, NY, USA; pp. 2809–2813. [Google Scholar]
- Qin, H.; Xu, T.; Li, T.; Chen, Z.; Feng, T.; Li, J. MUST: The First Dataset and Unified Framework for Multispectral UAV Single Object Tracking. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 10–17 June 2025; pp. 16882–16891. [Google Scholar]
- Pu, S.; Song, Y.; Ma, C.; Zhang, H.; Yang, M.H. Learning recurrent memory activation networks for visual tracking. IEEE Trans. Image Process. 2020, 30, 725–738. [Google Scholar] [CrossRef] [PubMed]
- Chen, Y.; Tang, Y.; Su, X.; Li, J.; Xiao, Y.; He, J.; Yuan, Q. PhTrack: Prompting for hyperspectral video tracking. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5533918. [Google Scholar] [CrossRef]
- Sun, W.; Tan, Y.; Li, J.; Hou, S.; Li, X.; Shao, Y.; Song, B. HotMoE: Exploring sparse mixture-of-experts for hyperspectral object tracking. IEEE Trans. Multimed. 2025, 27, 4072–4083. [Google Scholar] [CrossRef]
- Li, Z.; Xiong, F.; Lu, J.; Wang, J.; Chen, D.; Zhou, J.; Qian, Y. Multi-domain universal representation learning for hyperspectral object tracking. Pattern Recognit. 2025, 162, 111389. [Google Scholar] [CrossRef]
- Tan, Y.; Sun, W.; Li, J.; Hou, S.; Li, X.; Wang, Z.; Song, B. HyperTrack: A unified network for hyperspectral video object tracking. IEEE Trans. Circuits Syst. Video Technol. 2025. [Google Scholar] [CrossRef]
- Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8971–8980. [Google Scholar]
- Wang, Q.; Zhang, L.; Bertinetto, L.; Hu, W.; Torr, P.H. Fast online object tracking and segmentation: A unifying approach. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1328–1338. [Google Scholar]
- Mayer, C.; Danelljan, M.; Yang, M.H.; Ferrari, V.; Van Gool, L.; Kuznetsova, A. Beyond SOT: Tracking multiple generic objects at once. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 4–8 January 2024; pp. 6826–6836. [Google Scholar]
- Mayer, C.; Danelljan, M.; Bhat, G.; Paul, M.; Paudel, D.P.; Yu, F.; Van Gool, L. Transforming model prediction for tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 8731–8740. [Google Scholar]
- Yao, L.; Fu, C.; Li, S.; Zheng, G.; Ye, J. SGDViT: Saliency-guided dynamic vision transformer for UAV tracking. arXiv 2023, arXiv:2303.04378. [Google Scholar]
- Ye, B.; Chang, H.; Ma, B.; Shan, S.; Chen, X. Joint feature learning and relation modeling for tracking: A one-stream framework. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer Nature: Cham, Switzerland; pp. 341–357. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).