Next Article in Journal
Evaluation of Atmospheric Preprocessing Methods and Chlorophyll Algorithms for Sentinel-2 Imagery in Coastal Waters
Previous Article in Journal
Assessment and Validation of FAPAR, a Satellite-Based Plant Health and Water Stress Indicator, over Uganda
Previous Article in Special Issue
Mask-Guided Teacher–Student Learning for Open-Vocabulary Object Detection in Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Low-Altitude Multi-Object Tracking via Graph Neural Networks with Cross-Attention and Reliable Neighbor Guidance

College of Intelligence Science and Technology, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(20), 3502; https://doi.org/10.3390/rs17203502
Submission received: 31 August 2025 / Revised: 14 October 2025 / Accepted: 17 October 2025 / Published: 21 October 2025

Highlights

What are the main findings?
  • A novel multi-object tracking framework, NOWA-MOT, is proposed that leverages stable group context and graph neural networks to resolve tracking ambiguities caused by occlusion and non-linear motion in UAV imagery.
  • The framework introduces a cascaded association mechanism using cross-graph attention for robust feature enhancement and reliably matched neighbors as anchors to guide the matching of more difficult, ambiguous targets.
What is the implication of the main findings?
  • The proposed approach achieves state-of-the-art performance on challenging UAV datasets (VisDrone, UAVDT), significantly reducing identity switches and improving tracking continuity in dense, dynamic scenes.
  • This context-aware tracking paradigm provides a more reliable foundation for downstream remote sensing applications, such as traffic flow analysis and smart city monitoring, by delivering higher-fidelity trajectory data.

Abstract

In low-altitude multi-object tracking (MOT), challenges such as frequent inter-object occlusion and complex non-linear motion disrupt the appearance of individual targets and the continuity of their trajectories, leading to frequent tracking failures. We posit that the relatively stable spatio-temporal relationships within object groups (e.g., pedestrians and vehicles) offer powerful contextual cues to resolve such ambiguities. We present NOWA-MOT (Neighbors Know Who We Are), a novel tracking-by-detection framework designed to systematically exploit this principle through a multi-stage association process. We make three primary contributions. First, we introduce a Low-Confidence Occlusion Recovery (LOR) module that dynamically adjusts detection scores by integrating IoU, a novel Recovery IoU (RIoU) metric, and location similarity to surrounding objects, enabling occluded targets to participate in high-priority matching. Second, for initial data association, we propose a Graph Cross-Attention (GCA) mechanism. In this module, separate graphs are constructed for detections and trajectories, and a cross-attention architecture is employed to propagate rich contextual information between them, yielding highly discriminative feature representations for robust matching. Third, to resolve the remaining ambiguities, we design a cascaded Matched Neighbor Guidance (MNG) module, which uniquely leverages the reliably matched pairs from the first stage as contextual anchors. Through MNG, star-shaped topological features are built for unmatched objects relative to their stable neighbors, enabling accurate association even when intrinsic features are weak. Our comprehensive experimental evaluation on the VisDrone2019 and UAVDT datasets confirms the superiority of our approach, achieving state-of-the-art HOTA scores of 51.34% and 62.69%, respectively, and drastically reducing identity switches compared to previous methods.

1. Introduction

Visual multi-object tracking (MOT) is a fundamental computer vision task with the aim of detecting objects in video sequences and assigning each target a consistent, unique identity. Driven by advancements in deep learning, MOT techniques have evolved significantly over the past decade. Compared to single-object tracking (SOT), MOT presents additional challenges, including the disappearance of existing objects, the emergence of new objects, dynamic variation in target count, and unreliable cues during trajectory association. Currently, most research [1,2,3,4,5] focuses on tracking objects in settings with stationary or horizontally moving cameras—such as handheld or vehicle-mounted cameras—which operate with limited perception ranges.
In recent years, unmanned aerial vehicles (UAVs) have been adopted across various domains, including search and rescue, smart cities, intelligent transportation, and remote sensing. Compared to stationary or horizontally moving cameras, UAV-mounted cameras offer higher perspectives and broader fields of view. However, they also introduce more complex challenges, such as small target sizes and non-linear camera motion. Consequently, there is a growing demand for innovative techniques capable of addressing the unique complexities in UAV scenarios.
As shown in Figure 1, The following challenges are primarily encountered in UAV-based MOT:
(1) Image blurring induced by high-speed UAV motion.
(2) Complex motion patterns resulting from UAV maneuvers (e.g., hovering, ascending, forward movement).
(3) Frequent occlusions caused by elevated viewing angles, where trees, buildings, and objects that lie in the same imaging plane.
(4) Indistinct texture information for small objects at high altitudes, increasing susceptibility to confusion with similar objects.
Previous MOT research [6,7,8] predominantly features the tracking-by-detection (TBD) paradigm, which involves two steps: detection and association. Algorithms typically employ detection models to identify objects of interest, followed by re-identification (Re-ID) modules to extract appearance features. Motion characteristics are estimated using Kalman filtering, and both features are combined for association. Through this method, target feature extraction and data association are primarily optimized. Considering the characteristics of UAV imagery, relying solely on features from individual objects often leads to inaccurate associations. The reliability of detected features diminishes, making it difficult to maintain visual consistency for occluded objects. Object detection bounding boxes can overlap, particularly when they are proximate, causing the extracted appearance features to be confused across objects. Furthermore, sudden non-linear motion (e.g., abrupt platform or target direction changes) significantly alters motion features, even for objects with a high detection confidence, rendering both the motion and appearance cues ineffective for matching.
Occlusion is a major factor contributing to MOT failure [9]. It reduces the visible area of a target, hindering the extraction of sufficiently discriminative features via Re-ID for appearance-based matching methods. For motion-based matching methods, deformation of the detection bounding box due to occlusion causes the Intersection over Union (IoU) between predicted and detected boxes to fall below the matching threshold. Additionally, occlusion lowers the detection confidence. Many cascade matching methods [6,8] prioritize high-confidence detections before processing low-confidence ones, but this approach prevents occluded objects from matching their true trajectories in the first matching stage due to their low confidence. Occlusion frequently occurs for dense objects that are spatially close and visually similar, increasing the likelihood of erroneous trajectory associations.
The core contribution of this study stems from a fundamental observation of how complex scenes are often structured [10,11,12]. Target occlusion does not happen in a vacuum; rather, it often occurs when a smaller, individual target is obscured by surrounding, more distinct objects or groups that exhibit consistent, coordinated motion patterns. For instance, a pedestrian might be temporarily hidden by a group of people moving in unison, or a car might be occluded by a convoy of trucks. In these scenarios, while the occluded target’s individual appearance and motion features become unreliable, the spatio-temporal topology of the surrounding group remains relatively stable. Our method is designed to systematically exploit this observation. We posit that by modeling and leveraging these stable group dynamics, we can infer the identity and trajectory of an occluded object with far greater accuracy than by relying on its degraded individual features alone. This principle of using contextual group information to resolve individual ambiguity forms the foundational basis of our proposed framework.
Motivated by these insights, this paper introduces a confidence adjustment and cross-attention feature enhancement modules. The former integrates multiple cues to enhance the confidence scores of low-confidence objects, partially recovering occluded and blurred objects. The latter concatenates appearance features with spatial neighbor features and employs cross-attention mechanisms alongside graph neural networks to obtain features that better discriminate between trajectories and detections.
Another key innovation of our approach lies in utilizing reliable matched pairs from the first round as neighbors in the second matching stage. By modeling the topological relationships between unmatched detections/trajectories and these neighbors, more robust features are obtained. Specifically, we design velocity and directional constraints to identify reliable neighbors around unmatched trajectories and detections, enabling object association via a star-shaped topological distance metric. This dual mechanism ensures effective secondary matching even under severe occlusion.
The contributions of this paper can be summarized as follows:
(1) A novel target confidence adjustment method, termed Low-Confidence Occlusion Recovery (LOR), is introduced. This technique incorporates mutual occlusion relationships, positional data, and detection confidence to re-evaluate initially low-confidence detections, improving their reliability. Consequently, the anticipated errors in multi-round associations are effectively mitigated.
(2) A Graph Cross-Attention Enhancement Mechanism (GCA) is proposed to refine object Re-ID features. Complementarily, an innovative topological similarity computation function is developed to facilitate the First-Round target association.
(3) A cascaded matching mechanism named Matched Neighbor Guidance (MNG) is designed, leveraging initial matches and neighboring trajectories to associate detections via directed graph feature propagation.
(4) The integrated approach enables NOWA-MOT to achieve state-of-the-art performance on the VisDrone2019 and UAVDT benchmarks.
The remainder of this paper is structured as follows: Section 2 reviews prior work on detection-based tracking, graph neural networks, and occlusion-aware methods. Section 3 details the proposed methodology, and Section 4 presents the experimental configuration and the results of comparative analyses with state-of-art trackers. Finally, Section 5 discusses the framework’s attributes, and future research directions are suggested.

2. Related Work

2.1. Tracking by Detection

Modern multi-object tracking systems predominantly employ a tracking-by-detection (TBD) framework comprising three sequential stages: object detection, feature extraction, and data association. The detection module provides per-frame bounding box coordinates, object categories, and confidence scores and appearance features are extracted using re-identification (Re-ID) models. Some methods [13,14] feature improved re-identification models to obtain more discriminative features, where motion characteristics are quantified via techniques such as Kalman filtering. OC-SORT [15] utilizes target observations to compute virtual trajectories during occlusion periods, thereby mitigating the accumulation of Kalman filter parameter errors caused by occlusion. The association module establishes cross-frame correspondences by assigning unique identifiers based on trajectory detection similarity metrics. Common metrics include IoU [16], appearance cosine similarity [17], and novel feature modalities incorporating topological structures [18] or hybrid cues [19].
In UAV-based tracking, motion compensation modules are commonly integrated to mitigate platform-induced motion errors and camera jitter. Target occlusion constitutes a critical challenge for detection-based MOT systems, where partial or complete occlusion causes reduced detection reliability. To enhance robustness, specialized methods have been developed: Chen et al. [20] proposed Arbitrack, which employs rotated object detectors to replace traditional axis-aligned bounding boxes, capturing target orientation for improved motion estimation and feature matching. Deng et al. [21] leveraged topological features to augment target representations in aerial imagery, addressing motion feature degradation from non-linear trajectories while predicting occlusion probabilities via temporal–topological consistency, thus enabling dynamic recovery of occluded objects. Song et al. [22] proposed SFTrack, which proactively utilizes low-confidence detections as tracking initiators and exhibits an enhanced tracking accuracy through the reintroduction of traditional appearance matching algorithms for data association.
Our Matched Neighbor Guidance (MNG) module provides a theoretically stronger alternative to traditional cascaded matching paradigms. Standard cascaded approaches typically process all remaining low-confidence detections in a second stage, attempting to associate them based on weaker cues like IoU. The fundamental limitation of this is that this matching occurs in the context of high uncertainty. MNG introduces a novel principle: guidance from the certain to the uncertain. It uniquely leverages the high-confidence, reliably matched pairs from the first stage as stable contextual anchors. By constructing star-shaped topological features for unmatched objects relative to this shared, reliable context, MNG establishes a robust frame of reference to resolve ambiguity. This anchor-based guidance provides a stronger associative signal than simply comparing two uncertain elements (an unmatched track and a low-confidence detection) in isolation.

2.2. Tracking with Graph Neural Networks

Graphs are widely used across disciplines to model relationships in complex unstructured data, representing inherent connections through nodes and edges. In multi-object tracking, this framework is formulated as a bipartite graph matching problem, where trajectories and detections serve as nodes and edges, respectively. Optimization methods including network flows [23], multi-cut [24], minimum clique graphs [25], and disjoint paths [26] have emerged from this paradigm. Although initially prevalent in offline methods [27,28], their extension to online MOT—requiring per-frame matching of heterogeneous trajectory detection features—faces core challenges regarding the development of effective association methods. Key challenges for graph-based tracking involve (1) significant growth in computational/storage requirements during extended tracking and (2) balancing feature propagation and smoothing. Excessive propagation causes blurring of target–background distinction, while insufficient propagation limits higher-order feature extraction.
The predominant method involves creating detection graphs and trajectory graphs for target matching through bipartite graph matching. GSM [29] generates a graph for each target, evaluating target similarity through graph comparison, while in the GNMOT [30] approach, appearance graphs and matching graphs are constructed for detections and trajectories, matching scores are computed for pairs, and the optimal match is determined. GMTracker [31] transforms the matching problem into a graph matching problem based on vertex mapping relationships and employs the implicit function theorem to ensure differentiability in the matching layer. SGT [32] utilizes edge classification to ascertain whether two detections belong to the same target, aiding in the recovery of low-confidence detections. The MotionTrack method introduces an interaction module to learn interaction-aware motion patterns from short-term trajectories. This module utilizes an asymmetric adjacency matrix to depict interactions between objects and integrates information through graph convolutional networks to make predictions.

2.3. Tracking Under Occlusion

Occlusion in multi-object tracking presents a significant challenge, including obstacle occlusion, inter-object occlusion, and both short-term and long-term occlusion. To address these issues, researchers have developed various methods. For instance, in [33], motion models were employed to predict the positions of occluded pedestrians over extended periods, while in [34], robust appearance models were constructed to re-identify occluded individuals.
This paper focuses on occlusion in the context of UAV-based ground observations, which are characterized by a bird’s-eye view, the presence of numerous small objects, weak texture information, infrequent long-term occlusion, and frequent short-term inter-group occlusion among objects with similar appearances—all under non-linear platform motion. You et al. [35] proposed a method in which a social topology matrix is constructed based on spatio-temporal constraints, incorporating directional and velocity information to associate objects with their neighbors. This approach leverages statistical consistency in group motion to suppress false positives and recover missed detections. Additionally, BUSCA [36] introduced a transformer-based plug-and-play module that identifies missed detections by leveraging adjacent trajectories, motion cues, and learned trajectory labels.
While the occlusion recovery mechanism within our GCA module shares a high-level goal with “soft recovery” methods in the detection field, its underlying principle is fundamentally different. Traditional soft recovery techniques typically act as score boosters, elevating the confidence of low-score detections based on simple spatio-temporal priors like IoU with predicted track locations. In contrast, our GCA module operates as a feature enhancer. Instead of merely adjusting confidence scores, it actively propagates rich contextual information between trajectory and detection graphs via a cross-attention mechanism. This process generates an enhanced, context-aware feature representation for occluded objects at the embedding level. By considering an object’s relationship with its neighbors, GCA provides a more discriminative feature for the matching task, addressing the root cause of ambiguity rather than just its symptoms.
Based on this analysis, we argue that effectively handling occlusion in aerial view images hinges on the ability to recover low-confidence detections while minimizing false positives and carefully reassessing the appearance features of low-confidence objects. This is particularly critical in tracking-by-detection (TBD) frameworks, where discarding small objects during the detection phase precludes subsequent tracking. To safely enhance occlusion recovery without increasing false positives, it is essential to incorporate multiple validation mechanisms based on the spatio-temporal context of the target.

3. Proposed Method

This section details the proposed NOWA-MOT multi-object tracking framework. Section 3.1 provides an overview of the overall framework architecture and then Section 3.2, Section 3.3 and Section 3.4 delve into the three core components: the Low-Confidence Occlusion Recovery (LOR) module, the First-Round Association with Graph Cross-Attention (GCA) module for initial association, and the Second-Round Association with Matched Neighbor Guidance (MNG) module for secondary association.

3.1. Overview

We propose NOWA-MOT, a robust tracker designed to comprehensively address the challenges of low-altitude UAV tracking. As illustrated in Figure 2, the framework extends the classic tracking-by-detection (TBD) paradigm.
LOR: As shown in Figure 2b, the LOR module preprocesses the raw detector output. The detection confidence is dynamically adjusted by integrating multiple cues, including IoU, RIoU, location constraints, and initial detection scores—with the aim of recovering low-scoring true objects obscured by occlusion. This enables their participation in high-priority initial matching, mitigating adverse effects from detector failures.
GCA: Figure 2c illustrates the initial association stage. We employ a Graph Cross-Attention mechanism to enhance target feature representation. Specifically, a detection graph (current frame high-confidence detections) and a trajectory graph (active historical trajectories) are constructed, and node features are interactively enhanced via cross-attention between the graphs. The enriched features incorporate both intrinsic appearance and spatio-temporal contextual information from neighbors, yielding highly discriminative representations even under partial occlusion, thereby enabling accurate initial matching.
MNG: Leveraging initial matching results as priors (Figure 2d), the MNG module selects reliably matched neighboring trajectories (“reliable neighbors”) for each unmatched trajectory. Star-shaped topological features centered on unmatched trajectories/detections are built by aggregating the neighbors’ motion and appearance features. A cost matrix integrating multiple cues is computed, and final associations are resolved using the Hungarian algorithm.

3.2. Low-Confidence Object Recovery

Inspired by [37], the primary function of this module, as depicted in the figure, is to identify objects with a low detection confidence but a high likelihood of being present. Inputs include the current frame’s detections D = d i i = 1 M , d i = l o c i , conf i , with M representing the number of detections, and trajectories T = T j j = 1 N , T j = { l o c j , s t a t u s j } , where N denotes the number of trajectories and s t a t u s j indicates their current state (active or inactive). Predicted trajectory positions for the current frame, D ^ 1 , D ^ 2 , , D ^ m , are derived using Kalman filtering. Detections with a low confidence are defined by the threshold τ low , set as D 1 l o w , D 2 l o w , , D j l o w . The new confidence score c ^ d i is calculated primarily from the detection confidence c d i and a similarity evaluation function S D i , T j , which measures the similarity between detections and trajectories. If this score exceeds the threshold τ high for high-confidence objects, the target is included in the high-confidence target set D 1 high , D 2 h i g h , , D i h i g h . The calculation formula is as follows:
c ^ d i = max c d i , min ( S ( D i , T j ) , 1 )
where c d i indicates the prior confidence of the target and c ^ d i signifies the adjusted confidence. If the confidence derived from the new algorithm surpasses the original confidence, the new confidence is adopted as the target confidence.
There are several methods for setting S D i , T j , with the most basic being IoU. In this study, a multi-cue combination of IoU, Recovery Intersection over Union (RIoU), and location similarity is utilized, as illustrated in Equation (2).
While the Intersection over Union (IoU) serves as a foundational metric for defining the similarity function S D i , T j , it often proves insufficient under challenging conditions such as occlusion. To create a more robust association score, we introduce a multi-cue formulation that synergistically combines three components: standard IoU, RIoU, and a measure of location similarity, as detailed in Equation (2). Through a grid search on the validation set, the optimal weights were determined to be λ 1 = 0.4 , λ 2 = 0.3 , and λ 3 = 0.3 .
S D i , T j = λ 1 IoU D i , T j + λ 2 RIoU D i , T j + λ 3 S loc D i , T j
λ 1 + λ 2 + λ 3 = 1
I o U D i , T j represents the ratio of the area of overlap between the detection box and the trajectory to the total area, which is defined as
I o U = B 1 B 2 B 1 B 2
where B 1 and B 2 represent the areas of the detection box and the tracking box, respectively.
To better address occlusion, we introduce a custom metric termed RIoU [38]. Its core idea is that an occluded object’s detection box is often smaller than its true size. RIoU simulates a “recovered,” non-occluded state by expanding the dimensions of both the detection and trajectory prediction boxes toward a common reference size. This preprocessing allows for the calculation of a more meaningful overlap score, even when partial occlusion results in a low direct IoU, thereby more accurately reflecting the object’s true positional relationship.
The calculation of RIoU involves two key steps. The first is determining the reference width and height of the D i and T j :
w r = max w T , w D h r = max h T , h D
where w T , h T and w D , h D denote the width and height of the D i and T j , respectively.
w ^ T = w T + β w r w T h ^ T = h T + β h r h T w ^ D = w D + β w r w D h ^ D = h D + β h r h D
The recovery strength is controlled by the hyperparameter β , which adjusts the box dimensions. This parameter dictates the extent of box expansion: β = 0 corresponds to no recovery (i.e., standard IoU). To determine its optimal value, we conduct systematic ablation studies on the validation set. The results indicate that setting β to 2 strikes an optimal balance between effectively recovering occluded targets and avoiding erroneous associations with nearby objects due to over-expansion. A schematic diagram of RIoU is shown in Figure 3.
The standard IoU between the adjusted boxes T ^ and D ^ is calculated:
R I o U = | T ^ D ^ | | T ^ D ^ |
S loc is designed to quantify the similarity of adjacent bounding boxes based on the core assumption that objects within the same local topology share similar spatial characteristics. We construct this feature using the aspect ratios and relative positions of neighboring objects. Specifically, we calculate the weighted average m i t of the aspect ratio and distance for target i:
m i t = 1 m cos θ i , j t σ j t
where σ j t represents the aspect ratio of the neighbor detection box for target i in frame t and denotes the angle from neighbor node j to target i. The final location similarity is then calculated according to Equation (9).
S loc = min ( m i t , m i t 1 ) max ( m i t , m i t 1 )
To ensure the reliability of the spatial context, neighbors are not just selected based on spatial proximity, but a critical filtering step is also included: we exclusively select neighbors from high-confidence detections. In practice, we select the five nearest high-confidence neighbors to construct the location feature. This strategy is designed to mitigate noise at the source, effectively preventing low-confidence false positives from corrupting the topological structure and thereby ensuring the robustness of the S loc feature.

3.3. First-Round Association with GCA

Compared to conventional matching approaches that rely solely on appearance similarity and IoU metrics, graph-based feature embedding produces higher-order discriminative representations that significantly enhance the robustness of detection–trajectory associations. As depicted in Figure 4, a three-stage process is executed in this module: (1) construction of a bipartite graph and initialization of node features; (2) feature enhancement via a multi-head graph attention mechanism; and (3) final association using a robust similarity metric and optimal assignment.

3.3.1. Graph Initialization

We separately construct a detection graph G d = ( V d , E d ) for the high-confidence detection set D high and a trajectory graph G t = ( V t , E t ) for the active trajectory set T t 1 , employing lightweight and efficient OSNet [39] as the Re-ID backbone network. Through its distinctive omni-scale residual modules, OSNet effectively captures multiscale features spanning fine to coarse granularities. This capability is particularly crucial for recognizing small objects in UAV perspectives. For conciseness, we denote the initial features as h i 0 for nodes in G d and as h j 0 for nodes in G t . Node features are initialized by concatenating topological and appearance features:
h i ( 0 ) = f topo pos i I i θ i + f app a p p i
where
pos i = x c w 0 , y c h 0 , w w 0 , h h 0
l i = dist d i , d k / w 0 , d k N i
θ i = d k 1 d i d k / 2 π , d k 1 , d k N i .
| | denotes concatenation, ( d k 1 d i d k ) represents the angular relationship between three objects, and f topo ( · ) is an MLP (implemented as two-layer MLPs) that encodes topological features in a higher-dimensional space to 256. This representation exhibits scale and rotation invariance, accommodating camera motion in UAV scenarios. f app ( · ) projects appearance features in dimensions comparable to topological features, and N i denotes the set of neighbors for the target object i. To construct this set, we select the K-nearest neighboring objects (where K = 5 in our experiments) based on the Euclidean distance from a pool of high-confidence detections and stable trajectories.
Edge Initialization: The edges between the detection graph nodes and the trajectory graph nodes are constructed under distance constraints, with edge features representing pairwise similarity to the weight Graph Cross-Attention. We employ multi-head attention to capture node relations from both appearance and spatial perspectives. The number of heads is set as 2, and the kth head’s attention coefficient e i j k is computed as
e i j k = LeakyReLU ω k T W k h i W k h j
where ω k T denotes learnable parameters, | denotes concatenation, and W k denotes shared parameters across graphs.

3.3.2. Cross-Graph Feature Propagation

As illustrated in Figure 4, we employ a Graph Cross-Attention mechanism for feature enhancement between trajectory and detection graphs, allowing detection and trajectory nodes to exchange contextual information. This process enhances the node features by aggregating information from their respective neighbors in the opposing set.
α i j k = exp e i j k m N i exp e i m k
Node feature propagation aggregates features via
h i l + 1 = σ 1 K k = 1 K j N i α i j k W k h j l + h i l
where σ ( · ) is the activation function.
The similarity between the augmented detection node h i and the trajectory node h j is
s i , j = DualSoftmax h i , h j + IoU B i , B j , if d E B i , B j < 2 w i + w j 0 , otherwise
where d E B i , B j is the distance between B i and B j . The dual-softmax operation normalizes the raw score matrix S along both its rows and columns and computes their geometric mean. This enforces both row- and column-wise competition, ensuring that a high score for a pair ( i , j ) indicates that detection i is a strong candidate for trajectory j and that trajectory j is a strong candidate for detection i.
Finally, the matching matrix M is estimated by the Hungarian algorithm, based on the similarity matrix S .
M = Hungarian ( S )

3.3.3. Training Strategy

The primary objective of the training process is to optimize the GCA module to produce powerful node embeddings that lead to accurate matching. A significant challenge arises due to the Hungarian algorithm, used during inference to obtain the final discrete assignments, being non-differentiable. This property prevents the flow of gradients, making it unsuitable for end-to-end training.
To overcome this, we adopt a strategy inspired by SuperGlue [40], which formulates the assignment as a differentiable optimal transport problem. We utilize the Sinkhorn algorithm, a well-established method for approximating optimal transport, to convert the raw similarity matrix S into a doubly stochastic matrix. This matrix can be interpreted as a “soft assignment” matrix, where each element represents the probability of a match between a detection and a trajectory. Crucially, the Sinkhorn operation is differentiable, allowing gradients to be back-propagated through it.
S ¯ = Sin khorn ( S )
With this differentiable pipeline, we can train the GCA module by supervising its output against the ground-truth matching assignments. Our total loss function is composed of two synergistic components: matching loss to supervise the correctness of the match predictions and Node Loss to enforce the quality of the underlying feature embeddings.
Matching Loss: To address the severe class imbalance inherent in the matching task (where non-matches vastly outnumber true matches), we replace the standard cross-entropy loss with a more powerful combination of Focal Loss and Dice Loss.
The Focal Loss component alleviates class imbalance by down-weighting the loss assigned to well-classified examples, allowing the model to focus on hard-to-classify ambiguous pairs. The Dice Loss component directly optimizes the intersection between predicted and ground-truth matches and is particularly effective at forcing the model to produce higher confidence scores for positive samples. The combined matching loss is defined as
L M = β 1 · 1 N + e i , j E ( α ( 1 p t ) γ log ( p t ) ) + β 2 · 1 2 T P + 1 Σ + 1
where p t is the predicted probability for a given match, N + is the number of positive samples, and α and γ are the focusing parameters for the Focal Loss. For the Dice Loss term, TP is the true positive count (sum of probabilities for ground-truth matches) and S i g m a is the sum of probabilities over all predicted and ground-truth matches.
Node Loss: The goal of Node Loss is to structure the embedding space by ensuring that features from nodes corresponding to the same object are pulled closer together (intra-class compactness), while features from different objects are pushed apart (inter-class separability). We retain the effective margin-based contrastive loss for this purpose:
L N = ( i , j ) P d h i , h j 2 + ( i , k ) N max 0 , m d h i , h k 2
where d ( · , · ) is the Euclidean distance between the final node embeddings, P is the set of positive pairs (correct matches), N is the set of negative pairs (incorrect matches), and m is the margin hyperparameter.
Total Loss: The final loss function is a weighted sum of the Edge and Node Losses, balanced by a hyperparameter:
L total = λ loss L M + ( 1 λ loss ) L N
To ensure a principled approach, the loss weights in Equations (20) and (22) are not set heuristically. Instead, they are determined through a systematic grid search on the VisDrone2019 validation set, with HOTA as the primary optimization metric.

3.3.4. First-Round Association Algorithm

Algorithm 1 delineates our primary matching pipeline, which consists of four main stages. (1) Graph Initialization: Given the current frame’s high-confidence detections D h i g h and active trajectories from the previous frame, two separate graphs are constructed: a detection graph G d and a trajectory graph G t . The initial feature for each node in both graphs is then computed using Equation (10). (2) Cross-Graph Feature Propagation: A multi-head graph attention mechanism is employed to iteratively update and enhance the node features. This allows contextual information to be exchanged between the detection and trajectory graphs. (3) Similarity Computation: After feature enhancement, a final similarity matrix S is computed between all detection and trajectory nodes as per Equation (18). (4) Matching: The optimal assignments are resolved using the Hungarian algorithm, which outputs the final matched pairs.
Algorithm 1 First-Round Association with Graph Cross-Attention
Require: High-confidence detections D high ; Active trajectories T t 1 ; Confidence threshold η
Ensure: Matched pairs M ; Unmatched detections U d ; Unmatched trajectories U t
  1:
                             ▹ Graph Initialization
  2:
  Construct detection graph G d ( V d , E d ) from D high
  3:
  Construct trajectory graph G t ( V t , E t ) from T t 1
  4:
  for each node v i V d V t  do
  5:
         Initialize node feature h i ( 0 )
  6:
  end for
  7:
                       ▹ Cross-Graph Feature Propagation
  8:
  for l 0 to L 1 do
  9:
         Update all node features to h l + 1 via cross-graph attention
10:
  end for
11:
                             ▹ Similarity Computation
12:
  Compute similarity matrix S between V d and V t
13:
                                  ▹ Matching
14:
   M , U d , U t Hungarian ( S )
15:
  return  M , U d , U t

3.4. Second-Round Association Based on Reliable Neighbor Guidance

Building upon the reliable matches from the initial round, our framework introduces a key strategic departure from methods like ByteTrack: we retain these matched pairs to serve as contextual anchors during the second-stage association, rather than processing only the unmatched elements. We posit that these existing matches, which exhibit stable trajectory links, provide critical contextual cues for resolving more ambiguous cases. The effectiveness of this principle is empirically validated in our ablation study.
Our tracking method proceeds in four stages. First, we input the state from the previous frame: matched tracks T match and detections D match , unmatched tracks T unmatch and detections D unmatch , and low-confidence detections D low . Second, for each unmatched tracklet, we select its four nearest matched neighbors based on proximity, velocity, and direction. Third, we search the predicted area using IoU matching and topological feature similarity. Finally, during track management, new tracks are initiated from remaining high-confidence detections and low-confidence ones are discarded. Tracks that remain unmatched for 30 consecutive frames are terminated. These stages are described in detail in the following sections.

3.4.1. Spatio-Temporally Constrained Neighbor Selection

Intuitively, neighboring objects (e.g., pedestrian groups or vehicle clusters) exhibit coherent motion patterns. We therefore select adjacent trajectories for each unmatched trajectory by enforcing three constraints: distance thresholds, velocity consistency, and directional alignment.
Distance Constraint: We let x i , y i and x j , y j denote the center positions of the target trajectory and candidate neighbor trajectory, respectively, with w i representing the bounding box width. For high-confidence matches established during initial association, the detection box size inherently encodes camera perspective information and implicitly conveys target depth. Consequently, a normalized Euclidean distance metric is constructed as follows:
S i , j d = x i x j 2 + y i y j 2 w i
Directional Constraint: Adjacent trajectories inherently exhibit compatible motion patterns with the target. We thus employ a cosine similarity-based distance metric to select trajectories with directional consistency:
S i , j dir = 1 Δ x t i · Δ x t j + Δ y t i · Δ y t j ( Δ x t i ) 2 + ( Δ y t i ) 2 · ( Δ x t j ) 2 + ( Δ y t j ) 2
where Δ x t i , Δ y t i and Δ x t j , Δ y t j denote the velocity vectors of the trajectories T i and T j , respectively. A smaller S i , j dir value indicates a higher degree of directional alignment.
Velocity Constraint: Within correlated groups, object velocities are constrained by neighboring objects, resulting in comparable motion magnitudes. The velocity similarity S i , j v between trajectory T i and T j is consequently quantified as
S i , j y = Δ x t i Δ x t j 2 + Δ y t i Δ y t j 2
A smaller S i , j v indicates a higher probability of cohesive group movement.
The similarity between matched trajectories T j and target trajectories T i in the current frame is ultimately calculated as
S T i , T j = S i , j d + S i , j dir + S i , j v
The matched trajectories are ranked based on this composite score, and the top four trajectories with the lowest aggregate score are selected as the reliable neighbors for T i . Subsequently, matching attempts are performed between the current frame’s unmatched detections and the unmatched trajectory to form temporary pairs. The formation of a valid match pair is determined by comparing their group similarity.

3.4.2. Star-Shaped Topological Feature Description

As illustrated in Figure 5, in this stage, a robust, context-aware representation is constructed for each remaining object by introducing a directed graph topology. A key distinction from the first association round is our strategy to leverage the reliably matched pairs as contextual anchors. Inspired by the work of You et al. [35], we find that topological features, which characterize the precise geometric relationships between a target and its neighboring nodes, provide a highly discriminative representation. This approach is particularly effective for distinguishing between objects that exhibit similar appearances and positions.
To implement this strategy, we formally define a star-shaped topological graph for each unmatched trajectory T j denoted by G T j and for each unmatched detection D i denoted by G D i . Both graphs are defined by their respective central node (the unmatched object) and a shared set of reliable matched neighbors N m a t c h e d . By comparing the structure of these two graphs, we can measure the similarity of the unmatched objects based on their relationship with the same stable context. The topological distance between these two subgraphs is then calculated using Equation (27).
D topo ( G T j , G D i ) = x j x i w o 2 + y j y i h o 2 + k N matched ( d k j d k i ) 2 + θ k j θ k i π 2
The topological similarity S topo D i , T j between the detection D i and tracking boxes T j is normalized using Equation (28):
S topo D i , T j = exp D topo G T j , G D i
Subsequently, Hungarian matching is performed between residual detections and trajectories using S topo ( D i , T j ) .
S assoc ( D i , T j ) = S topo ( D i , T j ) + IoU ( D i , T j )

3.4.3. Second-Round Association Algorithm

As shown in Algorithm 2, given initial matched pairs and unmatched detections/trajectories, we first select reliable neighbors around unmatched detections using velocity–spatial consistency constraints. Star-shaped topologies are then constructed centering on unmatched detections and trajectories. Topological similarity metrics compare unmatched trajectories, resolved via Hungarian matching. Unlike conventional two-stage methods, this approach leverages prior matched knowledge and physical geometric constraints to reduce identity switches, ensuring reliable tracking continuity in dynamic UAV videos. The lightweight design does not require additional networks, guaranteeing deployability on UAV platforms. Unmatched detections initialize new trajectories in T t , while unmatched trajectories are marked as inactive and added to T t . Using Kalman filtering, we predict their positions for 30 frames, purging persistently unassociated trajectories.
Algorithm 2 Reliable Neighbor-Guided Second Association
Require:  M 1 : Matched pairs from first round U d : Unmatched detections U t : Unmatched tracks
Ensure:  M 2 : Matched pairs from second round U d , U t : Remaining unmatched sets
  1:
   M 2
  2:
  Initialize similarity matrix S assoc
  3:
  for each T j U t do                   ▹ Select reliable neighbors
  4:
         Compute S ( T j , T k ) for all T k M 1
  5:
          N j Top - 4 matched neighbors for T j
  6:
  end for
  7:
  for each pair ( T j , D i ) U t × U d do           ▹ Compute association similarity
  8:
         Construct subgraphs G T j and G D i using N j
  9:
         Compute S assoc ( j , i ) using Equations (27)–(29)
10:
  end for
11:
                               ▹ Perform matching
12:
   M 2 , U d , U t Hungarian ( S assoc )
13:
                              ▹ Track Management
14:
  for each T j U t  do
15:
         Update T j state using motion from N j
16:
         Terminate if age ( T j ) > 30
17:
  end for
18:
  return M 2 , U d , U t

4. Experiments

4.1. Implementation Details

4.1.1. Datasets and Metrics

To verify the effectiveness of the proposed method, experiments were conducted on two commonly used unmanned aerial vehicle (UAV) multi-target tracking datasets: VisDrone2019 [41] and UAVDT [42]. VisDrone2019 contains 56 training video sequences, 7 validation sequences, and 17 test evaluation sequences. These videos cover various scenes, including sports stadiums, commercial streets, highways, and suburban areas. When evaluating multi-target tracking tasks, the official evaluation tool for this dataset only considers five object types: people, cars, vans, trucks and buses. In contrast, UAVDT only tracks a single target category: vehicles. This dataset contains 50 sequences (30 for training and 20 for testing), mainly covering scenarios such as squares, intersections, and highways under different lighting conditions (such as sunny, nighttime, and foggy). It should be noted that we refer to the true value labels provided by SFtrack [22] for the revised test set of UAVDT.
This study features common multi-object tracking evaluation criteria used in previous work. The evaluation metrics are as follows: high-order tracking accuracy (HOTA), multi-target tracking accuracy (MOTA), IDF1, false positive (FN), false positive (FP), and ID switching (IDs). HOTA is a high-order metric used to measure tracking performance, and can better reflect the performance of tracking algorithms, while MOTA is a widely used MOT evaluation index, mainly related to detection performance (FP and FN) and ID stability.

4.1.2. Training Details

We use PyTorch Geometric 3.12 to construct graph neural networks for NOWA-MOT and train it on one NVIDIA RTX 4090 GPU for 20 epochs, with a batch size of 1. The Adam optimizer is applied with β 1 = 0.9 and β 2 = 0.999 , and the initial learning rate is 3 × 10 5 and weight decay is 1 × 10 5 . For the Re-ID module, we use an OSNet network for appearance feature extraction and pre-trained it on VisDrone and UAVDT datasets. The normalized crop size for the Re-ID module is set as (256, 128).

4.1.3. Inference Details

Due to the TBD structure of the proposed algorithm, the detection performance was adopted as the primary evaluation metric. For fairness, YoloX-m was retrained as the detector on both VisDrone and UAVDT datasets, with identical detection results provided to all comparison algorithms. Detection score thresholds were set to 0.6 (high) and 0.1 (low) and motion compensation (CMC) was employed to enhance tracking. Following the robust configuration established in BoT-SORT, we utilized the Enhanced Correlation Coefficient (ECC) algorithm to register images between consecutive frames. Specifically, the ECC method estimates the optimal affine transformation matrix that aligns the previous frame with the current one. This transformation is then applied to the trajectory locations predicted by the Kalman filter, effectively warping them into the current frame’s coordinate system before the association step. This process significantly mitigates the tracking errors caused by camera motion rather than true object movement. Target trajectories are initialized directly from the first frame’s detections. For subsequent frames, the tracker assigns detections to corresponding trajectories.

4.2. Comparison with State of the Art

NOWA-MOT was compared with other multi-object tracking (MOT) methods on UAVDT and VisDrone2019-MOT datasets. MOT methods fall into two paradigms: tracking by detection (TBD) (e.g., ByteTrack [8], BoT-SORT [6], UAVMOT [43]) and joint detection and tracking (JDT) (e.g., FairMOT [44], DroneMOT [45], GLOA [46]). Since the detector performance significantly impacts MOT metrics, identical detector weights were used for TBD methods on the same dataset. For JDT methods, retraining was performed on trackers with open-source code; results from original papers were adopted for non-open-source methods. The OSNet appearance feature extraction network was utilized for trackers requiring appearance features.

4.2.1. Results on UAVDT

NOWA-MOT was trained on the UAVDT training set and evaluated on its test set. As shown in Table 1, NOWA-MOT achieves superior performance in terms of HOTA, MOTA, and IDF1, reaching 62.69%, 69.01%, and 80.58%, indicating significant improvements in tracking accuracy and identity consistency. Compared with the two-stage tracker ByteTrack, NOWA-MOT substantially reduced ID switching (IDs: 305 → 131) and false positives (FPs: 43,056 → 36,098), achieving a 1.44% higher HOTA and demonstrating its ability to enhance tracking accuracy via multi-cue fusion. Compared with the UAV-specific tracker UAVMOT, NOWA-MOT achieved 2.5% and 8.6% gains in MOTA and IDF1, respectively. These results demonstrate that NOWA-MOT efficiently maintains tracking accuracy and identity consistency under non-linear UAV motion, outperforming competing methods.

4.2.2. Results on VisDrone

NOWA-MOT was trained on the VisDrone2019 training set and evaluated on its test development set, with the metrics, including HOTA, MOTA, IDF1, and IDs, compared against those of state-of-the-art methods. As shown in Table 2, NOWA-MOT achieved an HOTA of 51.34% and an IDF1 of 67.33%, surpassing all existing methods. VisDrone contains five target categories, with numerous small, irregularly moving objects (e.g., pedestrians, non-motor vehicles), leading to more frequent trajectory switches and association failures than for UAVDT. Compared with ByteTrack, the HOTA and MOTA of NOWA-MOT were improved by 3.96% and 4.79%, respectively, and compared to BOT-SORT, its HOTA and IDF1 were increased by 1.05% and 1.61%, respectively. It also achieved the highest IDF1 (highlighting robust identity preservation) and the lowest IDs (trajectory switches), demonstrating effectiveness in generating coherent trajectories—especially in complex environments. Compared to Unigraph (which also uses topological features), the HOTA and MOTA of NOWA-MOT were improved by 3.68% and 2.86%, respectively. These results confirm NOWA-MOT’s significant advantages in balancing detection accuracy and identity preservation in challenging UAV tracking scenarios.

4.3. Ablation Studies

In this section, we demonstrate the importance of each module and design strategy in NOWA-MOT through ablation experiments. Subsequently, we explore the sensitivity of the parameters to graph initialization and association. All models in this section were trained on the VisDrone2019 training set and evaluated on the VisDrone2019 test-dev set.

4.3.1. Effectiveness of Each Module

Using ByteTrack as the baseline (first row), we evaluated the effectiveness of each module; the results are shown in Table 3.
LOR: Without LOR, detections with confidence below 0.6 are excluded from the first matching step, as in ByteTrack. Incorporating LOR improves the HOTA and MOTA by 0.3% and 1.85%, respectively. This is because in low-altitude scenarios, the confidence scores fluctuate significantly due to occlusion and motion. Directly excluding low-confidence detections may lead to erroneous matches in crowded scenes. However, IDF1 slightly decreases by 0.48%, indicating that the module may introduce false positives (FPs) and cause track fragmentation. When combined with GCA, the performance is further improved, with HOTA, MOTA, and IDF1 increasing by 2.59%, 3.55%, and 5.5%, respectively, over the baseline. The combination of LOR and MNG also brings gains, though less pronounced, suggesting functional overlap in handling low-confidence and occluded objects.
GCA: GCA improves the HOTA and MOTA by 2.24% and 2.95%, respectively, demonstrating that graph convolution with spatio-temporal feature enhancement effectively aggregates neighborhood information and improves feature discriminability. The IDF1 also increases by 1.44%, indicating that topological features provide more stable association in dynamic drone scenarios. The combination of GCA and MNG yields the second best performance, highlighting their complementarity.
MNG: Using MNG alone improves the HOTA and IDF1 by 0.75% and 1.47%, respectively, showing its effectiveness in improving second-stage matching and trajectory stability.
The ablation study demonstrates that the GCA module is the primary computational bottleneck, reducing the FPS from 37.25 to 21.71. This is attributed to its computationally intensive graph convolution and cross-attention mechanisms. In contrast, the LOR and MNG modules are highly efficient, introducing minimal overhead as they rely on lightweight geometric calculations; MNG’s efficiency is further enhanced by reusing cached motion variables.
These results demonstrate the importance of each module in enhancing tracking performance and maintaining identity consistency.

4.3.2. Composition Analysis of the LOR Module

To further analyze the contribution of each component in LOR, we conducted fine-grained experiments, the results of which are shown in Table 4. Adding RIoU to IoU improves the MOTA by 0.44% but decreases the IDF1 by 1.2%, suggesting increased fragmentation. Incorporating S loc leads to improvements of 0.63% in HOTA, 1.41% in MOTA, and 0.73% in IDF1 (compared to the second row). Overall, the results indicate that combining multiple similarity metrics improves tracking, while RIoU should be carefully integrated to avoid adverse effects.

4.3.3. Analysis of Message Passing Steps

The number of graph neural network layers (L) impacts feature aggregation: too few layers limit neighborhood aggregation, while too many leads to over-smoothing. We tested the performance with different numbers of layers, as shown in Table 5. The results in Table 5 show that the model performs best when L = 2 , achieving a balance between aggregating contextual semantic information and avoiding feature over-smoothing.

4.3.4. Impact of Data Augmentation on the Graph Network Performance

Data augmentation exerts a significant impact on the graph network. As shown in the table, we incorporate data augmentation during the training of the graph neural network. To simulate the input of a real detector, in our approach, detector outputs are compared with the ground truth, Hungarian matching based on IoU is applied, and ground truth IDs are assigned to the corresponding detector results. This process generates a detector-based training set that more closely resembles the actual output distribution of the detector. The results are shown in Table 6 and show that our method exhibits improvements in the HOTA and MOTA of 1.33% and 1.47%, respectively.

4.3.5. Impact of Spatio-Temporal Geometric Constraints on Secondary Association

To validate the necessity of the spatio-temporal geometric constraints in MNG, we compare our method with a nearest neighbor approach. As shown in Table 7, the HOTA and IDF1 decrease by 0.89% and 3.48%, respectively, without constraints. These constraints prevent the selection of inversely moving targets, which is crucial for accurately computing the topological similarity.

4.3.6. Computational Efficiency Analysis

To evaluate the computational efficiency of our proposed framework, we compare its inference speed (Frames Per Second, FPS) against several representative state-of-the-art trackers. As shown in Table 8, our NOWA-MOT achieves the highest tracking accuracy with a HOTA score of 51.34, surpassing all compared methods.
This superior performance is achieved with an inference speed of 18.05 FPS. While trackers like ByteTrack (37.25 FPS) and MM-tracker (31.89 FPS) offer faster processing speeds, they exhibit lower tracking accuracy. Specifically, compared to the high-speed baseline ByteTrack, NOWA-MOT provides a substantial 3.96% absolute improvement in HOTA. Although BOT-SORT and MM-tracker are closer in accuracy, our method still holds a clear advantage. This analysis highlights a deliberate trade-off in our design, prioritizing tracking robustness and accuracy over raw speed. The resulting inference rate remains practical for many real-world deployment scenarios where high performance is the primary concern.

4.3.7. Parameter Sensitivity Analysis

We conducted a hyperparameter sensitivity analysis to evaluate the impact of the number of neighbors (N) on the performance of the GCA and MNG modules, with the results presented in Table 9. For the GCA module, we varied N from 3 to 9. The performance, measured by HOTA, peaks at our chosen value of N = 5 (51.34 HOTA). Using fewer neighbors (N = 3) provides insufficient contextual information, leading to a drop in accuracy, while using too many (N = 9) introduces noise from irrelevant objects, which also degrades performance. This confirms that N = 5 offers the best balance of contextual richness and noise for feature enhancement.
Similarly, for the MNG module, we tested N from 2 to 8. The model achieves optimal performance at N = 4, which provides the most stable set of contextual anchors for resolving ambiguous associations. Using too few neighbors (N = 2) makes the topological feature less reliable, while using too many (N = 6 or 8) can incorporate misleading motion patterns from more distant objects, harming accuracy. This analysis validates our choice of hyperparameters and demonstrates that the model’s performance is robust around these optimal values.

4.4. Qualitative Comparisons with State-of-the-Art Trackers

Figure 6 presents a qualitative comparison between our method and other state-of-the-art trackers. Red circles indicate the occurrence of ID switches, while green circles denote missed detections. In this video sequence, the upper part of the image contains dense small objects with similar appearances, while targets moving around the roundabout exhibit significant non-linear motion.
As observed, both ByteTrack and UAVMOT exhibit a noticeable number of ID switches, whereas our method maintains correct identities. This is because motion-based features often fail to meet the matching threshold under non-linear motion. Moreover, due to the small size of the targets, the cosine similarity of appearance features also falls short of the matching threshold, leading the tracker to mistakenly treat them as new objects. In contrast, our approach exhibits a relatively stable topological similarity. Based on the knowledge that objects do not appear out of nowhere, we associate targets with similar neighborhood contexts, thus achieving more robust data association.
Multiple trackers also exhibited missed detections of small objects. This occurred because it is difficult to associate detection boxes with existing tracks or initialize them as new trajectories due to occlusions and small object sizes. Our method addresses this issue through the LOR module, which enhances the confidence scores of low-confidence detections using contextual cues, allowing them to be associated with existing tracks or initialized as new ones.

4.5. Discussion

NOWA-MOT demonstrates good adaptability to the specific requirements of multi-object tracking in low-altitude scenarios; however, there are areas that warrant improvement. In particular, a targeted long-term occlusion strategy was not developed. This decision was made based on the enhanced field of view at low altitudes, where prolonged occlusion between objects is less frequent. Consequently, specific enhancements to streamline the algorithm and enhance its real-time performance have not been pursued. Moreover, our approach employs feature aggregation through graph neural networks in the second step of trajectory association. While graph neural networks are effective in capturing point relationships, the interactions between trajectories are intricate and diverse. Future enhancements could involve utilizing mechanisms like hypergraphs to link multiple nodes through hyperedges, facilitating exploration of feature transfers within groups and enhancing motion information extraction. This approach could lead to more precise trajectory predictions and enable progressive matching strategies from the group to the individual level using hypergraph matching methods. Furthermore, it is important to note that our methodology relies on the YOLOX detector. Should a different detector be employed, adjustments to certain hyperparameters may be necessary to align with the detector’s performance.
Our approach is limited by its core assumption that topological neighbors share similar spatial features. This assumption fails in complex scenarios such as “stereo traffic” overpasses, where vehicles on different levels appear adjacent in the 2D projection but are in reality independent. As shown in Figure 7, other challenging cases include low-light conditions that degrade detection, scenes with highly non-linear motion and occlusion like basketball courts, and situations with too few observable neighbors. These conditions can lead to tracking failures, evidenced by an increase in ID switches. Addressing these issues requires future work, such as incorporating targeted image enhancement and leveraging visual-language models to integrate scene priors.

5. Conclusions

This paper introduces a novel multi-object tracking algorithm to address challenges that arise due to a low-altitude perspective, such as small target sizes, similar appearances, and frequent mutual occlusion. This method leverages multiple cues to recover low-confidence occluded objects. Subsequently, Graph Cross-Attention is applied to enhance the features of both detection and trajectory nodes, enabling first-stage matching based on the similarity between detection and trajectory nodes. Furthermore, a novel cascaded second-stage matching strategy is introduced. Unlike previous, relatively isolated cascaded matching methods, the proposed approach utilizes already matched detection–trajectory pairs to identify the neighboring trajectories of unmatched ones based on motion characteristics such as velocity and direction. These neighboring trajectories are then used to construct star-shaped topological features, leading to more robust matching results. Thanks to these improvements, NOWA-MOT’s HOTA on the VisDrone test-dev set reached 51.34%, an increase of 3.96% compared to Bytetrack’s 47.38%.

Author Contributions

Conceptualization, H.Q.; methodology, H.Q.; software, H.Q. and X.G.; validation, H.Q., R.G. and B.D.; formal analysis, H.Q.; investigation, S.S.; resources, S.S.; data curation, H.Q.; writing—original draft preparation, H.Q.; writing—review and editing, H.Q. and X.S.; visualization, H.Q. and B.D.; supervision, H.Q.; project administration, H.Q.; funding acquisition, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Graduate innovation program of National University of Defense Technology, grant number: XJZH2024016.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Guo, Y.; Li, Y.; Wang, S.; Sun, K.; Liu, M.; Wang, Z. Pedestrian multi-object tracking combining appearance and spatial characteristics. Expert Syst. Appl. 2025, 272, 126772. [Google Scholar] [CrossRef]
  2. Li, M.; Zhang, Y.; Jia, Y.; Yang, Y. Advancing multi-object tracking through occlusion-awareness and trajectory optimization. Knowl.-Based Syst. 2025, 310, 112930. [Google Scholar] [CrossRef]
  3. Liu, Y.; Liu, X.; Jiang, Z.; Liu, J. Co-MOT: Exploring the Collaborative Relations in Traffic Flow for 3D Multi-Object Tracking. IEEE Trans. Intell. Transport. Syst. 2025, 26, 4744–4756. [Google Scholar] [CrossRef]
  4. Seidenschwarz, J.; Brasó, G.; Serrano, V.C.; Elezi, I.; Leal-Taixé, L. Simple Cues Lead to a Strong Multi-Object Tracker. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 13813–13823. [Google Scholar] [CrossRef]
  5. Tang, Z.; Naphade, M.; Liu, M.Y.; Yang, X.; Birchfield, S.; Wang, S.; Kumar, R.; Anastasiu, D.; Hwang, J.N. CityFlow: A City-Scale Benchmark for Multi-Target Multi-Camera Vehicle Tracking and Re-Identification. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 8789–8798. [Google Scholar] [CrossRef]
  6. Aharon, N.; Orfaig, R.; Bobrovsky, B.Z. Bot-sort: Robust associations multi-pedestrian tracking. arXiv 2022, arXiv:2206.14651. [Google Scholar]
  7. Du, Y.; Zhao, Z.; Song, Y.; Zhao, Y.; Su, F.; Gong, T.; Meng, H. StrongSORT: Make DeepSORT Great Again. IEEE Trans. Multimed. 2023, 25, 8725–8737. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Sun, P.; Jiang, Y.; Yu, D.; Weng, F.; Yuan, Z.; Luo, P.; Liu, W.; Wang, X. ByteTrack: Multi-object Tracking by Associating Every Detection Box. In Proceedings of the Computer Vision—ECCV 2022, Tel Aviv, Israel, 23–27 October 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer: Cham, Switzerland, 2022; pp. 1–21. [Google Scholar] [CrossRef]
  9. Sun, Z.; Wei, G.; Fu, W.; Ye, M.; Jiang, K.; Liang, C.; Zhu, T.; He, T.; Mukherjee, M. Multiple Pedestrian Tracking Under Occlusion: A Survey and Outlook. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 1009–1027. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Liang, Y.; Leng, J.; Wang, Z. SCGTracker: Spatio-temporal correlation and graph neural networks for multiple object tracking. Pattern Recognit. 2024, 149, 110249. [Google Scholar] [CrossRef]
  11. Qin, Z.; Zhou, S.; Wang, L.; Duan, J.; Hua, G.; Tang, W. MotionTrack: Learning Robust Short-Term and Long-Term Motions for Multi-Object Tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 17939–17948. [Google Scholar] [CrossRef]
  12. Xu, X.; Ren, W.; Sun, G.; Ji, H.; Gao, Y.; Liu, H. GroupTrack: Multi-Object Tracking by Using Group Motion Patterns. In Proceedings of the 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 14–18 October 2024; pp. 4896–4903. [Google Scholar] [CrossRef]
  13. Liu, S.; Shen, X.; Xiao, S.; Li, H.; Tao, H. A Multi-Scale Feature-Fusion Multi-Object Tracking Algorithm for Scale-Variant Vehicle Tracking in UAV Videos. Remote Sens. 2025, 17, 1014. [Google Scholar] [CrossRef]
  14. Fu, H.; Guan, J.; Jing, F.; Wang, C.; Ma, H. A real-time multi-vehicle tracking framework in intelligent vehicular networks. China Commun. 2021, 18, 89–99. [Google Scholar] [CrossRef]
  15. Cao, J.; Pang, J.; Weng, X.; Khirodkar, R.; Kitani, K. Observation-centric sort: Rethinking sort for robust multi-object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 9686–9696. [Google Scholar]
  16. Bochinski, E.; Eiselein, V.; Sikora, T. High-Speed tracking-by-detection without using image information. In Proceedings of the 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Lecce, Italy, 29 August–1 September 2017; pp. 1–6. [Google Scholar] [CrossRef]
  17. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar] [CrossRef]
  18. Zhang, J.; Wang, M.; Jiang, H.; Zhang, X.; Yan, C.; Zeng, D. STAT: Multi-Object Tracking Based on Spatio-Temporal Topological Constraints. IEEE Trans. Multimed. 2024, 26, 4445–4457. [Google Scholar] [CrossRef]
  19. Yang, M.; Han, G.; Yan, B.; Zhang, W.; Qi, J.; Lu, H.; Wang, D. Hybrid-sort: Weak cues matter for online multi-object tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 6504–6512. [Google Scholar]
  20. Chen, Y.; Wang, J.; Zhou, Q.; Hu, H. ArbiTrack: A Novel Multi-Object Tracking Framework for a moving UAV to Detect and Track Arbitrarily Oriented Targets. IEEE Trans. Multimed. 2025, 27, 5387–5397. [Google Scholar] [CrossRef]
  21. Deng, C.; Wu, J.; Han, Y.; Wang, W.; Chanussot, J. Learning a robust topological relationship for online multi-object tracking in uav scenarios. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5628615. [Google Scholar] [CrossRef]
  22. Song, I.; Lee, J. SFTrack: A Robust Scale and Motion Adaptive Algorithm for Tracking Small and Fast Moving Objects. In Proceedings of the 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Abu Dhabi, United Arab Emirates, 14–18 October 2024; pp. 10870–10877. [Google Scholar] [CrossRef]
  23. Zhang, L.; Li, Y.; Nevatia, R. Global data association for multi-object tracking using network flows. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar] [CrossRef]
  24. Tang, S.; Andriluka, M.; Andres, B.; Schiele, B. Multiple People Tracking by Lifted Multicut and Person Re-identification. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3701–3710. [Google Scholar] [CrossRef]
  25. Roshan Zamir, A.; Dehghan, A.; Shah, M. GMCP-Tracker: Global Multi-object Tracking Using Generalized Minimum Clique Graphs. In Computer Vision—ECCV 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7573, pp. 343–356. [Google Scholar] [CrossRef]
  26. Tang, S.; Andres, B.; Andriluka, M.; Schiele, B. Subgraph decomposition for multi-target tracking. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5033–5041. [Google Scholar] [CrossRef]
  27. Cetintas, O.; Brasó, G.; Leal-Taixé, L. Unifying Short and Long-Term Tracking with Graph Hierarchies. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 22877–22887. [Google Scholar] [CrossRef]
  28. Brasó, G.; Leal-Taixé, L. Learning a Neural Solver for Multiple Object Tracking. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 6246–6256. [Google Scholar] [CrossRef]
  29. Liu, Q.; Chu, Q.; Liu, B.; Yu, N. GSM: Graph Similarity Model for Multi-Object Tracking. In Proceedings of the International Joint Conference on Artificial Intelligence, Yokohama, Japan, 11–17 July 2020. [Google Scholar]
  30. Li, J.; Gao, X.; Jiang, T. Graph Networks for Multiple Object Tracking. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020. [Google Scholar]
  31. He, J.; Huang, Z.; Wang, N.; Zhang, Z. Learnable Graph Matching: Incorporating Graph Partitioning with Deep Feature Learning for Multiple Object Tracking. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 5295–5305. [Google Scholar] [CrossRef]
  32. Hyun, J.; Kang, M.; Wee, D.; Yeung, D.Y. Detection Recovery in Online Multi-Object Tracking with Sparse Graph Tracker. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023; pp. 4839–4848. [Google Scholar] [CrossRef]
  33. Lin, J.; Liang, G.; Zhang, R. LTTrack: Rethinking the Tracking Framework for Long-Term Multi-Object Tracking. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 9866–9881. [Google Scholar] [CrossRef]
  34. Gao, R.; Wang, L. MeMOTR: Long-term memory-augmented transformer for multi-object tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 9901–9910. [Google Scholar]
  35. You, S.; Yao, H.; Xu, C. Multi-Object Tracking With Spatial-Temporal Topology-Based Detector. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 3023–3035. [Google Scholar] [CrossRef]
  36. Vaquero, L.; Xu, Y.; Alameda-Pineda, X.; Brea, V.M.; Mucientes, M. Lost and Found: Overcoming Detector Failures in Online Multi-object Tracking. In Computer Vision—ECCV 2024; Leonardis, A., Ricci, E., Roth, S., Russakovsky, O., Sattler, T., Varol, G., Eds.; Springer Nature: Cham, Switzerland, 2025; Volume 15131, pp. 448–466. [Google Scholar] [CrossRef]
  37. Stanojević, V.; Todorović, B. BoostTrack++: Using tracklet information to detect more objects in multiple object tracking. arXiv 2025, arXiv:2408.13003. [Google Scholar] [CrossRef]
  38. Jin, H.; Nie, X.; Yan, Y.; Chen, X.; Zhu, Z.; Qi, D. AHOR: Online Multi-Object Tracking With Authenticity Hierarchizing and Occlusion Recovery. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 8253–8265. [Google Scholar] [CrossRef]
  39. Zhou, K.; Yang, Y.; Cavallaro, A.; Xiang, T. Omni-Scale Feature Learning for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3701–3711. [Google Scholar] [CrossRef]
  40. Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. SuperGlue: Learning Feature Matching With Graph Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  41. Chen, G.; Wang, W.; He, Z.; Wang, L.; Yuan, Y.; Zhang, D.; Zhang, J.; Zhu, P.; Van Gool, L.; Han, J.; et al. VisDrone-MOT2021: The Vision Meets Drone Multiple Object Tracking Challenge Results. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 2839–2846. [Google Scholar] [CrossRef]
  42. Yu, H.; Li, G.; Zhang, W.; Huang, Q.; Du, D.; Tian, Q.; Sebe, N. The Unmanned Aerial Vehicle Benchmark: Object Detection, Tracking and Baseline. Int. J. Comput. Vis. 2020, 128, 1141–1159. [Google Scholar] [CrossRef]
  43. Liu, S.; Li, X.; Lu, H.; He, Y. Multi-Object Tracking Meets Moving UAV. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 8866–8875. [Google Scholar] [CrossRef]
  44. Zhang, Y.; Wang, C.; Wang, X.; Zeng, W.; Liu, W. FairMOT: On the Fairness of Detection and Re-identification in Multiple Object Tracking. Int. J. Comput. Vis. 2021, 129, 3069–3087. [Google Scholar] [CrossRef]
  45. Wang, P.; Wang, Y.; Li, D. DroneMOT: Drone-based Multi-Object Tracking Considering Detection Difficulties and Simultaneous Moving of Drones and Objects. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 7397–7404. [Google Scholar] [CrossRef]
  46. Shi, L.; Zhang, Q.; Pan, B.; Zhang, J.; Su, Y. Global-Local and Occlusion Awareness Network for Object Tracking in UAVs. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 8834–8844. [Google Scholar] [CrossRef]
  47. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar] [CrossRef]
  48. Yao, M.; Wang, J.; Peng, J.; Chi, M.; Liu, C. FOLT: Fast Multiple Object Tracking from UAV-captured Videos Based on Optical Flow. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 3375–3383. [Google Scholar] [CrossRef]
  49. Ren, L.; Yin, W.; Diao, W.; Fu, K.; Sun, X. SuperMOT: Decoupling Motion and Fusing Temporal Pyramid Features for UAV Multiobject Tracking. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 14188–14202. [Google Scholar] [CrossRef]
  50. Yao, M.; Peng, J.; He, Q.; Peng, B.; Chen, H.; Chi, M.; Liu, C.; Benediktsson, J.A. MM-Tracker: Motion Mamba for UAV-platform Multiple Object Tracking. AAAI 2025, 39, 9409–9417. [Google Scholar] [CrossRef]
  51. Shim, K.; Ko, K.; Yang, Y.; Kim, C. Focusing on Tracks for Online Multi-Object Tracking. In Proceedings of the 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 10–17 June 2025; pp. 11687–11696. [Google Scholar] [CrossRef]
Figure 1. Challenges in UAV-based MOT.
Figure 1. Challenges in UAV-based MOT.
Remotesensing 17 03502 g001
Figure 2. The overall architecture of NOWA-MOT is as follows: Tracklets and the detections at the tth frame are fed into a backbone network and a Re-ID module to extract motion and appearance features. Subsequently, the LOR module restores omitted objects with a low detection confidence. Following this, the GCA module enhances features between detections and tracklets via Graph Cross-Attention mechanisms and accomplishes first-stage association. Finally, the MNG module performs second-stage matching by leveraging reliably matched neighbors as a reference.
Figure 2. The overall architecture of NOWA-MOT is as follows: Tracklets and the detections at the tth frame are fed into a backbone network and a Re-ID module to extract motion and appearance features. Subsequently, the LOR module restores omitted objects with a low detection confidence. Following this, the GCA module enhances features between detections and tracklets via Graph Cross-Attention mechanisms and accomplishes first-stage association. Finally, the MNG module performs second-stage matching by leveraging reliably matched neighbors as a reference.
Remotesensing 17 03502 g002
Figure 3. Visualization of the RIoU process.
Figure 3. Visualization of the RIoU process.
Remotesensing 17 03502 g003
Figure 4. Framework of the First-Rround Association using GCA.
Figure 4. Framework of the First-Rround Association using GCA.
Remotesensing 17 03502 g004
Figure 5. Star-shaped topological feature description.
Figure 5. Star-shaped topological feature description.
Remotesensing 17 03502 g005
Figure 6. Qualitative results of UGT. The frames in this figure are from uav0000355_00001_v in the VisDrone2019-MOT-test-dev set.
Figure 6. Qualitative results of UGT. The frames in this figure are from uav0000355_00001_v in the VisDrone2019-MOT-test-dev set.
Remotesensing 17 03502 g006
Figure 7. Examples of failure cases. IDSW indicates that the target is assigned a different identity compared to its track in the previous frame. Each target with IDSW is annotated accordingly.
Figure 7. Examples of failure cases. IDSW indicates that the target is assigned a different identity compared to its track in the previous frame. Each target with IDSW is annotated accordingly.
Remotesensing 17 03502 g007
Table 1. Comparison with state-of-the-art online MOT methods on UAVDT. Metrics with ↑ (↓) indicate higher (lower) values are better. The best results are indicated in bold. Results marked with * are taken from the original paper. This convention applies to all subsequent tables.
Table 1. Comparison with state-of-the-art online MOT methods on UAVDT. Metrics with ↑ (↓) indicate higher (lower) values are better. The best results are indicated in bold. Results marked with * are taken from the original paper. This convention applies to all subsequent tables.
MethodHOTA ↑MOTA ↑IDF1 ↑MOTP ↑MT ↑ML ↓FP ↓FN ↓IDs ↓
SORT [47]41.2438.9943.7174.2849341033,037172,6282350
DeepSORT[17]61.4868.4578.3880.0381711436,67070,044850
ByteTrack [8]61.2567.1978.8679.9383010943,05668,509305
UAVMOT  [43]57.7466.5171.9880.0482211343,19168,8692100
StrongSORT [7]61.3768.3278.1780.1280811336,51770,485985
FLOT * [48]48.568.380.124,105107,630800
DroneMOT * [45]50.1069.6074.5063817857,411112,548129
GLOA * [46]49.6068.9079.8062622055,822115,567433
BOT-SORT  [6]61.9967.8679.7380.4576312731,03578,439184
SuperMOT * [49]50.6070.6075.7067422760,587107,538407
MM-tracker [50]58.8762.4876.1379.9972316831,64995,958295
Tracktrack [51]57.6258.1670.9677.6872316830,547101,132781
Ours62.6969.0180.5880.1381411436,09869,407131
Table 2. Comparison with state-of-the-art online MOT methods on VisDrone2019.
Table 2. Comparison with state-of-the-art online MOT methods on VisDrone2019.
MethodHOTA ↑MOTA ↑IDF1 ↑MOTP ↑MT ↑ML ↓FP ↓FN ↓IDs ↓
SORT [47]33.2932.1821.6168.5131851164,54885,4535728
DeepSORT [17]45.7543.2755.9175.0684924950,11860,4364616
ByteTrack [8]47.3845.6860.3774.8987228545,21162,8002274
UAVMOT [43]39.4336.1051.0074.2052057427,983115,9252775
StrongSORT [7]46.0343.1656.4075.0784625149,71660,6725013
Unigraph [21]47.6647.6160.0175.6289529042,52761,5322306
FOLT * [48]42.156.977.624,105107,630800
DroneMOT* [45]43.758.671.468939741,99886,1771112
GLOA * [46]39.146.276.158182418,715158,0434426
BOT-SORT [6]50.2949.5565.7276.9071647114,26088,377483
SuperMOT * [49]51.7066.7077.289240730528795791105
MM-tracker [50]50.5450.6165.5975.1271544315,93483,781558
Tracktrack [51]50.7551.5467.0375.8681033727,60869,5581202
Ours51.3450.4767.3376.4088328923,23976,524793
Table 3. Ablation study for each module of NOWA-MOT.
Table 3. Ablation study for each module of NOWA-MOT.
LORGCAMNGHOTA ↑MOTA ↑IDF1 ↑IDs ↓FPS ↑
47.3845.6860.37227437.25
47.6847.5359.89214035.44
49.6248.6361.81110521.71
48.1347.5661.84156434.09
49.9749.2365.8786219.38
48.3447.6362.25147732.68
50.8849.5665.8893719.76
51.3450.4767.3379318.05
Table 4. Ablation study on the components of the LOR module.
Table 4. Ablation study on the components of the LOR module.
IoURIoU S loc HOTA ↑MOTA ↑IDF1 ↑IDs ↓
47.3845.6860.372274
47.0546.1259.162377
47.6847.5359.892140
Table 5. Analysis of the impact of different numbers of message passing layers.
Table 5. Analysis of the impact of different numbers of message passing layers.
LHOTA ↑MOTA ↑IDF1 ↑
145.5447.5657.78
249.6248.6361.81
346.1748.2458.48
Table 6. Impact of data augmentation on the graph network performance.
Table 6. Impact of data augmentation on the graph network performance.
AugmentationHOTA ↑MOTA ↑IDF1 ↑
w/o48.2947.1660.01
w/49.6248.6361.81
Table 7. Ablation study on spatio-temporal geometric constraints in secondary association.
Table 7. Ablation study on spatio-temporal geometric constraints in secondary association.
ConstraintsHOTA ↑MOTA ↑IDF1 ↑
Nearest Neighbor47.2447.5458.36
Proposed (w/ Constraints)48.1347.5661.84
Table 8. Algorithm speed ablation study.
Table 8. Algorithm speed ablation study.
MethodsHOTA ↑MOTA ↑IDF1 ↑FPS ↑
ByteTrack [8]47.3845.6860.3737.25
BOT-SORT  [6]50.2949.5565.7224.16
MM-tracker [50]50.5450.6165.5931.89
NOWA-MOT51.3450.4767.3318.05
Table 9. Hyperparameter sensitivity analysis for the number of neighbors (N) in GCA and MNG modules on the VisDrone dataset.
Table 9. Hyperparameter sensitivity analysis for the number of neighbors (N) in GCA and MNG modules on the VisDrone dataset.
ModuleNHOTA ↑MOTA ↑IDF1 ↑
GCA349.1350.0866.53
551.3450.4767.33
750.7250.3566.84
947.2948.0265.21
MNG249.5148.2462.75
451.3450.4767.33
649.6449.3964.91
848.4249.1163.48
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qian, H.; Sun, X.; Guo, R.; Su, S.; Ding, B.; Guo, X. Low-Altitude Multi-Object Tracking via Graph Neural Networks with Cross-Attention and Reliable Neighbor Guidance. Remote Sens. 2025, 17, 3502. https://doi.org/10.3390/rs17203502

AMA Style

Qian H, Sun X, Guo R, Su S, Ding B, Guo X. Low-Altitude Multi-Object Tracking via Graph Neural Networks with Cross-Attention and Reliable Neighbor Guidance. Remote Sensing. 2025; 17(20):3502. https://doi.org/10.3390/rs17203502

Chicago/Turabian Style

Qian, Hanxiang, Xiaoyong Sun, Runze Guo, Shaojing Su, Bing Ding, and Xiaojun Guo. 2025. "Low-Altitude Multi-Object Tracking via Graph Neural Networks with Cross-Attention and Reliable Neighbor Guidance" Remote Sensing 17, no. 20: 3502. https://doi.org/10.3390/rs17203502

APA Style

Qian, H., Sun, X., Guo, R., Su, S., Ding, B., & Guo, X. (2025). Low-Altitude Multi-Object Tracking via Graph Neural Networks with Cross-Attention and Reliable Neighbor Guidance. Remote Sensing, 17(20), 3502. https://doi.org/10.3390/rs17203502

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop