Next Article in Journal
Exploring Monthly Variation of Gait Asymmetry During In-Hand Trot in Thoroughbred Racehorses in Race Training
Previous Article in Journal
Spatiotemporal Risk Assessment of H5 Avian Influenza in China: An Interpretable Machine Learning Approach to Uncover Multi-Scale Drivers
Previous Article in Special Issue
AI and Data Analytics in the Dairy Farms: A Scoping Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Framework for Multi-Object Tracking in Space Animal Behavior Studies

1
Technology and Engineering Center for Space Utilization, Chinese Academy of Sciences, Beijing 100094, China
2
Key Laboratory of Space Utilization, Chinese Academy of Sciences, Beijing 100094, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Animals 2025, 15(16), 2448; https://doi.org/10.3390/ani15162448
Submission received: 20 July 2025 / Revised: 18 August 2025 / Accepted: 19 August 2025 / Published: 20 August 2025
(This article belongs to the Special Issue Artificial Intelligence as a Useful Tool in Behavioural Studies)

Simple Summary

This study presents a deep learning-based multi-object tracking framework tailored for tracking animals in space environments, addressing challenges like microgravity-induced erratic movements, frequent occlusions, and high visual similarity among individuals. The framework decouples appearance and motion features via dual-stream inputs; employs modality-specific encoders fused through a heterogeneous graph network to model cross-modal spatio-temporal relationships; and integrates an object re-detection module for identity continuity during occlusions or rapid movements. Validated on public datasets of space-observed Drosophila and zebrafish, it outperforms existing methods, highlighting artificial intelligence’s potential in behavioral studies under extreme space conditions to support space life sciences research.

Abstract

In space environments, microgravity, high radiation, and weak magnetic fields induce behavioral alterations in animals, resulting in erratic movement patterns that complicate tracking. These challenges impede accurate behavioral analysis, especially in multi-object scenarios. To address this issue, this study proposes a deep learning-based multi-object tracking (MOT) framework specifically designed for space animals. The proposed method decouples appearance and motion features through dual-stream inputs and employs modality-specific encoders (MSEs), which are fused via a heterogeneous graph network to model cross-modal spatio-temporal relationships. Additionally, an object re-detection module is integrated to maintain identity continuity during occlusions or rapid movements. This approach is validated using public datasets of space-observed Drosophila and zebrafish, with experimental results demonstrating superior performance compared with existing tracking methods. This work highlights the potential of artificial intelligence as a valuable tool in behavioral studies, enabling reliable animal tracking and analysis under extreme space conditions and supporting future research in space life sciences.

1. Introduction

Space-based life science experiments are increasingly critical for elucidating biological adaptation mechanisms in altered gravity environments and advancing innovative biotechnologies [1,2]. NASA’s Open Science platforms provide access to extensive behavioral datasets from rodent studies conducted aboard the International Space Station (ISS) [3], with recent analyses demonstrating that habitat factors such as elevated microgravity significantly alter movement patterns and activity rhythms—findings that inform the optimization of future animal housing and mission design. Parallel efforts by Japan’s National Space Development Agency (NASDA) include developing aquatic animal experiment facilities for NASA space shuttle missions, yielding valuable insights into life support systems and in-flight research protocols [4].
Building on these international advancements, China has enhanced its space life science capabilities through the completion of the China Space Station (CSS), which offers a long-term, stable orbital platform for investigating animal physiological and behavioral responses to microgravity, high radiation, and weak magnetic fields [5,6,7]. Among model organisms used in such research, Drosophila and zebrafish are particularly valuable due to their genetic homology with humans, short life cycles, and well-characterized behavioral repertoires—traits that facilitate elucidation of fundamental biological adaptation mechanisms in space [8]. Zebrafish specifically have emerged as a key species for testing integrated life support systems in long-duration missions, where development of biological life support systems is imperative. This context frames the 43-day closed aquatic ecosystem experiment aboard the CSS, which successfully maintained carefully selected zebrafish (screened for behavioral, physiological, and social compatibility) alongside Ceratophyllum and microorganisms, achieving the main scientific objectives [9].
Advances in experimental methodologies have allowed high-resolution imaging systems aboard the CSS to generate massive video datasets, capturing detailed behavioral dynamics of animals over extended periods. These rich behavioral datasets require automated, accurate, and robust tracking tools to circumvent the limitations of traditional manual or contact-based observation methods, which are labor-intensive and impractical in space environments [10,11,12].
Tracking animals in space environments presents unique and largely unaddressed challenges. The synergistic effects of microgravity, radiation, and altered magnetic fields induce erratic and unpredictable movement patterns, posing significant obstacles to conventional tracking algorithms [13,14]. Three key challenges are particularly prominent: first, animals within the same culture unit often exhibit high appearance similarity due to synchronized development, complicating individual differentiation—especially during trajectory crossings; second, microgravity induces complex motion patterns (e.g., sudden accelerations and sharp turns) that exceed the modeling capacity of traditional motion-based tracking approaches; and third, dense group interactions cause frequent occlusions, leading to trajectory fragmentation and identity switches in tracking outputs.
To address these challenges, this study proposes a novel deep learning-based MOT framework specifically designed for the unique scenarios of space animal tracking, including microgravity-induced erratic motion, high appearance similarity, and frequent occlusions. The framework integrates a heterogeneous graph neural network architecture with decoupled appearance–motion modeling and a cross-modal re-identification mechanism.
The proposed framework is evaluated using the publicly available SpaceAnimals Dataset, which comprises space-observed Drosophila and zebrafish captured under real space environmental conditions. Experimental results demonstrate that this method significantly outperforms state-of-the-art MOT approaches in both tracking accuracy and robustness. The key contributions of this study are as follows:
  • A multi-modal feature fusion framework: A deep learning architecture is proposed that separates and integrates appearance and motion features of space animals via a heterogeneous graph network, enhancing MOT performance in extreme space environments.
  • A motion decoupling method: A local polynomial approximation method is introduced to decompose motion components, enabling accurate estimation of speed and acceleration and improving tracking robustness for space animals under microgravity.
  • A cross-modal re-detection module: A cross-modal re-detection method is designed to align appearance and motion features for identity maintenance, facilitating recovery of lost tracks during occlusions or rapid movements of space animals.
This work underscores the potential of artificial intelligence (AI) as a powerful tool in animal behavioral studies, providing reliable solutions for space animal tracking and behavior analysis in orbital and deep-space environments. It further opens new avenues for future applications in space life sciences and the investigation of animal behavioral genetics.

2. Related Work

Recent advancements in AI and computer vision have significantly enhanced MOT, facilitating accurate and non-intrusive behavioral analysis in complex environments [15]. MOT methods are primarily categorized into detection-based tracking (DBT) and joint detection and tracking (JDT). DBT leverages object detectors such as YOLO [16,17] or Faster R-CNN [18], followed by association strategies based on appearance or motion [19]. However, appearance-based methods are susceptible to occlusions and lighting variations, while motion-based approaches often fail under abrupt movements [20].
JDT methods integrate detection and tracking within a unified framework. For instance, Transformer-based models like TransTrack [21] enable simultaneous detection and tracking via global attention, albeit with high computational overhead. FairMOT [22], built on CenterNet [23], enhances identity preservation in crowded scenes by incorporating Re-ID features into the detection process.
Despite substantial progress in MOT, existing algorithms encounter notable challenges when applied to animal behavioral analysis. Specifically, many animal species exhibit minimal inter-individual visual differences, rendering appearance-based tracking methods prone to identity mismatches and tracking errors [24,25,26]. To address this limitation, researchers have increasingly adopted motion-based tracking approaches [27,28,29].
Manoukis et al. [30] introduced a constant-velocity Markov process augmented with stochastic perturbations to improve tracking accuracy in densely populated animal groups. Angarita [31] enhanced tracking continuity by incorporating a global temporal smoothness optimization strategy. Wu et al. [32] addressed one-to-one matching constraints through a combination of particle filtering and multi-view tracking. Wang et al. [33] further advanced this line of research by integrating long short-term memory (LSTM) networks with particle filtering to more effectively model animal movement dynamics. Recently, Yin et al. [34] demonstrated the potential of Transformer-based architectures in capturing intricate motion patterns within animal groups, highlighting the efficacy of attention mechanisms for long-term tracking in complex environments. WildLive [35] enables real-time, onboard animal tracking in drones through optimized YOLO detection and sparse optical flow, achieving 17.8 fps. Tested on over 200K annotated wildlife videos, the system outperforms existing methods and supports autonomous wildlife monitoring. Tang et al. [36] introduced the anti-drift pose tracker (ADPT), a Transformer-based method that significantly reduces drift and outperforms existing tools in accuracy, achieving up to 99.72% identity accuracy in social mouse interactions while enabling efficient, real-time, end-to-end analysis.

3. Method

This paper presents an MOT framework based on a Multi-modal Heterogeneous Graph Transformer (MHGT) for space animals. As illustrated in Figure 1, the framework fuses spatio-temporal information from appearance and motion features to enable joint detection and tracking, comprising three core components: motion decoupling, cross-modal feature fusion, and a unified detection–tracking module.

3.1. Motion Decoupling

The proposed motion decoupling framework is designed to effectively separate meaningful motion patterns from noise and artifacts in space animal tracking scenarios. It adopts a three-stage hierarchical architecture: motion modeling, polar transformation with adaptive thresholding, and dual-threshold filtering. These stages operate sequentially to extract reliable motion features for subsequent multi-modal fusion.
The first stage models local motion via second-order Taylor expansion to approximate intensity variations within small spatial neighborhoods. Specifically, for a pixel location ( x , y ) , the image intensity is expressed as:
I ( x , y ) = 1 2 x y A x x A x y A x y A y y x y + b x b y x y + c
where A R 2 × 2 denotes the Hessian matrix capturing local curvature, b R 2 represents the gradient vector, and c is the intensity at the center. For two consecutive frames I t and I t + Δ t , the displacement ( Δ x , Δ y ) is estimated by solving the regularized optimization problem.
min A , b , c ( x , y ) Ω I t ( x , y ) I t + Δ t ( x + Δ x , y + Δ y ) 2 2 + λ α A F 2 + β b 2 2
where Ω denotes a 5 × 5 local window and  α = 0.1 and β = 0.01 are empirically determined to mitigate overfitting.
The resulting displacement vectors are transformed into polar coordinates to enhance the capture of motion magnitude and direction.
ρ = Δ x 2 + Δ y 2 θ = arctan 2 ( Δ y , Δ x )
Local statistics of the motion magnitude ρ are computed within sliding windows to derive the mean μ ρ and standard deviation σ ρ . Based on these statistics, adaptive thresholds are defined as:
τ low = μ ρ + k 1 σ ρ , τ high = μ ρ + k 2 σ ρ
with k 1 = 1.0 and k 2 = 2.5 being selected to effectively differentiate between noise and actual motion.
In the final filtering stage, each pixel is categorized into one of three classes based on its motion magnitude.
M ( x , y ) = 0 if ρ ( x , y ) < τ low 1 if τ low ρ ( x , y ) τ high 2 if ρ ( x , y ) > τ high
Here, M ( x , y ) = 0 suppresses static background regions, M ( x , y ) = 1 marks uncertain motion, and  M ( x , y ) = 2 indicates confident motion. The final motion features are retained as:
F motion = { ( ρ ( x , y ) , θ ( x , y ) ) M ( x , y ) > 0 }
which forms a sparse yet informative representation. This reduces computational overhead significantly while preserving essential motion cues for appearance–motion fusion.

3.2. Cross-Modal Feature Fusion

To effectively integrate appearance and motion information in space animal tracking, this study proposes a cross-modal feature fusion framework based on a heterogeneous graph network [37]. The framework employs a four-stage pipeline: feature extraction, graph construction, attention-based message passing, and deformable decoding. Each stage is specifically designed to progressively model complex inter- and intra-modality relationships.
The process initiates with multi-scale feature extraction via the PVTv2 [38] backbone, which processes each input frame through four hierarchical stages with downsampling ratios r { 4 , 8 , 16 , 32 } . At each stage s, pyramidal features M k a / m , s are derived for both appearance and motion modalities. These 3D feature maps are flattened and linearly projected into 2D query matrices as:
D k a / m , s = Flatten ( M k a / m , s ) W p s + p s
where W p s denotes learnable projection weights and  p s represents positional encodings. The latter are computed using sinusoidal functions across L = 8 multi-scale frequency bands to capture spatial context.
Based on the extracted features, this paper proposes a Multi-Modal Heterogeneous Graph Network (MHGN), denoted as G k = ( V k , E k ) . Here, V k is the node set defined as V k = D k a , D k m , T k 1 a , T k 1 m , where D k a and D k m correspond to the appearance and motion queries of the current frame, while T k 1 a and T k 1 m denote the trajectory memory from the previous frame (initialized to zero). The edge set E k = E t , E s , E h comprises three edge types: E t (temporal edges), connecting D k a / m and T k 1 a / m to model temporal consistency; E s (spatial edges), capturing spatial proximity via relative positional bias; and E h (heterogeneous edges), linking D k a and D k m to enable cross-modal interaction.
In the attention encoding stage, context-aware query representations are generated by processing this graph structure through three specialized attention mechanisms: temporal attention, spatial attention, and heterogeneous attention. The topological structure of the proposed heterogeneous graph network is illustrated in Figure 2. As shown in Figure 3 and Figure 4, the encoder models spatio-temporal dependencies and learns inter-modality mappings via modality-specific attention. Meanwhile, the feature fusion process described in Algorithm 1 outlines the key steps of the model selection pipeline.
First, modality-specific temporal attention updates node features via intra-modal self-attention:
Attn t ( Q , K , V ) = Softmax Q W Q t ( K W K t ) d V W V t
where W Q t , W K t , W V t denote modality-specific parameters.
Second, cross-modality spatial attention captures inter-modal spatial dependencies using relative position bias:
Attn s ( Q , K , V ) = Softmax Q W Q s ( K W K s ) d + ϕ i j V W V s
with ϕ i j = MLP ( Δ x i j , Δ y i j ) encoding relative spatial offsets.
Third, heterogeneous relation modeling aggregates information across distinct node types:
Attn h ( Q , K , V ) = m a , m Softmax Q W Q h ( K m W K h ) d V m W V h
This multi-level attention framework enables the model to learn enriched cross-modal representations while preserving modality-specific attributes. The refined features are subsequently fed into a deformable attention decoder for final query updating:
T ˜ k = m = 1 M W m q R m A m q · D ˜ k ( x q + Δ x m q , y q + Δ y m q )
where M = 8 denotes the number of attention heads, R m represents reference points, ( Δ x m q , Δ y m q ) are learnable offsets, and  A m q corresponds to attention weights.
Algorithm 1 Cross-modal feature fusion
Animals 15 02448 i001
The integrated fusion framework thereby achieves robust integration of appearance and motion cues, facilitating accurate and context-aware tracking of animals in complex spatial environments.

3.3. Unified Detection–Tracking Framework

The proposed framework seamlessly integrates object detection and tracking into a single end-to-end trainable architecture, enabling real-time, accurate, and robust animal tracking in spatial environments. Detection and tracking operate in parallel using shared multi-modal features: detection outputs provide initial object hypotheses, while tracking queries maintain temporal continuity across frames.
The detection head generates object predictions from fused multi-modal features D ˜ k a / m R N d × h . These features are first reshaped and processed via a series of convolutional layers to produce three key outputs: a center heatmap H k (with sigmoid activation) indicating object presence, bounding box coordinates B k = ( x , y , w , h ) , and refinement offsets O k (with Tanh activation) for precise localization. The detection branch is formulated as follows.
H k = σ ( Conv 3 × 3 ( ReLU ( Conv 1 × 1 ( D ˜ k a / m ) ) ) ) B k = Conv 1 × 1 ( ReLU ( Conv 3 × 3 ( D ˜ k a / m ) ) ) O k = Tan h ( Conv 1 × 1 ( ReLU ( Conv 3 × 3 ( D ˜ k a / m ) ) ) )
For tracking, a set of learned tracking queries T ˜ R N t × h encodes historical object states. These queries are processed through a displacement prediction network implemented as a multi-layer perceptron:
Δ k | k 1 = W 2 ReLU ( W 1 T ˜ + b 1 ) + b 2
where W 1 R h × 2 h and W 2 R 2 h × 2 denote learnable weights. Updated object positions are estimated as:
p k = p k 1 + Δ k | k 1 + ϵ k
with ϵ k N ( 0 , Σ ) modeling motion uncertainty.
To associate newly detected objects with existing tracks, a cross-frame affinity matrix A R N d × N t is computed to measure similarity between detection and tracking queries. The affinity score between detection U i and track V j is derived using a learned scoring function:
A i j = 1 1 + exp ( s i j ) , s i j = f θ ( U i , V j )
where f θ is defined as:
f θ ( U , V ) = W 2 ReLU ( W 1 [ U ; V ; U V ; | U V | ] + b 1 ) + b 2 .
This affinity matrix is subsequently used to solve the bipartite matching problem via the Hungarian algorithm:
M ^ = arg max M P ( N d , N t ) ( i , j ) M A i j .
To enhance detection reliability, a cross-modal confidence fusion strategy is introduced to leverage the complementary strengths of appearance and motion cues. The refined detection confidence is calculated as:
D e t ˜ k i = D e t k i 1 + exp ( γ c k , i h ) , γ = 2.0 ,
where c k , i h represents the harmonic mean of modality-specific confidences:
c k , i h = 2 c k a c k m c k a + c k m + ϵ
The entire framework is trained using a multi-task loss function that jointly optimizes detection, tracking, and association:
L = L c f + L b s + L r + L t d + L m a t c h
where L c f denotes the center-focused loss for penalizing heatmap prediction errors, L b s quantifies bounding box regression accuracy via SmoothL1 loss, L r ensures precise refinement of object centers, L t d aligns predicted and ground truth motion displacements using IoU, and  L m a t c h optimizes the data association matrix through binary cross-entropy.
This unified design enables the model to maintain consistent long-term object identities while adapting to dynamic and challenging visual conditions in spatial animal tracking scenarios.

4. Experiments

4.1. Experimental Data

This study utilizes the SpaceAnimal Dataset [39], the first publicly available dataset specifically developed for the non-contact behavioral analysis of multiple animal species in space environments. The dataset includes two model organisms with well-characterized cognitive behaviors: Drosophila and zebrafish, selected as experimental subjects to validate the proposed method. The Drosophila cohort was transported to the CSS via the Tianzhou-8 cargo spacecraft on 15 November 2024, successfully completing three generations under microgravity conditions. The zebrafish cohort was delivered to the CSS earlier by the Shenzhou-18 manned mission on 25 April 2024, achieving a breakthrough in on-orbit spawning.
High-resolution video data were collected throughout the life cycles of both species. Researchers selected representative time periods, performed detailed manual annotations, and constructed a multi-object tracking dataset. The Drosophila subset contains 20 video segments, totaling 2500 image frames and 27,500 individual instances, while the zebrafish subset comprises 8 video segments, with 1757 image frames and 7028 individual instances.
The dataset was partitioned into training, validation, and test subsets using a temporal split strategy to prevent data leakage and ensure realistic evaluation of model generalization. Specifically, the split ratio was 60% for training, 10% for validation, and 30% for testing, based on distinct experimental sessions. To avoid temporal overlap, all videos from the same session were assigned to the same subset. The training set was used for model learning, the validation set for hyperparameter tuning and early stopping, and the test set exclusively for final evaluation without involvement in the training process.

4.2. Metrics

This paper evaluates tracking performance using widely adopted MOT metrics, including multiple object tracking accuracy (MOTA) [40] and identification F1 score (IDF1) [41], which are defined as follows:
MOTA = 1 t ( FP t + FN t + IDSW t ) t GT t
IDF 1 = 2 IDTP 2 IDTP + IDFP + IDFN
where MOTA quantifies overall tracking performance by aggregating frame-wise errors—including false positives (FPs), false negatives (FNs), and identity switches (IDSWs)—normalized by the total number of ground truth objects (GTs). Higher MOTA scores indicate better alignment between predicted and true trajectories.
IDF1 evaluates the accuracy of identity preservation over time, computed using ID true positives (IDTPs), ID false positives (IDFPs), and ID false negatives (IDFNs). Values closer to 1 denote stronger consistency in individual object identification across frames.
Additionally, we report trajectory fragmentation (Frag), which counts the total number of trajectory interruptions during tracking. Lower Frag values indicate that the tracker maintains more continuous and stable trajectories, a critical attribute for long-term behavioral analysis.

4.3. Implementation Details

All experiments were conducted using PyTorch on an Ubuntu 18.04 OS. The machine was equipped with an Intel(R) Core(TM) i9 CPU @ 3.30 GHz, 128 GB of memory, and 2× GTX 3090 GPUs.
The model was trained with the following specifications (Table 1):

5. Results

5.1. Comparison with State-of-the-Art Methods

Table 2 presents comparative results on the SpaceAnimal Dataset, where our method is evaluated against six state-of-the-art tracking approaches: CenterTrack [42], TransCenter [43], TrackFormer [44], ByteTrack [45], MOTRv2 [46], and Hybrid-SORT [47]. The assessment spans two model species (Drosophila and zebrafish) using three metrics: MOTA, IDF1, and Frag (trajectory fragmentation).
In the Drosophila subset, our method achieves MOTA = 88.21%, IDF1 = 85.06%, and Frag = 42, outperforming all comparative methods. It surpasses ByteTrack—the second-highest performer in MOTA (75.21%)—by +13.0%, indicating substantial improvements in handling high-density, fast-moving targets under microgravity. TrackFormer exhibits the lowest MOTA (67.48%), likely due to its limited capability for motion modeling in dynamic environments.
For identity preservation, our method outperforms the second-ranked MOTRv2 (75.35%) by +9.71% in IDF1, demonstrating the efficacy of our cross-modal re-detection mechanism in maintaining identity consistency during occlusions and rapid movements. Hybrid-SORT and TrackFormer show notably lower IDF1 scores (66.23% and 66.25%, respectively), indicating deficiencies in long-term identity maintenance. Our method also achieves the lowest trajectory fragmentation (Frag = 42), significantly outperforming ByteTrack (91) and CenterTrack (98), which highlights its ability to sustain continuous, stable tracking of small, fast-moving targets.
Among comparative methods, ByteTrack performs relatively well in both MOTA and Frag but exhibits a substantially lower IDF1 score (76.50%) compared with our method, indicating limitations in temporal identity association. CenterTrack follows a similar pattern, with strong detection performance but weaker identity management.
In the zebrafish subset, our method again achieves superior performance with MOTA = 82.21%, IDF1 = 74.26%, and Frag = 36. It outperforms MOTRv2—the second-highest performer in MOTA (78.14%)—by +4.07%, demonstrating robustness in handling larger, more complex motion patterns typical of aquatic organisms. TransCenter exhibits the lowest MOTA (60.20%), likely due to its sensitivity to appearance degradation under low-contrast or motion-blurred conditions.
Our method also attains the highest IDF1 score (74.26%), surpassing MOTRv2 (64.24%) by +10.02%. This highlights its superiority in distinguishing visually similar individuals and maintaining identity across long sequences—a critical requirement for group behavior analysis. CenterTrack and ByteTrack show comparable IDF1 scores (60.14% and 62.95%, respectively), indicating moderate identity consistency but remaining inferior to our method.
For trajectory continuity, our method achieves the lowest Frag = 36, significantly outperforming MOTRv2 (78) and CenterTrack (85). This underscores its ability to generate smooth, uninterrupted trajectories—essential for capturing fine-grained behavioral patterns such as schooling and social interaction.
Among comparative methods, MOTRv2 performs best on zebrafish, particularly in MOTA and Frag, but its lower IDF1 score indicates limitations in identity preservation. CenterTrack and ByteTrack offer balanced performance but are outperformed by our method across all three metrics.
Across both species, several key trends emerge. ByteTrack and CenterTrack provide strong baseline performance in MOTA and Frag but struggle with identity preservation, especially in long-term tracking. MOTRv2 excels on zebrafish, likely due to its motion modeling capabilities, but underperforms on Drosophila, suggesting limited adaptability to smaller, faster-moving targets. TrackFormer and TransCenter show consistently lower performance across all metrics, indicating that their Transformer-based architectures may be ill-suited to the unique visual and motion characteristics of spatial animal tracking. Hybrid-SORT, which integrates detection with Kalman filtering, exhibits moderate performance but fails to match our method in identity consistency and trajectory smoothness.
Figure 5 and Figure 6 compare ground truth trajectories with the tracking results of the proposed method for Drosophila and zebrafish under microgravity. Despite complex motion patterns, predicted trajectories exhibit high consistency with true paths, confirming accurate tracking performance.
In summary, our method outperforms comparative approaches across all metrics and both species, demonstrating superior tracking accuracy, stronger identity consistency, and more stable trajectory estimation. These improvements are particularly pronounced in microgravity environments, where irregular motion patterns and suboptimal visual conditions are explicitly addressed by our architecture through a heterogeneous graph network and cross-modal identity re-detection mechanism.
We report the average inference speed, model size, and FLOPs of our framework on the SpaceAnimal Dataset using two NVIDIA RTX GPUs: an inference speed of 28.5 FPS (frames per second) for 800 × 800 resolution videos, a model size of 53.4 MB, and 18.7  GFLOPs per frame. These metrics confirm that our method achieves real-time performance with reasonable computational requirements, making it suitable for long-term behavioral monitoring in resource-constrained environments such as space stations.

5.2. Ablation Study

We conducted an ablation study to evaluate the contribution of each proposed component—motion decomposition (Motion), Multi-Modal Heterogeneous Graph Network (MHGN), and cross-modal re-detection (ReDet)—on the Drosophila and zebrafish subsets of the SpaceAnimal Dataset. The results are summarized in Table 3, where the baseline refers to the core tracking framework without any additional modules. Components are incrementally introduced, with performance evaluated using MOTA, IDF1, and the MT/ML metric (which quantifies tracking performance by comparing the count of consistently tracked targets [MT:   80 % lifespan coverage] against frequently lost targets [ML:   20 % lifespan coverage]).
The baseline achieves MOTA = 74.41% and IDF1 = 79.59% on Drosophila and MOTA = 74.12% and IDF1 = 60.14% on zebrafish. While detection performance is acceptable, identity consistency and robustness are limited—particularly under complex microgravity motion—as indicated by moderate MT/ML ratios (61/12 and 58/15) and frequent identity switches.
Adding motion decomposition improves both tracking accuracy and identity preservation. On Drosophila, MOTA increases by +7.22% to 81.63%, and IDF1 rises by +2.86% to 82.45%. On zebrafish, improvements are more pronounced: MOTA = 79.09% (+4.97%) and IDF1 = 67.79% (+7.65%). MT/ML ratios also improve to 73/8 and 65/11, reflecting fewer track losses and more complete trajectories. These results demonstrate that motion decomposition enhances motion estimation and tracking stability in dynamic environments.
Integrating MHGN further improves performance by fusing appearance and motion features. On Drosophila, MOTA = 86.45% (+4.82%) and IDF1 = 84.36% (+1.91%); on zebrafish, MOTA = 80.74% (+1.65%) and IDF1 = 71.22% (+3.43%). MT/ML ratios reach 82/5 and 72/8, indicating stronger association capability and fewer lost tracks. These findings confirm that MHGN enhances feature representation and robustness under occlusion and visual ambiguity.
Finally, introducing ReDet to maintain identity and recover lost tracks during occlusions or fast motion yields further gains: on Drosophila, MOTA = 88.21% (+1.76%) and IDF1 = 85.06% (+0.70%); on zebrafish, MOTA = 82.21% (+1.47%) and IDF1 = 74.26% (+3.04%). MT/ML ratios reach 86/4 and 78/6, demonstrating that ReDet effectively mitigates identity drift and track loss.
Across both species, each component contributes to performance improvement in a distinct, complementary manner. Motion decomposition enhances motion modeling and trajectory estimation, particularly in dynamic, unpredictable environments. MHGN strengthens feature fusion and association robustness by integrating multi-modal cues within a graph-based structure. ReDet ensures identity consistency and improves re-identification capability, especially during occlusions or rapid movements.
Together, these modules form a cohesive architecture that significantly outperforms the baseline and achieves state-of-the-art performance on the SpaceAnimal Dataset. The ablation results provide strong evidence that the design choices are well motivated and effective in addressing the unique challenges of animal tracking in spatial environments.

6. Discussion

6.1. Key Findings

This study presents a novel deep learning-based MOT framework tailored for tracking animals in extreme space environments, addressing the unique challenges posed by microgravity, high radiation, and weak magnetic fields. The experimental results on the SpaceAnimal Dataset—encompassing Drosophila and zebrafish—demonstrate that our method outperforms state-of-the-art MOT approaches across critical metrics, including MOTA, IDF1, and Frag.
The ablation study further confirms the contributions of individual components: motion decomposition enhances trajectory stability by accurately estimating speed and acceleration under microgravity; MHGN strengthens feature representation by leveraging inter-modal dependencies; and the cross-modal re-detection module mitigates identity switches during occlusions or rapid movements. Together, these components form a cohesive architecture that balances accuracy and robustness, addressing the limitations of existing methods in handling space-specific challenges.

6.2. Limitations

The proposed method, while effective in tracking Drosophila and zebrafish under space conditions, has limitations that restrict its broader applicability. Scalability to larger animal models is a key challenge, as the current design is optimized for small organisms with relatively simple locomotor patterns. Larger animals like rodents exhibit more complex behaviors, including diverse postures and interactions, which may strain the motion decoupling module and cross-modal fusion mechanisms, with non-rigid motion potentially being insufficiently modeled by the current local polynomial approximation-based motion decomposition. Additionally, the method relies on high-quality input, degrading under severe motion blur, low lighting, or occlusions, and assumes a static camera setup, struggling with viewpoint changes or dynamic backgrounds.
Hardware and computational constraints further limit deployment, as the method, despite achieving 28.5 FPS on 800 × 800 resolution, requires significant resources like dual GTX 3090 GPUs, which may not align with spacecraft power and weight limitations. The heterogeneous graph network’s computational cost, especially during attention encoding, hinders real-time use in resource-constrained scenarios, and prolonged occlusions or highly similar animal appearance and motion patterns can cause identity switches.

7. Conclusions

To address the challenges of multi-object tracking for spatial animals in microgravity environments, this paper proposes a novel MOT method. The framework integrates a motion decoupling module to extract motion components (e.g., velocity and acceleration), which—along with appearance features—serve as dual-input modalities. These features are encoded and fused via a heterogeneous graph network, which effectively integrates cross-modal spatiotemporal information to enhance feature representation. A cross-modal re-detection module is further introduced to maintain identity consistency and recover lost tracks during occlusions or fast movements.
Experimental results on the SpaceAnimal Dataset demonstrate that the proposed method achieves state-of-the-art performance, significantly outperforming existing approaches in key tracking metrics (including MOTA and IDF1). The ablation study confirms the effectiveness of each component in improving tracking accuracy, robustness, and identity preservation. For future work, we aim to generalize the framework to a broader range of multi-object tracking tasks, evaluating its adaptability and performance across diverse application scenarios.

Author Contributions

Conceptualization, Z.Z. and S.L.; methodology, Z.Z.; software, Z.Z. and Y.L.; validation, Z.Z. and Y.C.; formal analysis, K.L. and S.G.; investigation, Z.Z. and Y.C.; resources, Y.L.; data curation, Y.C. and K.L.; writing—original draft preparation, Z.Z.; writing—review and editing, S.L. and Y.L.; visualization, Z.Z.; supervision, S.L.; project administration, S.L.; funding acquisition, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Space Station Science Data Center (No.Y6140711WN) and the Microgravity Multidisciplinary Database at the National Basic Science Data Center (No.NBSDC-DB-17).

Institutional Review Board Statement

The study did not require ethical approval because it exclusively utilized publicly available video datasets of animal behavior collected in previous space missions. No new animal experiments were conducted as part of this research, and all data were obtained from open-access repositories with no involvement of live animals, interventions, or direct observations by the authors. Therefore, institutional ethical review was not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The SpaceAnimals datasets presented in this study can be accessed at the following link: https://doi.org/10.1038/s41597-025-05111-8.

Acknowledgments

The authors would like to thank the developers of MMTracking for providing a powerful and flexible framework for multi-object tracking research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rutter, L.; Barker, R.; Bezdan, D.; Cope, H.; Costes, S.V.; Degoricija, L.; Fisch, K.M.; Gabitto, M.I.; Gebre, S.; Giacomello, S.; et al. A New Era for Space Life Science: International Standards for Space Omics Processing. Patterns 2020, 1, 100148. [Google Scholar] [CrossRef]
  2. Clément, G. Introduction to Space Life Sciences. In Fundamentals of Space Medicine; Springer: New York, NY, USA, 2025; pp. 1–53. [Google Scholar]
  3. Beheshti, A.; Shirazi-Fard, Y.; Choi, S.; Berrios, D.; Gebre, S.G.; Galazka, J.M.; Costes, S.V. Exploring the Effects of Spaceflight on Mouse Physiology Using the Open Access NASA GeneLab Platform. J. Vis. Exp. 2019, 143, 58447. [Google Scholar] [CrossRef]
  4. Uchida, S.; Masukawa, M.; Kamigaichi, S. NASDA Aquatic Animal Experiment Facilities for Space Shuttle and ISS. Adv. Space Res. 2002, 30, 797–802. [Google Scholar] [CrossRef] [PubMed]
  5. Gu, Y. The China Space Station: A New Opportunity for Space Science. Natl. Sci. Rev. 2022, 9, nwab219. [Google Scholar] [CrossRef] [PubMed]
  6. Gao, M.; Zhao, G.; Gu, Y. Recent Progress in Space Science and Applications of China’s Space Station in 2020–2022. Chin. J. Space Sci. 2022, 42, 503–510. [Google Scholar] [CrossRef]
  7. Yang, J.; Liu, K.; Zhao, M.; Li, S. Video Process Detection for Space Electrostatic Suspension Material Experiment in China’s Space Station. Eng. Appl. Artif. Intell. 2024, 131, 107804. [Google Scholar] [CrossRef]
  8. Jiang, S.; Zhao, Y. China’s National Space Station: Opportunities, Challenges, and Solutions for International Cooperation. Space Policy 2021, 57, 101439. [Google Scholar] [CrossRef]
  9. Jia, C.; Zhang, X.; Sun, H.; Gu, P.; Yuan, Y.; Gao, W.; Zheng, W.; Yu, J.; Zhang, T.; Wang, G. Zebrafish Selection Strategy for the First Zebrafish Cultivation Experiment on the Chinese Space Station. Life Sci. Space Res. 2025, 46, 43–52. [Google Scholar] [CrossRef]
  10. Palma, O.; Plà-Aragonés, L.M.; Mac Cawley, A.; Albornoz, V.M. AI and Data Analytics in the Dairy Farms: A Scoping Review. Animals 2025, 15, 1291. [Google Scholar] [CrossRef]
  11. Hoffman, B.; Cusimano, M.; Baglione, V.; Canestrari, D.; Chevallier, D.; DeSantis, D.L.; Jeantet, L.; Ladds, M.A.; Maekawa, T.; Mata-Silva, V.; et al. A Benchmark for Computational Analysis of Animal Behavior, Using Animal-Borne Tags. Mov. Ecol. 2024, 12, 78. [Google Scholar] [CrossRef]
  12. Congdon, J.V.; Hosseini, M.; Gading, E.F.; Masousi, M.; Franke, M.; MacDonald, S.E. The Future of Artificial Intelligence in Monitoring Animal Identification, Health, and Behaviour. Animals 2022, 12, 1711. [Google Scholar] [CrossRef] [PubMed]
  13. Ohnishi, T. Life Science Experiments Performed in Space in the ISS/Kibo Facility and Future Research Plans. J. Radiat. Res. 2016, 57 (Suppl. S1), i41–i46. [Google Scholar] [CrossRef]
  14. Hong, X.; Zhang, Y.; Bailey, J.; Hoehn, R.; Stodieck, L.; Globus, R.K. Effects of Spaceflight Aboard the International Space Station on Mouse Estrous Cycle and Ovarian Gene Expression. NPJ Microgravity 2021, 7, 11. [Google Scholar] [CrossRef] [PubMed]
  15. Amosa, T.I.; Sebastian, P.; Izhar, L.I.; Ibrahim, O.; Ayinla, L.S.; Bahashwan, A.A.; Bala, A.; Samaila, Y. Multi-Camera Multi-Object Tracking: A Review of Current Trends and Future Advances. Neurocomputing 2023, 552, 126558. [Google Scholar] [CrossRef]
  16. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  17. Guo, Y.; Wu, Z.; You, B.; Chen, L.; Zhao, J.; Li, X. YOLO-SDD: An Effective Single-Class Detection Method for Dense Livestock Production. Animals 2025, 15, 1205. [Google Scholar] [CrossRef] [PubMed]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  19. Wang, Z.; Zheng, L.; Liu, Y.; Li, Y.; Wang, S. Towards Real-Time Multi-Object Tracking. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 107–122. [Google Scholar]
  20. Mayer, C.; Danelljan, M.; Paudel, D.P.; Van Gool, L. Learning Target Candidate Association to Keep Track of What Not to Track. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 13444–13454. [Google Scholar]
  21. Sun, P.; Cao, J.; Jiang, Y.; Zhang, R.; Xie, E.; Yuan, Z.; Wang, C.; Luo, P. Transtrack: Multiple Object Tracking with Transformer. arXiv 2020, arXiv:2012.15460. [Google Scholar]
  22. Zhang, Y.; Wang, C.; Wang, X.; Zeng, W.; Liu, W. Fairmot: On the Fairness of Detection and Re-Identification in Multiple Object Tracking. Int. J. Comput. Vis. 2021, 129, 3069–3087. [Google Scholar] [CrossRef]
  23. Zhou, X.; Zhuo, J.; Krähenbühl, P. Objects as Points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  24. Fazzari, E.; Romano, D.; Falchi, F.; Stefanini, C. Animal Behavior Analysis Methods Using Deep Learning: A Survey. Expert Syst. Appl. 2025, 289, 128330. [Google Scholar] [CrossRef]
  25. Lattal, K.A.; Fernandez, E.J. Grounding Applied Animal Behavior Practices in the Experimental Analysis of Behavior. J. Exp. Anal. Behav. 2022, 118, 186–207. [Google Scholar] [CrossRef] [PubMed]
  26. Lauer, J.; Mathis, A.; Bethge, V.; Lopes, M.; Ullrich, B.; Spether, C.; Mamidanna, P.; Yartsev, M.; Mathis, M.W. Multi-Animal Pose Estimation, Identification and Tracking with DeepLabCut. Nat. Methods 2022, 19, 496–504. [Google Scholar] [CrossRef] [PubMed]
  27. Walter, T.; Couzin, I.D. TRex, a Fast Multi-Animal Tracking System with Markerless Identification, and 2D Estimation of Posture and Visual Fields. eLife 2021, 10, 64000. [Google Scholar] [CrossRef]
  28. Dell, A.I.; Bender, J.A.; Branson, K.; Couzin, I.D.; de Polavieja, G.G.; Noldus, L.P.J.J.; Pérez-Escudero, A.; Perona, P.; Straw, A.D.; Wikelski, M. Automated Image-Based Tracking and Its Application in Ecology. Trends Ecol. Evol. 2014, 29, 417–428. [Google Scholar] [CrossRef]
  29. Ardekani, R.; Biyani, A.; Dalton, J.E.; Saltz, J.B.; Arbeitman, M.N.; Tower, J.; Nuzhdin, S.; Tavaré, S. Three-Dimensional Tracking and Behaviour Monitoring of Multiple Fruit Flies. J. R. Soc. Interface 2013, 10, 20120547. [Google Scholar] [CrossRef]
  30. Manoukis, N.C.; Butail, S.; Diallo, M.; Ribeiro, J.M.C.; Paley, D.A. Stereoscopic Video Analysis of Anopheles gambiae Behavior in the Field: Challenges and Opportunities. Acta Trop. 2014, 132, S80–S85. [Google Scholar] [CrossRef]
  31. Angarita-Jaimes, N.C.; Parker, J.E.A.; Abe, M.; Mashauri, F.; Martine, J.; Towers, C.E.; McCall, P.J.; Towers, D.P. A Novel Video-Tracking System to Quantify the Behaviour of Nocturnal Mosquitoes Attacking Human Hosts in the Field. J. R. Soc. Interface 2016, 13, 20150974. [Google Scholar] [CrossRef]
  32. Wu, H.S.; Zhao, Q.; Zou, D.; Chen, Y.Q. Automated 3D Trajectory Measuring of Large Numbers of Moving Particles. Opt. Express 2011, 19, 7646–7663. [Google Scholar] [CrossRef]
  33. Wang, S.H.; Su, H.F.; Cheng, X.E.; Liu, Y.; Quo, A.; Chen, Y.Q. Tracking the 3D Position and Orientation of Flying Swarms with Learned Kinematic Pattern Using LSTM Network. In Proceedings of the International Conference on Multimedia and Expo, Hong Kong, China, 10–14 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1225–1230. [Google Scholar]
  34. Yin, C.; Liu, X.; Zhang, X.; Wang, S.; Su, H. Long 3D-POT: A Long-Term 3D Drosophila-Tracking Method for Position and Orientation with Self-Attention Weighted Particle Filters. Applied Sci. 2024, 14, 6047. [Google Scholar] [CrossRef]
  35. Ngoc, D.N.; Richardson, T.S.; Watson, M.; Meier, K.; Kline, J.M.; Reid, S.; Maalouf, G.; Hine, D.; Mirmehdi, M.; Burghardt, T. WildLive: Near Real-time Visual Wildlife Tracking onboard UAVs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops: Computer Vision for Animal Behavior Tracking and Modeling (CV4Animals), Nashville, TN, USA, 11–15 June 2025. [Google Scholar]
  36. Tang, G.; Han, Y.; Sun, X.; Zhang, R.; Han, M.H.; Liu, Q.; Wei, P. Anti-Drift Pose Tracker (ADPT), a Transformer-Based Network for Robust Animal Pose Estimation Cross-Species. eLife 2025, 13, RP95709. [Google Scholar] [CrossRef]
  37. Hu, Z.; Dong, Y.; Wang, K.; Sun, Y. Heterogeneous Graph Transformer. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 2704–2710. [Google Scholar]
  38. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 568–578. [Google Scholar]
  39. Li, S.; Liu, K.; Wang, H.; Yang, R.; Li, X.; Sun, Y.; Zhong, R.; Wang, W.; Li, Y.; Sun, Y.; et al. Pose Estimation and Tracking Dataset for Multi-Animal Behavior Analysis on the China Space Station. Sci. Data 2025, 12, 766. [Google Scholar] [CrossRef]
  40. Bernardin, K.; Stiefelhagen, R. Evaluating Multiple Object Tracking Performance: The CLEAR MOT Metrics. EURASIP J. Image Video Process. 2008, 2008, 246309. [Google Scholar] [CrossRef]
  41. Ristani, E.; Solera, F.; Zou, R.; Cucchiara, R.; Tomasi, C. Performance Measures and a Data Set for Multi-Target, Multi-Camera Tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 17–35. [Google Scholar]
  42. Zhou, X.; Koltun, V.; Krähenbühl, P. Tracking Objects as Points. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 474–490. [Google Scholar]
  43. Xu, Y.; Ban, Y.; Delorme, G.; Gan, C.; Rus, D.; Alameda-Pineda, X. TransCenter: Transformers with Dense Representations for Multiple-Object Tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 7820–7835. [Google Scholar] [CrossRef] [PubMed]
  44. Meinhardt, T.; Kirillov, A.; Leal-Taixe, L.; Feichtenhofer, C. Trackformer: Multi-Object Tracking with Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8844–8854. [Google Scholar]
  45. Zhang, Y.; Sun, P.; Jiang, Y.; Yu, D.; Weng, F.; Yuan, Z.; Luo, P.; Liu, W.; Wang, X. Bytetrack: Multi-Object Tracking by Associating Every Detection Box. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 1–21. [Google Scholar]
  46. Zhang, Y.; Wang, T.; Zhang, X. Motrv2: Bootstrapping End-to-End Multi-Object Tracking by Pretrained Object Detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22056–22065. [Google Scholar]
  47. Yang, M.; Han, G.; Yan, B.; Zhang, W.; Qi, J.; Lu, H.; Wang, D. Hybrid-SORT: Weak Cues Matter for Online Multi-Object Tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 26–27 February 2024; Volume 38, No. 7. pp. 6504–6512. [Google Scholar]
Figure 1. Framework of the proposed MOT method for space animals.
Figure 1. Framework of the proposed MOT method for space animals.
Animals 15 02448 g001
Figure 2. The topological structure of the proposed heterogeneous graph network.
Figure 2. The topological structure of the proposed heterogeneous graph network.
Animals 15 02448 g002
Figure 3. The structure of the MHGN encoding.
Figure 3. The structure of the MHGN encoding.
Animals 15 02448 g003
Figure 4. The structure of the MHGN decoding.
Figure 4. The structure of the MHGN decoding.
Animals 15 02448 g004
Figure 5. Visualization comparison of tracking results of Drosophila.
Figure 5. Visualization comparison of tracking results of Drosophila.
Animals 15 02448 g005
Figure 6. Visualization comparison of tracking results of zebrafish.
Figure 6. Visualization comparison of tracking results of zebrafish.
Animals 15 02448 g006
Table 1. Training hyperparameters.
Table 1. Training hyperparameters.
ParameterValue
OptimizerAdamW
Initial Learning Rate2 × 10 4
Batch Size16
Training Epochs300
Learning Rate ScheduleCosine Annealing
Warmup Steps1000
Weight Decay0.05
Input Resolution512 × 512
AugmentationRandom Flip, Rotation (± 30 )
Table 2. Performance benchmarking of MOT algorithms for space animals.
Table 2. Performance benchmarking of MOT algorithms for space animals.
MethodDrosophilaZebrafish
MOTA ↑ IDF1 ↑ Frag ↓ MOTA ↑ IDF1 ↑ Frag ↓
CenterTrack [42]74.41%79.59%9874.12%60.14%85
TransCenter [43]72.58%74.12%11360.20%63.13%102
TrackFormer [44]67.48%66.25%9658.26%59.13%115
ByteTrack [45]75.21%76.50%9175.90%62.95%82
MOTRv2 [46]61.93%75.35%12578.14%64.24%78
Hybrid-SORT [47]70.62%66.23%9872.34%61.25%91
Ours88.21%85.06%4282.21%74.26%36
Table 3. Component ablation study.
Table 3. Component ablation study.
ConfigurationDrosophilaZebrafish
MOTA ↑ IDF1 ↑ MT ↑/ML ↓ MOTA ↑ IDF1 ↑ MT ↑/ML ↓
Baseline74.41%79.59%61/1274.12%60.14%58/15
+ Motion81.63%82.45%73/879.09%67.79%65/11
++ MHGN86.45%84.36%82/580.74%71.22%72/8
+++ ReDet88.21%85.06%86/482.21%74.26%78/6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Z.; Li, S.; Lv, Y.; Liu, K.; Cao, Y.; Guo, S. A Deep Learning Framework for Multi-Object Tracking in Space Animal Behavior Studies. Animals 2025, 15, 2448. https://doi.org/10.3390/ani15162448

AMA Style

Zhou Z, Li S, Lv Y, Liu K, Cao Y, Guo S. A Deep Learning Framework for Multi-Object Tracking in Space Animal Behavior Studies. Animals. 2025; 15(16):2448. https://doi.org/10.3390/ani15162448

Chicago/Turabian Style

Zhou, Zhuang, Shengyang Li, Yixuan Lv, Kang Liu, Yuxuan Cao, and Shicheng Guo. 2025. "A Deep Learning Framework for Multi-Object Tracking in Space Animal Behavior Studies" Animals 15, no. 16: 2448. https://doi.org/10.3390/ani15162448

APA Style

Zhou, Z., Li, S., Lv, Y., Liu, K., Cao, Y., & Guo, S. (2025). A Deep Learning Framework for Multi-Object Tracking in Space Animal Behavior Studies. Animals, 15(16), 2448. https://doi.org/10.3390/ani15162448

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop