Next Article in Journal
Long-Term Experiences of Basic Education in Laboratory Animal Science
Next Article in Special Issue
Collaborative Optimization of Model Pruning and Knowledge Distillation for Efficient and Lightweight Multi-Behavior Recognition in Piglets
Previous Article in Journal
Gastrointestinal Nematode Control in Lithuanian Sheep Farms: Insights from a Questionnaire Survey
Previous Article in Special Issue
Detecting Equine Gaits Through Rider-Worn Accelerometers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SDGTrack: A Multi-Target Tracking Method for Pigs in Multiple Farming Scenarios

College of Mechanical and Electrical Engineering, Fujian Agriculture and Forestry University, Fuzhou 350002, China
*
Author to whom correspondence should be addressed.
Animals 2025, 15(11), 1543; https://doi.org/10.3390/ani15111543
Submission received: 25 April 2025 / Revised: 18 May 2025 / Accepted: 21 May 2025 / Published: 24 May 2025

Simple Summary

This study presents a novel method for the effective detection and tracking of pigs in unknown environments and complex scenarios. Using the CSTrack model as the baseline, we enhance it by incorporating an environment-aware adaptive module and optimizing the target association strategy to address the model’s limited tracking ability in unknown scenes. Experimental results show that, compared to several advanced models, this method performs excellently in various complex environments, meeting the tracking requirements in unknown scenarios and providing robust technical support for the precise management of pigs.

Abstract

In pig farming, multi-object tracking (MOT) algorithms are effective tools for identifying individual pigs and monitoring their health, which enhances management efficiency and intelligence. However, due to the considerable variation in breeding environments across different pig farms, existing models often struggle to perform well in unfamiliar settings. To enhance the model’s generalization in diverse tracking scenarios, we have innovatively proposed the SDGTrack method. This method improves tracking performance across various farming environments by enhancing the model’s adaptability to different domains and integrating an optimized tracking strategy, significantly increasing the generalization of group pig tracking technology across different scenarios. To comprehensively evaluate the potential of the SDGTrack method, we constructed a multi-scenario dataset that includes both public and private data, spanning ten distinct pig farming environments. We only used a portion of the daytime scenes as the training set, while the remaining daytime and nighttime scenes were used as the validation set for evaluation. The experimental results demonstrate that SDGTrack achieved a MOTA score of 80.9%, an IDSW of 24, and an IDF1 score of 85.1% across various scenarios. Compared to the original CSTrack method, SDGTrack improved the MOTA and IDF1 scores by 16.7% and 33.3%, respectively, while significantly reducing the number of ID switches by 94.6%. These findings indicate that SDGTrack offers robust tracking capabilities in previously unseen farming environments, providing a strong technical foundation for monitoring pigs in different settings.

1. Introduction

With population and economic growth, the demand for pork continues to rise [1], presenting new opportunities and challenges for the swine industry. Rising labor costs and inefficiencies in pig farming highlight the need for intelligent farming practices, making continuous tracking and monitoring of pigs in complex environments increasingly important [2]. Although recent advances in computer vision have contributed to pig monitoring, most existing methods are validated only in controlled or idealized scenarios, failing to reflect the complexity of real-world farming conditions. In practical applications, environmental variations often reduce detection accuracy and cause frequent target loss, severely limiting the generalization and robustness of current models [3]. Therefore, it is essential to develop a target tracking model with strong cross-scenario adaptability to ensure reliable identification and stable performance under dynamic farming conditions.
Early methods for pig identification and positioning primarily relied on Radio Frequency Identification (RFID) technology, using electronic ear tags for automated tracking within a reader’s range [4]. This approach effectively managed individual pigs, particularly in tracking feeding and drinking behaviors. UHF-RFID technology was later applied to monitor feeder visits in growing-finishing pigs, underscoring its potential for animal behavior monitoring [5]. Subsequent advancements involved High-Frequency RFID (HF RFID) systems that used advanced multiplexers to connect multiple antennas to a single reader, enabling the precise monitoring of multiple pigs simultaneously [6,7]. Additionally, an RFID-based real-time alert system was developed to detect potential health issues by tracking individual pigs’ feeding and drinking behaviors, demonstrating high specificity in behavior identification [8]. However, the limitations of RFID technology have become increasingly apparent over time. High equipment and maintenance costs have become a heavy burden for long-term applications. Additionally, the technology may induce stress responses in pigs [2], complicating management and indirectly affecting data completeness and analytical accuracy.
With the development of deep learning, video-based multi-object tracking (MOT) methods like Tracking-by-Detection (TBD) and Joint Detection and Embedding (JDE) have become increasingly prominent [9]. TBD methods, which separate detection and association tasks, have been applied in pig farming to improve tracking accuracy and stability. Enhancements to two-stage trackers, like Faster R-CNN [10] and Mask R-CNN [11], have improved mask fitting and occlusion handling in pig tracking applications [12,13]. However, due to the limited speed of two-stage detectors, single-stage detectors have gained popularity for their efficiency. In complex environments, the optimized YOLOv5-Byte method integrates behavioral information and ID values to enhance detection accuracy and maintain continuous tracking, while effectively reducing ID association errors [14]. Similarly, using YOLOX-S and YOLOv5s as detectors has resulted in improved MOTA and IDF1 scores and a substantial reduction in ID switches [15]. For complex tracking scenarios, an MPC-YD method based on YOLOv5 was proposed, achieving high detection precision and reliable multi-object tracking performance [16].
Compared to two-stage tracking methods, the Joint Detection and Embedding (JDE) approach [17] integrates detection and re-identification tasks into a single framework. This integration eliminates information gaps between detection and tracking found in traditional methods and streamlines the workflow by reducing redundant computations. Consequently, JDE significantly enhances processing efficiency while maintaining high accuracy, making it well-suited for real-time multi-object tracking. Building on Joint Detection and Embedding algorithms, Guo et al. proposed a weighted association algorithm that improves tracking performance [17]. Their use of FairMOT combined with the weighted association algorithm achieved optimal tracking performance for pigs, with IDF1 and MOTA scores reaching 90.3% and 90.8%, respectively. Guo et al. compared three state-of-the-art deep learning-based MOT methods [18]: Joint Detection and Embedding (JDE), FairMOT, and YOLOv5s with DeepSORT [19]. By incorporating the improved weighted association method, they analyzed the continuous monitoring capabilities of each model for pig tracking. The results showed that the FairMOT method with the suggested weighted association achieved the highest IDF1 score, the fewest ID switches, and the fastest execution rate.
While existing research has made significant progress in tracking within specific scenarios, the performance of these models often drops sharply when applied to new or unknown environments due to their limited generalization capabilities. This is primarily because most models rely heavily on background information specific to the training scenarios, lacking adaptability to diverse environments. In pig farming, scene diversity manifests in various aspects, such as lighting conditions and scene changes [20]. These factors pose significant challenges for models handling cross-scene tasks, as changes in background information can drastically affect identification and tracking performance in different farming environments. To address the issue of poor model performance in unknown scenarios, we have innovatively proposed SDGTrack, which introduces adaptive modules and improves existing tracking strategies. This enables the model to distinguish and learn background information across different scenes. Experimental results show that SDGTrack not only performs well in known scenarios but also maintains high tracking accuracy in unknown environments.
The key contributions of this work are threefold:
(1)
An environment-aware adaptive module was proposed to enhance the model’s performance across various scenarios.
(2)
A target association strategy was designed to effectively reduce target mismatches and misassignments.
(3)
Comparative experiments with other leading trackers on MOT data across multiple scenarios validated the ability of SDGTrack to extend from a single environment to multiple scenarios.

2. Materials and Methods

2.1. Materials

2.1.1. Data Acquisition

The experimental data comprises two main components. The first part consists of data collected by our team using a Xiaomi CW400 high-definition surveillance camera (4 megapixels, resolution: 2560 × 1440; manufactured by Xiaomi Corporation, Beijing, China). Recordings were conducted on 5 and 17 November 2023, at a pig housing unit in the experimental garden of Fujian Agriculture and Forestry University. Each pen within the pig house measured 2.6 m × 1.6 m and was equipped with a feeding trough, exhaust fan, surveillance system, and bowl-type drinker. Two three-way crossbred pigs were housed per pen, one white and one dark-colored, each weighing approximately 50 kg, were housed per pen. The camera was mounted on the ceiling at a height of 2.2 m, providing a vertical top-down view that fully covered the pen area. The camera was configured for continuous recording throughout the day; when ambient light levels fell below 5 lux, infrared illumination was automatically enabled.
To enhance environmental diversity and evaluate the model’s generalization ability, we additionally incorporated several publicly available datasets and privately collected video sequences [21,22,23,24], these supplemental datasets were captured under various farming conditions, covering both daytime and nighttime illumination.

2.1.2. Dataset Construction

To support model training and evaluation, we constructed a multi-scenario pig tracking dataset by combining publicly available datasets and privately collected video resources. The dataset comprises ten distinct video sequences, including eight daytime scenes and two nighttime scenes, covering a variety of pig farming environments. The dataset is publicly available at https://github.com/hurry-baby/SDGTrack (accessed on 18 May 2025).
All ten videos were first preprocessed and standardized to a frame rate of 30 frames per second. From each video, we extracted a 1 min segment and subsequently annotated them following the Multiple Object Tracking (MOT) format [25]. The videos were then exported into frame-level image sequences, resulting in a total of 18,000 annotated images.
From the full dataset, we selected seven daytime sequences representing different farm settings as the training set, yielding 12,600 images reflecting diverse pig activity patterns under various daytime conditions. The remaining three sequences, which were not included in training, were designated as the test set. This test set includes one additional daytime scenario not seen during training (Daytime-Out Scene), and two nighttime scenarios: one representing the night version of an environment present in the training set (Night-In scene), and another from an entirely different pig house (Night-Out scene), used to evaluate the model’s generalization ability under both lighting and environmental changes. Specific information on the composition of the data is shown in Table 1:
The Night-In Scene tests the model’s performance under changing lighting conditions. Although the model has been trained in the same environment during the day, the nighttime scenes present significant changes in lighting and background information. The Night-Out scene evaluates the model’s generalization to a completely unfamiliar night environment. This test is more stringent than the first, as the model must not only adapt to lighting changes but also to a completely different farm environment. Finally, the daytime scenes from a different farm assess the model’s performance in unfamiliar daytime conditions. Although the lighting is similar to the training set, the differing farm layouts are likely to impact the model’s performance. As shown in Figure 1, the three test scenarios represent distinct environmental changes to evaluate the model’s tracking performance in unknown or changing environments.

2.2. Methods

2.2.1. Basic JDE and CSTrack Methods

The Joint Detection and Embedding (JDE) model [26] effectively integrates object detection and identity re-identification (re-ID) into a single network architecture. By utilizing joint learning, the JDE model not only accurately localizes objects but also associates identity information across consecutive frames, significantly improving tracking efficiency and reducing computational complexity. The baseline network of JDE is derived from YOLOv3’s [27] Darknet-53 and Feature Pyramid Network (FPN).The CSTrack model [28] builds on JDE by further optimizing both detection and re-ID tasks. It addresses the issue of “over-competition” between these tasks, where competition in representation learning can cause confusion and negatively impact overall performance. Architecturally, CSTrack enhances JDE by incorporating a feature decoupling module, creating separate feature maps tailored to each task, allowing detection and re-ID to learn independently and improve performance.
For detection, CSTrack replaces the original YOLOv3 framework with the faster and more advanced YOLOv5. Additionally, it introduces a scale-aware attention network that leverages spatial and channel attention to capture appearance information across different scales, optimizing re-ID feature representation and further enhancing re-ID performance.

2.2.2. SDGTrack Tracking Model

To address the limitations of traditional models in tracking pigs across different farm environments, this study proposes SDGTrack, a multi-object tracking solution that employs single-domain adaptive generalization. The framework is built upon CSTrack [28], as illustrated in Figure 2, where (a) shows the feature extraction structure, including the Backbone and Neck components, and (b) presents the prediction module proposed in this study. To reduce performance discrepancies across environments, we first introduce an enhanced domain-aware attention (DAA) module. This module learns feature information from multiple environments, enabling the model to selectively emphasize relevant background features and suppress less useful ones, thereby achieving adaptive representation for unseen domains.
Additionally, acknowledging the reduced reliability of re-identification (Re-ID) features in varying validation environments, we proposed the Re-Byte approach, building upon ByteTrack [29]. By splitting the matching task into high- and low-confidence categories, the model enhances its focus on low-confidence detections while efficiently completing the association process through a combination of IoU (Intersection over Union) distance and identity feature (ID feature) distance.
SDGTrack learns the three tasks of classification, bounding box regression, and appearance features in parallel, like the JDE [26] model. The total loss function is defined as Equation (1).
L t o t a l = i M j = α , β , γ 1 2 ( 1 e s j i L j i + s j i )
where the loss function consists of three components: the classification loss L α , the bounding box regression loss L β , and the embedded learning loss L γ . s j i stands for task-specific-independent uncertainty and is a tunable network parameter. M is the number of predictor heads.

2.2.3. Domain-Aware Attention Module

To enable the model to adaptively learn shared feature information across different scenarios, we introduce an improved domain-aware attention module. This module achieves this by adaptively weighting multiple Effective Squeeze-and-Excitation (ESE) blocks [30]. The architecture is illustrated in Figure 3. The domain-aware attention module first performs global pooling on the input feature map to aggregate spatial information and generate a global feature vector. Then, global pooling and fully connected operations are applied, followed by another fully connected layer and a softmax function to produce domain-sensitive weights. These weights are used to adaptively weight the domain-related features generated by the three ESE modules, as shown in the following equation:
S D D A = softmax ( W C F a ν g ( F i ) )
where W C is the fully connected weight matrix, F a ν g is the global average pooling operation, and the softmax function ensures that the sum of the elements in the generated weight vector S D D A is 1, so that it can be used for weighting and combining. The generated domain-sensitive weight vector S D D A is used to weight the output S C produced by the splicing of the three ESE modules to form a domain-adaptive response vector S D C , which is computed as follows:
S D C = S D D A S C
Next, the domain adaptive response vector S D C is used to adaptively weight the channels of the original input feature map F i , and the model’s perception of the environment is realized by converting S D D A to the scaling factor of the input features of F i through the sigmoid function. The calculation formula is as follows:
f i = F s c a l e ( F i , σ ( S D C ) )
where F s c a l e ( ) is a multiplication operation performed on the channels of the feature map and σ is a sigmoid function.
This approach, unlike the hard attention mechanism, allows cross-domain information sharing and improves the effectiveness of feature representation. With a domain-aware attention network, the model can better adapt to different environments and achieve robustness and accuracy for multi-target tracking tasks.

2.2.4. Re-Byte

To improve the model’s generalization ability in dynamic scenarios, we optimized the matching strategy in the Multiple Object Tracking (MOT) algorithm. Traditional tracking methods often rely on high-confidence detection boxes. However, in changing or unknown environments, detection accuracy may vary significantly, resulting in poor target matching when accuracy is affected by scene changes. Inspired by ByteTrack, we proposed the Re-Byte method. By dividing detection boxes into high and low thresholds, we enrich the target matching information, reducing the negative impact of scene changes, lighting variations, and occlusions on subsequent tracking tasks. This enhances the model’s generalization across complex conditions. Figure 4 illustrates the improved Re-Byte tracking process, which integrates multiple factors for target matching, significantly enhancing the model’s adaptability and robustness in challenging environments.
In the initial matching stage, we first screened all candidate detection boxes and categorized them into high- and low-threshold groups. The primary focus of this stage is to match high-threshold detection boxes with the currently tracked targets. We used a Kalman filter [31] to predict the potential positions and trajectories of all tracked targets in the current frame. The IoU distance [32] between the detection boxes and the predicted trajectories was then calculated. Finally, the Hungarian algorithm [33] was applied to achieve optimal matching between the detection boxes and the predicted trajectories. The IoU calculation formula is shown in Equation (5):
I o U = A B A B
where A denotes the detected target frame, B denotes the predicted frame obtained from the tracked target using the Kalman filtering method, and the value domain of I o U is [0, 1].
Unlike ByteTrack, we introduce an additional matching step in the tracking process to enhance performance. In this second matching phase, we re-associate unmatched detection boxes with predicted tracks. This association is based on the Euclidean distance [34] to evaluate the similarity between them. The goal is to address cases where high-confidence targets are not well-associated due to low overlap between bounding boxes caused by occlusion or other factors. The formula for calculating the Euclidean distance is shown in Equation (6).
d track = ( x pred x det ) 2 + ( y pred y det ) 2
where ( x det , y det ) is the center point coordinate of the detection frame, and ( x pred , y pred ) is the center point coordinate of the predicted trajectory.
For tracks that remain unmatched after the first two steps, we apply a third association using low-threshold detection frames. In this step, we combine the IOU distance and embedded feature distance through weighted fusion, with the latter calculated using cosine similarity. By considering both feature distances together, the model effectively leverages the inherent characteristics of targets, enhancing its performance in multi-object tracking tasks across various scenarios.

2.2.5. Evaluation Metrics

To comprehensively analyze the tracking accuracy of our model across different scenarios, we utilized the metrics derived from the MOT challenge based on pedestrian datasets [25]. Additionally, we incorporated the evaluation metrics used in the JDE [26] and FairMOT [35] methods to assess our proposed approach alongside other methods. The key metrics include MOTA, IDF1, MOTP, HOTA, MT, ML, IDS, FP, FN, and FPS. Among these, Multi-Object Tracking Accuracy (MOTA) is considered the most critical metric in multi-object tracking, as it evaluates the overall accuracy of a multi-object tracking algorithm. The calculation formula for MOTA is provided below.
M O T A = 1 F N + F P + I D S W G T
where G T represents the number of ground truth objects. The maximum value of M O T A is 1, while its minimum value can reach negative infinity. F N and F P denote the numbers of missed detections and false positives, respectively. I D S W counts the instances of ID switches, indicating how often the tracker incorrectly assigns or changes object IDs. The smaller the I D S W , the better the tracker maintains object identity. I D F 1 measures the tracker’s consistency in preserving object identity by combining ID precision and ID recall. The formula for calculating I D F 1 is as follows:
I D F 1 = 2 I D T P 2 I D T P + I D F P + I D F N
where I D T P represents the number of correctly matched objects, I D F P denotes the number of incorrectly matched objects, and I D F N refers to the number of missed objects. Multiple Object Tracking Precision ( M O T P ) is a metric used to measure the positional error in tracking, and its expression is as follows:
M O T P = t , i d t . i t c t
where c t represents the number of detection boxes that successfully match with ground truth in frame t , while d t . i measures the distance between matched pairs. Higher Order Tracking Accuracy (HOTA) is a unified metric that jointly evaluates detection and association performance across multiple IoU thresholds. It is computed as follows:
H O T A = 1 A α A D e t α A s s A α
where A = { 0.05 , 0.10 , ,   0.95 } denotes a set of fixed IoU thresholds, and Det α , AssA α denote the detection and association accuracy at threshold α , respectively.
All evaluation metrics used in this study are summarized in Table 2. The goal is to provide a comprehensive and detailed assessment of the model’s tracking performance in complex scenarios. The table lists various terms used to evaluate tracking accuracy and clearly indicates the ideal trends for each metric. An upward arrow next to a metric name signifies that higher values correspond to better model performance, while a downward arrow indicates that lower values are preferable.

3. Results and Analysis

3.1. Experimental Platforms

We use an NVIDIA GeForce RTX 4090 chip with 24 GB of RAM (NVIDIA, Santa Clara, CA, USA) as the graphics card for core computation. The CPU model is Intel Core i7-13700X, 3.40 GHz. The version of the CUDA compiler is 11.8, the version of Python is 3.8.18, and the version of PyTorch used in the work is 2.1.1. All experiments were performed on this device.

3.2. Comparative Experiments with Different MOT Algorithms

In order to evaluate the effectiveness of our proposed SDGTrack approach, we compared it to several state-of-the-art multi-target tracking benchmark models. To ensure the fairness and consistency of the comparison, we matched all the model configurations involved in the comparison to those in the CSTrack model, which mainly include the model size and the specific parameters of the data enhancement strategy. The data are all in our home-grown dataset, as shown in Figure 1.
The comparison results of each model are shown in Table 3, and our proposed SDGTrack model demonstrates outstanding performance across the majority of evaluation metrics. In particular, it achieves a MOTA of 80.9%, indicating strong overall tracking accuracy by minimizing false positives, false negatives, and identity switches. The model also attains an IDF1 score of 85.1%, reflecting its excellent ability to maintain consistent object identities throughout the tracking sequence. Furthermore, SDGTrack significantly reduces identity switches, with a total of only 24, representing a 94.6% reduction compared to the baseline. This reduction highlights the model’s robustness in preserving target identities across frames. In addition, SDGTrack achieves a HOTA score of 83.0%, which demonstrates its balanced performance in both spatial detection accuracy and temporal identity association. Although FairMOT [35] achieved the highest frame rate at 40.5 FPS, its MOTA and IDF1 scores were considerably lower (55.6% and 54.6%, respectively), indicating weaker tracking stability and accuracy. In comparison, our proposed SDGTrack achieved a competitive speed of 27.5 FPS while delivering significantly better accuracy. Overall, SDGTrack consistently outperformed other models across all evaluation scenarios.
In addition to evaluating the overall performance of the models, we also analyzed the metrics of the top four models in three different scenarios to understand their precision advantages and characteristics. As shown in Table 4, in the nighttime scenario within the daytime training dataset, SDGTrack achieved a MODA of 78.1%, an IDF1 of 81.9%, and a HOTA of 80.0%. In comparison, the best-performing baseline model, FairMOT, obtained a MODA of 68.8%, an IDF1 of 56.8%, and a HOTA of 62.5%, with our model surpassing it by 9.3%, 25.1%, and 17.5%, respectively. In the nighttime scenario without the daytime training dataset, SDGTrack achieved a MODA of 69.6%, an IDF1 of 82.6%, and a HOTA of 75.8%. In contrast, the best-performing baseline model, CSTrack, achieved a MODA of 52.8%, an IDF1 of 44.4%, and a HOTA of 48.4%, with our model surpassing it by 16.8%, 38.2%, and 27.4%. In the daytime scenario without the daytime training dataset, SDGTrack recorded a MODA of 95%, an IDF1 of 90.8%, and a HOTA of 92.9%. In comparison, the best-performing baseline model, CSTrack, achieved a MODA of 91.1%, an IDF1 of 69.4%, and a HOTA of 79.5%, with our model outperforming by 3.9%, 21.4%, and 13.4%.
Through a comparative analysis of the test results across the three scenarios, SDGTrack demonstrated outstanding performance not only in known environments but also maintained strong performance despite environmental changes such as transitions between daytime and nighttime. Additionally, in completely unknown scenarios, SDGTrack showed significant improvements in metrics such as HOTA, MOTA, IDF1, and ID switches compared to other models, indicating superior cross-scenario generalization. This highlights its effectiveness and robustness in multi-object tracking tasks.

3.3. Ablation Experiment

To further validate the effectiveness of SDGTrack in complex and dynamic environments, we conducted ablation experiments. These experiments were based on our custom dataset, with the test set covering three distinct scenarios, as illustrated in Figure 1. By comparing performance across these scenarios, we aimed to better analyze the model’s tracking capabilities under varying environmental conditions. We compared the results of our model with the baseline model CSTrack across the three scenarios, representing the most significant environmental variations. The results showed that SDGTrack achieved the best performance across all scenarios, with detailed metrics provided in Table 5:
As shown in Table 5, it is evident that the application of the domain-aware attention (DAA) network module led to MOTA improvements of 10.9%, 14.8%, and 2.1% in the three scenarios, resulting in an overall increase of 9.27%. For the IDF1 metric, there was an overall improvement of 26.8% across the three scenarios. Regarding HOTA, the DAA module also brought notable improvements of 9.7%, 11.7%, and 8.0% in the corresponding scenarios, with an overall enhancement of 19.9%. These results demonstrate that the DAA module contributes to improved tracking accuracy and target identity recognition and retention. Additionally, the improvements to the Re-Byte matching scheme resulted in substantial gains across all three scenarios. Overall, incorporating Re-Byte led to a 10.0-point increase in MOTA, a 30.9-point increase in IDF1, a 20.7-point increase in HOTA, and a reduction in IDs from 447 to 54, significantly reducing confusion and errors during tracking and enhancing the continuity and consistency of tracking trajectories.
Comparative analysis indicates that SDGTrack’s adaptive mechanisms and optimized tracking strategies provide enhanced robustness in handling complex environmental changes. Specifically, the model dynamically adjusts its detection and tracking strategies based on different scenarios, effectively reducing background interference. By integrating ID embedding features into the matching process, it improves the precision of pig identification and tracking. To visually demonstrate the improvement in ID embedding discrimination between our method and the baseline model, we conducted a visualization analysis. As shown in Figure 5, SDGTrack performed well in both the differentiation of ID embedding features and the feature association capabilities obtained through cosine similarity. Specifically, Figure 5c,d illustrate the pairwise relationships between targets, where the strong association of each target with itself is represented by a prominent dark red diagonal line in the visualization, while areas outside the diagonal appear in clear dark blue. The results indicate that incorporating embedding features into the matching process is crucial for enhancing multi-scale target detection and handling occlusions.
The experimental results demonstrate that both modules improve the model’s feature extraction and re-identification performance in unknown environments. Specifically, leveraging the DAA module enhances adaptability across different environments, effectively increasing MOTA accuracy. The Re-Byte scheme, which focuses on low-confidence targets, significantly improves the IDF1 score and reduces the number of ID switches. Our proposed SDGTrack model achieved a 16.6-point gain in MOTA, a 33.2-point gain in IDF1, a 25.2-point gain in HOTA, and reduced IDs from 447 to 24. These results indicate that SDGTrack significantly enhances tracking robustness and accuracy in new environments compared to the original model.

4. Discussion

Traditional tracking models perform well in specific scenarios but often suffer a sharp performance drop in unfamiliar environments. This is largely because they rely heavily on background information from specific scenarios during design and training, lacking adaptability to different environments. In pig farms, scene diversity manifests in various aspects, such as lighting conditions and environmental changes. These factors present substantial challenges in handling cross-scene tasks, particularly in different farming environments where changes in background information can significantly affect recognition and tracking. Therefore, the ability to distinguish between different backgrounds is crucial for model generalization. Our SDGTrack method addresses this issue by introducing adaptive modules and optimizing existing tracking strategies.
We compared the tracking performance of our model with two other representative tracking solutions, including a two-stage tracking method based on the YOLO model with different tracking strategies [36,37,38,39] and the fastest anchor-free FairMOT scheme. The results, shown in Figure 6, reveal that SDGTrack not only achieved the best tracking performance across all three scenarios but also recorded the fewest ID switches.
In general, all models performed best under well-lit daytime conditions, while their performance significantly declined in the two nighttime scenarios—particularly in the Night-Out scene. This trend indicates that illumination plays a critical role in detection accuracy and identity preservation. Poor lighting at night reduces target visibility, decreases detection confidence, and increases the difficulty of feature extraction, thereby negatively affecting overall tracking performance.
Beyond lighting conditions, background complexity and pig density also had a noticeable impact on model performance. In the Daytime-Out scenario, despite similar lighting to the training data, differences in pen layout and spatial structure made background modeling more difficult, leading to degraded detection accuracy and increased identity confusion. In the Night-Out scenario, where pig density was the highest, frequent occlusions occurred, compounded by complex background textures and variable lighting, resulting in more false detections and degraded identity matching.
In contrast, SDGTrack demonstrated superior adaptability under these challenging conditions, owing to its enhanced domain-aware attention mechanism and more robust identity association strategy. These findings further confirm that designing a highly robust multi-object tracking model requires comprehensive consideration of key environmental factors, including lighting conditions, background complexity, and object density.
To provide a more intuitive demonstration of the superior performance of the SDGTrack model, we present its tracking results across three different environments, as shown in Figure 7. The results clearly indicate that SDGTrack consistently maintains strong tracking performance and re-identification capabilities across various scenarios. Even in the presence of occlusions and other complex situations, the model demonstrates high accuracy. However, it is worth noting that in cases of extreme occlusion, there is still room for further optimization and improvement.
Although the SDGTrack model has made significant progress in diverse scenarios, several issues remain that need to be addressed. Firstly, in extremely complex or dynamically changing environments, the model may still experience a decline in tracking accuracy, particularly when handling large-scale occlusions or fast-moving targets. Secondly, the current model’s training and inference speed remain a challenge, especially for real-time applications. Enhancing computational efficiency and reducing latency are critical areas for future optimization. Additionally, SDGTrack currently focuses primarily on individual tracking within group-housed pig environments and has not yet addressed more complex behavior patterns or tracking tasks involving different species. The model still faces limitations in capturing more intricate social behaviors and environmental adaptability in dynamic settings. Future improvements will aim to enhance the model’s robustness in handling occlusion and motion dynamics, incorporate more advanced behavior modeling strategies, and improve real-time processing capabilities to support broader tracking applications.
It is worth emphasizing that SDGTrack not only achieves technical improvements but also demonstrates strong practical potential in handling complex and variable farming environments. In real-world pig farming scenarios, environmental factors such as lighting fluctuations, diverse pen structures, and changes in animal density can significantly impact the stability of traditional detection and tracking systems. With its robust adaptability to occlusions, lighting variations, and complex backgrounds, SDGTrack enables consistent and accurate tracking of individual pigs even under dynamically changing conditions. This capability supports more reliable health monitoring, behavioral analysis, and anomaly detection, especially in farms where environmental conditions vary frequently.
Moreover, the domain-adaptive architecture of SDGTrack provides a solid foundation for scaling to larger or multi-species farming systems. With moderate retraining or data adaptation, the framework can be extended to tracking tasks involving other livestock such as cattle or sheep. Notably, compared to sensor-based monitoring methods, SDGTrack can run efficiently on mid-range GPUs without relying on high-performance computing platforms. The system requires only standard RGB video input and does not depend on expensive sensors or dedicated hardware, significantly reducing deployment costs and minimizing direct interference with animals. These advantages offer strong economic feasibility and practical applicability, facilitating the advancement of precision livestock farming and improving both productivity and animal welfare.

5. Conclusions

To address the poor performance of existing tracking models in complex and dynamic farming environments, this paper proposes a novel multi-object tracking model, SDGTrack, designed to enhance generalization and tracking stability across diverse scenarios. First, the method incorporates domain-aware attention (DAA) strategies to strengthen the model’s adaptive learning capabilities for environmental information, thereby improving its generalization to unseen target domains. Additionally, by integrating our Re-Byte scheme, which combines IOU distance and ID feature distance for association, we not only optimize tracking accuracy but also significantly reduce the number of ID switches, allowing the tracking box to more precisely follow the target object. The improved model shows a marked enhancement in its ability to generalize from a single source domain to other unseen target domains. Experimental results confirm that the SDGTrack method demonstrates outstanding tracking performance across diverse farming environments, whether in bright daytime or dim nighttime scenarios. The method achieves significant improvements in key evaluation metrics, such as Multi-Object Tracking Accuracy (MOTA) and Identity F1 Score (IDF1), clearly proving its strong cross-domain generalization capabilities. Ultimately, SDGTrack provides a reliable and scalable technical foundation for real-world deployment in livestock and poultry tracking tasks under variable and challenging environmental conditions, contributing to the advancement of intelligent animal farming systems.

Author Contributions

Conceptualization, T.L. and J.H.; methodology, T.L.; software, D.J.; validation, J.Z.; formal analysis, D.J. and T.L.; investigation, D.Z.; resources, J.H.; data curation, J.Z.; writing—original draft preparation, T.L.; writing—review and editing, D.Z.; visualization, D.J.; supervision, D.J.; project administration, J.Z.; funding acquisition, J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Fujian Province Agricultural Key Core Technology Research Project (KMY24405XC).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Soare, E.; Chiurciu, I.-A. Study on the Pork Market Worldwide. Comput. Electron. Agric. 2017, 2017, 321–326. [Google Scholar]
  2. Tzanidakis, C.; Simitzis, P.; Arvanitis, K.; Panagakis, P. An Overview of the Current Trends in Precision Pig Farming Technologies. Livest. Sci. 2021, 249, 104530. [Google Scholar] [CrossRef]
  3. Tangirala, B.; Bhandari, I.; Laszlo, D.; Gupta, D.K.; Thomas, R.M.; Arya, D. Livestock Monitoring with Transformer. arXiv 2021, arXiv:2111.00801. [Google Scholar] [CrossRef]
  4. Floyd, R.E. RFID in Animal-Tracking Applications. IEEE Potentials 2015, 34, 32–33. [Google Scholar] [CrossRef]
  5. Adrion, F.; Kapun, A.; Eckert, F.; Holland, E.-M.; Staiger, M.; Götz, S.; Gallmann, E. Monitoring Trough Visits of Growing-Finishing Pigs with UHF-RFID. Comput. Electron. Agric. 2018, 144, 144–153. [Google Scholar] [CrossRef]
  6. Maselyne, J.; Adriaens, I.; Huybrechts, T.; De Ketelaere, B.; Millet, S.; Vangeyte, J.; Van Nuffel, A.; Saeys, W. Measuring the Drinking Behaviour of Individual Pigs Housed in Group Using Radio Frequency Identification (RFID). Animal 2016, 10, 1557–1566. [Google Scholar] [CrossRef]
  7. Maselyne, J.; Saeys, W.; De Ketelaere, B.; Mertens, K.; Vangeyte, J.; Hessel, E.F.; Millet, S.; Van Nuffel, A. Validation of a High Frequency Radio Frequency Identification (HF RFID) System for Registering Feeding Patterns of Growing-Finishing Pigs. Comput. Electron. Agric. 2014, 102, 10–18. [Google Scholar] [CrossRef]
  8. De Bruijn, B.G.C.; De Mol, R.M.; Hogewerf, P.H.; Van Der Fels, J.B. A Correlated-Variables Model for Monitoring Individual Growing-Finishing Pig’s Behavior by RFID Registrations. Smart Agric. Technol. 2023, 4, 100189. [Google Scholar] [CrossRef]
  9. Xu, L.; Huang, Y. Rethinking Joint Detection and Embedding for Multiobject Tracking in Multiscenario. IEEE Trans. Ind. Inf. 2024, 20, 8079–8088. [Google Scholar] [CrossRef]
  10. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  11. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-Cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  12. He, H.; Qiao, Y.; Li, X.; Chen, C.; Zhang, X. Optimization on Multi-Object Tracking and Segmentation in Pigs’ Weight Measurement. Comput. Electron. Agric. 2021, 186, 106190. [Google Scholar] [CrossRef]
  13. Sun, L.; Zou, Y.; Li, Y.; Cai, Z.; Li, Y.; Luo, B.; Liu, Y.; Li, Y. Multi Target Pigs Tracking Loss Correction Algorithm Based on Faster R-CNN. Int. J. Agric. Biol. Eng. 2018, 11, 192–197. [Google Scholar] [CrossRef]
  14. Tu, S.; Cai, Y.; Liang, Y.; Lei, H.; Huang, Y.; Liu, H.; Xiao, D. Tracking and Monitoring of Individual Pig Behavior Based on YOLOv5-Byte. Comput. Electron. Agric. 2024, 221, 108997. [Google Scholar] [CrossRef]
  15. Tu, S.; Zeng, Q.; Liang, Y.; Liu, X.; Huang, L.; Weng, S.; Huang, Q. Automated Behavior Recognition and Tracking of Group-Housed Pigs with an Improved DeepSORT Method. Agriculture 2022, 12, 1907. [Google Scholar] [CrossRef]
  16. Huang, Y.; Xiao, D.; Liu, J.; Tan, Z.; Liu, K.; Chen, M. An Improved Pig Counting Algorithm Based on YOLOv5 and DeepSORT Model. Sensors 2023, 23, 6309. [Google Scholar] [CrossRef]
  17. Guo, Q.; Sun, Y.; Min, L.; van Putten, A.; Knol, E.F.; Visser, B.; Rodenburg, T.; Bolhuis, L.; Bijma, P.; de With, P.H.N. Video-Based Detection and Tracking with Improved Re-Identification Association for Pigs and Laying Hens in Farms. In Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Virtual, 6–8 February 2022; SciTePress: Setúbal, Portugal, 2022; pp. 69–78. [Google Scholar]
  18. Guo, Q.; Sun, Y.; Orsini, C.; Bolhuis, J.E.; De Vlieg, J.; Bijma, P.; De With, P.H.N. Enhanced Camera-Based Individual Pig Detection and Tracking for Smart Pig Farms. Comput. Electron. Agric. 2023, 211, 108009. [Google Scholar] [CrossRef]
  19. Wojke, N.; Bewley, A.; Paulus, D. Simple Online and Realtime Tracking with a Deep Association Metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
  20. Zhang, L.; Gray, H.; Ye, X.; Collins, L.; Allinson, N. Automatic Individual Pig Detection and Tracking in Pig Farms. Sensors 2019, 19, 1188. [Google Scholar] [CrossRef]
  21. Gao, Y.; Yan, K.; Dai, B.; Sun, H.; Yin, Y.; Liu, R.; Shen, W. Recognition of Aggressive Behavior of Group-Housed Pigs Based on CNN-GRU Hybrid Model with Spatio-Temporal Attention Mechanism. Comput. Electron. Agric. 2023, 205, 107606. [Google Scholar] [CrossRef]
  22. Shirke, A.; Saifuddin, A.; Luthra, A.; Li, J.; Williams, T.; Hu, X.; Kotnana, A.; Kocabalkanli, O.; Ahuja, N.; Green-Miller, A.; et al. Tracking Grow-Finish Pigs Across Large Pens Using Multiple Cameras. arXiv 2021, arXiv:2111.10971. [Google Scholar]
  23. T. Psota, E.; Schmidt, T.; Mote, B.; C. Pérez, L. Long-Term Tracking of Group-Housed Livestock Using Keypoint Detection and MAP Estimation for Individual Animal Identification. Sensors 2020, 20, 3670. [Google Scholar] [CrossRef]
  24. Bergamini, L.; Pini, S.; Simoni, A.; Vezzani, R.; Calderara, S.; D’Eath, R.; Fisher, R. Extracting Accurate Long-Term Behavior Changes from a Large Pig Dataset. In Proceedings of the 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Virtual, 8–10 February 2021; SciTePress: Setúbal, Portugal, 2021; pp. 524–533. [Google Scholar]
  25. Milan, A.; Leal-Taixe, L.; Reid, I.; Roth, S.; Schindler, K. MOT16: A Benchmark for Multi-Object Tracking. arXiv 2016, arXiv:1603.00831. [Google Scholar] [CrossRef]
  26. Wang, Z.; Zheng, L.; Liu, Y.; Li, Y.; Wang, S. Towards Real-Time Multi-Object Tracking. In Computer Vision—ECCV 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; Volume 12356, pp. 107–122. ISBN 978-3-030-58620-1. [Google Scholar]
  27. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  28. Liang, C.; Zhang, Z.; Zhou, X.; Li, B.; Zhu, S.; Hu, W. Rethinking the Competition Between Detection and ReID in Multiobject Tracking. IEEE Trans. Image Process. 2022, 31, 3182–3196. [Google Scholar] [CrossRef] [PubMed]
  29. Zhang, Y.; Sun, P.; Jiang, Y.; Yu, D.; Weng, F.; Yuan, Z.; Luo, P.; Liu, W.; Wang, X. ByteTrack: Multi-Object Tracking by Associating Every Detection Box. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer Nature: Cham, Switzerland, 2022; Volume 13682, pp. 1–21. ISBN 978-3-031-20046-5. [Google Scholar]
  30. Lee, Y.; Park, J. CenterMask: Real-Time Anchor-Free Instance Segmentation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 13903–13912. [Google Scholar]
  31. Welch, G.; Bishop, G. An Introduction to the Kalman Filter; University of North Carolina at Chapel Hill, Department of Computer Science: Chapel Hill, NC, USA, 1995. [Google Scholar]
  32. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  33. Kuhn, H.W. The Hungarian Method for the Assignment Problem. Nav. Res. Logist. 1955, 2, 83–97. [Google Scholar] [CrossRef]
  34. Danielsson, P.-E. Euclidean Distance Mapping. Comput. Graph. Image Process. 1980, 14, 227–248. [Google Scholar] [CrossRef]
  35. Zhang, Y.; Wang, C.; Wang, X.; Zeng, W.; Liu, W. FairMOT: On the Fairness of Detection and Re-Identification in Multiple Object Tracking. Int. J. Comput. Vis. 2021, 129, 3069–3087. [Google Scholar] [CrossRef]
  36. Aharon, N.; Orfaig, R.; Bobrovsky, B.-Z. BoT-SORT: Robust Associations Multi-Pedestrian Tracking. arXiv 2022, arXiv:2206.14651. [Google Scholar]
  37. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple Online and Realtime Tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
  38. Cao, J.; Pang, J.; Weng, X.; Khirodkar, R.; Kitani, K. Observation-Centric SORT: Rethinking SORT for Robust Multi-Object Tracking. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 9686–9696. [Google Scholar]
  39. Maggiolino, G.; Ahmad, A.; Cao, J.; Kitani, K. Deep OC-Sort: Multi-Pedestrian Tracking by Adaptive Re-Identification. In Proceedings of the 2023 IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8 October 2023; pp. 3025–3029. [Google Scholar]
Figure 1. Diagram of the multi-scenario pig dataset. This dataset contains a total of ten pig farming scenarios, including eight daytime and two nighttime scenes. The goal of SDGTrack is to train the model on daytime scenarios so that it can generalize to other daytime and nighttime environments.
Figure 1. Diagram of the multi-scenario pig dataset. This dataset contains a total of ten pig farming scenarios, including eight daytime and two nighttime scenes. The goal of SDGTrack is to train the model on daytime scenarios so that it can generalize to other daytime and nighttime environments.
Animals 15 01543 g001
Figure 2. SDGTrack. (a) is the feature extractor including backbone and neck; (b) illustrates our proposed prediction structure of SDGTrack.
Figure 2. SDGTrack. (a) is the feature extractor including backbone and neck; (b) illustrates our proposed prediction structure of SDGTrack.
Animals 15 01543 g002
Figure 3. DAA module.
Figure 3. DAA module.
Animals 15 01543 g003
Figure 4. Re-Byte tracking flow chart.
Figure 4. Re-Byte tracking flow chart.
Animals 15 01543 g004
Figure 5. Visualizes of the discriminative capability of ID embeddings: (a) detection results; (b) visualization of ID embeddings for detected objects; (c) pairwise relationships between ID embeddings of detected objects in the current frame; (d) visualization of ID embeddings in the target sequence template; (e) pairwise relationships between ID embeddings in the target sequence template; (f) matching relationships between ID embeddings of detected objects in the current frame and those in the target sequence template.
Figure 5. Visualizes of the discriminative capability of ID embeddings: (a) detection results; (b) visualization of ID embeddings for detected objects; (c) pairwise relationships between ID embeddings of detected objects in the current frame; (d) visualization of ID embeddings in the target sequence template; (e) pairwise relationships between ID embeddings in the target sequence template; (f) matching relationships between ID embeddings of detected objects in the current frame and those in the target sequence template.
Animals 15 01543 g005
Figure 6. Tracking results for a typical tracking model.
Figure 6. Tracking results for a typical tracking model.
Animals 15 01543 g006
Figure 7. Tracking performance of SDGTrack in three scenarios.
Figure 7. Tracking performance of SDGTrack in three scenarios.
Animals 15 01543 g007
Table 1. Composition of the constructed dataset.
Table 1. Composition of the constructed dataset.
DatasetScenario TypeScene CountTotal ImagesDescription
TrainingDaytime712,600Data from 7 different daytime pig life scenarios
TestNight-In Scene11800A nighttime sequence captured in the same location as the training set.
Night-Out Scene11800Nighttime data recorded in a farming scenario completely different from the training set.
Daytime-Out Scene11800A daytime recording from another farm not included in the training.
Table 2. Evaluation metrics and terms for MOT methods.
Table 2. Evaluation metrics and terms for MOT methods.
MetricDescription
HOTA↑Combined accuracy of detection and identity tracking.
MOTA↑Evaluates the overall accuracy of the multi-object tracking algorithm.
IDF1↑Combines correctly detected objects (IDTP), false positives (IDFP), and missed objects (IDFN) into a single metric.
MOTP↑Measures the precision of the tracker in estimating the positions of targets.
MT↑Assesses the proportion of targets that can be consistently tracked throughout the process.
ML↓Evaluates the proportion of targets that are lost during the multi-object tracking process.
IDS↓Represents the total number of ID switches
FP↓False Positive (FP) refers to negative samples incorrectly predicted as positive by the model, also known as the false alarm rate.
FN↓False Negative (FN) refers to positive samples incorrectly predicted as negative by the model, also known as the miss rate.
FPS↑FPS represents the frame rate of the entire tracking framework.
↑ indicates performance metrics where higher values are preferred; ↓ indicates metrics where lower values are preferred.
Table 3. Comparative analysis of different models.
Table 3. Comparative analysis of different models.
MethodHOTA↑
(%)
MOTA↑
(%)
IDF1↑
(%)
MT↑
(%)
ML↓
(%)
FP↓FN↓IDS↓FPS↑
SORT (yolox)50.845.456.8179701927,57912637.5
DeepSORT (yolox)48.844.253.9189869326,65017112.3
Deep-OC-SORT (yolov8)55.451.559.5191157529,72256724.1
BoT-SORT (yolov8)51.549.653.5161127531,18859623.2
OC-SORT (yolov8)55.351.559.4181257529,72057220.2
ByteTrack47.346.648.0179549828,12835837.4
FairMOT55.155.654.6255756120,33733840.5
CSTrack57.764.351.8266147620,71944721.6
SDGTrack (Ours)83.080.985.1383339413,1632427.5
↑ indicates performance metrics where higher values are preferred; ↓ indicates metrics where lower values are preferred.
Table 4. Performance comparison of various models in different testing scenarios.
Table 4. Performance comparison of various models in different testing scenarios.
MethodDatasetHOTA↑
(%)
MOTA↑
(%)
IDF1↑
(%)
MT↑
(%)
ML↓
(%)
FP↓FN↓IDS↓FPS↑
Deep-OC-Sort (yolov8)night-in scene67.964.171.953064362515.7
night-out scene33.028.238.519022,75929827.9
daytime-out scene83.990.078.213057552724436.0
FairMOTnight-in scene62.568.856.86176047808522.0
night-out scene39.033.545.574626614,9729225.7
daytime-out scene80.090.570.712053558516140.5
CSTracknight-in scene56.264.748.853161917916.8
night-out scene48.452.844.4104114213,84917423.2
daytime-out scene79.591.169.411033367919445.6
SDGTracknight-in scene80.078.181.990155833711117.2
night-out scene75.869.682.6172167893421041.3
daytime-out scene92.995.090.8120176500348.8
↑ indicates performance metrics where higher values are preferred; ↓ indicates metrics where lower values are preferred.
Table 5. Ablation analysis of the SDGTrack model.
Table 5. Ablation analysis of the SDGTrack model.
MethodDatasetDAA Re-ByteHOTA↑
(%)
MOTA↑
(%)
IDF1↑
(%)
IDS↓
CSTracknight-in scene🗴   🗴56.264.748.879
night-out scene48.452.844.4174
daytime-out scene79.591.169.4194
Total57.764.351.8447
SDGTracknight-in scene✓   🗴75.975.676.290
night-out scene70.167.672.6166
daytime-out scene87.593.282.2208
Total77.676.778.6464
night-in scene🗴   ✓75.471.379.720
night-out scene68.162.374.429
daytime-out scene93.293.093.55
Total78.474.382.754
night-in scene✓   ✓80.078.181.911
night-out scene75.869.682.610
daytime-out scene92.995.090.83
Total82.980.985.124
↑ indicates performance metrics where higher values are preferred; ↓ indicates metrics where lower values are preferred.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, T.; Jie, D.; Zhuang, J.; Zhang, D.; He, J. SDGTrack: A Multi-Target Tracking Method for Pigs in Multiple Farming Scenarios. Animals 2025, 15, 1543. https://doi.org/10.3390/ani15111543

AMA Style

Liu T, Jie D, Zhuang J, Zhang D, He J. SDGTrack: A Multi-Target Tracking Method for Pigs in Multiple Farming Scenarios. Animals. 2025; 15(11):1543. https://doi.org/10.3390/ani15111543

Chicago/Turabian Style

Liu, Tao, Dengfei Jie, Junwei Zhuang, Dehui Zhang, and Jincheng He. 2025. "SDGTrack: A Multi-Target Tracking Method for Pigs in Multiple Farming Scenarios" Animals 15, no. 11: 1543. https://doi.org/10.3390/ani15111543

APA Style

Liu, T., Jie, D., Zhuang, J., Zhang, D., & He, J. (2025). SDGTrack: A Multi-Target Tracking Method for Pigs in Multiple Farming Scenarios. Animals, 15(11), 1543. https://doi.org/10.3390/ani15111543

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop