Next Article in Journal
Development and Validation of a Robust and Straightforward LC-MS Method for Measuring Taurine in Whole Blood and Plasma of Dogs and Reference Intervals Calculation
Next Article in Special Issue
Multi-Sensor Integration and Machine Learning for High-Resolution Classification of Herbivore Foraging Behavior
Previous Article in Journal
Novel Insertion/Deletion Polymorphisms and Genetic Studies of the Shadow of Prion Protein (SPRN) in Raccoon Dogs
Previous Article in Special Issue
Bridging the Gap Between Computational Efficiency and Segmentation Fidelity in Object-Based Image Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Automated Chicken Monitoring: Dataset and Machine Learning Methods for Visual, Noninvasive Reidentification

1
Faculty Electronics & Computer Science, Aalen University, 73430 Aalen, Germany
2
School of Computing, Engineering and Built Environment, Glasgow Caledonian University, Glasgow G4 0BA, UK
3
Department of Computer Science, University of the Western Cape, Cape Town 7535, South Africa
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Animals 2025, 15(1), 1; https://doi.org/10.3390/ani15010001
Submission received: 4 November 2024 / Revised: 14 December 2024 / Accepted: 18 December 2024 / Published: 24 December 2024

Simple Summary

The chicken is the world’s most farmed animal. Artificial-Intelligence-based reidentification offers a stress-free alternative to traditional methods, such as leg bands. In this paper, we introduce the first public dataset designed to reidentify individual chickens in images using artificial intelligence. Additionally, we provide an overview of existing public datasets for reidentifying other animals. On the introduced dataset, we test different artificial intelligence models to visually identify individual chickens. We evaluate two scenarios: using a single image per individual and using multiple images. For scenarios with multiple images per individual, we achieve accuracy rates of between 95.1% and 100%. Additionally, we adapt an artificial intelligence model originally designed for other animals and test its performance on chickens. Our findings reveal that this approach is highly effective when data are limited.

Abstract

The chicken is the world’s most farmed animal. In this work, we introduce the Chicks4FreeID dataset, the first publicly available dataset focused on the reidentification of individual chickens. We begin by providing a comprehensive overview of the existing animal reidentification datasets. Next, we conduct closed-set reidentification experiments on the introduced dataset, using transformer-based feature extractors in combination with two different classifiers. We evaluate performance across domain transfer, supervised, and one-shot learning scenarios. The results demonstrate that transfer learning is particularly effective with limited data, and training from scratch is not necessarily advantageous even when sufficient data are available. Among the evaluated models, the vision transformer paired with a linear classifier achieves the highest performance, with a mean average precision of 97.0%, a top-1 accuracy of 95.1%, and a top-5 accuracy of 100.0%. Our evaluation suggests that the vision transformer architecture produces higher-quality embedding clusters than the Swin transformer architecture. All data and code are publicly shared under a CC BY 4.0 license.

1. Introduction

1.1. Background

Chickens struggle to recognize other individuals after visible changes are applied to the comb or plumage [1]. Much like chickens are able to use visual cues to differentiate each other, artificial intelligence (AI) is capable of utilizing image or video inputs for reidentification (re-ID) purposes. Animal re-ID, the task of identifying individual animals within one (or sometimes several) species, finds applications in various fields, particularly in livestock management [2,3,4,5,6,7,8,9,10] and in wildlife conservation efforts, where monitoring endangered species is crucial [11,12,13,14]. Re-ID falls into one of two categories: closed set and open set. In closed-set re-ID, all individuals are known from the beginning, and those to be identified can be matched with identities of a predefined set. In open-set re-ID, the identity of the individual in question may not necessarily be part of a predefined set. It is possible to encounter completely new, undocumented individuals. Such individuals must be annotated as a new identity and, upon subsequent encounters, be accurately matched. Visual traits play a pivotal role in animal re-ID within computer vision, serving as essential markers for distinguishing individuals. Conversely, little interindividual variability poses a challenge to the re-ID task. However, the task is complex and extends beyond mere visual cues. Factors such as lighting, perspective, body changes over time, and partially obscured body parts pose additional challenges [15]. The task of re-ID is closely related to tracking, where individuals are detected and tracked across various video frames. During tracking, individuals often need to be reidentified after leaving and re-entering the field of vision. For instance, the challenge of individual chickens disappearing between frames or being occluded in later frames can be effectively addressed by chicken re-ID, which accurately recognizes the returning individual as the same chicken that left the frame.

1.2. Motivation

Automated tracking holds great potential in precision livestock farming [16]. These systems may allow for the observation of social structures and behavior, enhance welfare, and lead to more efficient animal management with minimal disruption to the livestock [17]. Chicken welfare assessment is increasingly focusing on individual animals rather than entire groups [18]. To effectively monitor, i.e., behavior, especially when dealing with many animals in a large group, it is crucial to accurately recognize and track each individual [18]. Moreover, EU regulations require individual identification to ensure traceability during disease outbreaks [19]. Traditional methods for re-ID, such as leg bands, wing tags, or backpacks with sensors, can cause significant stress to the animals and have been shown to negatively affect behavior, the immune system, and body weight [20,21]. AI-based re-ID offers a noninvasive and efficient alternative to traditional identification methods. Furthermore, identifying individual chickens is essential for behavioral studies [22], as it plays a crucial role not just in experiment oversight but also in ensuring accurate statistical analysis. The stress induced by traditional methods not only affects the animals’ well-being but also compromises the validity of research findings. AI-based identification maintains the integrity of a study while minimizing the stress on the animals. Moreover, ref. [23] recommends continuing exploring the ethological complexity of chickens in settings that are noninvasive and nonharmful, not only in commercial farming settings but also in more naturalistic settings.
Despite the significant potential of re-ID technologies, there is a notable lack of datasets that form the foundation for such advancements. Notably, as far as we are aware, no publicly available re-ID dataset for chickens currently exists, underscoring the need for development in this area. Public datasets for individual animal re-ID are scarce generally [17,24], in particular well-annotated datasets [8]. The practice of openly sharing data and code should be encouraged to enhance result comparability, but not all research data are currently made public.

1.3. Contributions

In broad terms, we performed visual closed-set re-ID of 50 individual chickens. The data used were cut-out crops of the chickens’ bodies, which excluded the background. The dataset was specifically constructed for this task, with images taken from a standing position to emphasize the plumage as the primary identifying feature.
In detail,
i
We present a summary of animal re-ID studies and give a comprehensive overview of publicly available datasets.
ii
We address the existing gap and introduce the first publicly available dataset for chicken re-ID: Chicks4FreeID. The dataset supports closed -and open-set re-ID, as well as semantic and instance segmentation tasks. We make this thoroughly documented dataset freely accessible to the research community and the public.
iii
We evaluate a species-agnostic state-of-the-art model on our dataset through two experiments. In the first experiment, we test the model using its frozen weights, which were not trained on chicken data. In the second experiment, we fine-tune the model to adapt it specifically to our dataset.
iv
We train two feature extractors from scratch in a standard supervised manner and test them on our dataset. Both models are based on transformer architectures.
v
We perform additional one-shot experiments with all previously mentioned models.
vi
Lastly, we make all associated code publicly available to ensure transparency and facilitate further research.

2. Related Work

2.1. Re-ID

While facial recognition is a prevalent method for reidentifying humans [25], the faces of animals can likewise serve as a means to reidentify individuals, as has previously been demonstrated for rhesus macaque [26], chimpanzee [27], cats [28], lions [29], dogs [30], giant pandas [13], and red pandas [14]. Notably, ref. [31] applied transformer-based similarity learning for chicken re-ID, using the chickens’ heads as the primary identifying feature, with a focus on a uniform group of white laying hens. Additionally, they examined the importance of different head features for visual re-ID, finding that the models prioritized characteristics such as the comb, wattles, and earlobes. However, animals frequently exhibit more distinctive visual traits beyond their faces. For example, natural markings such as stripes [32,33,34,35] and scale patterns [36] have served as prominent identifiers. Also, specific body parts can contribute to distinguishing individuals, such as the fins of dolphins [37] and sharks [38]. Similarly to how fingerprints differentiate humans, the nose prints of dogs have been utilized to uniquely identify individual dogs [39]. However, some studies aimed at identifying animals with little interindividual variability. Species exhibiting minimal or subtle visual distinctions between individuals are, for instance, (polar) bears [40,41] and elephants [42].

2.2. State of the Art

To further advance the field and aid the research community, ref. [43] released the WildlifeDatasets toolkit: an open-source toolkit for animal re-ID. It gathers publicly available animal re-ID datasets in one place in an effort to make them more easily accessible and to improve usability. Included are various tools, i.e., for data handling and processing, algorithms relevant to the task of re-ID, pretrained models, as well as evaluation methods. They address the prevailing absence of standardization across the literature and facilitate the comparability and reproducibility of results. Within their work, they also introduce a new state-of-the-art model, the MegaDescriptor, notably the first foundation model for animal re-ID. Likewise, ref. [44] presented an open-source re-ID method initially developed for sea stars, which was successfully extended to seven mammalian species without adjustments. They also reported state-of-the-art results. Moreover, ref. [45] introduced Tri-AI, a system designed for the rapid detection, identification, and tracking of individuals from a wide range of primate species. The system is capable of processing both video footage and still images.

2.3. Datasets

A review of the existing resources revealed fewer than 40 publicly available datasets for animal re-ID. This led to the conclusion that a significant number of animal species are not yet covered, including chickens. Birds in general seem to be under-represented in this domain, with only a couple of datasets available [46,47]. In fact, a noticeable focus lies on marine life [48,49,50,51,52,53,54,55,56,57,58]. However, cattle are the most frequently featured species [3,59,60,61,62,63], with much of the data collected by the same group of researchers. Despite chickens being the most widely farmed animal globally, no public re-ID dataset specifically for chickens was found. Table 1 provides a summary of the publicly accessible datasets found for animal re-ID, arranged by year. Each entry details the name of the dataset (“Dataset”), the associated publication (“Publ.”), and species focus (“Species”). “IDs” denotes the number of unique identities present within the dataset. Additionally, the total number of annotated animal instances within all images of each dataset is noted (“Annot.”). An indication (*) of whether the data were derived from video sources is given as well. For ease of access, a direct link to each dataset is provided (“Avail. at”). Although all of the datasets are publicly accessible, some are released under licenses that are relatively restrictive.

3. The Chicks4FreeID Dataset

3.1. Overview

The Chicks4FreeID dataset contains top-down view images of individually segmented and annotated chickens. It predominantly features female chickens (hens). For simplicity and to align with colloquial language, we refer to hens as simply “chickens” and male chickens as “roosters”. Some images also feature ducks. Each image is accompanied by a color-coded semantic segmentation mask that classifies pixel values by animal category (chicken, rooster, duck) and background, as well as binary segmentation mask(s) for the animal instance(s) depicted. Additionally, the dataset includes preprocessed cut-out crops (detailed in Section 3.4) of the respective animal instances. The Chicks4FreeID dataset can be utilized for various tasks, including re-ID, semantic segmentation, instance segmentation, and possibly anomaly detection. For each task, a distinct subset configuration was created [64]. Figure 1 gives a first overview of the dataset.

3.2. Collection

The data were collected manually, with no images derived from any existing dataset. Eleven private households in southern Germany were visited to photograph chickens. A total of 677 images were captured using two similar models of cameras: the “CyberShot DSC-RX100 VI” and the “CyberShot DSC-RX100 I” from Sony. The resolution of the images stands at 3648 × 5472 pixels. Each image includes at least one chicken, ensuring no images without chickens are part of the dataset. The animals were allowed to move freely without constraints to prevent distress. As a result, other individuals frequently entered the frame. The chickens were photographed from a standing position, aiming to capture the plumage from a top-down perspective. However, since the animals were free to move, their distance from the camera varied, which affected the view of their plumage. Thus, the resulting images do not show a clean top-town view. Most individuals were photographed outdoors under natural light, except for four individuals, which were indoors in a coop during winter. Data collection took approximately one year. However, all photos of a given individual were taken on the same day.

3.3. Annotation

All annotations were manually created by a human annotator using Labelbox [93] under a free educational license.
Instance Masks By drawing an almost pixel-perfect polygon outline, a binary instance mask was created for each animal in an image. This process resulted in a total of 1270 instance masks. In each mask, the instance is encoded in white, whereas the black area is considered background. Each instance includes the comb, head, beak, and plumage. In contrast, feet and scattered feathers are classified as part of the background, along with any other visible objects. Feet were intentionally excluded to prevent identification through leg rings. We created instance masks instead of bounding boxes to aid the creation of precise cut-out crops for re-ID. Bounding boxes often include other individuals, which could confuse the model, as well as background elements, which could introduce unintended clues. By using instance masks, we ensure a focus on the animal’s features, avoiding reliance on background information.
Identities Each instance was assigned an identity. There were 54 distinct identities. Prior to the photography sessions, the visible features of each animal were meticulously studied and recorded in a notebook. This was conducted to ensure the correct identification of each individual during annotation. Ground truth annotation was therefore performed using expert knowledge, without the use of any algorithms. In cases where the human annotator could not assign an identity, the instance in question was labeled as identity “Unknown”. It is essential to clarify that the label “Unknown” does not imply the presence of a new, undocumented individual. Instead, it represents an unidentified individual from the closed set, more precisely, from the annotated coop. A key strength of the Chicks4FreeID dataset is its expert ground truth labeling for all instances, ensuring reliable verification of results even in open-set Re-ID tasks. This sets it apart from datasets that include completely unannotated instances or rely on ground truth labels generated through clustering or AI methods, which may introduce labeling errors.
Animal Category Each instance was assigned to one of 3 animal categories. These are “chicken”, “rooster”, and “duck”. Roosters, and particularly ducks, are exceptions in this dataset, which is primarily composed of female chickens (hens). For simplicity and to align with colloquial language, instances of female chickens were categorized as “chicken” while male chickens were categorized as “rooster”.
Visibility Rating It was common for animals to be partially obscured by other animals or objects in the images. Additionally, they were often not fully contained within the image frame. Acknowledging the resulting varying visibility of the instances, each was assigned one of 3 visibility ratings: “bad”, “good”, or “best” (for examples, see Figure A1; for full overview, see Table A1 and Table A2). The “best” rating includes instances that fully display the animal from the desired top-down perspective and those where only an insignificant part is missing, such as the very tip of the tail feathers. Instances that include only small parts of the animal and with an undesired perspective fall under the “bad” rating. All remaining instances that do not qualify as “bad” or “best” are rated as “good”.
Coops The location (one of 11 coops) of each photograph was documented during the capture process. As a result, each image was annotated with the coop it belonged to. It is therefore straightforward to match the animals in the images to their respective coops.
Semantic Segmentation Masks Furthermore, 677 color-coded semantic segmentation masks were created, 1 for each image. They include four possible classes: “background”, “chicken”, “duck” and “rooster”. Every mask includes “background” and “chicken”, while “duck” and “rooster” are included if present. While the semantic segmentation masks are not directly involved in the re-ID process, they were added to the dataset due to the availability of existing annotations, expanding the dataset’s utility to support additional semantic segmentation tasks.

3.4. Preprocessing

The Chicks4FreeID dataset contains preprocessed data for the task of re-ID. These are cut-out crops of isolated animal instances with a solid black background. By removing visual cues like feet with rings or environmental features, it is ensured that the re-ID relies solely on the animals’ physical characteristics. The following steps describe the preprocessing procedure to obtain the cut-out crops for the re-ID task. For all individuals captured in an image, a bounding box is created based on the instance masks. In the first step, both the image and the mask are cropped (to the area of interest contained in the bounding box) to focus solely on the individual (see Figure 2: Step 1). The cropped mask is then used to remove the background from the cropped image (Step 2). Finally, the resulting image is adjusted to a square shape for ease of use and consistency (Step 3). The resulting resolutions remain as is, with no resizing taking place.

3.5. Dataset Statistics

The dataset contains 677 images featuring 54 individual animals. The 677 images were captured across 11 different coops, and the 54 individual animals were divided into three categories: “chicken”, “rooster”, and “duck”. There are 50 “chicken”, 2 “rooster”, and 2 “duck” identities. For simplicity and to align with common usage, hens were categorized as “chicken” while the two roosters were given their own category. Each image is paired with a semantic segmentation mask and an instance segmentation mask for each animal depicted. In total, there are 1270 instance segmentation masks. From the instance masks and the corresponding images, 1270 cut-out crops of the individual animals were generated, specifically for use in the re-ID process. There are 1215 “chicken” instances, 15 “rooster” instances, and 40 “duck” instances. The instances also received a visibility rating: 798 “best”, 201 “good”, and 271 “bad” instances. Among the 271 “bad” instances, 69 are labeled “Unknown” as they could not be certainly identified by the annotator. No “Unknown” instances are present among those rated as “best” or “good”. Therefore, the ground truth identity for all instances with these ratings is known. Figure 3 illustrates the number of instances for each individual, along with their visibility ratings. The figure starts with the individual having the most instances rated as “best” and is arranged in descending order. The chicken identity with the most “best” ratings is “Mirmir”, with 27 instances, while “Isolde” has the fewest, with 4 “best” rated instances. The dataset comprises chicken individuals with both uniform and nonuniform plumage. The diversity in appearance possibly aids individual identification, with differences in colors and patterns serving as markers. As a result, this renders the dataset easier to solve than one containing only uniform-plumage individuals. Nevertheless, some individuals in the dataset exhibit uniform plumage, including four that are solid white, four that are solid black, four in shades of gray, and six in shades of orange. Examples of these plumage colors are illustrated in Figure A2. Detailed information about the individuals and the corresponding annotations is provided in Table A3.

4. Materials and Methods

4.1. Hardware

The training and evaluation were conducted on 64 GB shared memory Apple M3 Max Chips (2023) running PyTorch 2.3.0 with MPS acceleration.

4.2. Data

For the experiments, we employed the subset “chicken-re-id-all-visibility”, which features 50 different individuals. It is hosted on Hugging Face [64], a widely used platform for machine learning datasets and models. The dataset encompasses 1146 “chicken” instances of all visibility ratings (793 “best”, 181 “good”, and 172 “bad”). It does not contain any “Unknown” instances nor instances of the animal category “rooster” or “duck”. As a result, it comprises 1146 pairs of preprocessed “chicken” cut-out crops and assigned identities. The subset was split into 916 train pairs and 230 test pairs (stratified). All 50 identities were included in the training set; this ensured that the testing set did not introduce any new identities. For a fair evaluation on all identities, the train/test split was stratified, i.e., each identity had the same fixed percentage of its cut-out crops allocated to the test set. Hence, the test set also included all of the 50 identities. As a consequence of the stratified split, identities with a higher total number of crops contributed more to the test set compared to identities with fewer crops, ensuring proportional representation across all identities. After accessing the described “chicken-re-id-all-visibility”subset, we additionally introduced a validation split (10% of the training set, stratified random split). Consequently, the dataset was finally split into 824 pairs for training (prior to applying any augmentation), 92 pairs for validation, and 230 pairs for testing.

4.3. Augmentation

Augmentation was applied dynamically during the training process to the training data: random rotation (360 degrees), random flip (horizontal and vertical), and RandAugment [94]. No data augmentation was applied to the test or validation set. To avoid data leakage, it is important to apply data augmentation only after a train-test split was established. This ensures that augmented versions of the same original image do not appear in both sets.

4.4. Feature Extractors

Vision Transformer We employed the ViT-B/16 [95] architecture, as implemented in [96]. The hyperparameters (ADAM + CosineWarmup) were inspired by the optimizer used in the lightly benchmarks [97] for their vision transformer backbones.
Swin Transformer We also utilized the swin_large_patch4_window12_384 architecture [98] as implemented in [99]. The hyperparameters (Stochastic Gradient Descent + Cosine Annealing, ArcFace [100] loss function) mirrored those used to build the MegaDescriptor, which also employs the same Swin transformer architecture. The Swin transformer itself is based on the vision transformer architecture [95]. The difference between the Swin Transformer and the vision transformer lies in how they handle image data: the Swin transformer uses a hierarchical structure with shifted windows to capture local and global features, while the vision transformer treats images as sequences of patches, relying on self-attention mechanisms throughout.
MegaDescriptor The employed MegaDescriptor-L-384 [43] (CC BY-NC 4.0 license [101]) is a state-of-the-art feature extractor for animal re-ID from the WildlifeDatasets toolkit (MIT license). It is based on the Swin transformer architecture [98] and was pretrained on diverse datasets featuring various animal species. A notable hyperparameter choice made by the MegaDescriptor-L384 authors is the ArcFace [100] loss function, which aims to aid in building meaningful embeddings. We selected the frozen MegaDescriptor-L-384 model over DINOv2 [102] and CLIP [103] due to its better performance on unseen animal domains, as reported by the authors. Their evaluation included cattle as an example of an unseen domain [43].

4.5. Classifiers

k-NN k-NN stores all the training data and calculated distances at the time of prediction. It works by comparing a new chicken’s embedding (feature vector) to all known embeddings. It then finds the k-closest embeddings in the feature space and assigns the most common identity among those neighbors. Our settings were inspired by [104]. We used a high value of k = 200 to provide a broader context, allowing us to consider not just the closest neighbor but the overall composition of the embedding space. For a correct prediction, other individuals must be sufficiently distant from the cluster of the “correct” individual. This k-NN approach prioritizes clustering quality over absolute discriminative power.
Linear Classifier To evaluate absolute discriminative power, we relied on a linear classifier. The employed linear classifier is a simple linear layer. It was trained for 90 epochs using the extracted embeddings (feature vectors) of the training data. Thereby, it learned to associate each feature vector with the correct identity by adjusting its weights, ultimately mapping each input to1one of the 50 identity classes. It was validated on the embeddings of the validation set.

4.6. Evaluation Metrics

We provide three of the most common metrics for closed set animal re-ID. These are mAP (mean average precision), top-1 accuracy (ratio of correct predictions versus total predictions), and top-5 accuracy (accuracy of the correct class being within the top 5 predictions) as implemented in TorchMetrics [105]. It is important to note that the mAP focuses more on the quality of the probability estimates, serving as a ranking metric that assesses “how well the model assigns high probabilities to the correct chicken.” This is similar to the top-5 accuracy, where the model’s prediction is considered correct if the correct answer is among the top five guesses. However, even though top-5 accuracy can provide some confidence in the model’s predictions, it does not fully capture the model’s actual performance in terms of the number of errors made. Therefore, the top-1 accuracy must be considered the most.

5. Experiments

The approach for closed-set re-ID involved two steps. First, a feature extractor generated embeddings for the cut-out crops. Second, the resulting feature vectors (embeddings) were then passed to a classifier to ultimately assign the identities. Feature extractor and classifier were trained separately. Figure 4 illustrates the training and evaluation process. We evaluated four feature extraction models: MegaDescriptor (both frozen and fine-tuned), vision transformer, and Swin transformer. All feature extractors were fed with images at an input resolution of 384 × 384 pixels (the square-shaped cut-out crops were resized). Additionally, we tested with a variation of two classifiers: k-NN and a linear classifier. In our experiments, we explored different training strategies: standard supervised learning (see Section 5.2) and one-shot learning (see Section 5.3). We employed both standard supervised learning and one-shot learning approaches to evaluate performance across different data availability scenarios. The one-shot setting simulated situations with very limited data while also providing insights into generalization capabilities by reducing the risk of overfitting. We also specifically compared the performance of the pretrained MegaDescriptor with its fine-tuned counterpart (see Section 5.1). Each experiment, as a whole, was repeated three times to ensure consistency.

5.1. Domain Transfer Experiment

This experiment had two main objectives. First, it aimed to evaluate the performance of the frozen pretrained MegaDescriptor on our Chicks4FreeID dataset, which represents a new, unseen domain. The MegaDescriptor was originally trained on diverse datasets featuring various animal species. However, chickens were not part of the original training data. We evaluated the MegaDescriptor as a feature extractor with its weights frozen, preserving its pretrained state. Second, the experiment sought to enhance performance by fine-tuning the MegaDescriptor on the Chicks4FreeID dataset. This optimally allows the model to adapt by leveraging the features it previously learned from other species and refining them through additional training. For this, we unfroze all the model’s layers and trained it for 200 additional epochs using the training set, as detailed in Section 4.2. For testing and validation, we utilized the test and validation sets, also described in Section 4.2.

5.2. Standard Supervised Learning Experiment

This experiment aimed to train a chicken feature extractor from scratch using the Chicks4FreeID dataset. We compared two transformer architectures; the Swin and the vision transformers.

5.3. One-Shot Experiment

This experiment aimed to evaluate all previously discussed models in a one-shot learning setting, where the training set was reduced to only one sample per individual. With 50 individuals in the dataset, the training set initially consisted of 50 samples, each representing one individual. All remaining samples of the previously established training set (see Section 4.2) were unused. These 50 samples were then augmented dynamically during training using the methods described in Section 4.3. The validation and test sets remained unchanged, as outlined in Section 4.2. All feature extractors, except the non-fine-tuned frozen MegaDescriptor, were trained on this one-shot training set. The classifiers were also trained/fitted solely on this reduced set. For the non-fine-tuned frozen MegaDescriptor, the only difference in the one-shot setting was therefore the classifier. Among the conducted experiments, this scenario most closely resembled an open-set re-ID setting, where a new, previously undocumented identity appears and initially little data are available. Consequently, the expected performance in an open-set scenario is likely to be similar the results observed in the one-shot setting, particularly for the non-fine-tuned frozen MegaDescriptor. However, the classifier would need to be retrained each time a new individual appears. To enable open-set re-ID and to identify unknown individuals without retraining, the classifier head could be replaced with a clustering algorithm, such as DBSCAN [106,107], or hierarchical clustering [108].

6. Results and Discussion

Table 2, Table 3 and Table 4 show the results on the test data for the domain transfer, standard supervised learning, and one-shot experiments, respectively. Figure 5 additionally illustrates the top-1 accuracy as bar charts. For runtime results and discussion, see Figure A3 in the Appendix A.
MegaDescriptor Shows Improved Performance with Fine-Tuning The fine-tuned MegaDescriptor model achieved a mAP of 96.0% and a top-1 accuracy of 91.6% when paired with a linear classifier (see Table 2). It outperformed the frozen MegaDescriptor in the domain transfer experiment. This aligns with expectations that a model performs not as well on an unseen domain. The fine-tuned MegaDescriptor also outperformed the Swin transformer, which we specifically trained on our data from scratch (see Table 3). This suggests the pretrained features are useful for identifying chickens, since fine-tuning seems preferable to training the same model from scratch.
Frozen MegaDescriptor is a Powerful Feature Extractor The frozen MegaDescriptor, when combined with a linear classifier, still performed well, with a mAP of 92.0% and a top-1 accuracy of 81.8% (see Table 2). Its performance was not far behind that of models specifically trained on our data (see Table 3). This is particularly impressive given that the frozen model had no exposure to chicken data. It was trained on a large dataset of many different animal species, which also included bird images. Since birds share common features like feathers, this overlap may have further contributed to the model’s strong performance, in addition to the large training dataset.
Trained-from-Scratch Vision Transformer Yields the Best Results The vision transformer trained from scratch, combined with a linear classifier, achieved the overall best results with a mAP of 97.0%, a top-1 accuracy of 95.1%, and a perfect top-5 accuracy of 100.0% (see Table 3).
Fine-Tuned MegaDescriptor Leads in One-shot Setting As expected, the performance dropped for all models in the one-shot setting (see Table 4). When trained with only a single image per individual, the fine-tuned MegaDescriptor achieved the best performance, with a 64.5% mAP, a 56.2% top-1 accuracy, and an 81.8% top-5 accuracy, closely followed by the vision transformer, which achieved a 61.9% mAP, a 52.3% top-1 accuracy, and a 80.8% top-5 accuracy. Both models were paired with a linear classifier for these evaluations. The fine-tuned MegaDescriptor likely outperformed the vision transformer in this scenario because it had already learned embeddings through pretraining before being fine-tuned in a one-shot manner. In contrast, the vision transformer had no prior learning, starting from scratch. This pretraining advantage gave the MegaDescriptor a head start in understanding features, likely contributing to its better performance. This underscores that in scenarios with very limited data, like the one-shot example, transfer learning (fine-tuning pretrained models) is more effective than training from scratch. Furthermore, we think that the improved discrimination capabilities make the fine-tuned MegaDescriptor more suitable in an open-set re-ID setting than the other models.
Evaluation Focus Varies Between Linear Classifiers and k-NN In all experiments, feature extractors consistently performed worse when paired with k-NN compared to when they were paired with a linear classifier. This was mainly due to the differences in the evaluation focus of the employed classifiers. The employed k-NN approach is more challenging by design as it aims to evaluate the clustering quality in embedding space, instead of absolute discriminative power. Additionally, the k-NN is a parameterless model, while the linear classifier has approximately hidden_dim × 50 parameters. k-NN only captures distances in the feature vector space, whereas a linear classifier can learn robust decision boundaries. Moreover, with a linear classifier, we can track training progress with a validation split, leading to better generalization compared to the fixed boundaries of k-NN.
Higher-quality Embedding Clusters with Vision Transformer Training the Swin transformer architecture, which also served as the basis for the MegaDescriptor, appeared to result in lower clustering quality of the embeddings compared to the vision transformer architecture, meaning the overall composition of the embedding space appeared to be better when employing a vision transformer. This difference is evident in Table 2 and Table 3 when paired with k-NN. This could be due to the simpler architecture and loss function, which could lead to reaching better performance faster.
Higher Accuracy with Vision Transformer Furthermore, the vision transformer consistently outperformed the Swin transformer when paired with a linear classifier. This could be because it is optimized to predict class probabilities directly, while the Swin transformer focuses on learning the relative distances between features in the feature space. This difference in optimization goals allows the vision transformer to better capture the overall structure of the data, leading to more accurate classification when paired with the same classifier.

7. Conclusions

7.1. Findings

We introduced the first open-source dataset for chicken re-ID to address the need for well-annotated, publicly available datasets for animal re-ID. Utilizing this dataset, we conducted a series of closed-set re-ID experiments. We combined various transformer-based feature extractors (MegaDescriptor, vision transformer, Swin transformer) with two classifiers (k-NN and linear classifier) and evaluated their performance across different settings (domain transfer, training from scratch, and one-shot learning).
We found that pretraining and domain transfer are critical when dealing with limited data per instance. The MegaDescriptor model proved well suited as a pretrained model in such scenarios. Pretraining on a different domain (e.g., animals other than chickens) enables a model to learn generalizable features that can later be fine-tuned to the specific target domain (e.g., chickens). Interestingly, if sufficient data are available, training a feature extractor from scratch is not necessarily preferable, as the fine-tuned MegaDescriptor outperformed the Swin transformer, which was trained from scratch.
Furthermore, our evaluation suggests that the vision transformer architecture produces higher-quality embedding clusters than the Swin transformer architecture.
Overall, transformer-based architectures as feature extractors achieved high accuracy and mAP on our dataset. The best results on our dataset, however, were achieved with the vision transformer trained from scratch combined with a linear classifier (mAP 97.0%, top-1 95.1%, top-5 100.0%).

7.2. Limitations and Future Work

We have not yet explored which features the models prioritize during the identification process, such as patterns, colors, shapes, specific areas of the plumage, or certain body parts. This presents an interesting direction for future research. The classifiers used for the experiments were straightforward and effective for the purposes of this study. While more advanced classifiers could potentially yield even better results, the main focus of this work was on evaluating the feature extractors. Another potential direction for future work is adapting the Chicks4FreeID dataset for open-set Re-ID experiments. Researchers could, for example, exclude certain individuals from the training set and introduce them later during evaluation. The dataset subset used for the experiments contained 1215 instance annotations of 50 individual chickens, offering a compact yet valuable resource. The versioning system of the dataset facilitates potential expansions and continuous improvements, ensuring its ongoing relevance and applicability for future research. For chicken breeds with minimal interindividual variability, such as those with uniform plumage, increasing both the number of individuals and instances per individual could further improve the re-ID process. This expansion would be particularly relevant for applications in industrial farming, where thousands of chickens of a single breed are typically kept. Furthermore, photographing individual chickens over an extended period proved challenging, as free-range chickens are often preyed upon by wild animals such as raccoons, foxes, or rats. Therefore, the decision was made to capture all images of a given individual in a single day. Unfortunately, this approach did not capture changes in appearance over time, such as molting. Future work could focus on increasing variability by photographing chickens under different lighting conditions, such as from dusk till dawn. It could also explore the use of different imaging techniques, like multispectral imaging. Additional improvements could include capturing different angles (front, back, side), documenting various stages of life, and recording changes over time. While challenging, these additions could provide a more comprehensive dataset. However, a dataset gains its true value when it is made publicly available. Therefore, we encourage the release of future datasets under open, unrestricted licenses.

Author Contributions

D.K.: conceptualization, data curation, investigation, methodology, project administration, visualization, writing—original draft preparation. T.S.: formal analysis, investigation, methodology, software, validation, visualization, writing—original draft preparation. U.K.: funding acquisition, resources, supervision, writing—review and editing. W.I.: supervision, writing—review and editing. All authors have read and agreed to the published version of this manuscript.

Funding

Publication funded by Aalen University of Applied Sciences.

Institutional Review Board Statement

This research project was reviewed according to the research ethics standards of Aalen University of Applied Sciences. A positive vote was given (date of approval: 23 October 2024, approval code: 14-241023).

Informed Consent Statement

Not applicable.

Data Availability Statement

The Chicks4FreeID dataset and the accompanying code (excluding imported libraries or models from external sources, which have their own licenses) are released under the CC BY 4.0 license. This license allows for the distribution, remixing, adaptation, and building upon the dataset in any medium or format. Users must give appropriate credit to the authors, include a link to the license, and clearly indicate if any changes were made. Commercial use of the dataset is permitted. Dataset: https://doi.org/10.57967/hf/2345; Code: https://github.com/DariaKern/Chicks4FreeID (accessed on 3 November 2024).

Acknowledgments

We are immensely thankful to the kind chicken owners who opened their coops for our research, allowing us to collect data and generously offering us fresh eggs. Each of your chickens has made a unique and valuable contribution to the advancement of science. Furthermore, we would like to acknowledge that AI-assisted tools were utilized for grammar and spelling checks.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of this study; in the collection, analyses, or interpretation of the data; in the writing of this manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
k-NNk-nearest neighbor
mAPmean average precision
re-IDreidentification

Appendix A. Annotations

Figure A1. Examples of visibility rating “best”, “good”, and “bad”.
Figure A1. Examples of visibility rating “best”, “good”, and “bad”.
Animals 15 00001 g0a1
Table A1. Full overview of all “chicken” annotations in the Chicks4FreeID dataset.
Table A1. Full overview of all “chicken” annotations in the Chicks4FreeID dataset.
CoopImagesIDBadBestGoodTotalCoopImagesIDBadBestGoodTotal
129Coop Total1628549......Camy37111
#Unknown110011 Samy820937
Chantal1506 Yin215219
Chayenne18110 Yuriko010010
Jaqueline1517742Coop Total142548
Mandy210315 Brownie124227
236Coop Total14391366 Spiderman018321
#Unknown4004847Coop Total2481565
Henny212418 Brunhilde111012
Shady314320 Fernanda015318
Shorty513624 Isolde141217
360Coop Total22581696 Mechthild018018
#Unknown5005968Coop Total148713114
Amalia36312 #Unknown1001
Edeltraut210315 Mavi217120
Erdmute212620 Mirmir127533
Oktavia412319 Nugget825235
Siglinde410115 Skimmy218525
Ulrike2801010140Coop Total5718936282
426Coop Total729541 #Unknown230023
Hermine412521 Beate322530
Matilda317020 Borghild718328
5116Coop Total8414148273 Eleonore616325
#Unknown220022 Henriette326433
Erna512421 Kristina321529
Heidi1020434 Margit218323
Isabella818733 Millie319426
Kathrin720532 Mona626638
Marina15241049 Sigrun123327
Monika11169361167Coop Total88013101
Regina515626 Gretel522431
Renate116320 Lena119020
646Coop Total16521280 Tina225734
#Unknown3003 Yolkoono014216
.....................Total677502417931811215
Table A2. Full overview of all “rooster” and “duck” annotations in the Chicks4FreeID dataset.
Table A2. Full overview of all “rooster” and “duck” annotations in the Chicks4FreeID dataset.
CoopIDCategoryBadBestGoodTotal
4Coop Total 2231540
EvelynDuck112922
MarleyDuck111618
5ElvisRooster61411
9JacksonRooster2114
Grand Total4 3052055

Appendix B. Plumage

Figure A2. Uniform plumage examples from left to right: solid white, solid black, shades of gray, shades of orange.
Figure A2. Uniform plumage examples from left to right: solid white, solid black, shades of gray, shades of orange.
Animals 15 00001 g0a2
Table A3. Overview of uniform plumage “chicken” individuals in the Chicks4FreeID dataset. Grand total of annotations of mixed plumage individuals is included for completeness.
Table A3. Overview of uniform plumage “chicken” individuals in the Chicks4FreeID dataset. Grand total of annotations of mixed plumage individuals is included for completeness.
PlumageIDCoopBadBestGoodTotal
Solid WhiteTotal 528538
Chantal11506
Chayenne118110
Jaqueline11517
Mandy1210315
Solid BlackTotal 5392165
Erdmute3212620
Ulrike328010
Isolde8141217
Fernanda8015318
Shades of GrayTotal 8661084
Erna5512421
Mavi9217120
Sigrun10123327
Yolkoono11014216
Shades of OrangeTotal 169017123
Henny2212418
Shady2314320
Shorty2513624
Brunhilde8111012
Mechthild8018018
Gretel11522431
Uniform PlumageGrand Total 3422353310
Mixed PlumageGrand Total 207570128905

Appendix C. Runtime

For efficiency and performance comparison, we report runtime speeds for each run in Figure A3.
Figure A3. Mean runtime results by experiment type and model (log scale).
Figure A3. Mean runtime results by experiment type and model (log scale).
Animals 15 00001 g0a3
Embedding training runtime and classifier training runtime are reported separately. Since each experiment was repeated three times, we report the mean values and error bars for each distinct experiment group. k-NN was the fastest classifier, taking around 1 min per run. This was consistent across both one-shot and full experiments, indicating that runtime speed was primarily dominated by overhead. Overall, excluding k-NN, training one-shot models was significantly faster than training with full datasets, by one order of magnitude, as seen in the logarithmic scale. The figure also shows that training only the classifiers (k-NN and linear) was much more efficient than training or fine-tuning the full pipeline. Training the comparatively large feature extractors took considerably longer. Another noticeable gap exists between the vision transformer and MegaDescriptor architectures (see Embedding Training bars for MegaDescriptor (fine-tuned) and Swin transformer versus vision transformer). This difference arises because the vision transformer avoids calculating shifting windows and uses a simpler loss function that requires only one image instead of two.

References

  1. Guhl, A.M.; Ortman, L.L. Visual Patterns in the Recognition of Individuals among Chickens. Condor 1953, 55, 287–298. [Google Scholar] [CrossRef]
  2. Andrew, W.; Hannuna, S.; Campbell, N.; Burghardt, T. Automatic individual holstein friesian cattle identification via selective local coat pattern matching in RGB-D imagery. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 484–488. [Google Scholar] [CrossRef]
  3. Li, S.; Fu, L.; Sun, Y.; Mu, Y.; Chen, L.; Li, J.; Gong, H. Cow Dataset. 2021. Available online: https://doi.org/10.6084/m9.figshare.16879780 (accessed on 3 November 2024).
  4. Andrew, W.; Greatwood, C.; Burghardt, T. Visual Localisation and Individual Identification of Holstein Friesian Cattle via Deep Learning. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 2850–2859. [Google Scholar] [CrossRef]
  5. Gao, J.; Burghardt, T.; Andrew, W.; Dowsey, A.W.; Campbell, N.W. Towards Self-Supervision for Video Identification of Individual Holstein-Friesian Cattle: The Cows2021 Dataset. arXiv 2021, arXiv:2105.01938. [Google Scholar]
  6. Andrew, W.; Gao, J.; Mullan, S.; Campbell, N.; Dowsey, A.W.; Burghardt, T. Visual identification of individual Holstein-Friesian cattle via deep metric learning. Comput. Electron. Agric. 2021, 185, 106133. [Google Scholar] [CrossRef]
  7. Zhang, T.; Zhao, Q.; Da, C.; Zhou, L.; Li, L.; Jiancuo, S. YakReID-103: A Benchmark for Yak reidentification. In Proceedings of the 2021 IEEE International Joint Conference on Biometrics (IJCB), Shenzhen, China, 4–7 August 2021; pp. 1–8. [Google Scholar] [CrossRef]
  8. Chan, J.; Carrión, H.; Mégret, R.; Rivera, J.L.A.; Giray, T. Honeybee reidentification in Video: New Datasets and Impact of Self-supervision. In Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2022)—Volume 5: VISAPP, INSTICC, Avenida de S. Francisco Xavier, Lote 7 Cv. C, Online, 6–8 February 2022; pp. 517–525. [Google Scholar] [CrossRef]
  9. Tausch, F.; Stock, S.; Fricke, J.; Klein, O. Bumblebee reidentification Dataset. In Proceedings of the 2020 IEEE Winter Applications of Computer Vision Workshops (WACVW), Snowmass Village, CO, USA, 1–5 March 2020; pp. 35–37. [Google Scholar] [CrossRef]
  10. Borlinghaus, P.; Tausch, F.; Rettenberger, L. A Purely Visual Re-ID Approach for Bumblebees (Bombus terrestris). Smart Agric. Technol. 2023, 3, 100135. [Google Scholar] [CrossRef]
  11. Kulits, P.; Wall, J.; Bedetti, A.; Henley, M.; Beery, S. ElephantBook: A Semi-Automated Human-in-the-Loop System for Elephant reidentification. In Proceedings of the 4th ACM SIGCAS Conference on Computing and Sustainable Societies, New York, NY, USA, 28 June–2 July 2021; pp. 88–98. [Google Scholar] [CrossRef]
  12. Moskvyak, O.; Maire, F.; Dayoub, F.; Armstrong, A.O.; Baktashmotlagh, M. Robust reidentification of Manta Rays from Natural Markings by Learning Pose Invariant Embeddings. In Proceedings of the 2021 Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, Australia, 29 November–1 December 2021; pp. 1–8. [Google Scholar] [CrossRef]
  13. Wang, L.; Ding, R.; Zhai, Y.; Zhang, Q.; Tang, W.; Zheng, N.; Hua, G. Giant Panda Identification. IEEE Trans. Image Process. 2021, 30, 2837–2849. [Google Scholar] [CrossRef]
  14. He, Q.; Zhao, Q.; Liu, N.; Chen, P.; Zhang, Z.; Hou, R. Distinguishing Individual Red Pandas from Their Faces. In Pattern Recognition and Computer Vision; Lin, Z., Wang, L., Yang, J., Shi, G., Tan, T., Zheng, N., Chen, X., Zhang, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 714–724. [Google Scholar]
  15. Vidal, M.; Wolf, N.; Rosenberg, B.; Harris, B.P.; Mathis, A. Perspectives on Individual Animal Identification from Biology and Computer Vision. Integr. Comp. Biol. 2021, 61, 900–916. [Google Scholar] [CrossRef]
  16. Neethirajan, S. ChickTrack—A quantitative tracking tool for measuring chicken activity. Measurement 2022, 191, 110819. [Google Scholar] [CrossRef]
  17. T. Psota, E.; Schmidt, T.; Mote, B.; C. Pérez, L. Long-Term Tracking of Group-Housed Livestock Using Keypoint Detection and MAP Estimation for Individual Animal Identification. Sensors 2020, 20, 3670. [Google Scholar] [CrossRef]
  18. Li, N.; Ren, Z.; Li, D.; Zeng, L. Review: Automated techniques for monitoring the behaviour and welfare of broilers and laying hens: Towards the goal of precision livestock farming. Animal 2020, 14, 617–625. [Google Scholar] [CrossRef]
  19. Food Traceability. European Commission—B-1049 Brussels. 2007. Available online: https://food.ec.europa.eu/system/files/2016-10/gfl_req_factsheet_traceability_2007_en.pdf (accessed on 13 August 2024).
  20. Dennis, R.L.; Fahey, A.G.; Cheng, H.W. Different Effects of Individual Identification Systems on Chicken Well-Being1. Poult. Sci. 2008, 87, 1052–1057. [Google Scholar] [CrossRef]
  21. Anderson, G.; Johnson, A.; Arguelles-Ramos, M.; Ali, A. Impact of Body-worn Sensors on Broiler Chicken Behavior and Agonistic Interactions. J. Appl. Anim. Welf. Sci. 2023, 1–10. [Google Scholar] [CrossRef] [PubMed]
  22. Stadig, L.M.; Rodenburg, T.B.; Ampe, B.; Reubens, B.; Tuyttens, F.A. An automated positioning system for monitoring chickens’ location: Effects of wearing a backpack on behaviour, leg health and production. Appl. Anim. Behav. Sci. 2018, 198, 83–88. [Google Scholar] [CrossRef]
  23. Marino, L. Thinking chickens: A review of cognition, emotion, and behavior in the domestic chicken. Anim. Cogn. 2017, 20, 127–147. [Google Scholar] [CrossRef] [PubMed]
  24. Lu, W.; Zhao, Y.; Wang, J.; Zheng, Z.; Feng, L.; Tang, J. MammalClub: An Annotated Wild Mammal Dataset for Species Recognition, Individual Identification, and Behavior Recognition. Electronics 2023, 12, 4506. [Google Scholar] [CrossRef]
  25. Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C.H. Deep Learning for Person reidentification: A Survey and Outlook. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2872–2893. [Google Scholar] [CrossRef]
  26. Witham, C.L. Automated face recognition of rhesus macaques. J. Neurosci. Methods 2018, 300, 157–165. [Google Scholar] [CrossRef]
  27. Freytag, A.; Rodner, E.; Simon, M.; Loos, A.; Kühl, H.S.; Denzler, J. Chimpanzee Faces in the Wild: Log-Euclidean CNNs for Predicting Identities and Attributes of Primates. In Pattern Recognition; Rosenhahn, B., Andres, B., Eds.; Springer: Cham, Switzerland, 2016; pp. 51–63. [Google Scholar]
  28. Lin, T.Y.; Kuo, Y.F. Cat Face Recognition Using Deep Learning; American Society of Agricultural and Biological Engineers (ASABE) Annual International Meeting: St. Joseph, MI, USA, 2018. [Google Scholar] [CrossRef]
  29. Dlamini, N.; Zyl, T.L.v. Automated Identification of Individuals in Wildlife Population Using Siamese Neural Networks. In Proceedings of the 2020 7th International Conference on Soft Computing and Machine Intelligence (ISCMI), Stockholm, Sweden, 14–15 November 2020; pp. 224–228. [Google Scholar] [CrossRef]
  30. Mougeot, G.; Li, D.; Jia, S. A Deep Learning Approach for Dog Face Verification and Recognition. In Proceedings of the PRICAI 2019: Trends in Artificial Intelligence, Cuvu, Fiji, 26–30 August 2019; Nayak, A.C., Sharma, A., Eds.; Springer International Publishing: Cham, Swtizerland, 2019; pp. 418–430. [Google Scholar] [CrossRef]
  31. Lamping, C.; Kootstra, G.; Derks, M. Transformer-Based Similarity Learning for Re-Identification of Chickens. Available online: https://doi.org/10.2139/ssrn.4886408 (accessed on 22 October 2024).
  32. Li, S.; Li, J.; Tang, H.; Qian, R.; Lin, W. ATRW: A Benchmark for Amur Tiger Re-identification in the Wild. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2590–2598. [Google Scholar] [CrossRef]
  33. Parham, J.; Crall, J.; Stewart, C.; Berger-Wolf, T.; Rubenstein, D. Animal population censusing at scale with citizen science and photographic identification. Aaai Spring Symp. Tech. Rep. 2017, SS-17-01–SS-17-08, 37–44. [Google Scholar]
  34. Haurum, J.B.; Karpova, A.; Pedersen, M.; Bengtson, S.H.; Moeslund, T.B. reidentification of Zebrafish using Metric Learning. In Proceedings of the 2020 IEEE Winter Applications of Computer Vision Workshops (WACVW), Snowmass Village, CO, USA, 1–5 March 2020; pp. 1–11. [Google Scholar] [CrossRef]
  35. Lahiri, M.; Tantipathananandh, C.; Warungu, R.; Rubenstein, D.I.; Berger-Wolf, T.Y. Biometric animal databases from field photographs: Identification of individual zebra in the wild. In Proceedings of the 1st ACM International Conference on Multimedia Retrieval, Vancouver, BC, Canada, 30–31 October 2008. [Google Scholar] [CrossRef]
  36. Adam, L.; Čermák, V.; Papafitsoros, K.; Picek, L. SeaTurtleID2022: A long-span dataset for reliable sea turtle reidentification. In Proceedings of the 2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 7131–7141. [Google Scholar] [CrossRef]
  37. Bouma, S.; Pawley, M.D.; Hupman, K.; Gilman, A. Individual Common Dolphin Identification via Metric Embedding Learning. In Proceedings of the 2018 International Conference on Image and Vision Computing New Zealand (IVCNZ), Auckland, New Zealand, 19–21 November 2018; pp. 1–6. [Google Scholar] [CrossRef]
  38. Hughes, B.; Burghardt, T. Automated Visual Fin Identification of Individual Great White Sharks. Int. J. Comput. Vision 2017, 122, 542–557. [Google Scholar] [CrossRef]
  39. Bae, H.B.; Pak, D.; Lee, S. Dog Nose-Print Identification Using Deep Neural Networks. IEEE Access 2021, 9, 49141–49153. [Google Scholar] [CrossRef]
  40. Zuerl, M.; Dirauf, R.; Koeferl, F.; Steinlein, N.; Sueskind, J.; Zanca, D.; Brehm, I.; Fersen, L.v.; Eskofier, B. PolarBearVidID: A Video-Based reidentification Benchmark Dataset for Polar Bears. Animals 2023, 13, 801. [Google Scholar] [CrossRef]
  41. Clapham, M.; Miller, E.; Nguyen, M.; Darimont, C.T. Automated facial recognition for wildlife that lack unique markings: A deep learning approach for brown bears. Ecol. Evol. 2020, 10, 12883–12892. [Google Scholar] [CrossRef] [PubMed]
  42. Körschens, M.; Denzler, J. ELPephants: A Fine-Grained Dataset for Elephant reidentification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 263–270. [Google Scholar] [CrossRef]
  43. Čermák, V.; Picek, L.; Adam, L.; Papafitsoros, K. WildlifeDatasets: An Open-Source Toolkit for Animal reidentification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 5953–5963. [Google Scholar]
  44. Wahltinez, O.; Wahltinez, S.J. An open-source general purpose machine learning framework for individual animal reidentification using few-shot learning. Methods Ecol. Evol. 2024, 15, 373–387. [Google Scholar] [CrossRef]
  45. Guo, S.; Xu, P.; Miao, Q.; Shao, G.; Chapman, C.A.; Chen, X.; He, G.; Fang, D.; Zhang, H.; Sun, Y.; et al. Automatic Identification of Individual Primates with Deep Learning Techniques. iScience 2020, 23, 32. [Google Scholar] [CrossRef] [PubMed]
  46. Ferreira, A.C.; Silva, L.R.; Renna, F.; Brandl, H.B.; Renoult, J.P.; Farine, D.R.; Covas, R.; Doutrelant, C. Bird individualID. 2020. Available online: https://github.com/AndreCFerreira/Bird_individualID (accessed on 3 November 2024).
  47. Kuncheva, L.I.; Williams, F.; Hennessey, S.L.; Rodríguez, J.J. Animal-Identification-from-Video. 2022. Available online: https://github.com/LucyKuncheva/Animal-Identification-from-Video (accessed on 3 November 2024).
  48. Adam, L.; Čermák, V.; Papafitsoros, K.; Picek, L. SeaTurtleID. 2022. Available online: https://www.kaggle.com/datasets/wildlifedatasets/seaturtleid2022 (accessed on 3 November 2024).
  49. Wahltinez, O. Sea Star Re-ID. 2023. Available online: https://lila.science/sea-star-re-id-2023/ (accessed on 3 November 2024).
  50. Me, W. Beluga ID. info@wildme.org. 2022. Available online: https://lila.science/datasets/beluga-id-2022/ (accessed on 3 November 2024).
  51. Cheeseman, T.; Southerland, K.; Reade, W.; Howard, A. Happywhale—Whale and Dolphin Identification. 2022. Available online: https://kaggle.com/competitions/happy-whale-and-dolphin (accessed on 3 November 2024).
  52. Nepovinnykh, E. SealID. Lappeenranta University of Technology, School of Engineering Science Yhteiset. 2022. Available online: https://doi.org/10.23729/0f4a3296-3b10-40c8-9ad3-0cf00a5a4a53 (accessed on 3 November 2024).
  53. Papafitsoros, K.; Adam, L.; Čermák, V.; Picek, L. SeaTurtleID. 2022. Available online: https://www.kaggle.com/datasets/wildlifedatasets/seaturtleidheads (accessed on 3 November 2024).
  54. Watch, W.T.; Conservation, L.O. Turtle Recall: Conservation Challenge. 2022. Available online: https://zindi.africa/competitions/turtle-recall-conservation-challenge/data (accessed on 3 November 2024).
  55. Trotter, C.; Atkinson, G.; Sharpe, M.; Richardson, K.; McGough, A.S.; Wright, N.; Burville, B.; Berggren, P. The Northumberland Dolphin Dataset 2020; Newcastle University: Newcastle upon Tyne, UK, 2020; Available online: https://doi.org/10.25405/data.ncl.c.4982342 (accessed on 3 November 2024).
  56. Humpback Whale Identification, Kaggle. 2018. Available online: https://kaggle.com/competitions/humpback-whale-identification (accessed on 3 November 2024).
  57. Khan, C.B.; Shashank; Kan, W. Right Whale Recognition, Kaggle. 2015. Available online: https://kaggle.com/competitions/noaa-right-whale-recognition (accessed on 3 November 2024).
  58. Holmberg, J.; Norman, B.; Arzoumanian, Z. Whale Shark ID. info@wildme.org. 2020. Available online: https://lila.science/datasets/whale-shark-id (accessed on 3 November 2024).
  59. Gao, J.; Burghardt, T.; Andrew, W.; Dowsey, A.W.; Campbell, N.W. Cows2021. 2021. Available online: https://doi.org/10.5523/bris.4vnrca7qw1642qlwxjadp87h7 (accessed on 3 November 2024).
  60. Andrew, W.; Gao, J.; Mullan, S.; Campbell, N.; Dowsey, A.W.; Burghardt, T. OpenCows2020. 2020. Available online: https://doi.org/10.5523/bris.10m32xl88x2b61zlkkgz3fml17 (accessed on 3 November 2024).
  61. Andrew, W.; Greatwood, C.; Burghardt, T. AerialCattle2017, University of Bristol. 2017. Available online: https://doi.org/10.5523/bris.3owflku95bxsx24643cybxu3qh (accessed on 3 November 2024).
  62. Andrew, W.; Greatwood, C.; Burghardt, T. FriesianCattle2017, University of Bristol. 2017. Available online: https://doi.org/10.5523/bris.2yizcfbkuv4352pzc32n54371r (accessed on 3 November 2024).
  63. Andrew, W.; Hannuna, S.; Campbell, N.; Burghardt, T. FriesianCattle2015, University of Bristol. 2016. Available online: https://doi.org/10.5523/bris.wurzq71kfm561ljahbwjhx9n3 (accessed on 3 November 2024).
  64. Kern, D.; Schiele, T.; Klauck, U.; Ingabire, W. Chicks4FreeID. 2024. Available online: https://huggingface.co/datasets/dariakern/Chicks4FreeID (accessed on 3 November 2024).
  65. Lu, W.; Zhao, Y.; Wang, J.; Zheng, Z.; Feng, L.; Tang, J. MammalClub. 2023. Available online: https://github.com/WJ-0425/MammalClub (accessed on 3 November 2024).
  66. He, Z.; Qian, J.; Yan, D.; Wang, C.; Xin, Y. Animal reidentification Algorithm for Posture Diversity. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  67. He, Z. Multi-Pose Dog Dataset. 2023. Mendeley Data, V1. Available online: https://data.mendeley.com/datasets/v5j6m8dzhv/1 (accessed on 3 November 2024).
  68. Dataset for: PolarBearVidID: A Video-Based Re-Identification Benchmark Dataset for Polar Bears. Available online: https://zenodo.org/records/7564529 (accessed on 3 November 2024).
  69. Kuncheva, L.I.; Williams, F.; Hennessey, S.L.; Rodríguez, J.J. A Benchmark Database for Animal reidentification and Tracking. In Proceedings of the 2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS), Genova, Italy, 5–7 December 2022; Volume 5, pp. 1–6. [Google Scholar] [CrossRef]
  70. Trust, B.P.C. Hyiena ID. Panthera Pardus CSV Custom Export. Retrieved from African Carnivore Wildbook 2022-04-28. 2022. Available online: https://lila.science/datasets/hyena-id-2022/ (accessed on 3 November 2024).
  71. Trust, B.P.C. Leopard ID. Panthera Pardus CSV Custom Export. Retrieved from African Carnivore Wildbook 2022-04-28. 2022. Available online: https://lila.science/datasets/leopard-id-2022/ (accessed on 3 November 2024).
  72. Nepovinnykh, E.; Eerola, T.; Biard, V.; Mutka, P.; Niemi, M.; Kunnasranta, M.; Kälviäinen, H. SealID: Saimaa Ringed Seal reidentification Dataset. Sensors 2022, 22, 7602. [Google Scholar] [CrossRef]
  73. Papafitsoros, K.; Adam, L.; Čermák, V.; Picek, L. SeaTurtleID: A novel long-span dataset highlighting the importance of timestamps in wildlife re-identification. arXiv 2022, arXiv:2211.10307. [Google Scholar] [CrossRef]
  74. Li, S.; Fu, L.; Sun, Y.; Mu, Y.; Chen, L.; Li, J.; Gong, H. Individual dairy cow identification based on lightweight convolutional neural network. PLoS ONE 2021, 16, 13. [Google Scholar] [CrossRef]
  75. Miele, V.; Dussert, G.; Spataro, B.; Chamaillé-Jammes, S.; Allainé, D.; Bonenfant, C. Revisiting animal photo-identification using deep metric learning and network analysis. Methods Ecol. Evol. 2021, 12, 863–873. [Google Scholar] [CrossRef]
  76. Miele, V.; Dussert, G.; Spataro, B.; Chamaillé-Jammes, S.; Allainé, D.; Bonenfant, C. Giraffe Dataset. 2020. Available online: https://plmlab.math.cnrs.fr/vmiele/animal-reid/ (accessed on 3 November 2024).
  77. Wang, L.; Ding, R.; Zhai, Y.; Zhang, Q.; Tang, W.; Zheng, N.; Hua, G. iPanda-50. 2021. Available online: https://github.com/iPandaDateset/iPanda-50 (accessed on 3 November 2024).
  78. Haurum, J.B.; Karpova, A.; Pedersen, M.; Bengtson, S.H.; Moeslund, T.B. AAU Zebrafish Re-Identification Dataset. 2020. Available online: https://www.kaggle.com/datasets/aalborguniversity/aau-zebrafish-reid (accessed on 3 November 2024).
  79. Guo, S.; Xu, P.; Miao, Q.; Shao, G.; Chapman, C.A.; Chen, X.; He, G.; Fang, D.; Zhang, H.; Sun, Y.; et al. AFD. Mendeley Data, Version 2. 2020. Available online: https://doi.org/10.17632/z3x59pv4bz.2 (accessed on 3 November 2024).
  80. Li, S.; Li, J.; Tang, H.; Qian, R.; Lin, W. ATRW (Amur Tiger Re-identification in the Wild). 2020. Available online: https://lila.science/datasets/atrw (accessed on 3 November 2024).
  81. Dlamini, N.; Zyl, T.L.v. Lion Face Dataset. Mara Masia Project, Kenya. 2020. Available online: https://github.com/tvanzyl/wildlife_reidentification/ (accessed on 3 November 2024).
  82. Trotter, C.; Atkinson, G.; Sharpe, M.; Richardson, K.; McGough, A.S.; Wright, N.; Burville, B.; Berggren, P. NDD20: A large-scale few-shot dolphin dataset for coarse and fine-grained categorisation. arXiv 2020, arXiv:2005.13359. [Google Scholar]
  83. Dlamini, N.; Zyl, T.L.v. Nyala Dataset. South African Nature Reserves. 2020. Available online: https://github.com/tvanzyl/wildlife_reidentification/ (accessed on 3 November 2024).
  84. Ferreira, A.C.; Silva, L.R.; Renna, F.; Brandl, H.B.; Renoult, J.P.; Farine, D.R.; Covas, R.; Doutrelant, C. Deep learning-based methods for individual recognition in small birds. Methods Ecol. Evol. 2020, 11, 1072–1085. [Google Scholar] [CrossRef]
  85. Mougeot, G.; Li, D.; Jia, S. Dog Face Dataset. 2019. Available online: https://github.com/GuillaumeMougeot/DogFaceNet (accessed on 3 November 2024).
  86. Lin, T.Y.; Kuo, Y.F. Cat Individual Images. 2018. Available online: https://www.kaggle.com/datasets/timost1234/cat-individuals (accessed on 3 November 2024).
  87. Schneider, J.; Murali, N.; Taylor, G.; Levine, J. Can Drosophila melanogaster tell who’s who? PLoS ONE 2018, 13, e0205043. [Google Scholar] [CrossRef]
  88. Dataset for: Can Drosophila Melanogaster Tell Who’s Who? Available online: https://borealisdata.ca/dataset.xhtml?persistentId=doi:10.5683/SP2/JP4WDF (accessed on 3 November 2024).
  89. Witham, C.L. MacaqueFaces. 2018. Available online: https://github.com/clwitham/MacaqueFaces (accessed on 3 November 2024).
  90. Parham, J.; Crall, J.; Stewart, C.; Berger-Wolf, T.; Rubenstein, D. Great Zebra and Giraffe Count ID. info@wildme.org. 2017. Available online: https://lila.science/datasets/great-zebra-giraffe-id (accessed on 3 November 2024).
  91. Freytag, A.; Rodner, E.; Simon, M.; Loos, A.; Kühl, H.S.; Denzler, J. Chimpanzee Faces in the Wild. Acknowledgements: Tobias Deschner, Laura Aporius, Karin Bahrke, Zoo Leipzig. 2016. Available online: https://github.com/cvjena/chimpanzee_faces (accessed on 3 November 2024).
  92. Holmberg, J.; Norman, B.; Arzoumanian, Z. Estimating population size, structure, and residency time for whale sharks Rhincodon typus through collaborative photo-identification. Endanger. Species Res. 2009, 7, 39–53. [Google Scholar] [CrossRef]
  93. Labelbox. “Labelbox”. 2024. Available online: https://labelbox.com (accessed on 3 November 2024).
  94. Cubuk, E.D.; Zoph, B.; Shlens, J.; Le, Q. RandAugment: Practical Automated Data Augmentation with a Reduced Search Space. In Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 18613–18624. [Google Scholar]
  95. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. 2021. Available online: https://openreview.net/forum?id=YicbFdNTTy (accessed on 8 May 2012).
  96. TorchVision: PyTorch’s Computer Vision Library. 2016. Available online: https://github.com/pytorch/vision (accessed on 3 November 2024).
  97. Lightly (Software Version 1.5.2). 2020. Available online: https://github.com/lightly-ai/lightly/blob/master/CITATION.cff (accessed on 3 November 2024).
  98. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
  99. Wightman, R. PyTorch Image Models. 2019. Available online: https://github.com/rwightman/pytorch-image-models (accessed on 3 November 2024).
  100. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. ArcFace: Additive Angular Margin Loss for Deep Face Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–19 June 2019. [Google Scholar]
  101. Čermák, V.; Picek, L.; Adam, L.; Papafitsoros, K. MegaDescriptor-L-384. 2024. Available online: https://huggingface.co/BVRA/MegaDescriptor-L-384 (accessed on 3 November 2024).
  102. Oquab, M.; Darcet, T.; Moutakanni, T.; Vo, H.; Szafraniec, M.; Khalidov, V.; Fernandez, P.; Haziza, D.; Massa, F.; El-Nouby, A.; et al. DINOv2: Learning Robust Visual Features without Supervision. arXiv 2024, arXiv:2304.07193. [Google Scholar]
  103. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models From Natural Language Supervision. PMLR 2021, 139, 8748–8763. [Google Scholar]
  104. Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised Feature Learning via Non-parametric Instance Discrimination. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3733–3742. [Google Scholar] [CrossRef]
  105. Detlefsen, N.S.; Borovec, J.; Schock, J.; Jha, A.H.; Koker, T.; Liello, L.D.; Stancl, D.; Quan, C.; Grechkin, M.; Falcon, W. TorchMetrics—Measuring Reproducibility in PyTorch. J. Open Source Softw. 2022, 7, 4101. [Google Scholar] [CrossRef]
  106. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the 2nd International Conference on Knowledge Discovery and Data Mining (KDD), Portland, OR, USA, 2–4 August 1996; pp. 226–231. [Google Scholar]
  107. Schubert, E.; Sander, J.; Ester, M.; Kriegel, H.P.; Xu, X. DBSCAN revisited, revisited: Why and how you should (still) use DBSCAN. Acm Trans. Database Syst. (Tods) 2017, 42, 19:1–19:21. [Google Scholar] [CrossRef]
  108. Nielsen, F. Introduction to HPC with MPI for Data Science; Hierarchical Clustering; Springer: Berlin/Heidelberg, Germany, 2016; Chapter 8; pp. 195–211. [Google Scholar]
Figure 1. Dataset overview.
Figure 1. Dataset overview.
Animals 15 00001 g001
Figure 2. Data preprocessing pipeline for subsequent re-ID.
Figure 2. Data preprocessing pipeline for subsequent re-ID.
Animals 15 00001 g002
Figure 3. Visibility distributions for all instances of each individual. Ducks and roosters are marked with an asterisk (*).
Figure 3. Visibility distributions for all instances of each individual. Ducks and roosters are marked with an asterisk (*).
Animals 15 00001 g003
Figure 4. Illustration of the training and evaluation process for the feature extractor and classifier, showcasing the linear classifier as an example in this workflow.
Figure 4. Illustration of the training and evaluation process for the feature extractor and classifier, showcasing the linear classifier as an example in this workflow.
Animals 15 00001 g004
Figure 5. Top-1 accuracy visualized in bar charts. The left bar chart combines the results in Table 2 and Table 3. The right bar chart illustrates the one-shot experiments in Table 4.
Figure 5. Top-1 accuracy visualized in bar charts. The left bar chart combines the results in Table 2 and Table 3. The right bar chart illustrates the one-shot experiments in Table 4.
Animals 15 00001 g005
Table 1. Publicly available animal re-ID datasets, arranged by date of publication. An asterisk (*) marks data derived from video footage. “n.a.” indicates that a corresponding publication could not be found.
Table 1. Publicly available animal re-ID datasets, arranged by date of publication. An asterisk (*) marks data derived from video footage. “n.a.” indicates that a corresponding publication could not be found.
YearPubl.DatasetIDsSpeciesAnnot.Avail. at
oursChicks4FreeID50, 2, 2chicken, duck, rooster1215, 40, 15[64]
2024[36]SeaTurtleID2022438sea turtle8729[48]
2023[24]Mammal Club (IISD)21811 terrestrial mammal species *33,612[65]
2023[66]Multi-pose dog dataset192dog1657[67]
2023[40]PolarBearVidID13polar bear *138,363[68]
2023[44]Sea Star Re-ID39, 56common starfish, Australian cushion star1204, 983[49]
2022[69]Animal-Identification-from-Video58, 26, 9pigeon *, pig *, Koi fish *12,671, 6184, 1635[47]
2022n.a.Beluga ID788beluga whale5902[50]
2022n.a.Happywhale15,58730 different species of whales and dolphins51,033[51]
2022n.a.Hyiena ID256spotted hyena3129[70]
2022n.a.Leopard ID430African leopard6805[71]
2022[72]SealID57Saimaa ringed seal2080[52]
2022[73]SeaTurtleIDHeads400sea turtle7774[53]
2022n.a.Turtle Recall100sea turtle2145[54]
2021[74]Cow Dataset13cow3772[3]
2021[5]Cows2021182Holstein-Friesian cattle *13,784[59]
2021[75]Giraffe Dataset62giraffe624[76]
2021[13]iPanda-5050giant panda6874[77]
2020[34]AAU Zebrafish Dataset6zebrafish *6672[78]
2020[45]Animal Face Dataset104041 primate species102,399[79]
2020[32]ATRW92Amur tiger *3649[80]
2020[29]Lion Face Dataset94lion740[81]
2020[82]NDD2044, 82bottlenose and white-beaked dolphin, white-beaked dolphin (underwater) *2201, 2201[55]
2020[29]Nyala Data237nyala1942[83]
2020[6]OpenCows202046Holstein-Friesian cattle *4736[60]
2019[84]Bird individualID30, 10, 10sociable weaver, great tit, zebra finch51,934[46]
2019[30]Dog Face Dataset1393dog8363[85]
2018[28]Cat Individual Images518cat13,536[86]
2018[87]Fruit Fly Dataset60fruit fly *2,592,000[88]
2018n.a.HumpbackWhaleID5004humpback whale15,697[56]
2018[26]MacaqueFaces34rhesus macaque *6280[89]
2017[4]AerialCattle201723Holstein-Friesian cattle *46,340[61]
2017[4]FriesianCattle201789Holstein-Friesian cattle *940[62]
2017[33]GZGC2056plains zebra and Masai giraffe6925[90]
2016[27]C-Tai78chimpanzee5078[91]
2016[27]C-Zoo24chimpanzee2109[91]
2016[2]FriesianCattle201540Holstein-Friesian cattle *377[63]
2015n.a.Right Whale Recognition447North Atlantic right whale4544[57]
2011[35]StripeSpotter45plains and Grevy’s zebra820[35]
2009[92]Whale Shark ID543whale shark7693[58]
Table 2. Domain transfer experiment results. The highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Table 2. Domain transfer experiment results. The highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Feature ExtractorTrainingEpochsClassifiermAPTop-1Top-5
MegaDescriptor [43]pretrained, frozen-k-NN 0.563 ± 0.011 0.609 ± 0.006 0.920 ± 0.025
MegaDescriptor [43]pretrained, frozen-linear 0.920 ± 0.008 0.818 ± 0.002 0.976 ± 0.003
MegaDescriptor [43]pretrained, fine-tuned200k-NN 0.835 ± 0.035 0.898 ± 0.026 0.976 ± 0.006
MegaDescriptor [43]pretrained, fine-tuned200linear 0.960   ±   0.009 0.916   ±   0.020 0.982   ±   0.007
Table 3. Results of standard supervised learning experiment. Highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Table 3. Results of standard supervised learning experiment. Highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Feature ExtractorTrainingEpochsClassifiermAPTop-1Top-5
Swin Transformer [98]from scratch200k-NN 0.728 ± 0.082 0.806 ± 0.060 0.966 ± 0.013
Swin Transformer [98]from scratch200linear 0.945 ± 0.019 0.884 ± 0.010 0.989 ± 0.004
Vision Transformer [95]from scratch200k-NN 0.923 ± 0.006 0.939 ± 0.006 1.000   ±   0.000
Vision Transformer [95]from scratch200linear 0.970   ±   0.014 0.951   ±   0.009 1.000   ±   0.000
Table 4. One-shot learning experiment results. Highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Table 4. One-shot learning experiment results. Highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Feature ExtractorTrainingEpochsClassifiermAPTop-1Top-5
MegaDescriptor [43]pretrained, frozen-k-NN 0.298 ± 0.035 0.341 ± 0.034 0.632 ± 0.037
MegaDescriptor [43]pretrained, frozen-linear 0.522 ± 0.013 0.473 ± 0.015 0.797 ± 0.014
MegaDescriptor [43]pretrained, fine-tuned200k-NN 0.464 ± 0.035 0.561 ± 0.034 0.782 ± 0.022
MegaDescriptor [43]pretrained, fine-tuned200linear 0.645   ±   0.027 0.562   ±   0.022 0.818   ±   0.036
Swin Transformer [98]from scratch200k-NN 0.289 ± 0.057 0.382 ± 0.058 0.620 ± 0.076
Swin Transformer [98]from scratch200linear 0.545 ± 0.039 0.492 ± 0.030 0.726 ± 0.015
Vision Transformer [95]from scratch200k-NN 0.387 ± 0.044 0.473 ± 0.051 0.785 ± 0.013
Vision Transformer [95]from scratch200linear 0.619 ± 0.020 0.523 ± 0.014 0.808 ± 0.012
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kern, D.; Schiele, T.; Klauck, U.; Ingabire, W. Towards Automated Chicken Monitoring: Dataset and Machine Learning Methods for Visual, Noninvasive Reidentification. Animals 2025, 15, 1. https://doi.org/10.3390/ani15010001

AMA Style

Kern D, Schiele T, Klauck U, Ingabire W. Towards Automated Chicken Monitoring: Dataset and Machine Learning Methods for Visual, Noninvasive Reidentification. Animals. 2025; 15(1):1. https://doi.org/10.3390/ani15010001

Chicago/Turabian Style

Kern, Daria, Tobias Schiele, Ulrich Klauck, and Winfred Ingabire. 2025. "Towards Automated Chicken Monitoring: Dataset and Machine Learning Methods for Visual, Noninvasive Reidentification" Animals 15, no. 1: 1. https://doi.org/10.3390/ani15010001

APA Style

Kern, D., Schiele, T., Klauck, U., & Ingabire, W. (2025). Towards Automated Chicken Monitoring: Dataset and Machine Learning Methods for Visual, Noninvasive Reidentification. Animals, 15(1), 1. https://doi.org/10.3390/ani15010001

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop