Next Article in Journal
Evaluation of Temporal Stability in Radiometric Calibration Network Sites Using Multi-Source Satellite Data and Continuous In Situ Measurements
Next Article in Special Issue
Empowering Wildlife Guardians: An Equitable Digital Stewardship and Reward System for Biodiversity Conservation Using Deep Learning and 3/4G Camera Traps
Previous Article in Journal
An Improved Parameter Estimation Method for High-Efficiency Multi-GNSS-Integrated Orbit Determination
Previous Article in Special Issue
Using Drones to Determine Chimpanzee Absences at the Edge of Their Distribution in Western Tanzania
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Removing Human Bottlenecks in Bird Classification Using Camera Trap Images and Deep Learning

1
School of Computer Science and Mathematics, Liverpool John Moores University, Byrom Street, Liverpool L3 3AF, UK
2
Department of Biosciences, Durham University, Durham DH1 3LE, UK
3
School of Mathematics and Statistics, University of St Andrews, St Andrews KY16 9SS, UK
4
Conservation and Research, Chester Zoo, Cheshire CH2 1LH, UK
5
Department of Biology, Faculty of Science, Utrecht University, 3584 Utrecht, The Netherlands
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(10), 2638; https://doi.org/10.3390/rs15102638
Submission received: 29 April 2023 / Revised: 12 May 2023 / Accepted: 15 May 2023 / Published: 18 May 2023
(This article belongs to the Special Issue Remote Sensing Applications to Ecology: Opportunities and Challenges)

Abstract

:
Birds are important indicators for monitoring both biodiversity and habitat health; they also play a crucial role in ecosystem management. Declines in bird populations can result in reduced ecosystem services, including seed dispersal, pollination and pest control. Accurate and long-term monitoring of birds to identify species of concern while measuring the success of conservation interventions is essential for ecologists. However, monitoring is time-consuming, costly and often difficult to manage over long durations and at meaningfully large spatial scales. Technology such as camera traps, acoustic monitors and drones provide methods for non-invasive monitoring. There are two main problems with using camera traps for monitoring: (a) cameras generate many images, making it difficult to process and analyse the data in a timely manner; and (b) the high proportion of false positives hinders the processing and analysis for reporting. In this paper, we outline an approach for overcoming these issues by utilising deep learning for real-time classification of bird species and automated removal of false positives in camera trap data. Images are classified in real-time using a Faster-RCNN architecture. Images are transmitted over 3/4G cameras and processed using Graphical Processing Units (GPUs) to provide conservationists with key detection metrics, thereby removing the requirement for manual observations. Our models achieved an average sensitivity of 88.79%, a specificity of 98.16% and accuracy of 96.71%. This demonstrates the effectiveness of using deep learning for automatic bird monitoring.

1. Introduction

The continuing loss of global biodiversity is a significant concern [1]. Birds are no exception to this phenomenon and a reduction in bird abundance can be a key indicator of adverse environmental change and declines in other species, such as insects [2]. Globally, many bird species have experienced significant population declines [3]. Several factors are implicated in this decline, including climate change [4], agricultural intensification, declining insect abundance and barriers to migration [5,6]. As a result, organisations and community science programs undertake comprehensive bird monitoring studies to ascertain population sizes while identifying the adverse effects of both natural and anthropogenic events [7].
Impacts on birds tend to vary geographically, and biodiversity in temperate and arctic regions is considered to be particularly vulnerable [8]. The long-term monitoring of populations is a critical component of successful conservation [9]. Given that many of the key drivers of decline operate at large scales, monitoring needs to take place over large geographical scales and for long durations. This process is often difficult, resource-intensive and costly [10]. In addition, it is estimated that over 85% of species-level information is provided by citizen science projects [11]. This introduces additional challenges due to variance in observers’ expertise and biases within monitoring protocols. This can result in both misclassification and inaccurate reporting [12,13]. As such, conservationists have resorted to the deployment of scalable and cost-effective monitoring technology to improve the modelling of bird populations.
Monitoring technologies such as camera traps and acoustic sensors enable the collection of important ecological data on abundance, distribution, and animal behaviour with-in ecosystems; this information can be used to stimulate and inform conservation strategies. Camera traps have seen increased usage in bird monitoring and can be complementary to other techniques, such as point counts [13] or acoustic monitoring. Camera traps provide multiple benefits, including unobtrusiveness, low cost, and the ability to conduct studies over large geographical scales and long observation periods. Figure 1 shows an example camera trap image of Passer domesticus (House sparrow) from the dataset used in this paper.
Advancements in camera trap technology, such as improved detection capabilities, extended battery life and wireless communication over 3/4G, have further extended their utility in ecology and conservation. However, their usage has introduced new challenges, including data management and storage, and the manual processing of large quantities of image data [14]. The number of camera traps used in a study can vary greatly, with the average deployment being 78 cameras [15]. As a result, some studies generate millions of images [16], causing significant bottlenecks in data processing and delaying or preventing the completion of studies [17]. The quantity of data collected through camera trap deployments is now one of the main limitations in monitoring studies. The data acquired, and its associated management, analysis and cataloguing requires significant resources to complete. These tasks often lag behind data acquisition, resulting in unused data, and reducing both the accuracy and effectiveness of surveys [18].
Here, we aim to address major challenges of camera trap data processing, through an automated object detection and classification pipeline capable of identifying and cataloguing a variety of different bird species using camera trap data. It is hoped that this approach will support large-scale and long-term bird surveys. The technology used in the approach aims to broaden observer participation by allowing citizen scientists and other community groups to participate in biodiversity monitoring. The current version of the pipeline can classify 10 different bird species using deep learning, although additional species could be included following the acquisition of adequate training data. The model is hosted on Conservation AI, which is a free platform that aims to facilitate the processing, management and storage of ecological data.
The remainder of the paper is structured as follows. In Section 2, we provide a background to current computer vision tools and their associated limitations, including those associated with automatic species classification. Section 3 details the proposed methodology, including the data description, pre-processing steps, hyperparameter selection and metrics used in the evaluation and inferencing of the trained model. The results for both model training and the deployment of the real-time inferencing trial are presented in Section 4. In Section 5, we discuss the results for both the training and model deployment. Section 6 provides a short conclusion, along with considerations for future work.

2. Background and Related Work

The development of computer vision tools for conservation applications is challenging and often impeded by several different factors. These include the availability of high-quality data for each species; high quality requires that data contain sufficient variation in both the species and environment (therefore supporting generalisation) and that the tagging process is accurate and robust. A wide variety of approaches utilise Machine Learning (ML) algorithms, such as Artificial Neural Networks (ANNs) for species classification within mammalian taxonomy. However, there is limited research within avian taxonomy, which is largely due to the limited utilisation of camera traps for bird surveys [19] and the complexity of species classification using ML approaches. As such, the automatic classification of different bird species has received much less interest among ML practitioners and conservationists. This leaves a significant gap in the tools available for processing data acquired from large-scale bird surveys. Additionally, it also limits the involvement of non-domain experts in bird monitoring surveys. The remainder of this section provides a discussion of some of the more common systems in operation today and highlights their limitations that we aim to address in this study.

2.1. Current Solutions

Historically, the identification of animal species within images has been undertaken manually by humans. However, there is now significant interest in fully or semi-automating this process. In existing approaches, researchers utilise both Deep Learning (DL) and non-DL methods to classify different animals from camera trap data. Most approaches simply detect the presence or absence of an animal to remove blank images and reduce the number of images that require manual processing. However, as this approach does not facilitate species-level classification, practitioners are still required to review large amounts of filtered data. In addition, the approach does little to alleviate the amount of data that has to be stored and managed.
The Microsoft Mega Detector (MD) is one ML tool for removing images which contain no animal objects. In [20], a generic detector was trained using a Faster-RCNN to classify animals, people and vehicles within camera trap data. The model uses the InceptionResNetv2 base network and the TensorFlow Object Detection API. The paper does not report any reliable performance metrics; however, the model has been independently assessed using camera trap data [21,22]. The results showed that it was possible to achieve precision = 98% and a recall = 93% using a Confidence Threshold (CT) of 0.65 for the Animal class. The results showed that reducing the CT to 0.1 increased the recall for the Animal class to 97%. However, this decreased precision to 95%, thereby reducing the number of false positives. The researchers noted that one of the biggest limitations of MD was its inability to classify individual species. While the MD is a useful tool, it has little benefit for bird surveys.
Other platforms, such as Wildlife Insights (WI) (https://www.wildlifeinsights.org/, accessed on 7 March 2023), do perform species-level classifications of mammals [23]. WI uses DL to detect 800 different animal species using the EfficientNet Architecture. The results show that the model has recall = 54% using CT 0.65 across all species, indicating a high false negative rate [21]. As a result, low recall values require expert review, leading to a process that is only semi-automated. WI provides a comprehensive platform for processing camera trap data. However, like the MD, it has limited utility for bird surveys. In addition, the platform does not facilitate the processing of real-time camera trap data; manual retrieval of data from the SD cards, processing the acquired data and manually uploading it to the WI platform constitute a significant overhead.
Several DL architectures are proposed in [23] to identify the presence or absence of a bird within images. The models were trained with 1200 images with a 125 × 125-pixel resolution, far lower than the resolution used in camera traps. The paper reported that the best-performing model achieved an accuracy of 95.5% after training the model over 60 epochs. No further evaluation metrics were reported, while no CTs were used to determine the precision and recall. As such, a full assessment of the model’s performance is lacking, limiting real-world deployment. Researchers in [24] proposed an approach that utilised sub-category recognition to derive local image features which are later used for classification. The evaluation was performed on the CUB-200-2011 dataset [25]. While the results showed an improvement over existing approaches, only accuracy was reported, and a full evaluation of the model and its inference performance is still required.
Other non-DL approaches have been used for bird classification. In [26], a Histogram of Oriented Gradients (HOG) was combined with Center-Symmetric Local Binary Pattern (CS-LBP) to generate a feature set, which was later classified using a Support Vector Machine (SVM). The classifier was trained to identify a single class (Crow) using a fixed resolution of 48 × 64 pixels, again far lower than the standard camera trap resolution. The paper reported an accuracy of 87% but no other metrics were provided. In addition, the absence of any additional classes means that the model has limited utility in bird surveys.

2.2. Limitations

The literature covers a wide variety of solutions for the automatic removal of blank images. However, very few aim to perform individual species classification due to similarities between species and the complex features that separate them. This is particularly true within avian taxonomy. Where solutions do exist, they are often impractical for real-world inference and do not consider the technical aspects of camera trap technology or the variety of environments in which they are deployed.
Convolutional Neural Network (CNN) approaches require a large corpus of high-quality labelled data that is reflective of the environment (or range of environments) in which the model is deployed. Most of the existing bird datasets used for modelling contain low-resolution images or cropped segments of the individual bird. This acts as a denoiser which can often attain high levels of performance during training but performs inadequately during real-world inference. The typical resolution for most camera traps is 1024 × 768. This means that existing models would experience significant drops in performance using larger-resolution images, as they are compressed or rescaled by the network. This results in extreme distortion of key features that describe the animal in question. Existing research focuses on the identification of a limited number of classes. This approach further reduces the practicality of existing models, as many species will go undetected or will likely be misclassified as an alternate species.
Another major challenge to consider is the deployment and automated inference of camera trap data. Large camera trap deployments generate significant amounts of data. This often impedes the study and prevents timely data analysis. The deployment of ML models, either centrally or on the edge, has not been sufficiently reported in the literature. The centralisation of inferencing will require the use of in-field communication, such as the Global System for Mobile Communications (GSM). While centralising models for inference has many benefits, it requires enhanced hardware to run complex model architectures. Many environments also lack access to GSM. In the remainder of this paper, we will discuss these limitations further and provide a first-step approach for the automatic classification of bird species using DL.

3. Materials and Methods

Here, we describe the dataset used in the study, along with the modelling approach taken and the metrics used to evaluate the trained model. We also discuss the data-tagging and pre-processing stages, including the proposed and associated technologies.

3.1. Data Collection and Description

The image dataset contains images of 10 distinct bird species found in Europe: Carduelis carduelis (European goldfinch), Chloris chloris (European greenfinch), Columba palumbus (Common wood pigeon), Corvus corone (Carrion crow), Cyanistes caeruleus (Eurasian blue tit), Erithacus rubecula (European robin), Fringilla coelebs (Common chaffinch), Garrulus glandarius (Eurasian jay), Passer domesticus (House sparrow) and Pica pica (Eurasian magpie) which were obtained through the iNaturalist website (https://www.inaturalist.org/, accessed on 15 May 2021) and private camera deployments.
In total, the dataset contained 32,982 image files. The mean resolution for the dataset was 972 × 769 pixels, conforming to the input resolution of most camera trap hardware. The overall resolution distribution for the entire dataset demonstrates no significant outliers that could negatively impact the training or inferencing of the model (Figure 2). In this case, no images were removed before tagging. Undertaking input resolution analysis of the dataset informs the aspect ratio coefficient used in the hyperparameter configuration for training the model.

3.2. Data Pre-Processing

The tagging of the data was undertaken using the Conservation AI tagging site. Bounding boxes are placed around objects to identify regions of interest. All tagged regions in each image were exported as Extensible Markup Language (XML) in the TensorFlow Pascal VOC format. Images considered to be of poor quality or unsuitable for model training were labelled as no good and systematically removed from the final dataset. The number of images per species varied from ~3000 to ~3750, constituting a slight class imbalance (Figure 3). The total number of tagged objects equated to 34,970.
Using a Python script, the images are randomly divided into 90% train and 10% validation based on the tagged labels. Using TensorFlow 2.2 and Pillow, the XML and associated images are converted into TFRecords (format for storing a sequence of binary records) for model training.

3.3. Model Selection

The Faster-RCNN is used to undertake both classification and object detection on images that contain the 10 different bird species. The Faster-RCNN performs object detection in two distinct stages [27]. The first stage uses a region proposal network to identify and extract regions of interest. This allows the model to estimate bounding box locations. The second stage conducts further processing to adjust the localisation of the bounding box by minimising the selected loss function and undertakes classification. Both the region proposal and object-detection tasks are undertaken using the same CNN. Figure 4 shows the Faster-RCNN architecture. For a complete description of the network, please refer to [28].

3.4. Transfer Learning

Transfer learning facilitates the adaptation of a pre-trained model and fine-tunes its learned parameters to support new objects of interest. This is an important technique because training CNNs on small datasets leads to poor performance due to low variance and reduced feature representation. Therefore, building on a much larger robust model that has been trained using a large corpus of images helps to counteract the issues associated with the low sample size used in this study. Transfer learning significantly reduces the amount of data required to train a model while maintaining a sufficient level of accuracy. This is an important aspect of this work that enables the development of such solutions to become attainable to a broad range of users with limited data, hardware, and financial support. The base model adopted for the transfer learning task is the Faster-RCNN Resnet 101 model, which has been pre-trained on the Common Objects in Context (COCO) dataset [29]. COCO is a large object-detection dataset containing 330 thousand images and 1.5 million object instances.

3.5. Modelling

Model training was conducted on an HP ProLiant ML 350 Gen. 9 server with 2x Intel Xeon E5-2640 v4 series processors and 768 GB of RAM. An additional GPU stack comprising six Nvidia Quadro RTX8000 graphics cards with a combined total of 288 GB of DDR5 RAM was installed. TensorFlow 2.2.0, CUDA 10.2 and cuDNN version 7.6 formed the software aspects of the training pipeline. In the pipeline.config file used by TensorFlow, the following training parameters were set:
  • To maintain both the resolution and aspect ratio resizer, minimum and maximum coefficients are set to 1024 × 1024 pixels, respectively. This minimises the scaling effect on the acquired data.
  • The default setting for the feature extractor coefficient is retrained to provide a standard 16-pixel stride length to retain the model features and improve training time.
  • The batch size coefficient is set to thirty-two, providing a balanced weight update without exceeding GPU memory limits.
  • The learning rate is set to 0.0004 to prevent large variations in response to the error.
To improve the generalisation of the model and prevent overfitting, the following augmentation techniques were applied (please note that the use of augmentation did not result in a larger dataset; images were randomly sampled during training):
  • Random_adjust_hue, which adjusts the hue of an image using a random factor.
  • Random_adjust_contrast, which adjusts the contrast of an image by a random factor.
  • Random_adjust_saturation, which adjusts the saturation of an image by a random factor.
  • Random_square_crop_by_scale, which was set with a scale_min of 0.6 and a scale_max of 1.3.
The Adam optimiser is implemented in ResNet 101 to minimise the loss function [30]. Unlike other optimisers such as Stochastic Gradient Decent (SGD), which maintains a single learning rate (alpha) throughout the entire training session, Adam calculates the moving average of the gradient m t /squared gradients v t based on the current gradient value g, and the parameters β 1 and β 2 to dynamically adjust the learning rate. Adam is described in [31] and is defined as:
m t = β 1 m t 1 + 1 β 1 g t
v t = β 2 v t 1 + 1 β 2 g t 2
where m t and v t are the estimates of the first and second moments of the gradient, respectively. Both m t and v t are initialised with 0. Biases are corrected by computing the first and second moment estimates [31] defined as:
m ^ t = m t 1 β 1 t
v ^ t = v t 1 β 2 t
Parameters are updated using the Adam update rule:
θ t + 1 = θ t n v ^ t + ϵ m ^ t .
The Rectified Linear Unit (ReLU) activation function was adopted during training to provide improvements over other functions, such as sigmoid or hyperbolic tangent (tanh) activations, which suffer from saturation changes around the mid-point of their input, which reduces the amount of available tuning. The use of these in deep multi-layered networks results in ineffective training caused by a vanishing gradient [32]. ReLU is defined as [33]:
g x = m a x 0 , x

3.6. Model Inferencing

The trained model was frozen and hosted using TensorFlow 2.5 and served through a public facing website developed by the authors. CUDA 11.2 and cuDNN 7.6.5 enable the GPU-accelerated inferencing aspect of the pipeline. Inferencing is undertaken on a custom-built server containing an Intel Xeon E5-1630v3 CPU, 256 GB of RAM and a NVidia Quadro RTX 8000 GPU. A NVidia Triton server 22.08 running in Docker with a Windows Subsystem for Linux 2 (WSL2) provided the software stack for inferencing the saved model. Due to the high specification of the GPU and the complexity of the model architecture, no model optimisation such as quantization or pruning was used.
A combination of both WIFI and ¾G cellular cameras was deployed near a set of eight birdfeeders in the UK and the Netherlands, as previously shown in Figure 1. The camera resolution was configured to 1920 × 1072 pixels with a Dots Per Inch (dpi) of 96. This configuration closely matches the aspect ratio resizer set in the pipeline.config used during training. The IR sensor was set to medium sensitivity, which when triggered automatically uploaded the acquired image to the platform using the Simple Mail Transfer Protocol (SMTP) for classification. Each camera was powered using a lithium battery (Battery-Li9, 7800 mAh) and recharged using solar panels throughout the duration of the study. Figure 5 shows one of the cameras used in the study.
The end-to-end inferencing pipeline (Figure 6) starts with the sensor and ends with the public facing Conservation AI site (https://www.conservationai.co.uk/, accessed on 5 March 2023) (Figure 7). Due to the use of standard protocols, the system can interface with a variety of different cameras for real-time inference. The SMTP server periodically polls the platforms camera trap email address and automatically schedules the images for classification using the Triton Server RESTAPI. Images are processed as they are received from the real-time camera traps. Where in-field communication is unavailable, images files can be batch-uploaded through the website, desktop client or REST API for offline inferencing.
The inferencing pipeline (Figure 8) begins with the transmission of acquired image files and associated meta data (Camera ID, image date/time) over 3/4G using SMTP. Images and payload data are automatically extracted from the email body using the imaplib library and sent for classification. Once the image is received by the inference server, it is queued for classification. The classified image and associated probability scores are returned and logged to a MySQL database against the user’s account, if the probability of a given prediction exceeds 50% confidence. If the probability score is below 50%, the image is classed as a blank (no animal present). Real-time alerts can also be configured, based on the detected species and the classification probability.

3.7. Evaluation Metrics Training

The performance of the trained model was evaluated using: RPNLoss/objectiveness to measure the model’s ability to generate suitable bounding boxes and categorise them as either a foreground or background object; RPNLoss/localisation to measure how well the RPN performs at generating bounding box regressor coordinates for foreground objects; BoxClassifierLoss/classification to measure the output layer/final classifier loss and describe the computed error for prediction and BoxClassifierLoss/localisation to measure the performance of the bounding box regressor. All these measures are combined to produce a TotalLoss metric. These metrics were obtained from TensorBoard 2.6 during training. The validation data was evaluated during training to derive the Mean Average Precision (mAP), which is the standard method for evaluating object detection models. mAP is defined as:
m A P = q = 1 Q A v e P q Q
where Q is the number of queries in the set and AveP(q) is the Average Precision (AP) for a given query q. The mAP is calculated on the bounding box locations for the final two checkpoints. Two Intersection over Union (IOU) thresholds, 0.50 and 0.75, are used to assess the overall performance of the model. The IOU is a useful metric for determining the accuracy of the bounding box location. This is achieved by measuring the percentage ratio of the overlap between the predicted bounding box and the ground truth bounding box, defined as:
I o U = A r e a   o f   O v e r l a p A r e a   o f   U n i o n
A threshold of @0.50 measures the overall detection accuracy of the model while the upper threshold of @0.75 measures localisation accuracy.

3.8. Evaluation Metrics Inference

Using the trained model, performance for inference conducted during the trial is evaluated by analysing images transmitted from the real-time cameras. It is important to undertake post-training analysis, as model performance can significantly differ once they are deployed to work on real-world tasks. Precision, Sensitivity (recall), Specificity, F-1 Score and Accuracy are used to evaluate the model’s performance on inference data. True Positive (TP), False Positive (FP), True Negative (TN) and False Negative (FN) are used to derive the different metrics. Precision is defined as:
P r e c i s i o n = T P T P + F P
Precision is used to ascertain the model’s ability to predict true positive detections. In the context of this study, it tells us that the animal detected and classified matched the ground truth while identifying any misclassifications in the image. Recall is defined as:
R e c a l l = T P T P + F N
Recall measures the proportion of true positives correctly identified by the model. In this context of object detection, it measures how well the model detected and correctly matched the ground truth labels. Recall also allows us to determine the number of false positives. The F-1 Score is defined as:
F 1   S c o r e = 2     P r e c i s i o n     S e n s i t i v i t y   ( R e c a l l ) P r e c i s i o n   +   S e n s i t i v i t y   ( R e c a l l )
The F-1 Score provides the harmonic means of both the precision and recall. A high F-1 Score means that the model has both high precision and recall. In the context of object detection, a high recall means that the model can accurately classify and localise objects in an image. Lastly, Accuracy is defined as:
A c c u r a c y = T P + T N T P + T N + F P + F N
Accuracy provides an overall assessment of the model’s ability to detect and correctly classify objects within the image, although this has limited bearing in unbalanced datasets and should therefore always be interpreted alongside other metrics defined in this section.

4. Results

We present the results in two parts: firstly, the training of the Faster-RCNN using the model configuration outlined in the methodology section; secondly, the inference results, which demonstrate the model’s performance in a real-world setting, using the 3/4G real-time cameras. Please note that the term ‘classes’ represents the individual bird species and therefore the terms ‘species’ and ‘classes’ are used interchangeably in this paper.

4.1. Training Results for Europe Birds Model

Using the tagged dataset, the data was randomly split into a training set, comprising 90% of the data (29,683 tags), and a validation set of 10% (3298 tags) to train and fine-tune the model. The model was trained over 58 epochs (30,000 steps) using a batch size of 32 to determine the best fit. During training, there was no overlap between the training and validation loss, indicating that no overfitting occurred.

4.1.1. RPN and Classifier Loss for Training Split

The model can detect candidate regions of interest with a high degree of accuracy (0.01325) while the RPN can sufficiently detect the localisation of an object (0.06868) (Table 1). This is encouraging, given the size of the tagged objects in relation to the background noise. The classification loss indicates that while the model can effectively localise an object, it is less accurate at correctly predicting the class (0.2977). The results of the box classifier localisation (0.05407) are comparable to the RPN results, affirming that the model can accurately identify candidate labels. The total loss (0.4337) combines the RPN and box classification values. The result demonstrates that the model’s predictions closely match the ground truth labels during training. While a total loss of 0.4337 is good for computer vision models, the model evidently found it difficult to classify some species of birds.

4.1.2. RPN and Classifier Loss for the Validation Split

The validation results are similar to the training results (Table 2). The loss for the RPN objectness (0.01212) is slightly lower than the training set, and the same is true for RPN Localisation (0.04031). Losses for the box classifier (0.2977) and box localisation (0.05407) are again slightly lower than the training metrics. The total loss (0.2828) is lower than the total loss for the training set. This suggests that the images contained within the validation split might have been slightly easier to localise and classify or that the training split contained greater levels of variance. There is no evidence that the model overfitted during training (the training loss does not increase while the validation continues to fall), although further training could have been undertaken to reduce the training loss further.

4.1.3. Precision and Recall Results for the Validation Split

Across all classes, the mAP (0.7856) shows that the model was able to detect and correctly identify the class (TP) on most occasions, with only 21.44% being false positives (FP) (Table 3). The mAP (Large), calculated on a subset of the dataset where the object was considered to be large, achieved 0.7902; hence the model was able to correctly identify the class when the object was close to the camera. This process was repeated for both mAP (Medium) and mAP (Small), for which the model’s performance dipped to 0.6296 and 0.4629, respectively. This indicates that the model struggles to correctly classify the species when the camera is further away from the object. Both [email protected] (0.9458) and [email protected] (0.8992) demonstrate that the model can effectively localise objects in the image and successfully map-predicted bounding boxes to the ground truth.

4.1.4. Precision and Recall Results for the Validation Split

The model can recall (Recall/AR@1) 81.94% of the objects when considering only the top-ranked detection for each image (Table 4). Similarly, Recall/AR@10 (0.8711) and Recall/AR@100 (0.8721) show the model’s performance when considering the top 10 and top 100 detections, respectively. The AR@100 (Large) can recall 85.57% of the large images, 70.53% of the medium images [AR@100 (Medium)] and 55.37% of the small images [AR@100 (Small)]. This underlines the model’s ability to correctly identify the species if it is closer to the camera, with performance decreasing significantly as objects get smaller (the bird is further away from the camera).

4.2. Trial Results for European Birds Model

The trained model was deployed to assess its performance during the trial period. Using eight different cameras and the inferencing pipeline outlined previously in the methodology section, the cameras transmitted a total of 42,973 images between 5 March 2021 and 28 February 2023. During the trial, 14,740 detections were made of the 10 species (Figure 9), and 28,233 blanks were recorded.
The 14,740 detections were used to generate ten inference evaluation subsets, each containing 25 randomly sampled images from each of the eight classes (excluding the Corvus corone and Fringilla coelebs classes, as these were insufficiently represented in the data). In addition, 25 blanks were randomly sampled and added to each fold to generate a total of 250 images per fold. A 5% margin of error and a confidence level of 95% was used to generate the inference evaluation dataset. The metrics were calculated for each of the 10 datasets and averaged to produce a final set for the model.

4.2.1. Performance Evaluation Results for Inference

The model archived a high level of accuracy across all classes, with an average range between 95.19% and 98.28%, and an overall accuracy of 96.71 across the 10 folds (Table 5). However, the model’s precision scores are more diverse. The precision ranges between 35.02% and 100%, with the Garrulus glandarius class having the lowest value. Overall, the model’s average precision (75.32) is good for an object-detection model. However, this indicates that there are issues with several classes (Carduelis carduelis, Chloris chloris and Garrulus glandarius). This requires further investigation of the training data and tagging process for these classes. The sensitivity of the model is high across all classes (between 76.67% and 95.24%). The worst-performing classes for sensitivity were Cyanistes caeruleus (76.67%) and Erithacus rubecula (85.73%), which need further investigation to ascertain the reason for the lower scores. However, it is likely a result of the model’s dependence on the colour of these two species and the data augmentation techniques applied during training. The average sensitivity of 88.79% across all folds indicates that the model can correctly identify a large proportion of actual positive cases. Specificity for the model is high across all classes (95.47% to 100%). The species with the lowest specificity was Chloris chloris; the highest was Erithacus rubecula. The model’s average specificity across all folds was 98.16%, indicating that the model was able to correctly identify most of the negative cases (distinguishing correctly between images that contain an object and images that are blank). The model’s F1-score varied across all the classes (between 47.61% and 93.49%). The species with the highest F1-score was Columba palumbus (92.51%); the lowest was Garrulus glandarius (47.61%). The average F1-score across the folds was 77.10%, which requires further investigation.

4.2.2. Confusion Matrix for Inference Data

Sources of confusion among the inference data across all folds (Table 6) largely align with the inference scores in Table 5. The model correctly predicts most of the classes with minimal error; the Columba palumbus, Chloris chloris and Passer domesticus classes performed the best. However, the model does misclassify examples across all the species, with Cyanistes caeruleus performing the worst. Again, this closely matches the results in Table 5 and is likely due to the model’s dependence on colour and the augmentation techniques used in training. These results highlight shortcomings in the model’s ability to accurately detect and classify some species which will need to be addressed in future training iterations.

5. Discussion

Bird monitoring provides an important biodiversity indicator but is a complex and time-consuming task, largely accessible only to skilled participants. Coupled with the need to analyse large volumes of data, this impedes the speed at which inferences can be used to inform interventions for the prevention of biodiversity loss. We demonstrate a practical alternative that uses deep learning and real-time 3/4G cameras for the automatic classification of bird species. The model was trained on 34,970 tags across 10 different bird species, using the Faster-RCNN architecture and transfer learning. The model allowed us to monitor eight different sites in both the UK and Netherlands over a two-year period.
During training, the model attained good results for the RPNLoss/objectness (0.01325) and RPNLoss/localisation (0.04031) for the training split. Both the validation RPNLoss/objectness and RPNLoss/localisation performed equally well, with a loss of 0.01212 and 0.04031, respectively. The BoxClassifierLoss/classification (0.2977) and BoxClassifierLoss/localisation (0.05407) for the training split performed reasonably well, although the higher classification loss indicates that the model struggled to correctly classify all species of birds. This was reflected in validation, during which the BoxClassifierLoss/classification and BoxClassifierLoss/localisation attained a loss of 0.2496 and 0.0121, respectively. By combining the metrics, the model was able to achieve a total loss of 0.4337 (train split) and 0.2828 (validation split), without overfitting. Given both the batch size and the number of tags used during training, the model would have benefited from a larger number of epochs.
The model performed well and attained good precision results on the validation split. The metrics demonstrate that the model can detect and localise objects with a high degree of accuracy. However, the model struggles with smaller and more distant objects, likely characterised by fewer features. This is reflected in the recall results for the validation split.
During the trial, the model inference performed well, achieving an average accuracy (96.71%), precision (75.32%), sensitivity (88.79%), specificity (98.16%) and an F1-score (77.10) across the 10 folds. However, the model struggled with some species, such as Chloris chloris and Carduelis carduelis, with average precisions of only 49.49% and 55.33%, respectively. Although the species was sometimes reasonably prominent within images (e.g., Figure 10), they were often distant from the cameras in the inference dataset. In contrast to other species, they also have a limited number of identifiable features at greater distances. However, the model is capable of detecting complex patterns when the bird is closer to the camera (e.g., Figure 11). It is probable that the model depends largely on the object’s colour to identify this class. The type of data augmentation used during training, during which colour was randomly altered to improve generalisation, will have compounded problems when the mode tries to identify this species.
Clearly, the model performs better with larger objects. With an average precision of 95.47%, performance for Columba palumbus was much better—likely a result of their large size and greater number of visible features (e.g., Figure 12 and Figure 13). The distance of the object from the camera clearly plays a significant role in the successful classification of a species. This means that careful consideration is needed when deploying cameras for small animals such as birds. Cameras must be deployed as close to the bird feeder as possible to increase the likelihood of a successful classification.
The inference results demonstrate that the model can detect an object with a high degree of accuracy. Likewise, it is also highly effective at removing images that do not contain a bird. During the trial, the model was able to remove blank images with an average accuracy of 98.28%. Practitioners should exercise caution when using the model for the classification of both Passer domesticus and Columba palumbus. The model will assign the closet matching class to species the model has not yet been trained on. It was noted during evaluation that the model sometimes mistook the Passer domesticus class for Prunella modularis (a similar-looking species). This is likely due to similarities between the species, occlusion (therefore reducing available features) and incorrect tagging in the training data. This is also true for Columba palumbus, where the model can predict different species within the same genus.
Overall, there has been limited research on the effectiveness of camera traps for monitoring small animals such as birds. The method could also be reduced if the camera fails to trigger, or the speed of the camera results in blurry or distorted images [34]. The inference test set contained several blurry images which negatively affected the results. This resulted from a low trigger speed on the deployed cameras. While this might have limited adverse effects on larger animals, distorted images can remove some of the detailed features of smaller animals which the model relies on for classification.

6. Conclusions

We presented an automatic bird classification system capable of identifying ten different species using both deep learning and real-time 3/4G cameras. This provides a possible solution to some of the issues associated with conducting large-scale bird monitoring projects, by reducing the requirements for domain expertise and the manual classification of camera trap data. Removing the bottlenecks in manual camera trap processing will speed up information provision and enable timely interventions for biodiversity. Using a transfer learning approach and the Faster-RCNN architecture, a model was trained and deployed in two countries to ascertain its effectiveness for monitoring bird populations. Images were transmitted in real-time and inferenced on the Conservation AI platform to automatically document the species of birds visiting each location. The results of the trial are encouraging and demonstrate that individual species can be detected with a high degree of accuracy. The model is also effective at accurately removing the large numbers of blank images that are transmitted because of motion detection. This is particularly important in bird studies, in which camera traps need to be set to higher levels of sensitivity due to the small size of the focal objects. By hosting the trained model on the Conservation AI platform, a unified and seamless process capable of automatically documenting bird biodiversity is provided. The solution should be seen as a foundational step that not only benefits domain experts but provides a gateway for anyone with an interest in contributing to the monitoring of bird populations. For example, it would be very easy to offer people the chance to deploy their own cameras in gardens, workspaces, and even communal parks to help broaden the bird-monitoring capabilities of organisations such as RSPB to better understand the biodiversity health of bird species in countries like the UK.
Our results identify important avenues for further research. While reducing the camera’s distance from the bird feeder will likely improve classification results for smaller bird species, alterations in the training data, tagging and hyperparameters are likely to yield greater improvements in accuracy. Removing data augmentation techniques should be trialled to yield significant gains in model performance. In addition, further work is required regarding the optimal camera setup and configuration for computer vision models in bird monitoring. Overall, we feel that the work demonstrates promising results and shows how deep learning and real-time inference can aid in the monitoring of birds, especially in those areas in which large numbers of skilled participants are unavailable. The model presented in this paper is available for use on the Conservation AI platform.

Author Contributions

Conceptualization, C.C. and P.F.; methodology, C.C. and P.F.; software, C.C. and P.F.; validation, C.C., P.F., S.W., S.N.L., N.D.W., P.A.S., C.S., N.M., J.M. and A.N.; formal analysis, C.C., P.F., S.W., P.A.S. and J.M.; investigation, C.C. and P.F.; resources, C.C. and P.F.; data curation, C.C. and P.F.; writing—original draft preparation, C.C. and P.F.; writing—review and editing, C.C., P.F., S.W., S.N.L., N.D.W., P.A.S., C.S., N.M., J.M. and A.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data is unavailable due to privacy and ethical restrictions.

Acknowledgments

The authors would like to thank ReoLink for donating the cameras and solar panels used in this study and Vodafone UK for sponsoring our communication. Finally, the authors would like to thank Rachel Chalmers for tagging all the data used in this study and for contributing so much to the Conservation AI platform.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mace, G.M.; Barrett, M.; Burgess, N.D.; Cornell, S.E.; Freeman, R.; Grooten, M.; Purvis, A. Aiming higher to bend the curve of biodiversity loss. Nat. Sustain. 2018, 1, 448–451. [Google Scholar] [CrossRef]
  2. Bowler, D.E.; Heldbjerg, H.; Fox, A.D.; de Jong, M.; Böhning-Gaese, K. Long-term declines of European insectivorous bird populations and potential causes. Conserv. Biol. 2019, 33, 1120–1130. [Google Scholar] [CrossRef]
  3. Peach, W.J.; Mallord, J.W.; Ockendon, N.; Orsman, C.J.; Haines, W.G. Depleted suburban house sparrow Passer domesticus population not limited by food availability. Urban Ecosyst. 2018, 21, 1053–1065. [Google Scholar] [CrossRef]
  4. Stephens, P.A.; Mason, L.R.; Green, R.E.; Gregory, R.D.; Sauer, J.R.; Alison, J.; Aunins, A.; Brotons, L.; Butchart, S.H.; Campedelli, T.; et al. Consistent response of bird populations to climate change on two continents. Science 2016, 352, 84–87. [Google Scholar] [CrossRef] [PubMed]
  5. Reif, J.; Vermouzek, Z. Collapse of farmland bird populations in an Eastern European country following its EU accession. Conserv. Lett. 2019, 12, e12585. [Google Scholar] [CrossRef]
  6. Wilson, S.; Saracco, J.F.; Krikun, R.; Flockhart, D.T.; Godwin, C.M.; Foster, K.R. Drivers of demographic decline across the annual cycle of a threatened migratory bird. Sci. Rep. 2018, 8, 7316. [Google Scholar] [CrossRef]
  7. Snäll, T.; Kindvall, O.; Nilsson, J.; Pärt, T. Evaluating citizen-based presence data for bird monitoring. Biol. Conserv. 2011, 144, 804–810. [Google Scholar] [CrossRef]
  8. Altdorff, D.; Borchard, N.; Young, E.H.; Galagedara, L.; Sorvali, J.; Quideau, S.; Unc, A. Agriculture in boreal and Arctic regions requires an integrated global approach for research and policy. Agron. Sustain. Dev. 2021, 41, 23. [Google Scholar] [CrossRef]
  9. Neate-Clegg, M.H.; Horns, J.J.; Adler, F.R.; Aytekin, M.Ç.K.; Şekercioğlu, Ç.H. Monitoring the world’s bird populations with community science data. Biol. Conserv. 2020, 248, 108653. [Google Scholar] [CrossRef]
  10. Stephenson, P.J. Technological advances in biodiversity monitoring: Applicability, opportunities and challenges. Curr. Opin. Environ. Sustain. 2020, 45, 36–41. [Google Scholar] [CrossRef]
  11. Pimm, S.L.; Alibhai, S.; Bergl, R.; Dehgan, A.; Giri, C.; Jewell, Z.; Joppa, L.; Kays, R.; Loarie, S. Emerging technologies to conserve biodiversity. Trends Ecol. Evol. 2015, 30, 685–696. [Google Scholar] [CrossRef] [PubMed]
  12. Pearce-Higgins, J.W.; Baillie, S.R.; Boughey, K.; Bourn, N.A.; Foppen, R.P.; Gillings, S.; Gregory, R.D.; Hunt, T.; Jiguet, F.; Lehikoinen, A.; et al. Overcoming the challenges of public data archiving for citizen science biodiversity recording and monitoring schemes. J. Appl. Ecol. 2018, 55, 2544–2551. [Google Scholar] [CrossRef]
  13. Fontúrbel, F.E.; Rodríguez-Gómez, G.B.; Fernández, N.; García, B.; Orellana, J.I.; Castaño-Villa, G.J. Sampling understory birds in different habitat types using point counts and camera traps. Ecol. Indic. 2020, 119, 106863. [Google Scholar] [CrossRef]
  14. Newey, S.; Davidson, P.; Nazir, S.; Fairhurst, G.; Verdicchio, F.; Irvine, R.J.; van der Wal, R. Limitations of recreational camera traps for wildlife management and conservation research: A practitioner’s perspective. Ambio 2015, 44, 624–635. [Google Scholar] [CrossRef] [PubMed]
  15. Young, S.; Rode-Margono, J.; Amin, R. Software to facilitate and streamline camera trap data management: A review. Ecol. Evol. 2018, 8, 9947–9957. [Google Scholar] [CrossRef] [PubMed]
  16. McShea, W.J.; Forrester, T.; Costello, R.; He, Z.; Kays, R. Volunteer-run cameras as distributed sensors for macrosystem mammal research. Landsc. Ecol. 2016, 31, 55–66. [Google Scholar] [CrossRef]
  17. Falzon, G.; Lawson, C.; Cheung, K.W.; Vernes, K.; Ballard, G.A.; Fleming, P.J.; Glen, A.S.; Milne, H.; Mather-Zardain, A.; Meek, P.D. ClassifyMe: A field-scouting software for the identification of wildlife in camera trap images. Animals 2019, 10, 58. [Google Scholar] [CrossRef] [PubMed]
  18. Harris, G.; Thompson, R.; Childs, J.L.; Sanderson, J.G. Automatic storage and analysis of camera trap data. Bull. Ecol. Soc. Am. 2010, 91, 352–360. [Google Scholar] [CrossRef]
  19. O’Brien, T.G.; Kinnaird, M.F. A picture is worth a thousand words: The application of camera trapping to the study of birds. Bird Conserv. Int. 2008, 18 (Suppl. S1), S144–S162. [Google Scholar] [CrossRef]
  20. Beery, S.; Morris, D.; Yang, S. Efficient pipeline for camera trap image review. arXiv 2019, arXiv:1907.06772. [Google Scholar]
  21. Vélez, J.; Castiblanco-Camacho, P.J.; Tabak, M.A.; Chalmers, C.; Fergus, P.; Fieberg, J. Choosing an Appropriate Platform and Workflow for Processing Camera Trap Data using Artificial Intelligence. arXiv 2022, arXiv:2202.02283. [Google Scholar]
  22. Fennell, M.J.; Beirne, C.W.; Burton, C. Use of object detection in camera trap image identification: Assessing a method to rapidly and accurately classify human and animal detections for research and application in recreation ecology. Glob. Ecol. Conserv. 2022, 35, e02104. [Google Scholar] [CrossRef]
  23. Mashuk, F.; Sattar, A.; Sultana, N. Machine Learning Approach for Bird Detection. In Proceedings of the 2021 Third International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV), Tirunelveli, India, 4–6 February 2021; IEEE: Piscataway, NJ, USA; pp. 818–822.
  24. Branson, S.; Van Horn, G.; Belongie, S.; Perona, P. Bird species categorization using pose normalized deep convolutional nets. arXiv 2014, arXiv:1406.2952. [Google Scholar]
  25. Wah, C.; Branson, S.; Welinder, P.; Perona, P.; Belongie, S. The Caltech-Ucsd Birds-200-2011 Dataset. 2011. Available online: https://authors.library.caltech.edu/27452/ (accessed on 5 March 2023).
  26. Mihreteab, K.; Iwahashi, M.; Yamamoto, M. Crow birds detection using HOG and CS-LBP. In Proceedings of the 2012 International Symposium on Intelligent Signal Processing and Communications Systems, Tamsui, Taiwan, 4–7 November 2012; IEEE: Piscataway, NJ, USA; pp. 406–409.
  27. Ren, S.; He, K.; Girshick, R.; Zhang, X.; Sun, J. Object detection networks on convolutional feature maps. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1476–1481. [Google Scholar] [CrossRef]
  28. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  29. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  30. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  31. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inf. Process. Syst. 2017, 30, 6629–6640. [Google Scholar]
  32. Kanai, S.; Fujiwara, Y.; Iwamura, S. Preventing gradient explosions in gated recurrent units. Adv. Neural Inf. Process. Syst. 2017, 30, 435–444. [Google Scholar]
  33. Ba, J.; Frey, B. Adaptive dropout for training deep neural networks. Adv. Neural Inf. Process. Syst. 2013, 26, 3084–3092. [Google Scholar]
  34. Ortmann, C.R.; Johnson, S.D. How reliable are motion-triggered camera traps for detecting small mammals and birds in ecological studies? J. Zool. 2021, 313, 202–207. [Google Scholar] [CrossRef]
Figure 1. Camera trap image of a Passer domesticus visiting one of the bird feeders used in the study.
Figure 1. Camera trap image of a Passer domesticus visiting one of the bird feeders used in the study.
Remotesensing 15 02638 g001
Figure 2. Image resolution distribution for the dataset. The resolution is shown for each image, highlighting any outliers that might affect model training.
Figure 2. Image resolution distribution for the dataset. The resolution is shown for each image, highlighting any outliers that might affect model training.
Remotesensing 15 02638 g002
Figure 3. Class distribution of tagged data showing the species name on the x axis and the number of tags per species on the y axis.
Figure 3. Class distribution of tagged data showing the species name on the x axis and the number of tags per species on the y axis.
Remotesensing 15 02638 g003
Figure 4. Faster RCNN architecture showing the CNN layers, Feature Map, RPN, RoI Pooling and fully connected layers.
Figure 4. Faster RCNN architecture showing the CNN layers, Feature Map, RPN, RoI Pooling and fully connected layers.
Remotesensing 15 02638 g004
Figure 5. 3/4G camera trap used for real-time inference. The camera was deployed near a bird feeder. Throughout the study the camera trap was continuously charged using a solar panel.
Figure 5. 3/4G camera trap used for real-time inference. The camera was deployed near a bird feeder. Throughout the study the camera trap was continuously charged using a solar panel.
Remotesensing 15 02638 g005
Figure 6. End-to-end inferencing pipeline for the Conservation AI platform.
Figure 6. End-to-end inferencing pipeline for the Conservation AI platform.
Remotesensing 15 02638 g006
Figure 7. Conservation AI platform (https://www.concervationai.co.uk, accessed on 5 March 2023).
Figure 7. Conservation AI platform (https://www.concervationai.co.uk, accessed on 5 March 2023).
Remotesensing 15 02638 g007
Figure 8. Inference processing pipeline.
Figure 8. Inference processing pipeline.
Remotesensing 15 02638 g008
Figure 9. Number of detections by species during the study duration.
Figure 9. Number of detections by species during the study duration.
Remotesensing 15 02638 g009
Figure 10. Example detections of Chloris chloris during the inference trial using one of the 3/4G cameras.
Figure 10. Example detections of Chloris chloris during the inference trial using one of the 3/4G cameras.
Remotesensing 15 02638 g010
Figure 11. Example detection of a Erithacus rubecula during the inference trial using one of the 3/4G cameras.
Figure 11. Example detection of a Erithacus rubecula during the inference trial using one of the 3/4G cameras.
Remotesensing 15 02638 g011
Figure 12. Example detections of Columba palumbus during the inference trial.
Figure 12. Example detections of Columba palumbus during the inference trial.
Remotesensing 15 02638 g012
Figure 13. Example detections of Pica pica during the inference trial.
Figure 13. Example detections of Pica pica during the inference trial.
Remotesensing 15 02638 g013
Table 1. TensorBoard Results for RPN and Classifier Loss on Training Split.
Table 1. TensorBoard Results for RPN and Classifier Loss on Training Split.
MetricValueStepSupport
RPNLoss/objectness0.0132530,00029,683
RPNLoss/localization0.0686830,00029,683
BoxClassifierLoss/classification0.297730,00029,683
BoxClassifierLoss/localisation0.0540730,00029,683
Total Loss0.433730,00029,683
Table 2. TensorBoard Results for RPN and Classifier Loss on Validation Split.
Table 2. TensorBoard Results for RPN and Classifier Loss on Validation Split.
MetricValueStepSupport
RPNLoss/objectness0.0121230,0003298
RPNLoss/localization0.0403130,0003298
BoxClassifierLoss/classification0.249630,0003298
BoxClassifierLoss/localisation0.012130,0003298
Total Loss0.282830,0003298
Table 3. TensorBoard Results for Precision.
Table 3. TensorBoard Results for Precision.
MetricValueStepSupport
Precision/mAP0.785630,0003298
Precision/mAP (Large)0.790230,0003298
Precision/mAP (Medium)0.629630,0003298
Precision/mAP (Small)0.462930,0003298
Precision/[email protected]0.945830,0003298
Precision/[email protected]0.899230,0003298
Table 4. TensorBoard Results for Recall.
Table 4. TensorBoard Results for Recall.
MetricValueStepSupport
Recall/AR@10.819430,0003298
Recall/AR@100.871130,0003298
Recall/AR@1000.872130,0003298
Recall/AR@100 (Large)0.875930,0003298
Recall/AR@100 (Medium)0.705330,0003298
Recall/AR@100 (Small)0.553730,0003298
Table 5. Inference Performance Metrics for Camera Deployment.
Table 5. Inference Performance Metrics for Camera Deployment.
Fold1Fold2Fold3Fold4Fold5Fold6Fold7Fold8Fold9Fold10Average
Carduelis carduelis
Accuracy91.58%96.39%96.39%97.81%96.35%97.80%96.00%99.37%98.82%98.19%96.87%
Precision85.00%93.75%93.75%70.00%30.00%33.33%12.50%50.00%60.00%25.00%55.33%
Sensitivity56.67%71.43%71.43%87.50%100.00%100.00%100.00%100.00%100.00%100.00%88.70%
Specificity98.13%99.42%99.42%98.29%96.30%97.78%95.98%99.36%98.80%98.18%98.17%
F1-Score68.00%81.08%81.08%77.78%100.00%50.00%22.22%66.67%75.00%40.00%66.18%
Chloris chloris
Accuracy92.55%95.90%93.58%92.75%98.40%94.68%97.11%96.91%95.45%97.02%95.44%
Precision36.84%52.94%42.11%26.32%80.00%44.44%54.55%61.54%46.15%50.00%49.49%
Sensitivity77.78%100.00%88.89%100.00%100.00%100.00%100.00%100.00%85.71%100.00%95.24%
Specificity93.30%95.70%93.82%92.55%98.30%94.44%97.01%96.75%95.86%96.93%95.47%
F1-Score50.00%69.23%57.14%41.67%88.89%61.54%70.59%76.19%60.00%66.67%64.19%
Columba palumbus
Accuracy93.55%96.39%95.63%91.33%97.37%97.80%94.38%94.58%94.38%96.45%95.19%
Precision97.96%96.36%95.74%93.22%98.61%96.55%96.72%92.31%92.06%95.16%95.47%
Sensitivity81.36%91.38%88.24%80.88%94.67%96.55%88.06%90.57%92.06%95.16%89.89%
Specificity99.21%98.53%98.48%96.88%99.13%98.39%98.20%96.46%95.65%97.20%97.81%
F1-Score88.89%93.81%91.84%86.61%96.60%96.55%92.19%91.43%92.06%95.16%92.51%
Cyanistes caeruleus
Accuracy91.58%98.42%95.63%95.72%96.35%97.27%97.27%94.92%96.91%97.11%96.12%
Precision65.52%92.00%95.83%100.00%100.00%100.00%100.00%94.74%100.00%92.31%94.04%
Sensitivity76.00%95.83%76.67%70.37%73.08%73.08%78.26%69.23%79.17%75.00%76.67%
Specificity93.94%98.80%99.35%100.00%100.00%100.00%100.00%99.34%100.00%99.36%99.08%
F1-Score70.37%93.88%85.19%82.61%84.44%84.44%87.80%80.00%88.37%82.76%83.99%
Erithacus rubecula
Accuracy94.57%97.91%97.77%96.24%98.93%96.74%97.67%98.74%97.67%96.45%97.27%
Precision89.66%95.83%95.83%100.00%100.00%95.83%92.00%100.00%100.00%95.83%96.50%
Sensitivity78.79%88.46%88.46%76.67%92.31%82.14%92.00%91.67%84.62%82.14%85.73%
Specificity98.01%99.39%99.35%100.00%100.00%99.36%98.64%100.00%100.00%99.29%99.40%
F1-Score83.87%92.00%92.00%86.79%96.00%88.46%92.00%95.65%91.67%88.46%90.69%
Garrulus glandarius
Accuracy93.05%97.40%97.77%97.81%97.88%98.89%100.00%94.58%98.25%99.39%97.50%
Precision7.69%37.50%33.33%20.00%33.33%33.33%100.00%10.00%25.00%50.00%35.02%
Sensitivity50.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%95.00%
Specificity93.51%97.35%97.74%97.80%97.86%98.88%100.00%94.55%98.24%99.39%97.53%
F1-Score13.33%54.55%50.00%33.33%50.00%50.00%100.00%18.18%40.00%66.67%47.61%
Blank
Accuracy97.75%98.94%98.87%98.35%98.93%99.44%97.67%98.13%97.11%97.60%98.28%
Precision100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%
Sensitivity83.33%92.00%91.67%88.00%92.00%96.00%84.00%88.00%80.00%84.00%87.90%
Specificity100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%100.00%
F1-Score90.91%95.83%95.65%93.62%95.83%97.96%91.30%93.62%88.89%91.30%93.49%
Passer domesticus
Accuracy96.13%96.89%96.15%93.72%94.87%95.19%96.00%96.91%96.00%95.88%95.77%
Precision94.44%97.22%88.57%87.50%85.71%97.56%92.11%97.06%95.12%91.43%92.67%
Sensitivity87.18%87.50%91.18%87.50%80.00%83.33%92.11%89.19%88.64%88.89%87.55%
Specificity98.59%99.35%97.30%95.80%97.58%99.28%97.08%99.20%98.47%95.88%97.85%
F1-Score90.67%92.11%89.86%87.50%82.76%89.89%90.91%92.96%91.76%90.14%89.85%
Pica pica
Accuracy96.67%97.91%98.31%95.21%98.93%99.44%97.67%98.13%97.67%98.19%97.81%
Precision25.00%50.00%80.00%38.46%77.78%85.71%50.00%75.00%75.00%57.14%61.41%
Sensitivity100.00%75.00%88.89%83.33%100.00%100.00%100.00%60.00%60.00%100.00%86.72%
Specificity96.63%98.40%98.82%95.60%98.89%99.42%97.62%99.35%99.35%98.15%98.22%
F1-Score40.00%60.00%84.21%52.63%87.50%92.31%66.67%66.67%66.67%72.73%68.94%
Overall Model
Accuracy94.16%97.35%96.71%95.44%97.56%97.47%96.83%97.14%96.94%97.49%96.71%
Precision66.90%79.51%79.79%70.61%78.38%76.31%76.69%76.21%74.92%73.84%75.32%
Sensitivity76.79%89.07%87.83%86.03%92.45%92.92%91.71%88.73%89.56%92.80%88.79%
Specificity96.81%98.55%98.18%97.44%98.67%98.62%98.21%98.41%98.22%98.54%98.16%
F1-Score66.23%81.39%80.79%71.39%80.91%79.39%78.43%76.64%77.77%78.11%77.10%
Table 6. Confusion Matrix for Inference. The bold diagonal number indicates the TP for each of the classes.
Table 6. Confusion Matrix for Inference. The bold diagonal number indicates the TP for each of the classes.
Carduelis carduelisChloris chlorisColumba palumbusCyanistes caeruleusErithacus rubeculaGarrulus glandariusBlankPasser domesticusPica pica
Carduelis carduelis63110804001
Chloris chloris1740300000
Columba palumbus4055200220828
Cyanistes caeruleus1425518704015
Erithacus rubecula128102322080
Garrulus glandarius0001014000
Blank029350218110
Passer domesticus13147041103450
Pica pica0040010049
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chalmers, C.; Fergus, P.; Wich, S.; Longmore, S.N.; Walsh, N.D.; Stephens, P.A.; Sutherland, C.; Matthews, N.; Mudde, J.; Nuseibeh, A. Removing Human Bottlenecks in Bird Classification Using Camera Trap Images and Deep Learning. Remote Sens. 2023, 15, 2638. https://doi.org/10.3390/rs15102638

AMA Style

Chalmers C, Fergus P, Wich S, Longmore SN, Walsh ND, Stephens PA, Sutherland C, Matthews N, Mudde J, Nuseibeh A. Removing Human Bottlenecks in Bird Classification Using Camera Trap Images and Deep Learning. Remote Sensing. 2023; 15(10):2638. https://doi.org/10.3390/rs15102638

Chicago/Turabian Style

Chalmers, Carl, Paul Fergus, Serge Wich, Steven N. Longmore, Naomi Davies Walsh, Philip A. Stephens, Chris Sutherland, Naomi Matthews, Jens Mudde, and Amira Nuseibeh. 2023. "Removing Human Bottlenecks in Bird Classification Using Camera Trap Images and Deep Learning" Remote Sensing 15, no. 10: 2638. https://doi.org/10.3390/rs15102638

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop