Next Article in Journal
Heat Stress Alters the Effect of Eimeria maxima Infection on Ileal Amino Acids Digestibility and Transporters Expression in Meat-Type Chickens
Previous Article in Journal
Determination of the Optimal Level of Dietary Zinc for Newly Weaned Pigs: A Dose-Response Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of a Slow Loris Computer Vision Detection Model

1
Key Laboratory of Livestock and Poultry Multi-Omics, Ministry of Agriculture and Rural Affairs, College of Animal Science and Technology, Sichuan Agricultural University, Chengdu 611130, China
2
Farm Animal Genetic Resources Exploration and Innovation Key Laboratory of Sichuan Province, Sichuan Agricultural University, Chengdu 611130, China
3
College of Information Engineering, Sichuan Agricultural University, Ya’an 625014, China
4
Sichuan Key Laboratory of Agricultural Information Engineering, Sichuan Agricultural University, Ya’an 625014, China
5
College of Life Science, Sichuan Agricultural University, Ya’an 625014, China
*
Authors to whom correspondence should be addressed.
Animals 2022, 12(12), 1553; https://doi.org/10.3390/ani12121553
Submission received: 16 May 2022 / Revised: 10 June 2022 / Accepted: 13 June 2022 / Published: 16 June 2022
(This article belongs to the Section Animal Welfare)

Abstract

:

Simple Summary

Slow lorises are nocturnal primates native to south-east Asia. All the species of slow loris have been listed in Appendix I of the Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES). It is difficult to artificially detect the slow loris due to its nocturnal habit and venomous bite. This article investigates the feasibility of computer vision for slow loris detection and proposes an improved YOLOv5 algorithm that contributes to formulating an available model for behavior recognition of this endangered taxon.

Abstract

The slow loris (Genus Nycticebus) is a group of small, nocturnal and venomous primates with a distinctive locomotion mode. The detection of slow loris plays an important role in the subsequent individual identification and behavioral recognition and thus contributes to formulating targeted conservation strategies, particularly in reintroduction and post-release monitoring. However, fewer studies have been conducted on efficient and accurate detection methods of this endangered taxa. The traditional methods to detect the slow loris involve long-term observation or watching surveillance video repeatedly, which would involve manpower and be time consuming. Because humans cannot maintain a high degree of attention for a long time, they are also prone to making missed detections or false detections. Due to these observational challenges, using computer vision to detect slow loris presence and activity is desirable. This article establishes a novel target detection dataset based on monitoring videos of captive Bengal slow loris (N. bengalensis) from the wildlife rescue centers in Xishuangbanna and Pu’er, Yunnan, China. The dataset is used to test two improvement schemes based on the YOLOv5 network: (1) YOLOv5-CBAM + TC, the attention mechanism and deconvolution are introduced; (2) YOLOv5-SD, the small object detection layer is added. The results demonstrate that the YOLOv5-CBAM + TC effectively improves the detection effect. At the cost of increasing the model size by 0.6 MB, the precision rate, the recall rate and the mean average precision (mAP) are increased by 2.9%, 3.7% and 3.5%, respectively. The YOLOv5-CBAM + TC model can be used as an effective method to detect individual slow loris in a captive environment, which helps to realize slow loris face and posture recognition based on computer vision.

1. Introduction

Slow loris (Nycticebus spp., Lorisidae) [1] is a group of nocturnal primates in Southeast Asia and China. All the species have been listed in Appendix I of the Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES). The illegal trade and habitat loss have caused a dramatic decline in the wild population of slow lorises [2]. Two species of the slow loris, namely the pygmy (N. pygmaeus) and Bengal slow loris (N. bengalensis), are distributed in China. Currently, less than 1200 individuals are restricted to 29 forest fragments in Yunnan Province. A large number of individuals were confiscated and kept in captivity due to the illegal harvesting and trade. Animal welfare plays an important role in captive husbandry and animal conservation [3]. Monitoring animal activity is one way to draw conclusions about animal welfare [4]. As the key link of protection and research of slow loris, the observation or identification plays a supporting role in many aspects of slow loris. However, animals often exhibit different behaviors because of people’s presence, which can affect the experimental results. At the same time, manual observation is a labor-intensive task, and it is often impossible for humans to continuously observe all day. Therefore, it is essential to develop an observation method that can work for a long time without human presence.
Object detection is an important research direction in computer vision, where the aim is to accurately identify the class and location of a specific target object in a given image [5]. Compared with the traditional detection methods, computer vision-based object detection has the advantages of speed, versatility and independence of subjective thought.
In recent years, the feature learning and transfer learning capabilities of DCNN (Deep Convolutional Neural Networks) have made significant progress in object detection algorithms, such as feature extraction, image representation, classification and recognition [6,7]. Object detection algorithms based on deep learning are mainly divided into two categories: two stage and one stage. The two stage object detection method first generates regions and then uses the CNNs (Convolutional Neural Networks) for sample classification. The common algorithm is the R-CNN (Region-based Convolutional Neural Network) series. One stage method directly extracts features to predict object class and location. The common algorithm is the YOLO (You Only Look Once) series.
With the large-scale application of deep learning [8] in the field of object detection, the accuracy and speed of object detection techniques have been rapidly improved. Object detection techniques have been widely used in the fields of pedestrian detection [9], face detection [10,11], text detection [12], traffic sign and signal detection and remote sensing image detection. Considerable research has also been conducted in the field of animals. For example, Huang et al. proposed an improved single multi-box detection method for cow condition scoring [13]. Hou et al. developed a CNN-based facial recognition model to identify individual giant pandas [14]. Schütz et al. used YOLOv4 to conduct individual detection and motion monitoring of red foxes [15]. Kalhagen et al. proposed a YOLO-Fish model for fish detection in real-time videos [16].
In order to solve the problems of traditional methods and improve the accuracy of slow loris detection, a novel manually annotated Bengal slow loris detection dataset is established in this study. Based on this dataset, an object detection method is proposed that improves YOLOv5. Improvements are made based on the YOLOv5 network and attention mechanism and deconvolution are introduced in the network. The experiments show that the method proposed in this article has a good effect on the detection of captive slow lorises.

2. Materials and Methods

2.1. Dataset Acquisition

The dataset used in this article was obtained from the wildlife rescue centers in Xishuangbanna and Pu’er, Yunnan Province. The information about each rescue center is shown in Table 1. Videos of Bengal slow loris’ activities were continuously acquired by installing night vision monitoring systems (TCNC9401S3E-2MP-I5S and TC-NC9501S3E-2MP-I3S infrared camera, Tiandy Technologies Co., Ltd., Tianjin, China) in their cages. More than 50TB of slow loris activity videos were obtained from April 2017 to June 2018. The obtained videos were extracted with equal frames and one frame was intercepted every three minutes. Because slow lorises are nocturnal animals, most of the videos were taken at night and some images were unclear due to the movement of slow lorises. This part of the data was added to the dataset to enhance the robustness of the dataset. After selection, 1237 images were obtained as the experimental dataset. The dataset is shown in Figure 1. The LabelImg was used to label the dataset. To record the percentage of time the slow loris spent alone and in group activities and thus better determine the degree of harmony of the slow loris group and the degree of an individual slow loris in the group, the status of slow loris was divided into two categories: single and socializing. The dataset was divided into training and validation sets according to 7:3.

2.2. Definition of the Slow Loris States

The two states of slow loris, single and socializing, are defined as follows.
Single: The state of activity in which the slow lorises themselves perform behaviors, such as resting and feeding alone.
Socializing: A state in which multiple slow lorises gather to perform behaviors, such as playing, allogrooming or fighting (distance < 0.3 m).
The two states are shown in Figure 2.

2.3. Experimental Environment and Hardware Configuration

The following hardware configurations were used: RTX A4000 16GB graphics card, 12-core Inter (R) Gold 5320 CPU (Central Processing Unit), and 32GB RAM (Random Access Memory). The experimental environments were: Ubuntu (Canonical Ltd., London, UK), CUDA 11.1 (NVIDIA Corporation, Santa Clara, CA, USA), PyTorch 1.8.1 (Facebook Artificial Intelligence Institute, New York, NY, USA), and Python 3.8.0. (Python Software Foundation, New Castle, DE, USA) All experiments in this article were performed on the GPU (Graphics Processing Unit). All experiments were trained using default parameters.

2.4. A YOLOv5 Network Introducing Attention Mechanism and Deconvolution

2.4.1. YOLO Series Models

YOLO is another framework, proposed by Redmon et al. in 2016, to resolve the speed problem in the deep learning object detection field after RCNN, fast-RCNN and faster-RCNN, which opened up a new idea of object detection [17,18]. The core idea of YOLO is to replace the two-stage algorithm of RoI (Region of Interest) + target detection with a set of one-stage algorithms of the network and to treat the object detection task as a regression problem; the position of the bounding box and the category it belongs to are directly returned to the output layer.
YOLOv1 uses the method of predefined candidate areas and divides the image into multiple grids. Although each grid can predict multiple bounding boxes, only the bounding box with the highest Intersection over Union (IoU) is selected as the object detection result. Hence, each grid can predict only one object at most. Only one target can be detected if objects occupy a small proportion of the screen; the image contains dense targets and each grid contains multiple targets. Although YOLOv1 is fast in detection, it is not accurate enough in localization and has a low recall rate. Redmon et al. introduced YOLOv2 [19] to improve the localization accuracy and the recall rate. Based on YOLOv1, Redmon et al. made many improvements to improve the mean average precision (mAP) in YOLOv2. However, the problem of identifying a small target is still a difficult task. After that, Redmon et al. proposed YOLOv3, summarizing some of their tentative improvements based on YOLOv2. In YOLOv3 [20], Redmon et al. used the residual model to further deepen the network structure and the Feature Pyramid Networks (FPN) architecture to achieve multi-scale detection. YOLOv3 made progress in the detection of small objects, but the detection of medium and larger sized objects was not ideal. To solve these problems, YOLOv4 and YOLOv5 were proposed later. Based on the original YOLO object detection architecture, many optimization strategies were adopted. There are different degrees of optimization in data processing, backbone network, network training, activation function and loss function. Both YOLOv5 and v4 use Cross Stage Partial Darknet (CSP-Darknet) as the backbone and extract rich informative features from input images [21]. Cross Stage Partial Networks (CSPNet) solve the problem of duplicate gradient information within network optimization. The networks respect the variability of the gradients by integrating feature maps from the beginning and the end of a network stage. Therefore, the number of parameters and floating-point operations per second (FLOPS) values of the model are reduced, improving the speed and accuracy of inference and reducing the model size.
The object detection network YOLOv5 has 4 versions, including YOLOv5s, YOLOv5m, YOLOv5l and YOLOv5x. The YOLOv5 model is divided into 4 parts: Input, Backbone, Neck and Output. Figure 3 shows the model structure of YOLOv5. The YOLOv5 is flexible, lightweight and faster with higher accuracy and better small target recognition than the other YOLO series. Therefore, YOLOv5 was selected in this study as the base framework of slow loris detection.

2.4.2. Attention Mechanisms

In recent years, attention mechanism has been widely used in various fields of deep learning [22] and attention models can be seen in various types of tasks such as image processing, speech recognition and natural language processing.
The Convolutional Block Attention Module (CBAM) is an attention mechanism module that combines spatial and channel [23]. The CBAM can achieve better results than the attention mechanism that only focuses on channels. The architecture of the CBAM is shown in Figure 4. Because CBAM incorporates spatial attention, it significantly improves the model’s ability to extract key information in space.
In the extraction pre-experiments of this study, the effect of object edge detection did not meet the expectations. After careful analysis, it was found that it was caused by the low clarity in the images and the lack of information about the edges of the objects. Because the CBAM can effectively integrate the spatial information of the images without significantly increasing the convolutional operations, the CBAM is added to the improved YOLOV5 model. The edge blurring problem is effectively solved by improving the information extraction ability of space, which improves the accuracy of the detection model.

2.4.3. Deconvolution

Convolution: Convolution is widely used in the field of image processing [24]. Different convolutional kernels can be used to achieve filtering, edge detection and image sharpening. In a CNN, the features in the image can be extracted by convolutional operation. The low-level convolutional layer can extract some features of the image, such as edges, lines and corners, whereas the high-level convolution layer can learn more complex features from the low-level convolutional layer. In the end, the classification and recognition of the image are obtained.
In the mathematical definition, a convolution is multiplying one variable with another one and adding it to the total sum. In digital image processing, the convolution operation is to let the convolution kernel slide over the image. The gray value of pixels in the image is multiplied by the value on the corresponding convolution kernel, and then all the multiplied values are added as the gray value of pixels in the image.
The following concepts play important roles in convolutional operations:
  • Padding: Before the convolution operation, the boundaries of the original matrix are filled with padding. Specifically, some values are padded on the boundary of the matrix to increase the size of the matrix. 0 is commonly chosen.
  • Stride: When sliding the convolution kernel, start with the top left corner of the input, and step over one column to the left or one row down at a time. The number of rows and columns in each slide is called Stride. During the convolution process, padding is used to avoid information loss and the step size (Stride) is also set to compress part of the information or make the output size smaller than the input one.
  • Channel (number of filters): The number of output channel layers is only related to the number of channels in the current filter.
In addition, there are concepts, such as input graph size, output graph size, and convolution kernel size. The formula of the 2D convolutional output image is calculated as follows (σ: the output image size, n × n : the input image size, k × k : the convolutional kernel size, p: the size of padding; s: the size of the stride):
σ = n + 2 p k s + 1
After the above, it is easy to see that the size and resolution ratio of the input image will be reduced after a series of convolution operations.
Deconvolution is also known as transposed convolution [25]. The forward propagation process of the convolution layer is the backward propagation process of the deconvolution layer and vice versa. The relationship between the input and output of the deconvolution is defined as:
σ = s n 1 + k 2 p
It can be seen that deconvolution allows the feature map to go from low to high dimensions [26], which is more random than giving the high-dimensional image directly. The network hyperparameters can be obtained by modifying the transposed convolution kernel. With the successful application of deconvolution on neural network visualization, it has been adopted in various works, such as scene segmentation and generative model.
In the proposed framework, deconvolution is introduced into the YOLOv5 network and is used to restore the feature map obtained by convolution to the pixel space. The patterns the feature map responds to most can be obtained, meaning the features extracted by the convolution operation can be known. The experimental results show that deconvolution effectively improves the accuracy of slow loris detection.

2.4.4. Network Structure

The CBAM and deconvolution are added to the YOLOv5 and YOLOv5-CBAM + TC is obtained with a better detection effect of slow loris. The network structure of YOLOv5-CBAM + TC is shown in Figure 5.

2.5. Evaluation Indicators for the Experiment

In this article, precision, recall and mAP are used as evaluation metrics for each model, which are commonly used in the field of object detection. The parameters in the calculation equation of each evaluation metric are defined in Table 2.
Precision is a statistic from the perspective of prediction results. It means the proportion of data is truly positive. That is, the percentage of “right search” or precision is calculated as:
precision = TP TP + FP
Recall is calculated from the real sample set. This means the proportion of positive samples recovered by the model to the total positive samples. That is, the percentage of “complete search”. The Recall is calculated as:
recall = TP TP + FN
The mAP is called mean average precision, which is the average value of AP (average precision) for each category. AP means the area under the precision–recall (PR) curve, and AP is also an indicator of P-R. The better the classifier, the higher the AP value.
The mAP is the average of AP of each category, and this value represents the comprehensive evaluation of the detection target.
[email protected] is the mean average precision when the threshold of IoU is 0.5.
[email protected]:0.95 is the average mAP when the threshold of IoU ranges from 0.5 to 0.95 in steps of 0.05.

3. Results

3.1. Model Comparison

In order to evaluate whether the proposed modified YOLOv5 network can better complete the task of detecting slow loris and whether it is better than the basic YOLOv5, the labeled dataset was used to train 3 networks (YOLOv5, YOLOv5-SD and YOLOv5-CBAM + TC). The dataset contained 1237 images, of which 70% were the training data and 30% were the validation data. The training was performed with 100 epochs, and the experimental results are shown in Figure 6 and Table 3.

3.2. Analysis

It can be seen from the experimental results that the YOLOv5-CBAM + TC outperforms the other two models in terms of precision, recall and mAP, and the YOLOv5-SD outperforms the YOLOv5 in terms of recall and mAP, but the precision decreases slightly and is not effective in detecting single slow loris. Therefore, YOLOv5-SD is not considered further and the YOLOv5-CBAM+TC model is selected.
Figure 7 shows the true labels of four images randomly selected from the slow loris dataset and the detection effects of the original YOLOv5 and the YOLOv5-CBAM + TC on the four images for lorises in different states.
It can be seen from Figure 7 that there are some slow lorises close to each other but not in a socially active state. The original YOLOv5 model appears to classify them as socially active (the left image in Figure 7b), whereas the improved YOLOv5-CBAM + TC model does not show this phenomenon. The results show that the YOLOv5-CBAM + TC is more suitable for loris detection in the nighttime environment. Embedding CBAM with deconvolution significantly improves the robustness and detection effect of the model, proving the effectiveness of the network.
It can be inferred from Table 4 that the proposed network has a higher mAP than the mainstream detection algorithms. The detection effect of YOLOv5-CBAM + TC is shown in Figure 8.

4. Discussion

In this article, the automatic detection model of Bengal slow loris in cages in wildlife rescue centers is studied and the YOLOv5 model is improved. The videos of the nocturnal activity of slow loris are obtained through a night vision monitoring system installed above the cages. The obtained videos are extracted frame by frame and the slow loris detection images are obtained using the improved YOLOv5 model. For the feasibility of the proposed approach, the following are the discussion points:
(1)
Because the dataset in this study is collected by a limited number of cameras located on top of the interior of the cage, the observation angle will be limited, and misplacement may occur, leading to incorrect detection results. Therefore, in the subsequent study, the camera positions will be adjusted and the number of cameras will be increased to avoid this problem as much as possible.
(2)
In terms of processing speed, YOLOv5 has a high processing speed for the reasons explained in Section 2.4.1. Although the improved YOLOv5-CBAM + TC model has 0.6 MB more than YOLOv5, it only takes 0.1 s to process a single image, which can meet the needs of practical applications.
(3)
The datasets used in this study were all human-collected and the collection locations were fixed with relatively simple and single backgrounds. Thus, the accuracy of the model may be reduced in complex environments, such as in the wild. Considering that the scenario currently applied is a case of slow lorises, the scenario in actual application is relatively simple. The authors will continue their research on YOLOv5-CBAM+TS and extend its application scenarios.
(4)
In terms of model generalization capability, YOLOv5 adopts a mosaic data enhancement strategy to improve the generalization capability and robustness of the model [27].
(5)
Compared with the application of computer vision in the detection of other mammals (such as elephant (Elephantidae) [28] and golden monkey (Rhinopithecus roxellana) [29]), the performance of the proposed YOLOv5 CBAM + TC model in the detection of slow loris exceeds the average level and meets the needs of practical applications.
(6)
The YOLOv5 CBAM + TC model was operated on a professional server in this study, but it can also be run smoothly on a common laptop, indicating that the model would be economical and practical in a real-world application.
The above discussion demonstrates that the proposed method is exploratory and helpful for slow loris conservation and has an important supporting role in realizing individual identification and behavior recognition of slow loris based entirely on deep learning.

5. Conclusions

Although computer vision has been widely applied to animal object detection [30,31], there are still fewer studies on slow loris detection, which significantly affects the protection of endangered species. In this study, computer vision is applied for individual identification and behavioral recognition of slow lorises. After establishing the slow loris dataset, the original YOLOv5 is improved and the YOLOv5-CBAM + TC model is obtained. The proposed model has better precision, recall and mAP than the original YOLOv5, and it has better performance in slow loris detection. The experimental results also demonstrate the effectiveness of the proposed YOLOv5-CBAM + TC. While providing an efficient and accurate method for slow loris detection for scientific researchers engaged in slow loris protection, this study also verifies the feasibility and immense development potential of computer vision in the field of animal protection. The results show that the improved YOLOv5 model has acceptable accuracy and speed in slow loris detection. This study lays a foundation for the welfare improvement of endangered animals, such as slow loris, which have received less attention and are relatively backward in protection measures. In the future, the authors will continue to optimize the YOLOv5-CBAM + TC model, expand its application range from captivity to the wild, and combine it with other deep learning technologies to realize the individual identification [32] and behavior recognition [33] of the slow loris.

Author Contributions

Conceptualization, Y.L., Y.X. and Q.N.; methodology, Y.L.; software, Y.L., Y.X., Y.Z. (Yuhui Zhu), Y.Z. (Yu Zhang), X.Y. (Xiao Yang) and Y.G.; validation, Y.L., Y.G., Y.Z. (Yuhui Zhu), Y.X. and Y.Z. (Yu Zhang); formal analysis, Y.X. and Y.L.; investigation, Y.G., Y.Z. (Yuhui Zhu), Y.X., X.Y. (Xiaoli Yao) and Y.L.; resources, J.M. and Q.N.; data curation, Y.X. and T.L.; writing—original draft preparation, Y.L. and Y.X.; writing—review and editing, Y.L., M.X. and Q.N.; visualization, Y.G., Y.X., Y.L., T.L. and X.Y. (Xiao Yang); supervision, J.M. and Q.N.; project administration, Y.L. and Q.N.; funding acquisition, Q.N. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Nature Science Foundation of China (No. 32070521) and Innovation and entrepreneurship training program for College Students.

Institutional Review Board Statement

The animal study protocol was approved by (IRB13627) and the Institutional Animal Care and Use Committee of the Sichuan Agricultural University, China under permit number DKY-2021302142 (date of approval: 10 October 2021).

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Acknowledgments

We are grateful to the entire staff of Dehong, Puer and Xishuangbanna Wildlife Rescue Center for facilitating the research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Munds, R.A.; Nekaris, K.A.I.; Ford, S.M. Taxonomy of the Bornean Slow loris, With New Species Nycticebus kayan (Primates, Lorisidae). Am. J. Primatol. 2013, 75, 46–56. [Google Scholar] [CrossRef] [PubMed]
  2. Nekaris, K.A.I.; Starr, C.R. Conservation and ecology of the neglected slow loris: Priorities and prospects. Endanger. Species Res. 2015, 28, 87–95. [Google Scholar] [CrossRef] [Green Version]
  3. Broom, D.M. Animal welfare: Concepts and measurement. J. Anim. Sci. 1991, 69, 4167–4175. [Google Scholar] [CrossRef]
  4. Broom, D.M. The scientific assessment of animal welfare. Appl. Anim. Behav. Sci. 1988, 20, 5–19. [Google Scholar] [CrossRef]
  5. Zhao, Z.Q.; Zheng, P.; Xu, S.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Rawat, W.; Wang, Z. Deep convolutional neural networks for image classification: A comprehensive review. Neural Comput. 2017, 29, 2352–2449. [Google Scholar] [CrossRef] [PubMed]
  7. Wang, Z.; Liu, J. A review of object detection based on convolutional neural network. In Proceedings of the 2017 36th Chinese Control Conference (CCC), Dalian, China, 26–28 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 11104–11109. [Google Scholar]
  8. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  9. Dollar, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian detection: An evaluation of the state of the art. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 743–761. [Google Scholar] [CrossRef]
  10. Tolba, A.S.; El-Baz, A.H.; El-Harby, A.A. Face recognition: A literature review. Int. J. Signal Process. 2006, 2, 88–103. [Google Scholar]
  11. Hu, G.; Yang, Y.; Yi, D.; Kittler, J.; Christmas, W.; Li, S.Z.; Hospedales, T. When face recognition meets with deep learning: An evaluation of convolutional neural networks for face recognition. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 7–13 December 2015; pp. 142–150. [Google Scholar]
  12. Ye, Q.; Doermann, D. Text detection and recognition in imagery: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 1480–1500. [Google Scholar] [CrossRef]
  13. Huang, X.; Hu, Z.; Wang, X.; Yang, X.; Zhang, J.; Shi, D. An improved single shot multibox detector method applied in body condition score for dairy cows. Animals 2019, 9, 470. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Hou, J.; He, Y.; Yang, H.; Connor, T.; Gao, J.; Wang, Y.; Zeng, Y.; Zhang, J.; Huang, J.; Zheng, B. Identification of animal individuals using deep learning: A case study of giant panda. Biol. Conserv. 2020, 242, 108414. [Google Scholar] [CrossRef]
  15. Schütz, A.K.; Schöler, V.; Krause, E.T.; Fischer, M.; Müller, T.; Freuling, C.; Conraths, F.; Stanke, M.; Homeier-Bachmann, T.; Lentz, H. Application of YOLOv4 for Detection and Motion Monitoring of Red Foxes. Animals 2021, 11, 1723. [Google Scholar] [CrossRef] [PubMed]
  16. Kalhagen, E.S.; Olsen, Ø.L. Hierarchical Fish Species Detection in Real-Time Video Using YOLO. Master’s Thesis, University of Agder, Kristiansand, Norway, 2020. [Google Scholar]
  17. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  18. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  19. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  20. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  21. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  22. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  23. Woo, S.; Park, J.; Lee, J.Y.; Kweon, S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  24. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef] [Green Version]
  25. Shi, W.; Caballero, J.; Theis, L.; Huszar, F.; Aitken, A.; Ledig, C.; Wang, Z. Is the deconvolution layer the same as a convolutional layer? arXiv 2016, arXiv:1609.07009. [Google Scholar]
  26. Pan, J.; Sayrol, E.; Giro-i-Nieto, X.; McGuiness, K.; O’Connor, N. Shallow and deep convolutional networks for saliency prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 598–606. [Google Scholar]
  27. Yao, J.; Qi, J.; Zhang, J.; Shao, H.; Yang, J.; Li, X. A real-time detection algorithm for Kiwifruit defects based on YOLOv5. Electronics 2021, 10, 1711. [Google Scholar] [CrossRef]
  28. Premarathna, K.S.P.; Rathnayaka, R.M.K.T.; Charles, J. An Elephant Detection System to Prevent Human-Elephant Conflict and Tracking of Elephant Using Deep Learning. In Proceedings of the 5th International Conference on Information Technology Research (ICITR), Moratuwa, Sri Lanka, 2–4 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
  29. Rui, S.; Xu, Z.; Ying, G.; Xinwen, Y.; Yan, C.; Yanan, H. Optimized Detection Method for Snub-Nosed Monkeys Based on Faster R-CNN. Laser Optoelectron. Prog. 2020, 57, 121022. [Google Scholar]
  30. Verma, G.K.; Gupta, P. Wild animal detection using deep convolutional neural network. In Proceedings of the 2nd International Conference on Computer Vision & Image Processing, Hong Kong, China, 29–31 December 2018; Springer: Singapore, 2018; pp. 327–338. [Google Scholar]
  31. Eikelboom, J.A.J.; Wind, J.; van de Ven, E.; Kenana, L.M.; Schroder, B.; de Knegt, H.J.; van Langevelde, F.; Prins, H.H.T. Improving the precision and accuracy of animal population estimates with aerial image object detection. Methods Ecol. Evol. 2019, 10, 1875–1887. [Google Scholar] [CrossRef] [Green Version]
  32. Khan, R.H.; Kang, K.W.; Lim, S.J.; Youn, S.D.; Kwon, O.J.; Lee, S.H.; Kwon, K.R. Animal Face Classification using Dual Deep Convolutional Neural Network. J. Korea Multimed. Soc. 2020, 23, 525–538. [Google Scholar]
  33. Zuerl, M.; Stoll, P.; Brehm, I.; Raab, R.; Zanca, D.; Kabri, S.; Happold, J.; Nille, H.; Prechtel, K.; Wuensch, S.; et al. Automated Video-Based Analysis Framework for Behavior Monitoring of Individual Animals in Zoos Using Deep Learning-A Study on Polar Bears. Animals 2022, 12, 692. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Partial images in the dataset.
Figure 1. Partial images in the dataset.
Animals 12 01553 g001
Figure 2. Example diagram of slow loris states.
Figure 2. Example diagram of slow loris states.
Animals 12 01553 g002
Figure 3. YOLOv5 structure diagram.
Figure 3. YOLOv5 structure diagram.
Animals 12 01553 g003
Figure 4. CBAM structure diagram.
Figure 4. CBAM structure diagram.
Animals 12 01553 g004
Figure 5. The network structure of YOLOv5-CBAM + TC.
Figure 5. The network structure of YOLOv5-CBAM + TC.
Animals 12 01553 g005
Figure 6. Experimental results of the three models: (a) precision; (b) recall; (c) [email protected]; (d) [email protected]:0.95.
Figure 6. Experimental results of the three models: (a) precision; (b) recall; (c) [email protected]; (d) [email protected]:0.95.
Animals 12 01553 g006
Figure 7. Comparison of YOLOv5-CBAM + TC and YOLOv5 prediction results: (a) true labels; (b) YOLOv5; (c) YOLOv5-CBAM+TC.
Figure 7. Comparison of YOLOv5-CBAM + TC and YOLOv5 prediction results: (a) true labels; (b) YOLOv5; (c) YOLOv5-CBAM+TC.
Animals 12 01553 g007
Figure 8. Detection effect of YOLOv5-CBAM + TC.
Figure 8. Detection effect of YOLOv5-CBAM + TC.
Animals 12 01553 g008
Table 1. Information about the wildlife rescue centers and the captive Bengal slow lorises.
Table 1. Information about the wildlife rescue centers and the captive Bengal slow lorises.
Captive SiteXishuangbannaPuer
Coordinate22.39276° N, 100.89636° E22.62198° N, 101.08916° E
Altitude (m)10601600
Annual mean temperature (°C)17.517.5
No. of individuals99
No. of enclosures11
Enclosure size (L × W × H) (m)5.7 × 4.2 × 3.53.5 × 2.1 × 2.0
No. of cameras32
Table 2. Parameter definitions.
Table 2. Parameter definitions.
Confusion MatrixPredicted Results
PositiveNegative
Expected ResultspositiveTP 1FN 2
negativeFP 3TN 4
1 True positive (TP): The prediction result is positive, and the prediction is correct. 2 False negative (FN): The prediction result is negative, but the prediction is incorrect. 3 False positive (FP): The prediction result is positive, but the prediction is incorrect. 4 True negative (TN): The prediction result is negative, and the prediction is correct.
Table 3. Experimental information of the three models.
Table 3. Experimental information of the three models.
ModelCategoryPrecisionRecall[email protected][email protected]:0.95Model Size
YOLOv5All0.9360.9160.9340.60914.4 MB
Single0.9550.9260.9560.554
Socializing0.9230.9060.9130.665
YOLOv5-SDAll0.9310.9430.950.57216.4 MB
Single0.9120.9240.9550.523
Socializing0.9510.9630.9440.622
YOLOv5-CBAM + TCAll0.9650.9530.9690.64215.0 MB
Single0.9560.9430.9640.568
Socializing0.9740.9630.9730.716
Table 4. Comparison between mainstream detection algorithms.
Table 4. Comparison between mainstream detection algorithms.
CategoryYOLOv5-CBAM + TCSSDCenterNetFaster-RCNN
mAP0.9690.9110.8890.939
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lei, Y.; Xiang, Y.; Zhu, Y.; Guan, Y.; Zhang, Y.; Yang, X.; Yao, X.; Li, T.; Xie, M.; Mu, J.; et al. Development of a Slow Loris Computer Vision Detection Model. Animals 2022, 12, 1553. https://doi.org/10.3390/ani12121553

AMA Style

Lei Y, Xiang Y, Zhu Y, Guan Y, Zhang Y, Yang X, Yao X, Li T, Xie M, Mu J, et al. Development of a Slow Loris Computer Vision Detection Model. Animals. 2022; 12(12):1553. https://doi.org/10.3390/ani12121553

Chicago/Turabian Style

Lei, Yujie, Ying Xiang, Yuhui Zhu, Yan Guan, Yu Zhang, Xiao Yang, Xiaoli Yao, Tingxuan Li, Meng Xie, Jiong Mu, and et al. 2022. "Development of a Slow Loris Computer Vision Detection Model" Animals 12, no. 12: 1553. https://doi.org/10.3390/ani12121553

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop