Next Article in Journal
A Dual Neural Architecture Combined SqueezeNet with OctConv for LiDAR Data Classification
Previous Article in Journal
Data Anomaly Detection for Internet of Vehicles Based on Traffic Cellular Automata and Driving Style
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mounting Behaviour Recognition for Pigs Based on Deep Learning

College of Information and Electrical Engineering, China Agricultural University, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(22), 4924; https://doi.org/10.3390/s19224924
Submission received: 11 October 2019 / Revised: 4 November 2019 / Accepted: 9 November 2019 / Published: 12 November 2019
(This article belongs to the Section Intelligent Sensors)

Abstract

:
For both pigs in commercial farms and biological experimental pigs at breeding bases, mounting behaviour is likely to cause damage such as epidermal wounds, lameness and fractures, and will no doubt reduce animal welfare. The purpose of this paper is to develop an efficient learning algorithm that is able to detect the mounting behaviour of pigs based on the data characteristics of visible light images. Four minipigs were selected as experimental subjects and were monitored for a week by a camera that overlooked the pen. The acquired videos were analysed and the frames containing mounting behaviour were intercepted as positive samples of the dataset, and the images with inter-pig adhesion and separated pigs were taken as negative samples. Pig segmentation network based on Mask Region-Convolutional Neural Networks (Mask R-CNN) was applied to extract individual pigs in the frames. The region of interest (RoI) parameters and mask coordinates of each pig, from which eigenvectors were extracted, could be obtained. Subsequently, the eigenvectors were classified with a kernel extreme learning machine (KELM) to determine whether mounting behaviour has occurred. The pig segmentation presented considerable accuracy and mean pixel accuracy (MPA) with 94.92% and 0.8383 respectively. The presented method showed high accuracy, sensitivity, specificity and Matthews correlation coefficient with 91.47%, 95.2%, 88.34% and 0.8324 respectively. This method can be an efficient way of solving the problem of segmentation difficulty caused by partial occlusion and adhesion of pig bodies, even if the pig body colour was similar to the background, in recognition of mounting behaviour.

1. Introduction

Intensive interaction among pigs is likely to bring negative effects to pig health as well reduce animal welfare. Mounting behaviour, which occurs to both male and female pigs during their entire lifetime—especially in oestrus [1,2]—is generally manifested as a pig places two front hoofs on the body or head of another pig which stays lying or dodges quickly. Most bruises, lameness and leg fractures are due to mounting behaviour [3] and those injuries will lead to serious economic losses in livestock farming. Therefore, timely detection and intervention of mounting behaviour will be able to increase animal welfare and further ensure pig health.
Traditional monitoring methods of animal behaviour rely mainly on human eye observation which consumes a lot of labour and involves subjective errors. With the development of image and video processing technology, automated video recognition techniques are increasingly applied to pig breeding enterprises. Optical flow vector [4] and fitted ellipse features in consecutive frames [5] were applied to monitoring of the locomotion of pigs by using a charge-coupled device (CCD) camera. For detecting lameness behaviour, joint angle waveform [6] and star skeleton model [7] were utilized from videos taken from the side of the pigs. In order to track the pigs, Kalman filter [8] and modified particle filtering [9] were investigated and had good robustness. Zhu [10] has proposed an automatic detection method for pigs’ respiratory behaviour based on the freeman code algorithm. By monitoring pigs’ appearance in their customary living area, several kinds of behaviours including drinking, feeding and excretory behaviour [11,12] were detected with high overall sensitivity and accuracy. Automatic video processing methods based on fitted ellipse and dense trajectories features [13] can even be utilized for monitoring lying behaviour. A 3D technique has developed rapidly in recent years and more in depth dimension information has been tapped for behaviour detection. Time of flight by Kinect2.0 is used to capture 2D depth images and 3D point cloud images. While Jonguk [14] had taken advantage of 2D depth images in presenting an automatic recognition algorithm for aggressive behaviour detection and Mateusz [15] extracted the cloud point information of pigs from 3D point cloud images and tracked it with ellipsoidal fitting.
Currently, deep learning has received wide attention due to its outstanding performance in computer vision. Sow position was detected from depth images and five classes of sow postures were classified with high overall accuracy using Faster Region-Convolutional Neural Networks (Faster R-CNN) [16,17]. In another survey, images of sows were segmented accurately by a Fully Convolutional Networks (FCN) [18] applied for detecting nursing behaviour. Deep learning has been used for locating pigs in feeding and drinking area and the occupation rate was implemented to measure the feeding and drinking behaviour of pigs using Faster R-CNN and GoogleNet [19,20]. Faster R-CNN was also adopted by [21] while solving the problem of the loss of pig tracking during visual tracking with good robustness and adaptability.
Considering the problems of epidermal injury and fracture caused by mounting behaviour, it is vital to observe the event in time in order to separate the pigs as soon as possible. Computer vision is capable of providing an apposite behaviour recognition method as a solution which is automatic, non-contact, low-cost, high profit and stress-free. An automatic recognition algorithm was implemented by fitting pigs to ellipses and estimating the major and minor axis of each ellipse in order to monitor mounting behaviour [22]. However, the author of this article also expressed that colour-based segmentation has its own weakness. Considering that it is difficult using colour-based segmentation to distinguish a pig’s body from the background and it is also not a simple task to segment the adhesive part among pigs, we applied the deep learning algorithm for pig segmentation. In this research, Mask R-CNN, a deep convolutional network optimized from Faster R-CNN with a branch for predicting an object mask in parallel with the existing branch for bounding box recognition [23], was proposed to segment pigs. Eigenvectors, such as mask perimeter, the half-length mask area divided by the midpoint of long side of the bounding-box and the distance between the centre points of the rectangle and so forth, were extracted for classification to recognize mounting behaviour. The algorithm was applied to male experimental miniature pigs in oestrus and the performance was considerable.

2. Material and Methods

2.1. Experimental Environment and Animal

The experiment was conducted at the China Experimental Miniature Pig Base in Zhuozhou. Four male pigs in oestrus were placed in a pen which was 2 m × 2.2 m in size. The pigs were three months old and about 25 kg in live weight. All lights were switched on during the experiment time, which lasted a week from June 11, 2018 to June 18, 2018. A GigE camera (Allied Vision Technologies, Manta G-282C, Nürnberg, Germany) was mounted on the elevating bracket at about 2.8 m high, pointing downward to get a top view of the pen (Figure 1a). The video acquisition software developed by AVT was selected in order to get video images at the size of 1936 × 1458. Taking into account the image quality and slow moving speed of pigs, the acquisition speed was set at 2 frames per second. All the frames were stored in a mobile hard disk (Western Digital 2Tb ultra, California, United States). Sample frames from the video sequence are shown in Figure 1b.
The pigs used in this trial were four minipigs pigs. This variety is small in size and white in appearance. It can be widely used in the fields of teratogenicity testing, drug metabolism, organ transplantation, skin transplantation test, and so forth. This type of pig has great medical and commercial value, therefore it is vital to detect harmful behaviour like mounting behaviour among pigs in time so that they can be isolated as soon as possible for the avoidance of epidermal damage.

2.2. Algorithm Overview

After collecting frames from video sequences, images were pre-processed and put into the network. The pigs’ images were segmented by Mask R-CNN, which performs better than Fully Convolutional Instance-aware Semantic Segmentation (FCIS) [24] on detecting overlapping instances.
The multi-dimensional eigenvectors were extracted from the parameters of bounding-boxes and the mask files generated from instance segmentation by the pig segmentation network based on Mask R-CNN. Eigenvectors were extracted from the segmentation results and then classified by extreme learning machine. According to the classification result, it is determined whether or not mounting behaviour has occurred. The procedure of this algorithm is shown in Figure 2.

3. Algorithm for Mounting Behaviour Detection

3.1. Image Pre-Processing and Labelling

After all the video files were observed, 1500 frames were chosen for the mounting behaviour detection experiment. For the purpose of making the sample versatile and contain pigs’ daily behavioural posture, the chosen frames contained a top-down view of feeding, walking, standing, lying, excreting, and mounting behaviours states of pigs in different illumination and cleanliness conditions. The dataset was split randomly into five folds, one fold reserved as test set and the other four reserved as a training set.
All the selected images were first determined by breeding experts to be divided into positive and negative samples based on whether mounting behaviour had occurred. According to the discriminant result, the frames containing mounting behaviour were taken as positive samples, the rest as negative samples. The research team used polygons to approximately mark the outlines of all the pigs in the images by Labelme toolbox. The JSON files were generated after the annotation. The network ground truth of input label names and masks was converted from JSON files using Labelme as well. The dataset processing procedure is shown in Figure 3a. Dataset allocation is shown in Figure 3b. Considering the sequence frame size obtained from the camera was too large for the network, it was converted to 640 × 480 in pixels.

3.2. Pig Segmentation by Mask R-CNN

3.2.1. Related Work

Mask R-CNN is a deep network for instance segmentation proposed by Kaiming He [23]. It is an extension of Faster R-CNN [25]. The prototype of the network dates back to the Region-based CNN (R-CNN) at the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) [26]. The R-CNN consists of selective search part which proposed 2000 region of interests, CNN and Support Vector Machine (SVM) which combines labels and bounding-box for object classification. Considering that the feature extraction for each candidate region is time consuming and cumbersome, Fast R-CNN [27], which uses a convolutional network to extract features, a RoIPool to extract each RoI corresponding feature in the feature map and an fully connected layer (FC) to classify, were created. However, Faster R-CNN applies a Region Proposal Network (RPN) to optimize the most time-consuming selective search portion, making the entire network a complete end to end network which makes Faster R-CNN flexible and robust. In this paper, we use Mask R-CNN to extract the bounding-boxes and masks of pigs.

3.2.2. The Architecture of Pig Detector

The architecture for the pig segmentation network based on Mask R-CNN is displayed in Figure 4. The entire pig network consists of three parts—(1) a backbone which was the combination of residual neural network and feature pyramid networks with 50 layers (ResNet50-FPN) [28] for feature extraction; (2) an RPN for RoIs proposal; (3) three branches for bbox regression, mask regression and category division.
The pig segmentation network selected ResNet50-FPN as the backbone network. As shown in Figure 4b, the backbone network retained five stages of ResNet50 and generated four-level feature maps as P2~P5. The feature maps were used in RPN for the RoIs proposal. RoIAlign was used for feature extraction. The RPN procedure was shown in Figure 4c, P6 was subsampled from P5 with stride 2. P2~P6 were used for RoIs proposal. The RPN initially determined whether the anchor belongs to the foreground or background and performed the first coordinate correction. In the three branches phrase showed in Figure 4d, in classification and bounding-box regression branches, the output category was firstly set to 2 (pig and background). The full connected layers performed a second parameter correction on the positive bounding-boxes. For obtaining the mask of pigs, the RoIPool was replaced with RoIAlign according to Mask R-CNN. RoIAlign cancels the rounding operation in the process of determining the coordinates in the feature map according to the proposals and the RoIs coordinates which were fixed size (7 × 7) according to the feature map like RoIPool did. Moreover, RoIAlign used bilinear interpolation to find the features corresponding to each block. After being processed by the RoIAlign layer, the mask branch started to use deconvolution to improve the resolution and reduce the channel. Finally, the 28 × 28 × 2 (pig and background) mask was regressed.

3.2.3. The Training and Testing Phase of Pig Segmentation Network

The training procedure of the pig segmentation network is shown in Figure 5.
For solving the problem of difficult training of a small-scale training dataset, the transfer learning method was applied to train the pig segmentation network. The experiment process was as shown.
Step 1:
The training dataset containing 1200 samples was randomly split into 5 subsets. One to four folds were reserved as a train set and the remaining was reserved as the validation set. The train set was put into the applied Mask R-CNN.
Step 2:
Load the pre-trained model using the Microsoft Common Objects in Context (MS COCO) dataset.
Step 3:
Modify the configuration parameters and the number of categories.
Step 4:
Start training and observe the change of the validation dataset loss curve.
Step 5:
Reset the parameters such as learning rate, weight decay and the anchor scale of RPN, et al.
Step 6:
Evaluate pig segmentation network using the validation dataset.
Step 7:
Repeat Step 5~Step 6, until the desired accuracy was achieved.
Step 8:
Update train set to 2–5 folds and validation set to the first fold. Repeat Step 2~Step 7 five times until the fivefold cross validation were finished and five models were obtained.
Step 9:
Using the test set to test five models respectively, the average accuracy and the average Mean Pixel Accuracy (MPA) values were obtained as evaluation metrics of this network.
A file contained each point coordinate of masks and another file contained each bounding-box’s parameter would be generated.

3.3. Mounting Behaviour Detection by Kernel-Extreme Learning Machine

3.3.1. Eigenvectors Extraction

After pig segmentation by the Mask R-CNN based network, the mask and bounding-box coordinates could be extracted for eigenvectors acquisition. We extracted the perimeter and the half-body area (HBA) of each pig in the mask as well as the distance between the centre point of every bounding-box in the image as an eigenvector.
(a)
HBA: The long side of the bounding-box was found and the area of the mask was divided into two parts S 1 ,   S 2 with the two long-side midpoint lines as shown in Figure 6a. When the pig’s mask became two parts due to the occurrence of mounting behaviour, the pig’s HBA was defined as S 11 + S 12 .
(b)
The distance between centre point of bounding-boxes (Bboxes) A, B, C, and D in Figure 6b are the centre points of the bounding-boxes of four pigs, respectively. The distance between four points was defined as l 1 , l 2 , …, l 6 . The centre point spacing of the bounding-box frame of each pig in each image was extracted as the eigenvectors.

3.3.2. Classification by Kernel-Extreme Learning Machine

In this article, we used the kernel-extreme learning machine (KELM) to classify the eigenvectors extracted by the Mask R-CNN based pig segmentation network. Kernel-extreme learning machine [29,30] is a kind of machine learning algorithm based on a feedforward neural network. The hidden layer node parameters can be random or artificially given and do not need to be adjusted. It has the characteristic advantages of a short training time and a high generalization ability.
The mounting behaviour detection could be summed up as a two-category problem. In order to establish the classification model of kernel-extreme learning machine, N n-dimensional eigenvectors were recorded as sample F = { F 1 , F 2 , , F n , } , and the classification result is T 1 (positive) and T 2 (negative) represented whether or not the mounting behaviour had occurred.
The mounting behaviour classifier network of the pigs is shown in Figure 6. ω i j is expressed as the weight between the input neuron i and the hidden layer neuron j , and b j is hidden layer neuron offset. The hidden layer activation function is g ( · ) and O 1 , O 2 , …, O m are hidden layer nodes.
The output T in the model can be expressed as: T = H β . H is the hidden layer output matrix and β is the weight between hidden layer and output layer.
H ( ω 1 , , ω m , b 1 , , b m , F 1 , , F m ) = [ g ( ω 11 F 1 + b 1 ) g ( ω 1 m F 1 + b m ) g ( ω 1 N F N + b 1 ) g ( ω N m F N + b m ) ] N × m
β = [ β 1 T β m T ] m × 2
T = [ T 1 T T N T ] N × 2
T i = j = 1 m β j g ( ω i j F i + b j ) ,   i = 1 , 2 , , N
Given an arbitrary small error ε ( ε > 0 ) and an arbitrary interval infinitely differentiable activation function g ( · ) , there is always an ELM with m hidden layers, arbitrarily assigned ω i j n and b j , there is:
H β T < ε
When hidden layer activation function g ( · ) can be guided, the ELM network does not need to change all the parameters and solves the Equation (5):
β ˜ = H + T ,
where H + is the generalized inverse matrix of the hidden layer output matrix T ; β ˜ is the output layer weight.
A kernel matrix for ELM could be defined as follows:
Ω E L M = H H T : Ω E L M i , j = h ( x i ) · h ( x j ) = K ( x i , x j ) .
Then, the output function of ELM classifier can be written compactly as:
f ( x ) = h ( x ) H T ( 1 C + H H T ) 1 T = [ K ( x , x 1 ) K ( x , x N ) ] ( 1 C + Ω E L M ) 1 T
Therefore, when training the mounting behaviour classifier, the feature mapping h ( x ) need not be known to users; instead its corresponding kernel K (u, v). The test results will be obtained by applying the initial hidden layer weights and the output layer weights obtained by the training.

3.3.3. The Training and Testing Phase of Extreme Learning Machine

K-fold cross validation has been widely used to evaluate the performance of machine learning classifiers. In order to seek the hyperparameters, we tuned the parameter based on fivefold cross validation and grid search. That is to say, the training dataset was randomly split into five subsets, and one of those sets was reserved as a test set and the others are reserved as a train set. This process would be repeated five times until getting five optimal models.
After obtaining five models, the models would be tested by the preserved test set. The average of the result would be used to evaluate the performance of the mounting behaviour recognition classifier.

3.4. Performance Evaluation of the Mounting Behaviour Recognition Method

The performance of the pig segmentation network is evaluated by the commonly used metric, Mean Pixel Accuracy (MPA). It is a standard measurement in the evaluation of image segmentation and it is derived from the currently segmented pixel:
M e a n   P i x e l   A c c u r a c y ( M P A ) = 1 k + 1 i = 0 k P i i j = 0 k P i j
where k is the total number of categories except the background, P i j is the total number of the pixel whose real class is i and predicted as j , and P i i is the total number of pixels whose real pixel class is i and collectedly predicted.
To evaluate the performance of the prediction model of mounting behaviour recognition, four metrics [31,32,33] were utilized—Specificity(SP), sensitivity(SN), accuracy(ACC), and the Matthews correlation coefficient (MCC), which are defined as follows:
S p e c i f i c i t y ( S P ) = T N T N + F P S e n s i t i v i t y ( S N ) = T P T P + F N A c c u r a c y ( A C C ) = T P + T N T P + F P + T N + F N M a t t h e w s   c o r r e l a t i o n   c o e f f i c i e n t ( M C C ) = T P × T N F P × F N ( T P + F N ) ( T P + F P ) ( T N + F P ) ( T N + F N )
where TP (true positive) represents the number of positive samples predicted to be positive; TN (true negative) represents the number of negative samples predicted to be negative; FP (false positive) represents the number of negative samples predicted to be positive; FN (false negative) represents the number of positive samples predicted to be negative.

4. Result

4.1. Experiment and Evaluation of Pig Segmentation

The experiment was done by using Tensorflow and Keras deep learning frameworks in python, with NVIDIA TITAN RTX (NVIDIA, Santa Clara, CA, USA) for acceleration. The datasets are strictly non-intersecting with each other. The network was initialized with an MS COCO pre-trained model. Then, all the layers were fine-tuned according to the new pig dataset. The RPN proposal number was 2000 and the positive ratio was 0.33. The learning rate was 0.0005. The weight decay was 0.00005 and the momentum was 0.9.
In order to compare the pig segmentation network effects, we chose ResNet50-FPN and ResNet101-FPN as backbone networks. The lumped loss changes and the mask loss changes during the training are shown in Figure 7. Among them, red curve indicated that the backbone network was ResNet50-FPN and the green curve represented ResNet101-FPN.
As can be seen from Figure 7, ResNet101 took fewer epochs than ResNet50 before convergence but took more time (15 min 24 s). ResNet50 took 18 s per epoch which was 6s less than ResNet101. It can be concluded that ResNet50-FPN converge faster than ResNet101-FPN as a backbone in the pig segmentation network. The pig segmentation network was tested with the test dataset. The result is shown in Figure 8.
The test samples were analysed. Among 300 samples, there were 1200 pigs including 305 pigs suspected of mounting behaviour, 113 pigs involved in interaction adhesion, 782 separated from each other. It was shown in Table 1 that in terms of pig detection, the pig recognition accuracy of ResNet50-FPN was 94.92% which was 3.26% higher than that of ResNet101-FPN network in five models trained by different train sets. In terms of pig segmentation, the MPA of ResNet50-FPN was 0.8383 which was 0.022 higher than that of ResNet101. In terms of running time, it took 0.012 more seconds for each image in the process of detection and segmentation in ResNet101. Some segmentation results based on Mask R-CNN with ResNet50-FPN backbone samples are shown in Figure 9.
Considering the above data, it could be inferred that the pig segmentation network was effective in pig detection and segmentation while partial occlusion and adhesion of the pig body occurred even if the pig bodies’ colour was similar to the background. The reason the MPA was not extremely high is due to the irregular phenotype of the mounted pig. In fact, both the mounted pig and other pigs were labelled ‘pig,’ which made the mask branch more difficult to converge. Although ResNet101-FPN has a stronger ability for complex fitting, it was too deep for one category and it easily caused the over fitting. Therefore, we chose ResNet50-FPN as the backbone of the pig segmentation network.
Considering the above data, it could be inferred that the pig segmentation network was effective in pig detection and segmentation while partial occlusion and adhesion of the pig body occurred even if the pig bodies’ colour was similar to the background. The reason the MPA was not extremely high is due to the irregular phenotype of the mounted pig. In fact, both the mounted pig and other pigs were labelled ‘pig,’ which made the mask branch more difficult to converge. Although ResNet101-FPN has a stronger ability for complex fitting, it was too deep for one category and it easily caused the over fitting. Therefore, we chose ResNet50-FPN as the backbone of the pig segmentation network.
Actually, when dealing with some images with overlapping pigs, the recognition rate while using the pig segmentation network was indeed better than that of the traditional algorithm. Figure 10 showed pigs extracted by the pig segmentation network and traditional threshold segmentation. Due to the illumination, the traditional method failed to recognize the pig in the bottom part of the image and identified some part of the bright ground as part of the pig’s body. Also, the two overlapping pigs were even regarded as one. It could be inferred that the pig segmentation network was superior to traditional methods in the recognition of overlapping pigs in the case where the background was very similar to the colour of the pig in images.
After analysis, we hold the opinion that the cause of the error was related to the stage of mounting behaviour. Considering that the images were taken from video, the samples could be extracted from any stage of mounting behaviour. During the early or middle stage of mounting, the pigs involved in the mounting behaviour had a large possibility of being recognized. However, if the image was taken during the late stage of mounting behaviour, it would be difficult to segment the outlines of the pigs because the overlapping area was extremely large.

4.2. Experiment and Evaluation of Mounting Behaviour Classifier

In order to keep the balance of positive and negative samples in the training set, we randomly reduced the negative samples in the original training set by 146, and 1054 eigenvectors were extracted for the training process of KELM. The KELM was done in Matlab (R2014a) (MathWorks, Natick, MA, USA) with an Intel Core i9-9900k CPU (Intel Corporation, Santa Clara, CA, USA). The accuracy of the test dataset classified by four different machine learning classifiers were compared and the results were shown in Table 2.
The commonly used four different machine learning algorithms were explored, including back propagation neural network (BP), Random forest (RF), Extreme learning machine (ELM) and Kernel-Extreme learning machine (KELM). Table 2 represents the performance of the four different classifiers on mounting behaviour recognition. The values were obtained by calculating the average values of the preserved test result. BP, RF, ELM and KELM respectively achieved an average MCC of 0.7917, 0.7962, 0.8158 and 0.8324. In particular, KELM showed a 1.67–4.07% higher MCC score than that of other three classifiers. By comparing with BP, RF and ELM, the accuracy and sensitivity of the KELM was 0.74–1.94% and 2.22–5.46% higher. However, the RF performed better in specificity with 1.19–2.25% compared with the other three classifiers. Hence, we considered the KELM classifier for mounting behaviour recognition of pigs.
It can be seen from the analysis results above that the proposed algorithm is efficient and has considerable accuracy (91.47%), sensitivity (95.2%), specificity (88.34%) and Matthews correlation coefficient (0.8324) in mounting behaviour recognition. It is noted that, because the pig segmentation network could not segment all the images accurately and completely, the dataset used to train KELM was not complete or accurate and some small parts were even missing. However, KELM still showed a good classification effect. Moreover, the proposed algorithm has strong portability. When the application environment changes, it will be capable of getting a considerable result after re-collecting new images for transfer learning.

5. Conclusions

In this study, we proposed a new algorithm for mounting behaviour recognition of pigs based on deep learning. The algorithm contained three parts—the pig segmentation network, eigenvectors extraction and KELM.
(a)
The pig segmentation network based on Mask R-CNN was applied and evaluated. The results showed that taking ResNet50-FPN as the backbone got better accuracy and Mean Pixel Accuracy which were 94.92% and 0.8383. This pig segmentation model can effectively solve the problem of segmentation difficulty caused by partial occlusion and adhesion of the pig body even if the pig bodies’ colour was similar to the background.
(b)
We proposed three features extracted in each image for getting the eigenvector—the perimeter of each pig, the half-body area (HBA) of each pig’s mask and the distance between the centre point of every bounding-box in one image.
(c)
The complete algorithm was evaluated by external validation and the experiment result showed that this method was efficient and the performance of the algorithm has considerable accuracy (91.47%), sensitivity (95.2%), specificity (88.34%) and Matthews correlation coefficient (0.8324) in mounting behaviour recognition.
Considering the information limitations of visual images, the deep image can be an alternative for further research into detecting mounting behaviour. The result should be even more impressive.

Author Contributions

Conceptualization, D.L. and Z.L.; methodology, D.L. and Z.L.; software, D.L. and K.Z.; validation, D.L. and K.Z.; formal analysis, D.L., K.Z. and Z.L.; investigation, D.L. and K.Z.; resources, Y.C. and D.L.; data curation D.L. and K.Z.; writing—original draft preparation, D.L.; writing—review and editing Y.C., Z.L., D.L. and K.Z.; project administration, Y.C. and D.L.; funding acquisition, Y.C.

Funding

This research was funded by the National Science and Technology Infrastructure project “National Research Facility for Phenotypic and Genotypic Analysis of Model Animals”, grant number ”4444-10099609”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Rydhmer, L.; Zamaratskaia, G.; Andersson, H.K.; Algers, B.; Guillemet, R.; Lundström, K. Aggressive and sexual behaviour of growing and finishing pigs reared in groups, without castration. Acta Agric. Scand. Sect. A 2006, 56, 109–119. [Google Scholar] [CrossRef]
  2. Hemsworth, P.H.; Tilbrook, A.J. Sexual behavior of male pigs. Horm. Behav. 2007, 52, 39–44. [Google Scholar] [CrossRef] [PubMed]
  3. Rydhmer, L.; Zamaratskaia, G.; Andersson, H.K.; Algers, B.; Lundström, K. Problems with aggressive and sexual behaviour when rearing entire male pigs. In Proceedings of the 55th Annual Meeting of the European Association for Animal Production, Bled, Slovenia, 5–9 September 2004. [Google Scholar]
  4. Gronskyte, R.; Clemmensen, L.H.; Hviid, M.S.; Kulahci, M. Pig herd monitoring and undesirable tripping and stepping prevention. Comput. Electron. Agric. 2015, 119, 51–60. [Google Scholar] [CrossRef]
  5. Kashiha, M.A.; Bahr, C.; Ott, S.; Moons, C.P.; Niewold, T.A.; Tuyttens, F.; Berckmans, D. Automatic monitoring of pig locomotion using image analysis. Livest. Sci. 2014, 159, 141–148. [Google Scholar] [CrossRef]
  6. Zhu, W.; Zhang, J. Identification of Abnormal Gait of Pigs Based on Video Analysis. In Proceedings of the 3rd International Symposium on Knowledge Acquisition and Modeling, Wuhan, China, 20–21 October 2010. [Google Scholar]
  7. Wu, Y. Detection of Pig Lame Walk Based on Star Skeleton Model. Master’s Thesis, Jiangsu University, Zhenjiang, China, 2014. [Google Scholar]
  8. Li, Z.Y. Study on Moving Object Detection and Tracking Technology in the Application of Pig Behavior Monitoring. Master’s Thesis, China Agricultural University, Beijing, China, 2013. [Google Scholar]
  9. Li, Y.Y.; Sun, L.Q.; Sun, X.X. Automatic tracking of pig feeding behavior based on particle filter with multi-feature fusion. Trans. CSAE 2017, 33 (Suppl. 1), 246–252. [Google Scholar]
  10. Zhu, W.; Wu, Z. Detection of Porcine Respiration Based on Machine Vision. In Proceedings of the 3rd International Symposium on Knowledge Acquisition and Modeling, Wuhan, China, 20–21 October 2010. [Google Scholar]
  11. Tan, H.L. Recognition Method of Identification and Drinking Behavior for Individual Pigs Based on Machine Vision. Master’s Thesis, Jiangsu University, Zhenjiang, China, 2017. [Google Scholar]
  12. Pu, X.F.; Zhu, W.X.; Lu, C.F. Sick pig behavior monitor system based on symmetrical pixel block recognition. Comput. Eng. 2009, 35, 250–252. [Google Scholar]
  13. Nasirahmadi, A.; Hensel, O. A new approach for categorizing pig lying behavior based on a Delaunay triangulation method. Animal 2017, 11, 131–139. [Google Scholar] [CrossRef] [PubMed]
  14. Lee, J.; Jin, L.; Park, D.; Chung, Y. Automatic Recognition of Aggressive Behavior in Pigs Using a Kinect Depth Sensor. Sensors 2016, 16, 631. [Google Scholar] [CrossRef] [PubMed]
  15. Mateusz, M.; Eric, T.P. Tracking of group-housed pigs using multi-ellipsoid expectation maximization. IET Comput. Vis. 2018, 12, 121–128. [Google Scholar]
  16. Zheng, C.; Zhu, X.; Yang, X.; Wang, L.; Tu, S.; Xue, Y. Automatic recognition of lactating sow postures from depth images by deep learning dector. Comput. Electron. Agric. 2018, 147, 51–63. [Google Scholar] [CrossRef]
  17. Xue, Y.; Zhu, X.; Zheng, C.; Mao, L.; Yang, A.; Tu, S.; Huang, N.; Yang, X.; Chen, P.; Zhang, N. Lactating sow postures recognition from depth image of videos based on improved Faster R-CNN. Trans. Chin. Soc. Agric. Eng. 2018, 34, 189–196. [Google Scholar]
  18. Aqing, Y.; Huasheng, H.; Xunmu, Z.; Xiaofan, Y.; Pengfei, C.; Shimei, L.; Yueju, X. Automatic recognition of sow nursing behaviour using deep learning-based segmentation and spatial and temporal feature. Biosyst. Eng. 2018, 175, 133–145. [Google Scholar]
  19. Qiumei, Y.; Deqin, X.; Sicong, L. Feeding behavior recognition for group-housed pigs with the Faster R-CNN. Comput. Electron. Agric. 2018, 155, 453–460. [Google Scholar]
  20. Qiumei, Y.; Deqin, X. Pig drinking behavior recognition based on machine vision. Trans. Chin. Soc. Agric. Mach. 2018, 49, 232–238. [Google Scholar]
  21. Sun, L.; Zou, Y.; Li, Y.; Cai, Z.; Li, Y.; Luo, B.; Liu, Y.; Li, Y. Multi target pigs tracking loss correction algorithm based on Faster R-CNN. Int. J. Agric. Biol. Eng. 2018, 11, 192–197. [Google Scholar] [CrossRef]
  22. Abozar, N.; Oliver, H.; Sandra, A.E.; Barbara, S. Automatic detection of mounting behaviours among pigs using image analysis. Comput. Electron. Agric. 2016, 124, 295–302. [Google Scholar]
  23. He, K.M.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–27 October 2017; pp. 2980–2988. [Google Scholar]
  24. Li, Y.; Qi, H.; Dai, J.; Ji, X.; Wei, Y. Fully convolutional instance-aware semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  25. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  26. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  27. Girshick, R. Fast R-CNN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  29. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. Neural Netw. 2004, 2, 985–990. [Google Scholar]
  30. Huang, G.; Zhu, Q.; Siew, C. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  31. Vinothini, B.; Sathiyamoorthy, S.; Adeel, M.; Gwang, L.; Balachandran, M.; Deok-Chun, Y. mACPred: A Support Vector Machine-Based Meta-Predictor for Identification of Anticancer Peptides. Int. J. Mol. Sci. 2019, 20, 1964. [Google Scholar]
  32. Balachandran, M.; Shaherin, B.; Tae, H.S.; Leyi, W.; Gwang, L. AtbPpred: A Robust Sequence-Based Prediction of Anti-Tubercular Peptides Using Extremely Randomized Trees. Comput. Struct. Biotechnol. J. 2019, 17, 972–981. [Google Scholar]
  33. Shaherin, B.; Balachandran, M.; Tae, H.S.; Gwang, L. SDM6A: A Web-Based Integrative Machine-Learning Framework for Predicting 6mA Sites in the Rice Genome. Mol. Ther. Nucleic Acids 2019, 18, 131–141. [Google Scholar]
Figure 1. (a) Scene of GigE Camera in the pen where the data acquired. (b) A sample frame of non-mounting behaviour. (c) A sample frame of mounting behaviour.
Figure 1. (a) Scene of GigE Camera in the pen where the data acquired. (b) A sample frame of non-mounting behaviour. (c) A sample frame of mounting behaviour.
Sensors 19 04924 g001
Figure 2. Mounting behaviour detection algorithm overview.
Figure 2. Mounting behaviour detection algorithm overview.
Sensors 19 04924 g002
Figure 3. (a) Dataset production process. (b) Dataset division.
Figure 3. (a) Dataset production process. (b) Dataset division.
Sensors 19 04924 g003
Figure 4. The architecture of pig segmentation network: (a) Overflow of the pig segmentation network procedure. (b) Close-up of the backbone. (c) Close-up of the Reign Proposal Network. (d) Close-up of the three branches.
Figure 4. The architecture of pig segmentation network: (a) Overflow of the pig segmentation network procedure. (b) Close-up of the backbone. (c) Close-up of the Reign Proposal Network. (d) Close-up of the three branches.
Sensors 19 04924 g004
Figure 5. The architecture of the pig segmentation network.
Figure 5. The architecture of the pig segmentation network.
Sensors 19 04924 g005
Figure 6. Feature extraction schematic diagram. (a) Half-body area schematic: When the pig’s mask is not cut off, half-body area (HBA) is S 1 and S 2 ; Conversely, S 11 + S 12 is the HBA when mounting behaviour occurs. (b) Center point spacing schematic: Four pig bounding-box centre points A, B, C, and D and the distance between them was defined as l 1 , l 2 , l 3 , l 4 , l 5 , and l 6 .
Figure 6. Feature extraction schematic diagram. (a) Half-body area schematic: When the pig’s mask is not cut off, half-body area (HBA) is S 1 and S 2 ; Conversely, S 11 + S 12 is the HBA when mounting behaviour occurs. (b) Center point spacing schematic: Four pig bounding-box centre points A, B, C, and D and the distance between them was defined as l 1 , l 2 , l 3 , l 4 , l 5 , and l 6 .
Sensors 19 04924 g006
Figure 7. Validation dataset loss curve during training. (a) Validation dataset loss curve (including classification, bounding-box regression and mask loss): the abscissa axis value represented epoch, and the ordinate axis value represented the loss value. (b) Mask loss curve of the validation dataset: the abscissa axis value represented epoch, and the ordinate axis value represented the loss value.
Figure 7. Validation dataset loss curve during training. (a) Validation dataset loss curve (including classification, bounding-box regression and mask loss): the abscissa axis value represented epoch, and the ordinate axis value represented the loss value. (b) Mask loss curve of the validation dataset: the abscissa axis value represented epoch, and the ordinate axis value represented the loss value.
Sensors 19 04924 g007
Figure 8. Performance of the ResNet50-FPN and the ResNet101-FPN based pig segmentation network. The accuracy represents the ratio of the images correctly recognized by the network (four pigs). MPA is the average value of each image. Time represents the average detection time of each image.
Figure 8. Performance of the ResNet50-FPN and the ResNet101-FPN based pig segmentation network. The accuracy represents the ratio of the images correctly recognized by the network (four pigs). MPA is the average value of each image. Time represents the average detection time of each image.
Sensors 19 04924 g008
Figure 9. Samples of pig segmentation result: (a) pig segmentation while mounting behaviour occurred; (b) pig segmentation without mounting behaviour; (c) incorrect segmentation.
Figure 9. Samples of pig segmentation result: (a) pig segmentation while mounting behaviour occurred; (b) pig segmentation without mounting behaviour; (c) incorrect segmentation.
Sensors 19 04924 g009aSensors 19 04924 g009b
Figure 10. Comparison between Mask R-CNN and traditional segmentation method: (a) Image segmentation by pig segmentation network. (b) Image segmentation by threshold (Otsu). (c) ellipse fitting.
Figure 10. Comparison between Mask R-CNN and traditional segmentation method: (a) Image segmentation by pig segmentation network. (b) Image segmentation by threshold (Otsu). (c) ellipse fitting.
Sensors 19 04924 g010
Table 1. Performance of the ResNet50-FPN and ResNet101-FPN based pig segmentation network.
Table 1. Performance of the ResNet50-FPN and ResNet101-FPN based pig segmentation network.
MethodsACCMPATime (s/pic)
ResNet50-FPN0.94920.83830.0746
ResNet101-FPN0.91660.81630.0866
Table 2. Comparison of the precision rate of machine learning classifiers.
Table 2. Comparison of the precision rate of machine learning classifiers.
MethodsACCMCCSNSP
BP neural network0.89530.79170.91610.8766
Random forest0.8980.79620.89740.8991
Extreme learning machine0.90730.81580.92980.8872
Kernel-ELM0.91470.83240.95200.8834

Share and Cite

MDPI and ACS Style

Li, D.; Chen, Y.; Zhang, K.; Li, Z. Mounting Behaviour Recognition for Pigs Based on Deep Learning. Sensors 2019, 19, 4924. https://doi.org/10.3390/s19224924

AMA Style

Li D, Chen Y, Zhang K, Li Z. Mounting Behaviour Recognition for Pigs Based on Deep Learning. Sensors. 2019; 19(22):4924. https://doi.org/10.3390/s19224924

Chicago/Turabian Style

Li, Dan, Yifei Chen, Kaifeng Zhang, and Zhenbo Li. 2019. "Mounting Behaviour Recognition for Pigs Based on Deep Learning" Sensors 19, no. 22: 4924. https://doi.org/10.3390/s19224924

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop