Analysis of Various Machine Learning Algorithms for Using Drone Images in Livestock Farms
Abstract
:1. Introduction
2. Materials and Methods
2.1. Machine Learning Model Based on UAV Data
2.2. Drone Technology Comparison and Solution Survey
2.3. Data Engineering
2.3.1. Data Process
2.3.2. Data Collection
2.3.3. Data Pre-Processing
2.3.4. Data Transformation
2.3.5. Data Preparation
2.3.6. Data Statistics
2.4. Machine Learning Models
2.4.1. Model Proposals
- 1.
- Improved YOLOv7:
- 2.
- Improved Mask RCNN:
- 3.
- Improved Faster RCNN:
- 4.
- Improved SSD (Single-Shot Multi-Box Detector):
- 5.
- Object Tracking Algorithm YOLOv7+ DeepSort:
2.4.2. Model Comparison and Justification
2.4.3. Model Evaluation Methods
2.4.4. Model Validation and Evaluation Results
2.5. Smart Surveillance System for Livestock Farms: Design, Implementation, Results
2.5.1. System Development and Design
System Development
System Boundary and Use Cases
System High-Level Data Analytics Requirements
2.5.2. System Design
System Architecture and Infrastructure
System Supporting Platforms and Cloud Environment
2.5.3. System Development
AI and Machine Learning Model Development
Implement Designed System
Input and Output Requirements, Supporting Systems and Solution APIs
3. Analysis of Model Execution and Evaluation Results
4. Discussion
4.1. Achievements
- Augmentation of the image and video data—Enables the user to take a closer look at the objects present on the farm for surveillance purposes;
- Livestock Detection—Classes like cow, sheep, and horse have been identified along with their behavior such as standing, sitting, and sleeping;
- In the output, livestock has been classified from an aerial view with appropriate class labels and confidence score;
- Object Tracking—The farm objects detected in the video are being tracked in each frame;
- Object counting results of various classes displayed in the video frames;
- Detection of other farm objects like buildings and farm equipment like trucks;
- Livestock behavior classification—Behavior of livestock is detected such as whether the livestock is standing or sleeping.
4.2. Constraints
- Optimal drone height—While training the model, it was observed that videos taken at heights more than 200–250 were not capturing the objects clearly, hence leading to inaccurate results. However, when videos were taken at 100–150 ft altitude, the objects were seen clearly and the model yielded better performance. Therefore, it was necessary to define the optimal drone height for collecting the data.
- A large dataset with high-resolution images is required for training the model on farming equipment. Due to the different angles, light settings and appearances of the truck can vary. To take care of these scenarios, we performed augmentations where brightness was increased to enhance the classifier results under different conditions.
- Training a deep Learning with a huge dataset is extremely time-consuming and challenging, even with the premium accounts of Google Pro.
4.3. System Quality Evaluation of Model Functions and Performance
4.4. System Visualization
- Step 1:
- Upload the drone-captured video and select the desired options.
- Step 2:
- Processed input video with detection and tracking are shown on the UI.
5. Conclusions
5.1. Summary
5.2. Prospect and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Data Availability Statement
Conflicts of Interest
References
- Bosquet, B.; Mucientes, M.; Brea, V.M. STDnet-ST: Spatio-temporal ConvNet for small object detection. Pattern Recognit. 2021, 116, 107929. [Google Scholar] [CrossRef]
- Zinke-Wehlmann, C.; Charvát, K. Introduction of smart agriculture. In Big Data in Bioeconomy; Springer International Publishing: Cham, Switzerland, 2021; pp. 187–190. [Google Scholar]
- Boruah, P. RetinaNet: The Beauty of Focal Loss—Towards Data Science.Medium. Available online: https://towardsdatascience.com/retinanet-the-beauty-of-focal-loss-e9ab132f2981 (accessed on 6 January 2022).
- Meng, X.; Chen, Y.; Xin, L. Adaptive decision-level fusion and complementary mining for visual object tracking with deeper networks. J. Electron. Imaging 2020, 4, 29. [Google Scholar] [CrossRef]
- Canuma, P. Data Science Project Management [The New Guide For ML Teams]. Neptune.Ai. Available online: https://neptune.ai/blog/data-science-project-management (accessed on 31 January 2022).
- Chen, C.; Zhang, Y.; Lv, Q.; Wei, S.; Wang, X.; Sun, X.; Dong, J. Rrnet: A hybrid detector for object detection in drone-captured images. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
- Li, C.; Yang, T.; Zhu, S.; Chen, C.; Guan, S. Density map guided object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 190–191. [Google Scholar]
- Commercial Drone Market Size, Share & Trends Report. Global-Commercial-Drones-Market. Historical Range: 2021–2028. 2021. Available online: https://www.grandviewresearch.com/industry-analysis/global-commercial-drones-market (accessed on 11 March 2024).
- Dike, H.U.; Zhou, Y. A robust quadruplet and faster region-based CNN for UAV video-based multiple object tracking in crowded environment. Electronics 2021, 10, 795. [Google Scholar] [CrossRef]
- Discover a Wide Range of Drone Datasets, senseFly. 2021. Available online: https://www.sensefly.com/education/datasets/?dataset=1502&industries%5B%5D=2 (accessed on 11 March 2024).
- Robicquet, A.; Sadeghian, A.; Alahi, A.; Savarese, S. Learning social etiquette: Human trajectory understanding in crowded scenes. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Part VIII 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 549–565. [Google Scholar]
- Li, W.; Mu, J.; Liu, G. Multiple object tracking with motion and appearance cues. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019; pp. 1–9. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Part I 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
- Li, M.; Zhang, Z.; Lei, L.; Wang, X.; Guo, X. Agricultural greenhouses detection in high-resolution satellite images based on convolutional neural networks: Comparison of faster R-CNN, YOLO v3 and SSD. Sensors 2020, 20, 4938. [Google Scholar] [CrossRef]
- Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8971–8980. [Google Scholar]
- Mandal, M.; Kumar, L.K.; Vipparthi, S.K. Mor-uav: A benchmark dataset and baselines for moving object recognition in uav videos. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2626–2635. [Google Scholar]
- Liu, M.; Wang, X.; Zhou, A.; Fu, X.; Ma, Y.; Piao, C. Uav-yolo: Small object detection on unmanned aerial vehicle perspective. Sensors 2020, 20, 2238. [Google Scholar] [CrossRef] [PubMed]
- Mueller, M.; Smith, N.; Ghanem, B. A benchmark and simulator for uav tracking. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Part I 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 445–461. [Google Scholar]
- Mevadata. Multiview Extended Video. 2021. Available online: https://mevadata.org/ (accessed on 11 March 2024).
- Imamura, S.; Zin, T.T.; Kobayashi, I.; Horii, Y. Automatic evaluation of cow’s body-condition-score using 3D camera. In Proceedings of the 2017 IEEE 6th Global Conference on Consumer Electronics (GCCE), Nagoya, Japan, 24–27 October 2017; pp. 1–2. [Google Scholar]
- Zhu, P.; Sun, Y.; Wen, L.; Feng, Y.; Hu, Q. Drone based rgbt vehicle detection and counting: A challenge. arXiv 2020, arXiv:2003.02437. [Google Scholar]
- Perera, A.G.; Law, Y.W.; Chahl, J. Drone-action: An outdoor recorded drone video dataset for action recognition. Drones 2019, 3, 82. [Google Scholar] [CrossRef]
- Papers with Code UAVDT Dataset. UAVDT Benchmark Dataset. 2018. Available online: https://paperswithcode.com/dataset/uavdt (accessed on 11 March 2024).
- Zhang, R.; Shao, Z.; Huang, X.; Wang, J.; Li, D. Object detection in UAV images via global density fused convolutional network. Remote Sens. 2020, 12, 3140. [Google Scholar] [CrossRef]
- Sun, S.; Akhtar, N.; Song, H.; Mian, A.; Shah, M. Deep affinity network for multiple object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 104–119. [Google Scholar] [CrossRef]
- Staaker. Best Surveillance Drones of 2022: Top Brands Reviewed. Staaker.Com. Available online: https://staaker.com/best-surveillance-drones/ (accessed on 13 November 2021).
- Klauser, F. Surveillance farm: Towards a research agenda on big data agriculture. Surveill. Soc. 2018, 16, 370–378. [Google Scholar] [CrossRef]
- Walter, A.; Finger, R.; Huber, R.; Buchmann, N. Smart farming is key to developing sustainable agriculture. Proc. Natl. Acad. Sci. USA 2017, 114, 6148–6150. [Google Scholar] [CrossRef] [PubMed]
- Triantafyllou, A.; Tsouros, D.C.; Sarigiannidis, P.; Bibi, S. An architecture model for smart farming. In Proceedings of the 2019 15th International Conference on Distributed Computing in Sensor Systems (DCOSS), Santorini, Greece, 29–31 May 2019; pp. 385–392. [Google Scholar]
- Mascarello, L.N.; Quagliotti, F.; Ristorto, G. A feasibility study of an harmless tiltrotor for smart farming applications. In Proceedings of the 2017 International Conference on Unmanned Aircraft Systems (ICUAS), Miami, FL, USA, 13–16 June 2017; pp. 1631–1639. [Google Scholar]
- Koshy, S.S.; Sunnam, V.S.; Rajgarhia, P.; Chinnusamy, K.; Ravulapalli, D.P.; Chunduri, S. Application of the internet of things (IoT) for smart farming: A case study on groundnut and castor pest and disease forewarning. CSI Trans. ICT 2018, 6, 311–318. [Google Scholar] [CrossRef]
- Zin, T.T.; Seint, P.T.; Tin, P.; Horii, Y.; Kobayashi, I. Body condition score estimation based on regression analysis using a 3D camera. Sensors 2020, 20, 3705. [Google Scholar] [CrossRef] [PubMed]
- Mukherjee, A.; Misra, S.; Sukrutha, A.; Raghuwanshi, N.S. Distributed aerial processing for IoT-based edge UAV swarms in smart farming. Comput. Netw. 2020, 167, 107038. [Google Scholar] [CrossRef]
- Wang, C.; Bochkovskiy, A.; Liao, H. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
- Hui, J. SSD Object Detection: Single Shot MultiBox Detector for Real-Time Processing. Date of Application: 10 October 2021. 2018. Available online: https://jonathan-hui.medium.com/ssd-object-detection-single-shot-multibox-detector-for-real-time-processing-9bd8deac0e06 (accessed on 11 March 2024).
- Noe, S.M.; Zin, T.T.; Tin, P.; Kobayashi, I. Automatic detection and tracking of mounting behavior in cattle using a deep learning-based instance segmentation model. Int. J. Innov. Comput. Inf. Control 2022, 18, 211–220. [Google Scholar]
- Tian, G.; Liu, J.; Yang, W. A dual neural network for object detection in UAV images. Neurocomputing 2021, 443, 292–301. [Google Scholar] [CrossRef]
- Aburasain, R.; Edirisinghe, E.A.; Albatay, A. Drone-based cattle detection using deep neural networks. In Intelligent Systems and Applications: Proceedings of the 2020 Intelligent Systems Conference (IntelliSys) Volume 1; Springer: Berlin/Heidelberg, Germany, 2021; pp. 598–611. [Google Scholar]
- Chandan, G.; Jain, A.; Jain, H. Real time object detection and tracking using Deep Learning and OpenCV. In Proceedings of the 2018 International Conference on Inventive Research in Computing Applications (ICIRCA), Coimbatore, India, 11–12 July 2018; pp. 1305–1308. [Google Scholar]
- Huang, W.; Zhou, X.; Dong, M.; Xu, H. Multiple objects tracking in the UAV system based on hierarchical deep high-resolution network. Multimed. Tools Appl. 2021, 80, 13911–13929. [Google Scholar] [CrossRef]
- Lu, J.; Yan, W.Q.; Nguyen, M. Human behaviour recognition using deep learning. In Proceedings of the 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Auckland, New Zealand, 27–30 November 2018; pp. 1–6. [Google Scholar]
- Kragh, M.; Jørgensen, R.N.; Pedersen, H. Object detection and terrain classification in agricultural fields using 3D lidar data. In Proceedings of the Computer Vision Systems: 10th International Conference, ICVS 2015, Copenhagen, Denmark, 6–9 July 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 188–197. [Google Scholar]
- Micheal, A.A.; Vani, K.; Sanjeevi, S.; Lin, C.H. Object detection and tracking with UAV data using deep learning. J. Indian Soc. Remote. Sens. 2021, 49, 463–469. [Google Scholar] [CrossRef]
- Pan, S.; Tong, Z.; Zhao, Y.; Zhao, Z.; Su, F.; Zhuang, B. Multi-object tracking hierarchically in visual data taken from drones. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
- Yang, S.; Chen, H.; Xu, F.; Li, Y.; Yuan, J. High-performance UAVs visual tracking based on siamese network. Vis. Comput. 2022, 38, 2107–2123. [Google Scholar] [CrossRef]
- Xu, Z.; Shi, H.; Li, N.; Xiang, C.; Zhou, H. Vehicle detection under UAV based on optimal dense YOLO method. In Proceedings of the 2018 5th International Conference on Systems and Informatics (ICSAI), Nanjing, China, 10–12 November 2018; pp. 407–411. [Google Scholar]
- Tong, Z.; Jieyu, L.; Zhiqiang, D. UAV target detection based on RetinaNet. In Proceedings of the 2019 Chinese Control And Decision Conference (CCDC), Nanchang, China, 3–5 June 2019; pp. 3342–3346. [Google Scholar]
- Yadav, V.K.; Yadav, P.; Sharma, S. An Efficient YOLOv7 and Deep Sort are Used in a Deep Learning Model for Tracking Vehicle and Detection. J. Xi’an Shiyou Univ. Nat. Sci. Ed. 2022, 18, 759–763. [Google Scholar]
- Yang, F.; Zhang, X.; Liu, B. Video object tracking based on YOLOv7 and DeepSORT. arXiv 2022, arXiv:2207.12202. [Google Scholar]
- Bondi, E.; Jain, R.; Aggrawal, P.; Anand, S.; Hannaford, R.; Kapoor, A.; Piavis, J.; Shah, S.; Joppa, L.; Dilkina, B.; et al. BIRDSAI: A dataset for detection and tracking in aerial thermal infrared videos. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1747–1756. [Google Scholar]
- Zhang, H.; Wang, G.; Lei, Z.; Hwang, J.N. Eye in the sky: Drone-based object tracking and 3d localization. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 899–907. [Google Scholar]
- GitHub—VisDrone/VisDrone-Dataset: The Dataset for Drone-Based Detection and Tracking Is Released, Including Both Image/Video, and Annotations. VisDrone2020 Challenge. 2020. Available online: https://github.com/VisDrone/VisDrone-Dataset (accessed on 11 March 2024).
- Zuo, M.; Zhu, X.; Chen, Y.; Yu, J. Survey of Object Tracking Algorithm Based on Siamese Network. J. Phys. Conf. Ser. 2022, 2203, 012035. [Google Scholar] [CrossRef]
Reference | Purpose | Model | Real-Time Prediction | Final Outcome | Training Data |
---|---|---|---|---|---|
[37] | High-quality detection of small targets | VGG | N | A 4.58 increase in precision rate was recorded compared to state-of-the-art method | VisDrone, UAVDT, MS COCO |
[38] | Detection of fauna | SSD-500, YOLOv3 | N | mAP—84.7 F1 score—93% Accuracy—89% | 1300 collected images |
[39] | Real-time object detection and tracking | SSD, MobileNets | Y | Confidence level—98.68% | 21 objects |
[40] | Dealing with small target detection and class imbalance challenges | HDHNet, Cascade RCNN, DeepSort | N | MOTA has an improvement of over 4% compared to to other methods | Visdrone2019 |
[9] | Detect multiple objects and track their trajectories | DQN, RCNN, SSD | N | 39% of MOTA, 75% of MOTP | UAVDT, Stanford Drone |
[41] | Human behavior recognition | YOLOv3 | Y | Accuracy—80.20% with 15 fps | Weizmann, INRIA, UT-Interaction |
[42] | Target detection and terrain classification | SVM classifier | Y | Accuracy—91.6% | Hand labeled data from local working farm dataset |
[43] | Target detection and tracking | DSOD, LTSM | N | Precision—96.13%, Recall—95.28% | VisDrone2019 |
[44] | Multi-object tracking hierarchically | SOT, HMTT | N | Mean Average Precision (mAP)—0.75 | VisDrone2019 |
[45] | High-performance UAVs visual tracking | SiamRPN, DAFSiamRPN | Y | Accuracy = 73.5%, precision = 54.1% | ILSVRC2017 VID, the Youtube-BB 904 video sequences |
[46] | Vehicle detection | YOLOv2, YOLOv3 | Y | mAP = 69.6%, FPS = 43 | The aerial dataset collected by a small UAV |
[47] | UAV target detection | Retina Net, Resnet101 | Y | mAp = 92.7%, FPS = 66.7 | PASCAL Dataset |
[48] | Tracking and detection of vehicles | YOLOv7, YOLOv5, DeepSort | Y | Tracking precision = 82.08%, FPS = 20.2 | COCO128, COCO 2017 and self-made custom dataset (car, bus, and truck) |
Model Name | Advantages | Disadvantages |
---|---|---|
Faster R-CNN [9] | - High speed | |
- More accurate | - Cannot be customized on a specific object detection task | |
- Requires less storage | ||
FS-SSD [35] | - At multiple scales, feature maps enhance the chances of object detection. | - Cannot detect small objects accurately |
YOLOv7 [34,48,49] | - High accuracy | |
- Highly generalizable | - Slow detection speed | |
- Multi-class prediction | ||
- It can detect objects in real time | ||
Retina Net [3] | - It can detect objects of small sizes | - The real-time detection criterion is not met |
RRNet [15] | - Can detect small objects even in a dense environment | - Challenging to suppress the high contrast and completely detect important features |
DMNet [25] | - Crops an image using an object density map to account for object distribution. | - Complex model |
STDnet-ST [1] | - Better accuracy | - Cannot detect tiny objects effectively |
GDF-Net [14] | - Detects multi-scale objects | - Unable to properly detect occluded objects |
Model | Advantages | Disadvantages |
---|---|---|
HDHNet [4] | - Excellent resolution is maintained. | - Incapable of managing real-time tracking |
- High network speed. | ||
DeepSort [43] | - Stable model. | - Lack of accuracy and robustness |
- Can track an object even with a high speed rate. | ||
Flow-Tracker [40,50] | - Can handle missing detection and fragmentation problems. | - ID switches, error detection |
TNT [18,51] | - Can handle occlusion data | - Effected by group plane estimation |
Deep Quadruplet Network (DQN) [38] | - Can track multiple objects | - Long training time |
Hierarchical Multi-Target Tracker [52,53] | - Assists in multiple object tracking in UAV dataset with improved accuracies as compared to the baseline models. | - Highly complex model and takes longer for training |
Deep Affinity Network (DAN) [40] | - Can track multiple objects with less training model training time. | - It is computationally intense and difficult to track smaller objects |
- Identifies objects at greater occlusion |
Drone | Cost | Flight Time | Control Range | Camera | Max Speed | Obstacle Avoidance |
---|---|---|---|---|---|---|
DJI Tello | $99 | 13 min | 100 m | 720 P HD | 17.8 mph | Yes |
Parrot Bebop 2 | $279 | 25 min | 300 m | FHD 1080 p | 37 mph | Yes |
Yuneec Typhoon Q500 4K | $484 | 25 min | 122 m | 4 k | 18 mph | Yes |
DJI Mavic Air 2 | $799 | 34 min | 10,000 m | 48 MP Photo | 42.3 mph | Yes |
DJI Phantom 4 | $2000 | 28 min | 5000 m | 12 MP 4 K camera | 45 mph | Yes |
Dataset Name | Data Count |
---|---|
DJI Phantom 4 (data collected by us) | 1140 |
Roboflow Universe | 13,118 |
Visdrone2019 | 5218 |
Total | 19,476 |
Dataset Name | Raw Data | Pre-processed Data | Transformed Data | Data Preparation | ||
---|---|---|---|---|---|---|
Training | Validation | Testing | ||||
Visdrone | 5218 | 5218 | 48,045 | 28,827 | 9609 | 9609 |
Roboflow | 13,118 | 13,118 | 102,000 | 61,200 | 20,400 | 20,400 |
DJI Phantom 4 | 1140 | 1140 | 1500 | 900 | 300 | 300 |
Total | 19,476 | 19,476 | 58,428 | 46,742 | 5843 | 5843 |
Model | Justification | Input–Accuracy | Merits and Drawbacks |
---|---|---|---|
YOLOv7 | YOLOv7, also referred to as Trainable Bag-of-Freebies, establishes a new standard for real-time object detectors. This most recent YOLO version features unique “compound scaling” and “extend” techniques that efficiently use computations and parameters. | Sheep—85.80% Cow—75.38% Horse—76.48% Truck—74.19% Person—66.4% Car—79.25% | Merits: able to detect tiny objects; high speed; high accuracy; ease of use. Drawback: requires high amounts of training |
Faster RCNN | Faster RCNN is a multiple-stage object detection model. It consists of two modules: RPN and Fast RCNN. To detect bounding boxes, the RPN module is used, whereas Fast RCNN is used for classifying the objects and fine-tuning the bounding boxes. | Sheep—66.54% Cow—60.2% Horse—67.18% Truck—44.43% Human—60.50% Car—68.21% | Merits: high accuracy; small object detection. Drawback: complex architecture where bounding box and classification are performed separately. |
SSD | SSD uses VGG16 as the base model and adds convolution layers. Feature extraction is also performed. It uses default boxes for location detection: L1 loss for localization and SoftMax loss for the confidence. | Sheep—38% Cow—36.2% Horse—41% Truck—26.5% Human—30.1% Car—45.12% | Merits: bounding boxes; real-time object detection; multi-scale feature maps improve the accuracy. Drawbacks: image resizing is required; not good for small-scale objects; high classification error. |
Mask R-CNN | Mask R-CNN is based on the basic architecture of the Faster R-CNN model. In addition, it consists of image segmentation as an additional layer. Segmentation masks are generated for each object. | Sheep—40.04% Cow—41.9% Horse—39.12% Truck—30.05% Human—33.8% Car—56.7% | Merits: accurate detection; high detection precision; object mask; pixel-to-pixel alignment. Drawbacks: requires a high training time; requires high processing configurations |
DeepSort | DeepSort combines association metrics and motion and appearance descriptors. It tracks the objects based on appearance and not only velocity and motion. | Able to track the object but not precision in counting the objects in the video. | Merits: time-efficient; robust; easier to implement. Drawback: complex architecture where bounding box and classification are performed separately. |
Model | Object | Training | Validation | Testing | mAP | Testing Time/Image (s) |
---|---|---|---|---|---|---|
YOLOv7+ DeepSort | Sheep | 1203 | 350 | 350 | 70.21% | 17.20 |
Cow | 1147 | 289 | 389 | 15.38% | 19.10 | |
Person | 1543 | 514 | 514 | 76.23% | 15.14 | |
Horse | 1924 | 584 | 584 | 61.23% | 16.13 | |
Farming Machineries | 516 | 172 | 172 | 63.03% | 16.42 | |
Faster RCNN | Sheep | 1203 | 350 | 350 | 58.1% | 24 |
Cow | 1147 | 389 | 389 | 60.2% | 22 | |
Person | 1543 | 514 | 514 | 67.20% | 21 | |
Horse | 1924 | 584 | 584 | 54.60% | 25 | |
Farming Machineries | 516 | 172 | 172 | 54.70% | 23 | |
Mask RCNN | Sheep | 1203 | 350 | 350 | 67.00% | 27.2 |
Cow | 1147 | 389 | 389 | 62.20% | 58.8 | |
Person | 1543 | 514 | 514 | 68.50% | 25.3 | |
Horse | 1924 | 584 | 584 | 55.01% | 26.4 | |
Farming Machineries | 516 | 172 | 172 | 62.00% | 24.5 | |
SSD | Sheep | 1203 | 350 | 350 | 56% | 20 |
Cow | 1147 | 389 | 389 | 44% | 25 | |
Person | 1543 | 514 | 514 | 45% | 27 | |
Horse | 1924 | 584 | 584 | 39% | 22 | |
Farming Machineries | 516 | 172 | 172 | 61% | 21 |
YOLO v7+DeepSort | Faster RCNN | ||||
---|---|---|---|---|---|
Feature | Classes | Previous Accuracy (%) | Current Accuracy (%) | Previous Accuracy (%) | Current Accuracy (%) |
Cattle Classification and Counting | Sheep | 85.50 | 89.54 | 66.54 | 71.57 |
Cow | 75.38 | 82.9 | 60.2 | 66.5 | |
Horse | 76.48 | 84.6 | 67.18 | 72.05 | |
Human Detection | Person | 66.4 | 80.6 | 60.50 | 69.34 |
Vehicle Detection | Truck | 74.19 | 81.3 | 44.43 | 62.12 |
Vehicle Detection | Car | 79.25 | 86.67 | 61.28 | 74.51 |
Behavior Detection | cow lying/standing | 79.02 | 87.6 | 45.50 | 68.51 |
Metrics | Livestock Classification and Tracking | Behavior Detection | Human Detection and Tracking | Vehicle Detection and Tracking |
---|---|---|---|---|
mAP@0.50 | 97.6 | 96.1 | 88.6 | 81.3 |
Precision | 97.5 | 96.2 | 80.9 | 83.7 |
Recall | 97.2 | 92.1 | 88.9 | 75.3 |
F-1 Score | 97.34 | 94.10 | 84.71 | 77.0 |
Detection Time (in seconds) | 0.71 | 0.83 | 0.52 | 0.75 |
Per Iteration Time | 1.21 | 1.33 | 1.22 | 1.25 |
Features | Environment | Device Specs | Average Run (In seconds) | Per Iteration Time |
---|---|---|---|---|
Livestock Classification and Counting | Google Colab Pro GPU: Tesla V100-SXM2 Memory: 54 GB RAM | HP (Windows) Processor: 11th Generation Intel® Core™ i5-1135G7 Memory: 12 GB | 0.71 | 1.18 |
Intrusion Detection | Google Colab Pro GPU: Tesla V100-SXM2 Memory: 54 GB RAM | MacBook Air Processor: 1.6 GHz Dual-Core Intel Core i5 Memory: 8 GB | 0.80 | 1.22 |
Vehicle Detection | Google Colab Pro GPU: Tesla V100-SXM2 Memory: 54 GB RAM | Dell Intel(R) Core (TM) i5-8250U CPU @ 1.60 GHz RAM-16 GB | 0.75 | 1.25 |
Livestock Monitoring | Google Colab Pro GPU: Tesla V100-SXM2 Memory: 54 GB RAM | MacBook Air Processor: 1.6 GHz Dual-Core Intel Core i5 Memory: 8 GB | 0.85 | 1.6 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gao, J.; Bambrah, C.K.; Parihar, N.; Kshirsagar, S.; Mallarapu, S.; Yu, H.; Wu, J.; Yang, Y. Analysis of Various Machine Learning Algorithms for Using Drone Images in Livestock Farms. Agriculture 2024, 14, 522. https://doi.org/10.3390/agriculture14040522
Gao J, Bambrah CK, Parihar N, Kshirsagar S, Mallarapu S, Yu H, Wu J, Yang Y. Analysis of Various Machine Learning Algorithms for Using Drone Images in Livestock Farms. Agriculture. 2024; 14(4):522. https://doi.org/10.3390/agriculture14040522
Chicago/Turabian StyleGao, Jerry, Charanjit Kaur Bambrah, Nidhi Parihar, Sharvaree Kshirsagar, Sruthi Mallarapu, Hailong Yu, Jane Wu, and Yunyun Yang. 2024. "Analysis of Various Machine Learning Algorithms for Using Drone Images in Livestock Farms" Agriculture 14, no. 4: 522. https://doi.org/10.3390/agriculture14040522
APA StyleGao, J., Bambrah, C. K., Parihar, N., Kshirsagar, S., Mallarapu, S., Yu, H., Wu, J., & Yang, Y. (2024). Analysis of Various Machine Learning Algorithms for Using Drone Images in Livestock Farms. Agriculture, 14(4), 522. https://doi.org/10.3390/agriculture14040522