Autonomous Vision-Based Object Detection and Tracking System for Quadrotor Unmanned Aerial Vehicles
Abstract
1. Introduction
- The precision of the majority of drone systems is constrained by space limitations and financial considerations, resulting in most micro-UAVs being outfitted with a singular visible camera device.
- It is imperative to acquire a precise state estimation of the target, and the UAV must be engineered to reliably monitor the maneuvering target without any preliminary information regarding it, thereby facilitating the realization of sensitive and robust tracking applications in various scenarios.
- Consequently, the implementation of highly efficient algorithms for the detection, tracking, and estimation of pertinent states is imperative within the UAV tracking system, ensuring real-time performance despite limited onboard computation capacities. This will subsequently enable autonomous tracking in both outdoor and indoor settings.
- We present a new algorithm that effectively handles noise, shadows, light reflections, and illumination interference by separating brightness from color information in the HSV color model. The algorithm operates autonomously without any user intervention or prior knowledge about the object and quickly adapts to changes in the surrounding environment.
- An Extended Kalman Filter (EKF)-based target state estimator is presented to estimate the states of the maneuvering target at each instant. It continuously provides an estimation regarding the target’s position and re-establish tracking when the target reappears within any region of the entire frame.
- A flight control algorithm is introduced to ensure stable tracking. This technique, based on visual tracking results, was formulated to enable the UAV to track a rapidly moving sphere while ensuring minimal power usage and enhanced real-time operational efficacy. A comprehensive framework of UAV tracking a moving target, implemented in both a simulation and a physical drone (DJI Tello drone).
2. System Overview
- Step 1: The DJI Tello’s camera captures an image, which is transmitted to the laptop for detection. The position of the yellow spherical object is estimated and sent to the next module as the initial pixel position for visual tracking.
- Step 2: A switching tracking strategy is used based on the estimated states of Step 1. The visual tracking method then calculates the current pixel position of the yellow spherical object in each video frame.
- Step 3: After determining the current pixel position in Step 2, the UAV’s attitude angles and velocities for the next moment are computed by the controller module. This adjusts the UAV’s flight control parameters to ensure that it follows the yellow spherical object, keeping it within the UAV’s field of view. This process enhances the accuracy of visual tracking in Step 2.
3. Methods
3.1. Object Detection
- Hue is the attribute of color, including red, blue, yellow color, etc., and its value is from 0 degrees to 360 degrees.
- Saturation is the level of color intensity, ranging from 0 to 100.
- Value is also a representation of the lightness of the color, which also falls within a range of 0–100.
3.2. Visual Tracking
3.2.1. System Observations
3.2.2. Visual Tracking by EKF
3.3. Controller
| Algorithm 1 PID Control |
|
| Algorithm 2 Tracking Controller |
|
4. Simulation Results
5. Experimental Results
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Valavanis, K.P.; Vachtsevanos, G.J. Handbook of Unmanned Aerial Vehicles; Springer Publishing Company, Incorporated: Princeton, NJ, USA, 2014. [Google Scholar]
- Shakhatreh, H.; Sawalmeh, A.H.; Al-Fuqaha, A.; Dou, Z.; Almaita, E.; Khalil, I.; Othman, N.S.; Khreishah, A.; Guizani, M. Unmanned aerial vehicles (UAVs): A survey on civil applications and key research challenges. IEEE Access 2019, 7, 48572–48634. [Google Scholar] [CrossRef]
- Konrad, T.; Gehrt, J.J.; Lin, J.; Zweigel, R.; Abel, D. Advanced state estimation for navigation of automated vehicles. Annu. Rev. Control 2018, 46, 181–195. [Google Scholar] [CrossRef]
- Eskandaripour, H.; Boldsaikhan, E. Last-mile drone delivery: Past, present, and future. Drones 2023, 7, 77. [Google Scholar] [CrossRef]
- Telli, K.; Kraa, O.; Himeur, Y.; Ouamane, A.; Boumehraz, M.; Atalla, S.; Mansoor, W. A comprehensive review of recent research trends on unmanned aerial vehicles (uavs). Systems 2023, 11, 400. [Google Scholar] [CrossRef]
- Singla, A.; Padakandla, S.; Bhatnagar, S. Memory-based deep reinforcement learning for obstacle avoidance in UAV with limited environment knowledge. IEEE Trans. Intell. Transp. Syst. 2019, 22, 107–118. [Google Scholar] [CrossRef]
- Padhy, R.P.; Verma, S.; Ahmad, S.; Choudhury, S.K.; Sa, P.K. Deep neural network for autonomous UAV navigation in indoor corridor environments. Procedia Comput. Sci. 2018, 133, 643–650. [Google Scholar] [CrossRef]
- Wu, S.; Li, R.; Shi, Y.; Liu, Q. Vision-based target detection and tracking system for a quadcopter. IEEE Access 2021, 9, 62043–62054. [Google Scholar] [CrossRef]
- Patel, H. Designing Autonomous Drone for Food Delivery in Gazebo/Ros Based Environments; Technical Report; Binghamton University: Binghamton, NY, USA, 2022. [Google Scholar]
- Blösch, M.; Weiss, S.; Scaramuzza, D.; Siegwart, R. Vision based MAV navigation in unknown and unstructured environments. In Proceedings of the 2010 IEEE International Conference on Robotics and Automation, Anchorage, AK, USA, 3–7 May 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 21–28. [Google Scholar]
- Jiang, S.; Jiang, C.; Jiang, W. Efficient structure from motion for large-scale UAV images: A review and a comparison of SfM tools. ISPRS J. Photogramm. Remote Sens. 2020, 167, 230–251. [Google Scholar] [CrossRef]
- Sun, Y.; Weng, Y.; Luo, B.; Li, G.; Tao, B.; Jiang, D.; Chen, D. Gesture recognition algorithm based on multi-scale feature fusion in RGB-D images. IET Image Process. 2023, 17, 1280–1290. [Google Scholar] [CrossRef]
- Ding, S.; Guo, X.; Peng, T.; Huang, X.; Hong, X. Drone detection and tracking system based on fused acoustical and optical approaches. Adv. Intell. Syst. 2023, 5, 2300251. [Google Scholar] [CrossRef]
- Kang, H.C.; Han, H.N.; Bae, H.C.; Kim, M.G.; Son, J.Y.; Kim, Y.K. HSV color-space-based automated object localization for robot grasping without prior knowledge. Appl. Sci. 2021, 11, 7593. [Google Scholar] [CrossRef]
- Zhao, X.; Pu, F.; Wang, Z.; Chen, H.; Xu, Z. Detection, tracking, and geolocation of moving vehicle from uav using monocular camera. IEEE Access 2019, 7, 101160–101170. [Google Scholar] [CrossRef]
- Yuan, Y.; Wu, Y.; Zhao, L.; Chen, J.; Zhao, Q. DB-Tracker: Multi-Object Tracking for Drone Aerial Video Based on Box-MeMBer and MB-OSNet. Drones 2023, 7, 607. [Google Scholar] [CrossRef]
- Liang, J.; Yi, P.; Li, W.; Zuo, J.; Zhu, B.; Wang, Y. Real-time visual tracking design for an unmanned aerial vehicle in cluttered environments. Optim. Control Appl. Methods 2025, 46, 476–492. [Google Scholar] [CrossRef]
- Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H. Fully-convolutional siamese networks for object tracking. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Heidelberg/Heidelberg, Germany, 2016; pp. 850–865. [Google Scholar]
- Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High performance visual tracking with siamese region proposal network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8971–8980. [Google Scholar]
- Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. Siamrpn++: Evolution of siamese visual tracking with very deep networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4282–4291. [Google Scholar]
- Rodríguez-Canosa, G.R.; Thomas, S.; Del Cerro, J.; Barrientos, A.; MacDonald, B. A real-time method to detect and track moving objects (DATMO) from unmanned aerial vehicles (UAVs) using a single camera. Remote Sens. 2012, 4, 1090–1111. [Google Scholar] [CrossRef]
- Kang, D.; Cha, Y.J. Autonomous UAVs for structural health monitoring using deep learning and an ultrasonic beacon system with geo-tagging. Comput. Aided Civ. Infrastruct. Eng. 2018, 33, 885–902. [Google Scholar] [CrossRef]
- Ali, R.; Kang, D.; Suh, G.; Cha, Y.J. Real-time multiple damage mapping using autonomous UAV and deep faster region-based neural networks for GPS-denied structures. Autom. Constr. 2021, 130, 103831. [Google Scholar] [CrossRef]
- Waqas, A.; Kang, D.; Cha, Y.J. Deep learning-based obstacle-avoiding autonomous UAVs with fiducial marker-based localization for structural health monitoring. Struct. Health Monit. 2024, 23, 971–990. [Google Scholar] [CrossRef]
- Cheng, H.; Lin, L.; Zheng, Z.; Guan, Y.; Liu, Z. An autonomous vision-based target tracking system for rotorcraft unmanned aerial vehicles. In Proceedings of the 2017 IEEE/RSJ international conference on intelligent robots and systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1732–1738. [Google Scholar]
- Kumar, M.; Mondal, S. Recent developments on target tracking problems: A review. Ocean Eng. 2021, 236, 109558. [Google Scholar] [CrossRef]
- Wan, M.; Gu, G.; Qian, W.; Ren, K.; Maldague, X.; Chen, Q. Unmanned aerial vehicle video-based target tracking algorithm using sparse representation. IEEE Internet Things J. 2019, 6, 9689–9706. [Google Scholar] [CrossRef]
- Li, Z.; Hovakimyan, N.; Dobrokhodov, V.; Kaminer, I. Vision-based target tracking and motion estimation using a small UAV. In Proceedings of the 49th IEEE Conference on Decision and Control (CDC), Atlanta, GA, USA, 15–17 December 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 2505–2510. [Google Scholar]
- Falanga, D.; Zanchettin, A.; Simovic, A.; Delmerico, J.; Scaramuzza, D. Vision-based autonomous quadrotor landing on a moving platform. In Proceedings of the 2017 IEEE International Symposium on Safety, Security and Rescue Robotics (SSRR), Shanghai, China, 11–13 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 200–207. [Google Scholar]
- Saavedra-Ruiz, M.; Pinto-Vargas, A.M.; Romero-Cano, V. Monocular Visual Autonomous Landing System for Quadcopter Drones Using Software in the Loop. IEEE Aerosp. Electron. Syst. Mag. 2021, 37, 2–16. [Google Scholar] [CrossRef]
- Cabrera-Ponce, A.A.; Martinez-Carranza, J. A vision-based approach for autonomous landing. In Proceedings of the 2017 Workshop on Research, Education and Development of Unmanned Aerial Systems (RED-UAS), Linköping, Sweden, 3–5 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 126–131. [Google Scholar]
- Coşkun, M.; Ünal, S. Implementation of tracking of a moving object based on camshift approach with a UAV. Procedia Technol. 2016, 22, 556–561. [Google Scholar] [CrossRef]
- Lee, D.; Park, W.; Nam, W. Autonomous landing of micro unmanned aerial vehicles with landing-assistive platform and robust spherical object detection. Appl. Sci. 2021, 11, 8555. [Google Scholar] [CrossRef]
- Keshun, Y.; Yingkui, G.; Yanghui, L.; Yajun, W. A novel physical constraint-guided quadratic neural networks for interpretable bearing fault diagnosis under zero-fault sample. Nondestruct. Test. Eval. 2025, 40, 1–31. [Google Scholar] [CrossRef]
- Chen, Z.; Yang, J.; Li, F.; Feng, Z.; Chen, L.; Jia, L.; Li, P. Foreign Object Detection Method for Railway Catenary Based on a Scarce Image Generation Model and Lightweight Perception Architecture. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 1-1. [Google Scholar] [CrossRef]
- Malik, M.H.; Zhang, T.; Li, H.; Zhang, M.; Shabbir, S.; Saeed, A. Mature tomato fruit detection algorithm based on improved HSV and watershed algorithm. IFAC-PapersOnLine 2018, 51, 431–436. [Google Scholar] [CrossRef]
- Boumehraz, M.; Habba, Z.; Hassani, R. Vision based tracking and interception of moving target by mobile robot using fuzzy control. J. Appl. Eng. Sci. Technol. 2018, 4, 159–165. [Google Scholar] [CrossRef]
- Foehn, P.; Brescianini, D.; Kaufmann, E.; Cieslewski, T.; Gehrig, M.; Muglikar, M.; Scaramuzza, D. Alphapilot: Autonomous drone racing. Auton. Robot. 2022, 46, 307–320. [Google Scholar] [CrossRef] [PubMed]
- Lei, B.; Liu, B.; Wang, C. Robust geometric control for a quadrotor UAV with extended Kalman filter estimation. Actuators 2024, 13, 205. [Google Scholar] [CrossRef]
- Ayala, A.; Portela, L.; Buarque, F.; Fernandes, B.J.; Cruz, F. UAV control in autonomous object-goal navigation: A systematic literature review. Artif. Intell. Rev. 2024, 57, 125. [Google Scholar] [CrossRef]
- Hulens, D.; Van Ranst, W.; Cao, Y.; Goedemé, T. Autonomous visual navigation for a flower pollination drone. Machines 2022, 10, 364. [Google Scholar] [CrossRef]
- Schnipke, E.; Reidling, S.; Meiring, J.; Jeffers, W.; Hashemi, M.; Tan, R.; Nemati, A.; Kumar, M. Autonomous navigation of UAV through GPS-denied indoor environment with obstacles. In Proceedings of the AIAA Infotech@Aerospace, Kissimmee, FL, USA, 5–9 January 2015; p. 0715. [Google Scholar]











Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gharsa, O.; Touba, M.M.; Boumehraz, M.; Agram, N. Autonomous Vision-Based Object Detection and Tracking System for Quadrotor Unmanned Aerial Vehicles. Sensors 2025, 25, 6403. https://doi.org/10.3390/s25206403
Gharsa O, Touba MM, Boumehraz M, Agram N. Autonomous Vision-Based Object Detection and Tracking System for Quadrotor Unmanned Aerial Vehicles. Sensors. 2025; 25(20):6403. https://doi.org/10.3390/s25206403
Chicago/Turabian StyleGharsa, Oumaima, Mostefa Mohamed Touba, Mohamed Boumehraz, and Nacira Agram. 2025. "Autonomous Vision-Based Object Detection and Tracking System for Quadrotor Unmanned Aerial Vehicles" Sensors 25, no. 20: 6403. https://doi.org/10.3390/s25206403
APA StyleGharsa, O., Touba, M. M., Boumehraz, M., & Agram, N. (2025). Autonomous Vision-Based Object Detection and Tracking System for Quadrotor Unmanned Aerial Vehicles. Sensors, 25(20), 6403. https://doi.org/10.3390/s25206403

