UAV-Based Transport Management for Smart Cities Using Machine Learning
Abstract
Highlights
- A smart transport management system based on UAV data integrating advanced machine learning and deep learning techniques is proposed to enhance road anomaly detection and severity classification.
- The system employs a comprehensive multi-stage framework, integrating a high-precision obstacle detection model, six specialized severity classification models, and an aggregation model to deliver accurate anomaly assessment, enabling strategic, data-driven road maintenance and enhanced transportation safety.
- A scalable and efficient solution is proposed to enhance road safety and optimize transportation management through intelligent anomaly detection and severity assessment.
- This framework sets a benchmark for future smart city initiatives by leveraging advanced machine learning techniques for proactive infrastructure maintenance and decision-making.
Abstract
1. Introduction
- A high-precision obstacle detection model capable of identifying multiple road obstacle categories. This model leverages advanced deep learning techniques and computer vision to ensure accurate detection.
- Six category-specific severity classification models, designed using a hybrid approach that combines the feature extraction capabilities of convolutional networks with the contextual understanding of Transformer-based architectures.
- A novel preprocessing technique that dynamically adjusts images based on altitude variations, improving detection accuracy in aerial transport management.
- A real-time interactive dashboard that visualizes the detected obstacles, facilitating automated risk assessment and data-driven intervention planning.
2. Related Work
2.1. Literature Survey
2.1.1. Vehicle and Animal Detection
2.1.2. Road Damage and Debris Detection
2.1.3. Garbage and Illegal Dumping Detection
2.1.4. Accident and Construction Detection
2.2. Technology Survey
Ref. | Objective | Models Used | Data Source | Area | Measurements | Evaluation Metrics |
---|---|---|---|---|---|---|
[1] | Cattle detection in desert farms | CNN (SSD-500, YOLOv3) | Custom | Desert Areas | Group-of-Animals Density Distribution, | F-score: 0.93, Accuracy: 0.89, map: 84.7 |
[2] | Deer detection and counting | YOLOv3, YOLOv4, YOLOv4-Tiny, SSD | Custom images—Manually Annotated | Dense Forests, Agricultural Areas | Number of Deer—Detection | Map: 70.45%, Precision: 86%, Recall: 75% |
[3] | Real-time UAV tracking for multi-target detection and tracking | YOLOv4, DeepSORT | Custom dataset of 3200 images | Various environments with UAVs | Tracking speed (69 FPS) | Tracking accuracy: 94.35% |
[4] | Small-scale moving target detection of aerial imagery | CNNs | Custom dataset of 10*10 pixel targets | Aerial environments | 10*10 Target size in all images | Accuracy |
[5] | Real-time object detection | YOLO, Fast YOLO | Custom datasets across various domains | Natural images, artwork | Detection speed—45 FPS for YOLO, 155 FPS for Fast YOLO | mAP, Localization error rate, False positive rate |
[6] | Moving Object Detection through UAV | YOLO, DCNN, Micro NN | KITTI Dataset | Traffic environments | Objects—Speed and Motion—3 FPS for Micro NN | Reducing error rates from simulations |
[7] | Livestock detection | Yolo v3, Faster RCNN, U-net | Custom dataset—89 Images | Aerial imagery captured by a quadcopter | Livestock—varying shapes, sizes, scales, and orientations | mAP |
[8] | Wild animal detection | DCNN, SVM, K-NN, Ensemble Tree | Camera dataset | Forests | Animal detection in cluttered environments, multilevel graph cut proposals | Accuracy: 91.4% |
[9] | Pavement Crack Detection | Faster RCNN, Preprocessing Algorithms | Crack Forest Dataset | Forests | Crack detection on Forest Roads | mAP |
[10] | Pavement crack detection | VGG-19, U-net DCNN (Residual Blocks) | Crack500 | Roads | Crack classification, crack segmentation | Classification Accuracy: 100%, Segmentation Accuracy: Improved by 10% |
[11] | Pavement distress detection using computer vision and deep learning | YOLO | Uav-pdd2023 | Highways, Provincial Roads, County Roads | Longitudinal cracks, transverse cracks, oblique cracks, alligator cracks, patching, and potholes | N/A |
[12] | UAV-based crack detection framework | Faster-RCNN, YOLOv5s, YOLOv7-tiny, YOLOv8s | DJI mini 2 UAV imagery | Urban Roads | Crack detection, road damage assessment | Faster-RCNN: highest accuracy, yolo models: fastest algorithms |
[13] | Domestic garbage detection using deep learning in complex multi-scenes | Skip-YOLO, YOLOv3 | Custom | Streets | Feature mapping, Convolution kernel, Dense convolutional blocks | Precision: 22.5%, recall: 18.6% |
[14] | Litter detection | CNN, Yolo V8 | Camera imagery | City Streets | Tree branches, Leaves, Plastic | F-score: 0.94, Accuracy: 0.90, map: 86.5 |
[15] | Waste dump detection | CNN, SSD | UAV imagery | Saint louis, Senegal | Varying shapes and sizes of waste dumping | F-score: 0.88, Accuracy: 0.85, map: 81.3 |
[16] | Illegal dumping detection | Adversarial Autoencoder, Vanilla GAN, WGAN, WGAN-GP | Satellite imagery | Various locations | Dumping site size, location | F-score: 0.90, Accuracy: 0.87, map: 82.4 |
[17] | Intelligent garbage detection using UAV and deep learning | CNN1, CNN2 | UAV images | Remote locations | Symmetry and homogeneity for image resizing | Accuracy: 94% |
[18] | Solid waste detection using machine learning and GIS data | K-means segmentation, Random Forest, Efficient Net | VHR images, GIS data | Streets | Classification of waste into 5 categories, Image segmentation, Rooftop removal for accuracy enhancement | Accuracy: 73.95–95.76% for class—Sure, Overall accuracy—80.18% |
[19] | Traffic and illegal dumping | Yolov5, Deep Sort | UAV video, coco dataset | San Jose | Detection of Illegal Dumping, Vehicle, and License Plate Identification | Accuracy: 97%, F-score: 0.95, Precision: 0.92, Recall: 0.91 |
[20] | Multi target detection | YOLOv4, CNN | COCO2017 | Objects | Detection of multiple targets—Humans, Cats, Stationery | mAP |
[21] | Graffiti detection | SSD MobileNet V2 | UAV imagery | Urban areas | Graffiti on walls and traffic signs detection | RPN loss |
[22] | Debris Object Detection Caused by Vehicle Accidents | Pretrained SSD, Faster RCNN | UAV Imagery | Traffic environments | Number of debris caused by Vehicle accidents | Accuracy Graphs, Evaluation matrix, Detection box score |
[23] | Improve construction safety using UAS and deep learning | Faster R-CNN, YOLOv3 | UAV imagery | Construction sites | Safety activity metrics, Hardhat detection | Precision: 93.1% for Faster R-CNN, 89.8% for YOLOv3 |
[24] | Public parking spaces monitoring | Object detection models | CCTV imagery | Parking areas | Number of vehicles | Accuracy, scalability |
[25] | Object detection in low-light conditions | Multi-enhancement networks | Camera imagery | Urban and Rural environments | Detection performance, response time | Accuracy |
[26] | Image classification using Transformers | Vision Transformers | CIFAR-10, CIFAR100, Pets, Flowers, Cars | Natural images | Comparison of VIT performance with typical CNNs | Accuracy |
[27] | Multiple Image classification | CNNs, Vision Transformers | Custom datasets | Noisy images | Comparison of CNNs and ViT on a large-scale dataset | Accuracy, mAP |
[28] | Attention Mechanisms in ViT | Vision Transformers | CIFAR 10/100, MINST, MINST-F | Natural and artificial images | Applied graph structures to the attention head of the Transformer | Accuracy |
[29] | AI Cloud Platform for Road inspection and analysis | YOLO, ViT, MB1-SSD | Custom UAV Imagery dataset | Urban and Rural environments | Density analysis, size-based severity classification, obstacle categorization | Precision, Accuracy, mAP, ROC, Confusion Matrix |
Proposed Work | UAV-based transport management for animal, construction, potholes, cracks, illegal dumping, and accidents | 2-staged Hybrid models combining Traditional CNNs with Transformer architectures | Custom UAV Imagery dataset | Roads, Streets and Highways | Density analysis, relative sizing, severity classification, obstacle categorization | High precision, severity levels: low/medium/high, scalable efficiency |
3. Data Engineering
3.1. Data Collection
3.2. Data Preprocessing
3.3. Data Transformation
- Rotation: This simulates different orientations of the data.
- Saturation: This accounts for variations in lighting conditions.
- Flipping: This shifts images vertically and horizontally to mimic perspective changes.
- Noise Injection: This adds random noise to help the model handle real-world noise.
3.4. Data Preparation
3.5. Data Statistics
4. Model Development
4.1. Stage 1—Obstacle Detection
4.2. Stage 2—Category-Specific Severity Classification
4.2.1. Accident Classification Model
4.2.2. Illegal Dumping Classification Model
4.2.3. Road Crack Classification Model
4.2.4. Road Construction Site Classification Model
4.2.5. Road Pothole Classification Model
4.2.6. Animal Obstruction Classification Model
4.3. Aggregation Module
5. Model Evaluation Methods and Results
5.1. Model-Oriented Evaluation Metrics
5.1.1. Confusion Matrix
- True positives (TPs): These are correctly predicted positive instances.
- True negatives (TNs): These are correctly predicted negative instances.
- False positives (FPs): These are incorrectly predicted positive instances.
- False negatives (FNs): These are incorrectly predicted negative instances.
5.1.2. ROC Curve
5.1.3. Accuracy
5.1.4. Precision
5.1.5. Recall
5.1.6. F1 Score
5.1.7. Mean Average Precision (mAP)
5.1.8. Loss
5.2. Stage 1–Obstacle Detection
5.3. Stage 2—Category-Specific Severity Classification
5.3.1. Accident Classification Model
5.3.2. Illegal Dumping Classification Model
5.3.3. Road Crack Classification Model
5.3.4. Road Construction Site Classification Model
5.3.5. Road Pothole Classification Model
5.3.6. Animal Obstruction Classification Model
5.4. City Inspection Analysis
5.4.1. Overview Dashboard
- Red Zone: This indicates high severity.
- Orange Zone: This indicates medium severity.
- Green Zone: This indicates low severity.
5.4.2. Category-Specific Dashboards
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
References
- Aburasain, R.Y.; Edirisinghe, E.A.; Albatay, A. Drone-Based Cattle Detection Using Deep Neural Networks. In Intelligent Systems and Applications; Advances in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2021; Volume 1250, pp. 598–611. [Google Scholar]
- Rančić, K.; Blagojević, B.; Bezdan, A. Animal Detection and Counting from UAV Images Using Convolutional Neural Networks. Drones 2023, 7, 179. [Google Scholar] [CrossRef]
- Hong, T.; Liang, H.; Yang, Q.; Fang, L.; Kadoch, M.; Cheriet, M. A Real-Time Tracking Algorithm for Multi-Target UAV Based on Deep Learning. Remote Sens. 2023, 15, 2. [Google Scholar] [CrossRef]
- Sun, W.; Yan, D.; Huang, J.; Sun, C. Small-scale moving target detection in aerial image by deep inverse reinforcement learning. Soft Comput. 2020, 24, 5897–5908. [Google Scholar] [CrossRef]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- He, Y.; Li, X.; Nie, H. A Moving Object Detection and Predictive Control Algorithm Based on Deep Learning. J. Phys. Conf. Ser. 2021, 2002, 012070. [Google Scholar] [CrossRef]
- Han, L.; Tao, P.; Martin, R.R. Livestock detection in aerial images using a fully convolutional network. Comput. Vis. Media 2019, 5, 123–134. [Google Scholar] [CrossRef]
- Verma, G.K.; Gupta, P. Wild Animal Detection Using Deep Convolutional Neural Network. In Proceedings of the 2nd International Conference on Computer Vision & Image Processing, Roorkee, India, 9–12 September 2017. [Google Scholar] [CrossRef]
- Zhao, M.; Shi, P.; Xu, X.; Xu, X.; Liu, W.; Yang, H. Improving the Accuracy of an R-CNN-Based Crack Identification System Using Different Preprocessing Algorithms. Sensors 2022, 22, 7089. [Google Scholar] [CrossRef] [PubMed]
- Elamin, A.; El-Rabbany, A. UAV-Based Pavement Crack Detection Using Deep Convolutional Neural Networks. In Proceedings of the FIG Congress 2022: Volunteering for the Future—Geospatial Excellence for a Better Living, Warsaw, Poland, 11–15 September 2022. [Google Scholar]
- Yan, H.; Zhang, J. UAV-PDD2023: A Benchmark Dataset for Pavement Distress Detection Based on UAV Images. Data Brief 2023, 235, 109692. [Google Scholar] [CrossRef] [PubMed]
- Chen, X.; Liu, C. A Pavement Crack Detection and Evaluation Framework for a UAV Inspection System Based on Deep Learning. Appl. Sci. 2023, 14, 1157. [Google Scholar] [CrossRef]
- Lun, Z.; Pan, Y.; Jiang, H. Skip-YOLO: Domestic Garbage Detection Using Deep Learning Method in Complex Multi-scenes. Int. J. Comput. Intell. Syst. 2023, 16, 139. [Google Scholar] [CrossRef]
- Ping, P.; Xu, G.; Kumala, E.; Gao, J. Smart Street Litter Detection and Classification Based on Faster R-CNN and Edge Computing. Int. J. Softw. Eng. Knowl. Eng. 2020, 30, 537–553. [Google Scholar] [CrossRef]
- Youme, O.; Bayet, T.; Dembele, J.M.; Cambier, C. Deep Learning and Remote Sensing: Detection of Dumping Waste Using UAV. Procedia Comput. Sci. 2021, 185, 361–369. [Google Scholar] [CrossRef]
- DumpingMapper: Illegal Dumping Detection from High Spatial Resolution Satellite Imagery. Available online: https://www.researchgate.net/publication/360435700_Accurate_Detection_of_Illegal_Dumping_Sites_Using_High_Resolution_Aerial_Photography_and_Deep_Learning (accessed on 5 August 2025).
- Verma, V.; Gupta, D.; Gupta, S. A Deep Learning-Based Intelligent Garbage Detection System Using an Unmanned Aerial Vehicle. Symmetry 2023, 14, 960. [Google Scholar] [CrossRef]
- Ulloa-Torrealba, Y.Z.; Schmitt, A.; Wurm, M.; Taubenböck, H. Litter on the Streets—Solid Waste Detection Using VHR Images. Eur. J. Remote Sens. 2023, 56, 2176006. [Google Scholar] [CrossRef]
- Pathak, N.; Biswal, G.; Goushal, M.; Mistry, V.; Shah, P.; Li, F.; Gao, J. Smart City Community Watch Camera-Based Community Watch for Traffic and Illegal Dumping. Smart Cities 2024, 7, 2232–2257. [Google Scholar] [CrossRef]
- Liu, J. Multi-target detection method based on YOLOv4 convolutional neural network. J. Phys. Conf. Ser. 2021, 1883, 012075. [Google Scholar] [CrossRef]
- Wang, S.; Gao, J.; Li, W.; Lu, S. Building Smart City Drone for Graffiti Detection and Clean-up. In Proceedings of the IEEE International Conference on Smart City Innovation, Leicester, UK, 19–23 August 2019. [Google Scholar] [CrossRef]
- Alam, H.; Valles, D.; Rabbany, A. Debris Object Detection Caused by Vehicle Accidents Using UAV and Deep Learning Techniques. In Proceedings of the IEEE Conference on Intelligent Systems, Vancouver, BC, Canada, 27–30 October 2020; pp. 3110–3115. [Google Scholar]
- Akinsemoyin, A.; Awolusi, I.; Chakraborty, D.; Al-Bayati, A.J.; Akanmu, A. Unmanned Aerial Systems and Deep Learning for Safety and Health Activity Monitoring on Construction Sites. Sensors 2023, 23, 6690. [Google Scholar] [CrossRef] [PubMed]
- Moroni, D.; Pieri, G.; Leone, G.R.; Tampucci, M. Smart cities monitoring through wireless smart cameras. In Proceedings of the 2nd International Conference, New York, NY, USA, 7 January 2019. [Google Scholar] [CrossRef]
- Awwad, Y.A.; Rana, O.; Perera, C. Anomaly Detection on the Edge Using Smart Cameras under Low-Light Conditions. Sensors 2024, 24, 772. [Google Scholar] [CrossRef] [PubMed]
- Wang, Y.; Deng, Y.; Zheng, Y.; Chattopadhyay, P.; Wang, L. Vision Transformers for Image Classification: A Comparative Survey. Technologies 2025, 13, 32. [Google Scholar] [CrossRef]
- Maurício, J.; Domingues, I.; Bernardino, J. Comparing Vision Transformers and Convolutional Neural Networks for Image Classification: A Literature Review. Appl. Sci. 2023, 13, 5521. [Google Scholar] [CrossRef]
- Kim, H.; Ko, B.C. Rethinking Attention Mechanisms in Vision Transformers with Graph Structures. Sensors 2024, 24, 1111. [Google Scholar] [CrossRef] [PubMed]
- Balivada, S.; Sha, Y.; Gao, J. Smart City Ai Cloud Platform for Road Inspection and Analysis. In Proceedings of the 11th IEEE International Conference on Big Data Computing Service and Machine Learning Applications (BigDataService), Tucson, AZ, USA, 21–24 July 2025; Available online: https://www.researchgate.net/publication/392312110_SMART_CITY_AI_CLOUD_PLATFORM_for_ROAD_INSPECTION_AND_ANALYSIS (accessed on 1 June 2025).
Dataset Source and Type | Category | Class | Quantity | Annotated | Severity Criteria | |
---|---|---|---|---|---|---|
Partial | Total | |||||
Drone Videos | Accidents | Severe Accident | 5 Videos | 9 Videos | Yes | ≥2 vehicles involved and road blockage |
Minor Accident | 4 Videos | Single vehicle, no major blockage | ||||
Public Image Dataset | Road Cracks | Longitudinal Crack | 1209 | 5047 | Yes | Area occupied by the crack |
Transverse Crack | 1187 | |||||
Oblique Crack | 1043 | |||||
Alligator Crack | 1013 | |||||
No Crack | 595 | |||||
Public Image Dataset | Road Potholes | Small Potholes | 1063 | 3458 | Yes | Depth < 2 cm, diameter < 30 cm |
Medium Potholes | 947 | Depth 2–5 cm | ||||
Large Potholes | 1098 | Depth > 5 cm or spanning lane | ||||
No Potholes | 350 | |||||
Custom Image Dataset | Construction Activities | Road Construction | 756 | 2224 | Yes | Area occupied by the construction activity |
Highway Construction | 802 | |||||
Bridge Construction | 321 | |||||
Roadside Building Construction | 345 | |||||
Mixed Image Dataset | Illegal Dumping | Low Level | 928 | 3064 | Yes | Small scattered debris, <0.5 m2 |
Medium Level | 1127 | Scattered debris, may slightly obstruct lane | ||||
High Level | 1009 | Large heap, obstructing lane | ||||
Custom Image Dataset | Wild Animals on Roads | Dogs | 192 | 2040 | Yes | Count and size of animals |
Cattle | 519 | |||||
Other Animals | 1329 |
UAV Data Category | Raw | Transformed | Prepared | ||
---|---|---|---|---|---|
Train | Validation | Test | |||
Accidents on road | 1023 | 10,057 | 7390 | 2111 | 1056 |
Illegal dumping on roads | 3064 | 27,576 | 19,303 | 5515 | 1758 |
Road cracks | 5047 | 45,423 | 31,796 | 9085 | 4542 |
Road construction activities | 2224 | 20,016 | 14,011 | 4003 | 2002 |
Road potholes | 3458 | 31,122 | 21,785 | 6224 | 3113 |
Animals on road | 2040 | 18,360 | 12,852 | 3672 | 1836 |
Category | mAP Score |
---|---|
Accident detection | 0.992 |
Illegal dumping detection | 0.362 |
Road crack detection | 0.645 |
Road construction site detection | 0.780 |
Road pothole detection | 0.693 |
Animal obstruction detection | 0.901 |
Model Name | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
Accident Classification Model | 90% | 95% | 88% | 91% |
Illegal Dumping Classification Model | 85% | 81% | 80% | 80% |
Road Crack Classification Model | 86% | 88% | 85% | 86% |
Road Construction Site Classification Model | 81% | 80% | 78% | 79% |
Road Pothole Classification Model | 87% | 85% | 83% | 84% |
Animal Obstruction Classification Model | 77% | 78% | 75% | 76% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Balivada, S.; Gao, J.; Sha, Y.; Lagisetty, M.; Vichare, D. UAV-Based Transport Management for Smart Cities Using Machine Learning. Smart Cities 2025, 8, 154. https://doi.org/10.3390/smartcities8050154
Balivada S, Gao J, Sha Y, Lagisetty M, Vichare D. UAV-Based Transport Management for Smart Cities Using Machine Learning. Smart Cities. 2025; 8(5):154. https://doi.org/10.3390/smartcities8050154
Chicago/Turabian StyleBalivada, Sweekruthi, Jerry Gao, Yuting Sha, Manisha Lagisetty, and Damini Vichare. 2025. "UAV-Based Transport Management for Smart Cities Using Machine Learning" Smart Cities 8, no. 5: 154. https://doi.org/10.3390/smartcities8050154
APA StyleBalivada, S., Gao, J., Sha, Y., Lagisetty, M., & Vichare, D. (2025). UAV-Based Transport Management for Smart Cities Using Machine Learning. Smart Cities, 8(5), 154. https://doi.org/10.3390/smartcities8050154