Early Wildfire Smoke Detection with a Multi-Resolution Framework and Two-Stage Classification Pipeline
Abstract
1. Introduction
- A composite multi-resolution representation that embeds high-resolution detail within a standard detector input using a single forward pass.
- A skyline-guided dynamic cropping strategy with reversible coordinate mapping for accurate localization in the original image space.
- A deployment-time confidence refinement mechanism to stabilize low-confidence early-smoke predictions.
- A comprehensive evaluation on early-smoke benchmarks (Early Smoke and Pyro-SDIS) demonstrating consistent gains in small-object and low-confidence regimes.
2. Related Work
2.1. Wildfire Monitoring Context
2.2. Deep Learning for Wildfire Monitoring
2.3. Small Object Detection and Resolution Manipulation
Our Approach
3. Methods
3.1. Overall Architecture
3.2. Composite Image Generation
- denotes the crop height,
- the horizontal and vertical crop offsets,
- and the scaling factors used for generating the intermediate-resolution and -width images,
- the dimensions of the intermediate resized image,
- and the vertical offsets introduced by cropping when mapping detections back to the original coordinate space.
| Algorithm 1 Composite Image Generation Procedure. |
Procedure GenerateComposite(I, , , ): Input: I (RGB image), , (detector input size), (intermediate resize width). Output: (composite for detector inference), M (metadata for coordinate restoration). 1. Global view (base canvas for detector input): (preserve aspect ratio), . 2. Intermediate view (used to select a smoke-prior ROI): , . 3. Define the ROI size to occupy the unused letterbox region: , . 4. Compute ROI location (skyline-guided, with fallback): . if is a default prior). . . Define . 5. Crop the intermediate view: , . 6. Construct the detector input (single tensor): , . 7. Return composite and restoration metadata: . return . |
3.3. Dynamic Cropping
3.4. Fusion: Remapping and Non-Maximum Suppression
3.5. Two-Stage Classifier
4. Experimental Setup
4.1. Datasets
4.1.1. Dataset Composition
4.1.2. Datasets Used
4.1.3. Dataset Construction
4.1.4. Dataset Description and Statistics
- 1.
- Early Smoke Dataset: 936 carefully filtered early-smoke images drawn from FASDD–UAV and D-Fire, emphasizing small, faint, low-contrast plumes with <2% image area under varying illumination and high background complexity.
- 2.
- Pyro-SDIS Dataset: 3523 deduplicated images from tower-mounted camera systems in predominantly forested scenes, containing far-distant, low-contrast smoke under diverse lighting and weather conditions, with manually curated YOLO annotations by Pyronear volunteers.
4.2. Training for Early Smoke Detection
4.2.1. Detector Training
4.2.2. Training for Early Smoke Classification
4.3. Evaluation Metrics
5. Results and Discussion
5.1. Overall Detection Performance
5.2. Deployment-Only Evaluation (Two-Stage Classifier)
Confidence Threshold Selection
5.3. Ablation Study
5.4. Generalizability Verification
Failure Analysis
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Arteaga, B.; Diaz, M.; Jojoa, M. Deep learning applied to forest fire detection. In Proceedings of the 2020 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT); IEEE: New York, NY, USA, 2020; pp. 1–6. [Google Scholar]
- Sun, J.; Qi, W.; Huang, Y.; Xu, C.; Yang, W. Facing the Wildfire Spread Risk Challenge: Where Are We Now and Where Are We Going? Fire 2023, 6, 228. [Google Scholar] [CrossRef]
- Vasconcelos, R.N.; Franca Rocha, W.J.; Costa, D.P.; Duverger, S.G.; Santana, M.M.d.; Cambui, E.C.; Ferreira-Ferreira, J.; Oliveira, M.; Barbosa, L.d.S.; Cordeiro, C.L. Fire detection with deep learning: A comprehensive review. Land 2024, 13, 1696. [Google Scholar] [CrossRef]
- Zhao, Y.; Ma, J.; Li, X.; Zhang, J. Saliency detection and deep learning-based wildfire identification in UAV imagery. Sensors 2018, 18, 712. [Google Scholar] [CrossRef] [PubMed]
- Wu, H.; Li, H.; Shamsoshoara, A.; Razi, A.; Afghah, F. Transfer learning for wildfire identification in UAV imagery. In Proceedings of the 2020 54th Annual Conference on Information Sciences and Systems (CISS); IEEE: New York, NY, USA, 2020; pp. 1–6. [Google Scholar]
- Zhao, L.; Zhi, L.; Zhao, C.; Zheng, W. Fire-YOLO: A small target object detection method for fire inspection. Sustainability 2022, 14, 4930. [Google Scholar] [CrossRef]
- Ahn, Y.; Choi, H.; Kim, B.S. Development of early fire detection model for buildings using computer vision-based CCTV. J. Build. Eng. 2023, 65, 105647. [Google Scholar] [CrossRef]
- Huang, P.; Chen, M.; Chen, K.; Zhang, H.; Yu, L.; Liu, C. A combined real-time intelligent fire detection and forecasting approach through cameras based on computer vision method. Process Saf. Environ. Prot. 2022, 164, 629–638. [Google Scholar] [CrossRef]
- Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision; IEEE: New York, NY, USA, 2015; pp. 1440–1448. [Google Scholar]
- Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
- Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar]
- Lv, C.; Zhou, H.; Chen, Y.; Fan, D.; Di, F. A lightweight fire detection algorithm for small targets based on YOLOv5s. Sci. Rep. 2024, 14, 14104. [Google Scholar] [CrossRef]
- Xiao, Z.; Wan, F.; Lei, G.; Xiong, Y.; Xu, L.; Ye, Z.; Liu, W.; Zhou, W.; Xu, C. Fl-yolov7: A lightweight small object detection algorithm in forest fire detection. Forests 2023, 14, 1812. [Google Scholar] [CrossRef]
- Zhang, L.; Wang, M.; Ding, Y.; Bu, X. MS-FRCNN: A multi-scale faster RCNN model for small target forest fire detection. Forests 2023, 14, 616. [Google Scholar] [CrossRef]
- Wang, Y.; Bashir, S.M.A.; Khan, M.; Ullah, Q.; Wang, R.; Song, Y.; Guo, Z.; Niu, Y. Remote sensing image super-resolution and object detection: Benchmark and state of the art. Expert Syst. Appl. 2022, 197, 116793. [Google Scholar] [CrossRef]
- Boroujeni, S.P.H.; Razi, A.; Khoshdel, S.; Afghah, F.; Coen, J.L.; O’Neill, L.; Fule, P.; Watts, A.; Kokolakis, N.M.T.; Vamvoudakis, K.G. A comprehensive survey of research towards AI-enabled unmanned aerial systems in pre-, active-, and post-wildfire management. Inf. Fusion 2024, 108, 102369. [Google Scholar] [CrossRef]
- Andrianarivony, H.S.; Akhloufi, M.A. Machine learning and deep learning for wildfire spread prediction: A review. Fire 2024, 7, 482. [Google Scholar] [CrossRef]
- Peng, Y.; Wang, Y. Automatic wildfire monitoring system based on deep learning. Eur. J. Remote Sens. 2022, 55, 551–567. [Google Scholar] [CrossRef]
- Afghah, F. Autonomous Unmanned Aerial Vehicle Systems in Wildfire Detection and Management-Challenges and Opportunities. In Proceedings of the Dynamic Data Driven Application Systems; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
- Lelis, C.A.S.; Roncal, J.J.; Silveira, L.; De Aquino, R.D.G.; Marcondes, C.A.C.; Marques, J.; Loubach, D.S.; Verri, F.A.N.; Curtis, V.V.; De Souza, D.G. Drone-Based AI System for Wildfire Monitoring and Risk Prediction. IEEE Access 2024, 12, 139865–139882. [Google Scholar] [CrossRef]
- Bailon-Ruiz, R.; Bit-Monnot, A.; Lacroix, S. Real-time wildfire monitoring with a fleet of UAVs. Robot. Auton. Syst. 2022, 152, 104071. [Google Scholar] [CrossRef]
- Phillips, W., III; Shah, M.; da Vitoria Lobo, N. Flame recognition in video. Pattern Recognit. Lett. 2002, 23, 319–327. [Google Scholar] [CrossRef]
- Gubbi, J.; Marusic, S.; Palaniswami, M. Smoke detection in video using wavelets and support vector machines. Fire Saf. J. 2009, 44, 1110–1115. [Google Scholar] [CrossRef]
- Mambile, C.; Kaijage, S.; Leo, J. Application of Deep Learning in Forest Fire Prediction: A Systematic Review. IEEE Access 2024, 12, 190554–190581. [Google Scholar] [CrossRef]
- Ghali, R.; Akhloufi, M.A. Deep learning approaches for wildland fires remote sensing: Classification, detection, and segmentation. Remote Sens. 2023, 15, 1821. [Google Scholar] [CrossRef]
- Akagic, A.; Buza, E. Lw-fire: A lightweight wildfire image classification with a deep convolutional neural network. Appl. Sci. 2022, 12, 2646. [Google Scholar] [CrossRef]
- Seydi, S.T.; Saeidi, V.; Kalantar, B.; Ueda, N.; Halin, A.A. Fire-Net: A Deep Learning Framework for Active Forest Fire Detection. J. Sens. 2022, 2022, 8044390. [Google Scholar] [CrossRef]
- Ramos, L.; Casas, E.; Bendek, E.; Romero, C.; Rivas-Echeverría, F. Computer vision for wildfire detection: A critical brief review. Multimed. Tools Appl. 2024, 83, 83427–83470. [Google Scholar] [CrossRef]
- Li, J.; Tang, H.; Li, X.; Dou, H.; Li, R. LEF-YOLO: A lightweight method for intelligent detection of four extreme wildfires based on the YOLO framework. Int. J. Wildland Fire 2023, 33, WF23044. [Google Scholar] [CrossRef]
- Bhargav, R.; Singh, P. Efficient UAV-Based Forest Fire Detection Using CNN and YOLOv8 Integration. In Proceedings of the 2025 6th International Conference on Recent Advances in Information Technology (RAIT); IEEE: New York, NY, USA, 2025; pp. 1–6. [Google Scholar]
- Jonnalagadda, A.V.; Hashim, H.A. SegNet: A segmented deep learning based Convolutional Neural Network approach for drones wildfire detection. Remote Sens. Appl. Soc. Environ. 2024, 34, 101181. [Google Scholar] [CrossRef]
- Khryashchev, V.V.; Larionov, R. Wildfire Segmentation on Satellite Images using Deep Learning. In Proceedings of the 2020 Moscow Workshop on Electronic and Networking Technologies (MWENT); IEEE: New York, NY, USA, 2020; pp. 1–5. [Google Scholar]
- Bouguettaya, A.; Zarzour, H.; Taberkit, A.M.; Kechida, A. A review on early wildfire detection from unmanned aerial vehicles using deep learning-based computer vision algorithms. Signal Process. 2022, 190, 108309. [Google Scholar] [CrossRef]
- Ghali, R.; Akhloufi, M.A.; Mseddi, W.S. Deep learning and transformer approaches for UAV-based wildfire detection and segmentation. Sensors 2022, 22, 1977. [Google Scholar] [CrossRef]
- Cheng, G.; Yuan, X.; Yao, X.; Yan, K.; Zeng, Q.; Xie, X.; Han, J. Towards large-scale small object detection: Survey and benchmarks. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 13467–13488. [Google Scholar] [CrossRef]
- Wang, J.; Yang, W.; Guo, H.; Zhang, R.; Xia, G.S. Tiny object detection in aerial images. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR); IEEE: New York, NY, USA, 2021; pp. 3791–3798. [Google Scholar]
- Jiang, L.; Yuan, B.; Du, J.; Chen, B.; Xie, H.; Tian, J.; Yuan, Z. MFFSODNet: Multiscale feature fusion small object detection network for UAV aerial images. IEEE Trans. Instrum. Meas. 2024, 73, 3381272. [Google Scholar] [CrossRef]
- Hu, M.; Li, Z.; Yu, J.; Wan, X.; Tan, H.; Lin, Z. Efficient-lightweight YOLO: Improving small object detection in YOLO for aerial images. Sensors 2023, 23, 6423. [Google Scholar] [CrossRef]
- Wang, G.; Chen, Y.; An, P.; Hong, H.; Hu, J.; Huang, T. UAV-YOLOv8: A small-object-detection model based on improved YOLOv8 for UAV aerial photography scenarios. Sensors 2023, 23, 7190. [Google Scholar] [CrossRef] [PubMed]
- Ozge Unel, F.; Ozkalayci, B.O.; Cigla, C. The Power of Tiling for Small Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops; IEEE: New York, NY, USA, 2019. [Google Scholar]
- Akyon, F.C.; Altinuc, S.O.; Temizel, A. Slicing aided hyper inference and fine-tuning for small object detection. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP); IEEE: New York, NY, USA, 2022; pp. 966–970. [Google Scholar]
- Zhang, H.; Hao, C.; Song, W.; Jiang, B.; Li, B. Adaptive slicing-aided hyper inference for small object detection in high-resolution remote sensing images. Remote Sens. 2023, 15, 1249. [Google Scholar] [CrossRef]
- Hao, C.; Zhang, H.; Song, W.; Liu, F.; Wu, E. Slinet: Slicing-aided learning for small object detection. IEEE Signal Process. Lett. 2024, 31, 790–794. [Google Scholar] [CrossRef]
- Telçeken, M.; Akgun, D.; Kacar, S. An evaluation of image slicing and YOLO architectures for object detection in UAV images. Appl. Sci. 2024, 14, 11293. [Google Scholar] [CrossRef]
- Muzammul, M.; Li, X.; Li, X. Enhancing Tiny Object Detection Using Guided Object Inference Slicing (GOIS): An efficient dynamic adaptive framework for fine-tuned and non-fine-tuned deep learning models. Neurocomputing 2025, 640, 130327. [Google Scholar] [CrossRef]
- Chen, Z.; Chen, G. STTSBI: A Fast Inference Framework for Small Object Detection in Ultra-High-Resolution Images. In Proceedings of the 2024 4th International Conference on Intelligent Technology and Embedded Systems (ICITES); IEEE: New York, NY, USA, 2024; pp. 129–135. [Google Scholar]
- Koyun, O.C.; Keser, R.K.; Akkaya, I.B.; Töreyin, B.U. Focus-and-Detect: A small object detection framework for aerial images. Signal Process. Image Commun. 2022, 104, 116675. [Google Scholar] [CrossRef]
- Wang, M.; Yue, P.; Jiang, L.; Yu, D.; Tuo, T.; Li, J. An open flame and smoke detection dataset for deep learning in remote sensing based fire detection. Geo-Spat. Inf. Sci. 2025, 28, 511–526. [Google Scholar] [CrossRef]
- De Venâncio, P.V.A.; Rezende, T.M.; Lisboa, A.C.; Barbosa, A.V. Fire detection based on a two-dimensional convolutional neural network and temporal analysis. In Proceedings of the 2021 IEEE Latin American Conference on Computational Intelligence (LA-CCI); IEEE: New York, NY, USA, 2021; pp. 1–6. [Google Scholar]
- Pyronear Team. Pyro-SDIS Dataset. Hugging Face. 2024. Available online: https://huggingface.co/datasets/pyronear/pyro-sdis (accessed on 9 January 2026).
- Sapkota, R.; Karkee, M. Ultralytics YOLO evolution: An overview of YOLO26, YOLO11, YOLOv8 and YOLOv5 object detectors for computer vision and pattern recognition. arXiv 2025, arXiv:2510.09653. [Google Scholar]













| Component | Specification | Manufacturer (City, Country) |
|---|---|---|
| GPU | NVIDIA RTX A5000 (24 GB) | NVIDIA Corporation (Santa Clara, CA, USA) |
| CPU & Memory | Intel Core i5-14400F; 64 GB RAM | Intel Corporation (Santa Clara, CA, USA) |
| OS | Ubuntu 24.04.3 LTS | Canonical Ltd. (London, UK) |
| CUDA version | 12.6 | NVIDIA Corporation (Santa Clara, CA, USA) |
| Python version | Python 3.10.18 | Python Software Foundation (Beaverton, OR, USA) |
| PyTorch version | 2.7.1 + cu126 | Meta Platforms, Inc. (Menlo Park, CA, USA) |
| Model | AP @0.5:0.95 | AP @0.5 | F1 | Precision | Recall | |
|---|---|---|---|---|---|---|
| GT-based trained | 0.420 | 0.852 | 0.212 | 0.832 | 0.881 | 0.787 |
| Skyline-trained | 0.397 | 0.845 | 0.113 | 0.812 | 0.832 | 0.793 |
| Model | AP @0.50:0.95 | AP @0.50 | F1 | Precision | Recall | |
|---|---|---|---|---|---|---|
| YOLOv12 | 0.400 | 0.843 | 0.127 | 0.838 | 0.902 | 0.782 |
| YOLOv11 | 0.389 | 0.795 | 0.222 | 0.821 | 0.839 | 0.803 |
| YOLOv10 | 0.379 | 0.814 | 0.153 | 0.806 | 0.775 | 0.840 |
| YOLOv9 | 0.421 | 0.858 | 0.219 | 0.835 | 0.864 | 0.809 |
| YOLOv8 | 0.420 | 0.852 | 0.212 | 0.832 | 0.881 | 0.787 |
| YOLOv5su | 0.409 | 0.825 | 0.102 | 0.811 | 0.802 | 0.819 |
| Model | AP @0.5:0.95 | AP @0.5 | F1 | Precision | Recall | |
|---|---|---|---|---|---|---|
| Baseline | 0.374 | 0.818 | 0.054 | 0.789 | 0.852 | 0.734 |
| Ours | 0.420 | 0.852 | 0.212 | 0.832 | 0.881 | 0.787 |
| Model | AP @0.5:0.95 | AP @0.5 | F1 | Precision | Recall | |
|---|---|---|---|---|---|---|
| YOLOv5su | 0.384 | 0.839 | 0.136 | 0.823 | 0.844 | 0.803 |
| YOLOv8s | 0.374 | 0.818 | 0.054 | 0.789 | 0.852 | 0.734 |
| YOLOv9s | 0.321 | 0.776 | 0.077 | 0.776 | 0.844 | 0.718 |
| YOLOv10s (SGD) | 0.367 | 0.804 | 0.210 | 0.785 | 0.816 | 0.755 |
| YOLOv11s | 0.379 | 0.792 | 0.134 | 0.770 | 0.852 | 0.702 |
| Ours | 0.420 | 0.852 | 0.212 | 0.832 | 0.881 | 0.787 |
| Experiment | AP @ 0.50:0.95 | AP @ 0.50 | F1 | Precision | Recall | |
|---|---|---|---|---|---|---|
| Baseline | 0.345 | 0.760 | 0.037 | 0.789 | 0.852 | 0.734 |
| Multiresolution | 0.404 | 0.813 | 0.188 | 0.832 | 0.881 | 0.787 |
| Multiresolution + Classifier | 0.410 | 0.820 | 0.188 | 0.832 | 0.881 | 0.787 |
| Multiresolution | Classifier | AP @ 0.50:0.95 | AP @ 0.50 | F1 | Precision | Recall | |
|---|---|---|---|---|---|---|---|
| X | X | 0.345 | 0.760 | 0.037 | 0.789 | 0.852 | 0.734 |
| X | O | 0.343 | 0.768 | 0.037 | 0.806 | 0.804 | 0.809 |
| O | X | 0.404 | 0.813 | 0.188 | 0.832 | 0.881 | 0.787 |
| O | O | 0.410 | 0.820 | 0.188 | 0.832 | 0.881 | 0.787 |
| Multiresolution | Classifier | Inference Time (ms) | FPS |
|---|---|---|---|
| X | X | 3.168 | 315.66 |
| X | O | 7.515 | 133.07 |
| O | X | 5.825 | 171.67 |
| O | O | 11.708 | 85.41 |
| Model | AP @0.5:0.95 | AP @0.5 | AP @0.75 | F1 | Precision | Recall | |
|---|---|---|---|---|---|---|---|
| YOLOv5su | 0.419 | 0.714 | 0.445 | 0.400 | 0.700 | 0.683 | 0.719 |
| YOLOv8s | 0.418 | 0.737 | 0.423 | 0.405 | 0.729 | 0.745 | 0.714 |
| YOLOv9s | 0.406 | 0.703 | 0.405 | 0.361 | 0.687 | 0.698 | 0.676 |
| YOLOv10s | 0.413 | 0.713 | 0.432 | 0.391 | 0.699 | 0.750 | 0.655 |
| YOLOv11s | 0.410 | 0.718 | 0.417 | 0.367 | 0.691 | 0.691 | 0.690 |
| Ours | 0.436 | 0.749 | 0.447 | 0.410 | 0.726 | 0.777 | 0.681 |
| Multiresolution | Classifier | AP @0.50:0.95 | AP @0.50 | AP @0.75 | F1 | Precision | Recall | |
|---|---|---|---|---|---|---|---|---|
| X | X | 0.386 | 0.664 | 0.404 | 0.365 | 0.724 | 0.758 | 0.692 |
| X | O | 0.401 | 0.667 | 0.424 | 0.375 | 0.734 | 0.708 | 0.762 |
| O | X | 0.410 | 0.691 | 0.426 | 0.386 | 0.726 | 0.777 | 0.681 |
| O | O | 0.416 | 0.703 | 0.431 | 0.394 | 0.726 | 0.777 | 0.681 |
| Multiresolution | Classifier | Average Inference Time (ms) | FPS |
|---|---|---|---|
| X | X | 3.238 | 308.83 |
| X | O | 6.741 | 148.35 |
| O | X | 5.630 | 177.62 |
| O | O | 13.818 | 72.37 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Jung, G.; Ahn, T.-H.; Min, B. Early Wildfire Smoke Detection with a Multi-Resolution Framework and Two-Stage Classification Pipeline. Fire 2026, 9, 92. https://doi.org/10.3390/fire9020092
Jung G, Ahn T-H, Min B. Early Wildfire Smoke Detection with a Multi-Resolution Framework and Two-Stage Classification Pipeline. Fire. 2026; 9(2):92. https://doi.org/10.3390/fire9020092
Chicago/Turabian StyleJung, Gihwan, Tae-Hyuk Ahn, and Byungseok Min. 2026. "Early Wildfire Smoke Detection with a Multi-Resolution Framework and Two-Stage Classification Pipeline" Fire 9, no. 2: 92. https://doi.org/10.3390/fire9020092
APA StyleJung, G., Ahn, T.-H., & Min, B. (2026). Early Wildfire Smoke Detection with a Multi-Resolution Framework and Two-Stage Classification Pipeline. Fire, 9(2), 92. https://doi.org/10.3390/fire9020092

