Multimodal Wildfire Classification Using Synthetic Night-Vision-like and Thermal-Inspired Image Representations
Abstract
1. Introduction
1.1. Background
1.2. Literature Review
1.3. Contribution and Novelty of Present Study
- ➢
- Unlike a large portion of existing works that formulate wildfire monitoring as an object detection or instance segmentation problem, the proposed approach is strictly designed as an image-level classification framework.
- ➢
- A hardware-independent multimodal dataset was constructed by generating synthetic night-vision-like, white-hot, and green-hot pseudo-thermal representations directly from RGB images.
- ➢
- A multimodal feature extraction and fusion framework was systematically evaluated using ImageNet-pretrained backbones (ResNet18, EfficientNet-B0, and DenseNet121).
- ➢
- The proposed framework was assessed in both binary (FLAME) and three-class (FireStage) classification settings, including early-stage fire scenarios.
- ➢
- The generalization ability of the approach was examined across datasets with substantially different sizes and characteristics.
- ➢
- Quantitative results revealed competitive performance across different backbone architectures and modality combinations, with detailed comparisons in terms of accuracy, precision, recall, F1-score, ROC/AUC, parameter count, and training time.
2. Materials and Methods
- (A)
- Preprocessing and synthesis of multimodal visual inputs from RGB images;
- (B)
- Modality-specific feature extraction followed by feature-level fusion and classification through a lightweight fully connected layer;
- (C)
- Quantitative performance evaluation using standard classification metrics.
2.1. Dataset and Features
2.2. Preprocessing and Augmentation
2.3. Data Synthesis
2.3.1. RGB-to-Grayscale Transformation
2.3.2. Contrast Enhancement for Night-Vision-like Modality
2.3.3. White-Hot and Green-Hot Pseudo-Thermal Modalities
2.3.4. Final Multimodal Sample Construction
2.3.5. Interpretation of Modality-Specific Intensity Distributions
2.4. Feature Extraction and Classification
2.4.1. Modality-Wise Feature Extraction
2.4.2. Multimodal Feature Fusion
2.4.3. Classification Head
2.5. Model Performance Evaluation
3. Results
4. Discussion and Limitation
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- de Almeida, R.V.; Crivellaro, F.; Narciso, M.; Sousa, A.I.; Vieira, P. Bee2Fire: A deep learning powered forest fire detection system. In Proceedings of the ICAART 2020—12th International Conference on Agents and Artificial Intelligence, Valletta, Malta, 22–24 February 2020; pp. 603–609. [Google Scholar]
- Benzekri, W.; El Moussati, A.; Moussaoui, O.; Berrajaa, M. Early forest fire detection system using wireless sensor network and deep learning. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 496–503. [Google Scholar] [CrossRef]
- Mohammad, M.B.; Bhuvaneswari, N.; Koteswari, C.P.; Priya, V.B. Hardware implementation of forest fire detection system using deep learning architectures. In Proceedings of the International Conference on Edge Computing and Applications (ICECAA) 2022, Tamilnadu, India, 13–15 October 2022; pp. 1198–1205. [Google Scholar]
- Ryu, J.; Kwak, D. A study on a complex flame and smoke detection method using computer vision detection and convolutional neural network. Fire 2022, 5, 108. [Google Scholar] [CrossRef]
- Shirwaikar, R.; Narvekar, A.; Hosamani, A.; Fernandes, K.; Tak, K.; Parab, V. Real-time semi-occluded fire detection and evacuation route generation: Leveraging instance segmentation for damage estimation. Fire Saf. J. 2025, 152, 104338. [Google Scholar] [CrossRef]
- Chaoxia, C.; Shang, W.; Zhang, F. Information-guided flame detection based on Faster R-CNN. IEEE Access 2020, 8, 58923–58932. [Google Scholar] [CrossRef]
- Casas, E.; Ramos, L.; Bendek, E.; Rivas-Echeverria, F. Assessing the effectiveness of YOLO architectures for smoke and wildfire detection. IEEE Access 2023, 11, 96554–96583. [Google Scholar] [CrossRef]
- Sathishkumar, V.E.; Cho, J.; Subramanian, M.; Naren, O.S. Forest fire and smoke detection using deep learning-based learning without forgetting. Fire Ecol. 2023, 19, 9. [Google Scholar] [CrossRef]
- Goncalves, A.M.; Brandao, T.; Ferreira, J.C. Wildfire detection with deep learning—A case study for the CICLOPE project. IEEE Access 2024, 12, 82095–82110. [Google Scholar] [CrossRef]
- Yan, C.; Wang, J. MAG-FSNet: A high-precision robust forest fire smoke detection model integrating local features and global information. Measurement 2025, 247, 116813. [Google Scholar] [CrossRef]
- Wang, W.; Huang, Q.; Liu, H.; Jia, Y.; Chen, Q. Forest fire detection method based on deep learning. In Proceedings of the International Conference on Cyber-Physical Social Intelligence (ICCSI) 2022, Nanjing, China, 18–21 November 2022; pp. 23–28. [Google Scholar]
- Zhu, W.; Niu, S.; Yue, J.; Zhou, Y. Multiscale wildfire and smoke detection in complex drone forest environments based on YOLOv8. Sci. Rep. 2025, 15, 2399. [Google Scholar] [CrossRef]
- Wang, X.; Wang, J.; Chen, L.; Zhang, Y. Improving computer vision-based wildfire smoke detection by combining SE-ResNet with SVM. Processes 2024, 12, 747. [Google Scholar] [CrossRef]
- Li, L.; Liu, F.; Ding, Y. Real-time smoke detection with Faster R-CNN. In Proceedings of the 2nd International Conference on Artificial Intelligence and Information Systems, Chongqing, China, 28–30 May 2021; pp. 1–5. [Google Scholar]
- Bahhar, C.; Ksibi, A.; Ayadi, M.; Jamjoom, M.M.; Ullah, Z.; Soufiene, B.O.; Sakli, H. Wildfire and smoke detection using staged YOLO model and ensemble CNN. Electronics 2023, 12, 228. [Google Scholar] [CrossRef]
- Li, Y.; Zhang, W.; Liu, Y.; Jing, R.; Liu, C. An efficient fire and smoke detection algorithm based on an end-to-end structured network. Eng. Appl. Artif. Intell. 2022, 116, 105492. [Google Scholar] [CrossRef]
- Wang, C.; Li, Q.; Liu, S.; Cheng, P.; Huang, Y. Transformer-based fusion of infrared and visible imagery for smoke recognition in commercial areas. Comput. Mater. Contin. 2025, 84, 5157–5176. [Google Scholar] [CrossRef]
- Wang, Y.; Wang, Y.; Khan, Z.A.; Huang, A.; Sang, J. Multi-level feature fusion networks for smoke recognition in remote sensing imagery. Neural Netw. 2025, 184, 107112. [Google Scholar] [CrossRef] [PubMed]
- Alkhammash, E.H. A comparative analysis of YOLOv9, YOLOv10, YOLOv11 for smoke and fire detection. Fire 2025, 8, 26. [Google Scholar] [CrossRef]
- Xue, Z.; Kong, L.; Wu, H.; Chen, J. Fire and smoke detection based on improved YOLOv11. IEEE Access 2025, 13, 73022–73040. [Google Scholar] [CrossRef]
- He, L.; Zhou, Y.; Liu, L.; Zhang, Y.; Ma, J. Research and application of deep learning object detection methods for forest fire smoke recognition. Sci. Rep. 2025, 15, 16328. [Google Scholar] [CrossRef]
- Niu, K.; Wang, C.; Xu, J.; Liang, J.; Zhou, X.; Wen, K.; Lu, M.; Yang, C. Early forest fire detection with UAV image fusion: A novel deep learning method using visible and infrared sensors. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 6617–6629. [Google Scholar] [CrossRef]
- Jin, P.; Cheng, P.; Liu, X.; Huang, Y. From smoke to fire: A forest fire early warning and risk assessment model fusing multimodal data. Eng. Appl. Artif. Intell. 2025, 152, 110848. [Google Scholar] [CrossRef]
- Shang, L.; Hu, X.; Huang, Z.; Zhang, Q.; Zhang, Z.; Li, X.; Chang, Y. YOLO-DKM: A flame and spark detection algorithm based on deep learning. IEEE Access 2025, 13, 117687–117699. [Google Scholar] [CrossRef]
- Arteaga, B.; Diaz, M.; Jojoa, M. Deep learning applied to forest fire detection. In Proceedings of the IEEE International Symposium on Signal Processing and Information Technology (ISSPIT) 2020, Louisville, KY, USA, 9–11 December 2020. [Google Scholar]
- Mohammed, R.K. A real-time forest fire and smoke detection system using deep learning. Int. J. Nonlinear Anal. Appl. 2022, 13, 2053–2063. [Google Scholar]
- Mohnish, S.; Akshay, K.P.; Ram, S.G.; Vignesh, A.S.; Pavithra, P.; Ezhilarasi, S. Deep learning based forest fire detection and alert system. In Proceedings of the International Conference on Communication, Computing and Internet of Things (IC3IoT) 2022, Chennai, India, 10–11 March 2022; pp. 1–5. [Google Scholar]
- Ban, Y.; Zhang, P.; Nascetti, A.; Bevington, A.R.; Wulder, M.A. Near real-time wildfire progression monitoring with Sentinel-1 SAR time series and deep learning. Sci. Rep. 2020, 10, 1322. [Google Scholar] [CrossRef]
- Rahul, M.; Saketh, K.S.; Sanjeet, A.; Naik, N.S. Early detection of forest fire using deep learning. In Proceedings of the IEEE REGION 10 CONFERENCE (TENCON) 2020, Osaka, Japan, 16–19 November 2020; pp. 1136–1140. [Google Scholar]
- Jiang, Y.; Wei, R.; Chen, J.; Wang, G. Deep learning of Qinling forest fire anomaly detection based on genetic algorithm optimization. UPB Sci. Bull. Ser.-Electr. Eng. Comput. Sci. 2021, 83, 75–84. [Google Scholar]
- Li, M.; Zhang, Y.; Mu, L.; Xin, J.; Yu, Z.; Liu, H.; Xie, G. Early forest fire detection based on deep learning. In Proceedings of the 3rd International Conference on Industrial Artificial Intelligence (IAI) 2021, Shenyang, China, 8–11 November 2021; pp. 1–5. [Google Scholar]
- Khan, S.; Khan, A. FFireNet: Deep learning based forest fire classification and detection in smart cities. Symmetry 2022, 14, 2155. [Google Scholar] [CrossRef]
- Gayathri, S.; Ajay Karthi, P.V.; Sunil, S. Prediction and detection of forest fires based on deep learning approach. J. Pharm. Negat. Results 2022, 13, 429–433. [Google Scholar] [CrossRef]
- Kang, Y.; Jang, E.; Im, J.; Kwon, C. A deep learning model using geostationary satellite data for forest fire detection with reduced detection latency. GISci. Remote Sens. 2022, 59, 2019–2035. [Google Scholar] [CrossRef]
- Ghosh, R.; Kumar, A. A hybrid deep learning model combining CNN and RNN to detect forest fires. Multimed. Tools Appl. 2022, 81, 38643–38660. [Google Scholar] [CrossRef]
- Tahir, H.U.A.; Waqar, A.; Khalid, S.; Usman, S.M. Wildfire detection in aerial images using deep learning. In Proceedings of the 2nd International Conference on Digital Futures and Transformative Technologies (ICoDT2) 2022, Rawalpindi, Pakistan, 24–26 May 2022. [Google Scholar]
- Peng, Y.; Wang, Y. Automatic wildfire monitoring system based on deep learning. Eur. J. Remote Sens. 2022, 55, 551–567. [Google Scholar] [CrossRef]
- Mashraqi, A.M.; Asiri, Y.; Algarni, A.D.; Abu-Zinadah, H. Drone imagery forest fire detection and classification using modified deep learning model. Therm. Sci. 2022, 26, 411–423. [Google Scholar] [CrossRef]
- Almasoud, A.S. Intelligent deep learning enabled wild forest fire detection system. Comput. Syst. Sci. Eng. 2023, 44, 1485–1498. [Google Scholar] [CrossRef]
- Alice, K.; Thillaivanan, A.; Koteswara Rao, G.R.; Rajalakshmi, S.; Singh, K.; Rastogi, R. Automated forest fire detection using atom search optimizer with deep transfer learning model. In Proceedings of the 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC), Salem, India, 4–6 May 2023; pp. 222–227. [Google Scholar]
- Xie, F.; Huang, Z. Aerial forest fire detection based on transfer learning and improved Faster R-CNN. In Proceedings of the 3rd International Conference on Information Technology, Big Data and Artificial Intelligence (ICIBA), Chongqing, China, 26–28 May 2023; pp. 1132–1136. [Google Scholar]






| Author | Dataset | Objective | Method | Performance |
|---|---|---|---|---|
| de Almeida, R.V. et al. (2020) [1] | Bee2Fire dataset (associated with the CICLOPE scenario) | Forest fire classification (fire/no-fire) | CNN (ResNet-18) | Specificity ≈ 99% |
| Benzekri, W. et al. (2020) [2] | Forest fire image sequence | Time-dependent fire detection | Sequence models based on RNN, LSTM, GRU | Accuracy ≈ 99.89% |
| Mohammad, M.B. et al. (2022) [3] | Forest fire images (edge device-focused) | Real-time fire detection on hardware | CNN (ResNet50, GoogLeNet, CNN-9, MobileNet, InceptionV3, AlexNet) | Accuracy ≈ 99.42% (ResNet50/GoogleNet-based) |
| Ryu, J. et al. (2022) [4] | Custom indoor/outdoor fire and smoke dataset | Mixed indoor/outdoor fire/smoke detection | Classic CV + CNN + InceptionV3-based classifier | ≈5–6% accuracy improvement over baseline methods |
| Shirwaikar, R. et al. (2025) [5] | Indoor semi-occluded fire dataset | Indoor disaster management; semi-occluded fire/spark detection + evacuation route | YOLOv8 (simplified) + instance segmentation-based damage estimation | Precision ≈ 0.73; F1 ≈ 0.81 |
| Chaoxia C. et al.(2020) [6] | BoWFire, PascalVOC and Corsician dataset | Outdoor fire images | Color and global information guided | Accuracy=99.50% |
| Casas, E. et al. (2023) [7] | Foggia fire/smoke CCTV dataset (outdoor environment, CCTV) | Comparison of different YOLO versions for early detection of smoke and wildfire | YOLOv5, YOLOv6, YOLOv7, YOLOv8, YOLO-NAS | F1 ≈ 0.95; recall ≈ 0.98 for the best model(s) |
| Sathishkumar et al. (2023) [8] | BoWFire + original forest fire image set (RGB) | Classification of forest fire and smoke images; transfer learning without forgetting old knowledge when switching to a new task | VGG16, InceptionV3, Xception + Learning without Forgetting (LwF) | Xception + LwF: ≈96.9% acc. (original), ≈91.4% acc. |
| Goncalves, A.M. et al. (2024) [9] | CICLOPE alarm images (Portugal; tower cameras) | Camera-based wildfire detection in a large area (2.7 M ha) forest | DenseNet-based feature extractor + detail-selective CNN classifier | Accuracy ≈ 99.7% (in CICLOPE alarm images) |
| Yan, C. et al. (2025) [10] | VIGP-FS forest smoke dataset | Local + global feature fusion for forest fire smoke detection | MAG-FSNet: CNN backbone + multi-scale attention + global feature fusion | Precision ≈ 88.4%; Recall ≈ 83.4%; mAP@0.5 ≈ 89.3% |
| Wang, W. et al. (2022) [11] | Forest fire image dataset | Forest fire object detection | YOLO-based CNN (forest fire detection) | Accuracy ≈ 83.9% |
| Zhu, W. et al. (2025) [12] | D-Fire drone dataset (drone-based) | UAV wildfire detection in complex forest environments | YOLOv8 (multiscale feature learning) | Precision ≈ 93.6%; Recall ≈ 88.5% |
| Wang, X. et al. (2024) [13] | Various public wildfire smoke image/video datasets | Outdoor wildfire smoke detection; classical CV + DL hybrid | SE-ResNet feature extractor + SVM classifier | Acc ≈ 98.99%; F1 ≈ 99% |
| Li, L. et al. (2021) [14] | Factory/indoor smoke dataset (ICAIIS 2021, real factory scenes) | Real-time detection of indoor (factory) smoke | Faster R-CNN-based smoke detector | Accuracy: ≈99.0% |
| Bahhar, C. et al. (2023) [15] | UAV and open datasets (wildfire, smoke; multi-source, RGB) | Staged YOLO architecture for UAV-based forest fire/smoke detection | Staged YOLO (different YOLO versions) + Ensemble CNN | Acc ≈ 99%; mAP ≈ 0.85 for smoke class |
| Li, Y. et al. (2022) [16] | Wildfire image dataset (~35 k images, tower/camera-focused) | Forest fire monitoring (smoke/fire) from surveillance towers | ResNet, EfficientNet-based end-to-end network + Grad-CAM visualization | AUROC ≈ 0.949 |
| Wang, C. et al. (2025) [17] | IR + visible (commercial area/business park) smoke dataset | Smoke detection in commercial/urban areas using IR + RGB fusion | Transformer-based Fusion (IR + Visible feature fusion) | Accuracy ≈ 90.9%; precision ≈ 98.4%; recall ≈ 92.4%; FP/FN < 5% |
| Wang, Y. et al. (2025) [18] | USTC_SmokeRS, E_SmokeRS, Aerial RS smoke datasets | Smoke detection in remote sensing images | ConvNeXt backbone + AFEM (attention feature enhancement module) + BFFM | Accuracy ≈ 98.9%; false alarm rate (FAR) ≈ 3.3% |
| Alkhammash, E.H. (2025) [19] | Smoke + D-Fire-like open-source datasets | Smoke/fire detection by comparing different YOLO generation models | Comparative analysis of YOLOv9, YOLOv10, and YOLOv11 | Precision ≈ 0.845; Recall ≈ 0.801 |
| Xue, Z. et al. (2025) [20] | Baidu Paddle wildfire + additional indoor/outdoor datasets | Indoor/outdoor multi-scenario fire/smoke detection | YOLOv11-DH3 (enhanced YOLOv11 derivative) | Precision ≈ 91.6%; Recall ≈ 90% |
| He, L. et al. (2025) [21] | WD + FFS forest fire smoke datasets | Analysis of different object detectors for forest fire smoke detection | YOLOv11x + optimized loss function (loss redesign) | Precison ≈ 0.949; Recall ≈ 0.850; highmAP@0.5 |
| Niu, K. et al. (2025) [22] | UAV fusion dataset (2752 RGB–IR image pairs) | Early forest fire detection using visible + IR fusion with UAV | YOLOv5s-based lightweight detector + image fusion | ≈10% increase in precision for small fire objects |
| Jin, P. et al. (2025) [23] | Multimodal dataset (3352 paired samples: imagery + environmental) | Smoke detection + early warning + risk assessment | YOLOv8n + MSDBlock (multiscale dual-branch block) | Accuracy ≈ 93.1%; ≈18.8% higher than traditional baselines |
| Shang, L. et al. (2025) [24] | Custom dataset (20,044 images; indoor, industrial, forest) | Multi-scenario (industrial + forest + indoor/outdoor) flame and spark detection | YOLO-DKM (improved YOLOv8; deformable conv + key modules) | Precision ≈ 82.1%; Recall ≈ 71.8% |
| Arteaga, B. et al. (2020) [25] | Forest fire images | Fire presence/absence classification | CNN (ResNet + VGG-based deep classifiers) | Accuracy ≈ 99.5% |
| Mohammed, R.K. (2022) [26] | Forest fire image dataset | Fire/smoke image classification | Inception-ResNet-based CNN | Accuracy ≈ 99.09% |
| Mohnish, S. et al. (2022) [27] | Forest fire image dataset | Fire image detection (bounding box level) | CNN-based detection (custom, single-stage) | Accuracy ≈ 92.20% |
| Ban, Y. et al. (2020) [28] | Sentinel-1 SAR time series (district/region-based) | Near real-time monitoring of fire progression | CNN-based time series model (on SAR images) | Accuracy ≈ 83.53% |
| Rahul, M. et al. (2020) [29] | Forest fire images | Fire/no-fire classification | CNN (ResNet50, VGG16, DenseNet121) | Accuracy ≈ 92.27% |
| Jiang, Y. et al. (2021) [30] | Qinling forest fire anomaly dataset | Forest anomaly/fire detection | CNN + BP NN, GA, SVM, GA-BP optimization | Accuracy ≈ 95% |
| Li, M. et al. (2021) [31] | Forest fire image dataset | Early forest fire detection (object detection) | h-EfficientDet (EfficientDet + h-EfficientDet architecture) | Accuracy ≈ 98.35% |
| Khan, S. et al. (2022) [32] | Forest fire detection dataset (Fire/No-Fire) | Fire/no-fire classification (smart city scenario) | FFireNet (MobileNetV2-based CNN) + MobileNetV2 comparison | Accuracy ≈ 98.42%; FFireNet + MobileNetV2 |
| Gayathri, S. et al. (2022) [33] | Forest fire image dataset | Forest fire prediction and detection | CNN-based classifier | Accuracy ≈ 96% |
| Kang, Y. et al. (2022) [34] | Geostationary satellite data (GEO; multi-temporal) | Early forest fire detection using GEO satellite data | CNN + Random Forest hybrid approach | Accuracy ≈ 98% |
| Ghosh, R.; Kumar, A. (2022) [35] | Forest fire image dataset | Spatial + temporal pattern learning for fire detection | CNN + RNN hybrid (e.g., LSTM layers) | Accuracy ≈ 99.62% |
| Tahir, H.U.A. et al. (2022) [36] | UAV wildfire imagery | Wildfire detection in UAV images | YOLOv5-based detection | F1 ≈ 94.44% |
| Peng, Y.; Wang, Y. (2022) [37] | Wildfire monitoring image dataset | Deep learning-based automatic wildfire monitoring system | CNN (SqueezeNet1.1, AlexNet, MobileNet, ResNet18, VGG16 comparisons) | Accuracy ≈ 99.28% |
| Mashraqi, A.M. et al. (2022) [38] | Drone imagery forest fire dataset | Fire classification from drone images | DIFFDC-MDL hybrid (LSTM-RNN + MobileNetV2) | Accuracy ≈ 99.38% |
| Almasoud, A.S. (2023) [39] | Forest fire image dataset | Smart, DL-based wild forest fire detection | IWFFDA-DL, ACNN-BLSTM + YOLOv3 combination | Accuracy ≈ 99.56% |
| Alice, K. et al. (2023) [40] | Forest fire image dataset | Automatic forest fire detection | Deep transfer learning (QRNN, ResNet50 + Atom Search Optimizer) | Accuracy ≈ 97.33% |
| Xie, F.; Huang, Z. (2023) [41] | UAV wildfire dataset | Fire/smoke detection in UAV images | Transfer Learning + Faster R-CNN (ResNet50 backbone + fusion and attention) | Accuracy ≈ 93.7% |
| Augmentation Type | Applied Set | Probability/Range | Purpose |
|---|---|---|---|
| Random Resized Crop | Training only | Scale: 0.75–1.0 | Scale and viewpoint variation |
| Horizontal Flip | Training only | 0.5 | Viewpoint invariance |
| Rotation | Training only | ±10° | Orientation robustness |
| Gaussian Blur | Training only | 0.25 | Sensor noise simulation |
| Color Jitter | Training only | 0.5 | Illumination variation |
| Normalization | Training and Validation | ImageNet mean/std | Stable optimization |
| Dataset | Split | Real Images | Epochs | Effective Training Samples |
|---|---|---|---|---|
| FireStage | Training | 632 | 10 | ~6320 |
| FireStage | Validation | 159 | - | 159 |
| FLAME | Training | 31,500 | 10 | ~315,000 |
| FLAME | Validation | 7875 | - | 7875 |
| Dataset Type | RGB | Night Vision | White | Green |
|---|---|---|---|---|
| FireStage Dataset | (3, 224, 224) | (3, 224, 224) | (3, 224, 224) | (3, 224, 224) |
| FLAME Dataset | (3, 224, 224) | (3, 224, 224) | (3, 224, 224) | (3, 224, 224) |
| Stage | Layer | Input Size | Output Size | Description |
|---|---|---|---|---|
| 1 | Fully connected | 4d (e.g., 2048) | 256 | Dimensionality reduction |
| 2 | Activation | 256 | 256 | Nonlinear transformation (ReLU fuction) |
| 3 | Dropout | 256 | 256 | Probability = 0.3 |
| 4a | Fully connected | 256 | 2 | Binary classification (No Fire, Fire) |
| 4b | Fully connected | 256 | 3 | Three-class classification (No Fire, Start Fire, Fire) |
| Metric | Definition | Formula |
|---|---|---|
| Accuracy | The ratio of examples correctly classified by the model to the total number of examples. | |
| Precision | The proportion of correct classes among the examples predicted as positive. | |
| Recall | The proportion of true positive examples that are correctly predicted. | |
| F1 Score | The harmonic mean that balances recall and precision. This metric is suitable for use in data imbalance. | |
| AUC | The Area Under the Receiver Operating Characteristic (ROC) Curve, representing the model’s ability to discriminate between classes across all classification thresholds. Higher values indicate better separability. | |
| ROC Curve | A graphical representation of the trade-off between the True Positive Rate (TPR) and the False Positive Rate (FPR) across different decision thresholds. |
| Dataset Type | Number of Images | Training Dataset Count | Test Dataset Count |
|---|---|---|---|
| FireStage Dataset | 791 | 632 | 159 |
| FLAME Dataset | 39,375 | 31,500 | 7875 |
| Dataset | Backbone | Modality | Acc | Prec | Recall | F1 | AUC | Params | Train Time (s) |
|---|---|---|---|---|---|---|---|---|---|
| FLAME | DenseNet121 | RGB | 0.996571 | 0.997999 | 0.996603 | 0.997300 | 0.999723 | 7.22 M | 1668 |
| FLAME | DenseNet121 | Green | 0.987429 | 0.991976 | 0.988209 | 0.990089 | 0.998736 | 7.22 M | 1271 |
| FLAME | DenseNet121 | Night | 0.983365 | 0.989355 | 0.984412 | 0.986878 | 0.998800 | 7.22 M | 1203 |
| FLAME | DenseNet121 | White | 0.990603 | 0.991820 | 0.993405 | 0.992612 | 0.999256 | 7.22 M | 1223 |
| FLAME | DenseNet121 | RGB + Green | 0.860190 | 0.998977 | 0.780775 | 0.876500 | 0.997340 | 7.48 M | 1804 |
| FLAME | DenseNet121 | RGB + Night | 0.833143 | 0.998918 | 0.738209 | 0.849000 | 0.996452 | 7.48 M | 1806 |
| FLAME | DenseNet121 | RGB + White | 0.995302 | 0.995214 | 0.997402 | 0.996307 | 0.999719 | 7.48 M | 1818 |
| FLAME | DenseNet121 | RGB + N + W + G | 0.943238 | 0.996947 | 0.913469 | 0.953384 | 0.997075 | 8.00 M | 6804 |
| FLAME | EfficientNet-B0 | RGB | 0.995429 | 0.997796 | 0.995004 | 0.996398 | 0.999822 | 4.34 M | 1579 |
| FLAME | EfficientNet-B0 | Green | 0.986032 | 0.994543 | 0.983413 | 0.988947 | 0.998722 | 4.34 M | 1157 |
| FLAME | EfficientNet-B0 | Night | 0.982603 | 0.989342 | 0.983213 | 0.986268 | 0.997983 | 4.34 M | 1227 |
| FLAME | EfficientNet-B0 | White | 0.989333 | 0.992788 | 0.990408 | 0.991597 | 0.998870 | 4.34 M | 1233 |
| FLAME | EfficientNet-B0 | RGB + Green | 0.810794 | 0.998864 | 0.703038 | 0.825240 | 0.995737 | 4.66 M | 1718 |
| FLAME | EfficientNet-B0 | RGB + Night | 0.841651 | 0.996301 | 0.753597 | 0.858118 | 0.993732 | 4.66 M | 1744 |
| FLAME | EfficientNet-B0 | RGB + White | 0.994032 | 0.995007 | 0.995604 | 0.995305 | 0.999671 | 4.66 M | 1723 |
| FLAME | EfficientNet-B0 | RGB + N + W + G | 0.846730 | 0.996600 | 0.761391 | 0.863260 | 0.994949 | 5.32 M | 6519 |
| FLAME | ResNet18 | RGB | 0.991746 | 0.997382 | 0.989608 | 0.993480 | 0.999608 | 11.31 M | 1548 |
| FLAME | ResNet18 | Green | 0.973206 | 0.992398 | 0.965228 | 0.978624 | 0.997508 | 11.31 M | 1157 |
| FLAME | ResNet18 | Night | 0.974222 | 0.974126 | 0.985612 | 0.979835 | 0.996641 | 11.31 M | 1300 |
| FLAME | ResNet18 | White | 0.989206 | 0.987512 | 0.995604 | 0.991541 | 0.998954 | 11.31 M | 1186 |
| FLAME | ResNet18 | RGB + Green | 0.962413 | 0.944486 | 0.999600 | 0.971262 | 0.995450 | 11.44 M | 1723 |
| FLAME | ResNet18 | RGB + Night | 0.975873 | 0.977959 | 0.984213 | 0.981076 | 0.997171 | 11.44 M | 1896 |
| FLAME | ResNet18 | RGB + White | 0.987556 | 0.981169 | 0.999600 | 0.990299 | 0.999438 | 11.44 M | 1707 |
| FireStage | DenseNet121 | RGB | 0.937107 | 0.923737 | 0.931411 | 0.927202 | 0.991925 | 7.22 M | 176 |
| FireStage | DenseNet121 | Green | 0.805031 | 0.811814 | 0.777452 | 0.789884 | 0.930476 | 7.22 M | 99 |
| FireStage | DenseNet121 | Night | 0.767296 | 0.769202 | 0.735745 | 0.745529 | 0.922191 | 7.22 M | 104 |
| FireStage | DenseNet121 | White | 0.798742 | 0.816472 | 0.755947 | 0.772497 | 0.927145 | 7.22 M | 100 |
| FireStage | DenseNet121 | RGB + Green | 0.748428 | 0.787677 | 0.788856 | 0.737219 | 0.968725 | 7.48 M | 129 |
| FireStage | DenseNet121 | RGB + Night | 0.842767 | 0.831060 | 0.854350 | 0.827998 | 0.974003 | 7.48 M | 132 |
| FireStage | DenseNet121 | RGB + White | 0.930818 | 0.925240 | 0.920658 | 0.922773 | 0.989529 | 7.48 M | 129 |
| FireStage | DenseNet121 | RGB + N + W + G | 0.893082 | 0.874812 | 0.900782 | 0.882311 | 0.979556 | 8.00 M | 441 |
| FireStage | EfficientNet-B0 | RGB | 0.905660 | 0.890713 | 0.889052 | 0.889729 | 0.966492 | 4.34 M | 170 |
| FireStage | EfficientNet-B0 | Green | 0.773585 | 0.773148 | 0.736722 | 0.744017 | 0.913076 | 4.34 M | 100 |
| FireStage | EfficientNet-B0 | Night | 0.773585 | 0.763612 | 0.764255 | 0.755687 | 0.922556 | 4.34 M | 102 |
| FireStage | EfficientNet-B0 | White | 0.798742 | 0.791329 | 0.768003 | 0.774298 | 0.916783 | 4.34 M | 101 |
| FireStage | EfficientNet-B0 | RGB + Green | 0.729560 | 0.774043 | 0.772890 | 0.724710 | 0.942445 | 4.66 M | 128 |
| FireStage | EfficientNet-B0 | RGB + Night | 0.867925 | 0.848146 | 0.868524 | 0.854238 | 0.960164 | 4.66 M | 131 |
| FireStage | EfficientNet-B0 | RGB + White | 0.930818 | 0.915282 | 0.915282 | 0.915282 | 0.961032 | 4.66 M | 125 |
| FireStage | EfficientNet-B0 | RGB + N + W + G | 0.861635 | 0.841486 | 0.864451 | 0.844856 | 0.960082 | 5.32 M | 416 |
| FireStage | ResNet18 | RGB | 0.924528 | 0.934972 | 0.903877 | 0.916292 | 0.984546 | 11.31 M | 169 |
| FireStage | ResNet18 | Green | 0.729560 | 0.743901 | 0.710166 | 0.720018 | 0.871148 | 11.31 M | 97 |
| FireStage | ResNet18 | Night | 0.723270 | 0.724347 | 0.706419 | 0.702695 | 0.874561 | 11.31 M | 104 |
| FireStage | ResNet18 | White | 0.742138 | 0.720044 | 0.738514 | 0.715710 | 0.900683 | 11.31 M | 97 |
| FireStage | ResNet18 | RGB + Green | 0.874214 | 0.851534 | 0.861193 | 0.854235 | 0.963100 | 11.44 M | 124 |
| FireStage | ResNet18 | RGB + Night | 0.823899 | 0.804250 | 0.821114 | 0.806133 | 0.943965 | 11.44 M | 132 |
| FireStage | ResNet18 | RGB + White | 0.905660 | 0.911584 | 0.867221 | 0.880889 | 0.978778 | 11.44 M | 122 |
| FireStage | ResNet18 | RGB + N + W + G | 0.880503 | 0.863298 | 0.890681 | 0.867819 | 0.962310 | 11.70 M | 416 |
| Study | Dataset | Method | Modality | Class | Accuracy (%) |
|---|---|---|---|---|---|
| Benzekri, W. et al. (2020) [2] | Fire Detection (image sequence) | LSTM/GRU | RGB (Sequence) | 2 | 99.89 |
| Goncalves, A.M. et al. (2024) [9] | FLAME Dataset | DenseNet + CNN | RGB | 2 | 99.70 |
| Wang, X. et al. (2024) [13] | Fire Detection (public wildfire smoke datasets) | SE-ResNet + SVM | RGB | 2 | 98.99 |
| Arteaga, B. et al. (2020) [25] | Fire Detection (image dataset) | CNN (ResNet + VGG) | RGB | 2 | 99.50 |
| Mohammed, R.K. (2022) [26] | Fire Detection (image dataset) | Inception-ResNet | RGB | 2 | 99.09 |
| Mohnish, S. et al. (2022) [27] | Fire Detection (image dataset) | CNN Detection | RGB | 2 | 92.20 |
| This Study (TAŞAR et al.) | FireStage Dataset | DenseNet | RGB | 3 | 93.71 |
| RGB + White | 93.08 | ||||
| RGB + Night + White + Green | 89.31 | ||||
| This Study (TAŞAR et al.) | FLAME Dataset | DenseNet | RGB | 2 | 99.66 |
| RGB + White | 99.53 | ||||
| RGB + Night + White + Green | 89.31 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Taşar, B.; Tatar, A.B.; Tanyildizi, A.K.; Yakut, O. Multimodal Wildfire Classification Using Synthetic Night-Vision-like and Thermal-Inspired Image Representations. Fire 2026, 9, 109. https://doi.org/10.3390/fire9030109
Taşar B, Tatar AB, Tanyildizi AK, Yakut O. Multimodal Wildfire Classification Using Synthetic Night-Vision-like and Thermal-Inspired Image Representations. Fire. 2026; 9(3):109. https://doi.org/10.3390/fire9030109
Chicago/Turabian StyleTaşar, Beyda, Ahmet Burak Tatar, Alper Kadir Tanyildizi, and Oğuz Yakut. 2026. "Multimodal Wildfire Classification Using Synthetic Night-Vision-like and Thermal-Inspired Image Representations" Fire 9, no. 3: 109. https://doi.org/10.3390/fire9030109
APA StyleTaşar, B., Tatar, A. B., Tanyildizi, A. K., & Yakut, O. (2026). Multimodal Wildfire Classification Using Synthetic Night-Vision-like and Thermal-Inspired Image Representations. Fire, 9(3), 109. https://doi.org/10.3390/fire9030109

