Next Article in Journal
Intelligent Firefighting Technology for Drone Swarms with Multi-Sensor Integrated Path Planning: YOLOv8 Algorithm-Driven Fire Source Identification and Precision Deployment Strategy
Next Article in Special Issue
Hierarchical Reinforcement Learning for Viewpoint Planning with Scalable Precision in UAV Inspection
Previous Article in Journal
LSTM-H: A Hybrid Deep Learning Model for Accurate Livestock Movement Prediction in UAV-Based Monitoring Systems
Previous Article in Special Issue
A Review on Deep Learning for UAV Absolute Visual Localization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advancing Early Wildfire Detection: Integration of Vision Language Models with Unmanned Aerial Vehicle Remote Sensing for Enhanced Situational Awareness

1
Fraunhofer Institute for Integrated Systems and Device Technology (IISB), Schottkystraße 10, 91058 Erlangen, Germany
2
Fraunhofer Institute for Integrated Circuits (IIS), Am Wolfsmantel 33, 91058 Erlangen, Germany
3
Institute of Power Electronics, Friedrich-Alexander-Universität Erlangen-Nürnberg, Fürther Straße 250, 90429 Nuremberg, Germany
*
Author to whom correspondence should be addressed.
Drones 2025, 9(5), 347; https://doi.org/10.3390/drones9050347
Submission received: 18 March 2025 / Revised: 17 April 2025 / Accepted: 30 April 2025 / Published: 3 May 2025

Abstract

:
Early wildfire detection is critical for effective suppression efforts, necessitating rapid alerts and precise localization. While computer vision techniques offer reliable fire detection, they often lack contextual understanding. This paper addresses this limitation by utilizing Vision Language Models (VLMs) to generate structured scene descriptions from Unmanned Aerial Vehicle (UAV) imagery. UAV-based remote sensing provides diverse perspectives for potential wildfires, and state-of-the-art VLMs enable rapid and detailed scene characterization. We evaluated both cloud-based (OpenAI, Google DeepMind) and open-weight, locally deployed VLMs on a novel evaluation dataset specifically curated for understanding forest fire scenes. Our results demonstrate that relatively compact, fine-tuned VLMs can provide rich contextual information, including forest type, fire state, and fire type. Specifically, our best-performing model, ForestFireVLM-7B (fine-tuned from Qwen2-5-VL-7B), achieved a 76.6% average accuracy across all categories, surpassing the strongest closed-weight baseline (Gemini 2.0 Pro at 65.5%). Furthermore, zero-shot evaluation on the publicly available FIgLib dataset demonstrated state-of-the-art smoke detection accuracy using VLMs. Our findings highlight the potential of fine-tuned, open-weight VLMs for enhanced wildfire situational awareness via detailed scene interpretation.

1. Introduction

Wildfires are burning large areas yearly and are responsible for substantial losses of life and enormous CO2 emissions. They are expected to increase in likelihood and severity in the near future due to factors like climate change [1]. In Germany, while precautionary measures have been established for residential areas and industrial plants through demand planning and pre-assessment of potential hazards, a comparable approach for forest fires is lacking. Consequently, emergency services often face unclear hazard situations upon alarm activation. The ever-changing vitality of vegetation further complicates the hazard potential, due to seasonal influences, pest infestations, or droughts. Moreover, forest fires are highly dynamic situations that can engage a large number of emergency personnel. Therefore, early and comprehensive information gathering is essential to enable timely reactions [2]. In addition to static techniques like watchtowers and wireless sensor networks, UAV-based automatic remote sensing platforms have been developed [3]. Traditionally, these platforms have utilized deep learning models such as Convolutional Neural Networks and Long Short-Term Memory networks to detect smoke through image classification tasks [4]. Recent breakthroughs in large language modeling and combining LLMs with vision capabilities have extended what is possible with computer vision [5,6]. Specifically, VLMs enable detailed captioning of images, combined with advanced reasoning, thereby enhancing scene understanding. VLMs are increasingly being used in real-world applications such as autonomous driving and robotics [7]. Moreover, multiple works have discussed using VLMs for general remote sensing tasks [8,9].
Our research goals were as follows:
  • Build a framework for detailed descriptions of forest fires in a structured output
  • Evaluate the performance of untrained closed, cloud-based VLMs
  • Fine-tune a lightweight open-weight dataset
  • Improve VLM-based detection performance on the FIgLib Dataset

1.1. Evolonic

Evolonic is a student research team at Friedrich-Alexander-University Erlangen and in close cooperation with the Fraunhofer IISB. The authors are current team members or have been in the past.

1.1.1. NF4 UAV

An electric VTOL (Vertical Takeoff and Landing) drone was previously designed and built by Evolonic. The drone features a separate lift and thrust powertrain, as shown in Figure 1. It combines the advantages of a rotary-wing drone, such as the lack of need for a runway and the ability to stop in midair, with the long-range and efficient flight of a fixed-wing aircraft [10]. As the powertrains for the two flight modes are separate, this also serves as an added layer of redundancy. The drone has a wingspan of just under three meters and a cruising speed of 65 km/h, which leads to a range of about 100 km per flight before it has to be recharged. Charging is carried out with a mobile base station that can be deployed to strategic hot spots. The UAV is equipped with a front-facing camera and an onboard companion computer in order to detect smoke. It also has the ability to send data, pictures, and control commands via LTE or a direct radio-link.

1.1.2. Automatic Operations

The operation of the system centers around the automatic surveillance of predefined forest areas with the UAVs. Depending on the forest fire danger index [11], the system will schedule more or fewer flights per day. The UAV will automatically take off from the base station and follow a preplanned patrol route. If no smoke is detected, the drone will fly back to the base station and recharge for the next flight. If smoke is detected by the onboard AI, an alert is automatically sent to the fire brigade dispatch center (Figure 2a). The drone will then fly towards the detected hotspot, thus increasing the location accuracy. To support the extinguishing efforts, the UAV will then continue to orbit overhead and provide a live feed of the situation (Figure 2b).

1.1.3. Software Architecture

A ROS2-based architecture was implemented, containing a node for smoke detection using YOLOv8 instance segmentation. This detection model was annotated and trained by Evolonic using previous forest fires, internet images, and data from fire brigades across Germany. Positive detections are sent to a web backend via MQTT together with environmental data, geo-referencing, and a time stamp. A web application, made available to dispatch centers and fire brigades, visualizes these data as a live stream and maps.

1.1.4. Test Flights

The system was evaluated in two separate field studies. The first of these was a durability test of the drone, which took place in southern Germany in the summer of 2023. This test was performed to ensure the long-term durability of the UAV. Over a span of about 30 days, almost 50 flights were conducted, from which no major system failures or material fatigue were observed. Only one servo for one of the ailerons was damaged during transportation, but this was detected during the preflight inspection. During this test, the stability of the connection system was also evaluated. Although LTE-connectivity can be difficult, especially in sparsely populated areas, the connection at the cruising altitude of 120 m above ground was proven to be very stable. While the overall durability test was a complete success, the detection pipeline could not be verified, due to a lack of forest fires during that time and at that location. Therefore, a second test was planned. Here, the goal was to measure the time and maximum distance needed to successfully detect a wildfire. A consortium of different research teams from the German Federal Institute for Material Research and Testing (BAM), the Otto von Guericke University Magdeburg, and OneSeven set up a patch of forest to deliberately set on fire, to study various parameters such as spreading speed, new extinguishing foams, and detection with long-range UAVs. The smoke detection system trained from previous data was able to detect smoke in images captured from this event.

1.2. Related Works

1.2.1. Wildfire Detection Methods

While in densely populated areas, most wildfires are reported by the public as visitors to the forest or people that live nearby, studies have shown that in remote locations, about 30% to 70% of fires are not detected for a long time period [11]. Therefore, specialized methods like towers staffed by human watchers have been established. Although these trained personnel perform very well in detecting even small fires at an early stage, cost concerns have driven the development of more automated systems [12]. One of these is the use of camera-based watchtowers. This technology utilizes visible light or infrared cameras in combination with machine learning algorithms to detect smoke or fire glow. Through triangulation, a reasonable estimation of the fire location can be calculated. However, this calculation is prone to errors, especially in hilly or mountainous terrain [12,13]. Satellites are also used in forest fire detection, which can monitor huge areas and be used for other purposes. The use of onboard AI has helped to combat false alarms, such as sun glints or water bodies. However, satellites are very dependent on their orbit times, making detection periods very long in most cases. Furthermore, they cannot provide a live picture of the situation. Clouds or smoke can also hinder fire detection, even with IR sensors. Another factor is that, while prices in the space industry have come down since the last century, they still pose a significant investment [14]. Wireless Sensor Networks (WSNs) utilize small, self-sufficient gas, humidity, or temperature sensors. These sensors are placed strategically every 100–300 m and communicate via low-power wireless networks or 4G/LTE to form a mesh. This technology is particularly effective in detecting small ground-level fires. However, due to the amount of sensors needed and the work required to place them, it is neither financially nor logistically feasible to place them over large areas [13,15]. In recent years, the use of UAVs in battling forest fires has increased. They can roughly be split into three categories. The first is mainly consumer or hobby-grade drones that are deployed as a short-distance lookout or for observation to navigate the fire brigade to a previously detected hot spot or provide a live feed of the ongoing extinguishing actions. These drones usually have a relatively short flight time and are steered by personnel on the ground. The second category includes specially built long-range, fixed-wing drones that automatically patrol preset areas and alert firefighters if a fire is detected. Lastly, some research has also been conducted on fighting fires with the help of drone swarms [14,16].

1.2.2. VLMs/LLMs for UAVs

Even if VLMs are still a relatively young research field, some work has already been carried out exploring how to use them in the context of UAVs. Sautenkov et al. [17] leveraged VLMs to automatically generate missions from text inputs and satellite images. Sakaino [18] showed how VLMs, among other Deep Learning models, are enhancing segmentation, captioning, and visibility under adverse weather conditions. Wei and Kulkarni [19] researched the potential of using VLMs for wildfires and evaluated certain models using the FIgLib dataset. But this technology can be significantly enhanced to provide fire brigades with a better understanding of the situation and to structure outputs for seamless integration into user interfaces like websites. These situation descriptions can contain detailed information about the fire and its environment, formatted simply and understandably. While some research has been performed on integrating VLMs into UAVs, currently, no method has been proposed for using this for wildfire detection with drones.

2. Materials and Methods

2.1. Wildfire Description

Our approach involves enhancing the information available to fire brigades after smoke detection. Therefore, we define several categories that should be answered by the forest fire VLM in the following section. Forest and vegetation fires are complex events that are influenced by a variety of factors. In addition to the classic elements of the fire triangle (oxygen, energy, fuel), environmental factors play a critical role. These are summarized in the so-called fire behavior triangle of topography, weather, and fuel properties [20]. The topography of the location and the resulting fire behavior can be derived from geo data. Important meteorological parameters such as solar radiation, wind direction, and wind speed can be determined from local weather station data. A central aspect is the nature of the fuel [21]. To better assess this, one of the labels used rates the vitality of the surrounding trees into categories of “vital”, “moderately vital”, “declining”, “dead”, “cannot be determined”, and “no forest fire visible” [22]. Trees that are dead or infested by pests, in particular, have a low resistance and can accelerate the spread of fire. Observations in Germany also show that forest fires spread particularly quickly in spring before flowering, as the trees are not yet fully supplied with water and have dried out over the winter. This information is an essential factor in predicting the further spread of the fire [23]. For further operational planning, it is also important to assess the potential risks to people and infrastructure. For this purpose, specific questions are asked, such as “Are people visible near the forest fire?” or “Is infrastructure visible near the forest fire?” [22]. Another decisive criterion is the inspection of the fire. As many fire brigades in Germany work voluntarily, false alarms repeatedly lead to volunteers being torn from their private or professional lives [24]. Possible false alarms can be caused, for example, by controlled burning of green cuttings, fog, and swirling dust. To ensure that a forest fire is recognized reliably, the following questions are therefore asked: “Can smoke from a forest fire be seen in the picture?”, “Can flames from a forest fire be seen in the picture?”, and “Can it be confirmed that this is an uncontrolled forest fire?”. In addition to fire detection, dynamic factors are also relevant, in order to assess the development and challenges of firefighting. The following parameters are recorded for this purpose: “What state is the forest fire currently in?”, “How big is the fire?”, “How intense is the fire?”, “What type of fire is it?”, and “Are there multiple sources of fire?” [20]. The full list of questions and corresponding answer options can be seen in Table 1. The combination of these parameters enables a comprehensive assessment of the situation and supports the emergency services in making well-founded decisions. An early and automated analysis of relevant influencing variables can significantly increase the effectiveness of firefighting.

2.2. VLMs

2.2.1. Architectures

Vision Language Models (VLMs) integrate a Large Language Model (LLM) with the ability to process visual data in the form of images or videos. Various architectures have been explored to achieve this integration, with two predominant approaches emerging in state-of-the-art VLMs: fully auto-regressive architectures, and cross-attention architectures. The first architecture comprises a pre-trained vision encoder and a multimodal adapter [25]. The vision encoder processes visual inputs, while the multimodal adapter maps the outputs of the encoder to the input format required by the LLM.

2.2.2. State-of-the-Art Models

We initiated our investigation by evaluating both closed-source and open-source Vision Language Models (VLMs) on our designated evaluation dataset. Closed-source VLMs serve as benchmarks for the current state of the art, while fine-tunable open-source models were of primary interest to our study. To select suitable VLMs for extracting detailed and structured information from forest fire images, we considered several key factors, such as the computational cost of running a model, either through API credits or on local hardware, results on common benchmarks, and the availability of established fine-tuning frameworks and associated code. Our goal was to identify open-source models that could operate within the 24 GB VRAM capacity of the available NVIDIA RTX 3090 GPU, with an ideal target of running on Evolonic’s 16 GB NVIDIA Jetson Orin NX. Assuming 16-bit weights, this constraint allowed for models up to approximately 10 billion total parameters; however, quantization techniques could potentially accommodate larger models. Various benchmarks exist for comparing VLMs, ranging from Optical Character Recognition (OCR) to mathematical reasoning and evaluating the tendencies of the model towards hallucinations [7]. The OpenVLM Leaderboard on Huggingface aggregates 31 such benchmarks, executable via the VLMEvalkit [26]. For compatibility with fine-tuning frameworks, we prioritized models that are compatible with LLaMA-Factory [27] or Unsloth [28], as well as custom training code provided by the model developers. Based on these criteria, we selected Google DeepMind’s Gemini 2.0 models and OpenAI’s GPT-4o family as closed-source references. For open-weight models, we chose the Qwen2.5-VL family, which includes models with 3B, 7B, and 72B parameters [6]. The InternVL 2.5 series of models [29] performs comparably to or slightly worse than Qwen2.5-VL and is currently unsupported by the aforementioned fine-tuning frameworks. All of these models have demonstrated strong performance on general benchmarks such as MMMU [30], and MME-RealWorld [31], which more closely resemble our forest fire description task.

2.2.3. Prompting and Structured Outputs

Instead of allowing the VLM to freely generate outputs, we use structured outputs to only allow a given JSON pattern as an answer. This JSON scheme contains the 11 answer categories derived in Section 2.1, each allowing an enumeration of 2 to 6 possible text answers to choose from. All of these fields have to be present in the output of the LLM, without allowing any additional fields. While binary classification is only used with the fields Smoke and Flames, the other fields also allow looser answers. Most answers accept “Cannot be determined” in case the situation cannot be fully resolved from the image. In the case of the field Uncontrolled, which answers if the forest fire is out of control or if it might be a controlled burning, the answer “Closer investigation required” is possible. The model is also prompted to answer subsequent fields with “No forest fire visible” in case the binary classification is negative.

2.2.4. Datasets

We worked with three different datasets to train and evaluate our approach. The first two are new datasets for the specific task of forest fire description. The dataset ForestFireInsights-Eval consists of 301 images, while the training set ForestFireInsights-Train consists of 1196 images. The imagery was sourced from four distinct origins: Evolonic drone footage, drone footage obtained from various fire brigades across Germany, a publicly accessible dataset from the University of Split [32,33], and Internet videos. Evolonic’s contributions include images captured during an actual wildfire in Tennenloher Forst near Erlangen, Germany (Figure 3), a controlled smoke test in Erlangen, and a forest fire simulation near Calvörde, Germany. The evaluation dataset is publicly available and exclusively features images sourced from Evolonic and the University of Split. In contrast, the training dataset incorporated frames extracted from internet videos and confidential images provided by fire brigades, many of which were shared under non-redistribution agreements. Annotations were created by the authors, containing the fields explained in Section 2.1. These keys were then put in a predefined JSON answer scheme and converted to a single text string. For evaluation purposes, the original JSON-like Python dictionary was also saved in the dataset. The annotation workflow included creating preliminary suggestions with an off-the-shelf VLM before manually editing all fields in a web-hosted Argilla environment.
The third dataset used in this work was the test set of the FIgLib dataset [4], consisting of 4880 images from stationary wildfire cameras in California. These were not manually annotated by us but contain a timestamp relative to the outbreak of the forest fire visible in the images. Only the smoke detection was therefore validated on this dataset, allowing a comparison with other state-of-the-art methods. We developed a script to convert the images to a format compatible with our evaluation code and published the converted dataset.
It was ensured that the training dataset did not contain any images from fire events that overlapped with those used in either one of the evaluation datasets. The distribution of annotations across all categories for all three datasets is documented in Appendix A, while Table 2 compares them in regard of their purpose and contents.

2.2.5. Training

Fine-tuning VLMs on custom datasets mostly relies on three different approaches: Full fine-tuning, Low-Rank Adaption (LoRA) [34], and Quantized Low-Rank Adaption (QLoRA) [35]. While the first approach fine-tunes all model weights, the latter only trains on a small subset, typically less than 1 %. Instead of relying on FP16 values, as with full fine-tuning and LoRA, QLoRA utilizes a 4-bit quantization of the base model [35]. This leads to a further reduction in required training memory compared to LoRA, which is already multiple times more memory-efficient than full fine-tuning. Training both LoRA and QLoRA produces adapters that can be merged with the base model afterwards. We used the LLaMA-Factory framework [27] for fine-tuning the Qwen2.5-VL models with the LoRA method. Different learning rates, batch sizes, and numbers of epochs were tested and evaluated, with the best results shown in Table 3.
Figure 4 shows the loss curves for the best training runs of the 3B and 7B models fine-tuned from Qwen2.5-VL, captured with weights and biases.

2.2.6. Evaluation

Evaluation was performed using a script made publicly available in our repository. First, text predictions from the model were generated on the samples of the evaluation dataset. The script offers two possible backends for this purpose: the Google Gemini API, and an OpenAI-compatible endpoint, which can also be used for local inference with compatible frameworks. We used vLLM [36] for this task, which allows running Qwen2.5-VL models and supports guided decoding backends for generating structured outputs. After all predictions had been generated, we validated the structured output, converted the text to a Python dictionary, and compared all keys to the human-annotated groundings from the evaluation dataset. For keys that contained an enumeration of possible values, only the accuracy of correct values was computed, while for the smoke detection, precision, recall, and  F 1  score were also computed. The metrics were computed using the following equations, also used by other reference works [4,19]:
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l
The separation of prediction generation also allowed adding other inference backends, without changing the actual evaluation code. All evaluations across inference methods and providers were made with a temperature setting of 0.0, which should improve the reproducibility of our results.
The context length depended on the number of tokens required for the text prompt, as well as the tokenized image. For Qwen VLMs and their fine-tuned versions, the number of tokens for our text prompt including the chat template was 526, while the number of image tokens was dependent on the image pixel resolution [6,37]:
Image Tokens = H e i g t h × W i d t h ( 14 × 14 ) × 4
For the evaluations performed with our fine-tuned VLM on our evaluation set, we set the context length of the model to 4500 tokens, as in the training. The FIgLib dataset contains images of a higher resolution (4 to 6 Megapixels), requiring a context length of nearly 9000 tokens. Due to the higher image resolution and larger number of entries, the evaluations on this dataset were performed with batched inference using a thread pool with up to 64 parallel workers. Wei and Kulkarni [19] obtained better FIgLib results with horizon cropping. As this approach is only applicable to fixed-horizon images, unlike UAV imagery, we did not implement this approach and solely compared with zero-shot evaluation. For the cloud-based models, we did not evaluate Gemini 2.0 Pro, as only 5 requests per minute were allowed at the time of writing.

2.2.7. Computing and API Checkpoints

Training and evaluation processes were conducted on a combination of local and rented cloud GPUs. Evaluation runs typically used an NVIDIA RTX 3090, RTX 4090, or RTX 4060 Ti for the ForestFireInsights-Eval dataset, requiring between 16 and 24 GB of VRAM. The parallelized evaluation runs on the larger FIgLib dataset required using an NVIDIA A100 with 80 GB of VRAM. Training runs were performed on similar hardware, as noted in Table 3. Closed weight models from Google DeepMind and OpenAI were employed in conjunction with their respective APIs. Table 4 shows the respective checkpoints or versions used for the cloud-based API models.

3. Results

3.1. ForestFireInsights-Eval

The evaluation dataset was tested using closed, cloud-based VLMs, open-weight models, and our fine-tuned models. We present the results for all output fields grouped by their coarse category. The first category encompasses binary detection of smoke and flames, and estimating whether the fire is uncontrolled or if further investigation is required. The percentages for these fields are visualized in Table 5.
The best zero-shot accuracy for smoke detection was achieved by Gemini 2.0 Pro at 95.7%. In the Uncontrolled category, both our fine-tuned models had an advantage over most other models, except for Gemini Flash 2.0. Here, Gemini models generally performed better than OpenAI and Qwen models. The best overall score was achieved by ForestFireVLM-7B at 90.0%, closely followed by the 3B variant and Gemini Flash 2.0.
Table 6 presents quantitative metrics for fire description and their corresponding scores. Our fine-tuned models achieved superior results across these categories. Specifically, the 7B model excelled in fire intensity and size, while the 3B model performed best for fire hotspots. Gemini demonstrated the highest performance among off-the-shelf models, with Gemini 2.0 Pro leading by a significant margin.
Table 7 displays the performance metrics for qualitative fire description, including fire state, fire type, and their average percentages. Our fine-tuned models, ForestFireVLM-7B and ForestFireVLM-3B, again demonstrated superior performance in both categories. Notably, ForestFireVLM-7B achieved the highest scores across all metrics, with 64.1% for fire state, 71.8% for fire type, and an average of 67.9%. The smaller variant, ForestFireVLM-3B, also performed well, with 61.5% for fire state, 68.1% for fire type, and an average of 64.8%. Among the off-the-shelf models, Gemini Pro 2.0 led, with scores of 53.5% for fire state, 66.5% for fire type, and an average of 60.0%, closely followed by Gemini Flash 2.0 Lite, with an average of 55.2%.
Table 8 illustrates the performance metrics for various environmental factors, including infrastructure nearby, people nearby, tree vitality, and their average percentages. Our fine-tuned models, ForestFireVLM-7B and ForestFireVLM-3B, exhibited strong performance across these categories. The 7B model achieved the highest overall average of 67.7%, with notable scores of 66.1% for people nearby and 62.8% for tree vitality. The smaller variant, ForestFireVLM-3B, led in infrastructure nearby, with a score of 74.4%, and maintained a competitive average of 64.8%. Among the off-the-shelf models, Gemini Flash 2.0 demonstrated commendable performance, with an average of 62.5%, while Gemini Pro 1.5 followed closely with 56.3%.
Table 9 presents an overview of the comprehensive performance metrics across all tasks described in the tables above, including detection, fire quantitative description, fire qualitative description, environmental factors, and their overall averages. The overall average here was formulated as the average across all output fields.

3.2. Examples

The following images were randomly picked from the evaluation dataset and show the outputs with ForestFireVLM-7B. In Figure 5, an example image from the forest fire in Tennenloher Forst near Erlangen is demonstrated. The human-annotated ground truths were set similarly for most categories, but with setting the Uncontrolled field to “Closer investigation required”, People Nearby to “Cannot be determined”, and the Tree Vitality to “Moderate Vitality”.
An image without any fire is shown in Figure 6, where the ForestFireVLM-7B correctly output “No forest fire visible” for all additional categories. These results matched our human-made annotations.
For the image in Figure 7, once again, most answers matched our annotations, with People Nearby being set to “Yes” and Tree Vitality set to “Moderate Vitality” in the human-made annotations.

3.3. FIgLib Test Dataset

We evaluated the zero-shot smoke detection capabilities on the test set of the Figlib dataset. Despite training with a context length of 4500 tokens, inferencing with a context length of 9000 tokens worked well and improved the model performance significantly compared to the baseline Qwen2.5-VL model. Table 10 shows that our fine-tuned models outperformed the other VLM-based zero-shot approaches.
Lastly, we compare the performance of our fine-tuned VLMs to other approaches from Dewangan et al. [4] and Wei and Kulkarni [19] in Table 11. SmokeyNet with a single frame combines a ResNet34 with a Vision Transformer, while the other variant uses three frames and an additional LSTM (Long Short-Term Memory) network.

3.4. Inference Times

We evaluated inference times using a subset of 50 images of the ForestFireInsights-Eval dataset and a maximum image size of 1M pixels using vLLM [36]. Table 12 compares the inference times of ForestFireVLM-3B, -7B, and -7B quantized to FP8 on different hardware options. These included the NVIDIA Jetson Orin NX onboard computer used in Evolonic drones, and NVIDIA RTX 4060 Ti and RTX 4090 consumer desktop GPUs. The 7B model requires at least 24 GB of memory in native FP16, hence it could not run on the Orin NX and RTX 4060 Ti with 16 GB of memory each. All inference times in the table show the total average inference times, with their corresponding prefill and decode times in brackets.

4. Discussion

4.1. Key Takeaways

We have shown that general-purpose VLMs can classify multiple key attributes of forest fires. Binary classification of smoke or fire visibility was reliable with all VLMs evaluated, while describing the fire and the environment was more challenging. We tested closed, cloud-based VLMs from OpenAI and Google Deepmind, and Qwen2.5-VL open-weight models on our evaluation dataset. Gemini Pro 2.0 scored an average of 65.5% on our benchmark, being the best model without further fine-tuning. We then fine-tuned Qwen2.5-VL models on our private training dataset with 1.2K images, and improved the evaluation to an average of 76.6%. We also validated the smoke detection performance of our approach with the FIgLib-Test dataset, gaining significantly better results for binary smoke classification than previous VLM-based zero-shot methods.

4.2. Model Performance

Most models demonstrated very good performance in binary flame and smoke detection, even without additional fine-tuning. Notably, the 3B version of Qwen2.5-VL outperformed its larger variants in smoke detection by a significant margin.
Our fine-tuned models, ForestFireVLM-7B and ForestFireVLM-3B, clearly outperformed the other models in most categories. The 7B model achieved an impressive overall average of 76.6%, with scores of 90.0% for detection, 77.7% for fire quantitative description, 67.9% for fire qualitative description, and 67.7% for environmental factors. The smaller variant, ForestFireVLM-3B, also demonstrated strong performance, with an overall average of 74.8%. Among the off-the-shelf models, Gemini Pro 2.0 showed commendable results with an overall average of 65.5%, while Gemini Flash 2.0 followed closely with 64.1%. Remarkably, all evaluated families of models were close to their individual peers. The Qwen2.5-VL models exhibited varied performance, with the 7B variant lagging behind its counterparts. Notably, Qwen2.5-VL-3B outperformed its larger variants in some metrics, showing a smaller performance gap compared to the 72B model than to the 7B model.
The examples showed both the good results of our fine-tuned model and the ambiguity of our ground truth labels. The decision between setting a definitive answer or “Cannot be determined” can be especially hard. Human-made annotations are, therefore, not perfect and will vary between different annotators and with different knowledge of the imaged forest fire events.
When looking at the FIgLib-Test dataset, the results were similar to our findings with the ForestFireInsights-Eval dataset. In comparison to the other methods, it is important to state that our approach only evaluated a single image. As the human experts could look at multiple images before deciding, and the best SmokeyNet variant utilized three frames at once, they are not directly comparable to our approach. The Horizon Tiling approach with LLaVA 7B split the image in multiple tiles across a fixed horizon line [19] and is therefore also not directly comparable with our method.
The inference time comparison showed that a comparably powerful edge device like the Jetson Orin NX was significantly slower than consumer GPUs. The reasons for this lie in the overall lower computation performance, significantly lower memory bandwidth, and potentially less optimized inference engines in the ARM-based Jetson ecosystem. Additionally, the tests where carried out with a relatively high image resolution and a correspondingly large number of input tokens.

4.3. Integration

A flowchart of how ForestFireVLM could be integrated with an UAV-based forest fire detection pipeline can be found in Figure 8. While image capturing and smoke detection can be carried out on an onboard computer in real time, the additional inference time of a VLM only requires processing images after a first detection, and on a dedicated GPU server. The Uncontrolled field of the VLM’s output, containing either “Yes”, “No”, or “Closer investigation required”, could then be used to decide if emergency personal should be notified via a web application.

4.4. Limitations and Future Work

One significant limitation lies in the annotation process of the forest fire images. All annotations were carried out by the authors themselves, who are not trained fire brigade professionals. Many annotations remain vague and could benefit from annotators who are more experienced with interpreting forest fires in all stages. The success of our fine-tuned models can be explained by its adaption to the annotator’s preferences in some part. However, our results confirm that VLMs can learn such specific annotations, while future works could improve the factuality by using annotations from actual domain experts. Another limitation regarding the FIgLib and University of Split datasets might be their contamination regarding the VLM pretraining datasets, as the vision encoders used in modern VLMs are trained on almost all publicly available image data. The strong results on Evolonic’s own, previously unpublished, drone footage demonstrate robustness against this. As demonstrated in this work, it is possible to fine-tune VLMs on a broad range of forest-fire-related tasks. With about 1.2K samples, we used a relatively small instruction fine-tuning dataset, which could be scaled up with more labeled data. Our approach is currently limited to a single image per inference, neglecting temporal features of the smoke. Our base VLMs support multi-frame inference, which could enable leveraging this in future work. This would allow the VLM to extract additional information from smoke motion and multiple viewpoints. This would increase the compute cost for both training and inference, adding to the already high latencies on edge devices. While running a 3B parameter 16-bit or 7B parameter quantized VLM is possible with 16 GB of memory, matching the available memory on Evolonic’s NVIDIA Orin NX onboard computers, the inference times are an order of magnitude higher compared to desktop machines. While memory usage might not be a limiting factor anymore, the latency of running onboard VLMs remains a challenge that has to be addressed.

5. Conclusions

We presented a family of state-of-the-art VLMs fine-tuned specifically for forest fire detection and structured description. Our contributions include
  • ForestFireVLM-7B and ForestFireVLM-3B, fine-tuned versions of their Qwen2.5-VL counterparts, publicly available for research and practical applications.
  • A framework for detailed descriptions of forest fires in a structured format.
  • Improving VLM-based detection performance on the FIgLib dataset.
  • A dedicated evaluation dataset for structured forest fire descriptions and accompanying code for future research in this domain.
While real-time inference with VLMs poses challenges for time-critical tasks such as forest fire detection, we posit that providing additional information can significantly aid fire brigades in their decision-making processes. Unlike previous approaches, our system provides actionable intelligence in a structured format, ready for integration into real-world emergency response workflows. By offering these resources, we aim to support further advancements in the application of VLMs for wildfire management and response. Our evaluation and inference code with additional data can be found on GitHub (https://github.com/leon-seidel/ForestFireVLM, accessed on 15 March 2025), while our models and evaluation datasets are hosted on HuggingFace (https://huggingface.co/collections/leon-se/forestfirevlm-67d3429a77d9a5fc6c7ce9f5, accessed on 15 March 2025).

Author Contributions

Conceptualization, L.S.; methodology, L.S., S.G. and T.R.; software, L.S.; validation, L.S. and S.G.; formal analysis, L.S.; investigation, L.S.; resources, L.S. and S.-N.I.; data curation, L.S. and S.G.; writing—original draft preparation, L.S., S.G. and S.-N.I.; writing—review and editing, L.S., S.G., T.R. and S.-N.I.; visualization, L.S. and S.G.; supervision, B.E. and M.M.; project administration, B.E. and M.M.; funding acquisition, B.E. and M.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The evaluation dataset ForestFireInsights-Eval, as well as a modified version of the FIgLib test dataset, are available on HuggingFace (https://huggingface.co/collections/leon-se/forestfirevlm-67d3429a77d9a5fc6c7ce9f5, accessed on 15 March 2025). The training dataset ForestFireInsights-Train has not been made publicly available for legal reasons.

Acknowledgments

The authors would like to express their gratitude to Adrian Sauer for his leadership as Project Lead. Special thanks are extended to Lorenz Einberger and Dominik Schuler for their work on the drone design and building, to Leonhard Kluge for his responsibility in the design and construction of the base station, and to Isabella Hufnagl for her design of the web application. Appreciation is also given to Oliver Grau for his expertise in electronics engineering, and to Simon Grau for improving our smoke detection and dataset. Thanks are also due to Felix Körwer for strategic initiatives, and to Lara Schindhelm and Sebastian Wiederhold for recording forest fire videos in Erlangen, along with all other Evolonic members who contributed to the project. Thanks also go to the Office for Fire and Civil Protection of the city of Erlangen and all fire brigade members for their collaboration.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LLMLarge Language Model
LoRALow-Rank Adaption
LSTMLong Short-Term Memory
OCROptical Character Recognition
UAVUnmanned Aerial Vehicle
VLMVision Language Model
VTOLVertical Takeoff And Landing
WSNWireless Sensor Network

Appendix A

The following figures contain the distributions of human-made annotations for their respective categories across the three datasets considered in this work. As the FIgLib-Test dataset is only used for verification of the smoke detection we only require annotations in the “Smoke” field here.
Figure A1. Annotation distribution for “Smoke” in all datasets.
Figure A1. Annotation distribution for “Smoke” in all datasets.
Drones 09 00347 g0a1
Figure A2. Annotation distribution for “Flames” in all datasets.
Figure A2. Annotation distribution for “Flames” in all datasets.
Drones 09 00347 g0a2
Figure A3. Annotation distribution for “Uncontrolled” in all datasets.
Figure A3. Annotation distribution for “Uncontrolled” in all datasets.
Drones 09 00347 g0a3
Figure A4. Annotation distribution for “Fire State” in all datasets.
Figure A4. Annotation distribution for “Fire State” in all datasets.
Drones 09 00347 g0a4
Figure A5. Annotation distribution for “Fire Type” in all datasets.
Figure A5. Annotation distribution for “Fire Type” in all datasets.
Drones 09 00347 g0a5
Figure A6. Annotation distribution for “Fire Intensity” in all datasets.
Figure A6. Annotation distribution for “Fire Intensity” in all datasets.
Drones 09 00347 g0a6
Figure A7. Annotation distribution for “Fire Size” in all datasets.
Figure A7. Annotation distribution for “Fire Size” in all datasets.
Drones 09 00347 g0a7
Figure A8. Annotation distribution for “Fire Hotspots” in all datasets.
Figure A8. Annotation distribution for “Fire Hotspots” in all datasets.
Drones 09 00347 g0a8
Figure A9. Annotation distribution for “Infrastructure Nearby” in all datasets.
Figure A9. Annotation distribution for “Infrastructure Nearby” in all datasets.
Drones 09 00347 g0a9
Figure A10. Annotation distribution for “People Nearby” in all datasets.
Figure A10. Annotation distribution for “People Nearby” in all datasets.
Drones 09 00347 g0a10
Figure A11. Annotation distribution for “Tree Vitality” in all datasets.
Figure A11. Annotation distribution for “Tree Vitality” in all datasets.
Drones 09 00347 g0a11

References

  1. Jones, M.W.; Kelley, D.I.; Burton, C.A.; Di Giuseppe, F.; Barbosa, M.L.F.; Brambleby, E.; Hartley, A.J.; Lombardi, A.; Mataveli, G.; McNorton, J.R.; et al. State of Wildfires 2023–2024. Earth Syst. Sci. Data 2024, 16, 3601–3685. [Google Scholar] [CrossRef]
  2. Kalabokidis, K.; Xanthopoulos, G.; Moore, P.; Caballero, D.; Kallos, G.; Llorens, J.; Roussou, O.; Vasilakos, C. Decision support system for forest fire protection in the Euro-Mediterranean region. Eur. J. For. Res. 2012, 131, 597–608. [Google Scholar] [CrossRef]
  3. Bouguettaya, A.; Zarzour, H.; Taberkit, A.M.; Kechida, A. A review on early wildfire detection from unmanned aerial vehicles using deep learning-based computer vision algorithms. Signal Process. 2022, 190, 108309. [Google Scholar] [CrossRef]
  4. Dewangan, A.; Pande, Y.; Braun, H.W.; Vernon, F.; Perez, I.; Altintas, I.; Cottrell, G.W.; Nguyen, M.H. FIgLib & SmokeyNet: Dataset and Deep Learning Model for Real-Time Wildland Fire Smoke Detection. Remote Sens. 2022, 14, 1007. [Google Scholar] [CrossRef]
  5. OpenAI; Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; et al. GPT-4 Technical Report. arXiv 2023, arXiv:2303.08774. [Google Scholar]
  6. Bai, S.; Chen, K.; Liu, X.; Wang, J.; Ge, W.; Song, S.; Dang, K.; Wang, P.; Wang, S.; Tang, J.; et al. Qwen2.5-VL Technical Report. arXiv 2025, arXiv:2502.13923. [Google Scholar]
  7. Li, Z.; Wu, X.; Du, H.; Nghiem, H.; Shi, G. Benchmark Evaluations, Applications, and Challenges of Large Vision Language Models: A Survey. arXiv 2025, arXiv:2501.02189. [Google Scholar]
  8. Tao, L.; Zhang, H.; Jing, H.; Liu, Y.; Yan, D.; Wei, G.; Xue, X. Advancements in Vision–Language Models for Remote Sensing: Datasets, Capabilities, and Enhancement Techniques. Remote Sens. 2025, 17, 162. [Google Scholar] [CrossRef]
  9. Li, X.; Wen, C.; Hu, Y.; Yuan, Z.; Zhu, X.X. Vision-Language Models in Remote Sensing: Current progress and future trends. IEEE Geosci. Remote Sens. Mag. 2024, 12, 32–66. [Google Scholar] [CrossRef]
  10. Tielin, M.; Chuanguang, Y.; Wenbiao, G.; Zihan, X.; Qinling, Z.; Xiaoou, Z. Analysis of Technical Characteristics of Fixed-Wing VTOL UAV. In Proceedings of the 2017 IEEE International Conference on Unmanned Systems (ICUS), Beijing, China, 27–29 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 293–297. [Google Scholar] [CrossRef]
  11. Schneider, D. Waldbrandfrüherkennung; W. Kohlhammer GmbH: Stuttgart, Germany, 2021. [Google Scholar] [CrossRef]
  12. Hsieh, R. Alberta Wildfire Detection Challenge: Operational Demonstration of Six Wildfire Detection Systems; Technical Report; TR 2023 n.1; FPInnovations: Thunder Bay, ON, Canada, 2023. [Google Scholar]
  13. Alkhatib, A.A.A. A Review on Forest Fire Detection Techniques. Int. J. Distrib. Sens. Netw. 2014, 10, 597368. [Google Scholar] [CrossRef]
  14. Mohapatra, A.; Trinh, T. Early Wildfire Detection Technologies in Practice—A Review. Sustainability 2022, 14, 12270. [Google Scholar] [CrossRef]
  15. Göttlein, A.; Laniewski, R.; Brinkschulte, C.; Schwichtenberg, H. Praxistest eines Waldbrand-Frühwarnsystems. AFZ der Wald, 31 May 2023. [Google Scholar]
  16. Blais, M.A.; Akhloufi, M.A. Drone Swarm Coordination Using Reinforcement Learning for Efficient Wildfires Fighting. SN Comput. Sci. 2024, 5, 314. [Google Scholar] [CrossRef]
  17. Sautenkov, O.; Yaqoot, Y.; Lykov, A.; Mustafa, M.A.; Tadevosyan, G.; Akhmetkazy, A.; Cabrera, M.A.; Martynov, M.; Karaf, S.; Tsetserukou, D. UAV-VLA: Vision-Language-Action System for Large Scale Aerial Mission Generation. arXiv 2025, arXiv:2501.05014. [Google Scholar]
  18. Sakaino, H. Dynamic Texts from UAV Perspective Natural Images. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Paris, France, 1–6 October 2023; pp. 2070–2081. [Google Scholar]
  19. Wei, T.; Kulkarni, P. Enhancing the Binary Classification of Wildfire Smoke Through Vision-Language Models. In Proceedings of the 2024 Conference on AI, Science, Engineering, and Technology (AIxSET), Laguna Hills, CA, USA, 30 September–2 October 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 115–118. [Google Scholar] [CrossRef]
  20. Pronto, L.; Held, A. Einführung in das Feuerverhalten. 2021. Available online: https://www.waldbrand-klima-resilienz.com/_files/ugd/bf6978_f7fbed679d1e438fb88025e159f21ff4.pdf (accessed on 15 March 2025).
  21. Saxena, S.; Dubey, R.; Yaghoobian, N. A Model for Predicting Ignition Potential of Complex Fuel in. Ph.D. Thesis, Florida State University, Tallahassee, FL, USA, 2023. [Google Scholar] [CrossRef]
  22. Cimolino, U. Analyse der Einsatzerfahrungen und Entwicklung von Optimierungsmöglichkeiten bei der Bekämpfung von Vegetationsbränden in Deutschland. Ph.D. Thesis, Universität Wuppertal, Wuppertal, Germany, 2014. [Google Scholar]
  23. Patzelt, S.T. Waldbrandprognose und Waldbrandbekämpfung in Deutschland—Zukunftsorientierte Strategien und Konzepte unter Besonderer Berücksichtigung der Brandbekämpfung aus der Luft. Ph.D. Thesis, Johannes Gutenberg Universität, Mainz, Germany, 2008. [Google Scholar] [CrossRef]
  24. Deutscher Feuerwehrverband. Anzahl der Feuerwehren; Deutscher Feuerwehrverband: Berlin, Germany, 2023. [Google Scholar]
  25. Laurençon, H.; Tronchon, L.; Cord, M.; Sanh, V. What matters when building vision-language models? In Proceedings of the 38th Conference on Neural Information Processing Systems (NeurIPS 2024), Vancouver, BC, Canada, 10–15 December 2024. [Google Scholar]
  26. Duan, H.; Fang, X.; Yang, J.; Zhao, X.; Qiao, Y.; Li, M.; Agarwal, A.; Chen, Z.; Chen, L.; Liu, Y.; et al. VLMEvalKit: An Open-Source Toolkit for Evaluating Large Multi-Modality Models. In Proceedings of the 32nd ACM International Conference on Multimedia, Melbourne, Australia, 28 October–1 November 2024. [Google Scholar]
  27. Zheng, Y.; Zhang, R.; Zhang, J.; Ye, Y.; Luo, Z.; Feng, Z.; Ma, Y. LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models. arXiv 2024, arXiv:2403.13372. [Google Scholar]
  28. Han, D.; Han, M.; Unsloth Team. Unsloth. 2023. Available online: http://github.com/unslothai/unsloth (accessed on 15 March 2025).
  29. Chen, Z.; Wang, W.; Cao, Y.; Liu, Y.; Gao, Z.; Cui, E.; Zhu, J.; Ye, S.; Tian, H.; Liu, Z.; et al. Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scaling. arXiv 2024, arXiv:2412.05271. [Google Scholar]
  30. Yue, X.; Ni, Y.; Zhang, K.; Zheng, T.; Liu, R.; Zhang, G.; Stevens, S.; Jiang, D.; Ren, W.; Sun, Y.; et al. MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024. [Google Scholar]
  31. Zhang, Y.F.; Zhang, H.; Tian, H.; Fu, C.; Zhang, S.; Wu, J.; Li, F.; Wang, K.; Wen, Q.; Zhang, Z.; et al. MME-RealWorld: Could Your Multimodal LLM Challenge High-Resolution Real-World Scenarios that are Difficult for Humans? arXiv 2024, arXiv:2408.13257. [Google Scholar]
  32. Center for Wildfire Research, University of Split. Available online: http://wildfire.fesb.hr/ (accessed on 13 March 2005).
  33. Krstinić, D.; Stipaničev, D.; Jakovčević, T. Histogram-based smoke segmentation in forest fire detection system. Inf. Technol. Control 2009, 38, 237–244. [Google Scholar]
  34. Hu, E.J.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Wang, L.; Chen, W. LoRA: Low-Rank Adaptation of Large Language Models. ICLR 2022, 1, 3. [Google Scholar]
  35. Dettmers, T.; Pagnoni, A.; Holtzman, A.; Zettlemoyer, L. QLoRA: Efficient Finetuning of Quantized LLMs. Adv. Neural Inf. Process. Syst. 2023, 36, 10088–10115. [Google Scholar]
  36. Kwon, W.; Li, Z.; Zhuang, S.; Sheng, Y.; Zheng, L.; Yu, C.H.; Gonzalez, J.E.; Zhang, H.; Stoica, I. Efficient Memory Management for Large Language Model Serving with PagedAttention. In Proceedings of the 29th Symposium on Operating Systems Principles, Koblenz, Germany, 23–26 October 2023. [Google Scholar]
  37. Wang, P.; Bai, S.; Tan, S.; Wang, S.; Fan, Z.; Bai, J.; Chen, K.; Liu, X.; Wang, J.; Ge, W.; et al. Qwen2-VL: Enhancing Vision-Language Model’s Perception of the World at Any Resolution. arXiv 2024, arXiv:2409.12191. [Google Scholar]
Figure 1. Rendering of the eVTOL drone that was used for this project with the base station.
Figure 1. Rendering of the eVTOL drone that was used for this project with the base station.
Drones 09 00347 g001
Figure 2. Isometric illustrations of (a) the drone detecting a wildfire, and (b) the drone monitoring the situation and providing an overview of the situation for fire brigades.
Figure 2. Isometric illustrations of (a) the drone detecting a wildfire, and (b) the drone monitoring the situation and providing an overview of the situation for fire brigades.
Drones 09 00347 g002
Figure 3. Sample image from ForestFireInsights-Eval: Real forest fire near Erlangen, captured by Evolonic using a DJI Mavic drone.
Figure 3. Sample image from ForestFireInsights-Eval: Real forest fire near Erlangen, captured by Evolonic using a DJI Mavic drone.
Drones 09 00347 g003
Figure 4. Smoothed training losses for the Qwen2.5-VL fine-tuned models trained on our dataset.
Figure 4. Smoothed training losses for the Qwen2.5-VL fine-tuned models trained on our dataset.
Drones 09 00347 g004
Figure 5. Answers from ForestFireVLM-7B for an image of a real forest fire in 2022.
Figure 5. Answers from ForestFireVLM-7B for an image of a real forest fire in 2022.
Drones 09 00347 g005
Figure 6. ForestFireVLM-7B correctly output “No forest fire visible” for all categories when no forest fire was in the image.
Figure 6. ForestFireVLM-7B correctly output “No forest fire visible” for all categories when no forest fire was in the image.
Drones 09 00347 g006
Figure 7. ForestFireVLM-7B captions for a frame from a test with the fire brigade Erlangen in 2022.
Figure 7. ForestFireVLM-7B captions for a frame from a test with the fire brigade Erlangen in 2022.
Drones 09 00347 g007
Figure 8. Flowchart of the proposed integration of our ForestFireVLM in a UAV-based smoke detection pipeline.
Figure 8. Flowchart of the proposed integration of our ForestFireVLM in a UAV-based smoke detection pipeline.
Drones 09 00347 g008
Table 1. Fields, questions, and answer options for the structured output generation.
Table 1. Fields, questions, and answer options for the structured output generation.
FieldQuestionOptions
SmokeIs smoke from a forest fire visible in the image?Yes, No
FlamesAre flames from a forest fire visible in the image?Yes, No
UncontrolledCan you confirm that this is an uncontrolled forest fire?Yes, Closer investigation required, No forest fire visible
Fire StateWhat is the current state of the forest fire?Ignition Phase, Growth Phase, Fully Developed Phase, Decay Phase, Cannot be determined, No forest fire visible
Fire TypeWhat type of fire is it?Ground Fire, Surface Fire, Crown Fire, Cannot be determined, No forest fire visible
Fire IntensityWhat is the intensity of the fire?Low, Moderate, High, Cannot be determined, No forest fire visible
Fire SizeWhat is the size of the fire?Small, Medium, Large, Cannot be determined, No forest fire visible
Fire HotspotsDoes the forest fire have multiple hotspots?Multiple hotspots, One hotspot, Cannot be determined, No forest fire visible
Infrastructure NearbyIs there infrastructure visible near the forest fire?Yes, No, Cannot be determined, No forest fire visible
People NearbyAre there people visible near the forest fire?Yes, No, Cannot be determined, No forest fire visible
Tree VitalityDescribe the vitality of the trees around the fire.Vital, Moderate Vitality, Declining, Dead, Cannot be determined, No forest fire visible
Table 2. Comparison of the datasets used in this work.
Table 2. Comparison of the datasets used in this work.
DatasetPurposeEnvironment Description
ForestFireInsights-TrainTraining (all fields)Forest fires across Germany and worldwide (UAV and airborne perspectives)
ForestFireInsights-EvalEvaluation (all fields)Forest fires in Germany and Croatia (UAV and airborne perspectives)
FIgLib-TestEvaluation (detections)Forest fires in California (tower-based perspectives)
Table 3. Hyperparameters and GPUs chosen for our models using the LLaMA-Factory framework.
Table 3. Hyperparameters and GPUs chosen for our models using the LLaMA-Factory framework.
SettingForestFireVLM-3BForestFireVLM-7B
Learning rate0.000050.0002
Epochs22
Batch Size11
Gradient Accumulation8 Steps16 Steps
GPUNVIDIA RTX 3090NVIDIA A100 80 GB
Table 4. Model checkpoints or versions from models used with Google’s or OpenAI’s API endpoints.
Table 4. Model checkpoints or versions from models used with Google’s or OpenAI’s API endpoints.
ModelVersion or Checkpoint
Gemini Pro 1.5Version 002
Gemini Flash 2.0Version 001
Gemini Flash 2.0 LiteVersion 001
Gemini Pro 2.0Experimental 2025-02-05
GPT-4o2024-08-06
GPT-4o mini2024-07-18
Table 5. Performance for smoke and fire detection, best results are bold.
Table 5. Performance for smoke and fire detection, best results are bold.
ModelFlames (%)Smoke (%)Uncontrolled (%)Average (%)
ForestFireVLM-7B98.095.077.190.0
ForestFireVLM-3B96.095.475.889.0
Gemini Pro 1.594.791.465.183.7
Gemini Flash 2.096.495.075.488.9
Gemini Flash 2.0 Lite96.793.758.883.1
Gemini Pro 2.090.495.753.279.7
GPT-4o95.794.747.579.3
GPT-4o mini95.094.743.977.9
Qwen2.5-VL-3B93.492.039.274.9
Qwen2.5-VL-7B92.779.732.268.2
Qwen2.5-VL-72B93.783.434.670.5
Table 6. Performance for quantitative fire metrics, best results are bold.
Table 6. Performance for quantitative fire metrics, best results are bold.
ModelFire Hotspots (%)Fire Intensity (%)Fire Size (%)Average (%)
ForestFireVLM-7B78.174.181.177.7
ForestFireVLM-3B82.169.880.477.4
Gemini Pro 1.569.148.551.556.4
Gemini Flash 2.058.834.251.248.1
Gemini Flash 2.0 Lite79.447.267.464.7
Gemini Pro 2.076.763.174.171.3
GPT-4o26.321.921.623.3
GPT-4o mini35.626.627.930.0
Qwen2.5-VL-3B55.837.938.244.0
Qwen2.5-VL-7B9.04.33.75.6
Qwen2.5-VL-72B28.621.921.624.0
Table 7. Performance for the qualitative fire description, best results are bold.
Table 7. Performance for the qualitative fire description, best results are bold.
ModelFire State (%)Fire Type (%)Average (%)
ForestFireVLM-7B64.171.867.9
ForestFireVLM-3B61.568.164.8
Gemini Pro 1.541.563.552.5
Gemini Flash 2.044.562.153.3
Gemini Flash 2.0 Lite46.563.855.2
Gemini Pro 2.053.566.560.0
GPT-4o29.952.241.0
GPT-4o mini31.955.843.9
Qwen2.5-VL-3B36.938.237.5
Qwen2.5-VL-7B12.034.623.3
Qwen2.5-VL-72B28.947.838.4
Table 8. Performance in environmental fields, best results are bold.
Table 8. Performance in environmental fields, best results are bold.
ModelInfrastructure Nearby (%)People Nearby (%)Tree Vitality (%)Average (%)
ForestFireVLM-7B74.166.162.867.7
ForestFireVLM-3B74.458.561.564.8
Gemini Pro 1.566.557.544.956.3
Gemini Flash 2.070.156.560.862.5
Gemini Flash 2.0 Lite51.237.530.639.8
Gemini Pro 2.060.143.243.548.9
GPT-4o31.952.839.941.5
GPT-4o mini33.247.541.940.9
Qwen2.5-VL-3B56.232.632.240.3
Qwen2.5-VL-7B51.244.920.939.0
Qwen2.5-VL-72B62.544.249.251.9
Table 9. Total model performances, best results are bold.
Table 9. Total model performances, best results are bold.
ModelDetection (%)Fire Quantitative (%)Fire Qualitative (%)Environmental (%)Overall (%)
ForestFireVLM-7B90.077.767.967.776.6
ForestFireVLM-3B89.077.464.864.874.8
Gemini Pro 1.583.756.452.556.363.1
Gemini Flash 2.088.948.153.362.564.1
Gemini Flash 2.0 Lite83.164.755.239.861.2
Gemini Pro 2.079.771.360.048.965.5
GPT-4o79.323.341.041.546.8
GPT-4o mini77.930.043.940.948.5
Qwen2.5-VL-3B74.944.037.540.350.2
Qwen2.5-VL-7B68.25.623.339.035.0
Qwen2.5-VL-72B70.524.038.451.946.9
Table 10. VLM-based zero-shot smoke detection results on the FigLib dataset, best results are bold.
Table 10. VLM-based zero-shot smoke detection results on the FigLib dataset, best results are bold.
ModelAccuracy (%)Precision (%)Recall (%) F 1  Score (%)
ForestFireVLM-7B78.598.657.272.4
ForestFireVLM-3B76.398.852.668.6
Gemini 1.5 Pro70.0100.039.256.3
Gemini 2.0 Flash74.196.949.165.2
Gemini 2.0 Flash Lite71.595.644.260.5
Qwen2.5-VL-3B70.491.344.159.5
Qwen2.5-VL-7B60.3100.019.532.7
PaliGemma [19]52.1100.03.05.7
Phi3 [19]52.6100.04.07.6
GPT-4o [19]74.595.250.666.1
LLaVA 7B [19]67.587.639.254.1
Table 11. Comparison to other methods on the FIgLib dataset, best results are bold.
Table 11. Comparison to other methods on the FIgLib dataset, best results are bold.
ModelAccuracy (%)Precision (%)Recall (%) F 1  Score (%)
ForestFireVLM-7B78.598.657.272.4
ForestFireVLM-3B76.398.852.668.6
Human (average of 3) [4]78.593.574.482.8
SmokeyNet (1 frame) [4]82.588.675.281.3
SmokeyNet (3 frames) [4]83.690.976.182.8
LLaVA (Horizon Tiling) [19]81.486.573.779.6
Table 12. Comparison of inference times using different GPUs.
Table 12. Comparison of inference times using different GPUs.
GPUForestFireVLM-3B (FP16)ForestFireVLM-7B (FP8)ForestFireVLM-7B (FP16)
Jetson Orin NX23.0 s (5.0/18.1 s)34.1 s (7.2/26.9 s)-
RTX 4060 Ti3.2 s (0.5/2.7 s)4.1 s (0.6/3.5 s)-
RTX 40901.1 s (0.1/1.0 s)1.5 s (0.2/1.3 s)2.0 s (0.2/1.8 s)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Seidel, L.; Gehringer, S.; Raczok, T.; Ivens, S.-N.; Eckardt, B.; Maerz, M. Advancing Early Wildfire Detection: Integration of Vision Language Models with Unmanned Aerial Vehicle Remote Sensing for Enhanced Situational Awareness. Drones 2025, 9, 347. https://doi.org/10.3390/drones9050347

AMA Style

Seidel L, Gehringer S, Raczok T, Ivens S-N, Eckardt B, Maerz M. Advancing Early Wildfire Detection: Integration of Vision Language Models with Unmanned Aerial Vehicle Remote Sensing for Enhanced Situational Awareness. Drones. 2025; 9(5):347. https://doi.org/10.3390/drones9050347

Chicago/Turabian Style

Seidel, Leon, Simon Gehringer, Tobias Raczok, Sven-Nicolas Ivens, Bernd Eckardt, and Martin Maerz. 2025. "Advancing Early Wildfire Detection: Integration of Vision Language Models with Unmanned Aerial Vehicle Remote Sensing for Enhanced Situational Awareness" Drones 9, no. 5: 347. https://doi.org/10.3390/drones9050347

APA Style

Seidel, L., Gehringer, S., Raczok, T., Ivens, S.-N., Eckardt, B., & Maerz, M. (2025). Advancing Early Wildfire Detection: Integration of Vision Language Models with Unmanned Aerial Vehicle Remote Sensing for Enhanced Situational Awareness. Drones, 9(5), 347. https://doi.org/10.3390/drones9050347

Article Metrics

Back to TopTop