Next Article in Journal
Decoding Non-Invasive Electroencephalography Signal via a Two-Discriminator Adversarial Network
Previous Article in Journal
Pulse Wave Velocity Estimation in a Controlled In Vitro Vascular Model: Benchmarking Machine Learning Approaches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Context-Aware Multi-Agent Architecture for Wildfire Insights

by
Ashen Sandeep
1,
Sithum Jayarathna
1,
Sunera Sandaruwan
1,
Venura Samarappuli
1,
Dulani Meedeniya
1 and
Charith Perera
2,*
1
Department of Computer Science and Engineering, University of Moratuwa, Moratuwa 10400, Sri Lanka
2
School of Computer Science and Informatics, Cardiff University, Cardiff CF24 3AA, UK
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(3), 1070; https://doi.org/10.3390/s26031070
Submission received: 29 December 2025 / Revised: 2 February 2026 / Accepted: 4 February 2026 / Published: 6 February 2026
(This article belongs to the Section Internet of Things)

Abstract

Wildfires are environmental hazards with severe ecological, social, and economic impacts. Wildfires devastate ecosystems, communities, and economies worldwide, with rising frequency and intensity driven by climate change, human activity, and environmental shifts. Analyzing wildfire insights such as detection, predictive patterns, and risk assessment enables proactive response and long-term prevention. However, most of the existing approaches have been focused on isolated processing of data, making it challenging to orchestrate cross-modal reasoning and transparency. This study proposed a novel orchestrator-based multi-agent system (MAS), with the aim of transforming multimodal environmental data into actionable intelligence for decision making. We designed a framework to utilize Large Multimodal Models (LMMs) augmented by structured prompt engineering and specialized Retrieval-Augmented Generation (RAG) pipelines to enable transparent and context-aware reasoning, providing a cutting-edge Visual Question Answering (VQA) system. It ingests diverse inputs like satellite imagery, sensor readings, weather data, and ground footage and then answers user queries. Validated by several public datasets, the system achieved a precision of 0.797 and an F1-score of 0.736. Thus, powered by Agentic AI, the proposed, human-centric solution for wildfire management, empowers firefighters, governments, and researchers to mitigate threats effectively.

1. Introduction

Wildfires have been a natural phenomenon over the years [1]. However, their frequency, intensity, and spatial extent have increased significantly in recent decades due to climate change and prolonged droughts [2]. Wildfires lead to natural disasters, such as landslides and flooding. causing environmental damage, public safety concerns, and economic damage [3]. In particular, large-scale land degradation and infrastructure loss cause significant financial losses. The exposure to delicate particulate matter from wildfire smoke can have immediate effects. Moreover, biomass combustion driven by wildfires contributes significantly to atmospheric carbon emissions [4,5]. As a result, this pollution reinforces climate change through feedback mechanisms [2]. These trends highlight the urgent need for more effective and adaptive wildfire monitoring and response strategies.
Traditional wildfire systems relied on physical models and rule-based systems [6,7,8]. As a result of their static nature, traditional wildfire systems faced limitations in handling complexity and performance. These systems typically rely on empirical and statistical analyses of fire weather indices, which have limited predictive power. With the limitations of these systems, traditional methods are rapidly being replaced by remote sensing technologies, Machine Learning (ML)-based fire detection algorithms, and advanced fire weather indices, resulting in significantly improved predictive accuracy [9,10].
With the transition to ML-based systems, Deep Learning (DL) plays a leading role. Studies have introduced approaches such as satellite image analysis with CNNs, integrating sensor systems with DL, and developing comprehensive systems that combine multimodal data [11,12,13,14]. Despite these advances, existing approaches exhibit several unresolved limitations. (1) Most systems process visual, meteorological, or textual data in isolation [15]. This leaves the problem of cross-modal integration largely unaddressed. (2) Many DL models function as opaque predictors, offering limited interpretability. This leads to a hindrance of trust and adoption in high-stakes decision-making contexts. (3) Prior MAS frameworks often assume stable communication infrastructures. However, the assumption does not hold in remote wildfire zones characterized by intermittent connectivity. (4) Current approaches rarely support human-in-the-loop reasoning. Nevertheless, it limits domain experts’ ability to interrogate, validate, or contextualize algorithmic recommendations.
While ML and DL have dominated artificial intelligence paradigms, emerging trends now leverage MAS integrated with LMM to enable collaborative, scalable, and context-aware decision-making in complex environments [16,17]. LMMs, alongside RAG, play a crucial role by augmenting reasoning capabilities through integration with external knowledge sources [18,19]. However, existing implementations typically operate under static and query-driven paradigms; thus, there remains a lack of mechanisms for dynamically orchestrating heterogeneous data streams in real time [6,20]. Consequently, their applicability to time-sensitive wildfire response scenarios remains limited. Notably, the proposed model must jointly interpret multimodal inputs, including aerial imagery, satellite data, and environmental sensor readings, under uncertain conditions.
The main objective of this study is to propose an orchestrated multi-agent framework with multimodal RAG pipelines to enable dynamic data integration, context-aware reasoning, and human-centered decision support for wildfire management. The proposed approach emphasizes coordinated reasoning across agents and modalities to support situational awareness in operational settings. Accordingly, this study addresses the following research questions:
(RQ1)
How can an orchestrator-based MAS dynamically integrate heterogeneous data sources (such as UAV imagery, satellite observations, and tabular environmental data) into a unified reasoning framework for wildfire response?
(RQ2)
What advantages does a multimodal MAS integrated with RAG pipelines offer over traditional wildfire prediction and management approaches?
Consequently, this study makes the following key contributions to the domain of wildfire management.

Statement of Novelty and Contributions

While recent works like WildfireGPT [15] and SmokeyNet [14] have applied AI to wildfire management, our work introduces three distinct architectural innovations that differentiate it from existing RAG-based and multi-agent paradigms:
  • Architectural Determinism via Decay-Weighted Routing: Unlike standard autonomous agents that rely on open-ended loops (which are prone to “getting stuck”), we introduce a novel orchestration policy π ( q , C t ) governed by a decay factor. This mathematically enforces task convergence, a critical requirement for safety-critical response systems.
  • Lossless Multimodal RAG: We address the information loss inherent in standard “caption-based” retrieval systems. By retrieving and processing raw visual artifacts (Base64) rather than text descriptions, our pipeline preserves the forensic granularity required to distinguish between smoke plumes and cloud cover.
  • Formalized State Engineering: We replace standard natural language prompting with a rigorous “Context Tuple” framework (p). This formalization constrains the stochastic nature of Large Multimodal Models (LMMs), ensuring that agent behavior is reproducible and auditable, a feature largely absent in generic generative AI frameworks.
By advancing multimodal, context-aware and human-centered decision-support mechanisms, this work contributes to the development of more resilient wildfire response systems. It aligns with the united nations sustainable development goals on sustainable cities (SDG11), climate action (SDG 13), and life on land (SDG 15) by advancing technologies for disaster resilience and ecosystem preservation [21].

2. Related Work

Early research in wildfire management primarily relied on physics-based and rule-based models to simulate fire behavior and estimate fire danger. These approaches typically model wildfire spread as a wave-propagation process, governed by meteorological variables such as wind speed, temperature, and fuel characteristics [22]. For instance, Tavakol et al. [6] demonstrated the use of UAV swarms that employ simple rule-based heuristics for autonomous wildfire suppression. From another point of view, several studies have focused on handling environmental uncertainty and improving response times through fuzzy logic and edge computing. Duarte et al. [7] developed a Fuzzy Inference System (FIS) to map forest fire susceptibility in the Amazon region, utilizing fuzzy rules to effectively model the ambiguity inherent in meteorological and land-use data. Similarly, Toledo-Castro et al. [8] proposed a dynamic fuzzy logic controller embedded within wireless sensor networks, enabling distributed nodes to autonomously validate fire outbreaks and reduce false alarms without relying on a central server. While such models offer valuable theoretical insights into wildfire dynamics, they exhibit certain limitations such as (1) an inability to consistently model fire behaviour across different spatial and temporal scales, (2) high data requirements and computational expense, and (3) challenges in validating model predictions against real-world fire phenomena. In real-world wildfire scenarios, where data are incomplete, delayed, or noisy, these requirements are rarely met. Consequently, physics-based models offer limited support for real-time decision-making during rapidly evolving fire events [23,24].
To overcome the limitations of traditional modeling approaches, recent studies have increasingly adopted data-driven methods based on DL [11,12,25,26]. A prominent example of this is “SmokeyNet” [14], a DL-based model that integrates Convolutional Neural Networks (CNNs), Long Short-Term Memory Networks (LSTMs), and Vision Transformers (ViT) to identify smoke patterns from multiple sources, including the HP- WREN camera network in the USA. Although such studies have shown high performance in specific tasks like smoke detection, they lack the cognitive capabilities to forensically identify the significance. Another study, WildfireGPT [15], primarily handles textual queries with climate projections and knowledge retrieval, lacking dynamic orchestration to fuse multimodal inputs like UAV imagery with sensor data in evolving scenarios, resulting in delayed contextual insights during time-sensitive events.
Another important consideration is the explainability in environmental data screening with remote sensing, which is crucial for building trust in AI-driven decisions that impact disaster management, climate monitoring, and ecosystem preservation [27,28,29]. The study by Klotz et al. [29] addressed the effectiveness of explainable AI (XAI) methods and evaluation metrics for remote sensing image scene classification. The study provided methodological insights and experimental validation, offering guidelines for selecting appropriate XAI methods, metrics, and hyperparameters tailored to remote sensing contexts. Similarly, Ahangama et al. [28] proposed XAI framework for precipitation nowcasting using spatio-temporal and multivariate image data to predict short-term rainfall patterns with pixel-level accuracy. The generated attribution maps have revealed the model’s focus on critical features like echo intensity and motion patterns across image sequences, enhancing trust and interpretability for meteorological decision-making in disaster-prone regions.
Recently, the emergence of Agentic AI and Generative AI has introduced new paradigms for autonomous and interactive disaster management [30,31,32]. MAS have been widely explored for decentralized coordination. For example, Zadeh et al. [20] integrated Deep Reinforcement Learning (DRL) into a high-level MAS to optimize fire tracking, while Mawanza [18] proposed a heterogeneous MAS combining UAVs and ground robots for comprehensive monitoring. Additionally, Kouzehgar et al. [19] explored Multi-Agent Reinforcement Learning (MARL) for swarm-based ocean monitoring, a technique with high transferability to wildfire detection scenarios.
Moreover, studies have addressed edge-based solutions as well. For instance, addressing the connectivity and bandwidth limitations of remote monitoring, Kalatzis et al. [33], introduced a hierarchical edge computing framework for UAVs. By processing visual and infrared data directly on ‘Edge’ nodes rather than transmitting raw streams to a cloud server, their system significantly reduces the latency typically associated with centralized processing.
While frameworks like WildfireGPT [15] utilize LLMs for textual queries, they rely on static retrieval workflows that lack dynamic cross-modal synthesis. Furthermore, standard multi-agent orchestration paradigms often assume stable connectivity and prioritize open-ended exploration. In contrast, our architecture introduces a ‘Strategic Real-Time’ constraint. We deviate from standard agent frameworks by implementing strict logical guardrails and a weighted intent-match formula that prioritizes safety and interpretability over the creative autonomy typically found in Generative AI agents.
To provide an unambiguous distinction between the proposed system and existing state-of-the-art frameworks (such as WildfireGPT and standard ReAct agents), we present a comparative analysis in Table 1. This comparison highlights the specific architectural deviations, such as the shift from recursive looping to decay-weighted routing and from caption-based retrieval to lossless artifact injection, that are necessary to ensure the reproducibility and precision required for wildfire insights.
Table 2 and Table 3 provide an overview of existing studies that have utilized different approaches over the years, in related domains.
As shown in Table 3, several limitations persist in existing wildfire management support studies. The primary gap is the lack of cross-modal integration, as most models handle visual or meteorological data independently. Another issue is that existing solutions provide limited contextual decision support, where DL models predict fire spread, but the interpretability or situational guidance for responders is not sufficient. Additionally, scalability and communication bottlenecks are challenges in current solutions. Several MAS architectures assume perfect connectivity, which is impractical in remote fire zones with unstable networks. Furthermore, there is minimal human-in-the-loop reasoning, failing to provide firefighters and planners with the transparent, conversational systems needed to justify and explain recommendations. Accordingly, the comparison of wildfire screening approaches reveal that numerous studies have employed traditional remote sensing techniques alongside ML and DL; however, the integration of MAS with RAG and LMMs remains underexplored and represents an emerging research frontier. In order to overcome these gaps, the proposed study enables agents to collaboratively interpret multimodal data, reasoning through natural language, and deliver context-aware decision support for wildfire management.

3. Methodology

3.1. Process Overview

The proposed framework employs a hierarchical MAS that decouples task orchestration from execution, as depicted in the system architecture shown in Figure 1.
A central orchestrator agent dynamically manages the full lifecycle of a user query, from ingestion to response generation. This design overcomes the constraints of linear pipelines by supporting conditional branching and adaptive tool selection. The process initiates when a multimodal query that combining natural language text with images such as satellite or UAV imagery is submitted via the VQA interface. The input is routed directly to the orchestrator agent, which acts as the core decision-making hub. Unlike rigid static workflows, the orchestrator leverages an LMM to parse the query’s semantic intent. It then selects the optimal execution path among subordinate agents, distinguishing simple retrieval tasks from those requiring multi-step reasoning. When external environmental context is required, the orchestrator invokes the data acquisition agent via the data acquisition tool. This specialized agent serves as the interface to the underlying RAG infrastructure. In order to handle heterogeneous data formats, the data acquisition agent is equipped with three distinct retrieval mechanisms:
  • CSV data retrieval pipeline: retrieves structured meteorological and historical wildfire data from a vector database using high-dimensional text embeddings.
  • Image data retrieval pipeline: retrieves visual analogs from satellite or aerial imagery databases, utilizing visual embeddings to identify patterns similar to the input image.
  • Multimodal RAG pipeline: synchronizes the retrieval of both text and image data when the query necessitates a combined environmental view.
This modular separation isolates raw data processing from higher-level reasoning logic. Once relevant data artifacts are retrieved and aggregated, control returns to the orchestrator, which activates the reasoning agent via the reasoning tool. The reasoning agent processes the aggregated context, comprising retrieved metadata and the original query, through an LMM to generate the final reasoned output. The orchestrator then formats this output and delivers it to the user interface, completing the interaction loop.

3.2. Materials and Datasets

This study leverages wildfire and environmental datasets, as shown in Table 4. The datasets are from public sources across multiple geographical regions, including Algeria [34], the United States [35,36], and Canada. These datasets were selected based on data availability and the significant ecological impact of wildfires in these areas.
To ensure a rigorous and fair comparison between different LMM backbones, a strict hold-out evaluation procedure has been employed. The primary datasets FLAME 3 and Algerian Forest Fires have been split into two mutually exclusive subsets. The first subset has been used only to populate the RAG vector indices ( I n d e x t a b , I n d e x i m g ) and select the few-shot exemplars ( ϵ ) used in context engineering. This subset represents the historical context that is accessible to the agents during their reasoning process. The second subset, referred to as the Evaluation Subset or Test Set, represents a reserved subset of instances used exclusively for final testing and benchmarking. This subset has been withheld from the vector database and used to simulate unseen wildfire scenarios used in testing.
The splitting of the datasets has been performed in a non-dynamic manner. A single test set was also created before any experimentation took place, and all models were evaluated using this same test set. This means that models such as GPT-4o, GPT-5-Nano, etc. have all been tested using the same unseen instances. This is in contrast to random re-sampling a test set for each different experimental run. This procedure ensures that all models have been tested under exactly the same conditions, so any variance in performance reported in Section 4 is solely due to the capabilities of each LMM backbone.

3.3. Data Preprocessing and Retrieving

The preprocessing pipeline addresses heterogeneous data sources, while ensuring quality, consistency, and compatibility with downstream retrieval and generation components. It is customized to the specific characteristics of the employed datasets.

3.3.1. Preprocessing and Retrieval Algorithms

Below are the algorithmic procedures for data preprocessing (Algorithm 1) and multimodal retrieval with generation (Algorithm 2). For a complete list of the mathematical notations and definitions used in these algorithms, please refer to Appendix B.
Algorithm 1 Multimodal RAG pipeline: preprocessing phase
Require: Tabular dataset D t a b , Image dataset D i m g
Ensure: Populated indices Index t a b , Index i m g
    Hyperparameters:
1:
τ m i s s 0.40          ▹ Threshold for dropping rows with missing values
2:
θ i q r 1.5             ▹ Multiplier for IQR-based outlier detection
3:
procedure PreprocessTabular( D t a b )
4:
      for each row r i D t a b  do
5:
          if  MissingRatio ( r i ) > τ m i s s  then
6:
               D t a b D t a b { r i }      ▹ Remove row if too much data is missing
7:
          else
8:
               r i Impute ( r i , method = { mean , mode } )
9:
               r i RemoveOutliers ( r i , θ i q r )
10:
              r i Normalize ( r i , Z - score )
11:
              t i SerializeToText ( r i )         ▹ Convert row data to text string
12:
              v i Model B A A I ( t i )             ▹ Generate 768-dim embedding
13:
            Index t a b . add ( v i , metadata ( r i ) )
14:
        end if
15:
    end for
16:
end procedure
17:
procedure PreprocessImages( D i m g )
18:
    for each image I j D i m g  do
19:
        if  ¬ IntegrityCheck ( I j )  or  IsDuplicate ( I j )  then
20:
            D i m g D i m g { I j }          ▹ Discard corrupt or duplicate images
21:
        else
22:
            I j Resize ( I j )               ▹ Normalize image resolution
23:
            u j Model C L I P ( I j )             ▹ Generate visual embedding
24:
            Index i m g . add ( u j , metadata ( I j ) )
25:
        end if
26:
    end for
27:
end procedure
Algorithm 2 Multimodal RAG pipeline: retrieval and generation phase
Require: User query q, Populated indices Index t a b , Index i m g
Ensure: Generated answer A
        Hyperparameters:
1:
k Top - k retrieval count        ▹ Number of items to retrieve per search
2:
procedure MultimodalGeneration(q)
3:
     q v e c Model B A A I ( q )          ▹ Embed user query for text search
4:
     C t a b Index t a b . SearchKNN ( q v e c , k )        ▹ Retrieve tabular context
5:
     q c l i p Model C L I P . EncodeText ( q )      ▹ Embed query for image search
6:
     C i m g Index i m g . SearchKNN ( q c l i p , k )        ▹ Retrieve visual context
7:
     C a u g PromptAugment ( q , C t a b , C i m g )       ▹ Combine retrieved contexts
8:
     A Model G e n ( C a u g )                ▹ Generate final reasoning
9:
    return A
10:
 end procedure
Figure 2 provides a high-level architectural view of the multimodal RAG pipeline, highlighting the interaction between preprocessing, vector indexing, multimodal retrieval, and LLM-based generation.
Furthermore, the retrieval depth parameter was set to k = 20 in order to maximize the level of retrieval in terms of recall within the context-retrieval phase. For wildfire management, it is important to maximize diversity in terms of historical analogs, especially in terms of retrieval numbers, in order to capture specific edge cases that could otherwise be lost with a lower number of retrieval counts. This design choice operates on the premise that the reasoning agent’s LMM (e.g., GPT-5-Nano) possesses sufficient attention mechanisms to effectively filter contextual noise from a larger pool of retrieved artifacts. As such, k = 20 was determined to be optimal in terms of retrieval in order to provide a wide enough scope of meteorological and visual scenarios within the context window of the prompt.

3.3.2. Pipeline Execution Flow

Algorithms 1 and 2 outline the process within pre-processing, retrieval and generation phases. The process starts with the processing of the raw tabular dataset D t a b , which contains meteorological and fire data, and the raw image dataset D i m g , comprising satellite or aerial imagery. The procedure iterates through the tabular dataset, removing rows r i with a missing value ratio exceeding the threshold τ m i s s of 0.40, followed by statistical imputation and the removal of outliers using the interquartile range multiplier θ i q r , which is 1.5. Subsequently, each valid row is serialized into a textual format t i and transformed into a vector embedding v i using the transformer model M o d e l B A A I (BAAI/bge-base-en-v1.5), which is then stored in the tabular vector index I n d e x t a b . In parallel, the system analyzes the image I j from the dataset D i m g , evaluating integrity checks that remove redundant and corrupted images, followed by resizing and the computation of the image embedding u j by the M o d e l C L I P (OpenAI CLIP ViT-B/32) transformer model, for storage in the image index  I n d e x i m g .
To address the challenge of aligning different embedding spaces (BAAI/bge-base-en-v1.5 embedding space and CLIP embedding space) without training a computationally expensive bridge encoder or projection layer, the system adopts a contextual prompt augmentation strategy instead of joint retrieval in a shared vector space. When a user provides a natural language query q, the textual component is converted into embeddings q v e c using M o d e l B A A I to query the tabular index I n d e x t a b . In order to increase the accuracy of these primary retrieval results, a re-ranking process is carried out using the Cohere 5.11.1 Re-rank model to produce the final tabular context C t a b . Meanwhile, the querying process is carried out in M o d e l C L I P to produce q c l i p to query visually similar elements C i m g from the image index I n d e x i m g . Due to the involvement of different modalities in different vector spaces, parallel streams are utilized in retrieval.
The alignment and synthesis of these heterogeneous data points take place during the augmentation step. Specifically, this process involves the deterministic construction of a composite input sequence, where the user query q, the serialized tabular records C t a b , and the retrieved visual artifacts C i m g are concatenated within a structured prompt template. A distinguishing feature of our multimodal RAG pipeline, compared to existing cross-modal retrieval frameworks, is the preservation of visual fidelity. Rather than generating intermediate textual captions, which acts as a lossy compression step, the system retrieves raw image artifacts in Base64 format. These are interleaved with tabular context in the augmented prompt C a u g , allowing the LMM to utilize its native cross-attention mechanisms on the original pixel data, yielding higher precision in identifying specific burn patterns.
Consequently, the proposed framework operates strictly at the representation level of information fusion. Unlike decision-level systems that aggregate independent probability scores from isolated unimodal models, our architecture interleaves raw visual encodings and serialized text within the LMM’s shared context window ( C a u g ). This enables the model to perform deep semantic fusion via cross-modal attention layers during the inference phase, mapping heterogeneous inputs into a unified latent representation space prior to generating the final reasoning output.

3.4. Structured Prompt Engineering and Context Formalization

In order to mitigate the stochastic nature of LMMs and ensure deterministic task execution, the proposed solution employs a structured prompt engineering framework. Rather than solely relying on natural language instructions, it formalizes the prompt space as a specialized instantiation of context engineering [38,39].
A pre-defined prompt p P is structured as a tuple p = ( ρ , γ , τ , ε , ϕ ) , where ρ R denotes the role specification, defining the agent’s identity and strictly bounded domain expertise to prevent out-of-scope task execution; γ G represents the goal state of the agent, defining the objective function the agent must optimize; τ T corresponds to the task vector, which is a sequence of executable instructions required to achieve γ ; ε E represents reference cases, providing ground-truth input–output pairs to guide the model’s reasoning trajectory. These cases are strictly curated from the training and validation subsets of the primary datasets (FLAME 3 and Algerian Forest Fires) to serve as few-shot exemplars. Specifically, we employ a static set of three exemplars for each agent. Unlike dynamic exemplar retrieval, which can result in different styles of reasoning depending on the retrieved exemplars, a static set guarantees deterministic behavioral convergence. In this way, the agent avoids prompt drift and ensures that it strictly adheres to the safety protocols with each cycle of execution. Importantly, to prevent data leakage and ensure the validity of the evaluation, there is no overlap between these reference cases and the independent test set used to generate the quantitative performance metrics presented in Section 4.4; ϕ C specifies operational constraints, including negative constraints (what not to do) and strict output formatting rules (e.g., JSON schemas) to ensure inter-agent communication compatibility. These context parameter values are provided for each agent before executing. While executing, agents can work along with these given context parameters.
This formalism is applied distinctively across the agentic workflow. For the orchestrator agent, the constraint set ϕ enforces a strict routing syntax to enable dynamic task decomposition without hallucinating non-existent tools. Similarly, for the reasoning agent, the task vector τ maintains a Chain-of-Thought (CoT) process to ensure that all outputs are explainable and grounded in the retrieved artifacts [40]. To resolve potential conflicts between heterogeneous data sources (e.g., a meteorological record suggesting low risk vs. visual evidence of a plume), the CoT prompt implements a ‘Hierarchical Evidence Weighting’ protocol. This instructs the reasoning agent to assign higher validity to direct observational evidence (raw visual artifacts from the Multimodal RAG pipeline) over static or predicted metadata (from the CSV pipeline) when identifying active hazards.

3.5. Proposed Multi-Agent Architecture

This study experimented with the following two architectural paradigms when implementing the proposed solutions.
  • Task-Based Approach
  • Orchestrator-Based Approach
Generally, the proposed MAS is formally defined as a tuple M = ( A , T , L , R ) , denoting Agent, Task, Tool and Rule, respectively, which are described as follows.
  • Agent A = { Orchestrator Agent , Data Acquisition Agent , Reasoning Agent } denotes the set of agents in the system. An Agent is an intelligent and automated unit powered by an LMM that performs specific tasks. Apart from the LMM, an agent must be provided with a role, a goal describing its instructions, a set of tasks to achieve the aforementioned goal, and a set of tools it can use to perform its assigned tasks. Agents can establish communication with other agents while maintaining their own memory of interactions. An output of an agent depends on the LMM it is supplied with. Therefore, the most suitable LMM for an agent may vary, and it largely depends on the tasks assigned to the agent. The proposed solution has three key agents that support its functionality.
    Data acquisition agents gather data for processing. This agent is supplied with data retrieval tools, including the CSV data retrieval tool, the Image data retrieval tool, and the multimodal RAG tool, and invoke them when necessary.
    The reasoning agent processes complex spatial–temporal patterns in wildfire behavior, analyzing the wildfire cases that have previously happened. Generally, this agent is invoked after the data acquisition agent is invoked, because the data acquisition agent provides the data that is retrieved from its multiple RAG pipelines. Then the reasoning agent provides the reasoning about the current situation with the data that it gets from the Data Acquisition agent.
    Orchestrator agent coordinates tasks and their execution within this MAS. As the central coordinator of the proposed framework, it enables full functionality by connecting other intelligent agents as tools. As shown in Figure 1, the data acquisition agent serves as the data acquisition tool, and the reasoning agent serves as the reasoning tool. The orchestrator agent facilitates communication between these sub-agents, allowing seamless task delegation. Its plan-and-execute nature provides adaptability to dynamic environments, making it ideal for domains such as wildfire management.
    Structurally, the system adopts a hub-and-spoke (Star) topology, enforcing strict isolation between subordinate agents. The data acquisition agent and reasoning agent operate in distinct environments and are invoked via independent API calls. Consequently, they possess no shared memory or lateral communication channels; the state of the orchestrator is opaque to them. This design ensures that the orchestrator agent acts as the sole source of truth for conversation history and global context, preventing information leakage and ensuring that all inter-agent data flow is explicitly filtered and routed through the central policy function.
  • Task T = { t 1 , t 2 , , t n } represents the set of tasks, which are dynamically planned and executed by the Orchestrator Agent.
  • Tool L = { CSV data retrieval tool , Image data retrieval tool , Multimodal RAG tool } represents the set of functions available within the system. Each tool is a specialized skill or capability that agents can invoke to execute specific actions. Beyond pre-built tools, custom tools can be developed and assigned to agents to extend their functionality. As shown in Figure 1, these tools are primarily utilized by the data acquisition agent and custom-built as follows.
    The CSV data-retrieval tool is an application interface that utilizes a CSV-based RAG Pipeline to retrieve raw metadata from indexed meteorological CSV data.
    The image data retrieval tool is an application interface that utilizes an Image Data Retrieval pipeline. It accepts both text and image inputs as queries and retrieves relevant images in base-64 format along with their metadata.
    The multimodal RAG tool utilizes the above retrieval pipelines in parallel. This ensures that the Reasoning Agent receives context that is both semantically and visually grounded.
  • Rule R represents the set of interaction rules governing agent coordination. These rules are specified independently of the agent prompts.

3.5.1. Task-Based Architecture

As shown in Figure 3, the task-based approach is modeled as a linear composite function. Let A d a t a and A r e a s o n denote the data acquisition agent and the reasoning agent, respectively. Then, the system response S to a posed user query q can be stated as in (1).
R = A reason A data ( q )
Figure 3 reflects the “sequential process”, where the output of the data acquisition task is passed as the input for the reasoning task. Here, the probability of a successful response P ( R s u c c e s s ) is strictly dependent on the joint probability of success for both sequential steps, as given in (2).
P ( R s u c c e s s ) = P ( s u c c e s s | A d a t a ) · P ( s u c c e s s | A r e a s o n )
Therefore, this task-based approach has the following limitations.
  • Error propagation: since the coupling is linear ( R = f ( g ( x ) ) ), any hallucination in the initial retrieval stage is propagated to the reasoning stage, which can lead to erroneous outcomes.
  • Computational inefficiency: as per (1), this approach forces the execution of A d a t a for every query. This results in an inevitable computational cost C o s t ( A d a t a ) , even when the retrieved data might not be of use (e.g., for general knowledge queries).
Accordingly, we propose an orchestrator-based approach to address the limitations of the task-based approach.

3.5.2. Orchestrator-Based Architecture

The orchestrator-based approach, as shown in Figure 1, redefines the system as a dynamic decision process. Unlike the fixed sequence in Figure 3, the orchestration agent node acts as a central policy function π that maps the current query q and current context at a given time t ( C t ) to an optimal target agent. The routing decision is modeled as an optimization problem where the orchestrator selects an agent a t a r g e t to maximize a relevance score S, such that π ( q , C t ) { a t a r g e t , T t o o l s } .
Let, I n t e n t M a t c h quantifies the semantic similarity between the user’s intent and the agent’s functional role; D a t a N e e d represents the necessity of specific modalities (e.g., a thermal image query explicitly weights the data acquisition agent); and α and β are adaptive weighting coefficients derived from the system prompt. Then, the selection of a t a r g e t , represented by the “Access Tools” arrows in Figure 1, is determined by (3).
a t a r g e t = argmax a A α · IntentMatch ( q , Role a ) + β · DataNeed ( q , Capability a )
The core novelty of our orchestration lies in the dynamic policy optimization that balances ‘Intent Match’ ( α ) against ‘Data Need’ ( β ). Unlike standard semantic routers that use a simple cosine similarity threshold, our Orchestrator dynamically instantiates these weighting coefficients based on query complexity. Crucially, we introduce a decay factor λ d e c a y (Algorithm 3, Line 20). This algorithmic modification prevents the infinite retrieval loops often observed in recursive agent frameworks, ensuring the system converges toward a reasoning state, a necessary architectural adaptation for emergency response systems.
The architectural decoupling of the data acquisition agent with the reasoning agent creates an artificial latency bottleneck that prevents the unverified discharge of assessment-level information. The application of the ‘decay-weighted’ routing policy π ( q , C t ) ensures the completion of the data retrieval cycle, which includes the cross-referencing of CSV artifacts, before the reasoning agent begins the process of creating an assessment-level response. This increases the time-to-insight for the user by approximately 176 s but acts as a logical gatekeeper to the unverified discharge of potentially hallucinated advice.
The routing logic relies on the dynamic maximization of the relevance score S a . As defined in Algorithm 3, Line 3, the weighting coefficients α (Intent Match importance) and β (Data Need importance) are not hard-coded constants but are dynamically instantiated by the Orchestrator’s LMM upon initial query analysis. The function π ( Q , C t ) represents the policy that maps the current query Q and context C t to the optimal agent.
Algorithm 3 Orchestrator dynamic policy optimization
Require: User query Q, Set of available agents A , Initial context C 0
Ensure: Final system response R
  Parameters: Decay rate λ d e c a y , Threshold τ
1:
  C t C 0
2:
  R NULL
3:
  α , β InitializeWeights ( Q )      ▹ Derived via LLM analysis of query intent
4:
 while R is NULL do
5:
      S m a x
6:
      a t a r g e t NULL
7:
     for each agent a A  do           ▹ The policy function π ( Q , C t )
8:
          I m a t c h ComputeIntentMatch ( Q , Role a )
9:
          D n e e d ComputeDataNeed ( Q , C t , Capability a )
10:
         S a ( α · I m a t c h ) + ( β · D n e e d )         ▹ Calculate optimization score
11:
        if  S a > S m a x  then
12:
              S m a x S a
13:
                 a t a r g e t a
14:
         end if
15:
    end for
16:
     if  a t a r g e t = = Data acquisition agent  then
17:
          T SelectTools ( a t a r g e t , Q )
18:
          D a t a ExecuteTools ( T )
19:
          C t C t D a t a          ▹ Update context with retrieved artifacts
20:
          β β · λ d e c a y            ▹ Decay data weight to force convergence
21:
    else if  a t a r g e t = = Reasoning agent  then
22:
          R GenerateResponse ( a t a r g e t , Q , C t )
23:
    else
24:
          R DirectResponse ( Orchestrator , Q )          ▹ Fallback for queries
25:
    end if
26:
 end while
27:
 return R
The InitializeWeights(Q) function represents a zero-shot prompt analysis, where the LMM assigns higher values to β the query Q contains explicit requests for external data (e.g., “Show me current wind speeds”) and higher values to α for logic-heavy queries.
To mitigate the inherent stochasticity of LMMs and ensure consistent routing behavior, strict deterministic guardrails were implemented for the orchestrator agent. First, the model’s temperature parameter was set to 0.0, effectively forcing the model to greedily sample the most probable token, thereby minimizing variance in weight generation ( α , β ). Second, the output space of the orchestrator was constrained using structured decoding and logit bias enforcement. That is the, bias occurs due to the unnormalized output score from the neural network’s final layer before applying activation functions like softmax to convert the output into probabilities. This restricts the generated weights to a discrete numerical range and enforces a strict JSON schema for tool selection, preventing format hallucinations. These mechanisms ensure that the policy function π ( q , C t ) remains mathematically stable across independent execution runs for identical queries.
In addition to the pseudocode provided by Algorithm 3, Table 5 shows the actual decision boundaries used by multi-agent system. It establishes the unique query conditions with the major weighting factors ( α , β ) and the corresponding execution flow, showing the enforcement of the Strategic Real-Time constraints.
Given that agents operate as independent API endpoints, the orchestrator agent implements a supervisor–worker validation protocol to handle inconsistent or low quality outputs. Since subordinate agents are stateless, the orchestrator can dynamically re-invoke them with refined parameters without resetting the entire system state. If an agent yields conflicting information, such as the data acquisition agent returning null meteorological data or if an agent exhibits poor reasoning (hallucinated formats), the orchestrator triggers a recursive correction loop. This mechanism allows the system to resolve conflicts by prioritizing grounded evidence over incomplete retrieval results, ensuring that temporary failures in one environment do not propagate as final system errors.

3.5.3. Comparison of Architecture

Table 6 shows a comparison between the two approaches, where the orchestrator-based method observes better maintainability, informing the decision to adopt it for scalable, adaptive wildfire management.

3.5.4. Application Implementation

Figure 4 illustrates the high-level application architecture, where an API Gateway routes user requests to the core Multi-Agent System for wildfire reasoning, alongside authentication and admin services. This backend integrates Large Multimodal Models (LMMs) for intelligence, utilizing Vector and NoSQL databases (MongoDB 7.0.28) for efficient multimodal data retrieval and secure session management.

4. Results

4.1. Applied Task Formulation

In order to properly assess and evaluate the effectiveness of the MAS, a multimodal wildfire reasoning and response task was formulated. The specific subtasks comprising this formulation (hazard identification, spread prediction, and response planning) are detailed in Table A1 in Appendix A. This task was designed to simulate the decision-making process of an incident commander under dynamic operational conditions. Adhering to the protocol defined in Section 3.2, a strict hold-out evaluation approach was maintained. The data is split in a manner such that none of the data used in creating the vector embeddings or few-shot prompts will ever enter the test set. To ensure a proper comparison between different LMM backbones, a common test set is used in all evaluations without dynamic re-sampling, ensuring that the reported metrics reflect true architectural differences rather than data variance.
The system’s objective extends beyond binary hazard detection (“Fire” vs. “No Fire”). For each test instance, the system is engaged in a process of multiple-step reasoning that entails the identification of the presence of live fire threats through a cross-checking of visual artifacts with meteorological information, prediction of the direction and rate of spread of fire based on wind direction information and terrain data, and the development of particularized mitigatory action plans based on synthesized information. While the quantitative measures of Accuracy, Precision, Recall, and F1-Score presented in this section of the paper primarily measure the system’s grounding and detection capacity, the quality of the system’s reasoning is measured through a qualitative assessment of the system’s generated responses.
With respect to the statistical analysis of variance, this research focused on deterministic behavior to ensure the level of reliability required for disaster response. As highlighted in the orchestration policy, the LMMs employed a temperature setting of 0.0 to ensure greedy sampling, thus eliminating the level of stochastic behavior within the generated outputs of the agents. As a result, the outputs are similar even if the models are executed several times on the same hold-out test set. Thus, the level of variance is negligible, and the evaluation strategy was focused on ensuring a strict architectural comparison, eliminating the impact of the random seed variations that are normally associated with generative models.
The reliability of the generated insights is validated through the architectural enforcement of the orchestrator’s recursive validation loop, not through the statistical averaging of stochastic repetitions. Unlike standard generative evaluations that quantify reliability via confidence intervals, this system utilizes an iterative self-correction mechanism to ensure the output converges to a stable, logic-compliant state. This approach effectively replaces the need for traditional confidence analysis based on repetition with architectural determinism. The system is designed to produce a optimal response for any given input. By having this operational consistency, the provided architecture guarantees that identical environmental variables trigger a uniform response.

4.2. Evaluation Metrics

We assessed the proposed solution utilizing a suite of metrics tailored to wildfire prediction and detection tasks. Core classification metrics, including accuracy, precision, recall, and F1-score, assess model reliability, guiding agent selection (e.g., GPT-5-Nano) to prioritize high precision in scenarios like analyzing California Rim Fire remote sensing data. This minimizes false alarms, reduces unnecessary resource deployment, and enhances interpretable decision-making for fire suppression strategies. Complementing these, latency evaluates real-time query processing in the conversational UI, ensuring timely responses to user prompts on FLAME 3 UAV imagery for rapid evacuation planning. Throughput measures the orchestrator-based architecture’s capacity to handle concurrent data streams from different sources, optimizing agent collaboration during peak wildfire events. Finally, resource utilization informs efficient allocation across LMM and RAG pipelines, enabling edge–device deployment for sustainable, field-deployable management without sacrificing contextual reasoning on diverse datasets such as the National USFS Fire Occurrence Point.

4.3. Web Application with a Visual Question–Answer System

The VQA application was developed with the objective of enabling natural and intuitive interactions, akin to human conversation, as illustrated in Figure 5. The interface features a clean and uncluttered layout, incorporating a sidebar on the left for efficient navigation through previous conversations. The central screen focuses on the ongoing dialogue, thereby maintaining user attention on the interaction. Constructed as a contemporary web application utilizing React 18.3.1 and TypeScript 5.8.3, along with the shadcn/UI 0.9.5 component library, the system guarantees high performance, reliability, and cross-device compatibility. The system follows a three-tier architecture model built using Vite 5.4.19 for optimized compilation and TanStack Query 5.83.0 for efficient server-state synchronization. The presentation layer captures multimodal inputs and communicates via RESTful APIs with the backend application layer, where the Orchestrator Agent manages execution logic, supported by a data layer that leverages vector databases for high-speed RAG retrieval.
The core functionality resides in the bottom input bar, offering user-friendly query submission options. These include typing detailed questions, uploading up to five images for contextual enhancement, or voicing queries aloud. Voice inputs are accurately transcribed into text, facilitating hands-free operation and on the move usability. Beyond basic querying, the system encompasses essential functional modules, including secure authentication, session management for conversation history persistence, and a markdown-based rendering engine. To enhance readability, the VQA’s responses are delivered in a structured and formatted manner. This multimodal approach that integrates text and images significantly increases the efficiency of the system and user accessibility. Case studies with specific user queries and responses of different models are available at our website [41].

4.4. Comparative Analysis of LLM

The proposed MAS was evaluated using five distinct LMM backbones to assess the trade-off between reasoning capability and operational efficiency. To ensure strictly comparable results, an identical, fixed test set was employed across all backbone evaluations. No model-specific data re-sampling or dynamic splitting was performed, ensuring that the performance metrics in Table 7 reflect architectural differences rather than data variances.
According to results shown in Table 7 and Figure 6, the GPT-4o model exhibits a symmetrical, balanced profile, maintaining equilibrium across Accuracy (0.700), Precision (0.752), and Recall (0.700). This distinct “diamond” shape indicates its suitability for general-purpose monitoring, where false negatives and false positives carry equal weight.
Conversely, GPT-5-Nano model demonstrates a skewed profile. While it achieves the highest system-wide Precision (0.797), it suffers from lower Recall (0.684). This asymmetry suggests the model adopts a conservative decision boundary, requiring high-confidence multimodal evidence before triggering an alert. In a wildfire context, this behavior is operationally advantageous for automated suppression systems, where false alarms incur high resource mobilization costs.
Inference latency is a critical performance metric for real-time deployment. As shown in Figure 7 visualizes this efficiency frontier. The results highlight that GPT-4.1-Mini, despite its high accuracy, incurs a significant latency penalty (213.02 s), rendering it suboptimal for rapid-response scenarios. In contrast, GPT-5-Nano occupies the optimal region of the graph (in purple), delivering superior precision with a 17% reduction in inference time (176.70 s) compared to the GPT 4.1-Mini variant. This balance validates the selection of GPT-5-Nano as the primary reasoning engine for the Orchestrator Agent.

4.5. Ablation Study and Architectural Validation

To quantify the dependency of the system on its multimodal components, an ablation study was performed by systematically disabling the CSV data retrieval, Image data retrieval, and multimodal RAG pipelines. Table 8 gives the quantitative performance observed across these configurations. The ablation analysis reveals a dependency on the multimodal RAG tool. As visualized in Figure 8, the removal of this component (green line) degrades performance by reducing accuracy from 0.684 to 0.471 and the F1-score from 0.736 to 0.515. This degradation confirms that the multimodal RAG pipeline functions as the main integration hub for the data acquisition agent. Without it, the system cannot effectively bridge the semantic gap between tabular data and unstructured visual inputs, leading to a breakdown in reasoning capabilities.
While the complete model incurs the highest inference latency (176.70 s), this cost is justified by the reliability gains. The “Multimodal Removed” configuration, while faster (116.94 s), yields unacceptable accuracy, validating the design choice to prioritize robust contextual integration over processing speed. Thus, the shrinkage of the “Multimodal Removed” polygon (green) reflects the system’s reliance on cross-modal synthesis for robust detection.
In contrast, independent removal of either CSV data retrieval tool or Image data retrieval tool resulted in only marginal performance drops ( Δ A c c u r a c y 0.02 0.04 ). This phenomenon is highlighted in the degradation heatmap shown in Figure 9. Here, the darker cells indicate a severe impact on performance. The high intensity of the bottom row confirms that removing the multimodal RAG tool is the most detrimental ablation. The top rows show lighter intensities (lower performance drop) compared to the bottom row. This suggests that the architecture possesses inherent redundancy; when a single-modality tool is disabled, the data acquisition agent can leverage the multimodal RAG tool as a fallback mechanism to retrieve the missing context by preventing total system failure.
To quantitatively evaluate the factual accuracy of the system’s reasoning and ensure it does not hallucinate safety-critical advice, precision is utilized as a direct proxy for groundedness, validated against the labeled ground-truth datasets (FLAME 3 and Algerian Forest Fires). In wildfire management, the hazardous hallucination impacts primarily as false positives, where the system makes a positive, soundly reasoned warning in the absence of a hazard. Thus, a high precision score means that the system’s positive predictions are always backed by facts, and it does not hallucinate or make up its predictions. In GPT-5-Nano, the precision was scored at 0.797, indicating that the system’s reasoning was mostly linked to the positive examples rather than being counterfeit. The efficacy of the proposed architecture in the mitigation of hallucinations is proven through the results obtained in the ablation study, as depicted in Table 8. If the LMMs were using internal knowledge to reason (hallucination), the removal of the retrieval tools would have little effect on the reasoning process. However, the sharp drop in the F1-score, from 0.736 to 0.515, when the Multimodal RAG tool is removed, confirms the strong contextual dependency of the LMM. This thus establishes the causality between the retrieved information and the generated insight, thereby proving that the reasoning process is grounded in the provided data artifacts and not the hallucinated information.
However, this rigorous grounding imposes specific limitations that must be acknowledged. The system exhibits a conservative decision boundary, evidenced by a lower recall of 0.684, indicating a trade-off where the model prioritizes the elimination of false alarms over maximal sensitivity. The system requires strong cross-modal evidence, such as the presence of “smoke”, as signified visually and meteorologically, before it concludes with a “Fire” determination. In this way, it may overlook fires at their infancy or unclear cases such as wisps of smoke carried by strong winds. The system relies on corroboration via the RAG model. Therefore data availability is crucial. The system cannot deduce whether it is seeing a fire using a black box CNN model. Instead, it relies on available metadata such as wind speed, humidity, historical data, etc. The system may default to a negative determination via the safety rules it applies if it cannot get this data. This underlines the need for a strong data backbone to support the high precision claimed. Moreover, positional drift poses a challenge, particularly in high-velocity situations where the system’s inference latency of approximately 176 s can result in a significant lag, thus requiring a “look-ahead” capability to synchronize the location received by the system with the actual location of the fire front.
Qualitatively, the multi-agent system’s error profile is shaped by systems architectural emphasis on high-precision reasoning over raw sensitivity. The most observed error category is False Negative, which stems from the model’s conservative decision boundary (Recall: 0.684). In some scenarios characterized by ambiguous visual cues, such as low-density smoke dispersed by high winds or thermally masked heat signatures, the reasoning agent try to reject the alert unless it is strongly corroborated by cross-modal evidence, at which time it leads to missed detections. High precision (0.797) minimizes the hallucinations, and contextual misalignment errors can be occurred if the retrieved historical analogs are semantically close but meteorologically differ. This potentially leads to accurate hazard detection but suboptimal spread prediction vectors.

5. Discussion

5.1. Lessons Learned

This study addressed the limitations of siloed data processing in wildfire management through the development of an orchestrator-based MAS. It was demonstrated that by integrating LMMs with specialized RAG pipelines, agentic workflows can effectively synthesize heterogeneous data, ranging from satellite imagery to historical fire records into coherent, context-aware decision support.
The findings presented in this research directly addressed the formulated research questions. RQ1 (Multimodal Integration) is achieved by the successful deployment of the data acquisition agent. It confirmed that a hierarchical agentic framework can dynamically retrieve and normalize disparate data formats, specifically tabular, visual, and sensor data, without the need for manual preprocessing. RQ2 (Operational Advantages) is justifiable by the comparative performance of the system that highlighted the value of the “Reasoning Agent.” Unlike traditional DL models that function as “black boxes,” the proposed architecture provides text-based reasoning. This shift from opaque probability scores to explained decisions was identified as critical for establishing trust with human incident commanders.

5.2. Comparative Analysis with State-of-the-Art Models

To validate the efficacy of the agentic approach, the system was benchmarked against established baselines in wildfire detection and management. Table 9 provides a high-level comparison of functional capabilities.
It was observed that DL models like SmokeyNet [14] excel at a high-speed pattern recognition but lack contextual awareness. Similarly, Fuzzy Inference Systems [7] and WSN controllers [8] provide efficient static monitoring yet struggle to adapt to dynamic, unstructured queries. In contrast, the proposed orchestrator-based system, while incurring higher latency (176.70 s), introduces a critical layer of semantic reasoning absent in these baselines, achieving a precision of 0.797. This performance profile is defined here as ‘Strategic Real-Time’. Compared to the standard manual coordination cycle (approx. 30 min), this latency delivers substantial operational acceleration.
To prevent this computation window from undermining the physical validity of responses, we integrated a predictive look-ahead mechanism. Instead of optimizing for the state at data ingestion ( t 0 ), the system requests wildfire spread predictions for a future time horizon ( t 0 + Δ t ), where Δ t covers the inference latency. As a result, generated resource allocation plans target the anticipated wildfire state at execution time, preserving decision intelligence despite the strategic latency.
To analytically estimate the operational impact of the predictive look-ahead mechanism, we conducted a sensitivity analysis simulating ‘Positional Drift’, which can be defined as the spatial error induced by inference latency, across a spectrum of wind velocities (20–80 km/h). We compared our approach against a Standard Non-Predictive Baseline, defined as a detection system that reports coordinates based solely on the data ingestion timestamp ( t 0 ), without compensating for the processing duration ( Δ t ).
As detailed in Table 10, the system’s benchmarked average inference latency ( Δ t ) of 176.70 s creates a significant spatial discrepancy that scales linearly with wind intensity. For instance, under a high-velocity benchmark of 60 km/h (ROS ≈ 100 m/min) typical of wind-sensitive shrublands, standard system latency results in a theoretical positional error of approximately 294.5 m ( 2.945 min × 100 m / min ). Table 10 demonstrates that across this entire operational range, integrating the t 0 + Δ t look-ahead projection mitigates systematic drift, reducing the spatial error to the marginal stochastic variance of the model’s inference time.
Figure 10 presents a performance comparison between the proposed MAS (utilizing GPT-5-Nano), a standard RAG Proxy, and a non-agentic LLM [14]. The results demonstrate that the orchestrator-based architecture significantly outperforms the standard RAG proxy, improving the F1-score from 0.515 to 0.736 and Precision from 0.569 to 0.797. Although the non-agentic model has a higher value of overall classification performance, including a precision of 0.898, it behaves like a ‘black box,’ with no specific cross-modal grounding applicable to disaster response tasks. Unlike the proposed MAS, where the outputs are validated by using the externally retrieved evidence, non-agentic approaches only use the parametric knowledge present in the model, resulting in a high chance of unverifiable hallucinations with no possible links to the original data.
Unlike generic agentic RAG systems that typically rely on open-ended ReAct loops and lossy image-to-text captioning, the proposed framework introduces a deterministically orchestrated architecture optimized for safety-critical environments. A key differentiator is the implementation of a decay-weighted routing policy ( π ( q , C t ) ), which mathematically forces task convergence to prevent the infinite retrieval loops often observed in standard autonomous agents. Furthermore, the system advances beyond standard multimodal RAG by utilizing a ‘lossless’ retrieval pipeline that injects raw Base64 visual artifacts directly into the LMMs context window, thereby preserving granular visual features that are discarded during the intermediate captioning steps of conventional frameworks. Finally, stochastic variability is strictly managed through a formalized context engineering tuple ( p = ( ρ , γ , τ , ϵ , ϕ ) ), which replaces loose natural language prompting with rigid state definitions to ensure the reproducibility required for disaster management.
We use the ‘Non-agentic LLM’ baseline to approximate the behaviour of black-box models like SmokeyNet [14]. While these models achieve superior precision (0.898) through strict visual pattern matching, they fail to distinguish benign events from genuine hazards. A prime example is a prescribed burn, where a black-box model simply detects ‘smoke’ and triggers an alert; our reasoning agent validates that visual input against active burn permits via the CSV RAG pipeline. Although this cross-verification yields a more conservative F1-score (0.736), it provides a critical operational advantage: effectively filtering false positives and contextualizing the event as ‘controlled’ rather than triggering the unnecessary mobilization typical of purely visual detection.
While recent studies have explored MAS for wildfire suppression, reliance is often placed on conceptual frameworks or simulations. For instance, Tavakol et al. [6] and Zadeh et al. [20] proposed swarm-based architectures for UAV coordination. While these approaches excel in coverage optimization, it was noted that they lack the cross-modal data synthesis capabilities of the orchestrator-based approach.
As detailed in Table 11, ML and DL models like SmokeyNet achieve higher raw performance metrics (e.g., Precision 89.84%) compared to the proposed system (Precision 79.7%). However, this disparity reflects the fundamental difference in operational scope. SmokeyNet is a specialized ‘narrow AI’ optimized solely for binary visual classification (Smoke vs. No Smoke) within a closed dataset. In contrast, the orchestrator-based MAS is designed for dynamic, open-world reasoning, integrating heterogeneous inputs such as wind speed, historical burn records, and thermal imagery.
In government decision-support scenarios, the marginal reduction in raw classification precision represents a necessary trade-off for achieving operational explainability. While black-box models may excel in pixel-level detection, they lack the semantic capacity to explain the cause or context of a fire. Our proposed system prioritizes the validity of final decisions such as evacuation planning, over the raw sensitivity of initial detection, ensuring incident commanders receive actionable intelligence rather than isolated alerts.
However, the tendency of LLM to hallucinate and the system’s overall accuracy being below 0.8 necessitate the need for the system to be deployed in an operationally rigorous manner. In the real-world scenario of natural disasters, the framework has the tendency to function more like a human-in-the-loop Decision Support System (DSS) rather than an autonomous system for decision-making, in which explainability becomes the key safety aspect of the system’s functioning. Unlike high-accuracy ‘black box’ models that may confidently misclassify benign events (false positives) without justification, the proposed system exposes its reasoning chain, allowing incident commanders to audit the logic, for instance, verifying if a fire spread prediction is physically consistent with the retrieved data. This transparency allows human experts to detect and discount hallucinations before taking action. To further enhance system reliability beyond the current human-in-the-loop paradigm, future work will integrate technical mitigations to address hallucinations. Implementing self-consistency mechanisms, where the model generates multiple reasoning chains to identify a consensus output, and filtering out stochastic errors can increase the power of the network. Integrating uncertainty quantification metrics that assign confidence scores to predictions, automatically flagging low-certainty outputs for mandatory manual verification, is another solution.

5.3. Challenges and Future Works

Despite these contributions, the proposed model is bottleneck with inference latency. As mentioned in the results, the reliance on LMMs introduces response delays (approximately 176 s for GPT-5-Nano), which may be prohibitive for real-time suppression schemes, where sub-second reaction times are required. Furthermore, the computational cost of continuous RAG retrieval presents a barrier to deployment on resource-constrained edge devices. To address these challenges, the proposed multimodal agentic AI framework with RAG-based VQA offers substantial potential for extension, building on its current strengths in synthesizing visual, textual, and geospatial data for wildfire insights while addressing limitations in computational efficiency and multimodal coverage.
This framework can be extended to incorporate an audio RAG module to process forest acoustic data, enabling classification of fire-specific sounds such as crackling flames and animal distress calls [44]. This would leverage audio-language models integrated with agentic workflows, allowing autonomous agents to triage auditory alerts in real time and fuse them with visual VQA outputs for richer situational awareness. Moreover, real-time multimodal integration will be pursued by coupling the system with streaming data sources such as satellite feeds, IoT sensors, and drone imagery [45]. Similarly, land cover and land use fusion will incorporate high-resolution datasets into the decision-making pipeline, empowering agents to assess vulnerability of infrastructure and biodiversity hotspots [13]. Consequently, agentic orchestration frameworks could enable dynamic risk analysis, where specialized agents negotiate insights such as smoke plume prediction via diffusion models, evacuation routing, and update the RAG knowledge base on-the-fly, enhancing responsiveness in operational wildfire centers. Additionally, a judge agent can be implemented to assess the outputs and decisions of other agents to ensure accuracy, reliability, and alignment with real-time priorities, thereby minimizing errors in critical scenarios [46].
Additionally, optimization strategies for scalability, including model distillation, quantized embeddings, and federated learning across edge devices in disaster-prone regions, can be designed and utilized to tackle computational overhead from pretrained models and vector databases [47,48,49]. This ensures low-latency inference without sacrificing accuracy, critical for large-scale deployment.
Importantly, chain-of-thought prompting is a crucial aspect, as it decomposes intricate reasoning processes into sequential intermediate steps, significantly boosting accuracy, interpretability, and performance on multimodal challenges [45,50].
Generally, black-box models often obscure the reasoning of the solution, making it difficult for stakeholders to validate outputs against ground realities or regulatory standards. By providing interpretable insights such as attention maps highlighting key image features or reasoning chains, XAI enhances model accountability, facilitates error debugging, and supports interdisciplinary collaboration among scientists, policymakers, and local communities through VQA systems [27,28,29]. Ultimately, this transparency empowers more reliable, ethical applications in high-stakes scenarios like rapid disaster response, where opaque predictions could lead to misguided interventions. Further, post-disaster assessment can integrate AI-driven MAS to enable real-time, predictive evaluations of burn severity, erosion risks, and recovery needs, enhancing proactive rehabilitation efforts [51].
Finally, the framework will evolve toward a standardized, interoperable architecture, facilitating multi-agenct collaboration. Extensions to other disasters, including floods via hydroacoustic RAG and United Nation’s sustainable development goals (SDG), particularly SDG 13 (climate action) and SDG 15 (life on land), will promote forest preservation as carbon sinks, climate resilience, and reduced deforestation through proactive, non-technical VQA access for responders [21]. These advancements position the system as a versatile tool for agentic AI in environmental disaster management.

6. Conclusions

This paper proposed a novel orchestrated multi-agent framework designed to transform the landscape of wildfire management through context-aware, multimodal reasoning. By integrating LMMs with specialized RAG pipelines, the proposed system successfully bridged the gap between heterogeneous data streams, including satellite imagery, meteorological tabular records, and ground-based footage, enabling a human-centric VQA interface. Experimental validation across diverse geographical datasets demonstrated a precision of 0.797 and an F1-score of 0.736. While inference latency remains a challenge for real-time suppression, the modularity of the architecture provides a robust foundation for future extensions. Finally, this work aligns with global sustainability goals by providing decision-makers with a proactive, scalable, and adaptive tool for ecosystem preservation and disaster resilience.

Author Contributions

Conceptualization, D.M. and C.P.; methodology, A.S., S.J., S.S., V.S. and D.M.; software, A.S., S.J., S.S. and V.S.; validation, A.S., S.J., S.S., V.S. and D.M.; formal analysis, A.S., S.J., S.S. and V.S.; investigation, A.S., S.J., S.S., V.S. and D.M.; resources, D.M. and C.P.; data curation, A.S., S.J., S.S., V.S. and D.M.; writing—original draft preparation, A.S., S.J., S.S. and V.S.; writing—review and editing, D.M. and C.P.; visualization, A.S., S.J., S.S. and V.S.; supervision, D.M. and C.P.; project administration, D.M. and C.P.; funding Acquisition, D.M. and C.P. All authors have read and agreed to the published version of the manuscript.

Funding

This publication does not support with funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: Algerian Forest Fires Dataset [34], California Rim and King Forest Fires data [36], National USFS Fire Occurrence Point data [37], and the FLAME 3 Dataset [35].

Acknowledgments

This work was supported in part by the University of Moratuwa, Sri Lanka, under the Senate Research Committee, SRC/LT/2025/25 Conference and Publishing Grant.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Multimodal Wildfire Reasoning and Response Tasks

The “Multimodal Wildfire Reasoning and Response” task formulated in Section 4.1 is composed of three distinct sub-tasks designed to simulate the decision-making pipeline of an incident commander. Table A1 details the input, objective, and output for each sub-task.
Table A1. Definitions of the sub-tasks comprising the Multimodal Wildfire Reasoning and Response evaluation.
Table A1. Definitions of the sub-tasks comprising the Multimodal Wildfire Reasoning and Response evaluation.
Task IDTask NameDefinition and Scope
Task 1Hazard IdentificationObjective: Detect the presence of active fire threats by cross-referencing visual artifacts with meteorological data.
(Detection)Input: Satellite/UAV imagery + Wind/Temperature data.
Output: Binary Classification (Fire/No Fire) and confidence score.
Task 2Spread PredictionObjective: Forecast the future direction and rate of spread (ROS) of the identified fire front.
Input: Wind direction, wind speed, and terrain features.
Output: Directional vector and estimated ROS (e.g., “rapid downslope expansion”).
Task 3Response PlanningObjective: Formulate particularized mitigatory action plans based on the synthesized threat assessment.
Input: Synthesized outputs from Task 1 and Task 2.
Output: Actionable recommendations (e.g., “initiate evacuation in Sector 4”).

Appendix B. Mathematical Notations

The following table summarizes the mathematical notations, variable definitions, and processing functions used throughout the proposed Multimodal RAG pipeline.
Table A2. Summary of mathematical notations and definitions used in the framework.
Table A2. Summary of mathematical notations and definitions used in the framework.
CategoryNotationDescription
Data Domains D t a b { r 1 , r 2 , , r n } , where each r i R d represents a meteorological record with d features (temperature, humidity, wind speed, etc.).
D i m g { I 1 , I 2 , , I m } , where each I j R H × W × C denotes satellite or aerial imagery with height H, width W, and C color channels.
Hyperparameters τ m i s s Missing value threshold for row removal (default: 0.40). Range [ 0 , 1 ] .
θ i q r Interquartile Range (IQR) multiplier for outlier detection (default: 1.5). θ i q r R + .
Embedding Models Model B A A I Text embedding function ( T R 768 ) using BAAI/bge-base-en-v1.5, mapping text strings to 768-dimensional vectors.
Model C L I P Multimodal embedding function ( I T R 512 ) using OpenAI CLIP, mapping images or text to 512-dimensional vectors.
Vector Indices Index t a b Vector database storing tabular embeddings { v i } with associated metadata.
Index i m g Vector database storing visual embeddings { u j } with associated metadata.
Retrieval ParamskTop-k retrieval count ( k N + ), specifying the number of most similar items to retrieve per search query.
Context VariablesqUser query in natural language ( q T ).
q v e c Text embedding of query q for tabular search ( q v e c R 768 ).
q c l i p CLIP embedding of query q for visual search ( q c l i p R 512 ).
C t a b Retrieved tabular context (top-k meteorological records).
C i m g Retrieved visual context (top-k similar images).
C a u g Augmented prompt combining q, C t a b , and C i m g .
AFinal generated answer.
Functions MissingRatio ( r i ) Computes fraction of missing values in row r i . Output [ 0 , 1 ] .
Impute ( r i ) Fills missing values using statistical methods (mean for numerical, mode for categorical).
RemoveOutliersRemoves values outside [ Q 1 θ · IQR , Q 3 + θ · IQR ] .
Normalize ( r i ) Applies Z-score normalization: z = x μ σ .
SerializeToTextConverts structured row data to natural language text t i T .
IntegrityCheckValidates image file integrity. Returns { true , false } .
IsDuplicate ( I j ) Detects duplicate images via perceptual hashing. Returns { true , false } .
Resize ( I j ) Normalizes image resolution to standard dimensions.
SearchKNN ( v , k ) Retrieves k nearest neighbors to vector v using cosine similarity.
PromptAugmentConstructs structured prompt template combining query q and retrieved contexts C t a b , C i m g .
Model G e n ( C a u g ) Large Multimodal Model that generates final reasoning A from augmented context.

References

  1. National Interagency Fire Center. Wildland Fire Summary and Statistics Annual Report 2024. 2024. Available online: https://www.nifc.gov/fire-information/statistics (accessed on 1 December 2024).
  2. Wasserman, T.N.; Mueller, S.E. Climate influences on future fire severity: A synthesis of climate-fire interactions and impacts on fire regimes, high-severity fire, and forests in the western United States. Fire Ecol. 2023, 19, 43. [Google Scholar] [CrossRef]
  3. Naser, M.; Kodur, V. Vulnerability of structures and infrastructure to wildfires: A perspective into assessment and mitigation strategies. Nat. Hazards 2025, 121, 9995–10015. [Google Scholar] [CrossRef]
  4. Sharma, D.; Kashyap, M.P.; Das, D.; Chatterji, B.; Modi, M.; Talukdar, N. Assessing fire-induced tree cover loss and its contribution to carbon emission in BRICS+ nations. Discov. Environ. 2026, 4, 7. [Google Scholar] [CrossRef]
  5. Von Scheffer, C.; Mauquoy, D.; Theurer, T.; Coathup, D.; Muirhead, D. ‘Fire Islands’: Holocene wildfire intensity as a critical determinant of carbon accumulation in South Atlantic peatlands. Quat. Sci. Rev. 2026, 374, 109759. [Google Scholar] [CrossRef]
  6. Tavakol Sadrabadi, M.; Peiró, J.; Innocente, M.S.; Rein, G. Conceptual design of a wildfire emergency response system empowered by swarms of unmanned aerial vehicles. Int. J. Disaster Risk Reduct. 2025, 124, 105493. [Google Scholar] [CrossRef]
  7. Duarte, M.; da Silva, T.A.; de Sousa, J.P.; de Castro, A.L.; Lourenço, R. Fuzzy Inference System for Mapping Forest Fire Susceptibility in Northern Rondônia, Brazil. Geogr. Environ. Sustain. 2024, 17, 83–94. [Google Scholar] [CrossRef]
  8. Toledo-Castro, J.; Caballero-Gil, P.; Rodríguez Pérez, N.; Santos-González, I.; Hernández-Goya, C.; Aguasca, R. Forest Fire Prevention, Detection, and Fighting Based on Fuzzy Logic and Wireless Sensor Networks. Complexity 2018, 2018, 1639715. [Google Scholar] [CrossRef]
  9. Cao, J.; Liu, X.; Xue, R. FireMM-IR: An Infrared-Enhanced Multi-Modal Large Language Model for Comprehensive Scene Understanding in Remote Sensing Forest Fire Monitoring. Sensors 2026, 26, 390. [Google Scholar] [CrossRef]
  10. Binlajdam, R.; Meedeniya, D.; Jayaweera, C.; Karakus, O.; Rana, O.; Ter Wengel, P.; Goossens, B.; Lertsinsrubtavee, A.; Mekbungwan, P.; Mishra, D.; et al. Review on Sustainable Forestry with Artificial Intelligence. ACM J. Comput. Sustain. Soc. 2025, 3, 1–48. [Google Scholar] [CrossRef]
  11. Andrianarivony, H.S.; Akhloufi, M.A. Machine learning and deep learning for wildfire spread prediction: A review. Fire 2024, 7, 482. [Google Scholar] [CrossRef]
  12. Saleh, A.; Zulkifley, M.A.; Harun, H.H.; Gaudreault, F.; Davison, I.; Spraggon, M. Forest fire surveillance systems: A review of deep learning methods. Heliyon 2024, 10, e23127. [Google Scholar] [CrossRef]
  13. Jayanetti, A.; Meedeniya, D.; Dilini, N.; Wickramapala, M.; Madushanka, H. Enhanced land cover and land use information generation from satellite imagery and foursquare data. In Proceedings of the 6th International Conference on Software and Computer Applications(ICSCA); ACM: Bangkok, Thailand, 2017; pp. 149–153. [Google Scholar] [CrossRef]
  14. Dewangan, A.; Pande, Y.; Braun, H.W.; Vernon, F.; Perez, I.; Altintas, I.; Cottrell, G.W.; Nguyen, M.H. FIgLib & SmokeyNet: Dataset and deep learning model for real-time wildland fire smoke detection. Remote Sens. 2022, 14, 1007. [Google Scholar] [CrossRef]
  15. Xie, Y.; Jiang, B.; Mallick, T.; Bergerson, J.D.; Hutchison, J.K.; Verner, D.R.; Branham, J.; Alexander, M.R.; Ross, R.B.; Feng, Y.; et al. WildfireGPT: Tailored Large Language Model for Wildfire Analysis. arXiv 2025, arXiv:2402.07877. [Google Scholar] [CrossRef]
  16. Du, S.; Li, J.; Noto, M. Wildfire scene recognition based on qwen2-wildfire. In Proceedings of the 2025 8th International Conference on Software Engineering and Information Management; ACM: Singapore Singapore, 2025; pp. 254–262. [Google Scholar] [CrossRef]
  17. Meedeniya, D.; Jayaweera, C. Blazing Trails: Cutting-Edge Technologies Revolutionizing Forest Fire Screening. In Pioneering Autonomous Technology A Deep Dive into Hyper Automation, 1st ed.; Swain, K., Pattnaik, P., Poonia, R., Nayak, S., Eds.; Elsevier Academic Press: Amsterdam, The Netherlands, 2026; Volume 143, Chapter 3; pp. 1–16. [Google Scholar] [CrossRef]
  18. Mawanza, L. Fault-tolerant dynamic formation control of the heterogeneous multi-agent system for cooperative wildfire tracking. Syst. Sci. Control Eng. 2025, 12, 2294991. [Google Scholar] [CrossRef]
  19. Kouzehgar, M.; Meghjani, M.; Bouffanais, R. Multi-Agent Reinforcement Learning for Dynamic Ocean Monitoring by a Swarm of Buoys. arXiv 2020, arXiv:2012.11641. [Google Scholar] [CrossRef]
  20. Zadeh, R.B.; Elmi, A.; Moghaddam, V.; MahmoudZadeh, S. A Conceptual High-Level Multiagent System for Wildfire Management. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5911415. [Google Scholar] [CrossRef]
  21. UNDP. Sustainable Development Goals. 2015. Available online: https://www.undp.org/sustainable-development-goals (accessed on 25 December 2025).
  22. Ángel Javaloyes, M.; Pendás-Recondo, E.; Sánchez, M. A general model for wildfire propagation with wind and slope. SIAM J. Appl. Algebra Geom. 2023, 7, 414–439. [Google Scholar] [CrossRef]
  23. Gao, X.; Cao, C.; Wang, S.; Xu, M.; Li, J.; Yang, X.; Yang, Y.; Hu, R.; Zhang, Y.; Wu, S.; et al. Remote sensing diagnosis of Forest fire risk based on state-trend characteristics using machine learning models. Ecol. Indic. 2026, 182, 114527. [Google Scholar] [CrossRef]
  24. Yel, S.G.; Küçüker, D.M.; Görmüş, E.T. Wildfire susceptibility mapping with multiple machine learning algorithms utilizing forest inventory and FIRMS data: A case study in Arsin, Trabzon, Türkiye. Int. J. Appl. Earth Obs. Geoinf. 2026, 146, 105091. [Google Scholar] [CrossRef]
  25. Uma Maheswara Rao, R.; Waila, P.; Mammen, P.C.; Muthu, R. Enabling Artificial Intelligence (AI) and Machine Learning (ML) Techniques for Managing Forest Fires. In Application of Machine Learning in Earth Sciences: A Practical Approach; Springer: Berlin/Heidelberg, Germany, 2026; pp. 429–459. [Google Scholar] [CrossRef]
  26. Mousa, M.H.; Algamdi, A.M.; Fouad, Y.; Elshewey, A.M. CNN-MLP framework for forest burned areas prediction using PSO-WOA algorithm. Sci. Rep. 2026, 16, 4982. [Google Scholar] [CrossRef]
  27. Saranya, A.; Subhashini, R. A systematic review of Explainable Artificial Intelligence models and applications: Recent developments and future trends. Decis. Anal. J. 2023, 7, 100230. [Google Scholar] [CrossRef]
  28. Ahangama, I.; Meedeniya, D.; Pradhan, B. Explainable Image Segmentation for Spatio-Temporal and Multivariate Image Data in Precipitation Nowcasting. Results Eng. 2025, 26, 105595. [Google Scholar] [CrossRef]
  29. Klotz, J.; Burgert, T.; Demir, B. On the Effectiveness of Methods and Metrics for Explainable AI in Remote Sensing Image Scene Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 27764–27780. [Google Scholar] [CrossRef]
  30. Pan, Y.; Yang, J.; Lu, M.; Bao, Q.; Zhu, T.; Yao, Q.; New, S.; Chen, D.; Shi, C.; Chen, L. Bridging the “Last-mile Gap” in Climate Services Delivery: A Dynamical-AI Hybrid Framework for Next-Month Wildfire Danger Prediction and Emergency Action. Adv. Atmos. Sci. 2026, 43, 706–722. [Google Scholar] [CrossRef]
  31. Aththanayake, S.; Mallikarachchi, C.; Wickramasinghe, J.; Kugarajah, S.; Meedeniya, D.; Pradhan, B. ResQConnect: An AI-Powered Multi-Agentic Platform for Human-Centered and Resilient Disaster Response. Sustainability 2026, 18, 1014. [Google Scholar] [CrossRef]
  32. Xie, Y.; Jiang, B.; Mallick, T.; Bergerson, J.; Hutchison, J.K.; Verner, D.R.; Branham, J.; Alexander, M.R.; Ross, R.B.; Feng, Y.; et al. MARSHA: Multi-agent RAG system for hazard adaptation. npj Clim. Action 2025, 4, 70. [Google Scholar] [CrossRef]
  33. Kalatzis, N.; Avgeris, M.; Dechouniotis, D.; Papadakis-Vlachopapadopoulos, K.; Roussaki, I.; Papavassiliou, S. Edge Computing in IoT ecosystems for UAV-enabled Early Fire Detection. In Proceedings of the 2018 IEEE International Conference on Smart Computing (SMARTCOMP), Taormina, Italy, 18–20 June 2018; pp. 106–114. [Google Scholar] [CrossRef]
  34. Abid, F. Algerian Forest Fires Dataset. 2019. Available online: https://archive.ics.uci.edu/ml/datasets/Algerian+Forest+Fires+Dataset (accessed on 12 December 2025).
  35. Hopkins, B.; ONeill, L.; Marinaccio, M.; Rowell, E.; Parsons, R.; Flanary, S.; Nazim, I.; Seielstad, C.; Afghah, F. FLAME 3 Dataset: Unleashing the Power of Radiometric Thermal UAV Imagery for Wildfire Management. arXiv 2024, arXiv:2412.02831. [Google Scholar] [CrossRef]
  36. Stavros, N.; Tane, Z.; Kane, V.; Veraverbeke, S.; McGaughey, R.; Lutz, J.A.; Ramirez, C.; Schimel, D.S. Remote Sensing Data Before and After California Rim and King Forest Fires, 2010–2015. 2016. Available online: https://www.earthdata.nasa.gov/data/catalog/ornl-cloud-king-rim-fire-analysis-1288-1 (accessed on 12 December 2025).
  37. Data, G. National USFS Fire Occurrence Point. 2025. Available online: https://catalog.data.gov/dataset/national-usfs-fire-occurrence-point-feature-layer-d3233 (accessed on 7 August 2025).
  38. Hua, Q.; Ye, L.; Fu, D.; Xiao, Y.; Cai, X.; Wu, Y.; Lin, J.; Wang, J.; Liu, P. Context Engineering 2.0: The Context of Context Engineering. arXiv 2025, arXiv:2510.26493. [Google Scholar] [CrossRef]
  39. Wu, S.; Qiao, Y.; He, S.; Zhou, J.; Wang, Z.; Li, X.; Wang, F. FireCLIP: Enhancing Forest Fire Detection with Multimodal Prompt Tuning and Vision-Language Understanding. Fire 2025, 8, 237. [Google Scholar] [CrossRef]
  40. Zhang, Z.; Yao, Y.; Zhang, A.; Tang, X.; Ma, X.; He, Z.; Wang, Y.; Gerstein, M.; Wang, R.; Liu, G.; et al. Igniting language intelligence: The hitchhiker’s guide from chain-of-thought reasoning to language agents. ACM Comput. Surv. 2025, 57, 1–39. [Google Scholar] [CrossRef]
  41. Sandeep, A.; Samarappuli, V.; Jayarathne, S.; Sandaruwan, S. AI-Powered Insight Engine for Wildfire Data Reasoning. 2025. Available online: https://sites.google.com/cse.mrt.ac.lk/fusionsense (accessed on 12 December 2025).
  42. Cruz, M.G.; Alexander, M.E. The 10% wind speed rule of thumb for estimating a wildfire’s forward rate of spread in forests and shrublands. Ann. For. Sci. 2019, 76, 44. [Google Scholar] [CrossRef]
  43. Andrews, P.L. The Rothermel Surface Fire Spread Model and Associated Developments: A Comprehensive Explanation and Guide; Technical Report RMRS-GTR-371; USDA Forest Service, Rocky Mountain Research Station: Fort Collins, CO, USA, 2018. [Google Scholar] [CrossRef]
  44. Meedeniya, D.; Ariyarathne, I.; Bandara, M.; Jayasundara, R.; Perera, C. A Survey on Deep Learning-based Forest Environment Sound Classification at the Edge. ACM Comput. Surv. 2023, 56, 66. [Google Scholar] [CrossRef]
  45. Faria, F.T.J.; Baniata, L.H.; Choi, A.; Kang, S. Towards Robust Chain-of-Thought Prompting with Self-Consistency for Remote Sensing VQA: An Empirical Study Across Large Multimodal Models. Mathematics 2025, 13, 3046. [Google Scholar] [CrossRef]
  46. Zhao, X.; Wang, H.; Dai, C.; Tang, J.; Deng, K.; Zhong, Z.; Kong, F.; Wang, S.; Morikawa, S. Multi-Stage Simulation of Residents’ Disaster Risk Perception and Decision-Making Behavior: An Exploratory Study on Large Language Model-Driven Social–Cognitive Agent Framework. Systems 2025, 13, 240. [Google Scholar] [CrossRef]
  47. Paranayapa, T.; Ranasinghe, P.; Ranmal, D.; Meedeniya, D.; Perera, C. A Comparative Study of Preprocessing and Model Compression Techniques in Deep Learning for Forest Sound Classification. Sensors 2024, 24, 1149. [Google Scholar] [CrossRef]
  48. Karim, M.M.; Van, D.H.; Khan, S.; Qu, Q.; Kholodov, Y. Ai agents meet blockchain: A survey on secure and scalable collaboration for multi-agents. Future Internet 2025, 17, 57. [Google Scholar] [CrossRef]
  49. Ranmal, D.; Ranasinghe, P.; Paranayapa, T.; Meedeniya, D.; Perera, C. ESC-NAS: Environment Sound Classification Using Hardware-Aware Neural Architecture Search for the Edge. Sensors 2024, 24, 3749. [Google Scholar] [CrossRef]
  50. Han, Z.; Wang, J.; Yan, X.; Jiang, Z.; Zhang, Y.; Liu, S.; Gong, Q.; Song, C. CoReaAgents: A Collaboration and Reasoning Framework Based on LLM-Powered Agents for Complex Reasoning Tasks. Appl. Sci. 2025, 15, 5663. [Google Scholar] [CrossRef]
  51. Chen, Z.; Asadi Shamsabadi, E.; Jiang, S.; Shen, L.; Dias-da Costa, D. Integration of large vision language models for efficient post-disaster damage assessment and reporting. Nat. Commun. 2026. [Google Scholar] [CrossRef]
Figure 1. High-level architectural diagram of the proposed solution.
Figure 1. High-level architectural diagram of the proposed solution.
Sensors 26 01070 g001
Figure 2. Architectural overview of the Multimodal RAG Framework.
Figure 2. Architectural overview of the Multimodal RAG Framework.
Sensors 26 01070 g002
Figure 3. System Architecture Overview Diagram of the Task-Based Approach.
Figure 3. System Architecture Overview Diagram of the Task-Based Approach.
Sensors 26 01070 g003
Figure 4. High-level Architectural Diagram of the application.
Figure 4. High-level Architectural Diagram of the application.
Sensors 26 01070 g004
Figure 5. Conversational User Interface.
Figure 5. Conversational User Interface.
Sensors 26 01070 g005
Figure 6. Radar graph for the performance of different models.
Figure 6. Radar graph for the performance of different models.
Sensors 26 01070 g006
Figure 7. Accuracy vs. Inference Time.
Figure 7. Accuracy vs. Inference Time.
Sensors 26 01070 g007
Figure 8. Visualization of ablation performance.
Figure 8. Visualization of ablation performance.
Sensors 26 01070 g008
Figure 9. Performance degradation heatmap.
Figure 9. Performance degradation heatmap.
Sensors 26 01070 g009
Figure 10. Performance of the Proposed MAS against Standard RAG and Non-agentic LLM baselines.
Figure 10. Performance of the Proposed MAS against Standard RAG and Non-agentic LLM baselines.
Sensors 26 01070 g010
Table 1. Architectural contrast between standard Agentic RAG paradigms and the proposed framework.
Table 1. Architectural contrast between standard Agentic RAG paradigms and the proposed framework.
FeatureStandard Agentic RAGProposed Framework (Ours)Architectural Advantage
Orchestration LogicRecursive ReAct Loops: Relies on open-ended ‘Reason + Act’ cycles that often loop indefinitely or hallucinate tools in complex scenarios.Decay-Weighted Routing: Implements a policy function π ( q , C t ) with a decay factor λ d e c a y that mathematically forces task convergence.Prevents infinite loops and ensures deterministic latency for safety-critical response.
Multimodal DataIntermediate Captioning: Converts images to text descriptions before processing, causing loss of granular visual details (e.g., smoke density).Lossless Artifact Injection: Retrieves and injects raw Base64 visual artifacts directly into the LMM context window.Preserves forensic visual fidelity required for distinguishing similar hazards (e.g., cloud vs. smoke).
Context ManagementUnstructured Logs: Appends raw conversation history to the prompt, leading to context drift and unauthorized tool usage.Formalized State Tuples: Uses a rigid tuple structure p = ( ρ , γ , τ , ϵ , ϕ ) to strictly define role boundaries and constraints.Guarantees reproducibility and prevents agents from acting outside safety guardrails.
Table 2. Overview of Related Studies on Wildfire Screening.
Table 2. Overview of Related Studies on Wildfire Screening.
StudyDescriptionApproachRegion
Rule-Based Fuzzy Logic AI Other
SmokeyNet (2022) [14]Multimodal smoke detectionCNN, LSTM, ViTUSA (HPWREN sites)
UAV Swarms for WER (2025) [6]Autonomous UAV wildfire suppressionSimple rulesMAS, Swarm RoboticsUnited Kingdom
High-Level MAS with DRL (2025) [20]MAS with DRL for fire trackingMAS, Deep RLGlobal
Heterogeneous MAS (2025) [18]UAV/ground robot monitoringMAS
MARL-based Systems (2020) [19]Swarm ocean monitoringMulti-Agent RLBedok Reservoir, Singapore
WildfireGPT (2025) [15]RAG-based LLM decision supportMAS, LLM, RAGUnited States
Fuzzy Fire Mapping (2024) [7]Fire susceptibility mappingFuzzy Inference SystemBrazil (Rondônia)
WSN Fire Controller (2018) [8]IoT/WSN Fire ControllerDynamic Fuzzy LogicIoT, WSNSpain
Edge-UAV System (2018) [33]UAV Early DetectionEdge ComputingEuropean South Region
Table 3. Comparison of Existing Systems.
Table 3. Comparison of Existing Systems.
StudyMAS ArchitectureUAV/Drone SwarmsMulti Modal FusionVQAContext AwarenessHistorical DataRAGMAS OrchestrationExplainable Decisions
SmokeyNet [14]Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001
UAV Swarms for WER [6]Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002
High-Level MAS with DRL [20]Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002
Heterogeneous MAS [18]Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002
WildfireGPT [15]Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001
MARL-based Systems [19]Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001Sensors 26 01070 i002
Fuzzy Fire Mapping [7]Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001
WSN Fire Controller [8]Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001
Edge-UAV System [33]Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i002Sensors 26 01070 i001Sensors 26 01070 i002
Proposed VQA-MAS System (Ours)Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001Sensors 26 01070 i001
Sensors 26 01070 i001 Satisfies; Sensors 26 01070 i002 Does Not Satisfy.
Table 4. Dataset features and reasons for using them.
Table 4. Dataset features and reasons for using them.
DatasetTypeDescriptionFeatures
Algerian Forest Fires [34]Tabular244 instances from Bejaia (northeast) and Sidi Bel-Abbes (northwest) regions in Algeria, 122 per region.Contains meteorological data for in CSV format, indexed in the vector DB, and retrieved from the CSV Retrieval pipeline
Remote Sensing Data Before and After California Rim and King Forest Fires, 2010–2015 [36]Satellite and thermal imagesHigh-resolution surface reflectance, thermal imagery, burn severity metrics, and LiDAR-derived structural measures from Sierra Nevada Mountains, California, USA, collected before/after 2013 Rim and 2014 King fires.Provides high-resolution multi-spectral and thermal imagery, indexed in vector DB and utilized in image retrieval pipeline.
National USFS Fire Occurrence Point [37]TabularIgnition points for USFS wildland fires, maintained at Forest/District level to track occurrence and origin.Provides historical US wildfire data for VQA to understand fire patterns and geographical risks.
FLAME 3 - Radiometric Thermal UAV Imagery for Wildfire Management [35]UAV images/thermal622 image quartets labeled Fire and 116 labeled No Fire from the surrounding forestry of the prescribed burn plot.Gives RGB-thermal image pairs for multimodal fusion algorithms.
Table 5. Orchestrator Decision Logic Matrix.
Table 5. Orchestrator Decision Logic Matrix.
Query ConditionDominant Weight FactorSelected AgentExecution Action Order
System Initialization (e.g., “New User Query q Received”)Policy Initialization ( π ( q , C 0 ) ) Orchestrator evaluates: is β (Data Need) or α (Intent Match) the priority?Orchestrator Agent1. Parse semantic intent of query q.
2. Instantiate dynamic weights ( α , β ).
3. Determine initial a t a r g e t .
4. Route control to the selected agent (typically Data Acquisition for complex queries).
Data Retrieval (External Context Required)High Data Need ( β > α )Data Acquisition Agent1. Execute RAG pipelines.
2. Update global context ( C t ).
3. Data validation: If data is null, re-invoke Data Acquisition Agent immediately.
4. Once data is sufficient, decay β ( β β · λ d e c a y ) to shift control.
Context Reasoning Sufficient Data AvailableHigh Intent Match ( α > β ) Prioritizes logic synthesis and answer generation.Reasoning Agent1. Ingest grounded context ( C t ).
2. Perform Chain-of-Thought analysis.
3. Draft response (R).
4. Submit R to Orchestrator Agent.
Ambiguous/Conflict (e.g., Conflicting visual vs. textual data)Recursive Correction Orchestrator detects low confidence or format violation.Recursive Loop1. Trigger Supervisor–Worker Protocol
2. Re-invoke subordinate agent with refined constraints
3. Filter Hallucinations
Table 6. A comparison between Task-based and Orchestrator-based approaches.
Table 6. A comparison between Task-based and Orchestrator-based approaches.
AbilityTask-Based ApproachOrchestrator-Based Approach
Dynamic task planningNoYes
No of crew neededSingleMultiple
Agent invocationSequential, fixedFlexible, orchestrated
MaintainabilityDirect, less flexibleModular, more maintainable
Task execution flowCrew executes tasksOrchestrator agent manages flow
Use case suitabilitySimple, linear flowsComplex, adaptive workflows
Table 7. Quantitative performance comparison of OpenAI and Gemini models across wildfire detection tasks.
Table 7. Quantitative performance comparison of OpenAI and Gemini models across wildfire detection tasks.
ModelAccuracyPrecisionRecallF1-ScoreInference Time (s)
GPT-4.1-Mini0.7000.7000.7000.700213.02
GPT-4.1-Nano0.6320.6620.6320.647160.21
GPT-4o0.7000.7520.7000.725190.57
GPT-50.4500.6040.4500.515155.25
GPT-5-Nano0.6840.7970.6840.736176.70
Gemini-2.5-Flash-Lite0.6000.6620.6000.629176.70
Table 8. Ablation results with the selected GPT-5-Nano LMM.
Table 8. Ablation results with the selected GPT-5-Nano LMM.
Ablation TypeAccuracyPrecisionRecallF1-ScoreInference Time (s)
Without CSV Data Retrieval tool (Only Image Data Retrieval tool and Multimodal RAG tool)0.6670.7220.6670.69394.50
Without Image Data Retrieval tool (Only CSV Data Retrieval tool and Multimodal RAG tool)0.6430.6510.6480.649107.96
Without Multimodal RAG tool (Only Image Data Retrieval tool and CSV Data Retrieval tool)0.4710.5690.4710.515116.94
With CSV Data Retrieval tool, Image Data Retrieval tool and Multimodal RAG tool0.6840.7970.6840.736176.70
Table 9. Functional comparison of the proposed MAS against existing approaches.
Table 9. Functional comparison of the proposed MAS against existing approaches.
ApproachPerformanceDetection SpeedMonitoring CapabilityManagement SupportScalability
WSN Fire Controller (2018) [8]N/A (real-time alerts)Real-time (WSN)Sensor-based variablesRisk alertsHigh (multi-hop routing)
Edge-UAV System (2018) [33]N/A (qualitative efficient management of CPU/RAM, battery life, and network resources based on initial experiments)Real-time (edge/fog)UAV detectionResource allocationHigh (hierarchical)
MARL-based Systems (2020) [19]Learning convergenceReal-time inferenceSpatial coverageMonitoringMedium (dataset dependent)
SmokeyNet (2022) [14]Accuracy: 83.49%Real-time inferenceImageSmoke DetectionMedium (dataset dependent)
Fuzzy Fire Mapping (2024) [7]AUC 0.879N/A (static mapping)Susceptibility via GISPrevention actionsMedium (climate-sensitive)
UAV Swarms for WER (2025) [6]N/A (conceptual)N/A (conceptual)Real-time via UAVsEvacuation/
Suppression
High (swarm-based)
High-Level MAS with DRL (2025) [20]N/A (conceptual)Real-time (DRL algorithms)UAV/IoT trackingDSS for decision-makingMedium (integrated data)
Heterogeneous MAS (2025) [18]Finite-time convergence (simulations)Finite-time trackingCooperative air-groundFault-tolerant trackingHigh (heterogeneous agents)
WildfireGPT (2025) [15]Correctness: 97.73% (case studies)Real-time inference (LLM-based)Data synthesis (climate projections/literature)Risk insights/
decision-making
High (LLM scalable)
Proposed MASPrecision 0.797, F1-score 0.736Low latency (orchestration)Multimodal synthesisContext-Aware ReasoningHigh (agentic scalability)
Table 10. Sensitivity analysis of Positional Drift: Comparing theoretical spatial error between a Standard Non-Predictive Baseline and the proposed look-ahead mechanism across a spectrum of wind intensities.
Table 10. Sensitivity analysis of Positional Drift: Comparing theoretical spatial error between a Standard Non-Predictive Baseline and the proposed look-ahead mechanism across a spectrum of wind intensities.
Wind Speed (km/h)Est. Rate of Spread (ROS) [m/min] 1Inference Latency ( Δ t ) [sec]Spatial Error (Baseline) [m] 2Spatial Error (Ours) [m]
20≈33176.7097.1Residual Variance
40≈66176.70194.3Residual Variance
60 (Benchmark)≈100176.70294.5Residual Variance
80≈133176.70391.6Residual Variance
1 ROS values are parameterized using a conservative wind-coupling ratio ( R O S 0.1 · V w i n d ) to simulate high-intensity propagation. This “10% Rule of Thumb” is a standard operational heuristic for shrublands [42], consistent with the wind-driven physics described in the Rothermel model (p. 14, [43]). 2 Calculated as linear displacement ( d = R O S × Δ t ) for a system without temporal compensation.
Table 11. Architectural comparison highlighting the shift from static inference in static models to dynamic, context-aware reasoning in the proposed orchestrator-based MAS.
Table 11. Architectural comparison highlighting the shift from static inference in static models to dynamic, context-aware reasoning in the proposed orchestrator-based MAS.
ApproachArchitectureData Handling StrategyReasoning ProcessPerformance Metrics
WSN Fire Controller (2018) [8]WSN with fuzzy controllerSensor (meteorological/gases)Fuzzy logic for alertsN/A (real-time alerts)
Edge-UAV System (2018) [33]Edge/fog/cloud hierarchyUAV sensor dataDynamic allocationN/A (qualitative efficient management of CPU/RAM, battery life, and network resources based on initial experiments)
MARL-based Systems (2020) [19]MARL networksContinuous 2D position states and movement actionsReinforcement learning with centralized trainingConvergence in 5000 episodes (CR-MARL)
SmokeyNet (2022) [14]Single-model CNNsStatic sequential image frames from fixed camerasFixed inference/Binary classificationPrecision: 89.84%, Recall: 76.45%, F1-score: 82.59%, Accuracy: 83.49%
Fuzzy Fire Mapping (2024) [7]Fuzzy inference with GISRemote sensing (temp/rainfall)Rule-based susceptibilityAUC 0.879
UAV Swarms for WER (2025) [6]Systems engineering with swarmsMultimodal (sensors/UAVs)Collaborative self organizationN/A (conceptual)
High-Level MAS with DRL (2025) [20]Hierarchical MASIntegrated historical/real-timeDRL for tracking/estimationN/A (conceptual)
Heterogeneous MAS (2025) [18]Fault tolerant formation controlSensor inputs for trackingFO-NFTSM/HOSMOFinite-time convergence
WildfireGPT (2025) [15]LLM Agent with RAG frameworkMulti-modal data sourcesMulti-round conversational reasoningCorrectness: 97.73%, Relevance: 98.20%, Entailment: 93.75%, Accessibility: 95.49%
Proposed Orchestrator- Based MAS (Ours)Orchestrator based with LMM/RAGDynamic multimodal (text/image)Context-aware via agentsPrecision 0.797, F1-score 0.736 (Section 5, Table 7)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sandeep, A.; Jayarathna, S.; Sandaruwan, S.; Samarappuli, V.; Meedeniya, D.; Perera, C. Context-Aware Multi-Agent Architecture for Wildfire Insights. Sensors 2026, 26, 1070. https://doi.org/10.3390/s26031070

AMA Style

Sandeep A, Jayarathna S, Sandaruwan S, Samarappuli V, Meedeniya D, Perera C. Context-Aware Multi-Agent Architecture for Wildfire Insights. Sensors. 2026; 26(3):1070. https://doi.org/10.3390/s26031070

Chicago/Turabian Style

Sandeep, Ashen, Sithum Jayarathna, Sunera Sandaruwan, Venura Samarappuli, Dulani Meedeniya, and Charith Perera. 2026. "Context-Aware Multi-Agent Architecture for Wildfire Insights" Sensors 26, no. 3: 1070. https://doi.org/10.3390/s26031070

APA Style

Sandeep, A., Jayarathna, S., Sandaruwan, S., Samarappuli, V., Meedeniya, D., & Perera, C. (2026). Context-Aware Multi-Agent Architecture for Wildfire Insights. Sensors, 26(3), 1070. https://doi.org/10.3390/s26031070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop