Next Article in Journal
An Investigation of the Process of Risk Coupling and the Main Elements of Coal-Mine Gas-Explosion Risk
Previous Article in Journal
Critical Criterion for Flame Spread Acceleration of Concave Surfaces with Different Curvatures
Previous Article in Special Issue
Applying Systems Thinking Concepts to Major Casualty Fires: Lessons Learned from Taiwan
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Generative AI as a Pillar for Predicting 2D and 3D Wildfire Spread: Beyond Physics-Based Models and Traditional Deep Learning

1
Geospatial Research Innovation Development (GRID), School of Built Environment, University of New South Wales (UNSW), Sydney, NSW 2052, Australia
2
School of Minerals and Energy Resources Engineering, University of New South Wales (UNSW), Sydney, NSW 2052, Australia
*
Author to whom correspondence should be addressed.
Fire 2025, 8(8), 293; https://doi.org/10.3390/fire8080293
Submission received: 3 June 2025 / Revised: 15 July 2025 / Accepted: 17 July 2025 / Published: 24 July 2025
(This article belongs to the Special Issue Fire Risk Assessment and Emergency Evacuation)

Abstract

Wildfires increasingly threaten human life, ecosystems, and infrastructure, with events like the 2025 Palisades and Eaton fires in Los Angeles County underscoring the urgent need for more advanced prediction frameworks. Existing physics-based and deep-learning models struggle to capture dynamic wildfire spread across both 2D and 3D domains, especially when incorporating real-time, multimodal geospatial data. This paper explores how generative artificial intelligence (AI) models—such as GANs, VAEs, and transformers—can serve as transformative tools for wildfire prediction and simulation. These models offer superior capabilities in managing uncertainty, integrating multimodal inputs, and generating realistic, scalable wildfire scenarios. We adopt a new paradigm that leverages large language models (LLMs) for literature synthesis, classification, and knowledge extraction, conducting a systematic review of recent studies applying generative AI to fire prediction and monitoring. We highlight how generative approaches uniquely address challenges faced by traditional simulation and deep-learning methods. Finally, we outline five key future directions for generative AI in wildfire management, including unified multimodal modeling of 2D and 3D dynamics, agentic AI systems and chatbots for decision intelligence, and real-time scenario generation on mobile devices, along with a discussion of critical challenges. Our findings advocate for a paradigm shift toward multimodal generative frameworks to support proactive, data-informed wildfire response.

1. Introduction

Wildfires and bushfires have emerged as one of the most destructive natural disasters of the 21st century, leaving a devastating trail across natural ecosystems, agricultural lands, and densely populated urban regions [1,2,3]. These fast-moving fires not only cause immense structural damage and loss of human life but also result in widespread economic disruption and long-term environmental degradation [4,5,6]. The release of smoke, fine particulate matter, and toxic gases during large-scale bushfires contributes significantly to air pollution, impacting public health and exacerbating climate change in cities across the globe [7,8]. Striking examples of this escalating crisis include the 2025 Los Angeles wildfire season—when the Palisades and Eaton fires swept through Southern California—and Australia’s 2019–2020 “Black Summer” fires [9,10,11,12]. The Los Angeles wildfire resulted in estimated economic losses exceeding USD 250 billion, destroyed thousands of homes, and displaced entire communities, causing profound physical and emotional distress among the affected population [13,14]. These events underscore the urgent need for accurate and timely wildfire forecasting systems. In particular, real-time bushfire propagation simulation-capable of predicting fire-spread pathways and identifying at-risk zones-plays a critical role in supporting firefighting operations, emergency evacuation planning, and long-term fire hazard mitigation strategies [15,16,17,18].
A variety of wildfire propagation models have been developed, each rooted in distinct computational approaches. Physics-based models (e.g., FARSITE, SPARK, Prometheus) simulate fire spread using physical laws of combustion, heat transfer, fuel conditions, and wind dynamics [19,20,21,22,23], offering detailed outputs but requiring extensive computation and environmental inputs—often limiting their use in real-time scenarios [24]. Empirical models, like McArthur’s Fire Danger Index, rely on historical fire behavior to produce rapid, region-specific predictions, but struggle with generalizability across diverse ecosystems [25]. Traditional machine learning models—such as decision trees, random forests (RF), and support vector machines (SVM)—have been widely applied for fire ignition prediction and risk classification [26,27,28,29]. However, their capacity to simulate dynamic fire progression remains limited.
In contrast, deep-learning models have recently gained momentum for 2D fire-spread forecasting, due to their ability to capture complex spatio-temporal relationships in fire behavior [30,31]. Architectures such as convolutional neural networks (CNNs), recurrent neural networks (RNNs), and U-Net variants have been applied to wildfire segmentation and short-term spread prediction, leveraging remote sensing, meteorological, and topographic data [32,33,34,35]. These deep-learning models (e.g., CNNs and RNNs) often outperform traditional physics-based simulators in runtime, scalability, and—in data-rich environments—predictive accuracy, as they learn complex patterns and variability directly from empirical observations rather than relying on explicit physical formulations [36,37]. However, deep-learning-based fire models still face key limitations that prevent them from enabling more advanced fire prediction capabilities with higher spatial and temporal resolution in real time. Most models remain constrained to 2D spatial domains [30] and lack integration and real-time augmentation of multimodal inputs, such as meteorological, topographic, and vegetation (fuel) data [36,38], which often change dynamically as fire propagates across the landscape. In addition, many existing models cannot simulate vertical fire dynamics, limiting their effectiveness during fast-evolving wildfire events [39,40].
Building on this trajectory, emerging generative AI models—such as variational autoencoders (VAEs), generative adversarial networks (GANs), and transformers—offer substantial advantages over earlier deep-learning architectures like CNNs and RNNs, particularly for complex environmental and urban modeling tasks [41,42,43]. These models already power a wide range of applications in everyday software for personal use, content creation, and entertainment, including image synthesis, text generation, music composition, and interactive conversational systems, where they deliver capabilities far beyond those of conventional deep-learning approaches [44]. Unlike traditional models that primarily specialize in classification or regression, generative AI systems learn intricate data distributions to produce entirely new, coherent, and contextually appropriate outputs [45]. This demonstrated ability to generate high-fidelity, multimodal content has sparked growing interest in adapting generative techniques to scientific and environmental domains, where similar benefits—such as robust uncertainty handling, scenario simulation, and the integration of heterogeneous data—hold significant transformative potential.
Motivated by this potential, this paper investigates how the convergence of generative AI techniques can drive next-generation bushfire prediction systems. We begin by conducting a systematic review of recent studies that apply generative AI to fire prediction and monitoring, utilizing a new paradigm that leverages large language models (LLMs) for literature synthesis, classification, and knowledge extraction. Through this review, we examine how these technologies support high-fidelity spatial modeling and improved scalability and accuracy, capabilities that are critical for enhancing situational awareness and decision making during fire emergencies [46,47]. We then outline five key future directions for advancing generative AI in wildfire management: developing unified multimodal frameworks that seamlessly integrate 2D and 3D data; designing conversational agentic AI systems for interactive, real-time wildfire intelligence; training interdisciplinary AI foundation models; enabling edge-based scenario generation on mobile and IoT devices; and advancing explainable AI interfaces to improve transparency and trust, accompanied by a discussion of critical challenges.

2. Previous Reviews in Fire-Spread Management

Several previous studies have conducted comprehensive literature reviews on predicting bushfire spread using traditional methods, including simulation techniques, machine learning, and deep-learning. While these approaches are not the primary focus of this paper—which centers on generative AI applications—they provide essential background on conventional fire-spread prediction and its underlying scientific rationale. This section begins by summarizing these reviews, highlighting how both simulation models and deep-learning approaches have been applied to fire propagation prediction. The overview of previous review articles aims to provide a comprehensive understanding of the existing AI applications in wildfire-spread prediction and to identify gaps where generative AI models can be leveraged to address challenges faced by traditional simulation and deep-learning approaches.

2.1. Wildfire Simulation Models and Traditional Machine Learning

Wildfire simulation models are traditional, well-established tools that simulate and predict bushfire spread by solving physical and empirical equations based on fuel, weather, and terrain conditions—originating from foundational models like Rothermel’s in the 1970s [48,49]—and they have long served as the backbone of operational fire management and planning systems. Based on their underlying rationale, these models can be classified into several categories, as presented in Table 1, and are summarized through multiple previous studies [49,50].
On the application level, a recent study conducts a comprehensive review of empirical and dynamic wildfire simulation models, focusing on their applications in predicting bushfire and wildfire spread across Australia [51]. It critically evaluates a suite of simulation systems including PHOENIX Rapidfire, SPARK, AUSTRALIS, REDEYE, IGNITE, and SiroFire, each leveraging distinct modeling techniques to simulate fire behavior, together with a variety of traditional machine-learning and deep-learning methods for wildfire prediction. PHOENIX Rapidfire, a deterministic simulator, uses a fire characterization model based on Huygen’s principle to predict spread, flame height, and intensity. While it excels in speed and ease of use, it is limited by reliance on predefined behavior models and underestimation of irregular fire shapes [52,53]. SPARK provides flexible, modular simulation driven by user-defined spread models and integrates well with GIS platforms, though its high computational demands pose practical constraints. AUSTRALIS employs a discrete-event simulation based on empirically derived rate-of-spread models; despite its efficiency, it underperforms in severe fire conditions [21]. REDEYE and IGNITE integrate geospatial and real-time data for risk prediction and hotspot mapping, respectively, yet their effectiveness is limited by data compatibility and platform-specific requirements. SiroFire models dynamic weather components and supports strategic planning but lacks adaptability for diverse fuel types [51]. Collectively, these simulators provide valuable predictive capabilities; however, each exhibits distinct limitations, ranging from computational inefficiencies and limited model adaptability to reduced accuracy under extreme conditions. This review highlights the need for hybrid modeling approaches that integrate traditional simulation techniques with machine learning and real-time data to enhance predictive accuracy, reduce simulation runtime, and strengthen operational resilience.

2.2. Deep Learning in Wildfire Prediction

Several prior studies have conducted comprehensive reviews of machine-learning models, including deep-learning approaches, for wildfire prediction. In this work, we concentrate on the most recent literature review to provide an up-to-date overview of the deep-learning models utilized and their respective application domains. As the application of machine learning and deep learning to wildfire prediction has already been extensively discussed and explored in numerous review papers over the past decades, this paper does not repeat those reviews. Instead, we focus specifically on reviewing, examining, and summarizing studies that apply emerging generative AI models, an area that has not been comprehensively covered in previous reviews.
In the following subsection, we leverage a LLM to generate a literature summary and bibliographic visualizations that highlight the evolving trends in deep-learning applications over time. Jain et al. [26] present a comprehensive scoping review of machine-learning (ML) applications in wildfire science and management, analyzing 300 studies published up to the end of 2019. ML usage is categorized across six core domains: fuel characterization, fire detection, climate interactions, risk prediction, fire behavior, and post-fire effects. Among modern approaches, deep-learning (DL) algorithms have shown strong potential for modeling wildfire spread, particularly due to their capacity to handle high-dimensional and multivariate data. CNNs are widely used for spatial fire detection and smoke recognition from imagery, while long short-term memory (LSTM) networks incorporate temporal dynamics to model fire growth. Deep neural networks (DNNs) have also been applied to map burned areas and forecast fire spread, often outperforming traditional models when sufficient annotated data are available. Despite their promise, the review emphasizes the importance of domain expertise and high-quality data for meaningful DL integration in operational forecasting.
A recent study systematically reviews DL applications in forest fire prediction, analyzing 55 key publications from 2017 to 2024 [54]. The review emphasizes the transformative role of deep learning in capturing complex spatio-temporal patterns associated with fire ignition and spread, offering significant advantages over traditional machine-learning approaches. CNNs are effectively used for analyzing satellite imagery to detect fire-prone areas and assess post-fire damage. LSTM and gated recurrent unit (GRU) networks model time-series data to forecast fire progression based on historical environmental conditions. Emerging generative AI models such as GANs have been used for synthetic data generation, while MLPs and U-Net architectures support fire risk estimation and boundary segmentation. Despite promising results, DL models face challenges such as data heterogeneity, limited inclusion of human activity factors, generalization across geographies, and the scarcity of high-quality labeled datasets. These limitations highlight the need for integrating diverse data sources and establishing standardized evaluation protocols to ensure model reliability in real-world wildfire prediction scenarios.

2.3. Limitations of the Existing Deep-Learning Applications in Wildfire Prediction

Based on a comprehensive review of existing literature at the intersection of deep learning and wildfire prediction, as well as an analysis of the inherent challenges and limitations of various deep-learning models stemming from their underlying mechanisms and data-driven assumptions, we identify several critical knowledge gaps and challenges that constrain the effectiveness and broader applicability of current deep-learning approaches in wildfire prediction.
L1. 
Quantification of Limited Uncertainty: Traditional models such as CNNs and RNNs typically produce deterministic outputs and struggle to quantify prediction uncertainty, an essential capability for wildfire management and emergency response applications involving stochastic environmental processes, such as abrupt changes in wind speed and directions [55,56].
L2. 
Weak Long-Term Dependency Modeling: RNNs and DNNs often struggle with vanishing gradients and limited temporal memory, which undermines their ability to capture long-range dependencies. This limitation makes them less effective in modeling the temporal progression of wildfires and in representing the long-term variability of underlying environmental processes such as fuel accumulation, climate patterns, and vegetation dynamics [57,58].
L3. 
Inadequate Multimodal Data Integration and Prediction: Traditional deep learning architectures are not inherently designed to fuse multimodal geospatial data sources, such as 2D GIS data (e.g., satellite imagery, fuel load maps, meteorological layers), 2.5D data (e.g., point clouds, digital surface models), and 3D GIS and Building Information Modeling (BIM) data (e.g., building models) [59]. As a result, generating multimodal fire-spread prediction outputs across different dimensions—from 1D time series to 3D spatial representations—within a unified deep-learning framework remains a significant challenge.
L4. 
Limited Data Augmentation Capabilities: Traditional deep-learning models, such as CNNs and RNNs, typically achieve strong performance only when trained on abundant, high-quality labeled data. They depend heavily on large volumes of annotated samples, which are often scarce in wildfire prediction tasks due to the rare, unpredictable, and spatially heterogeneous nature of fire events [32,39]. However, many of these models lack the capability to perform data augmentation or generate synthetic training samples, limiting their predictive accuracy and generalizability in data-sparse or unseen regions. This limitation poses a significant challenge for developing robust wildfire.
L5. 
Missing Data and Poor Data Quality Challenges: Environmental datasets frequently contain missing or incomplete information caused by sensor malfunctions, occlusions, or data transmission failures [60]. Such data gaps hinder accurate modeling and prediction. Traditional deep-learning models often assume complete input data or rely on simplistic imputation methods that fail to capture the underlying spatio-temporal dependencies critical for wildfire dynamics. These limitations reduce model robustness and prediction accuracy in real-world scenarios where data are inherently noisy or sparse [61].
L6. 
Lack of Explainability: Deep-learning models such as CNNs and RNNs often operate as “black boxes,” providing limited insight into how predictions are made and offering little transparency or trustworthiness [62,63].
While previous studies have provided structured and comprehensive reviews, most recent research remains focused on traditional machine-learning and deep-learning methods, or at most covers only a specific type of generative AI model for wildfire prediction. In contrast, comprehensive reviews and in-depth discussions that encompass the broader spectrum of emerging generative AI models, along with explanations of their underlying algorithmic principles in this domain, remain scarce and largely underexplored.
In the following section, we examine recent emerging studies that apply generative AI models to wildfire-spread prediction, discussing the underlying rationale for using generative approaches and highlighting their advantages over traditional deep-learning methods in addressing the aforementioned limitations. Our discussion then expands to explore how popular generative AI models can outperform conventional deep-learning approaches in certain predictive analytics tasks. Ultimately, we aim to offer new perspectives on the potential of generative AI in next-generation emergency response systems, enabling faster, more reliable, and higher-resolution predictions of wildfire spread.

3. Emerging Generative AI Models and Their Advantages

Since around 2018 and especially in the early 2020s, generative AI models have experienced rapid advancements and widespread adoption across diverse fields, marking a significant shift in the landscape of artificial intelligence [64]. Broadly defined, generative AI refers to a class of machine-learning models that are capable of learning the underlying distribution of data and generating new, realistic content—such as text, images, audio, or even simulations—that resembles the training data [65,66,67]. These models and their architectures have been increasingly experimented with and adopted to solve complex problems across various scientific disciplines.

3.1. Generative AI Applications in Environmental and Urban Sciences

The recent emerging generative AI models go beyond traditional predictive analytics by creating novel outputs, making them highly valuable in data-scarce environments or in domains requiring scenario generation, synthesis, or simulation. Generative AI encompasses several prominent categories, including GANs, VAEs, autoregressive models, diffusion models, and flow-based models [64,66,68]. Within these categories, more specialized types have emerged: GANs consist of a generator and discriminator in adversarial training to synthesize realistic data [65]; VAEs combine probabilistic encoding and decoding for efficient latent space learning [66]; autoregressive models like GPT-4 predict future tokens in sequences for high-quality text generation [64]; diffusion models such as stable diffusion and denoising diffusion probabilistic models (DDPMs) generate content by gradually denoising random noise into coherent samples [67,69]; and flow-based models leverage invertible transformations for exact likelihood estimation and sample generation [70].
Each generative AI model and its architecture offers distinct mathematical properties and advantages depending on the application context. The rise of these models has revolutionized data science and AI by enabling machines to not only interpret and predict but also to imagine and create, thereby unlocking new opportunities across fields including hazard prediction, urban planning, climate science, decision support, design automation, and digital twin construction [47,71]. In the environmental hazard domain, [47] presents a comprehensive review of how generative AI addresses longstanding challenges in data availability, quality, and resolution. These models learn high-dimensional probability distributions from limited samples, making them ideal for data-scarce applications such as geohazards, hydrometeorology, and climate-related analysis. By generating physically consistent synthetic data—ranging from downscaled meteorological fields to simulated landslide and seismic imagery—generative AI enhances forecasting accuracy, susceptibility mapping, and early warning systems, driving more reliable and scalable hazard modeling frameworks.
In the urban management sector, [71] conducts a scoping review on integrating generative AI into urban digital twins, highlighting its transformative role in automating the generation of high-quality urban data, hypothetical planning scenarios, and 3D city models. These capabilities help overcome major challenges related to data sparsity, simulation scalability, and design complexity. Through the synthesis of multi-modal urban data and simulation of complex dynamics, generative AI enhances real-time decision support, predictive analytics, and participatory planning across sectors such as transportation, energy, water, and infrastructure. The convergence of generative AI with digital twin technologies represents a paradigm shift toward intelligent, adaptive, and inclusive urban solutions. Furthermore, growing research extends generative AI applications into other urban subsystems, including logistics optimization, intelligent transportation systems, and domain-specific knowledge generation [72,73,74].

3.2. Theoretical Foundations of Fire-Spread Modeling for Generative AI Integration

Generative AI provides powerful computational tools to simulate, model, and predict natural phenomena, but it still relies on existing physical laws and fundamental principles. To explore how emerging generative AI architectures can enhance or redefine current wildfire-spread simulations and their core processes, it is crucial to first revisit the fundamental principles underlying fire-spread modeling. At their core, most wildfire-spread simulators are built upon two foundational computational paradigms: Huygens’ principle-based wavefront expansion methods and grid-based local interaction models. Each embodies a distinct conceptual strategy to replicate the spatio-temporal dynamics of fire growth, offering complementary strengths and insights for different use cases [75].
Huygens’ principle-based methods simulate wildfire spread by borrowing ideas from wave physics, focusing on how the physical process propagates over time. They treat every point along the fire edge as a new ignition point that pushes the fire outward, often forming an ellipse shaped by wind, slope, and fuel [76]. This helps capture how fires stretch and move in different directions. Many widely used systems, like Prometheus, FARSITE, and FlamMap, rely on this approach to keep updating the fire boundary over time [22,77]. On the other hand, grid-based models focus on the characterization and connectivity of individual spatial or volumetric units created by dividing the landscape into a grid of small squares (2D cells) or cubes (3D voxels). This grid provides a spatial and temporal framework over which physical processes, such as wildfire, can propagate [23]. Each unit records its combustion state—unburned, burning, or burned—and captures local fuel and environmental characteristics. Fire spreads by following local transition rules influenced by fuel type, wind, slope, and stochastic effects [75]. Tools like Cell2Fire and hybrid models that combine grid methods with physics or machine learning build on this idea [78,79]. While often exploratory, grid-based models are valuable for studying detailed fire patterns, testing new hypotheses, and extending simulations into 3D to better capture fire behavior.
From a computational standpoint, generating large-scale grid-based models often requires substantial resources, including high RAM and GPU memory. Initializing large, connected grids in parallel environments can also incur significant time costs due to communication and load-balancing overhead. Moreover, both Huygens’ principle-based and grid-based methods typically depend on integrating multiple layers of GIS and meteorological data—such as topography, fuel attributes, and weather conditions—to achieve accurate physical or empirical simulations. Consequently, simulating wildfire spread over large areas or extended timeframes becomes highly resource-intensive, significantly increasing inference latency and constraining the operational feasibility of these approaches for rapid or wide-area fire-behavior forecasting.
To provide a clearer theoretical foundation regarding how these generative approaches can be integrated into the Huygens’ principle-based methods and the grid-based model, we first outline the core mathematical formulations that govern their learning objectives and elucidate their ability to capture both local and non-local fire-spread dynamics. Conceptually, VAEs, diffusion models, and transformers can be seamlessly integrated with these core fire-spread simulation principles by operating over localized kernels of neighboring cells surrounding active burning fronts. This architecture enables the models to learn spatial and temporal transitions of adjacent unburned cells, effectively guided by the propagation dynamics derived from physics-based simulation outputs or real-world data. Simultaneously, by embedding fuel characteristics and environmental covariates within their latent representations or as embeddings, these models capture the complex, multifactorial behavior underlying wildfire propagation, thus offering a more computationally efficient and rapid means to model and predict fire spread beyond traditional physical rules [80,81].
Conceptually, VAEs optimize a stochastic latent-variable model by maximizing the evidence lower bound (ELBO), which balances reconstruction accuracy against divergence from a prior distribution and thus enables learning complex distributions over fine-scale spatial fire-spread patterns. In contrast, diffusion models learn to reverse a gradual noising process through a denoising objective, allowing them to generate plausible next-timestep fire perimeters defined by Huygens’ principle and conditioned on environmental factors [82]. Meanwhile, transformers use self-attention mechanisms to dynamically capture long-range spatial and temporal dependencies, making them well-suited to learn how fires propagate across heterogeneous landscapes without explicitly encoding all physical interactions.
In the following section, we examine the generalized mathematical formulations of each generative AI architecture and explore, at a theoretical level, how they can be integrated with the core principles of fire-spread prediction. While the mathematical formulations may vary across different variants of each architecture, these variants are typically developed to meet specific modeling objectives or domain-specific requirements. Additionally, although GANs are widely used for generative tasks, we do not include them in this section due to the generality of their mathematical formulation—since both the generator and discriminator can adopt arbitrary architectures, there is no canonical structure to analyze in the context of fire-spread prediction.

3.2.1. VAEs

Diving into the details, VAEs optimize a stochastic latent-variable model by maximizing the ELBO, which is represented in the following equation that balances reconstruction fidelity with regularization toward a prior, enabling VAEs to have the potential ability learn distributions over plausible local fire-spread patterns [66].
L VAE = E q ϕ ( z | x ) log p θ ( x | z ) D KL q ϕ ( z | x ) p ( z ) ,
In the Equation (1), x represents the observed data—such as spatial snapshots of wildfire perimeters or voxel grids showing burned states—while z denotes latent variables that capture hidden influences like wind, fuel, and terrain effects on fire spread. The encoder q ϕ ( z | x ) , parameterized by ϕ , approximates the posterior distribution over these latent factors given the observed fire data, and the decoder p θ ( x | z ) , parameterized by θ , reconstructs or generates fire-spread patterns from them. The model optimizes the ELBO, balancing the reconstruction term E q ϕ ( z | x ) [ log p θ ( x | z ) ] , which ensures the decoded spread closely matches the observed data, and the KL divergence term D K L ( q ϕ ( z | x ) p ( z ) ) , which regularizes the latent space toward a simple prior p ( z ) . In wildfire applications, this enables VAEs to learn meaningful compressed representations of local fire dynamics and to generate realistic alternative spread scenarios that remain consistent with underlying physical and environmental processes.

3.2.2. Diffusion Models

Diffusion models learn to reverse a gradual noising process, often through a simplified denoising objective (as shown in Equation (2)), allowing them to have the ability to generate realistic next-step fire perimeters conditioned on environmental features [67].
L DM = E x , ϵ , t ϵ ϵ θ ( x t , t ) 2 ,
In this equation, x denotes the original data sample, which in wildfire modeling could represent a spatial map or voxel grid of a current fire perimeter or fuel state. The variable t indexes the timestep within the diffusion process, tracking the progression of noise addition or removal. Correspondingly, x t is the noisy version of the data at step t, simulating partially corrupted fire perimeter states. The term ϵ represents the actual noise sampled and applied during the forward diffusion process, while ϵ θ ( x t , t ) is the model’s prediction of this noise, parameterized by θ . By minimizing the difference between the true noise and the predicted noise across all samples, noise levels, and timesteps—captured by the expectation E x , ϵ , t —the model learns to effectively reverse the noising process. In wildfire applications, this can potentially enable the generation of realistic next-timestamp fire perimeters conditioned on environmental features, by iteratively refining uncertain or noisy predictions into plausible spread scenarios that align with learned data distributions and underlying physical patterns.

3.2.3. Transformers

Finally, transformers use self-attention to model long-range spatial or temporal dependencies (as depicted in Equation (3)), which enables dynamic context aggregation across heterogeneous landscapes or multi-scale temporal sequences [68].
Attention ( Q , K , V ) = softmax Q K d k V ,
In this equation, Q, K, and V represent the query, key, and value matrices, respectively. These are learned projections of the input data that capture different spatial or temporal features. In wildfire-spread modeling, for example, Q might encode the characteristics of a focal location or time step—such as current fuel moisture, wind conditions, or ignition state—while K and V capture information from surrounding spatial cells or earlier temporal states. The operation softmax ! Q K d k computes attention weights by measuring the similarity between the query and the keys, normalized by the dimension d k to ensure stable gradients. These attention weights are then used to compute a weighted sum of the values V, resulting in a context-aware representation that dynamically integrates long-range dependencies. In the context of fire spread, this mechanism could be incorporated into a grid-based framework to learn how a cell’s or voxel’s future state is influenced by conditions across heterogeneous landscapes and multi-step temporal interactions, effectively complementing classical Huygens’ principle and traditional grid-based simulators by automatically capturing complex, data-driven spatial and temporal relationships.

3.3. Advantage over Traditional Deep-Learning Models

Building on the limitations identified in the existing wildfire prediction literature, we conducted a comparative analysis of traditional deep-learning models and generative AI approaches. As summarized in Section 2.3, our findings highlight the key advantages of generative AI models in overcoming the limitations of conventional methods. These advantages are outlined below, with a detailed comparison presented in Table 2, which maps each limitation (L1–L6) discussed in the previous section to the corresponding strengths of generative AI.
  • Richer Uncertainty Modeling: In contrast to many traditional deep-learning models, generative AI models such as VAEs and diffusion models produce probabilistic outputs that inherently capture uncertainty in predictions—an essential capability for high-risk applications like wildfire forecasting [66,67]. This feature can be leveraged to address the limitation L1.
  • Better Long-Term Dependencies: Transformer-based architectures outperform RNNs in modeling long-range dependencies in sequential data, making them well-suited for tracking wildfire dynamics over extended time periods [64,68]. The capability could be harnessed to tackle the limitation L2.
  • Multimodal Data Fusion: Generative AI models, particularly those based on transformers and diffusion techniques, excel at integrating heterogeneous data sources (e.g., satellite imagery, meteorological variables, and point clouds), enabling more robust 2D and 3D wildfire forecasting through a unified framework [69,83] to address the limitation L3.
  • Data Augmentation and Synthesis: Generative AI models such as GANs and diffusion models can produce synthetic yet realistic wildfire progression data, providing valuable training samples in data-scarce scenarios, facilitating missing data imputation to enhance data quality, and supporting the simulation of extreme or hypothetical conditions [65,84]. This capability can be leveraged to address Limitations L4 and L5, as well as to enable scenario generation for simulating hypothetical wildfire events.
  • AI Explainability through Latent Space: Many generative AI models, such as VAEs and GANs, rely on latent spaces and latent vectors to operate, where complex data relationships are encoded in lower-dimensional representations [65,66]. These latent variables can be visualized and analyzed to enhance the explainability, interpretability, and trustworthiness of the models in tasks such as classification, prediction, clustering, and data generation [85], thereby addressing Limitation L6.
The unique advantages of generative AI models are systematically inferred from their architectural principles and algorithmic rationale, particularly in the context of data-driven wildfire prediction. Applications of these models often rely on standard environmental and urban GIS datasets—such as shapefiles, raster imagery, and 3D LiDAR point clouds—which are also widely used in generative AI research across other domains. In Section 5, we present a comprehensive and critical review of studies applying generative AI to wildfire prediction, highlighting their strengths for bushfire management. Our subsequent review of generative AI models aims to explore future opportunities for leveraging their strengths to revolutionize bushfire prediction, as well as to discuss the associated challenges. The insights gained from this review are presented in Section 3.2.

4. Review Strategy

To ensure methodological rigor and transparency, this study adopts a systematic review strategy combined with narrative synthesis. By systematically formulating search queries and applying them across well-established academic databases—including IEEE Xplore and Scopus—we identified and screened relevant literature in a structured, reproducible manner (as depicted in Figure 1). This approach enabled us to comprehensively map existing research on the application of artificial intelligence to wildfire-spread prediction, while also allowing us to iteratively narrow our scope to focus on the emerging role of generative AI models. Following initial retrieval and screening, we employed a narrative synthesis to qualitatively analyze and compare selected studies, examining their methodological frameworks, data sources, and reported predictive performances. This process provided nuanced insights into how diverse deep-learning and generative architectures have been leveraged to advance fire propagation modeling.
In addition to the identified articles, we employed an LLM-powered tool from our previous work [86] to characterize the literature by extracting taxonomies from each paper’s abstract using a combined process of scientific discourse tagging (SDT) and named entity recognition (NER). These taxonomies were then classified against an existing body of knowledge. Specifically, we used sentence transformers to align each article’s extracted taxonomy with domain knowledge by applying a cosine similarity threshold, comparing entities to definitions from a domain ontology built through comprehensive literature reviews and authoritative textbooks. A similarity threshold of 0.7 was chosen to ensure robust alignment, classifying papers under specific methodologies or application areas only when their extracted taxonomy showed high semantic relevance to existing definitions. For example, this process enabled classification of literature into application areas such as fire-spread prediction, detection and monitoring, and risk assessment, as illustrated in Figure 1.

5. Generative AI Applications in Wildfire Management

Through the following subsections, we present a structured, in-depth review of 11 existing studies that have explored the use of generative AI models for wildfire management. This review is organized according to the application areas shown in Figure 1, which were identified through keyword-based literature characterization. Our primary aim is to examine how generative AI has been applied to support fire prediction, while also highlighting related areas such as fire monitoring and risk mappin, as summarized in Figure 2. We focus specifically on evaluating the prediction accuracy, inference time, and computational efficiency of these generative AI models compared to traditional methods used as benchmarks.

5.1. Fire-Spread Prediction

We identified seven existing studies that have explored or prototyped the use of various generative AI models to support fire-spread prediction and fire-behavior modeling. These studies leverage three different architectures, namely GANs, VAEs, and generative transformers, to simulate wildfire dynamics, generate synthetic fire scenarios, or enhance predictive performance. Collectively, they demonstrate the emerging potential of generative AI in capturing the complex spatio-temporal patterns of wildfire propagation and improving the accuracy, scalability, and adaptability of fire-behavior forecasting systems.

5.1.1. GANs

GANs and their variants have proven valuable in environmental and urban research by generating synthetic data to augment limited and imbalanced datasets [71]. This capability is especially critical in wildfire research, where collecting real-world fire data is often constrained by safety risks, high costs, and logistical challenges. By simulating underrepresented fire scenarios—such as passive and active crown fires—GANs help improve the diversity and robustness of training data, leading to more accurate classification of fire behavior and better support for operational decision making. We identified two studies that effectively employ GANs for data augmentation to enhance wildfire-spread prediction through integration with machine-learning and simulation models. Sadegh et al. [87] presents a machine-learning framework for predicting wildfire-spread sustainability and crown fire occurrence in semiarid shrublands of southern Australia. The study uses a TGAN to generate synthetic fire records, addressing the challenge of small training datasets commonly encountered in this domain. The inclusion of synthetic data led to substantial performance improvements, increasing classification accuracy by up to 20% for spread sustainability and 4% for crown fire prediction. This enhanced dataset was used to train and evaluate several classifiers—including SVM, multilayered neural networks (MLP), and multinomial naive Bayes (MNB)—with the SVM model paired with TGAN-generated data achieving the highest accuracy. The results demonstrate the potential of combining generative AI with traditional supervised-learning models to improve generalization and reliability in data-scarce wildfire prediction settings. Trained on a dataset comprising 61 experimental fires in Australian semiarid shrublands—augmented by 27 TGAN-generated samples for spread sustainability and 13 for crown fire cases—the TGAN-enhanced SVM achieved 90% accuracy for predicting sustained fire propagation and 80% accuracy for active crown fire occurrence on an independent evaluation set of 29 fires. Compared to logistic regression models developed on the same data, this represented improvements of 15 percentage points for spread sustainability and 4 points for crown fire prediction, along with markedly higher sensitivities (1.0 vs. 0.77 and 0.55, respectively). Although the study did not report training or inference times, it clearly demonstrated that integrating TGAN-based data augmentation substantially boosts predictive performance and robustness relative to traditional statistical approaches, especially under data-limited conditions.
Sadegh et al. [88] extend this approach to Canadian conifer forests, focusing on predicting wildfire propagation types, including surface, passive crown, and active crown fires. In this study, TGAN is again employed to address class imbalance by generating synthetic data for underrepresented fire types, significantly boosting prediction accuracy and F1-scores—especially for difficult-to-classify categories like passive crown fires. The study evaluates four additional machine-learning models: two ensemble methods (Random Forest and XGBoost) and two AutoML approaches (TabPFN and AutoGluon), with TabPFN consistently achieving the best results when trained on GAN-augmented data. These findings reinforce the value of GAN-based data generation as a means of overcoming real-world data limitations and enhancing the predictive capabilities of wildfire-behavior models. Trained on a dataset of 113 experimental fire cases collected over 60 years in Canadian conifer forests, including 52 surface fires, 23 passive crown, and 38 active crown fires, the study demonstrated that using TGAN to generate synthetic records dramatically improved prediction performance. With balanced datasets created by adding up to 26 synthetic passive crown and 13 active crown fire samples, the best generative AI-related model, TabPFN, achieved 91% accuracy and a 93% F1-score on independent evaluation data in the binary classification task, outperforming traditional logistic regression baselines. In the more challenging multi-class setting (surface vs. passive crown vs. active crown), the addition of GAN data elevated TabPFN’s accuracy from 61% to 74% and its F1-score from 50% to 65%, particularly improving predictions of passive crown fires. Although exact training or inference times were not provided, the study clearly highlighted how TGAN-based data augmentation substantially enhances the spatial-temporal prediction fidelity of wildfire models under data-limited scenarios.
In summary, previous studies have shown that integrating GAN-based data augmentation can substantially outperform traditional simulation and machine-learning approaches in fire-spread prediction and behavior modeling [87,88]. By leveraging TGANs to generate synthetic wildfire records, these works effectively addressed class imbalance and limited data, achieving up to 90–91% accuracy and F1-scores exceeding 93%, notably outperforming logistic regression baselines. In more complex multi-class settings involving surface, passive crown, and active crown fires, GAN-augmented models improved accuracy by over 13 percentage points and F1-scores by 15 points, significantly enhancing the detection of underrepresented fire types. These findings highlight the strength of GAN-driven data generation in boosting the fidelity, robustness, and generalizability of wildfire forecasting models beyond what conventional statistical or standard deep-learning methods can achieve under data-scarce conditions.

5.1.2. VAEs

The use of VAEs and their variants is emerging in the wildfire prediction sector to generate realistic spatio-temporal simulations of fire spread, serving as synthetic training data for downstream forecasting models. These generative models are particularly effective for tasks that require the synthesis of physically consistent fire evolution sequences from high-dimensional, multi-modal inputs, such as topography, vegetation density, and weather conditions, where real data may be scarce, incomplete, or expensive to obtain through traditional physics-based simulations.
Cheng et al. [89] introduces a generative AI framework to overcome the computational demands of traditional physics-based wildfire simulations. Using a vector-quantized variational autoencoder (VQ-VAE) trained on cellular automata (CA) data, the model generates high-fidelity 3D sequences of wildfire spread that capture essential geophysical influences like vegetation and slope. This approach accelerates data generation by over four orders of magnitude compared to CA and MTT simulators. The synthetic wildfire scenarios are then used to train a POD-LSTM surrogate model, which, with the inclusion of VQ-VAE data, achieves notably higher prediction accuracy and structural similarity on both simulated and real events. This demonstrates the VQ-VAE’s effectiveness in reducing simulation costs, enriching training datasets, and improving the realism and generalizability of wildfire forecasts. Specifically, the VQ-VAE was trained on 40 CA-generated fire-spread sequences from the Chimney Fire region in California, each spanning 8 days (16 temporal snapshots) at 128 × 128 spatial resolution, and was then used to generate 500 additional synthetic fire scenarios. This resulted in surrogate models trained on VQ-VAE–augmented data achieving significantly lower relative RMSE and higher structural similarity index (SSIM) than models trained solely on the limited CA data, for both unseen synthetic fires and the real Chimney Fire event observed via MODIS and VIIRS satellites. In terms of computational efficiency, the VQ-VAE generated 8-day wildfire-spread sequences in just 0.3 s, representing a speed-up of 4–5 orders of magnitude compared to traditional CA and MTT simulators, effectively overcoming the major runtime bottlenecks of physics-based wildfire modeling while preserving crucial spatial-temporal dynamics.

5.1.3. Transformer

Transformers have emerged as powerful deep-learning architectures for wildfire prediction tasks, particularly in modeling fire spread, classifying risk levels, and generating predictive spatial outputs from complex spatio-temporal data. Their capacity to capture long-range dependencies and contextual interactions across diverse input modalities, such as satellite imagery, weather data, topography, and vegetation maps, makes them especially well-suited for wildfire forecasting tasks that require spatial precision and temporal consistency to support timely decision making and emergency responses. Li and Rad [90] present a transformer-based hybrid model—Attention Swin U-Net with Focal Modulation (ASUFM)—designed for next-day wildfire-spread forecasting across North America. ASUFM integrates spatial attention and focal modulation layers into a Swin transformer U-Net backbone, producing predictive fire masks from multivariate remote sensing data. Though not a conventional generative model, ASUFM simulates realistic fire-spread scenarios and addresses challenges such as spatial resolution, class imbalance, and temporal consistency. Trained on an expanded NDWS dataset (2012–2023), the model achieved state-of-the-art results in Dice score and PR-AUC, outperforming U-Net and other transformer variants. Additional techniques, including weighted loss functions, skip connections, and cosine learning rate scheduling, further enhanced the model’s generalization and accuracy. On the extended NDWS dataset covering over 31,000 samples from 2012–2023 across all of North America at 1 km spatial resolution, ASUFM achieved a Dice score of 0.4066, precision of 0.4345, recall of 0.4096, and a PR-AUC of 0.3974, substantially outperforming traditional encoder-decoder baselines such as U-Net (Dice 0.3493, PR-AUC 0.2945) and even pure Swin U-Net variants. This highlights the transformer’s superior ability to model long-range dependencies and capture complex spatial fire propagation patterns. Although the study did not report explicit training or inference times, it noted using large-scale GPU infrastructure (NVIDIA A100-80G or RTX A6000) with advanced techniques like focal modulation and spatial attention to balance precision-recall under heavy class imbalance. The expanded spatial and temporal coverage of the dataset ensures that ASUFM’s learned representations generalize across diverse geographic terrains and multi-year fire regimes, demonstrating its promise for continent-scale, next-day wildfire-spread prediction.
Deepa et al. [91] propose a contrastive-learning framework for early forest fire prediction that combines a contrastive vision transformer (CViT) with a pool former module. CViT functions as a powerful feature extractor via self-supervised contrastive learning and multi-head attention, while the pool former improves prediction efficiency by modeling spatial dependencies without heavy matrix operations. Though it does not employ traditional generative AI, the framework enhances feature robustness under variable environmental conditions through preprocessing and data augmentation strategies. The study utilized the publicly available Forest Fire Big Data dataset from Kaggle, comprising 1832 images of forest scenes under diverse environmental and weather conditions, covering both fire and no-fire instances. In terms of quantitative performance, the CViT-pool former method achieved 92.8% accuracy, 89.4% precision, 91.4% recall, and a 90.2% F1-score, clearly outperforming faster R-CNN (accuracy 89.0%, F1 88.9%) and ResNet (accuracy 87.4%, F1 84.5%). While the study did not report explicit training or inference times, it demonstrated that transformer-based contrastive learning significantly boosts predictive robustness and spatial discrimination compared to traditional deep CNN and RNN baselines.
Anne et al. [92] introduces a real-time wildfire prediction system that integrates a CNN–transformer hybrid model with a blockchain-based infrastructure for enhanced data security and traceability. The CNN extracts spatial features from drone imagery, while the transformer captures temporal patterns such as smoke visibility and fire direction. Although it does not use generative AI in the traditional sense, the system generates fire probability maps that simulate the spread and severity of wildfires, improving alert accuracy and resolution. It also incorporates a rule-based assistant for decision support and achieves 93.18% prediction accuracy, outperforming CNN and ResNet-42 models, albeit with a slightly higher training time. The system was trained and tested on forest fire images from Algeria’s Béjaïa region, comprising 540 training images and 100 test images, each at a spatial resolution of 244 × 244 pixels. In quantitative terms, the CNN–transformer model achieved an accuracy of 93.18%, precision of 91%, recall of 97%, and an F1-score of 94%, surpassing both standalone CNN and CNN ResNet-18 baselines. While it required a training time of approximately 30 min—slightly longer than the 20 min for CNN ResNet-18—this additional computational cost was offset by substantially higher predictive performance, particularly in modeling complex temporal features like fire evolution and direction. These results highlight the efficacy of integrating transformer-based temporal learning in enhancing the spatial-temporal resolution of wildfire risk maps.
In contrast, Li et al. [93] present a true generative AI approach through the development of Sim2Real-Fire, a large-scale, multi-modal dataset paired with the S2R-FireTr transformer model. S2R-FireTr forecasts and backtracks binary fire masks by learning from 1 million simulated wildfire sequences and generalizing to 1000 real-world cases. By leveraging cross-attention across five aligned modalities—topography, vegetation, fuel types, weather, and satellite imagery—it generates physically plausible fire-spread scenarios even under temporally incomplete conditions. Trained on 1 million multi-modal simulated wildfire scenarios with global coverage at 30 m spatial resolution, and evaluated on 1000 annotated real-world wildfire cases, S2R-FireTr achieved an AUPRC of 72.9%, F1-score of 69.6%, and IoU of 56.4% on real-world data. These results substantially outperformed traditional physical simulators—FARSITE (AUPRC 55.9%), WFDS (61.2%), and WRF-SFIRE (63.0%)—as well as leading deep-learning baselines like Rainformer and Earthformer. While exact training or inference times were not detailed, the approach demonstrated orders-of-magnitude efficiency gains by bypassing computationally intensive physics-based simulations, reinforcing its value for large-scale, real-time wildfire forecast and backtracking tasks.
Together, these studies pave the way for broader adoption of advanced transformer-based and generative AI methods, which have likewise shown remarkable improvements over traditional simulation tools and conventional deep learning in wildfire forecasting. For example, Ref. [90] reported that ASUFM, trained on 31,000 samples across North America, achieved superior Dice and PR-AUC metrics compared to U-Net and Swin U-Net, effectively modeling complex spatial fire dynamics at continental scale. Similarly, Ref. [91] found that a contrastive vision transformer pipeline outperformed CNN and RNN baselines, reaching over 92% accuracy and a 90% F1-score on diverse forest imagery. Ref. [92] further demonstrated that integrating CNNs with transformers achieved 93% accuracy and a 94% F1-score on Algerian wildfire data, capturing temporal evolution beyond what CNNs alone could model. Finally, Ref. [93] showed that a transformer trained on 1 million globally simulated wildfire scenarios attained AUPRC gains of 9–17 points over physics-based models like FARSITE and WFDS, enabling near real-time large-scale forecasting. Collectively, these results underscore the transformative advantages of GANs, VAEs, and transformers in enhancing the precision, scalability, and efficiency of wildfire-spread prediction and behavior modeling.

5.2. Wildfire Detection and Monitoring

In this review, we found that vision transformers have so far been applied primarily to support wildfire detection and monitoring, particularly for generating high-resolution segmentation masks and spatial representations of active fire regions. These models excel in tasks requiring fine-grained spatial understanding and temporal consistency, such as detecting smoke plumes and segmenting active fire zones, due to the visual complexity of input data (e.g., RGB UAV imagery, thermal maps) and the operational need for accurate, real-time predictions to support early firefighting efforts. Our review identified three notable studies that incorporate transformer-based architectures to facilitate wildfire detection and monitoring. Falcão et al. [94] focuse on early detection of wildfire smoke plumes using a deep-learning ensemble framework that integrates EfficientNetV2 (CNN), DeiT, and Swin TransformerV2 (both vision transformers). While generative models are not explicitly employed, the ensemble architecture—coupled with a neural network-based meta-classifier—demonstrates strong performance under challenging conditions like haze, fog, and low-contrast smoke. The study leverages transfer-learning and data augmentation techniques to improve generalization, achieving an average accuracy of 96.46% and an AUPRC of 95.14%, improving upon the best individual base model by approximately 2.1% in accuracy and 1.2% in AUPRC. Precision remained comparably high to the strongest base learner (EfficientNetV2), while recall and F1-scores demonstrated consistent gains. Although the study did not benchmark against traditional machine-learning or classical computer vision methods, it provided clear computational insights—training was performed via fine-tuning with transfer learning on a Google Colab platform using NVIDIA T4 GPUs, and inference time was measured at approximately 58 ms per batch of 16 images, compared to around 17 ms for individual models. This moderate increase in latency was justified by substantially enhanced robustness, allowing real-time deployment across hundreds of camera feeds operating at practical frame rates for wildfire monitoring.
Ghali et al. (2022) [95] target fire classification and segmentation using UAV imagery. Although it does not incorporate conventional generative AI, the framework produces fine-grained segmentation masks, effectively achieving a generative outcome. The authors develop an ensemble classifier with EfficientNet-B5 and DenseNet-201 for fire detection and employ three segmentation models—EfficientSeg, TransUNet, and TransFire—two of which are transformer-based. These models address complex challenges such as detecting small fires and delineating fire boundaries in noisy, cluttered backgrounds. The proposed transformer-based segmentation models demonstrated remarkable quantitative gains. TransUNet-R50-ViT achieved an accuracy and F1-score of 99.9%, outperforming the CNN-based U-Net by nearly 1 percentage point, and effectively capturing fine fire boundaries even in cluttered backgrounds. TransFire also excelled with an F1-score of 99.82%, confirming the advantage of vision transformers in extracting detailed fire regions. In terms of computational performance, TransUNet completed inference in 0.51 s per image, slightly higher than U-Net’s 0.29 s, while TransFire processed images at 1.0 s, reflecting a modest increase in latency for substantially improved segmentation fidelity. Although the study did not benchmark these models against traditional machine-learning approaches, it clearly demonstrated the superior precision and robustness of transformer-based architectures in high-resolution wildfire segmentation using UAV imagery.
Ghali et al. (2021) [96] explores the use of vision transformers—TransUNet and medical transformer (MedT)—to segment wildfire regions from RGB imagery for early detection and boundary delineation. These transformers perform generative tasks by producing binary segmentation masks that can be used to simulate wildfire-spread scenarios. The models address core challenges such as fine boundary detection, long-range dependency modeling, and feature misclassification. Trained on the CorsicanFire dataset, TransUNet and MedT achieve F1-scores of 97.7% and 96.0%, respectively, outperforming conventional architectures like U-Net, U2-Net, and EfficientSeg. Despite slightly longer inference times (1.2 s for TransUNet and 2.72 s for MedT), the trade-offs are justified by improvements in spatial granularity, generalization, and robustness under diverse environmental conditions. In terms of quantitative performance, TransUNet achieved an F1-score of 97.7% with its hybrid ResNet50-ViT backbone, while MedT reached 96.0%, each substantially exceeding the benchmarks set by U-Net (92.0%), U2-Net (82.9%), and EfficientSeg (94.3%). Precision and recall were jointly reflected in these high F1-scores, underscoring the transformers’ capability to minimize both false positives and false negatives in fire boundary segmentation. Although the study did not report training time in detail, inference times were noted at 1.2 s for TransUNet and 2.72 s for MedT, representing a moderate computational overhead compared to simpler models, yet yielding superior segmentation fidelity and finer detection of small or diffuse fire regions.
In terms of performance, transformer-based models have consistently outperformed traditional simulation tools and conventional deep-learning approaches in both fire-spread detection and boundary segmentation. For instance, Ref. [94] demonstrated that stacking vision transformers with CNNs boosted average accuracy to 96.46% and AUPRC to 95.14%, surpassing the best individual models by notable margins while maintaining practical inference times suitable for real-time monitoring. Similarly, Refs. [95,96] showed that transformer-based segmentation networks like TransUNet and TransFire achieved exceptional F1-scores up to 99.9%, outperforming CNN-based baselines such as U-Net by 5–15 percentage points, with only modest increases in inference time. These results highlight the superior ability of transformers to capture fine spatial details and long-range dependencies, enhancing the precision, robustness, and spatial fidelity of wildfire detection and boundary modeling beyond what traditional or purely CNN-based methods can achieve.

5.3. Wildfire Risk Mapping

Wildfire risk mapping plays a critical role in proactive fire management, enabling early warning systems, resource allocation, and strategic response planning across vulnerable landscapes. In this review, only one study using transformer-based method has been discovered.
One study Limber et al. [97] developed a residual transformer model to forecast wildfire potential across California by emulating the Wildland Fire Potential Index (WFPI), aiming to enhance short-term risk prediction and early warning for fire management. The model was trained to generate daily WFPI maps using Daymet meteorological data, MODIS-derived NDVI, and static fuel classifications from the Scott and Burgan fire models. As a generative forecasting emulator, it produces new spatial fire risk scenarios up to seven days ahead, addressing the high computational demands of physical models while ensuring rapid, spatially coherent forecasts. The residual connection leverages temporal autocorrelation in WFPI to boost short-term accuracy and stability.
As a performance summary, the model achieved spatial correlation coefficients of 0.85–0.98 across four weekly forecasts in July 2023, showing a slight tendency to underpredict extreme risk values. Precision remained high for one- and two-day forecasts, with modest degradation from days three to seven due to error propagation. Bayesian hyperparameter optimization and HPC resources enabled efficient tuning and training—requiring 42 h on 48 GPUs for tuning and 24 h on 24 GPUs for training—while full inference for four weeks of statewide daily forecasts was completed in just 6.5 min on a 128-core CPU. Although no direct comparisons were made to traditional or other ML models, the transformer substantially improved forecast speed and quality over existing USGS/USFS WFPI approaches, underscoring its promise for large-scale, data-driven wildfire risk prediction.

6. Discussion and Future Directions

Our targeted literature review identified and critically examined 11 studies that specifically explore the application of generative AI models in wildfire science. This relatively small number underscores that the integration of generative methods in this domain remains nascent, presenting substantial opportunities for further research and methodological innovation. Among the works reviewed, the predominant focus has been on leveraging GAN, VAE, and transformer architectures to enhance bushfire prediction capabilities [65,66,68]. Based on the theoretical foundation of diffusion models discussed in Section 3.2.2, we suggest that diffusion-based methods represent a promising future direction for fire-spread modeling, even though they have not yet been explored in current studies. These clear research gaps highlight a fertile area for advancing wildfire science through the development of more sophisticated, data-driven generative approaches capable of simulating complex spatio-temporal fire dynamics and supporting robust decision-making.
Building on the preceding review of current generative AI applications in wildfire modeling and prediction, our analysis of 11 studies shows that generative AI models can achieve accuracies around 90%, with some demonstrating computational performance exceeding traditional methods by several orders of magnitude. Combined with recent trends in LLM-powered agentic AI and distributed edge computing, these advantages open promising new research directions that remain largely unexplored. This section synthesizes key insights from the literature, outlines future avenues for investigation, and highlights major challenges along with potential strategies to address them.

6.1. Future Research Directions for Generative AI-Powered Wildfire Applications

This section outlines a series of forward-looking research directions for how the existing generative AI-powered fire-spread prediction applications can be further extended with more resaerch and development efforts to reshape wildfire prediction and management across scales and settings (as illustrated in Figure 3).

6.1.1. A Unified Simulation Framework for 2D and 3D Wildfire Dynamic

Previous studies have highlighted the importance of using multimodal data to improve bushfire prediction, while also noting the significant technical challenges involved in fusing heterogeneous datasets [31,34]. Traditional wildfire modeling has primarily relied on 2D spatial inputs—such as satellite imagery, temperature maps, and vegetation indices—even though fire behavior is inherently three-dimensional, shaped by terrain elevation, vertical fuel structures, atmospheric dynamics, and built environments. Conventional deep-learning approaches typically handle 2D and 3D data in isolation, requiring separate models for different modalities, which limits integration and adaptability.
Despite the experimental use of transformers and VAEs for fire-spread prediction in the 11 reviewed studies, current research has not explored their potential as multimodal generative frameworks for unifying diverse data sources. From a data science standpoint, architectures like VAEs, diffusion models, transformers are inherently well-suited to integrate heterogeneous inputs by learning joint representations across modalities. VAEs can encode geospatial, meteorological, and 3D structural information into shared latent spaces governed by probabilistic distributions, while transformers leverage self-attention to dynamically capture cross-modal and long-range dependencies [98,99]. This capability removes the need for explicit feature concatenation or separate fusion pipelines, enhancing scalability and allowing models to holistically capture the complex interplay of factors that drive wildfire propagation. By jointly learning how dynamic weather patterns interact with terrain and fuel characteristics, such approaches could improve predictive accuracy and enable scenario-based forecasting under varying meteorological conditions—critical for informed emergency response planning and efficient resource allocation. These considerations highlight a promising research agenda that leverages the unique strengths of multimodal generative AI to advance holistic wildfire simulation.

6.1.2. Chatbots for Wildfire Decision Intelligence

Recent trends across various scientific disciplines, including transportation and climate science, increasingly leverage the integration of LLMs with retrieval-augmented generation (RAG) to develop advanced conversational AI systems [73,100]. These multi-agent frameworks enable interactive, multi-turn dialogue that dynamically retrieves domain-specific data and generates context-aware, user-adaptive insights. Despite this growing popularity in other fields, such paradigms have yet to be explored for supporting wildfire decision intelligence.
This represents a compelling research direction. Future work could explore how conversational AI assistants—built on fine-tuned LLMs, RAG mechanisms, and an agentic AI paradigm—might transform wildfire prediction and management [73,101]. In such a framework, specialized agents could handle sub-tasks such as fire-spread simulation, atmospheric analysis, or evacuation planning, all coordinated by a dialogue manager to ensure coherent, user-centered interactions. RAG would ground the assistant’s responses in real-time wildfire data and scientific knowledge, drawing on sources like localized fire weather indices and historical ignition patterns.
By enabling natural language interaction, these AI assistants could empower urban planners, emergency responders, and community members to access relevant data, simulation outputs, and optimized decision solutions in an accessible format. This contrasts sharply with traditional systems that often rely on manual data processing and technical interfaces dense with domain-specific jargon. As an adaptive platform that learns continuously from expert feedback and evolving data streams, a wildfire-focused conversational AI system could facilitate more proactive, transparent, and inclusive approaches to fire risk assessment and emergency response. Advancing this vision represents an important future research agenda, leveraging the strengths of emerging LLM-RAG architectures to address longstanding challenges in wildfire decision support.

6.1.3. AI Foundation Model for Interdisciplinary Wildfire Management

The emergence of AI foundation models—large-scale, pre-trained systems typically built on transformer architectures—has become an increasingly popular trend across numerous scientific domains, including remote sensing, urban logistics optimization, and GIS [102,103]. These models serve as general-purpose backbones trained on diverse datasets to support a wide range of downstream tasks, enabling cross-domain knowledge transfer, zero-shot inference, and rapid fine-tuning. Despite their growing success in other disciplines, foundation models remain largely unexplored in the context of wildfire science.
This gap represents a promising direction for future research. A foundation model for wildfire management could leverage generative AI architectures—such as multimodal transformers, or diffusion models—to integrate heterogeneous datasets across spatial, temporal, and semantic dimensions. Trained on inputs like satellite imagery, LiDAR scans, weather time series, terrain maps, fuel assessments, and incident reports, such a model could generate predictive outputs and scenario-based simulations under varying conditions, supporting applications ranging from fire-spread forecasting to resource deployment planning.
Unlike traditional task-specific deep-learning pipelines that often rely on separate models for different data modalities—such as CNNs for 2D raster GIS data—an AI foundation model would offer a unified, adaptive, and scalable framework. By inherently capturing cross-modal relationships, such a model could robustly generalize across diverse geographic regions, fire regimes, and management objectives. Integrating multimodal data within a single generative architecture would enable simultaneous analysis of evolving 2D and 3D spatial features, detection of smoke and vulnerable infrastructure from CCTV feeds, assessment of wildfire impacts on wildlife, and even prediction of housing price shifts in fire-prone areas, alongside automated synthesis of textual reports and advisories. Advancing this line of research represents a significant opportunity to enhance the deployability, scalability, and maintainability of AI-powered wildfire decision support, establishing a transformative paradigm for interdisciplinary wildfire management.

6.1.4. Real-Time Fire Scenario Generation on Mobile Devices

Generative AI models such as VAEs and transformers have shown considerable promise as computationally efficient surrogates for simulating wildfire spread. Unlike physics-based models that rely on intensive numerical solvers, these architectures can emulate spatio-temporal fire dynamics from historical or synthetic data, producing high-resolution wildfire scenarios with minimal computational overhead. Once trained, they offer near-instant inference, making them attractive for time-critical applications such as emergency response, evacuation planning, and real-time firefighting operations, and potentially deployable on mobile devices with limited computing resources.
However, despite the reviewed 11 studies demonstrating the computational efficiency and superior inference speeds of generative AI models, this research remains largely at an experimental stage. None of these studies have explored the bright potential of deploying such lightweight generative models directly on mobile platforms and edge devices. This represents an important gap and a promising direction for future research.
Edge and mobile deployment could distribute computational and data analytics workloads across devices such as smartphones, UAVs, sensor controllers, and CCTV camera systems, reducing reliance on remote infrastructure and mitigating communication delays—especially critical in disaster-affected or connectivity-limited regions. Prior work has shown that convolutional and transformer-based models can be optimized for low-power hardware, such as NVIDIA Jetson Nano or Intel Movidius, enabling onboard inference without cloud dependence [104,105]. Lightweight ViTs-based fire segmentation models are increasingly compatible with embedded systems, supporting high-frequency updates and real-time forecasting on resource-constrained devices [106,107].
Future work could leverage these developments to embed generative wildfire models within mobile edge computing frameworks, enabling on-the-fly scenario generation using the most current field-collected data—such as wind speed, terrain, vegetation, and humidity. This would empower frontline responders with localized fire trajectory predictions, dynamic risk maps, and adaptive containment strategies, all processed directly on-site. Advancing this research agenda could establish a scalable, resilient, and connectivity-independent paradigm for delivering real-time wildfire intelligence.

6.1.5. Trustworthy Wildfire Prediction via Explainable AI and Interactive Visual Analytics

Recent advances in generative AI have opened new avenues for developing explainable and interactive wildfire prediction systems that go beyond traditional black-box models. Many state-of-the-art generative architectures, such as VAEs and transformers, rely on latent spaces to encode complex environmental data into compressed representations. These latent variables offer a powerful mechanism for uncovering how models process input features and arrive at predictions or classifications [85,108]. Notably, although the 11 reviewed studies employed VAE- and transformer-based approaches that inherently learn latent representations and embeddings, none explored explainable AI techniques that visualize or interpret these latent spaces to clarify the models’ decision-making processes.
This gap highlights a compelling direction for future research. By developing visual analytics interfaces that directly examine latent spaces, stakeholders—including emergency managers, urban planners, and community members—could better understand how inputs such as fuel conditions, topography, and weather patterns collectively influence predicted fire spread. Interactive exploration of these low-dimensional representations would facilitate detection of meaningful patterns, anomalies, or risk clusters, enabling non-technical users to interrogate AI-driven forecasts and fostering greater transparency and trust. Such approaches stand in contrast to existing wildfire modeling tools that often produce opaque outputs requiring specialized interpretation.
While these explainable AI systems could be embedded within broader platforms—such as urban digital twins or smart city dashboards—the core research agenda centers on advancing methodologies for visualizing, interpreting, and interacting with the internal decision processes of generative AI wildfire models. Integrating real-time data streams from IoT sensors, UAVs, or crowdsourcing platforms, these systems could continuously update their analyses and present the latest insights in an intuitive, user-centered format. Pursuing this line of research promises to significantly enhance the interpretability, reliability, and adoption of AI-based wildfire prediction tools, paving the way for more transparent and participatory fire risk management.

6.2. Challenges and Potential Solutions

Despite the promising potential of generative AI in wildfire prediction, several critical challenges must be addressed to ensure its reliability, scalability, and scientific validity in real-world fire management applications. Table 3 summarizes these key challenges along with potential solutions, which are further elaborated in the following subsections.

6.2.1. Stochasticity Challenges in Fire Prediction

While advances in computational modeling and data-driven learning have markedly improved our ability to simulate wildfire spread, it is critical to acknowledge the inherent stochasticity and irreducible unpredictability that characterize wildfire behavior [50,111]. Unlike controlled physical systems, wildfires evolve within highly dynamic and often chaotic natural environments, where small variations in initial conditions—such as localized wind gusts, abrupt changes in humidity, or micro-scale fuel heterogeneity—can lead to dramatically different outcomes over time [112,113]. This sensitivity to initial and boundary conditions underscores a fundamental limitation: even the most sophisticated models, whether physics-based or generative, cannot entirely eliminate uncertainty in predicting fire evolution.
Moreover, wildfires are influenced by a multitude of coupled processes across scales, including turbulent convection, ember lofting and spotting, and rapid transitions in combustion regimes, which are intrinsically probabilistic and only partially observable [50]. For example, fine-scale wind vortices can transport embers far beyond modeled fire perimeters, triggering new ignitions in a manner that defies deterministic forecasting. Similarly, fuel moisture content can fluctuate over short distances due to microclimatic effects, introducing additional variability that is difficult to capture with static or coarse-grained input data [113].
Consequently, while generative AI offers powerful new avenues for representing complex spread dynamics and for sampling from learned distributions that mirror observed variability, it does not circumvent the fundamental unpredictability embedded in wildfire phenomena. Instead, these models should be viewed as tools to better characterize the probabilistic ranges of likely outcomes and to improve computational efficiency, rather than to produce more precise point forecasts. This perspective emphasizes the need for ensemble simulations, scenario-based planning, and robust uncertainty quantification as integral components of any wildfire prediction framework [113,114]. By explicitly accounting for the stochastic nature of fire behavior, researchers and practitioners can make more informed decisions that appropriately balance forecast precision with the intrinsic volatility of wildfire systems.

6.2.2. Computational Challenges

Training large-scale generative AI models—particularly diffusion models, transformers, and other autoregressive architectures—presents substantial computational challenges. These models typically require high-resolution spatial-temporal data spanning vast geographic regions and extended fire seasons, leading to massive, memory-intensive datasets. As emphasized by Manduchi et al. [115], training such models at scale demands access to powerful GPUs or TPUs, careful memory management, and optimization strategies to maintain training stability across distributed systems.
Moreover, inference with diffusion models often involves iterative denoising steps, while transformer-based language models follow a sequential token generation process, both of which result in high latency. This is particularly problematic in real-time or near-real-time wildfire forecasting applications, where rapid decision making is essential. Even recent improvements like FlashAttention and latent-space diffusion training [115,116] only partially mitigate these bottlenecks. Additionally, robust training requires extensive data augmentation, domain adaptation, and fine-tuning, which further increases the computational burden [116].
To improve computational efficiency, emerging strategies such as model quantization, low-bit training, distillation, and lossy latent-space modeling have shown promise [115]. However, these approaches often come with trade-offs in terms of accuracy, robustness, or interpretability—making them challenging to adopt in high-stakes domains such as environmental hazard forecasting and disaster response.

6.2.3. Evaluation Challenges

Evaluating the outputs of generative AI models in wildfire prediction poses significant methodological challenges, particularly due to the lack of standardized, interpretable, and domain-specific evaluation frameworks. Unlike classification or regression tasks, where performance can be quantified using well-established metrics, the assessment of synthetic wildfire scenarios—especially those generated under creative or hypothetical conditions—remains highly subjective and context-dependent [115,116]. Widely used metrics such as Fréchet inception distance (FID) or inception score (IS), originally developed for image synthesis, are ill-suited for evaluating spatio-temporal fidelity, physical realism, or consistency with known fire-behavior dynamics in environmental simulations [115,116].
Recent studies emphasize the urgent need for developing “domain-specific evaluation metrics” tailored to complex generative tasks. These include assessing spatial spread accuracy, temporal progression, and physical plausibility relative to empirical wildfire data and simulation-based fire-behavior models [117]. Additionally, generative AI models frequently suffer from pathologies such as “mode collapse, sample hallucination, and memorization”, which compromise their generalizability and reliability when trained on sparse or biased fire datasets [115]. These issues are particularly critical in “high-stakes applications” such as real-time emergency response or predictive decision support, where misleading outputs can result in dangerous operational consequences.
The evaluation challenge is further exacerbated by the absence of ground truth for rare or extreme wildfire events, which limits model validation using traditional benchmarks. As [115,118] argue, current benchmarks fail to capture domain-specific constraints and interpretability, especially in dynamic, real-world environments. To address these shortcomings, researchers have advocated for the integration of “human-in-the-loop evaluation”, “uncertainty quantification”, and “physics-informed priors” into both model training and assessment pipelines [117,119]. Furthermore, a growing body of literature calls for “hybrid metrics” that combine statistical quality measures with expert-driven assessments and simulation-based validations, offering a more holistic evaluation of generative outputs [115,116,117].

6.2.4. Energy and Environmental Challenges

One significant but often under-discussed challenge in developing generative AI models for wildfire prediction lies in their substantial energy consumption and environmental footprint. Training foundation models—particularly those based on large-scale transformer architectures—requires extensive computational resources, often involving thousands of GPU hours and high-performance computing clusters [120]. This process results in considerable electricity usage, much of which is still powered by fossil fuels in many regions, leading to high levels of carbon emissions. As highlighted by [103], the environmental cost of training a single large language model can surpass the annual carbon footprint of several individuals. While these models offer valuable capabilities for simulating fire dynamics and generating proactive mitigation strategies, their development raises important concerns regarding sustainability and ecological responsibility, especially in the context of wildfire prediction, where climate change and ecosystem degradation are already pressing issues.
To mitigate these impacts, it is essential to explore strategies such as model distillation, sparsity optimization, transfer learning from pre-trained models, and leveraging green AI principles that emphasize energy-efficient training and inference [109,110]. Integrating these strategies not only aligns the development of AI models with climate-conscious values but also ensures that the tools designed to protect the environment do not inadvertently contribute to its degradation.

7. Conclusions

This review systematically examined the emerging role of generative AI models—including VAEs, GANs, transformers, and diffusion architectures—in advancing bushfire prediction, monitoring, and risk mapping. While only 11 studies were identified that specifically explore these approaches for wildfire science, their collective findings demonstrate that generative AI can achieve prediction accuracies near 90% and offer computational efficiencies far surpassing traditional methods. Despite these advances, the use of generative AI in this domain remains nascent compared to its widespread adoption in other scientific and engineering fields.
Building on these observations, we proposed several promising research directions to shape the next generation of wildfire decision support systems. These include developing unified multimodal frameworks that seamlessly integrate 2D and 3D data, designing conversational agentic AI systems to deliver interactive, real-time wildfire intelligence, training interdisciplinary AI foundation models, enabling edge-based scenario generation on mobile and IoT devices, and advancing explainable AI interfaces for improved transparency and trust.
Addressing these frontiers will require tackling notable challenges such as managing inherent wildfire stochasticity, reducing computational and energy footprints, and establishing rigorous, domain-specific evaluation standards. By harnessing the unique strengths of generative AI while embedding principles of explainability, sustainability, and interdisciplinary integration, future research can build robust, adaptive, and human-centered systems that significantly enhance wildfire prediction and emergency management capabilities.

Author Contributions

Conceptualization: H.X. and S.Z.; Methodology: H.X.; Investigation, H.X.; Writing—original draft preparation, H.X.; Writing—review and editing, S.Z., R.L. and I.C.; Visualization, H.X.; Supervision, S.Z. and I.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by a project titled BushFireAI at the University of New South Wales (UNSW Sydney), Australia, under grant number PS73170.

Acknowledgments

The authors would like to express their sincere gratitude to the Fire and Rescue NSW (FRNSW), NSW Rural Fire Service (NSW RFS), and Australasian Fire and Emergency Service Authorities Council (AFAC)—the National Council for fire and emergency services in Australia and New Zealand—for their valuable expert support and guidance throughout this research. Their insights and expertise have significantly contributed to the development and relevance of this work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bowman, D.M.; Kolden, C.A.; Abatzoglou, J.T.; Johnston, F.H.; van der Werf, G.R.; Flannigan, M. Vegetation fires in the Anthropocene. Nat. Rev. Earth Environ. 2020, 1, 500–515. [Google Scholar] [CrossRef]
  2. Doerr, S.H.; Santín, C. Global trends in wildfire and its impacts: Perceptions versus realities in a changing world. Philos. Trans. R. Soc. B Biol. Sci. 2016, 371, 20150345. [Google Scholar] [CrossRef] [PubMed]
  3. Kotozaki, Y. The increase of wildfires and their impact: The importance of prevention and measures through the example of Ofunato. Front. For. Glob. Chang. 2025, 8, 1589796. [Google Scholar] [CrossRef]
  4. Goralnick, E.; Nadeau, K.; Moyal-Smith, R.; Szema, A. Long-Term Health Implications of the Los Angeles Wildfires and Insights From Military Burn Pit Exposures. JAMA Intern. Med. 2025. [Google Scholar] [CrossRef] [PubMed]
  5. Amiri, A.; Gumiere, S.; Bonakdari, H. Firestorm in California: The new reality for wildland-urban interface regions. Urban Clim. 2025, 62, 102528. [Google Scholar] [CrossRef]
  6. NASA Scientific Visualization Studio. Overview Maps of 2025 Los Angeles Fires. 2025. Available online: https://svs.gsfc.nasa.gov/5568/ (accessed on 14 July 2025).
  7. Johnston, F.H.; Henderson, S.B.; Chen, Y.; Randerson, J.T.; Marlier, M.; DeFries, R.S.; Kinney, P.; Bowman, D.M.; Brauer, M. Estimated global mortality attributable to smoke from landscape fires. Environ. Health Perspect. 2012, 120, 695–701. [Google Scholar] [CrossRef]
  8. Urbanski, S. Wildland fire emissions, carbon, and climate: Emission factors. For. Ecol. Manag. 2014, 317, 51–60. [Google Scholar] [CrossRef]
  9. Brown, T.; Shelton, J. Confluence of fire and people. Nat. Sustain. 2025, 8, 329–330. [Google Scholar] [CrossRef]
  10. Han, S.Y.; Lee, Y.; Yoo, J.; Kang, J.Y.; Park, J.; Myint, S.W.; Cho, E.; Gu, X.; Kim, J.S. Spatial Disparities in Fire Shelter Accessibility: Capacity Challenges in the Palisades and Eaton Fires. arXiv 2025, arXiv:2506.06803. [Google Scholar]
  11. Woolcott, O.O. Los Angeles County in flames: Responsibilities on fire. Lancet Reg. Health-Am. 2025, 42, 101005. [Google Scholar] [CrossRef]
  12. Natural Hazards Research Australia. Understanding the Black Summer Bushfires Through Research: A Summary of Key Findings from the Bushfire and Natural Hazards CRC; Technical Report 10.2022; Natural Hazards Research Australia: Melbourne, Australia, 2023; Available online: https://www.naturalhazards.com.au/black-summer (accessed on 15 January 2023).
  13. EPA. California Wildfires Response. 2025. Available online: https://www.defense.gov/spotlights/California-Wildfire-Response/ (accessed on 15 January 2025).
  14. Gazette, H. Death, Destruction, and Trauma of L.A. Wildfires: Los Angeles Paradise Fire. 2025. Available online: https://news.harvard.edu/gazette/story/2025/01/death-destruction-and-trauma-of-l-a-wildfires-los-angeles-paradise-fire/ (accessed on 15 January 2025).
  15. Baptiste Filippi, J.; Bosseur, F.; Mari, C.; Lac, C.; Le Moigne, P.; Cuenot, B.; Veynante, D.; Cariolle, D.; Balbi, J.H. Coupled atmosphere-wildland fire modelling. J. Adv. Model. Earth Syst. 2009, 1, 11. [Google Scholar] [CrossRef]
  16. Barton, J.; Nishino, A.; Kohtake, N.; Westrin, H.; Shimada, M.; Guo, R.; Ichikawa, R.; Li, J.; Canbulat, I.; Zlatanova, S. Near-realtime Location Specific Messaging During Extreme Bushfire Events. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2024, 48, 9–14. [Google Scholar] [CrossRef]
  17. Wang, Z.; Zlatanova, S. Safe route determination for first responders in the presence of moving obstacles. IEEE Trans. Intell. Transp. Syst. 2019, 21, 1044–1053. [Google Scholar] [CrossRef]
  18. Wang, Z.; Zlatanova, S.; Moreno, A.; Van Oosterom, P.; Toro, C. A data model for route planning in the case of forest fires. Comput. Geosci. 2014, 68, 1–10. [Google Scholar] [CrossRef]
  19. Finney, M.A. FARSITE, Fire Area Simulator—Model Development and Evaluation; Number 4; US Department of Agriculture, Forest Service, Rocky Mountain Research Station: Fort Collins, CO, USA, 1998.
  20. Mell, W.; Jenkins, M.A.; Gould, J.; Cheney, P. A physics-based approach to modelling grassland fires. Int. J. Wildland Fire 2007, 16, 1–22. [Google Scholar] [CrossRef]
  21. Miller, C.; Hilton, J.; Sullivan, A.; Prakash, M. SPARK—A bushfire spread prediction tool. In Proceedings of the International Symposium on Environmental Software Systems, Melbourne, Australia, 25–27 March 2015; Springer: Cham, Switzerland, 2015; pp. 262–271. [Google Scholar]
  22. Tymstra, C.; Bryce, R.; Wotton, B.; Taylor, S.; Armitage, O. Development and Structure of Prometheus: The Canadian Wildland Fire Growth Simulation Model; Information Report NOR-X-417; Natural Resources Canada, Canadian Forest Service, Northern Forestry Centre: Edmonton, AB, Canada, 2010.
  23. Barton, J.; Gorte, B.; Eusuf, M.; Zlatanova, S. A voxel-based method to estimate near-surface and elevated fuel from dense LiDAR point cloud for hazard reduction burning. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, 6, 3–10. [Google Scholar] [CrossRef]
  24. Dipierro, S.; Valdinoci, E.; Wheeler, G.; Wheeler, V.M. A simple but effective bushfire model: Analysis and real-time simulations. SIAM J. Appl. Math. 2024, 84, 1504–1514. [Google Scholar] [CrossRef]
  25. Noble, I.; Gill, A.; Bary, G. McArthur’s fire-danger meters expressed as equations. Aust. J. Ecol. 1980, 5, 201–203. [Google Scholar] [CrossRef]
  26. Jain, P.; Coogan, S.C.; Subramanian, S.G.; Crowley, M.; Taylor, S.; Flannigan, M.D. A review of machine learning applications in wildfire science and management. Environ. Rev. 2020, 28, 478–505. [Google Scholar] [CrossRef]
  27. Bot, K.; Borges, J.G. A systematic review of applications of machine learning techniques for wildfire management decision support. Inventions 2022, 7, 15. [Google Scholar] [CrossRef]
  28. Nur, A.S.; Kim, Y.J.; Lee, J.H.; Lee, C.W. Spatial prediction of wildfire susceptibility using hybrid machine learning models based on support vector regression in Sydney, Australia. Remote Sens. 2023, 15, 760. [Google Scholar] [CrossRef]
  29. Ismail, F.N.; Woodford, B.J.; Licorish, S.A.; Miller, A.D. An assessment of existing wildfire danger indices in comparison to one-class machine learning models. Nat. Hazards 2024, 120, 14837–14868. [Google Scholar] [CrossRef]
  30. Andrianarivony, H.S.; Akhloufi, M.A. Machine Learning and Deep Learning for Wildfire Spread Prediction: A Review. Fire 2024, 7, 482. [Google Scholar] [CrossRef]
  31. Shadrin, D.; Illarionova, S.; Gubanov, F.; Evteeva, K.; Mironenko, M.; Levchunets, I.; Belousov, R.; Burnaev, E. Wildfire spreading prediction using multimodal data and deep neural network approach. Sci. Rep. 2024, 14, 2606. [Google Scholar] [CrossRef]
  32. Ghali, R.; Akhloufi, M.A. Deep learning approaches for wildland fires using satellite remote sensing data: Detection, mapping, and prediction. Fire 2023, 6, 192. [Google Scholar] [CrossRef]
  33. Vargas, J.A.S. Development of a Wildfire Risk Prediction System Based on Deep Learning Methods and Remote Sensing. Master’s Thesis, Institute for Geoinformatics, University of Münster, Münster, Germany, 2025. [Google Scholar]
  34. Zakari, R.Y.; Malik, O.A.; Wee-Hong, O. An Enhanced Wildfire Spread Prediction Using Multimodal Satellite Imagery and Deep Learning Models. Remote Sens. Appl. Soc. Environ. 2025, 39, 101632. [Google Scholar] [CrossRef]
  35. Alizadeh, N.; Mahdianpari, M.; Hemmati, E.; Marjani, M. Fusionfirenet: A Cnn-Lstm Model for Short-Term Wildfire Hotspot Prediction Utilizing Spatio-Temporal Datasets. Remote Sens. Appl. Soc. Environ. 2025, 39, 101632. [Google Scholar] [CrossRef]
  36. Radke, D.; Hessler, A.; Ellsworth, D. FireCast: Leveraging Deep Learning to Predict Wildfire Spread. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 4575–4581. [Google Scholar]
  37. Chatterjee, S.S.; Lindsay, K.; Chatterjee, N.; Patil, R.; Callafon, I.A.D.; Steinbach, M.; Giron, D.; Nguyen, M.H.; Kumar, V. Prescribed Fire Modeling using Knowledge-Guided Machine Learning for Land Management. In Proceedings of the 2024 SIAM International Conference on Data Mining (SDM), Houston, TX, USA, 18–20 April 2024; SIAM: Philadelphia, PA, USA, 2024; pp. 589–597. [Google Scholar]
  38. Abdollahi, A.; Yebra, M. Challenges and Opportunities in Remote Sensing-Based Fuel Load Estimation for Wildfire Behavior and Management: A Comprehensive Review. Remote Sens. 2025, 17, 415. [Google Scholar] [CrossRef]
  39. Xu, Z.; Li, J.; Cheng, S.; Rui, X.; Zhao, Y.; He, H.; Xu, L. Wildfire risk prediction: A review. arXiv 2024, arXiv:2405.01607. [Google Scholar]
  40. He, Y. Study on Wildfire Dynamics and Cooking Stove Pollution: Experimental Analysis and Numerical Modeling. Ph.D. Thesis, UC Riverside, Riverside, CA, USA, 2024. [Google Scholar]
  41. Bhatia, A.; Eaturu, A.; Vadrevu, K.P. Deep Learning Models for Fire Prediction: A Comparative Study. In Remote Sensing of Land Cover and Land Use Changes in South and Southeast Asia; CRC Press: Boca Raton, FL, USA, 2025; Volume 1, pp. 222–241. [Google Scholar]
  42. Zhang, Q.; Wang, T. Deep learning for exploring landslides with remote sensing and geo-environmental data: Frameworks, progress, challenges, and opportunities. Remote Sens. 2024, 16, 1344. [Google Scholar] [CrossRef]
  43. Lifelo, Z.; Ding, J.; Ning, H.; Dhelim, S. Artificial intelligence-enabled metaverse for sustainable smart cities: Technologies, applications, challenges, and future directions. Electronics 2024, 13, 4874. [Google Scholar] [CrossRef]
  44. Li, J.; Zhang, C.; Zhu, W.; Ren, Y. A comprehensive survey of image generation models based on deep learning. Ann. Data Sci. 2025, 12, 141–170. [Google Scholar] [CrossRef]
  45. Sordo, Z.; Chagnon, E.; Ushizima, D. A Review on Generative AI For Text-To-Image and Image-To-Image Generation and Implications To Scientific Images. arXiv 2025, arXiv:2502.21151. [Google Scholar]
  46. Dilo, A.; Zlatanova, S. A data model for operational and situational information in emergency response: The Dutch case. Appl. Geomat. 2011, 3, 207–218. [Google Scholar] [CrossRef]
  47. Ma, Z.; Mei, G.; Xu, N. Generative deep learning for data generation in natural hazard analysis: Motivations, advances, challenges, and opportunities. Artif. Intell. Rev. 2024, 57, 160. [Google Scholar] [CrossRef]
  48. Rothermel, R.C. A Mathematical Model for Predicting Fire Spread in Wildland Fuels; Intermountain Forest & Range Experiment Station, Forest Service: Fort Collins, CO, USA, 1972; Volume 115.
  49. Finney, M.A.; Cohen, J.D.; McAllister, S.S.; Jolly, W.M. On the need for a theory of wildland fire spread. Int. J. Wildland Fire 2012, 22, 25–36. [Google Scholar] [CrossRef]
  50. Sullivan, A.L. Wildland surface fire spread modelling, 1990–2007. 1: Physical and quasi-physical models. Int. J. Wildland Fire 2009, 18, 349–368. [Google Scholar] [CrossRef]
  51. Singh, H.; Ang, L.M.; Paudyal, D.; Acuna, M.; Srivastava, P.K.; Srivastava, S.K. A Comprehensive Review of Empirical and Dynamic Wildfire Simulators and Machine Learning Techniques used for the Prediction of Wildfire in Australia. Technol. Knowl. Learn. 2025, 30, 935–968. [Google Scholar] [CrossRef]
  52. Pugnet, L.; Chong, D.; Duff, T.; Tolhurst, K. Wildland–urban interface (WUI) fire modelling using PHOENIX Rapidfire: A case study in Cavaillon, France. In Proceedings of the 20th International Congress on Modelling and Simulation, Adelaide, Australia, 1–6 December 2013; pp. 1–6. [Google Scholar]
  53. Richards, G.D. A general mathematical framework for modeling two-dimensional wildland fire spread. Int. J. Wildland Fire 1995, 5, 63–72. [Google Scholar] [CrossRef]
  54. Mambile, C.; Kaijage, S.; Leo, J. Application of Deep Learning in Forest Fire Prediction: A Systematic Review. IEEE Access 2024, 12, 190554–190581. [Google Scholar] [CrossRef]
  55. Gal, Y.; Ghahramani, Z. Dropout as a Bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the International Conference on Machine Learning (ICML), New York City, NY, USA, 19–24 June 2016; pp. 1050–1059. [Google Scholar]
  56. Lakshminarayanan, B.; Pritzel, A.; Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  57. Hochreiter, S.; Bengio, Y.; Frasconi, P.; Schmidhuber, J. Gradient flow in recurrent nets: The difficulty of learning long-term dependencies. In A Field Guide to Dynamical Recurrent Neural Networks; IEEE Press: Piscataway, NJ, USA, 2001; pp. 237–243. [Google Scholar]
  58. Chen, D.; Cheng, S.; Hu, J.; Kasoar, M.; Arcucci, R. Explainable Global Wildfire Prediction Models using Graph Neural Networks. arXiv 2024, arXiv:2402.07152. [Google Scholar]
  59. Li, J.; Hong, D.; Gao, L.; Yao, J.; Zheng, K.; Zhang, B.; Chanussot, J. Deep learning in multimodal remote sensing data fusion: A comprehensive review. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102926. [Google Scholar] [CrossRef]
  60. Shen, H.; Li, X.; Cheng, Q.; Zeng, C.; Yang, G.; Li, H.; Zhang, L. Missing information reconstruction of remote sensing data: A technical review. IEEE Geosci. Remote Sens. Mag. 2015, 3, 61–85. [Google Scholar] [CrossRef]
  61. Li, X.; Luo, S.; He, Y.; Liu, Y. MisGAN: Learning from incomplete data with generative adversarial networks. In Proceedings of the International Conference on Learning Representations (ICLR 2019), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  62. Rudin, C. Stop explaining black box machine learning models for high-stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [PubMed]
  63. Doshi-Velez, F.; Kim, B. Towards a rigorous science of interpretable machine learning. arXiv 2017, arXiv:1702.08608. [Google Scholar]
  64. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. In Proceedings of the Advances in Neural Information Processing Systems 33 (NeurIPS 2020), Virtual, 6–12 December 2020. [Google Scholar]
  65. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  66. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  67. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. In Proceedings of the Advances in Neural Information Processing Systems 34, Vancouver, BC, Canada, 6–12 December 2020; pp. 6840–6851. [Google Scholar]
  68. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  69. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference On Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 10684–10695. [Google Scholar]
  70. Dinh, L.; Sohl-Dickstein, J.; Bengio, S. Density estimation using Real NVP. In Proceedings of the International Conference on Learning Representations (ICLR 2017), Toulon, France, 24–26 April 2017. [Google Scholar]
  71. Xu, H.; Omitaomu, F.; Sabri, S.; Zlatanova, S.; Li, X.; Song, Y. Leveraging generative AI for urban digital twins: A scoping review on the autonomous generation of urban data, scenarios, designs, and 3D city models for smart city advancement. Urban Inform. 2024, 3, 29. [Google Scholar] [CrossRef]
  72. Tupayachi, J.; Xu, H.; Omitaomu, O.A.; Camur, M.C.; Sharmin, A.; Li, X. Towards next-generation urban decision support systems through AI-powered construction of scientific ontology using large language models—A case in optimizing intermodal freight transportation. Smart Cities 2024, 7, 2392–2421. [Google Scholar] [CrossRef]
  73. Xu, H.; Yuan, J.; Zhou, A.; Xu, G.; Li, W.; Ban, X.J.; Ye, X. Generative Artificial Intelligence–Powered Multi-Agent Paradigm for Smart Urban Mobility: Opportunities and Challenges for Integrating Large Language Models and Retrieval-Augmented Generation with Intelligent Transportation Systems. In Urban Human Mobility; CRC Press: Boca Raton, FL, USA, 2025; pp. 123–137. [Google Scholar]
  74. Taiwo, R.; Yussif, A.M.; Zayed, T. Making waves: Generative artificial intelligence in water distribution networks: Opportunities and challenges. Water Res. X 2025, 28, 100316. [Google Scholar] [CrossRef]
  75. Sun, J.; Qi, W.; Huang, Y.; Xu, C.; Yang, W. Facing the wildfire spread risk challenge: Where are we now and where are we going? Fire 2023, 6, 228. [Google Scholar] [CrossRef]
  76. Anderson, D.; Catchpole, E.; De Mestre, N.; Parkes, T. Modelling the spread of grass fires. ANZIAM J. 1982, 23, 451–466. [Google Scholar] [CrossRef]
  77. Mitsopoulos, I.; Mallinis, G.; Karali, A.; Giannakopoulos, C.; Arianoutsou, M. Mapping fire behaviour in a Mediterranean landscape under different future climate change scenarios. In Proceedings of the International Conference ADAPTtoCLIMATE, Nicosia, Cyprus, 27–28 March 2014. [Google Scholar]
  78. Rui, X.; Hui, S.; Yu, X.; Zhang, G.; Wu, B. Forest fire spread simulation algorithm based on cellular automata. Nat. Hazards 2018, 91, 309–319. [Google Scholar] [CrossRef]
  79. Xu, Y.; Li, D.; Ma, H.; Lin, R.; Zhang, F. Modeling forest fire spread using machine learning-based cellular automata in a GIS environment. Forests 2022, 13, 1974. [Google Scholar] [CrossRef]
  80. Yang, Y.; Jin, M.; Wen, H.; Zhang, C.; Liang, Y.; Ma, L.; Wang, Y.; Liu, C.; Yang, B.; Xu, Z.; et al. A survey on diffusion models for time series and spatio-temporal data. arXiv 2024, arXiv:2404.18886. [Google Scholar]
  81. Joshi, S. Introduction to Diffusion Models, Autoencoders and Transformers: Review of Current Advancements. HAL 2025, hal-04999764. [Google Scholar] [CrossRef]
  82. Zhihan, G. Exploring Deep Learning for Earth System Forecasting. Ph.D. Thesis, Hong Kong University of Science and Technology, Hong Kong, China, 2024. [Google Scholar]
  83. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; PmLR: Irvine, CA, USA, 2021; pp. 8748–8763. [Google Scholar]
  84. Dhariwal, P.; Nichol, A. Diffusion models beat gans on image synthesis. Adv. Neural Inf. Process. Syst. 2021, 34, 8780–8794. [Google Scholar]
  85. Xu, H.; Boyaci, A.; Lian, J.; Wilson, A. Explainable AI for Multivariate Time Series Pattern Exploration: Latent Space Visual Analytics with Temporal Fusion Transformer and Variational Autoencoders in Power Grid Event Diagnosis. arXiv 2024, arXiv:2412.16098. [Google Scholar]
  86. Xu, H.; Li, X.; Tupayachi, J.; Lian, J.J.; Omitaomu, O.A. Automating Bibliometric Analysis with Sentence Transformers and Retrieval-Augmented Generation (RAG): A Pilot Study in Semantic and Contextual Search for Customized Literature Characterization for High-Impact Urban Research. In Proceedings of the 2nd ACM SIGSPATIAL International Workshop on Advances in Urban-AI, Atlanta, GA, USA, 29 October–1 November 2024; pp. 43–49. [Google Scholar]
  87. Khanmohammadi, S.; Arashpour, M.; Golafshani, E.M.; Cruz, M.G.; Rajabifard, A. An artificial intelligence framework for predicting fire spread sustainability in semiarid shrublands. Int. J. Wildland Fire 2023, 32, 636–649. [Google Scholar] [CrossRef]
  88. Khanmohammadi, S.; Cruz, M.G.; Perrakis, D.D.; Alexander, M.E.; Arashpour, M. Using AutoML and generative AI to predict the type of wildfire propagation in Canadian conifer forests. Ecol. Inform. 2024, 82, 102711. [Google Scholar] [CrossRef]
  89. Cheng, S.; Guo, Y.; Arcucci, R. A generative model for surrogates of spatial-temporal wildfire nowcasting. IEEE Trans. Emerg. Top. Comput. Intell. 2023, 7, 1420–1430. [Google Scholar] [CrossRef]
  90. Li, B.S.; Rad, R. Wildfire spread prediction in North America using satellite imagery and vision transformer. In Proceedings of the 2024 IEEE Conference on Artificial Intelligence (CAI), Singapore, 25–27 June 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1536–1541. [Google Scholar]
  91. Deepa, N.; Husain, S.O.; Ezhil, G.; Indhumathi, V.; Revathi, R. An Effective Forest Fire Prediction Based on Contrastive Vision Transformer with Pool Former. In Proceedings of the 2024 International Conference on Integrated Intelligence and Communication Systems (ICIICS), Karnataka, India, 22–23 November 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–5. [Google Scholar]
  92. Annane, B.; Lakehal, A.; Alti, A. Secured Forest Fire Prediction Using Blockchain and CNN Transformers. In Proceedings of the 2024 International Conference on Information and Communication Technologies for Disaster Management (ICT-DM), Setif, Algeria, 19–21 November 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–7. [Google Scholar]
  93. Li, Y.; Li, K.; Guohui, L.; Ji, C.; Wang, L.; Zuo, D.; Guo, Q.; Zhang, F.; Wang, M.; Lin, D.; et al. Sim2real-fire: A multi-modal simulation dataset for forecast and backtracking of real-world forest fire. Adv. Neural Inf. Process. Syst. 2024, 37, 1428–1442. [Google Scholar]
  94. Falcão, G.; Fernandes, A.M.; Garcia, N.; Aidos, H.; Tomás, P. Stacking Deep Learning Models for Early Detection of Wildfire Smoke Plumes. In Proceedings of the 2023 31st European Signal Processing Conference (EUSIPCO), Helsinki, Finland, 4–8 September 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1370–1374. [Google Scholar]
  95. Ghali, R.; Akhloufi, M.A.; Mseddi, W.S. Deep learning and transformer approaches for UAV-based wildfire detection and segmentation. Sensors 2022, 22, 1977. [Google Scholar] [CrossRef]
  96. Ghali, R.; Akhloufi, M.A.; Jmal, M.; Souidene Mseddi, W.; Attia, R. Wildfire segmentation using deep vision transformers. Remote Sens. 2021, 13, 3527. [Google Scholar] [CrossRef]
  97. Limber, R.; Hargrove, W.W.; Hoffman, F.M.; Kumar, J. Forecast of Wildfire Potential Across California USA Using a Transformer. In Proceedings of the 2024 IEEE International Conference on Big Data (BigData), Washington, DC, USA, 15–18 December 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 4342–4350. [Google Scholar]
  98. Li, Y.; Yang, Z.; Yang, Z.; Li, X.; Liu, W.; Li, Q. Multimodal Disentangled Fusion Network via VAEs for Multimodal Zero-Shot Learning. IEEE Trans. Comput. Soc. Syst. 2025; early access. [Google Scholar]
  99. Xu, P.; Zhu, X.; Clifton, D.A. Multimodal learning with transformers: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 12113–12132. [Google Scholar] [CrossRef] [PubMed]
  100. Xie, Y.; Jiang, B.; Mallick, T.; Bergerson, J.D.; Hutchison, J.K.; Verner, D.R.; Branham, J.; Alexander, M.R.; Ross, R.B.; Feng, Y.; et al. WildfireGPT: Tailored Large Language Model for Wildfire Analysis. arXiv 2025, arXiv:2402.07877. [Google Scholar]
  101. Xu, H.; Sun, Y.; Tupayachi, J.; Omitaomu, O.; Zlatanov, S.; Li, X. Towards the Autonomous Optimization of Urban Logistics: Training Generative AI with Scientific Tools via Agentic Digital Twins and Model Context Protocol. arXiv 2025, arXiv:2506.13068. [Google Scholar]
  102. Li, X.; Xu, H.; Tupayachi, J.; Omitaomu, O.; Wang, X. Empowering cognitive digital twins with generative foundation models: Developing a low-carbon integrated freight transportation system. arXiv 2024, arXiv:2410.18089. [Google Scholar]
  103. Myers, D.; Mohawesh, R.; Chellaboina, V.I.; Sathvik, A.L.; Venkatesh, P.; Ho, Y.H.; Henshaw, H.; Alhawawreh, M.; Berdik, D.; Jararweh, Y. Foundation and large language models: Fundamentals, challenges, opportunities, and social impacts. Clust. Comput. 2024, 27, 1–26. [Google Scholar] [CrossRef]
  104. Hu, Z.; Shi, S.; Ye, Y.; He, W.; Yang, Z.; Li, X.; Feng, S. Edge Computing-Based Wildfire Detection and Optimization Algorithm. In Proceedings of the 2024 7th International Conference on Advanced Algorithms and Control Engineering (ICAACE), Shanghai, China, 1–3 March 2024. [Google Scholar]
  105. Mahdi, A.S.; Mahmood, S.A. An Edge Computing Environment for Early Wildfire Detection. Ann. Emerg. Technol. Comput. (AETiC) 2022, 6, 56–68. [Google Scholar] [CrossRef]
  106. Spiller, D.; Thangavel, K.; Sasidharan, S.T.; Amici, S.; Ansalone, L.; Sabatini, R. Wildfire Segmentation Analysis from Edge Computing for On-board Real-time Alerts Using Hyperspectral Imagery. In Proceedings of the 2022 IEEE International Conference on Metrology for eXtended Reality, Artificial Intelligence and Neural Engineering (MetroXRAINE), Rome, Italy, 26–28 October 2022. [Google Scholar]
  107. Lee, S.I.; Koo, K.; Lee, J.H.; Lee, G.; Jeong, S.; O, S.; Kim, H. Vision Transformer Models for Mobile/Edge Devices: A Survey. Multimed. Syst. 2024, 30, 109. [Google Scholar] [CrossRef]
  108. Liu, Y.; Jun, E.; Li, Q.; Heer, J. Latent space cartography: Visual analysis of vector space embeddings. Comput. Graph. Forum 2019, 38, 67–78. [Google Scholar] [CrossRef]
  109. Salehi, S.; Schmeink, A. Data-centric green artificial intelligence: A survey. IEEE Trans. Artif. Intell. 2023, 5, 1973–1989. [Google Scholar] [CrossRef]
  110. Barbierato, E.; Gatti, A. Toward green AI: A methodological survey of the scientific literature. IEEE Access 2024, 12, 23989–24013. [Google Scholar] [CrossRef]
  111. Morvan, D. Physical phenomena and length scales governing the behaviour of wildfires: A case for physical modelling. Fire Technol. 2011, 47, 437–460. [Google Scholar] [CrossRef]
  112. Tedim, F.; Leone, V.; Amraoui, M.; Bouillon, C.; Coughlan, M.; Delogu, G.; Fernandes, P.; Ferreira, C.; McCaffrey, S.; McGee, T.; et al. Defining extreme wildfire events: Difficulties, challenges, and impacts. Fire 2018, 1, 9. [Google Scholar] [CrossRef]
  113. Hilton, J.E.; Miller, C.; Sullivan, A.L.; Rucinski, C. Effects of spatial and temporal variation in environmental conditions on simulation of wildfire spread. Environ. Model. Softw. 2015, 67, 118–127. [Google Scholar] [CrossRef]
  114. Rochoux, M.C.; Collin, A.; Zhang, C.; Trouvé, A.; Lucor, D.; Moireau, P. Front shape similarity measure for shape-oriented sensitivity analysis and data assimilation for Eikonal equation. ESAIM Proc. Surv. 2018, 63, 258–279. [Google Scholar] [CrossRef]
  115. Manduchi, L.; Pandey, K.; Meister, C.; Bamler, R.; Cotterell, R.; Däubener, S.; Fellenz, S.; Fischer, A.; Gärtner, T.; Kirchler, M.; et al. On the challenges and opportunities in generative AI. arXiv 2024, arXiv:2403.00025. [Google Scholar]
  116. Bandi, A.; Adapa, P.V.S.R.; Kuchi, Y.E.V.P.K. The power of generative AI: A review of requirements, models, input–output formats, evaluation metrics, and challenges. Future Internet 2023, 15, 260. [Google Scholar] [CrossRef]
  117. Fui-Hoon Nah, F.; Zheng, R.; Cai, J.; Siau, K.; Chen, L. Generative AI and ChatGPT: Applications, challenges, and AI-human collaboration. J. Inf. Technol. Case Appl. Res. 2023, 25, 277–304. [Google Scholar] [CrossRef]
  118. Sun, Y.; Jang, E.; Ma, F.; Wang, T. Generative AI in the wild: Prospects, challenges, and strategies. In Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 11–16 May 2024; pp. 1–16. [Google Scholar]
  119. Yan, L.; Greiff, S.; Teuber, Z.; Gašević, D. Promises and challenges of generative artificial intelligence for human learning. Nat. Hum. Behav. 2024, 8, 1839–1850. [Google Scholar] [CrossRef]
  120. Zhou, J.; Chen, Y.; Hong, Z.; Chen, W.; Yu, Y.; Zhang, T.; Wang, H.; Zhang, C.; Zheng, Z. Training and serving system of foundation models: A comprehensive survey. IEEE Open J. Comput. Soc. 2024, 5, 107–119. [Google Scholar] [CrossRef]
Figure 1. Search queries used to identify and acquire literature from IEEExplore and Scopus databases.
Figure 1. Search queries used to identify and acquire literature from IEEExplore and Scopus databases.
Fire 08 00293 g001
Figure 2. Summary of recent studies applying generative AI to diverse areas in bushfire modeling and prediction.
Figure 2. Summary of recent studies applying generative AI to diverse areas in bushfire modeling and prediction.
Fire 08 00293 g002
Figure 3. A summary of five future visions for applying generative AI to revolutionize wildfire prediction and management, ranging from the development of multimodal 2D and 3D wildfire modeling to the implementation of cognitive digital twins with explainable AI capabilities.
Figure 3. A summary of five future visions for applying generative AI to revolutionize wildfire prediction and management, ranging from the development of multimodal 2D and 3D wildfire modeling to the implementation of cognitive digital twins with explainable AI capabilities.
Fire 08 00293 g003
Table 1. Fire simulation models based on different underlying principles.
Table 1. Fire simulation models based on different underlying principles.
Model TypeCore PrincipleExamplesStrength
Physical ModelsBased on first principles of physics and chemistry (e.g., combustion thermodynamics, heat transfer, fluid dynamics).FIRETEC, WFDS, FIRESTAR, IUSTI, GrishinHigh fidelity, models full fire–fuel–atmosphere interaction, scientific rigor.
Quasi-Physical ModelsIncludes physical processes (e.g., energy conservation, heat transfer) but omits combustion chemistry, often uses simplified fire shape assumptions.UoS (Spain), LEMTA, FIRESTAR-liteBalance between physical realism and computational feasibility.
Empirical ModelsDerived purely from statistical regression of observed fire behavior (no physical basis); field or lab-based.McArthur FDRS, CSIRO Grass Meter, CFBP (Canada)Easy to use, computationally light, good for operational tools.
Quasi-Empirical ModelsEmpirical models informed or supported by a physical framework (e.g., use physical insights to design empirical terms).Rothermel model, BEHAVE, Noble-McArthur modelWidely used in practice, moderate complexity.
Mathematical Analogue ModelsUse abstract mathematical constructs (e.g., cellular automata, percolation theory, wavelet propagation) not rooted in real fire physics.Cellular Automata models, Huygens’ wavelet, Prometheus, SiroFireFlexible, suitable for exploratory simulation, fast prototyping.
Table 2. Comparison of traditional machine-learning, deep-learning, and generative AI models.
Table 2. Comparison of traditional machine-learning, deep-learning, and generative AI models.
FeatureTraditional ML (e.g., SVM, RF)Deep Learning (e.g., CNN, LSTM)Generative AI (e.g., VAE, Transformer, Diffusion Model)
Data generation capabilityCannot generate dataPredictive onlyCan generate new, realistic, diverse data.
Representation learningManual feature engineeringLearns hierarchical featuresLearns latent and semantic representations.
Handling missing data/imputationBasic imputation (e.g., mean)Regression-based or interpolation onlyLearns to impute based on data distribution.
Few-shot/zero-shot learningRequires full trainingRequires full trainingSupported by large-scale transformers.
Multimodal learningNeeds manual integrationSeparate networks for each modalityUnified models handle text, image, video, etc.
Uncertainty quantificationVia Bayesian methods or ensemblesDeterministic outputBuilt-in probabilistic frameworks (e.g., VAEs).
Synthetic data augmentationNot supportedRequires manual engineeringEasily supports realistic data generation.
Scenario simulationNot applicableNot applicableSimulates realistic and hypothetical conditions.
Latent space manipulationNot availableNot interpretableSupports interpolation and control.
Creativity and generative powerNoneNoneHigh—generates novel outputs and scenario.
Data efficiencyNeeds many labeled samplesHigh data demandSome support few-shot learning via pretraining.
InterpretabilityOften interpretableDifficult to interpret hidden layersLatent space can be visualized and interpretable.
Training complexitySimple to trainNeeds tuning and GPU supportComplex training and high computational cost.
Use in scientific simulationLimited (basic regression)Used in some modelingStrong in data-driven and uncertain modeling.
Table 3. Challenges and potential solutions associated with applying generative AI models in wildfire prediction.
Table 3. Challenges and potential solutions associated with applying generative AI models in wildfire prediction.
ChallengeDescriptionPotential Solution
Stochastic nature of wildfire behavior Wildfire spread exhibits inherent unpredictability due to sensitivity to initial conditions, microclimate fluctuations, turbulent convection, and ember-driven spotting, which no model can fully eliminate. Employ ensemble simulations, scenario-based planning, and robust uncertainty quantification to capture probabilistic ranges of outcomes, supporting risk-informed decision making rather than exact forecasts.
ComputationalTraining and deploying large-scale generative AI models for wildfire prediction is resource-intensive due to high-resolution data demands, heavy memory usage, and slow inference.Techniques such as quantization, low-bit training, distillation, and latent-space modeling can improve efficiency, though often at the cost of accuracy or interpretability.
EvaluationEvaluating generative AI outputs is difficult due to the lack of standardized, interpretable, and domain-specific metrics for assessing the realism and utility of synthetic spatio-temporal fire scenarios.Solutions include developing domain-specific and hybrid metrics, incorporating human-in-the-loop evaluation, uncertainty quantification, and physics-informed priors.
Energy and environmentalGenerative AI model development has a significant environmental impact due to high energy consumption and carbon emissions from large-scale training.Mitigation strategies include model distillation, sparsity optimization, transfer learning, and adopting energy-efficient “green AI” practices [109,110].
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, H.; Zlatanova, S.; Liang, R.; Canbulat, I. Generative AI as a Pillar for Predicting 2D and 3D Wildfire Spread: Beyond Physics-Based Models and Traditional Deep Learning. Fire 2025, 8, 293. https://doi.org/10.3390/fire8080293

AMA Style

Xu H, Zlatanova S, Liang R, Canbulat I. Generative AI as a Pillar for Predicting 2D and 3D Wildfire Spread: Beyond Physics-Based Models and Traditional Deep Learning. Fire. 2025; 8(8):293. https://doi.org/10.3390/fire8080293

Chicago/Turabian Style

Xu, Haowen, Sisi Zlatanova, Ruiyu Liang, and Ismet Canbulat. 2025. "Generative AI as a Pillar for Predicting 2D and 3D Wildfire Spread: Beyond Physics-Based Models and Traditional Deep Learning" Fire 8, no. 8: 293. https://doi.org/10.3390/fire8080293

APA Style

Xu, H., Zlatanova, S., Liang, R., & Canbulat, I. (2025). Generative AI as a Pillar for Predicting 2D and 3D Wildfire Spread: Beyond Physics-Based Models and Traditional Deep Learning. Fire, 8(8), 293. https://doi.org/10.3390/fire8080293

Article Metrics

Back to TopTop