Skip Content
You are currently on the new version of our website. Access the old version .
SensorsSensors
  • Article
  • Open Access

5 December 2025

A Quantum-Hybrid Framework for Urban Environmental Forecasting Integrating Advanced AI and Geospatial Simulation

,
,
and
1
Information Technology Faculty, Turiba University, Graudu Street 68, LV-1058 Riga, Latvia
2
Institute of Information Technology, Riga Technical University, Kalku Street 1, LV-1658 Riga, Latvia
3
Computer Engineering and Electronics Department, Kremenchuk Mykhailo Ostrohradskyi National University, Universitetska Street 20, 39600 Kremenchuk, Ukraine
*
Author to whom correspondence should be addressed.
This article belongs to the Section Environmental Sensing

Abstract

The paper examines the development of forecasting and modeling technologies for environmental processes using classical and quantum data analysis methods. The main focus is on the integration of deep neural networks and classical algorithms, such as AutoARIMA and BATS, with quantum approaches to improve the accuracy of forecasting environmental parameters. The research is aimed at solving key problems in environmental monitoring, particularly insufficient forecast accuracy and the complexity of processing small data with high discretization. We developed the concept of an adaptive system for predicting environmental conditions in urban agglomerations. Hybrid forecasting methods were proposed, which include the integration of quantum layers in LSTM, Transformer, ARIMA, and other models. Approaches to spatial interpolation of environmental data and the creation of an interactive air pollution simulator based on the A* algorithm and the Gaussian kernel were considered. Experimental results confirmed the effectiveness of the proposed methods. The practical significance lies in the possibility of using the developed models for operational monitoring and forecasting of environmental threats. The results of the work can be applied in environmental information systems to increase the accuracy of forecasts and adaptability to changing environmental conditions.

1. Introduction

The escalating pace of global urbanization concentrates populations in dense metropolitan areas, intensifying the need for accurate, high-resolution environmental monitoring. Urban atmospheres are complex systems governed by the intricate interplay of natural meteorological phenomena and a multitude of anthropogenic activities, resulting in a high-dimensional, non-linear dynamic environment [1]. The timely prediction of air quality parameters—such as pollutant concentrations, temperature, and humidity—is paramount for mitigating public health risks, informing sustainable urban planning, and executing effective environmental policies [2]. Achieving this requires a paradigm shift from static monitoring to dynamic, predictive environmental intelligence.
A promising paradigm for this shift is the Urban Digital Twin, a concept that involves creating a dynamic, virtual replica of a city by integrating real-time sensor data, predictive models, and interactive simulations [3]. Such a system can provide unprecedented insights into urban environmental processes, enabling proactive management and “what-if” scenario analysis. The development of a robust Urban Digital Twin hinges on the ability to process vast, heterogeneous data streams and to model their complex, often non-linear, interdependencies. Existing Urban Digital Twin frameworks often focus heavily on static visual representation or architectural data, lacking dynamic, diverse predictive engines capable of processing real-time sensor data [4]. While studies such as Verde et al. [4] demonstrate the utility of digital simulation for evaluating climate adaptation strategies, the need for frameworks that tightly couple these simulations with advanced AI-driven forecasting agents still exists.
Presented research addresses the critical challenge of converting raw, high-dimensional data streams from urban sensor networks into actionable environmental intelligence. By integrating quantum-hybrid computing with geospatial simulation, this study proposes a method to enhance the computational processing of sensor data, thereby improving the accuracy and reliability of urban monitoring systems.
The manuscript proposes a comprehensive framework that serves as a foundational step toward this vision, integrating advanced computational techniques to enhance forecasting and modeling capabilities for municipal-level environmental monitoring.
The field of environmental forecasting has evolved significantly, moving from first-principle models to data-driven approaches. Traditionally, Chemistry-Transport Models (CTMs) have been the cornerstone of air quality prediction. These models numerically solve complex differential equations describing the physical and chemical processes in the atmosphere [1]. While powerful, CTMs are computationally expensive and their accuracy is often limited by uncertainties in the underlying emissions inventories and meteorological inputs [1].
To overcome these limitations, statistical and machine learning models have gained prominence. Classical time-series methods, such as the AutoRegressive Integrated Moving Average (ARIMA) model, have been widely used for their simplicity and interpretability in capturing linear dependencies and seasonal patterns in environmental data [5,6]. However, they often struggle with the strong non-linearities inherent in environmental systems [1].
The advent of deep learning has revolutionized data-driven environmental science [7,8]. Recurrent Neural Networks (RNNs), particularly Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) architectures, are adept at modeling temporal dependencies and are widely used for forecasting time-series data like dissolved oxygen levels or drought indices [2]. Convolutional Neural Networks (CNNs), traditionally used for image analysis, have been adapted to extract spatiotemporal features from gridded environmental data [7,8]. More recently, the Transformer architecture, with its self-attention mechanism, has shown exceptional capability in capturing long-range dependencies in sequential data, making it a powerful tool for tasks like climate prediction and long-term drought forecasting [9,10].
The next frontier in computational science is Quantum Machine Learning (QML), an emerging field that seeks to leverage the principles of quantum mechanics to solve complex computational problems [11,12]. Quantum computing, through phenomena like superposition and entanglement, offers a fundamentally different way to process information, with the theoretical potential to model the intricate correlations and high-dimensional state spaces characteristic of environmental systems more efficiently than classical computers [13]. Current research in applied QML focuses on hybrid quantum-classical models, where a classical neural network performs initial feature processing, and a parameterized or Variational Quantum Circuit (VQC) acts as a specialized co-processor [14].
Recent research highlights a significant shift from theoretical QML to applied environmental modeling.
Thus, research [15] demonstrated a quantum machine learning approach to spatiotemporal emission modeling, comparing a quantum quanvolutional neural network against a classical ConvLSTM. These findings indicated that quantum models could achieve lower loss and higher accuracy in emission concentration forecasts, suggesting that quantum layers can effectively capture complex spatial features in environmental satellite data. Similarly, in the agricultural domain, work [16] proposed a hybrid quantum deep learning model combining Bi-LSTM with quantum feature processing for yield forecasting. Their results showed that quantum circuits could explore high-dimensional feature spaces more efficiently than classical methods, significantly improving prediction accuracy (R2 ≈ 0.99).
The broader applicability of “quantum-like” data modeling in applied sciences has also been reviewed in [17], arguing that quantum probability frameworks can better model uncertainty and complex system behaviors than traditional approaches. Furthermore, research [18] presents successful application of Hybrid Quantum-Classical Neural Networks (H-QNN) to image classification tasks, outperforming classical CNNs with fewer parameters, which supports the hypothesis that quantum-hybrid architectures can offer computational efficiency advantages in processing high-dimensional sensor data.
These hybrid approaches are being actively explored for various tasks, including financial and scientific time-series forecasting, demonstrating the potential to achieve comparable or superior performance with fewer parameters.
Despite these advancements, several critical gaps remain in the application of advanced computational methods to urban environmental monitoring. First, the existing research often treats forecasting, spatial analysis, and simulation as isolated disciplines. There is a clear need for a unified, conceptual framework that integrates these components into a cohesive analytical pipeline, especially one that can accommodate emerging technologies like QML and Large Language Model (LLM)-based analysis. The fragmented presentation of these elements in many studies hinders the development of a holistic environmental intelligence system [2].
Second, while QML holds significant theoretical promise, its practical application to real-world problems is still in its infancy. Comprehensive, large-scale empirical studies that benchmark hybrid quantum models against a wide array of state-of-the-art classical models are rare. In particular, their performance on noisy, limited, and non-stationary environmental sensor data–the kind typically encountered in practical deployments–is not well understood. The field requires rigorous empirical validation to move beyond theoretical promise and assess the true utility of near-term quantum devices [11].

2. Analysis of the State-of-the-Art in the Field

Many current studies examine the effectiveness of using quantum neural networks for time series forecasting in the financial sector.
The authors in [19] investigate the use of parameterized quantum circuits (PQCs) as quantum neural networks for time series forecasting. They compare the performance of PQCs with classical bidirectional long short-term memory networks and find that quantum models can achieve similar or better accuracy with fewer parameters and faster training.
The study [20] considers the application of one-dimensional quantum convolutional neural networks (QCNNs) for time series prediction. The use of Fourier transform allows for control over the design of QCNNs, which improves their performance. The authors demonstrate that even with a limited number of parameters, quantum circuits are able to effectively model complex functions, which reduce training time.
In [21], the Quantum Gramian Angular Field method was proposed, which combines quantum computing with deep learning for financial time series forecasting. The research involves transforming stock return data into two-dimensional images using specially designed quantum circuits, allowing convolutional neural networks (CNNs) to be used for forecasting. The results show significant improvements in accuracy compared to classical methods.
The authors in [22] propose a method for learning temporal data using parameterized quantum circuits that have a structure similar to recurrent neural networks. In this model, some of the qubits store past data, while others are used to predict and encode new input data. The study demonstrates the ability of quantum circuits to effectively learn and predict time series.
Quantum computing is also actively used for the analysis of environmental parameters and forecasting, allowing us to compare their efficiency with classical algorithms.
One such study [23] proposes the integration of quantum computing with the Internet of Vehicles (IoV) for environmental monitoring and rapid response to environmental threats. Using quantum sensors installed on vehicles, the system provides accurate air quality assessment. Proposed Quantum Mesh Network Fabric dynamically adjusts the quantum network topology according to traffic, maintaining the integrity of quantum states and ensuring reliable data transmission. Additionally, the use of a variational quantum classifier and quantum entanglement techniques allows for reduced delays in the transmission of danger signals, which contributes to the prompt informing of emergency services and the population.
Another study [24] focuses on the use of quantum algorithms for stock price prediction. The authors conduct a series of experimental simulations using both classical and quantum hardware. They use quantum annealing for feature selection and principal component analysis to reduce the dimensionality of the data. The prediction problem is transformed into a classification problem, where a quantum support vector machine is used to predict price movements (up or down). The results are compared with classical models by analyzing the accuracy of the predictions and the F-measure.
Separately, deep neural networks, such as LSTM, GRU, CNN, and Transformer, are actively used to predict environmental indicators.
One study [25] compares the performance of different deep learning models, including CNN, TCN, LSTM, GRU, and BiRNN, for multi-step prediction of dissolved oxygen levels in water. Using data from the Yangtze River from 2012 to 2016, the authors found that the GRU model performed best in terms of RMSE, MAE, and coefficient of determination, highlighting its effectiveness in predicting ecological parameters.
Another study [26] focused on long-term drought assessment based on geospatial satellite data. The authors first employ an LSTM model (three layers, 256 hidden cells) to fill gaps in satellite soil moisture data, using inputs like temperature, precipitation, ET, and runoff. Subsequently, basin-specific LSTM models (also three layers, 256 hidden cells) are developed to forecast soil moisture. The most promising output of this work is the integration of the LSTM models with an interpretable AI method called Expected Gradients (EG), which allows the framework to quantify the specific contribution of each input feature (such as precipitation or temperature) to the soil moisture predictions over time, thereby providing a crucial understanding of the mechanisms driving drought development.
However, classical architectures continue to evolve. Recent studies have proposed sophisticated hybrid classical models to address non-linearity in sensor data. For instance, work [27] presents developed an IoT-based monitoring system utilizing a hybrid LSTM-GRU model for real-time power forecasting. This work demonstrated that combining different recurrent architectures can mitigate the specific limitations of standalone LSTM or GRU models, achieving superior accuracy in real-time IoT scenarios. In the financial domain, which shares time-series characteristics with environmental data, work [28] introduces a decomposition-ensemble model (CEEMDAN-SE combined with ARIMA-CNN-LSTM). This approach effectively handled non-stationarity by decomposing complex signals into high- and low-frequency components, a methodology relevant to processing volatile environmental sensor readings.
Some of the studies analyze the use of mathematical modeling methods for pollution assessment based on spatial interpolation, in particular using the inverse weighted distance method, for the analysis of environmental data and time series of changes in environmental parameters.
Work [29] presents a course on environmental geographic information systems (GIS), which includes an overview of the concepts and methods used to analyze and model environmental data. The course structure and examples of its practical use for solving environmental problems are presented.
The paper [30] is devoted to the use of remote sensing and open source software for working with GIS in ecology. The authors describe approaches to spatial data analysis, including tools for creating pollution maps and monitoring the state of the environment.
The source [31] discusses the basics of using GIS for environmental modeling and engineering. It describes methods of spatial analysis, interpolation, and prediction of environmental parameters. It provides examples of creating thematic maps and developing models for planning environmental protection measures.
Some basic research is devoted to computer modeling of changes in pollutant concentrations in local areas, which allows for effective assessment of the impact of man-made sources on the environment.
In [32], the authors propose a new parameterization of the concentration flux using fractional calculus for modeling the dispersion of pollutants within the planetary boundary layer of the atmosphere. The mathematical approach and its application to environmental monitoring problems are discussed in detail.
The paper [33] is devoted to the atmospheric transport and diffusion model HYSPLIT, developed by NOAA. The authors describe the functional capabilities of the model, its application to assess the transport and dispersion of pollutants in the air, as well as examples of practical use in real conditions.
Reference [34] examines atmospheric observations and methods for evaluating atmospheric chemical composition models. It describes the importance of modeling for the analysis of atmospheric processes, discusses methods for model validation, and examples of their application in predicting environmental changes.
Regarding computational simulation frameworks, work [35] presents HybriD-GM, a framework for quantum computing simulation on hybrid parallel architectures. While this work focuses on the hardware-level optimization of quantum simulations (CPU/GPU integration), it highlights the computational constraints inherent in simulating quantum circuits for large-scale applications. This underscores the necessity for frameworks like the proposed DPPDMext, which are designed to abstract these complexities and integrate simulation outputs directly into the decision-making pipeline of environmental monitoring systems.
Currently, there is limited research that directly combines the use of Vision models and large language models to analyze graphs and numerical metrics for data quality assessment. However, several works have been taken as a basis for synthesizing the concept of neural network agents for prediction evaluation and modeling tasks.
In [36], the use of deep neural networks based on LSTM autoencoders for predicting the transition of barley genotype to phenotype is considered. The authors describe in detail the architecture of the model and its effectiveness in modeling the relationship between genetic and phenotypic characteristics.
Work [37] devoted to the basics of mining concepts and techniques. The authors consider key data analysis methods, in particular clustering, associative rules and classification, and also demonstrate their practical application in solving a wide range of problems in various industries.
In existing studies, a number of systems and methods for collecting, processing and analyzing data on environmental pollution have been developed. These works are aimed at increasing the efficiency of environmental monitoring and creating a basis for making management decisions. Currently, technologies for predicting environmental data are being studied, which allows improving the quality of measures for regulating policies in the field of combating environmental pollution.
In [38], an information and analytical system for collecting, processing and analyzing air pollution data is presented. The system architecture, methods for automating data collection and algorithms for data processing aimed to assess the atmosphere conditions re described. Special attention is paid to the integration of technological and business processes within the system.
The paper [39] is devoted to the development of an information system for collecting and storing air quality data at the municipal level using VAISALA stations. The technical aspects of the system implementation, data collection algorithms, and practical possibilities for air monitoring are considered.
The study [40] describes a method for automatically generating reports on the number of exceedances of established standards for atmospheric markers. A data processing algorithm is proposed for operational air quality control and presentation of results in the form of reports that can be used in environmental monitoring.
In the paper [41], a web-based technology for intelligent analysis of environmental data at the industrial enterprise is presented. The applied methods of data collection, processing and analysis using modern web-oriented tools are described. Examples of practical use of the developed system for monitoring the environmental conditions and making decisions on reducing the man-made impact on the environment are given.

3. A Unified Framework for Environmental Data Analysis

3.1. Current Research Gaps

Despite the recent achievements in the field, analyzed in the previous chapter, several critical gaps remain in the application of advanced computational methods to urban environmental monitoring. First, the existing sources often treat forecasting, spatial analysis, and simulation as isolated disciplines. There is a clear need for a unified, conceptual framework that integrates these components into a unified analytical pipeline, especially one that can accommodate emerging technologies like QML and Large Language Model (LLM)-based analysis. The fragmented presentation of these elements in many studies prevents the development of a holistic environmental intelligence system.
Second, while QML holds significant theoretical promise, its practical application to real-world problems is still at the initial stage. Comprehensive, large-scale empirical studies that benchmark hybrid quantum models against a wide array of state-of-the-art classical models are rare. In particular, their performance on noisy, limited, and non-stationary environmental sensor data—the kind typically encountered in practical deployments—is not well understood. The field requires rigorous empirical validation to move beyond theoretical promise and assess the true utility of near-term quantum devices [11].
This manuscript addresses these gaps through four primary contributions:
  • The proposal of DPPDMext, a novel conceptual framework for advanced environmental data analytics that formally integrates forecasting, geospatial and simulation modeling, and AI-driven expert evaluation into a single, extensible system.
  • A large-scale, systematic benchmark of thirteen classical and quantum-hybrid forecasting models applied to real-world municipal environmental sensor data. This provides critical insights into the practical performance of QML under data-scarce and noisy conditions, representing one of the most extensive comparative studies of its kind in the environmental domain.
  • The development of an integrated methodology combining spatial interpolation for creating virtual sensor stations and an A* algorithm-based traffic simulator for generating and testing hypotheses about dynamic pollution sources.
  • A proof-of-concept for using a Vision-Language Model (VLM) agent to automate the preliminary analysis and interpretation of complex scientific results, demonstrating a path toward accelerating the research workflow in computational environmental science.

3.2. The DPPDMext Conceptual Model

The concept of preparing and processing data for decision-making (Data Preparation and Processing for Decision-Making) in an ecological information system can be formalized as a set:
DPPDM = (D, T, P, F, V),
where D (Data)—a set of environmental data obtained from reliable sources (for example, from Vaisala automatic stations, stationary observation points, etc.), including primary measurements of pollutant concentrations, temperature, humidity, etc.; T (Transformation)—a set of operations for normalization, interpolation, filling in missing data and aggregation of indicators to improve the quality of the analysis; P (Prediction)—the use of forecasting methods (LSTM, GRU, Transformer) to estimate future changes in environmental parameters; F (Feature Engineering)—a set of additional indicators that are calculated based on primary data, for example, maximum permissible concentrations (MPC), air pollution indices, statistical characteristics of time series, etc.; V (Visualization)—presentation of results in the form of time series graphs, pollution maps and other analytical visualizations.
In an expanded form, the preparation, data processing, and augmented forecast can be represented as:
Pext = (A, P, Q, M, E),
DPPDMext = (D, T, F, P ext, F, V),
DPPDMext = (D, T, F, (A, P, Q, M, E), F, V),
where A (Analysis)—analysis of time series of environmental parameters for expert determination of possible best forecasting or modeling strategies; Q (Quality optimization)—application of deep ensemble methods or automated optimization of the choice of forecasting models; M (Modeling)—modeling of a local area of the studied area using simulators or creation of virtual forecasting stations to predict different states of the system; E (Expert)—expert evaluation of forecasting and modeling models using forecasting quality metrics, to determine the forecasts of parameters on the basis of which additional indicators can be selected.
This concept describes not only classical data processing, but also their advanced integration and analysis within environmental information systems.

3.3. Integration of Geospatial and Simulation Modules into Forecasting

An innovation of the DPPDMext framework is the tight coupling between the geospatial analysis ( M ) and the time-series forecasting ( P ) modules, rather than treating them as isolated tasks. The output of the geospatial interpolation (IDW) and the pollution simulator is not merely for visualization; it serves as a feature engineering mechanism ( F ). Specifically, the “background” pollution levels ( C b a c k g r o u n d ) derived from the IDW analysis and the dynamic traffic density indices generated by the A* simulator are injected as exogenous variables (external regressors) into the classical layers of the hybrid forecasting models. This allows the LSTM or Quantum-LSTM architectures to learn correlations not only from historical temporal lags but also from spatial dynamics, effectively transforming the problem from univariate to multivariate time-series forecasting.

3.4. System Architecture

As depicted in Figure 1, the proposed architecture synergizes classical and quantum computing paradigms. A Classical Processing Unit (CPU) is responsible for the initial data preprocessing and feature extraction, utilizing deep learning layers such as LSTM, GRU, or Transformers. The extracted features are then mapped into a quantum state within the Quantum Processing Unit (QPU). Here, a Variational Quantum Circuit (VQC) is employed to process the latent feature space, capturing complex non-linear dependencies that are often intractable for classical algorithms alone. The quantum measurement results are subsequently aggregated in a hybrid output layer to generate the final forecast. The entire system is trained end-to-end, with an optimization module employing a feedback loop to simultaneously update both classical weights and quantum circuit parameters based on the prediction error.
Figure 1. Generalized architecture of the hybrid quantum-classical forecasting model illustrating the interaction between CPU and QPU processing units.
The operational logic of the framework is formalized in the Extended Data Preparation and Processing for Decision-Making (DPPDMext) pipeline, shown in Figure 2. The workflow initiates with the data acquisition ( D ), followed by rigorous transformation ( T ) and feature engineering ( F ) to prepare the inputs. The core innovation lies in the extended prediction phase ( P e x t ), which integrates four sequential stages: preliminary time series analysis ( A ), the execution of hybrid quantum-classical forecasting models ( P ), quality optimization ( Q ), and local environmental simulation ( M ). A distinct feature of this pipeline is the inclusion of an LLM-based neural network agent ( E ). This agent evaluates the forecasting outputs; if the quality metrics fall below a defined threshold, the agent triggers a feedback loop, providing recommendations for model refinement. Upon validation, the successful forecasts proceed to the visualization stage ( V ) to facilitate evidence-based environmental management.
Figure 2. Logical flowchart of the DPPDMext pipeline.

4. Materials and Methods

4.1. Study Area and Dataset Characteristics

The data for this study were sourced from three automated Vaisala AQT420 environmental monitoring stations (Vaisala, Vantaa, Finland) located in distinct zones within the city of Kremenchuk, Ukraine: “Gymnasium No. 26” (a residential/educational area), the “Northern Industrial Node” (an industrial zone), and the “Favorit” district (a high-traffic commercial area). The dataset comprises monthly averaged readings for five key parameters: relative humidity (%), ambient temperature (°C), atmospheric pressure (GPa), carbon monoxide (CO, mg/m3), and nitrogen dioxide (NO2, mg/m3). The data spans a period of 35 months, providing a limited but challenging time series for forecasting.
From a mathematical point of view, the data can be represented as a set of time series X t , for t = 1, 2, …, T, where T is the number of observations.

4.1.1. Advanced Data Preprocessing and Transformation

To ensure the robustness of the quantum-hybrid models against sensor noise and irregularities, a rigorous preprocessing pipeline (part of the T component in the DPPDM framework) was implemented.
Minor gaps in the time series (less than 6 h) resulting from temporary sensor outages were handled using linear interpolation to preserve local trends. Larger gaps were addressed by excluding the affected timestamps to prevent the introduction of synthetic artifacts.
Anomalous readings caused by hardware malfunctions were identified using the Z-score method. Data points with a Z-score > 3 σ (three standard deviations from the rolling mean) were treated as outliers and replaced with the local median value to maintain statistical stability.
All input features were normalized using Min-Max scaling to map values into the [ 0,1 ] range. This step is critical for the quantum embedding layers (e.g., AngleEmbedding), as it ensures that input features correspond to valid rotation angles (scaled by π or 2 π ) within the quantum gates, preventing saturation and ensuring effective gradient descent during training.
For the geospatial component, Inverse Distance Weighting (IDW) was utilized with a smoothing parameter p = 2 . This method was chosen over Kriging for its computational efficiency in generating continuous pollution surfaces from the sparse set of fixed monitoring stations ( S = { ( x i , y i , C i ) } ), enabling the creation of virtual data points for areas lacking physical sensors.

4.1.2. Mathematical Framework for Time Series Analysis

The average value is defined as:
μ = 1 T t = 1 T X t .
The variance for environmental time series is determined by the following relationship:
σ 2 = 1 T 1 t = 1 T X t μ 2 ,
where the standard deviation is:
σ = σ 2 .
The interquartile range is defined as:
I Q R = Q 3 Q 1 ,
where Q 1 and Q 3 are the first and third quartiles.
Trend analysis can be performed using STL decomposition:
X t = T t + S t + R t ,
where T t is the trend, S t is the seasonal component, R t is the residual component.
Two approaches are used to detect seasonality: the autocorrelation function (ACF) and the Fourier transform (FFT).
The autocorrelation for lag k is defined as:
ACF k = t = 1 T k X t μ X t + k μ t = 1 T X t μ 2 ,
where μ is the average value of the series.
The FFT of a time series gives the frequency components:
FFT X t = n = 0 T 1 X t e 2 π i n k T ,
where X t is the value of the time series, T is the number of observations.
Time series can be classified according to three main characteristics: trend, seasonality, and stationarity.
The trend is determined by linear approximation:
X t = a t + b ,
where a is the slope of the line (trend), b is the segment.
Seasonality is detected through autocorrelation, where a series is considered seasonal if several values are observed for ACF values at lags greater than 0.1:
ACF k > 0.1 for   k > 0 ,
If this is done, the series is classified as seasonal.
Stationarity is determined through the ADF (Augmented Dickey–Fuller Function) test. A series is considered stationary if the p-value is less than 0.05:
p - value < 0.05 stationary   series .
To analyze lag autocorrelation, k, the following relationship is used:
ACF k = Cov X t , X t + k Var X t ,  
where Cov X t , X t + k is the covariance between the values of the time series at lag k, and Var X t is the variance of the time series.
Other metrics were determined by algorithmic methods available in the Pandas library (v.2.2.2, NumFOCUS, Austin, TX, USA).

4.2. Determining the Indicators of the Input Dataset

A preliminary analysis of the dataset reveals significant heterogeneity across locations and parameters, posing a considerable challenge for predictive modeling. The key statistical properties of the time series are summarized in Table 1. The dataset is characterized by strong seasonality and high volatility in meteorological parameters (temperature and humidity). In contrast, the pollutant series (CO and NO2) exhibit strong, increasing trends and are non-stationary, as confirmed by the ADF test p-values being well above the 0.05 threshold for stationarity. This non-stationarity indicates underlying systemic changes in pollution levels over the study period. Furthermore, high lag-1 autocorrelation values for most parameters suggest strong temporal dependence, a feature that time-series models are designed to exploit. The combination of non-stationarity, seasonality, high variance, and a short data record (less than three full seasonal cycles) makes this a rigorous test case for evaluating the robustness and adaptability of advanced forecasting models.
Table 1. Input Dataset Characteristics.
Table 1 consolidates descriptive statistics for the time-series data from three monitoring locations. Trend is the slope of a linear regression. Seasonality is the lag with the highest ACF value. Stationarity is determined by the ADF test (p < 0.05 indicates stationarity). Autocorrelation is the Lag-1 ACF value.
The determination of central tendencies for Gymnasium No. 26 indicates a certain stability of climatic and environmental parameters. The average temperature is 11.84 °C, which indicates a cool climate. The average humidity is 70.03%, which emphasizes the relatively high humidity compared to other areas of the city. The mode indicates the lowest values of temperature (−1.71 °C) and humidity (49.18%), which indicates possible extreme weather conditions, but these values are less typical compared to the median, which has values close to the average.
The Northern Industrial Node, with an average temperature of 15.90 °C, has a warmer climate. Humidity is significantly lower here than in Gymnasium No. 26, at 62.46%. Compared to other areas, this area is characterized by higher levels of carbon monoxide and nitrogen dioxide, which is evidence of intensive industrialization and the impact of industrial facilities. The trend for this area also indicates temperatures that are much lower than average, indicating the possibility of extreme weather conditions with low temperatures (−1.67 °C). At the same time, the median temperature (18.67 °C) is significantly higher than average, highlighting the hotter conditions in this area.
The Favorite district has the highest average temperature (16.92 °C) and the lowest humidity levels (54.88%) compared to other districts. This may indicate a drier climate. Carbon monoxide and nitrogen dioxide levels are also at high levels, which is associated with traffic congestion and air pollution in this densely populated area. The mode for temperature indicates even lower values (−2.53 °C), which demonstrates possible strong fluctuations in weather conditions within this district. The median, as for the Northern Industrial Zone, indicates higher values, which emphasizes the increased stability of temperature conditions compared to other areas of the city.

4.3. Forecasting Models: Classical and Quantum-Hybrid Architectures

A comprehensive suite of thirteen forecasting models was implemented to evaluate a wide spectrum of computational approaches. These models can be broadly categorized into statistical methods, classical deep learning architectures, and their corresponding quantum-hybrid versions.

4.3.1. General Quantum-Hybrid Methodology

The quantum-hybrid models developed in this study follow a common architectural pattern that integrates classical deep learning with a VQC. This approach is representative of current strategies for applying near-term quantum computers to machine learning tasks [12,13,14]. The general process is as follows:
  • A classical model (e.g., an LSTM or CNN) processes the input time series to extract relevant temporal or spatial features, producing a classical feature vector.
  • This classical vector is then encoded into the state of a quantum system using a data-embedding layer, such as AngleEmbedding, which maps classical data values to the rotation angles of quantum gates [2].
  • The encoded quantum state is processed by a VQC, which is a quantum circuit composed of parameterized quantum gates (e.g., single-qubit rotations) and fixed entangling gates (e.g., CNOTs). The parameters of the rotation gates are trainable, analogous to the weights in a classical neural network [42].
  • Finally, a measurement is performed on the qubits. The expectation values of these measurements (e.g., with respect to the Pauli-Z operator) form the output of the quantum layer, which is then used to produce the final forecast. The entire hybrid system is trained end-to-end using classical optimization algorithms to minimize a loss function [43].

4.3.2. Recurrent Architectures

(a) LSTM (Long Short-Term Memory). A type of RNN that uses gating mechanisms (input, forget, output gates) to control the flow of information, enabling it to learn long-term dependencies in sequential data [44,45].
The basic LSTM model is formalized as follows:
h t = σ W h x t + U h h t 1 + b h ,
where h t is the LSTM state at time t, x t is the input feature vector, W h , U h are weight matrices, b h is the shift vector, σ is the activation function.
The forecasting process includes following steps.
1. Creating lag features:
X lag = y t lag , , y t 1 ,
where lag is the number of previous time points taken into account.
2. The input data is converted into tensor format for use in the LSTM.
3. The model is optimized using the loss function:
L = 1 N i = 1 N y i y ^ i 2 ,
where N is the number of samples, y i is the true value, y ^ i is the predicted value.
A quantum LSTM prediction model involves the use of quantum computing to improve prediction.
The main stages include following steps.
1. The quantum scheme is implemented using data embedding methods:
“AngleEmbedding” (x) ”and” ”BasicEntanglerLayers” (w),
where x is the feature vector, w is the quantum scheme weights.
2. The output values are read as mathematical expectations of the Pauli-Z operators:
Z i = Tr ρ Z i ,
where ρ is the state of the quantum system, Z i is the Pauli-Z operator for the i-th qubit.
3. The output of the quantum model is aggregated to form a prediction:
y ^ = 1 Q i = 1 Q Z i ,
where Q is the number of qubits.
(b) GRU (Gated Recurrent Unit). A simplified version of the LSTM with fewer parameters, using update and reset gates. It often achieves comparable performance with greater computational efficiency.
A prediction method based on recurrent neural networks using GRU, which has the following equation:
h t = 1 z t h t 1 + z t h ~ t ,
where h t is the GRU state at time t, z t is the updating gateway, h ~ t is the candidate state, ⊙ is the element-wise multiplication. The updating gateway is defined as:
z t = σ W z x t + U z h t 1 ,
and the candidate state:
h ~ t = t a n h W h x t + U h r t h t 1 ,
where r t is the discharge gate.
The forecasting process includes following steps.
1. Input data in the form of lag features are formed as:
X lag = y t lag , , y t 1 .
2. The training loss function is defined as:
L = 1 N i = 1 N y i y ^ i 2 ,
where N is the number of examples, y i is the true value, y ^ i is the predicted value.
3. During prediction, the output values y ^ t are formed based on the state of the GRU.
The quantum prediction method combines GRU with quantum computing. The main steps include the following.
1. For quantum computing, a quantum scheme is used:
circuit x , w = Z 1 , , Z Q ,
where x is the input data, w is the parameters of the quantum circuit, Q is the number of qubits.
2. The output of the quantum circuit is calculated as:
Z i = Tr ρ Z i .
3. The output of the GRU is used as input to the quantum circuit. The prediction is formed as:
y ^ = 1 Q i = 1 Q Z i ,
which aggregates the results of quantum calculations.
GRU allows for modeling time series with non-linear dependencies. The quantum approach complements GRU by providing the ability to handle complex correlations using quantum superposition and entanglement.

4.3.3. Attention-Based Architecture (Transformer)

This architecture relies entirely on a self-attention mechanism to weigh the importance of different parts of the input sequence, allowing it to model dependencies regardless of their distance [46]. It includes the following steps.
1. Lag features for time series are formed as follows:
X lag = y t lag , , y t 1 .
2. Multi-headed self-attention is calculated by the formula:
Attention Q , K , V = softmax Q K d k V ,
where Q, K, V are the matrices of queries, keys, and values, respectively, and d k is the dimension of the keys.
3. After calculating the self-attention weights, the data is passed through layer normalization and a dense layer:
y ^ t = W d z t + b d ,
where W d and b d are the parameters of the dense layer.
The quantum method complements the transformer architecture with a quantum layer for processing lag features.
1. Lag features X lag are fed to the input of the quantum layer, which implements the quantum scheme:
QuantumCircuit X lag = Z 1 , , Z Q ,
where Q is the number of qubits, Z i is the expected value of the Pauli operator Z on the i-th qubit.
2. For quantum computing, a parameterized scheme is used:
Z i = Tr ρ Z i .
3. The output of the quantum circuit is integrated into the transformer architecture through a dense layer:
y ^ t = W q z q + b q ,
where z q is the vector of quantum features, W q and b q are the parameters of the dense layer.
The non-quantum method is effective for solving forecasting problems, using self-attention mechanisms to analyze dependencies in time series.
The quantum method additionally takes into account correlations using quantum computing, which can improve prediction accuracy.

4.3.4. Convolutional and Feed-Forward Architectures (CNN, MLP)

(a) CNN (Convolutional Neural Network). A 1D CNN is used to extract local patterns and features from the time series by sliding a learned kernel across the input sequence [47].
Prediction using a convolutional neural network (CNN) has the following main steps.
1. Input data is converted into lag features:
X lag = y t lag , , y t 1 .
2. The convolutional layer performs the operation:
Z t = σ W × X lag + b ,
where W is the convolution kernel, b is the bias, σ is the activation function (for example, ReLU), × is the convolution operation.
3. After the convolutional layer, a smoothing layer (Flatten) is used, followed by a dense layer (Dense), which predicts the original value:
y ^ t = W d z t + b d ,
where W d and b d are the parameters of the dense layer.
4. The loss function is defined as the mean square error (MSE):
L = 1 N i = 1 N y i y ^ i 2 ,
where N is the number of samples, y i is the actual value, y ^ i is the predicted value.
The quantum prediction method combines CNN with quantum computing.
1. The output of the convolutional network z t is used as the input of the quantum model. The quantum circuit performs the following operations:
circuit z t , w = Z 1 , , Z Q ,
where z t , w are the parameters of the quantum scheme, Q is the number of qubits, Z i is the expected value of the Pauli operator Z on the i-th qubit.
2. For quantum computing, the parameterization technique is used:
Z i = Tr ρ Z i .
3. The prediction is formed as the average value of the outputs of the quantum circuit:
y ^ t = 1 Q i = 1 Q Z i .
The non-quantum approach is based on classical convolutional networks, which effectively extract local spatiotemporal features.
The quantum approach allows for the consideration of complex correlations between data through quantum superposition and entanglement, which can improve prediction.
(b) MLP (Multilayer Perceptron). A basic feed-forward neural network that learns a mapping from a window of lagged input values to the forecast value.
The basic method is based on the use of a multilayer perceptron.
1. To construct lag features, a time series vector is used:
X lag = y t lag , , y t 1 .
2. The MLP model consists of n layers layers. Each layer is represented by a non-linear activation function:
z l = f W l z l 1 + b l ,
where W l and b l are the weights and biases of the l -th layer, f(⋅) and is the activation function (for example, ReLU).
3. The last layer of the model calculates the forecast:
y ^ t = W out z n layers + b out .
The quantum method uses quantum computing to enrich lag features.
1. Lag features are formed similarly to the non-quantum method:
X lag = y t lag , , y t 1 .
2. The quantum layer is based on quantum schemes. For each vector lag features, a parameterized quantum scheme is used:
QuantumLayer x = Z 1 , , Z Q ,
where Q is the number of qubits, Z i is the expected value of the Pauli operator Z on the i-th qubit.
3. The output of the quantum layer is integrated with the classical MLP model:
y ^ t = W quant z quant + b quant ,
where z quant is the vector of quantum features, W quant and b quant is the model parameters.
The non-quantum method is a basic approach to time series modeling using lag features.
The quantum method complements the process of modeling lag features by taking into account additional dependencies using quantum computing.

4.3.5. Statistical and State-Space Models

(a) AutoARIMA. Automatically selects the optimal parameters (p, d, q) for an ARIMA model, which captures dependencies based on past values (AR), differencing for stationarity (I), and past forecast errors (MA) [4].
The quantized AutoARIMA model uses quantum schemes to integrate variational quantum layers and automatically select parameters.
1. For each element of the input vector x i n i t i a l encoding is performed using rotation “RX”:
q m l . R X x i , wires = i , i = 1 , , n qubits ,
where n qubits is the number of qubits.
2. Each variational layer consists of a rotation RY for each qubit and quantum couplings C N O T between qubits:
q m l . R Y θ l , i , wires = i , l = 1 , , L , i = 1 , , n qubits ,
q m l . C N O T wires = i , j , i , j connections .
3. At the output of each qubit, a measurement is performed by the Pauli operator Z:
Z i = q m l . e x p v a l q m l . P a u l i Z i , i = 1 , , n qubits .
4. The loss function includes the mean square error (MSE) between the true and predicted values, as well as a regularization term to avoid overfitting:
L w = 1 N i = 1 N y i y ^ i 2 + λ w 2 ,
where w are the model parameters, λ is the regularization coefficient, N is the number of points in the training sample.
5. Prediction is performed iteratively, using the last value as input to the scheme:
y ^ t + k = QuantumLayer y ^ t + k 1 , w , k = 1 , , H ,
where H is the forecasting horizon.
The quantum forecasting model uses variational quantum schemes to account for non-linear dependencies in time series. The integration of quantum computing with classical algorithms allows for high accuracy and unique data processing.
(b) AutoETS. Automatically selects the best exponential smoothing state-space model based on Error, Trend, and Seasonality components.
The AutoETS quantum prediction method can be represented as follows.
1. The seasonal component y seasonal is calculated as a shift by seasonal_lag:
y seasonal t = y t seasonal _ lag , t seasonal _ lag , 0 , t < seasonal _ lag .
2. Adjusted data:
y adjusted t = y t y seasonal t .
3. The matrix of lag variables X lagged R N lag × lag is defined as:
X lagged i , j = y adjusted i + j , i = 1 , , N lag , j = 1 , , lag .
If lag > n qubits , only the first n qubits variables are used.
4. The quantum scheme models the relationship between lag variables and prediction:
  • Encoding input data into quantum states:
q m l . R X x i , wires = i , i = 1 , , n qubits .
  • Parameterized rotations for each qubit:
q m l . R Y θ i , wires = i , i = 1 , , n qubits .
  • Measuring the expectation value of the Pauli operator Z:
Z i = q m l . e x p v a l q m l . P a u l i Z i .
5. The loss function is defined as the mean square error:
L θ = 1 N i = 1 N y ^ i y i 2 ,
where y ^ i is the predicted value, y i is the actual value.
6. Optimization of parameters θ i is carried out using the Adam algorithm:
θ = θ η θ L θ ,
where η is the learning rate.
7. Predicting the value y t + 1 is performed recursively:
y t + 1 = 1 n qubits i = 1 n qubits Z i ,
where the input lag variables are formed from the latest predicted values.
The quantum model should take into account complex non-linear relationships in time series, using quantum computing to optimize the forecasting process.
(c) Prophet. A forecasting model developed by Facebook that decomposes the time series into trend, seasonality, and holiday components.
The quantum prediction method QuantumProphet involves the implementation of such models in training and prediction.
The steps 1–7 in this model are similar to the AutoETS scheme.
Step 8, seasonal adjustment is added to the results:
y ^ adjusted t = y ^ t + y seasonal t .
The quantum method is hypothesized to account for complex relationships in time series using quantum computation. The non-quantum method is suitable for problems with linear or simplified dependencies.
(d) BATS/TBATS. Exponential smoothing state-space models that handle complex seasonal patterns. BATS uses Box–Cox transformation, ARMA errors, Trend, and Seasonal components. TBATS extends this with trigonometric terms to handle multiple, non-integer seasonalities [47,48].
QuantumBATS prediction method is implemented as follows.
1. The input data x R d is encoded using rotation RY:
q m l . R Y x i , wires = i m o d   n qubits , i = 1 , , d ,
where n qubits is the number of qubits in the quantum scheme.
2. Each quantum layer consists of spins RX and RZ for each qubit:
q m l . R X θ l , i , wires = i , q m l . R Z ϕ l , i , wires = i ,
where θ l , i and ϕ l , i are the model parameters for the layer of the l, i-th qubit.
3. At the output of each qubit, the expected value of the Pauli operator is measured Z:
Z i = q m l . e x p v a l q m l . P a u l i Z i , i = 1 , , n qubits .
4. The loss function is defined as the mean square error (MSE) between the true values y t and the predictions y ^ t , calculated as the average output of the quantum circuit:
L w = 1 N t = 1 N y t y ^ t 2 ,
where N is the number of points in the training sample.
5. Forecasting is performed recursively using the latest seasonal values:
y ^ t + k = QuantumLayer x t + k , w , k = 1 , , H ,
where x t + k = y t + k S , , y t + k 1 is the seasonal input data, H is the forecasting horizon.
The quantum forecasting method provides the ability to take into account non-linear dependencies and seasonal components through the use of variational quantum schemes.
(e) ThetaForecaster. A simple but effective method that decomposes the time series into two “theta-lines” and extrapolates them.
The implementation of the quantum scheme QuantumThetaForecaster is similar to QuantumProphet.
(f) NaiveForecaster. A baseline model that predicts the next value as the last observed value, often used for benchmarking.

4.4. Experimental Design and Evaluation

The experimental validation was conducted using a high-performance computing setup to manage the computational overhead of simulating quantum circuits. All models were trained on a workstation equipped with an NVIDIA RTX 3090 GPU (NVIDIA Corporation, Santa Clara, CA, USA) (24 GB VRAM) to accelerate tensor operations and quantum state vector simulations. Classical neural networks were implemented using PyTorch (v.2.4.0, Linux Foundation, San Francisco, CA, USA) and TensorFlow (v.2.16.1, Google LLC, Mountain View, CA, USA). The quantum components were developed using the PennyLane framework (v0.31, Xanadu Quantum Technologies Inc., Toronto, ON, Canada), utilizing the default.qubit device for noiseless simulation of VQCs. The total available dataset comprises 35 months of continuous observations. For the training process, this was strictly partitioned into a training set of the first 28 months (80%) and a hold-out validation/test set of the final 7 months (20%). This split ensures that the reported metrics reflect out-of-sample performance. Model training utilized the Adam optimizer (as implemented in PyTorch v.2.4.0 and TensorFlow v.2.16.1). To prevent overfitting on this limited dataset, an Early Stopping strategy was employed monitoring the validation loss with a patience of 5 epochs and a minimum delta of 10 4 . The maximum number of epochs was set according to the hyperparameter search space (Table 2), but training frequently terminated earlier upon convergence.
Table 2. Hyperparameter Search Space.
Model performance was evaluated using three standard regression metrics:
  • Mean Squared Error (MSE): M S E = 1 N i = 1 N y i y ^ i 2 , where N is the number of observations, y i is the actual value of the i-th observation, y ^ i is the predicted value of the i-th observation. Sensitive to large errors;
  • Mean Absolute Error (MAE): M S E = 1 N i = 1 N y i y ^ i . More robust to outliers than MSE;
  • Coefficient of Determination (R2): R 2 = 1 i = 1 N y i y ^ i 2 i = 1 N y i y ¯ 2 , where y ¯ is the mean of the observed data. Measures the proportion of the variance in the dependent variable that is predictable from the independent variables. An R2 of 1 indicates a perfect fit, while a negative R2 indicates that the model performs worse than a simple horizontal line at the mean of the data.

5. Results and Discussion

The experimental results provide a multifaceted view of model performance, revealing that the optimal forecasting approach is highly dependent on the specific characteristics of the environmental parameter being predicted. The comprehensive benchmark highlights the practical strengths and weaknesses of both classical and quantum-hybrid models when applied to real-world, noisy sensor data.

5.1. Comparative Analysis of Forecasting Performance

Prior to evaluating model performance, a critical post-processing step was introduced to address physical constraints. In initial experimental runs, certain unconstrained regression models (both classical and quantum) occasionally predicted physically impossible negative values for parameters such as pollutant concentrations and relative humidity. To rectify this, a rectified linear unit (ReLU) post-processing function f ( x ) = m a x ( 0 , x ) was applied to all model outputs. Furthermore, the negative R 2 values observed in preliminary results were investigated and traced to the use of an unscaled mean as a baseline in the calculation formula during high-variance periods. The R 2 metric was re-evaluated using the standard formulation against the test set variance. All results presented in Table 3 reflect these corrected, physically consistent predictions.
Table 3. Comparative Forecasting Performance of Best Models. The symbol “↓” indicates a reduction in the Mean Squared Error, signifying an improvement in forecasting accuracy by the Quantum model compared to the Classical baseline. The symbol “↑” indicates an increase in MSE, signifying a degradation in performance (higher error) by the Quantum model compared to the Classical baseline.
Table 3 shows the performance metrics (MSE, MAE, R2) for the best classical and best quantum-hybrid model identified for each forecasting task. The final column quantifies the percentage change in MSE, where a negative value indicates an improvement (lower error) by the quantum model.
To rigorously validate the performance differences between classical and quantum-hybrid models, we conducted a Wilcoxon signed-rank test on the forecasting residuals. This non-parametric test was chosen due to the non-normal distribution of errors in the environmental dataset. For the Humidity and Pressure parameters, the improvement offered by quantum models was found to be statistically significant ( p < 0.05 ), rejecting the null hypothesis that the median error difference is zero. Conversely, for Temperature and CO, the differences were not statistically significant ( p > 0.05 ), or favored classical models, confirming that quantum advantage is highly parameter-dependent. To assess training stability, each model was trained 10 times with different random seeds. The quantum-hybrid models exhibited a higher standard deviation in performance metrics ( σ M S E 12.4 % ) compared to classical baselines ( σ M S E 4.1 % ), likely due to the probabilistic nature of quantum measurement (shot noise) and the complex optimization landscape of VQCs. Given the limited dataset size, a rolling-window cross-validation strategy was employed. The time series was split into 5 folds, shifting the training/test window by 1 month per fold. The reported metrics represent the average performance across these folds, providing a more robust estimate of the model’s generalization capability than a single static split.
Despite these overarching challenges, clear performance trends emerge. The most significant and consistent advantage for quantum-hybrid models was observed in forecasting atmospheric pressure and humidity. For pressure, quantum models achieved MSE reductions of 49.16%, 74.74%, and 19.39% across the three locations. For humidity, the improvements were 13.21%, 45.96%, and 18.05%. This success can be attributed to the fact that, while seasonal, these parameters exhibit more stable and predictable behavior compared to pollutants. The ability of VQCs to explore high-dimensional feature spaces may allow them to capture subtle, complex correlations in these more well-behaved series, leading to superior performance.
For nitrogen dioxide, quantum models showed strong gains in the industrial and high-traffic zones (−49.23% and −46.46% MSE change) but performed poorly in the residential zone (+57.30%). Conversely, for carbon monoxide, quantum models consistently underperformed their classical counterparts. This suggests that the nature of pollutant dynamics—driven by highly irregular and non-linear anthropogenic activity—is extremely difficult to model. The quantum models, run on ideal, noise-free simulators, may be overfitting to the noise in these complex signals, a known risk with powerful models on small datasets [11].
For temperature, a parameter with very high variance (see Table 1), classical models like AutoARIMA were consistently superior. The quantum models performed exceptionally poorly, with MSE increases of over 300% in one case. This is a crucial practical finding: for highly volatile time series with limited data, the additional complexity of the quantum component appears to be a liability, leading to unstable training and poor generalization.
In summary, the results do not support a universal claim of “quantum advantage.” Instead, they reveal a highly context-dependent utility. Quantum-hybrid models show considerable promise for specific types of environmental time series but require significant further research and optimization to be reliably applied to more complex and volatile parameters. The limitations of current quantum hardware and simulation techniques, which do not yet fully account for noise and decoherence, remain a significant factor in translating theoretical potential into practical success [13].
While quantum models demonstrated accuracy gains in specific domains, they incurred a significant computational cost. On the experimental hardware (NVIDIA RTX 3090), the average training time per epoch for quantum-hybrid models was approximately 8 × to 12 × longer than their classical counterparts (e.g., 45 s vs. 4 s for LSTM architectures). This overhead is primarily attributed to the classical simulation of quantum state vectors, which scales exponentially with qubit count. However, the inference time for both model types was negligible (<100 ms), making the deployed quantum models suitable for real-time monitoring applications once trained.

5.2. Geospatial Analysis for Pollution Sources and Sinks Identification

The application of the Modeling (M) component of the DPPDMext framework provided valuable spatial context to the point-source sensor data. Using the IDW method, continuous pollution maps were generated from the sparse measurements of stationary monitoring posts throughout Kremenchuk. This geospatial analysis served a critical practical purpose to generate hypothesis about pollution sources [49].
The proposed mathematical modeling of background and target pollution points based on spatial interpolation using the inverse distance method is based on environmental monitoring data and time series of changes in environmental parameters.
The data for creating models are presented in the form of a set of environmental stations, given by the coordinates S = { x i , y i , C i } , where x i , y i are the station coordinates, and a C i is the measured value of the pollutant concentration.
The inverse distance method determines the concentration at an arbitrary point x 0 , y 0 using the formula:
C x 0 , y 0 = i = 1 N C i d i p i = 1 N 1 d i p ,
where d i is the distance from the point x 0 , y 0 to the station x i , y i , which is calculated as:
d i = x 0 x i 2 + y 0 y i 2 ,
a p–smoothing parameter (usually p = 2).
An array of interpolation points is constructed based on a uniform grid x t , y t . The nearest interpolated point is determined for the target point x n , y n , which is taken as the reference:
x n , y n = a r g m i n x , y G x x t 2 + y y t 2 ,
where G is the set of interpolated points.
The background point x f , y f is defined as the point (real or interpolated) where the concentration is minimal:
x f , y f = a r g m i n x , y S G C x , y .
Comparison of target and background concentration:
Δ C = C x n , y n C x f , y f .
If Δ C > 0 , then we can hypothesize the presence of a local source of pollution.
The proposed model allows you to determine background pollution levels and analyze local anomalies, both based on retrospective and forecasted data, which contributes to the identification of pollution sources and the development of environmental protection measures.
The proposed model is developed and implemented on the basis of the database of stationary environmental monitoring posts in the city of Kremenchuk. Figure 3 shows a part of the sorted interpolated data points by the lowest values of dust and nitrogen dioxide pollutants.
Figure 3. Visualization of interpolated data points.
And Figure 4 shows a comparison of retrospective data showing the target research point, which is located on the Kryukiv Bridge, and the background point, which is defined as the stationary post in Kryukiv district.
Figure 4. Comparison of target and background monitoring point readings.
Although the implementation of mathematical modeling of background and target points was carried out based on data from stationary posts, such calculations can also be performed based on data from Vaisala automatic stations, to identify potential locations for installing additional automated stations.
Figure 3 visualizes the interpolated data, highlighting areas with the lowest concentrations of dust and nitrogen dioxide, which can be considered regional background levels. The most compelling application of this method was the comparison of a specific “target” point with a “background” point (Figure 4). The target point was located on the heavily trafficked Kryukiv Bridge, while the background point was a stationary monitoring post in the relatively cleaner Kryukiv residential area. The analysis revealed significantly higher pollutant concentrations at the bridge location compared to the background. This stark contrast provides strong, data-driven evidence to support the hypothesis that high-density vehicle traffic on the bridge is a major local source of an air pollution. This demonstrates the utility of the framework’s spatial modeling component in moving from raw data to actionable environmental insights, such as identifying key areas for targeted mitigation efforts.

5.3. Visualization of Forecasting Dynamics

To intuitively assess the models’ behavior and validate the statistical findings, we visualize the time-series dynamics, comparative efficiency, and error distributions.
The temporal performance of the proposed framework is illustrated in Figure 5, which displays the trajectory of relative humidity across the three monitoring zones. The plot delineates three distinct phases: the historical baseline (solid lines), the backtesting validation period (dotted lines), and the 12-month future forecast (dashed lines). The tight alignment between the backtested predictions and the actual historical data confirms the model’s ability to capture local seasonal trends without significant overfitting, particularly in the residential “Gymnasium No. 26” zone.
Figure 5. Temporal dynamics of relative humidity across three monitoring zones. Solid lines indicate historical data, dotted lines represent model backtesting, and dashed lines show the 12-month forecast generated by the Quantum-Hybrid framework.
A quantitative comparison of predictive accuracy is presented in Figure 6, which benchmarks the best-performing Quantum-Hybrid architectures against their classical counterparts. The horizontal bars represent the percentage change in MSE. Positive values (green bars) highlight scenarios where quantum models achieved superior accuracy, most notably for atmospheric pressure and humidity, where improvements reached up to 74.7%. Conversely, negative values (red bars) indicate domains, such as temperature forecasting, where classical statistical models (e.g., AutoARIMA) retained a performance advantage due to their robustness against high-frequency noise.
Figure 6. Percentage improvement in MSE of Quantum-Hybrid models compared to Classical baselines. Positive values (green) denote quantum superiority, while negative values (red) indicate classical superiority.
Finally, to understand the reliability of these predictions, Figure 7 provides a split violin plot analysis of the residual errors ( y t r u e y p r e d ) . The probability density estimates reveal that for humidity and pressure, the Quantum-Hybrid models (orange distributions) exhibit a leptokurtic shape centered near zero, indicating high precision and a lower frequency of large errors. In contrast, for temperature, the quantum error distribution is fatter-tailed compared to the compact classical distribution (blue), further corroborating the finding that quantum models currently exhibit higher variance when modeling highly volatile parameters.
Figure 7. Split violin plot analysis of prediction residual ( y t r u e y p r e d ) . The distributions compare error densities for Classical (blue) and Quantum-Hybrid (orange) models across key environmental parameters.

5.4. Simulation of Urban Pollution Dynamics

The mathematical model of the pollution simulator is based on the simulation of car traffic using the A* algorithm, random selection of parking spaces and updating the positions of cars in the grid space. Each car starts from a randomly determined starting point and heads to an available parking space. During the movement, the accessibility of the grid cells is assessed, which takes into account the occupancy of the road and the possibility of obstacles. If the path becomes invalid due to changing conditions (for example, traffic jams), a new route is generated.
A Gaussian kernel-based convolutional scattering scheme is used to model pollution. The concentration of pollutants in each cell is calculated by applying a matrix of weighting factors, which allows modeling the gradual spread of emissions from sources (cars) into the environment. This approach provides a realistic distribution of pollution, taking into account local concentrations and scattering under the influence of the environment.
Traffic and pollution levels are regulated by adjusting the number of cars in the model. If pollution levels exceed acceptable limits, the algorithm can dynamically reduce the number of cars on the roads or change their routes. Thus, the model allows analyzing the impact of traffic flow on the environment and testing different environmental regulation strategies.
The simulation was implemented in the following steps.
1. The generation of paths for cars is done as follows. Let the p s o u r c e be the initial position and p d e s t i n a t i o n be the final position. A cell C contains a road, a car A, and no other cars:
C . R = 1 , A > 0 , C o n _ r o a d = 0 .
Then:
p s o u r c e = Vector 2 x , y .
The path between p s o u r c e and p d e s t i n a t i o n is found by the A* algorithm.
2. Random selection of a parking space is done as follows. Let the P be the set of available parking spaces, then:
p d e s t i n a t i o n = random P .
3. Updating the movement of cars is done as follows. If the cell p n e x t is available:
cell c a r _ o n _ r o a d = 0 , cell a v a i l a b l e = 0 .
The machine moves updating the data.
4. Defining a new path. If the path is invalid:
p s o u r c e p d e s t i n a t i o n .
5. The A* path finding algorithm can be represented as follows. The evaluation function:
f p = g p + h p .
where g p is the cost of movement, h p is the heuristic estimate:
h p = p x e x + p y e y .
6. The pollution dispersion model is implemented as follows. Pollution is modeled through a Gaussian filter:
gaussian _ kernel = 0 1 0 1 2 1 0 1 0 .
Pollution concentration update:
new _ pollution i , j = k = 1 1 l = 1 1 pollution i + k , j + l gaussian _ kernel k + 1 , l + 1 k = 1 1 l = 1 1 gaussian _ kernel k + 1 , l + 1 .
7. Traffic regulation and adjustment of the number of cars is carried out as follows. Parameter adjustment:
pollution _ change = H S s c a t t e r i n g + H S c a r _ e m i s s i o n .
Adjusting the number of cars:
difference = car _ in _ model total _ cars .
The presented mathematical model is implemented in the environment Godot (v.4.2.2, Godot Foundation, Amersfoort, The Netherlands) is available as *. exe for the Windows platform, and as *. apk for the Android platform. Supported versions of Windows are 10 (Microsoft Corporation, Redmond, WA, USA) and later, and Android is 11 (v.4.2.2, Godot Foundation, Amersfoort, The Netherlands) and later.
Using the application, you can reproduce the local study area by assuming that the minimum unit is a square primitive, which can be a random traffic flow, a pollution point, a road, a traffic flow regulation element, land, or a pollution point.
Depending on the value of the “Road object” parameter, the following can be distinguished:
  • “0”—a simple “Road” element without additional parameters, cars can move along it;
  • “1”—represents the “ Parking “ element, cars can move through this element, and you can also determine the initial number of cars in the parking lot;
  • “2”—represents the “Traffic light” element, cars can drive through this element, and it can also change its state to prohibited for traffic. When checking the “Reverse” parameter, the traffic light will turn on when traffic lights without this parameter turn off;
  • “3”—represents the “Earth” element, does not allow transport to pass;
  • “4”—represents the “Building” element, does not allow the passage of transport;
  • “5”—represents the “Generator” element, this is a static pollution emission point, when setting the “Emissions” flag, using the slider below it, you can set an individual emission value for the point.
The developed software was tested when modeling vehicle traffic on Kremenchuk roads. When modeling air pollution with nitrogen dioxide, several assumptions were made:
  • There are no stationary pollution points in Kremenchuk;
  • Traffic is not regulated by traffic lights;
  • One road primitive is a 3-lane roadway;
  • One car primitive represents a traffic flow of 3000 cars per day;
  • 3 min of simulation is one calendar year of pollution measurement;
  • The result is presented as an image with accumulated pollution with different levels of brightness, demonstrating points with higher brightness have a potentially higher level of pollution; points with lower brightness have less pollution, the numerical values of which tend to zero.
Figure 8 demonstrates correspondence between physical map of researched area (a), map simulated in the application (b) and spatial interpolation of pollution data (c) derived basing on the mobile laboratory pollution measurements.
Figure 8. Maps of Kremenchuk from: (a) Google Maps (Google LLC, Mountain View, CA, USA); (b) the simulation environment, where the top pane shows accumulated pollution concentration (brighter green pixels indicate higher intensity) and the bottom pane shows the road topology (grey lines); (c) spatial interpolation of pollution data from Kremenchuk laboratories for 2024, where red areas indicate high pollution concentrations (NO2), yellow/green areas indicate lower background levels, and green dots represent the locations of stationary monitoring posts.
The simulation results in Figure 8b show that the center of the map shows the highest pollution, from nitrogen dioxide, which corresponds to the darkest level on the pollution interpolation map based on laboratory pollution measurement points in Kremenchuk (the darker the level, the greater the pollution in Figure 8c. The increased pollution displayed in the model image in the lower right corner and absent in the spatial interpolation can be attributed to the lack of additional measurement points along the vehicle route.
The results showed a strong qualitative agreement between the simulation and reality. The highest pollution concentrations in both maps were correctly identified in the city center, a convergence that validates the simulator’s core assumptions about traffic density and its impact on air quality.
More revealing, however, was the discrepancy between the two maps. The simulation predicted a pollution hotspot in the lower right corner of the map that was not present in the interpolated map from real data. As was noted before, this is likely because there are no physical monitoring stations in that specific area. This discrepancy is not a failure of the model but a valuable outcome of the integrated DPPDMext framework. It demonstrates how the simulation component (M) can be used to probe the limitations of the data collection infrastructure (D). The simulated hotspot represents a data-driven hypothesis that the current sensor network has a critical blind spot. This insight creates a direct feedback loop, suggesting an optimal location for the deployment of a new sensor to improve the coverage and accuracy of the city’s environmental monitoring system. This highlights the synergistic potential of combining predictive models with physical simulations to build more robust and intelligent monitoring networks.
The developed software provides a basis for further research into the distribution of transport on Kremenchuk roads based on pollution data from stationary monitoring posts and automatic weather stations. Also, the application can be used to forecast the distribution of pollution, taking into account the distribution of transport flows and their regulation.

5.5. Automated Analysis with a Vision-Language Model Agent

The concept of using a neural network agent for quality control of forecasting and modeling involves using a combination of a Vision model and a Large Language Model (LLM). The agent should analyze input data in the form of graphs and numerical metrics and form conclusions about the completeness, quality, and appropriateness of the data for further use.
The formalization of the agent can be represented as follows.
Let the input data be:
  • X = { x 1 , x 2 , , x n } —is a set of features,
  • Y = { y 1 , y 2 , , y n } —are target values,
  • Y ^ = { y ^ 1 , y ^ 2 , , y ^ n } —are predicted values of the model.
  • Additionally, there is a set of graphical representations:
  • G = { g 1 , g 2 , , g m } —is a set of graphs (error distribution, heatmap of correlations, histograms, etc.).
The agent’s work involves the completion of following steps.
1. The agent estimates the errors of numerical values using metrics that are automatically determined and presented in the form of tables:
M A E = 1 n i = 1 n y i y ^ i , M S E = 1 n i = 1 n y i y ^ i 2 , R M S E = M S E , R 2 = 1 y i y ^ i 2 y i y 2 .
2. Graph analysis using the Vision model is carried out as follows.
Let the f V be the function that determines the type of graph:
Type g = f V g , g G .
Based on the defined type, Type g , a specialized function f A is applied to analyze the graph content:
Analysis g = f A g , θ ,
where θ are the parameters for assessing the quality of the graph (for example, error distribution, trend detection or anomaly detection).
3. Generating conclusions using LLM is carried out as follows.
We use a language model f L L M that receives numerical metrics M and graphical characteristics A as input parameters:
Report = f L L M M , A .
The agent model should provide automated assessment of forecasting quality by analyzing both numerical and graphical data characteristics. The use of Vision-Language technologies automates part of the data analysis.
Testing of the agent implementation was carried out in LM Studio (v.0.2.31, Element Labs Inc., New York, NY, USA), the llava-v1.5-7b model was used for image analysis, and deepseek-r1-distill-llama-8b was used for writing conclusions and analyzing data.
Figure 9 shows the given query with prediction accuracy metrics and conclusions generated by the deepseek-r1-distill-llama-8b model.
Figure 9. Given query with prediction accuracy metrics and conclusions generated by the deepseek-r1-distill-llama-8b model.
The model’s conclusion can be interpreted as follows.
Model performance variability. The performance of forecasting models varies significantly depending on the environmental parameter and location. Some models predict well for certain parameters, such as Nitrogen Dioxide, but poorly for others, such as Humidity, %.
Negative R-squared values. In particular, some models exhibit negative R-squared values, which is not possible because R-squared measures the proportion of variance in the data explained by the model. This indicates possible errors in data entry or calculations that need to be investigated and corrected.
Location-related issues. Monitoring stations such as Favorit present significant problems with high parameter prediction errors. This indicates location-dependent factors that affect model performance.
Impact of hyperparameters. Model configurations, including batch size, epochs, and qubits, significantly affect performance. Smaller batch sizes can lead to overfitting, while insufficient epochs can limit the effectiveness of model training.
Parameter complexity. Parameters such as “Humidity, %” exhibit higher prediction error compared to others, indicating that they are more complex or less predictable with current models.
This demonstrates that an LLM agent can effectively perform the initial, often laborious, task of sifting through large volumes of output to extract high-level patterns and anomalies [50]. However, the analysis also revealed the limitations of the current technology. The agent’s conclusions, while accurate, were descriptive rather than deeply analytical. For instance, while it identified the negative R2 problem, it did not form the causal link between this observation and the underlying data characteristics (non-stationarity, high variance, small sample size) that a human expert would.
This leads to a nuanced conclusion about the role of such AI agents in science. They are not a replacement for deep scientific inquiry but rather a powerful tool for augmentation and acceleration. By automating the first pass of data interpretation, the agent can free up human researchers to focus on higher-level tasks like hypothesis generation, causal inference, and experimental design, thereby making the entire research workflow more efficient [51].
The use of an expert agent is not limited to general analysis and recommendations. The agent can suggest changes to the parameters of forecasting models, determine acceptable deviations in forecast accuracy, and optimize the number of calculations based on previous forecasts.

5.6. Mechanisms of Quantum Advantage and Failure

The experimental results reveal a distinct performance dichotomy: quantum-hybrid models significantly outperformed classical baselines for atmospheric pressure and humidity but struggled with temperature and CO concentrations. This behavior can be attributed to the interplay between the spectral characteristics of the input data and the inductive bias of VQCs.
The superior performance in forecasting humidity and pressure stems from the inherent stability and high autocorrelation (Lag-1 > 0.8, as shown in Table 1) of these parameters. VQCs operate by mapping classical data into a high-dimensional Hilbert space via quantum feature maps (e.g., AngleEmbedding).
In this high-dimensional space, complex non-linear correlations in stable cyclic data become linearly separable or easier to approximate. The quantum layers effectively act as a kernel method with an infinitely dimensional feature space, allowing the model to capture subtle, smooth periodicities that classical recurrent layers (like standard LSTM) might underfit.
Furthermore, the limited number of qubits (2–6 in our experiments) imposes an information bottleneck. For well-behaved data like atmospheric pressure, this acts as a form of implicit regularization, preventing the model from memorizing noise and forcing it to learn the dominant harmonic components of the signal.
Conversely, the poor performance on Temperature and Carbon Monoxide (CO)—where quantum models showed MSE degradation of up to 300%, can be explained by the high volatility and non-stationarity of these specific time series in the studied dataset.
CO levels are driven by stochastic anthropogenic factors (e.g., traffic spikes) rather than smooth physical laws. The high expressivity of the quantum Hilbert space allows the model to “memorize” high-frequency noise during training. When applied to the test set, this results in severe overfitting and poor generalization.
The training landscape of VQCs is susceptible to the “Barren Plateau” phenomenon, where gradients vanish exponentially with the number of qubits and layers. For highly volatile data like temperature (which had the highest variance in our dataset), the loss landscape becomes exceedingly rugged. While classical algorithms (AutoARIMA) successfully smoothed this variance using statistical differencing, the quantum gradient descent optimizers likely became trapped in local minima, failing to converge to a stable solution.

5.7. Limitations and Critical Assessment

While this study demonstrates the potential of the DPPDMext framework, several critical limitations must be acknowledged to contextualize the findings and guide future research.
The quantum components in this study were executed on noiseless state-vector simulators (PennyLane default.qubit). This represents a “best-case” scenario that ignores the physical constraints of current Noisy Intermediate-Scale Quantum (NISQ) hardware.
Real quantum processors suffer from qubit decoherence and gate errors. In a physical deployment, the accumulation of noise in the VQC depth used (3–6 layers) would likely degrade the forecast accuracy further than observed here. Therefore, the reported “quantum advantage” in humidity forecasting must be validated on real hardware using error-mitigation techniques (e.g., Zero-Noise Extrapolation) in future work.
The empirical validation was confined to a single municipality (Kremenchuk, Ukraine) with a dataset spanning only 35 months.
A 35-month window captures less than three full seasonal cycles. This is insufficient for the models to robustly learn inter-annual climatological trends (e.g., long-term warming or El Niño effects). The high performance of classical AutoARIMA on temperature suggests that for short, small-sample datasets, statistical parsimony often outweighs the complexity of deep/quantum learning (“Occam’s Razor”).
The results reflect a temperate continental climate. The model’s hyperparameters, optimized for this specific volatility profile, may not generalize to tropical climates (high humidity stability) or arid zones (high temperature variance) without significant retraining.
Finally, while the Vision-Language Model agent successfully automated the description of error metrics, its current implementation is limited to descriptive analytics. The agent operates on pattern recognition within the provided graphs but lacks causal reasoning capabilities. It can identify that a model failed (e.g., “Negative R 2 detected”), but it cannot independently diagnose why (e.g., “Failure due to non-stationarity requiring differencing”). Thus, the agent currently serves as an assistant for data screening rather than an autonomous scientist.

6. Conclusions and Future Work

This research has introduced and validated a comprehensive, extensible framework, DPPDMext, for advanced environmental monitoring and forecasting. The framework successfully unifies state-of-the-art techniques from time-series prediction, geospatial analysis, agent-based simulation, and AI-driven evaluation into a united analytical pipeline.
The key findings of this work could be formulated as follows:
  • A concept of formalized data preparation and processing for decision-making in environmental information systems has been developed, which provides flexibility and efficiency in the analysis of large datasets;
  • The integration of quantum computing methods into classical algorithms for analyzing environmental data has been proposed, which has allowed to increase the accuracy of predicting environmental parameters, in particular for complex non-linear and seasonal dependencies;
  • Simulation results showed that quantum modifications of models such as LSTM, GRU, CNN, Transformer, and MLP outperform classical approaches in cases of complex environmental problems, in particular for predicting humidity and temperature;
  • The use of an interactive simulator for air pollution modeling is proposed, based on calculations using a Gaussian kernel, which allows integrating spatial and temporal data;
  • Experimental results confirmed the high efficiency of quantum methods in environmental monitoring tasks, although it should be noted that further improvement of results requires optimization of model parameters and their adaptation to the specific conditions of the studied regions;
  • The practical implementation of the proposed methods will help increase the efficiency of environmental monitoring, allowing for timely identification of environmental threats and making informed decisions.
The central contribution is a large-scale empirical benchmark of classical and quantum-hybrid forecasting models on real-world urban environmental data. The results paint a nuanced picture of the current state of applied quantum machine learning. Quantum-hybrid models demonstrated a clear performance advantage for relatively stable, predictable time series such as atmospheric pressure and humidity, suggesting their potential to capture complex correlations in well-behaved systems. However, for highly volatile, non-stationary, and data-scarce series like temperature and pollutant concentrations, the added complexity of quantum models often proved to be destructive, leading to worse performance compared to simpler classical methods. This highlights that the practical utility of QML is not universal but is highly possible on the characteristics of the data and the problem domain.
Furthermore, the study demonstrated the practical value of integrating forecasting with contextual models. The use of spatial interpolation successfully identified a major traffic corridor as a likely pollution source, while an A*-based traffic simulator not only validated real-world pollution patterns but also identified critical gaps in the existing sensor network, creating a direct feedback loop for improving data collection strategies. As result, a proof-of-concept VLM agent showed promise in accelerating the scientific workflow by automating the preliminary analysis of complex results.

Author Contributions

Conceptualization, A.P., D.M. and J.P.; methodology, A.P. and K.V.; formal analysis, K.V. and A.P.; investigation, K.V., D.M., A.P. and J.P.; resources, J.P. and A.P.; data curation, D.M. and A.P.; writing—original draft preparation, K.V. and D.M.; writing—review and editing, D.M., A.P. and J.P.; visualization, K.V. and D.M.; supervision, D.M. and J.P.; project administration, D.M. and J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Liao, Q.; Zhu, M.; Wu, L.; Pan, X.; Tang, X.; Wang, Z. Deep learning for air quality forecasts: A review. Curr. Pollut. Rep. 2020, 6, 399–409. [Google Scholar] [CrossRef]
  2. Vadurin, K.; Perekrest, A.; Bakharev, V.; Shendryk, V.; Parfenenko, Y.; Shendryk, S. Towards Digitalization for Air Pollution Detection: Forecasting Information System of the Environmental Monitoring. Sustainability 2025, 17, 3760. [Google Scholar] [CrossRef]
  3. Peldon, D.; Banihashemi, S.; LeNguyen, K.; Derrible, S. Navigating urban complexity: The transformative role of digital twins in smart city development. Sustain. Cities Soc. 2024, 111, 105583. [Google Scholar] [CrossRef]
  4. Verde, S.; Dell’Acqua, F.; Losasso, M. Environmental Data, Modeling and Digital Simulation for the Evaluation of Climate Adaptation and Mitigation Strategies in the Urban Environment. Sustainability 2024, 16, 2179. [Google Scholar] [CrossRef]
  5. Yadav, D.K.; Soumya, K.; Goswami, L. Autoregressive integrated moving average model for time series analysis. In Proceedings of the 2024 International Conference on Optimization Computing and Wireless Communication (ICOCWC), Debre Tabor, Ethiopia, 29–30 January 2024; IEEE: New York, NY, USA, 2024; pp. 1–6. [Google Scholar]
  6. Lai, Y.; Dzombak, D.A. Use of the autoregressive integrated moving average (ARIMA) model to forecast near-term regional temperature and precipitation. Weather Forecast. 2020, 35, 959–976. [Google Scholar] [CrossRef]
  7. Hsieh, W.W. Evolution of machine learning in environmental science—A perspective. Environ. Data Sci. 2022, 1, e3. [Google Scholar] [CrossRef]
  8. Gyaneshwar, A.; Mishra, A.; Chadha, U.; Raj Vincent, P.M.D.; Rajinikanth, V.; Pattukandan Ganapathy, G.; Srinivasan, K. A Contemporary Review on Deep Learning Models for Drought Prediction. Sustainability 2023, 15, 6160. [Google Scholar] [CrossRef]
  9. Oyarzabal, R.S.; Santos, L.B.; Cunningham, C.; Broedel, E.; Lima, G.R.; Cunha-Zeri, G.; Peixoto, J.S.; Anochi, J.A.; Garcia, K.; Costa, L.C.; et al. Forecasting drought using machine learning: A systematic literature review. Nat. Hazards 2025, 121, 9823–9851. [Google Scholar] [CrossRef]
  10. Nandgude, N.; Singh, T.P.; Nandgude, S.; Tiwari, M. Drought Prediction: A Comprehensive Review of Different Drought Prediction Models and Adopted Technologies. Sustainability 2023, 15, 11684. [Google Scholar] [CrossRef]
  11. Rivera-Ruiz, M.A.; Mendez-Vazquez, A.; López-Romero, J.M. Time series forecasting with quantum machine learning architectures. In Proceedings of the Mexican International Conference on Artificial Intelligence, Monterrey, Mexico, 24–29 October 2022; Springer Nature: Cham, Switzerland, 2022; pp. 66–82. [Google Scholar]
  12. Ahmed, A.S.A.; Kurnaz, S. Quantum computing and artificial neural network for classification and time series prediction. In Proceedings of the 2022 International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA), Ankara, Turkey, 9–11 June 2022; IEEE: New York, NY, USA, 2022; pp. 1–4. [Google Scholar]
  13. Morstyn, T.; Wang, X. Opportunities for quantum computing within net-zero power system optimization. Joule 2024, 8, 1619–1640. [Google Scholar] [CrossRef]
  14. Choudhary, P.K.; Innan, N.; Shafique, M.; Singh, R. HQNN-FSP: A Hybrid Classical-Quantum Neural Network for Regression-Based Financial Stock Market Prediction. arXiv 2025, arXiv:2503.15403. [Google Scholar]
  15. Zheng, K.; Van Griensven, J.; Fraser, R. A Quantum Machine Learning Approach to Spatiotemporal Emission Modelling. Atmosphere 2023, 14, 944. [Google Scholar] [CrossRef]
  16. Setiadi, D.R.I.M.; Susanto, A.; Nugroho, K.; Muslikh, A.R.; Ojugo, A.A.; Gan, H.-S. Rice Yield Forecasting Using Hybrid Quantum Deep Learning Model. Computers 2024, 13, 191. [Google Scholar] [CrossRef]
  17. Lipovetsky, S. Quantum-like Data Modeling in Applied Sciences: Review. Stats 2023, 6, 345–353. [Google Scholar] [CrossRef]
  18. Ranga, D.; Prajapat, S.; Akhtar, Z.; Kumar, P.; Vasilakos, A.V. Hybrid Quantum–Classical Neural Networks for Efficient MNIST Binary Image Classification. Mathematics 2024, 12, 3684. [Google Scholar] [CrossRef]
  19. Chen, S.; Yoo, S.; Fang, Y.-L. Quantum Long Short-Term Memory. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 8622–8626. [Google Scholar]
  20. Juarez-Osorio, S.L.; Rivera-Ruiz, M.A.; Mendez-Vazquez, A.; Rodriguez-Tello, E.; López Romero, J.M. Fourier Series Guided Design of Quantum Convolutional Neural Networks for Enhanced Time Series Forecasting. Phys. Scr. 2024, 100, 106009. [Google Scholar] [CrossRef]
  21. Xu, Z.; Wang, Y.; Feng, X.; Wang, Y.; Li, Y.; Lin, H. Quantum-enhanced forecasting: Leveraging quantum gramian angular field and CNNs for stock return predictions. Financ. Res. Lett. 2024, 67, 105840. [Google Scholar] [CrossRef]
  22. Takaki, Y.; Mitarai, K.; Negoro, M.; Fujii, K.; Kitagawa, M. Learning temporal data with a variational quantum recurrent neural network. Phys. Rev. A 2021, 103, 052414. [Google Scholar] [CrossRef]
  23. Nahar, A.; Mondal, K.K.; Das, D.; Buyya, R. Qiov: A quantum-driven internet-of-vehicles-based approach for environmental monitoring and rapid response systems. arXiv 2024, arXiv:2403.18622. [Google Scholar]
  24. Kea, K.; Kim, D.; Huot, C.; Kim, T.-K.; Han, Y. A Hybrid Quantum-Classical Model for Stock Price Prediction Using Quantum-Enhanced Long Short-Term Memory. Entropy 2024, 26, 954. [Google Scholar] [CrossRef]
  25. Qin, H. Comparison of deep learning models on time series forecasting: A case study of dissolved oxygen prediction. arXiv 2019, arXiv:1911.08414. [Google Scholar] [CrossRef]
  26. Duan, Y.; Bo, Y.; Yao, X.; Chen, G.; Liu, K.; Wang, S.; Yang, B.; Li, X. A Deep Learning Framework for Long-Term Soil Moisture-Based Drought Assessment Across the Major Basins in China. Remote Sens. 2025, 17, 1000. [Google Scholar] [CrossRef]
  27. Rahman, M.M.; Joha, M.I.; Nazim, M.S.; Jang, Y.M. Enhancing IoT-Based Environmental Monitoring and Power Forecasting: A Comparative Analysis of AI Models for Real-Time Applications. Appl. Sci. 2024, 14, 11970. [Google Scholar] [CrossRef]
  28. Dong, Z.; Zhou, Y. A Novel Hybrid Model for Financial Forecasting Based on CEEMDAN-SE and ARIMA-CNN-LSTM. Mathematics 2024, 12, 2434. [Google Scholar] [CrossRef]
  29. Kerski, J. Teaching and Learning Geography with a Web GIS Approach. In Geographic Information Systems (GIS) for Disaster Management; Kumar, S., Champati Ray, P.K., Singh, R.B., Pandey, A.C., Eds.; Springer Nature: Cham, Switzerland, 2024; pp. 113–135. [Google Scholar]
  30. Toms, M. Remote Sensing and GIS for Ecologists: Using Open Source Software. Bird Study 2016, 64, 116. [Google Scholar] [CrossRef]
  31. Brimcombe, A. GIS, Environmental Modeling and Engineering, 2nd ed.; CRC Press: Boca Raton, FL, USA, 2020. [Google Scholar]
  32. Goulart, A.G.; Lazo, M.J.; Suarez, J.M.S. A new parameterization for the concentration flux using the fractional calculus this model the dispersion of contaminants in the Planetary Boundary Layer. Phys. A Stat. Mech. Its Appl. 2019, 518, 38–49. [Google Scholar] [CrossRef]
  33. Stein, A.F.; Draxler, R.R.; Rolph, G.D.; Stunder, B.J.B.; Cohen, M.D.; Ngan, F. NOAA’s HYSPLIT Atmospheric Transportation and Dispersion Modeling System. Bull. Am. Meteorol. Soc. 2015, 96, 2059–2077. [Google Scholar] [CrossRef]
  34. Brasseur, G.P.; Jacob, D.J. Atmospheric Observations and Model Evaluation. In Modeling of Atmospheric Chemistry; Cambridge University Press: Cambridge, UK, 2017; pp. 436–486. [Google Scholar] [CrossRef]
  35. Avila, A.; Santos, H.; Cruz, A.; Xavier-de-Souza, S.; Lucca, G.; Moura, B.; Yamin, A.; Reiser, R. HybriD-GM: A Framework for Quantum Computing Simulation Targeted to Hybrid Parallel Architectures. Entropy 2023, 25, 503. [Google Scholar] [CrossRef]
  36. Wang, G.; Xuan, J.; Wang, P.; Li, C.; Lu, J. LSTM autoencoder-based deep neural networks for barley genotype-to-phenotype prediction. In Australasian Joint Conference on Artificial Intelligence; Springer Nature: Singapore, 2024; pp. 342–353. [Google Scholar]
  37. Han, J.; Pei, J.; Tong, H. Data Mining: Concepts and Techniques; Morgan Kaufmann: Burlington, MA, USA, 2022. [Google Scholar]
  38. Zavaliiev, A.; Vadurin, K.; Perekrest, A.; Bakharev, V. Information and analytical system for collecting, processing and analyzing data on air pollution. Autom. Technol. Bus. Process. 2024, 16, 72–82. [Google Scholar] [CrossRef]
  39. Vadurin, K.; Perekrest, A.; Bakharev, V.; Deriyenko, A.; Ivashchenko, A.; Shkarupa, S. An information system for collecting and storing air quality data from municipal level Vaisala stations. Infocommun. Comput. Technol. 2023, 2, 38–49. [Google Scholar] [CrossRef]
  40. Vadurin, K.; Perekrest, A.; Bakharev, V. Development of a method of automatic reporting on the number of exceedances of the established standards of atmospheric air markers. Infocommun. Comput. Technol. 2023, 2, 50–59. [Google Scholar] [CrossRef]
  41. Perekrest, A.; Mamchur, D.; Zavaleev, A.; Vadurin, K.; Malolitko, V.; Bakharev, V. Web-based technology of intellectual analysis of environmental data of an industrial enterprise. In Proceedings of the 2023 IEEE 5th International Conference on Modern Electrical and Energy System (MEES), Kremenchuk, Ukraine, 27–30 September 2023; IEEE: New York, NY, USA, 2023; pp. 1–7. [Google Scholar] [CrossRef]
  42. Esparza-Gómez, J.M.; Luque-Vega, L.F.; Guerrero-Osuna, H.A.; Carrasco-Navarro, R.; García-Vázquez, F.; Mata-Romero, M.E.; Olvera-Olvera, C.A.; Carlos-Mancilla, M.A.; Solís-Sánchez, L.O. Long Short-Term Memory Recurrent Neural Network and Extreme Gradient Boosting Algorithms Applied in a Greenhouse’s Internal Temperature Prediction. Appl. Sci. 2023, 13, 12341. [Google Scholar] [CrossRef]
  43. Li, Y.-C.; Zhang, Y.-F.; Xu, R.-Q.; Zhou, R.-G.; Dong, Y.-L. HQRNN-FD: A Hybrid Quantum Recurrent Neural Network for Fraud Detection. Entropy 2025, 27, 906. [Google Scholar] [CrossRef]
  44. Wang, J.; Qiang, X.; Ren, Z.; Wang, H.; Wang, Y.; Wang, S. Time-Series Well Performance Prediction Based on Convolutional and Long Short-Term Memory Neural Network Model. Energies 2023, 16, 499. [Google Scholar] [CrossRef]
  45. El Zaar, A.; Mansouri, A.; Benaya, N.; Bakir, T.; El Allati, A. Hybrid Transformer-CNN architecture for multivariate time series forecasting: Integrating attention mechanisms with convolutional feature extraction. J. Intell. Inf. Syst. 2025, 63, 1233–1264. [Google Scholar] [CrossRef]
  46. Hall, T.; Rasheed, K. A Survey of Machine Learning Methods for Time Series Prediction. Appl. Sci. 2025, 15, 5957. [Google Scholar] [CrossRef]
  47. Ray, M.; Sahoo, K.C.; Abotaleb, M.; Ray, S.; Sahu, P.K.; Mishra, P.; Ghazi Al Khatib, A.; Sankar Das, S.; Jain, V.; Balloo, R. Modeling and forecasting metrological factors using BATS and TBATS models for Keonjhar district of Orissa. Mausam 2022, 73, 555–564. [Google Scholar] [CrossRef]
  48. Puindi, A.C.; Silva, M.E. Dynamic structural models with covariates for short-term forecasting of time series with complex seasonal patterns. J. Appl. Stat. 2021, 48, 804–826. [Google Scholar] [CrossRef] [PubMed]
  49. Huang, N.; Lin, T.; Guan, J.; Zhang, G.; Qin, X.; Liao, J.; Liu, Q.; Huang, Y. Identification and Regulation of Critical Source Areas of Non-Point Source Pollution in Medium and Small Watersheds Based on Source-Sink Theory. Land 2021, 10, 668. [Google Scholar] [CrossRef]
  50. de Zarzà, I.; de Curtò, J.; Roig, G.; Manzoni, P.; Calafate, C.T. Emergent Cooperation and Strategy Adaptation in Multi-Agent Systems: An Extended Coevolutionary Theory with LLMs. Electronics 2023, 12, 2722. [Google Scholar] [CrossRef]
  51. Liu, P.; Wang, L.; Li, J. Unlocking the Potential of Explainable Artificial Intelligence in Remote Sensing Big Data. Remote Sens. 2023, 15, 5448. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.