Next Article in Journal
Multiscale Fractal Evolution Mechanism of Pore Heterogeneity in Hydrocarbon Source Rocks: A Thermal Simulation Experiment in the Xiamaling Formation
Next Article in Special Issue
Multifractal Carbon Market Price Forecasting with Memory-Guided Adversarial Network
Previous Article in Journal
Quantum Creation of a Friedmann-Robertson-Walker Universe: Riesz Fractional Derivative Applied
Previous Article in Special Issue
Graph-Based Stock Volatility Forecasting with Effective Transfer Entropy and Hurst-Based Regime Adaptation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

UK Carbon Price Dynamics: Long-Memory Effects and AI-Based Forecasting

by
Zeno Dinca
1,
Camelia Oprean-Stan
2,* and
Daniel Balsalobre-Lorente
3,4,5
1
Department of International Business and Economics, Bucharest University of Economic Studies, 010374 Bucharest, Romania
2
Faculty of Economic Sciences, Lucian Blaga University of Sibiu, 550324 Sibiu, Romania
3
Department of Applied Economics I, University of Castilla La Mancha, 16002 Cuenca, Spain
4
UNEC Research Methods Application Center, Azerbaijan State University of Economics (UNEC), Istiqlaliyyat Str. 6, Baku 1001, Azerbaijan
5
Economic Research Center (WCERC), Western Caspian University, Baku 1001, Azerbaijan
*
Author to whom correspondence should be addressed.
Fractal Fract. 2025, 9(6), 350; https://doi.org/10.3390/fractalfract9060350
Submission received: 12 April 2025 / Revised: 24 May 2025 / Accepted: 25 May 2025 / Published: 27 May 2025

Abstract

This study examines the price dynamics of the UK Emission Trading Scheme (UK ETS) by integrating advanced computational methods, including deep learning and statistical modelling, to analyze and simulate carbon market behaviour. By analyzing long-memory effects and price volatility, it assesses whether UK carbon prices align with theoretical expectations from carbon pricing mechanisms and market efficiency theories. Findings indicate that UK carbon prices exhibit persistent long-memory effects, contradicting the Efficient Market Hypothesis, which assumes price movements are random and fully reflect available information. Furthermore, regulatory interventions exert significant downward pressure on prices, suggesting that policy uncertainty disrupts price equilibrium in cap-and-trade markets. Deep learning models, such as Time-series Generative Adversarial Networks (TGANs) and adjusted fractional Brownian motion, outperform traditional approaches in capturing price dependencies but are prone to overfitting, highlighting trade-offs in AI-based forecasting for carbon markets. These results underscore the need for predictable regulatory frameworks, hybrid pricing mechanisms, and data-driven approaches to enhance market efficiency. By integrating empirical findings with economic theory, this study contributes to the carbon finance literature and provides insights for policymakers on improving the stability and effectiveness of emissions trading systems.

1. Introduction

As the evidence of climate change continues to mount, it is clear that this challenge is multifaceted and cannot be easily addressed through traditional policy approaches. To reduce climate-related risks, many countries have implemented national and regional emissions trading systems. These mechanisms aim to make polluters internalize the external costs of their emissions [1]. The goal is to incentivize emissions reductions by making it economically advantageous for companies to invest in cleaner technologies and practices [2]. The European Union Emissions Trading System (EU ETS), established in 2005, has served as a model for similar initiatives worldwide. Following Brexit, the UK launched its own UK Emissions Trading Scheme (UK ETS) in January 2021, providing an independent framework to regulate carbon emissions through market-based mechanisms. The UK ETS operates similarly to the EU ETS as a cap-and-trade system but exhibits distinct market dynamics due to regulatory divergence, supply mechanisms, and domestic economic factors [3]. UK carbon permits, officially known as UK allowances (UKAs), are tradable certificates within the UK ETS. Each UKA represents the right to emit one tonne of carbon dioxide (CO2) or equivalent greenhouse gases. Unlike other commodity markets, carbon permits are influenced not only by supply and demand fundamentals but also by regulatory cycles, volatility clustering [4], economic policy uncertainty [5,6], and climate policy uncertainty [7]. Accurate carbon price forecasting is essential for increasing market transparency, supporting decarbonization strategies, and reducing financial risk from carbon exposure.
Despite the importance of carbon price forecasting, existing models face significant limitations. Carbon markets, as a nexus of environmental and financial considerations, present data characteristics (volatility clustering, mean reversion, and structural breaks) similar to those in financial markets [8]. The self-similarity in carbon price fluctuations suggests that long-term memory effects fundamentally shape market behaviour [9]. However, previous research has largely overlooked these persistent structures, leading to suboptimal predictive accuracy and potentially hindering market efficiency [10]. Regarding the process of synthetic data generation, conventional approaches, while effective in certain aspects, may struggle to capture the full complexity and nuances of real-world carbon price time series [11]. Studies using synthetic data often rely on limited validation techniques, primarily focusing on basic statistical properties [12]. This raises concerns about the reliability and generalizability of findings based on synthetic data. This study addresses these research gaps by integrating advanced computational techniques—particularly deep learning models such as Time-series Generative Adversarial Networks (TGANs)—to analyze and replicate UK carbon price dynamics.
Several key factors drive the decision to focus this study on the UK ETS market. First, regulatory distinctions following Brexit have led to a divergence between the UK ETS and the EU ETS, resulting in separate allowance auctions, price floors, and regulatory interventions [13]. These differences create unique price behaviours that cannot be directly inferred from EU ETS trends. Second, the UK ETS operates as a relatively new market compared to the EU ETS, leading to distinct liquidity constraints and price formation mechanisms. Examining price dynamics in a smaller, less liquid market provides valuable insights applicable to other emerging carbon trading systems. Lastly, the UK economy features a unique energy mix, industrial composition, and carbon reduction commitments that set it apart from other ETS markets. The impact of macroeconomic variables, energy policy decisions, and carbon reduction targets on UK carbon pricing necessitates a focused analysis [14]. By examining the UK ETS, this study not only enhances the understanding of an increasingly important carbon market but also provides a framework applicable to other emissions trading systems globally. The methodology employed here—combining long-memory modelling with deep learning approaches—can be extended to other carbon markets to improve forecasting accuracy and policy evaluation.
This research aims to explore the structural and predictive dynamics of UK carbon markets by addressing three key objectives. First, it investigates the persistence of long-memory effects in carbon prices, challenging the Efficient Market Hypothesis (EMH) [15]. By quantifying self-similar price behaviour using the Hurst exponent and fractional Brownian motion (fBm), this study aims to enhance the understanding of non-random price dependencies often overlooked by traditional models. Second, this study evaluates the effectiveness of deep learning models, particularly TGAN, in capturing the intricate temporal dependencies and volatility inherent in carbon price data. Finally, it compares the predictive accuracy of traditional econometric models with advanced machine learning approaches to assess their suitability for carbon market forecasting. By exploring innovative methodologies, this research seeks to refine forecasting methodologies and provide a more comprehensive understanding of carbon market price dynamics.
This research makes several novel contributions to the field of carbon market analysis by integrating advanced methodologies and providing empirical insights into UK carbon price dynamics. Unlike previous studies focusing on short-term price fluctuations [8], this study emphasizes structural price properties. It explores the role of long-range dependencies in shaping future price behaviour, challenging conventional assumptions of market efficiency. This study uses Rescaled Range (R/S) analysis to estimate the Hurst exponent and applies fractional Brownian motion (fBm) to generate synthetic data. These methods provide empirical evidence of self-similarity and long-term dependence in UK carbon prices. This finding challenges the Efficient Market Hypothesis in carbon trading, as price signals may not always align with fundamental abatement costs. A key innovation is the adjusted fBm model, which uses spectral scaling from the empirical covariance matrix to better match real market dynamics. This study addresses forecasting limitations related to missing data and seasonality by applying linear interpolation and Locally Estimated Scatterplot Smoothing (LOESS)-based decomposition [16]. This improves the quality and consistency of the input data.
To enhance forecasting capabilities, the research employs TGAN-based synthetic data generation, demonstrating its effectiveness in capturing complex market dependencies and improving scenario analysis. A novel validation framework is also introduced, where a Long Short-Term Memory (LSTM) regression model is used to assess the predictive power of synthetic data beyond conventional statistical comparisons. To validate the relevance of the proposed methodology, deep learning-based predictions are benchmarked against those generated by a traditional econometric model. This model is the Autoregressive Integrated Moving Average with Exogenous Variables (ARIMAX), which serves as a secondary reference. Despite differences in the underlying modelling frameworks, this comparison provides a contextual basis for assessing the predictive performance of synthetic data-driven forecasting methods. This study evaluates the advantages of deep learning models, such as TGAN, over traditional econometric approaches. In doing so, it advances synthetic data modelling for carbon markets and offers practical insights for both policymakers and market participants. Although Artificial Intelligence (AI)-based models like TGAN and long-memory processes are not new in financial forecasting, their use for synthetic data generation and predictive validation in carbon markets, especially the UK ETS, is still limited. This study bridges that gap by integrating these methods into a framework tailored to emissions trading dynamics.
The remainder of this study is structured as follows: Section 2 reviews the literature on carbon pricing models, highlighting their strengths and limitations. Section 3 presents the theoretical analysis and research hypotheses. Section 4 details the methodology, including data sources, preprocessing techniques, and modelling frameworks. Section 5 provides the empirical results and analysis, and Section 6 concludes with the study’s findings and discusses the research implications.

2. Literature Review

The study of carbon pricing mechanisms has gained increasing attention due to their critical role in climate policy and market efficiency [7]. Research on carbon pricing employs various methods, each with distinct strengths and limitations, often combining several approaches to gain a comprehensive understanding of this complex policy instrument. Traditional time-series models, such as ARIMA (Autoregressive Integrated Moving Average), ARIMAX (Autoregressive Integrated Moving Average with Exogenous Variables), and GARCH (Generalized Autoregressive Conditional Heteroskedasticity), remain popular due to their transparency, ease of interpretation, and strong statistical underpinnings [17,18]. These models are particularly useful for identifying trends, volatility, and autoregressive patterns in carbon markets. However, their reliance on linear assumptions and stationarity constraints limits their capacity to capture the non-linear, dynamic, and evolving nature of carbon pricing. Additionally, these models often fail to accommodate regime shifts, structural breaks, and complex seasonal interactions commonly observed in carbon markets.
Computable General Equilibrium (CGE) models offer a broader macroeconomic perspective by simulating the economy-wide impacts of carbon pricing policies [19]. They are grounded in economic theory and are useful for policy evaluation and scenario analysis. However, their accuracy strongly depends on parameter calibration, especially elasticity assumptions. This leads to significant variation across studies and raises concerns about robustness and replicability.
In contrast, machine learning (ML) and deep learning (DL) models such as Long Short-Term Memory (LSTM), Temporal Convolutional Networks (TCNs), Support Vector Regression (SVR), and Spiking Neural Networks (SNNs) provide powerful tools to model non-linear, high-dimensional, and non-stationary time series [20,21,22,23]. These models can detect subtle temporal interactions and adaptive patterns that traditional models may miss. However, they suffer from a lack of interpretability, potential overfitting, high computational cost, and sensitivity to hyperparameter tuning [24]. Moreover, their performance often depends on the availability of large, high-quality datasets—an ongoing challenge in emerging carbon markets.
Experimental and quasi-experimental approaches, such as difference-in-differences, randomized controlled trials, and event studies, provide rigorous empirical evidence on the causal effects of carbon pricing policies [25,26,27]. While these methods are essential for policy evaluation, they are difficult to implement in real-world contexts due to ethical, political, and logistical constraints. Additionally, quasi-experimental designs may not fully control for confounding variables and can be sensitive to methodological decisions, such as the choice of event windows.
While the literature on carbon price modelling spans a diverse set of methodological paradigms—ranging from traditional econometrics to computable general equilibrium (CGE) frameworks and machine learning (ML)/deep learning (DL) models—it rarely offers a structured comparative synthesis that evaluates their respective strengths and limitations in an integrated manner. Such a synthesis is crucial to justify the need for hybrid approaches in complex policy domains like carbon markets. To address this gap and substantiate our methodological framework, Table 1 presents a comparative matrix of the three dominant modelling paradigms: (i) traditional econometric models, (ii) CGE-based macroeconomic simulations, and (iii) ML/DL-based data-driven architectures. The comparison spans theoretical rigour, interpretability, predictive capacity, robustness, and policy relevance—dimensions essential to the modelling of price dynamics in emerging emissions trading systems (ETS).
Table 1 clearly reveals that no single paradigm is sufficient in isolation to model the non-linearities, regulatory feedbacks, and temporal complexities inherent in carbon pricing. Therefore, this study adopts a hybrid modelling framework that integrates the theoretical transparency of traditional models with the adaptive and high-dimensional capabilities of deep learning. This integration ensures both empirical robustness and practical relevance for stakeholders in carbon markets. This integrated perspective is implemented by benchmarking the performance of transformer-based Generative Adversarial Networks (TGAN) and fractional Brownian motion (fBm) against a traditional ARIMAX baseline. The comparative framework enables a systematic evaluation of the forecasting improvements introduced by deep learning methods, while preserving the interpretability and theoretical coherence characteristic of traditional models. This hybrid approach is particularly well-suited to emerging carbon markets, where the need for predictive accuracy must be balanced with policy relevance and methodological transparency.

3. Theoretical Background and Hypotheses Development

3.1. Introduction to Carbon Pricing Mechanisms

Carbon pricing mechanisms are essential for addressing climate change by internalizing the external costs of greenhouse gas emissions. These mechanisms can be broadly classified into two theoretical models: Pigouvian taxation (carbon tax) and cap-and-trade systems. Each model has distinct characteristics, advantages, and challenges, critical for policymakers aiming to reduce carbon emissions effectively.
The Pigouvian tax, named after economist Arthur Pigou, is designed to correct the market failure caused by negative externalities, such as carbon emissions [28]. By taxing carbon emissions, the government incentivizes emitters to reduce their carbon footprint, thereby internalizing the external costs associated with pollution. This approach is grounded in the principle that polluters should pay for the damage they cause to the environment, which aligns with the “polluter pays” principle [29]. Research indicates that a well-structured carbon tax can yield a “double dividend” [30]. The first dividend is the environmental benefit from reduced emissions, while the second is the potential for economic growth by reinvesting tax revenues into green technologies and infrastructure. For instance, studies have shown that carbon taxes can stimulate innovation in low-carbon technologies, leading to long-term economic benefits [31]. However, the effectiveness of a carbon tax can be influenced by its design, including the tax rate and the scope of emissions covered [32]. Moreover, implementing a carbon tax can face public resistance, often due to concerns about its regressive nature, where lower-income households disproportionately bear the burden of increased energy costs [33]. To mitigate this, policymakers can consider revenue recycling strategies, where tax revenues are used to offset costs for vulnerable populations or to fund renewable energy projects [34]. In a carbon tax system, price stability results from the government setting a fixed tax per ton of carbon emitted. This ensures that carbon prices remain stable and predictable, allowing businesses to make informed investment decisions more confidently. However, while the tax provides economic predictability, it does not directly cap emissions, meaning that firms can continue polluting if willing to pay the tax. As a result, while carbon pricing remains stable, the actual level of emissions remains uncertain, as it depends on how firms respond to the financial incentive to reduce pollution.
The UK Emissions Trading Scheme (UK ETS), which serves as the empirical basis for this study, operates under a cap-and-trade model, where firms trade emission allowances within a government-imposed cap. In contrast to carbon taxes, cap-and-trade systems establish a market for carbon emissions by setting a limit (cap) on total emissions and allowing companies to buy and sell allowances for emissions. This market-based approach incentivizes companies to reduce their emissions below their allocated cap, as they can sell excess allowances for profit. The cap-and-trade model is rooted in the Coase theorem, which suggests that market mechanisms can lead to efficient outcomes in the presence of externalities, provided property rights are well-defined [35]. Cap-and-trade systems have been implemented in various regions with mixed results. While they can effectively reduce emissions, their success often depends on the robustness of the regulatory framework and the initial allocation of allowances [32]. For example, studies have shown that cap-and-trade can significantly reduce emissions when combined with complementary policies, such as renewable energy mandates [36]. However, challenges such as market volatility and the potential for “leakage”, where emissions are transferred to regions with less stringent regulations, can undermine the effectiveness of cap-and-trade systems [37]. Unlike a carbon tax, which sets a fixed price on emissions, a cap-and-trade system ensures that total emissions remain capped while allowing prices to fluctuate based on market conditions. In this system, market forces determine the price of emission allowances through supply and demand, creating a dynamic pricing mechanism. Ideally, in a well-functioning market, the price of these allowances should reflect the marginal cost of abatement—the cost required to reduce one additional unit of emissions—ensuring an economically efficient approach to emissions reduction.
In conclusion, both carbon taxes and cap-and-trade systems aim to reduce greenhouse gas emissions but do so through different mechanisms. Carbon taxes provide price certainty, allowing businesses to plan for future costs, while cap-and-trade offers certainty in emission reductions but can lead to price volatility in allowance markets [32]. The choice between these mechanisms often depends on political, economic, and social contexts, as well as the specific environmental goals of a region. Empirical evidence suggests that a hybrid approach, integrating elements of both carbon taxation and cap-and-trade, may yield the best outcomes [31]. For instance, combining a carbon tax with a cap-and-trade system can stabilize prices while ensuring that emission reduction targets are met [38]. Furthermore, the design of these mechanisms should consider equity implications to ensure that the burden of climate policies does not disproportionately affect low-income communities [34]. Policymakers must carefully design these mechanisms to maximize environmental benefits while minimizing economic disruptions and ensuring social equity.

3.2. Efficient Market Hypothesis (EMH) and Carbon Pricing

The effectiveness of a cap-and-trade system depends on how efficiently carbon prices reflect market fundamentals. According to the EMH, prices in a well-functioning market should fully reflect all available information at any given time [15]. Within this theoretical framework, carbon prices should adjust instantly and unbiasedly to new information, ensuring allocative efficiency in emissions trading schemes (ETS). Several canonical conditions underpin this idealized efficiency: high liquidity, broad participation, information transparency, and regulatory stability [39].
However, while EMH offers a foundational lens, its assumptions often fail to hold empirically in carbon markets. Studies have documented departures from random walk behaviour in carbon price series, indicating serial correlation, long memory, and mean-reverting structures [40,41,42]. These findings are inconsistent with the weak-form EMH and suggest the presence of persistent price dynamics potentially exploitable by informed traders. Moreover, carbon markets are relatively young, politically contingent, and structurally mediated—features that limit the applicability of classical market efficiency assumptions.
Recent literature advances alternative explanations for these inefficiencies rooted in behavioural and institutional economics. A growing body of research models carbon price dynamics using regime-switching frameworks, where markets alternate between low- and high-volatility states due to shifts in regulatory signals, policy uncertainty, or macroeconomic shocks [43]. These non-linear dynamics deviate from the EMH’s assumption of a single, stable equilibrium process.
Moreover, the financialization of carbon markets has introduced speculative pressures, with increased participation from financial actors seeking portfolio diversification or arbitrage opportunities [44]. This can decouple carbon prices from underlying abatement cost fundamentals, reducing the environmental integrity of price signals. Investor heterogeneity—in expectations, risk tolerance, and time horizons—further contributes to asymmetric responses to information, while behavioural biases such as herding, overreaction to policy news, or confirmation bias can generate price anomalies [45].
Institutional design also plays a central role. Factors such as allocation mechanisms (e.g., free versus auctioned allowances), banking and borrowing rules, and market segmentation across jurisdictions influence how information is processed and disseminated [46]. Inadequate coordination between carbon and related markets—particularly electricity—can further distort carbon price formation and weaken market efficiency [47].
Taken together, these observations suggest that carbon markets operate under hybrid informational and institutional regimes that blend partial efficiency with structural, behavioural, and regulatory frictions. These conditions underscore the need for advanced modelling strategies to accommodate regime dependence, non-linearity, and path dependence. In this context, strict adherence to the EMH may be insufficient for understanding real-world carbon price behaviour, motivating the integration of alternative theoretical perspectives into carbon market modelling.

3.3. Hypotheses Development

3.3.1. Self-Similar Behaviour of Carbon Permit Prices

From a theoretical perspective, self-similar processes are characterized by the property that their statistical characteristics are invariant to changes in temporal and spatial scales, suggesting the presence of long-range dependence and the potential for improved forecasting [27]. This hypothesis, if supported by empirical evidence, could provide valuable insights into the fundamental drivers of carbon prices, which are crucial for the effective implementation of carbon pricing schemes and the transition to a low-carbon economy. Previous studies have examined self-similar behaviour in various financial markets, including equity, foreign exchange, and commodity markets [48,49]. These studies have found that many financial time series exhibit long-range dependence, which can be quantified using the Hurst exponent, a measure of the degree of self-similarity. Applying this approach to the carbon permit market could shed light on the underlying dynamics of this emerging asset class and its potential implications for risk management and investment strategies.
The theoretical analysis and empirical evidence from previous studies suggest that this line of research holds promise for enhancing the understanding of carbon market dynamics and informing the design of effective policies and investment strategies [50]. This study seeks to expand on these foundations by examining the presence of long-term dependence on UK carbon permit prices. Leveraging methodologies from prior research, we hypothesize the following:
Hypothesis 1:
UK carbon prices exhibit self-similar behaviour independent of trend and seasonal influences, thereby challenging the assumptions of the Efficient Market Hypothesis.

3.3.2. Capturing the Complexity of Carbon Prices: A Comparative Analysis of Traditional and Advanced Deep Learning Techniques

Traditional pricing models, such as standard Brownian motion, have long been employed to analyze carbon prices. These models are rooted in stochastic processes designed to capture random walk behaviours. However, their reliance on simplistic assumptions about price distributions and underlying market dynamics often results in limited predictive power, leading to potential inaccuracies and suboptimal decision making [51]. This is particularly problematic given the multifaceted influences on carbon markets, including regulatory changes, technological advancements, and geopolitical events, which traditional models may fail to account for adequately. Generative Adversarial Networks (GANs), with their ability to model complex distributions and capture intricate patterns, offer a robust alternative. The ability of GANs to learn the underlying distribution of the data and generate realistic samples can potentially provide a more comprehensive and accurate representation of the dynamics and volatility structure of carbon prices [52].
This study investigates the potential of advanced deep learning techniques, specifically TGAN, to more effectively capture and replicate the complex characteristics of UK carbon prices. For this, we hypothesize the following:
Hypothesis 2:
Advanced deep learning techniques, such as TGAN, can effectively capture and replicate the underlying properties of UK carbon prices.

4. Methodology

4.1. Data Collection

Daily front-month continuation closing price data for the United Kingdom Emission Energy Allowances Future contracts (UKAFMc1) traded on the Intercontinental Exchange were collected for 2022 and 2023 in GBP as a proxy for the UK carbon market. These data were sourced from the Thomson Reuters Eikon platform.
This study focuses on the 2022–2023 period, as it offers a particularly informative context for analyzing carbon price dynamics, due to multiple relevant policy and market factors. Since the launch of the UK ETS in January 2021, the initial months were marked by price discovery and transitional adjustments. However, by 2022, the market had gained more stability, making it a reliable timeframe for assessing long-term price trends. Additionally, this period encompasses significant policy interventions (presented in Section 5.1), which influenced price expectations and volatility. These policy changes provide a natural setting for examining the market’s response to regulatory adjustments.
Furthermore, the UK carbon market during this time faced substantial macroeconomic and energy market shocks, particularly the European energy crisis following Russia’s invasion of Ukraine, which had direct implications for carbon price volatility. Studying this period allows for exploring how external disruptions affect market dynamics. Moreover, given that the UK ETS is still relatively new, high-frequency trading data are more consistent for recent years, whereas older data may have gaps or distortions due to the transition from the EU ETS. Focusing on 2022–2023, this study ensures the use of high-quality, uninterrupted datasets while capturing a phase of regulatory evolution, economic shocks, and market stabilization, providing a solid foundation for meaningful analysis.
To address missing data points caused by irregular reporting intervals, exchange holidays, or trading cessation days before contract expiration, linear interpolation was used to estimate values for gaps in the time series. This method assumes a linear relationship between adjacent data points, which is a reasonable assumption given that carbon trading activity is suspended during non-trading days, meaning that price movements resume from the last observed trend once trading restarts. Additionally, we examined the nature of missing data by analyzing the distribution and frequency of trading gaps. The dataset reveals that most missing values correspond to weekends and official UK market holidays, ensuring that interpolation does not introduce artificial trends or distort the underlying price structure. The validity of this approach was further assessed by comparing interpolated values with post-holiday price movements, confirming that no abrupt deviations or structural inconsistencies were introduced.

4.2. Time-Series Decomposition and Long-Term Memory Analysis

To analyze the statistical characteristics of the original time series, several decomposition techniques were utilized to eliminate trends and seasonal patterns from the data. Before this, Simple Linear Regression (SLR), Polynomial Regression, and Fourier Transform Regression models were applied to gain deeper insights into trends and seasonal behaviours. The decomposition of the time series into sinusoidal components through the Fourier Transform Regression model facilitated identifying and quantifying underlying seasonal trends, offering a more accurate and robust framework for analyzing and forecasting price dynamics in carbon markets.
Two modelling approaches were investigated:
(i)
Sequential (SQ)–fitting a trend model (SLR or Polynomial), calculating residuals, and then fitting a Fourier Series model to the residuals to capture seasonality, a process described by Equations (1)–(4), where t stands for time, y(t) is the observed price value at time t, p is the polynomial order (p of 1 is descriptive of an SLR), k is the harmonic order and bi, cj, and dj are the regression coefficients.
y t r e n d t = a 0 + i = 1 p b i t i
y s e a s o n a l t = j = 1 k c j cos 2 π j t P + d j sin 2 π j t P + ϵ
y s e a s o n a l t r e s i d u a l s t = y t   y t r e n d t
y S Q t = y t r e n d t + y s e a s o n a l t
(ii)
Mixed (MM)—utilizing a single regression model that incorporates both Fourier harmonics for seasonality and the linear time variable for trend as predictors (Equation (5)).
y M M t = a 0 + i = 1 p b i t i + j = 1 k C j cos 2 π j t P + D j sin 2 π j t P + ϵ
Locally Estimated Scatterplot Smoothing (LOESS) [16] was applied to decompose the time series of UKAFMc1 prices into trend, seasonal, and residual components. Seasonal-Trend decomposition using LOESS is particularly effective for analyzing time series data with complex seasonal patterns, making it well-suited for data exhibiting intricate seasonal behaviours and evolving trends. Theoretically, the residual component should contain only random noise; however, in practice, it may retain subtle time-dependent patterns that the decomposition process may not fully capture. The data were deconstructed using assumed seasonal periods of 30 days (monthly) and 90 days (quarterly). To ensure the validity of subsequent analyses, the Augmented Dickey–Fuller (ADF) test was performed on the residuals to assess their stationarity and confirm the absence of trends or seasonal patterns.
To determine the presence of long-term memory or persistence in the time series, the Hurst parameter using Rescaled Range (R/S) analysis [53,54] was estimated. R/S analysis was applied to residuals after the decomposition of seasonal and trend components. This approach ensures that the Hurst exponent reflects intrinsic long-term dependencies rather than external patterns. Decomposed application aligns with modern adaptations of R/S analysis, as seen in works like [55] or [56], where detrended data were used for more precise memory assessment.

4.3. Synthetic Data Generation

After extracting the statistical parameters, three synthetic time series of equal length to the original time series were generated. The first synthetic time series was simulated using fBm, a generalized form of standard Brownian motion. The technique employed is the Davies–Harte [57] method, advantageous due to its computational efficiency and effectiveness in generating long-range dependent processes, ensuring that the statistical properties of the simulated data closely match those of the true fractional Brownian motion process.
To further refine the fBm model, a second synthetic time series was generated using a similar procedure with an additional step involving the application of the autocovariance matrix to the fractional Brownian motion noise. This process, Adjusted fBm, inspired by spectral whitening methods in signal processing [58], utilizes the eigen decomposition of the covariance matrix by whitening the synthetic time series with the eigenvectors and subsequently scaling it using the square root of the eigenvalues. This approach results in a time series that effectively captures the variance patterns present in the original data [59].
In order to generate synthetic carbon price data that preserve the temporal structure and memory effects of the original series, this study employed a Time-series Generative Adversarial Network (TGAN). At a conceptual level, TGAN is a deep learning model designed specifically for time-series data [60]. It builds on the original Generative Adversarial Network (GAN) framework [61] to capture complex non-linear dependencies and temporal dynamics that classical modelling approaches often overlook [62]. In other words, TGAN leverages deep neural networks to learn the patterns in historical carbon prices and produce new, realistic price trajectories. Prior research has shown that TGAN can outperform classical models like Geometric Brownian Motion (GBM) in reproducing the statistical properties (marginal distributions) and temporal behaviour of financial time series [63]. This makes TGAN a compelling choice for modelling carbon price dynamics, as GBM and other simple stochastic models cannot easily capture features like autocorrelation, regime shifts, or long-memory effects present in real price data.

4.3.1. TGAN Architecture and Data Preparation

The TGAN architecture comprises four interconnected layers: (i) generator, (ii) discriminator, (iii) embedding network, and (iv) recovery network. Figure 1 illustrates how real carbon price data are embedded, used to generate synthetic sequences, and validated through an adversarial learning process. The TGAN training process combines the standard unsupervised GAN training with an additional supervised learning signal, along with the autoencoder reconstruction objective. All four components (generator, discriminator, embedder, and recovery) are trained together in a multi-task fashion, although in practice, training can be staged for stability (e.g., pre-training the autoencoder). In essence, the two core components of the TGAN model, the generator and the discriminator, operate in a competitive learning process (the typical GAN setup). The generator creates artificial data sequences that mimic real historical patterns, while the discriminator attempts to distinguish between real and synthetic sequences. Through iterative training, the generator learns to produce increasingly realistic time-series data, and the process reaches an equilibrium when the discriminator can no longer differentiate between real and synthetic data. What differentiates TGAN from a standard GAN is its ability to learn complex temporal relationships by incorporating recurrent memory units and an embedding mechanism into the architecture. Moreover, TGAN introduces an autoencoder structure, consisting of an embedding network and a recovery network that works alongside the GAN’s generator and discriminator. The embedding network encodes the original time series into a latent representation (feature space), and the recovery network decodes the generator’s output back to the original data space. By generating data in this latent space, the model ensures that essential temporal features (such as seasonality or slow mean-reverting tendencies) are captured and later reconstructed in the output.
To capture the temporal dependencies and self-correlations inherent in the original time series data, we integrated Long Short-Term Memory (LSTM) layers into the architecture of GAN. LSTM is a type of Recurrent Neural Network known for its capability to remember information over long time lags [64], which is essential for time series like carbon prices that exhibit both short-term fluctuations and long-term trends.
We represent the carbon price series in a matrix format (Equation (6)) to facilitate training the TGAN. If N is the length of the time series and T is the number of time lags considered for each training sample (plus one, to include the current time step), then each row of the matrix corresponds to one sequence of length T extracted from the original series. Each such T-length sequence (each matrix row) serves as a training sample for the LSTM-based generator and discriminator. Prior to training, the time-series values are scaled using a min–max normalizer [65] to ensure stable training.
X = x t x t 1 x t 2 x t T x t + 1 x t x t 1 x t + 1 T x t + 2 x t 1 x t x t + 2 T         x N x N 1 x N 2 x N T ,         X x R 0 x 1

4.3.2. Training Strategy and Loss Functions

The training strategy and the loss components involved are briefly summarized below. In the autoencoder structure, the embedding network (E acts as an encoder) takes the real carbon price sequence data and compresses it into a latent representation (denoted H ). This network learns an informative summary of the time series, extracting the essential temporal relationships in a lower-dimensional (or feature-enriched) space. Formally, the embedder maps the original data X to latent features H (Equation (7)), where dimensions d i m ( H ) can be chosen larger than, equal to, or smaller than the original feature dimension to best capture the dynamics.
H = E X ,     H = N T d i m H
The generator (G) receives input from the latent space and produces synthetic latent sequences. The generator is trained via a joint unsupervised and supervised learning process. In the unsupervised adversarial mode, the generator transforms a random noise tensor Z (Equation (8)) into a synthetic latent sequence H ~ that should resemble latent encodings of real data (Equation (9)). The random noise input Z is not a single vector but a sequence of random vectors of dimension d i m ( Z ) , with the same time length as the training sequences, giving the generator the necessary degrees of freedom to generate sequential patterns. For training, we segment the carbon price series into sequences of length T time steps (e.g., T = 20), forming a training matrix as presented in Equation (6). The unsupervised training is identical to a standard GAN, where the generator tries to produce data that fools the discriminator using the noisy tensor Z as input. Equation (10) represents the generator’s unsupervised loss function ( L G u ).
Z = N T d i m Z ,       Z i , j , k     N 0 ,   1
H ~ = G Z ,     H ~ = N T
L G u = t = 1 N log 1 D X t ~ N
In contrast, supervised training is realized by generating data from the latent representation obtained by transforming X into H through the embedding network. Specifically, given an embedded real sequence H (where H t denotes the latent vector at time step t of the real sequence), the generator is asked to predict the next-step latent state. At each time t, using the previous latent state H t 1 as input, G produces a predicted latent vector for the current step, denoted G H t 1 . The LSTM-based generator G then learns to map either random sequences Z or embedded sequences H to synthetic outputs H ~ that follow the patterns of the real data in the latent space. The supervised loss ( L G s ) is then defined as the mean squared error (L2 norm) between the generator’s prediction and the actual embedding H t at that time step (Equation (11)).
L G s = t = 1 N 1 H t G H t 1 2 N 1
The overall training objective for the generator combines the unsupervised adversarial loss and the supervised loss described above. The objective of the generators’ gradient descent is to minimize η     L G s + L G u where the first term η controls for the influence of the supervised loss on the training process.
The recovery network (R acts as a decoder) takes the synthetic latent sequence H ~ produced by the generator and reconstructs it back to the original data space, producing a synthetic carbon price sequence X ~ (Equation (12)). This network is the mirror of the embedder and ensures that the generated latent patterns translate into realistic sequences in the original scale and units. The pair of embedding and recovery networks acts as an encoder–decoder (autoencoder), allowing the model to retain temporal structure: any long-term pattern encoded in H should be present in X ~ after reconstruction. Notably, the synthetic data X ~ is not generated directly in the original feature space of X , but via this latent space pipeline (embed → generate → recover), which forces the generator to respect the learned structure of real data. The recovery loss is computed by comparing the original vectors with those reconstructed from the generated data. The reconstruction is performed via the recovery network. The overall formulation of the recovery loss is presented in Equation (13).
X ~ = R H ~ ,     X ~ = N T
L R u = t = 1 N X t X t ~ 2 N 1
The adjustment of the embedding and recovery network relies on the supervised loss and the recovery loss. The formula is λ     L G s + L R u , where the first term λ once again controls the influence of the supervised loss. In the original paper, Yoon et al. [61] observed no overall benefits in reducing the effect of the supervised loss on the two optimizations, prompting us to neglect the existence of λ and η in our model.
The discriminator (D) develops the ability to distinguish between real and synthetic data, providing adversarial feedback to the generator to improve the quality and realism of the generated sequences (Equation (14)). It outputs a probability D(⋅) representing how likely a given sequence is real.
D X   0 ,   1   a n d   D X ~   0 ,   1
The model was optimized using the gradient descent of the Binary Cross-Entropy (BCE) loss. Equation (15) is the generalized form of the BCE loss for a single sample where y is the true label of the original data point and y ^ is the predicted probability. Equation (16) illustrates how BCE loss is applied in estimating the discriminator’s loss for an entire batch.
L B C E y ,   y ^ = y · l o g y ^ + 1 y · l o g 1 y ^
L D = 1 N t = 1 N log D X t + log 1 D X t ~
By combining these elements, TGAN effectively extends the standard GAN framework to handle chronological properties. During training, the discriminator and generator are trained simultaneously in an adversarial process, leading to an equilibrium where both models are optimized.

4.3.3. Comparative Evaluation and Model Selection

We performed an ablation study to justify the use of TGAN over simpler generative approaches. Specifically, we compared three architectures of increasing complexity: (i) a plain GAN (no embedding network, generator trained only adversarially), (ii) the full TGAN model as described above, and (iii) a variant of TGAN where the standard BCE-based discriminator is replaced by a Wasserstein GAN critic with gradient penalty [63] (a technique often used to stabilize GAN training). All three models were built with the same base architecture (the generator and discriminator each had the same number of LSTM layers and neurons) and were trained on 20-step sequences of the carbon price data for 500 epochs to ensure a fair comparison. To evaluate synthetic data quality, we adopted two evaluation measures inspired by Yoon et al. [61] and other similar studies [66,67].
First, the discriminative test (real vs. synthetic distinguishability), which measures how well an external classifier can distinguish synthetic data from real data. If the synthetic data are very realistic, the classifier should perform no better than random guessing (50% accuracy). We trained a simple one-layer LSTM classifier (with dropout regularization) to classify sequences as “real” or “synthetic”. We then computed its classification accuracy on a held-out set. The discriminative score is defined as the absolute difference |accuracy − 0.5|. An ideal outcome is an accuracy of 50% (score of 0), meaning the classifier cannot tell real from synthetic at all.
Second, the predictive test (train on synthetic, test on real) evaluates the utilitarian quality of the synthetic data—whether models trained on synthetic data can generalize to real data. We trained a predictive model on the synthetic dataset and then tested its performance on real data. Specifically, we used a one-layer LSTM regression model (with dropout) that learns to predict the next time step given the previous 20 steps. We trained this model using synthetic sequences (from each GAN variant) and then evaluated its prediction error on real carbon price sequences (the same task on real data). We report standard forecast error metrics, Mean Absolute Error (MAE), and Mean Squared Error (MSE), for each case. If the synthetic data sufficiently capture the true dynamics, the model trained on synthetic should achieve low error on real data (close to a model trained on real data). Using these metrics, we found that the TGAN-generated data were the most similar to the original data.

4.4. Synthetic Data Validation

To assess the statistical properties of the generated synthetic time series in comparison to the original series, a comprehensive analysis was conducted involving three key steps. (1) The autocorrelation function was calculated for each synthetic time series across various lag values. Key metrics, such as the area under the curve (AUC) and the mean autocorrelation, were extracted to provide insights into the temporal dependencies within each synthetic series. Aggregate statistics were obtained by computing the coefficients for multiple pairs between the synthetic data points and their lagged counterparts. (2) The similarity between the original and synthetic time series was evaluated by computing the cross-correlation between the detrended and deseasonalized original data and the corresponding synthetic data. Positive and negative lag values were considered to capture lead and lag relationships. The AUC and mean cross-correlation values were extracted to quantify the degree of similarity. (3) To further assess the predictive capabilities of the synthetic time series, a two-layer LSTM neural network was trained using an expanding window time-series-split cross-validation approach [68] to predict the residuals of the original time series using synthetic data. The model’s prediction error was evaluated to determine the extent to which the synthetic time series could replicate the dynamic behaviour of the original data.

5. Empirical Results and Discussion

5.1. Descriptive Statistics

This subsection provides a descriptive overview of UK carbon allowance price dynamics between 2022 and 2023. It identifies key patterns of price volatility, seasonal shifts, and the role of regulatory interventions. Particular attention is given to the sharp price decline in 2023, which is linked to verified emissions reductions, increased renewable energy deployment, and announced policy changes affecting allowance supply.
After interpolating the absent time data, the resulting time series exhibits non-monotonic behaviour, characterized by alternating periods of upward and downward trends, as depicted in Figure 2. The figure suggests that the UK ETS market experienced significant price volatility in 2022, reaching peak levels before entering a prolonged downward trend throughout 2023, consistent with market uncertainties and fluctuations in supply-demand dynamics. The observed decline in UK ETS allowance prices during 2023 can be attributed to several key factors. One major factor influencing this decline was verified emissions under the UK ETS, which fell by 12.5% in early 2023, largely due to an accelerated shift from fossil fuels to renewable energy sources in electricity generation [69]. This decline in emissions reduced the demand for carbon allowances, aligning with the start of a sustained downward trend, reinforcing the connection between reduced emissions and lower demand for allowances.
Additionally, in the first and second quarters of 2023, the UK experienced a decline in its carbon-intensive energy consumption and an increase in its renewable energy generation. For example, in the first quarter of 2023, gas demand was 6.1% lower than gas demand in the same quarter of the previous year, whereas renewables’ share of total electricity generation reached a record high of 47.8% [70]. These combined factors reduced the need for carbon-intensive energy production companies to purchase ETS allowances, lowering prices during March and April 2023. Moreover, the UK government’s announcement in July 2023 regarding the release of an additional 53.5 million carbon allowances between 2024 and 2027. These allowances, originally unused from 2021 to 2023, created a market perception of oversupply, accelerating the price decline. Thus, the observed price behaviour in Figure 2 aligns with market fundamentals, where both regulatory actions (such as allowance releases) and structural changes (such as emissions reductions) played a significant role in shaping UK ETS price trends.
The descriptive statistics of the time series in Table 2 indicate that the UK ETS market exhibited significant price fluctuations between 2022 and 2023, as reflected in a wide price range and a high standard deviation, suggesting notable variability in allowance prices. The market also showed a marginal negative average price change daily, as indicated by the mean of the first difference, with daily changes displaying noticeable dispersion. Over 30 days, the average percentage change was slightly negative, suggesting a minimal price decrease. Similarly, the 90-day average percentage change indicates a stronger downward trend over a longer horizon. While the overall price level fluctuated, the 30-day and 90-day rolling means remained slightly higher than the average, suggesting a slight upward trend in average prices. However, the negative average percentage change suggests that despite the slight increase in rolling means, the overall trend in carbon prices remained downward.

5.2. Time Series Decomposition

This subsection investigates the underlying trend and seasonal structures of the carbon price series using polynomial regression, Fourier-based models, and LOESS decomposition. The goal is to isolate deterministic components from the time series and assess whether residual structures exhibit long-range dependencies. Emphasis is placed on comparing monthly (30-day) and quarterly (90-day) seasonality and interpreting the implications of the residual dynamics.
In order to gain deeper insights into trends and seasonal behaviours of the data, Simple Linear Regression (SLR), Polynomial Regression, and Fourier Transform Regression models are applied. Two modelling approaches were investigated: (i) Sequential (SQ) and (ii) Mixed (MM). The obvious difference lies in the more rigid structure of SQ that effectively isolates the trend component from the seasonal component. This distinction in the modelling approach is reflected in the regression statistics presented in Table 3 and Table 4.
The polynomial regression models in Table 3 demonstrate increasing R2 values from Simple Linear Regression up to the fifth degree, suggesting improved fit with added complexity. However, the sixth-degree model exhibits severe overfitting, evidenced by a negative R2 and drastically higher AIC/BIC values, indicating instability. While the fifth-degree polynomial achieves the highest R2, the fourth-degree model is preferable due to its lower AIC and BIC, presenting a better balance between fit and parsimony.
In Table 4, the low Fourier R2 values for the SQ models suggest a limited ability to capture seasonal patterns. MM models generally perform well in trend modelling, with MM3 achieving the highest R2 (0.8706), comparable to the best linear models. While Simple Linear Regression initially exhibits better AIC/BIC values than polynomial models, this trend reverses when combined with seasonal regression. The fifth-degree polynomial model produces the best coefficient estimates, suggesting that it provides the most accurate representation of the underlying trend and seasonal patterns in the data. Notably, SQ1 and MM1 include insignificant coefficients in their sinusoidal terms, suggesting that 90-day periods explain the periodic pattern much better than the 30-day periods. MM3 is the only mixed model approaching the goodness of fit of the fifth degree, although it performs better in terms of AIC, BIC, and score when compared to the equivalent 3rd-degree polynomial model, having a lower value in all of them. Finally, Shapiro–Wilk tests reveal non-normality in the residuals of all models, implying that even after accounting for trend and seasonality, the time series may still possess non-Gaussian properties indicative of long-term dependence. These results motivate further investigation into the underlying patterns, which can yield a deeper understanding of the price properties of carbon allowances if the trend and seasonality are removed.
LOESS decomposition was employed to separate the time series into three components: trend, seasonality, and residuals. This analysis helps in understanding both the systematic trends and the cyclical behaviours present in the data, supporting more informed forecasting and decision making. The decomposition results are presented in Figure 3, comparing 30-day (monthly) and 90-day (quarterly) seasonal effects. The trend shows a significant long-term decline, especially from early 2023 onward. Seasonal effects differ between monthly and quarterly models, with the monthly model capturing more frequent fluctuations and the quarterly model highlighting broader cycles. Residuals suggest that while most patterns are explained, some unpredictable behaviours remain, potentially due to market shocks or policy changes.
To facilitate the discussion, the residuals from the monthly and quarterly decompositions are referred to as M-SEAS (monthly seasonal decompositions) and Q-SEAS (quarterly seasonal decompositions), respectively. Table 5 demonstrates that the properties of the residuals vary depending on the selected seasonal configuration. A Shapiro–Wilk test at a significance level of 0.05 indicates that the residuals from the monthly decomposition are approximately normally distributed, exhibiting behaviour similar to random Gaussian noise. In contrast, the residuals from the quarterly decomposition are non-normally distributed, suggesting that the time series may possess characteristics beyond those captured by seasonality and trend. Both M-SEAS and Q-SEAS are confirmed to be stationary based on the results of the Augmented Dickey–Fuller (ADF) test, confirming their suitability for time series modelling. The Hurst parameter for both series was estimated using Rescaled Range (R/S) analysis. As expected, M-SEAS, which fails to differ substantially from random noise, has a Hurst parameter close to 0.5, indicating near-random behaviour with minimal persistence. In contrast, Q-SEAS displays clear long-memory behaviour, suggesting that past trends are likely to persist or recur over time.

5.3. Configuring the Synthetic Data Generation Through the Deep Learning Model

This subsection focuses on configuring the TGAN model for optimal synthetic data generation. An ablation study is conducted to compare three generative architectures, with performance evaluated across both monthly and quarterly residuals. Bayesian optimization is applied to tune hyperparameters, and model complexity is adjusted to reflect the statistical properties of the input series—particularly the persistence of long-memory structures in the quarterly decomposition. WTGAN refers to a Wasserstein-based adaptation of TGAN, designed to improve convergence stability by minimizing the Wasserstein distance between real and synthetic time-series distributions.
For determining the best GAN architecture through the ablative analysis, we have used both the Q-SEAS and the M-SEAS datasets, showcasing the study’s results in Table 6. As mentioned, the TGAN displays the greatest ability to replicate the original data.
Hyperparameter tuning is a required step in any quantitative analysis that employs complex neural networks. From among the traditional methods, we had the option of adopting either grid search or random search, both having the advantage of being fairly simple to implement. The lack of heuristic reasoning is cited as the main limitation of this method [71,72], and the method appears to underperform in comparison to Bayesian Search, Genetic Algorithms, or other task-specialized optimization methods [71,72,73,74]. Grid Search, while considered the most widely used, was noted on the other hand to be a poor choice for high-dimensional hyperparameters due to the high computational cost, suitable instead for 1D or 2D hyperparameters [74]. Grid Search was found to underperform, for example, Bayesian Search in comparison tests [75]. To simplify the optimization process, we have decided to fix the learning rates for the embedding and recovery networks at 10−4 and tune only the learning rates for the generator and discriminator and the size of batches used for training the TGAN on the M-SEAS and Q-SEAS time series using a simple Bayesian Search. The number of lags was heuristically decided using the Autocorrelation Function (ACF) and the Partial Autocorrelation Function (PACF) as guidance. For the initiation of the Gaussian Process, we used we generated 10 hyperparameters by randomly sampling from a uniform distribution values bounded between 10−2 and 10−5 for the learning rate and 10 to 100 for the batch size. We updated the Gaussian Process another 20 times until we obtained a matrix of 30 different hyperparameter combinations. The objective function used for this search algorithm was the distance between the Hurst Parameter of the real dataset and that of the generated data. After fitting the Gaussian Process on the updated hyperparameter matrix, we select the hyperparameter configuration with the highest mean performance (exploitation). For simplicity, we round the learning rates afterwards to the closest power of 10 and the batch size to the closest multiple of 10.

5.4. Synthetic Data Generation, Validation, and Comparison

This subsection presents a comparative evaluation of three synthetic data generation methods—fBm, adjusted fBm, and TGAN—based on their ability to replicate key statistical and predictive properties of the original series. The analysis includes autocorrelation and cross-correlation diagnostics, as well as LSTM-based out-of-sample testing. Findings highlight that while all synthetic models outperform standard Brownian motion, adjusted fBm and TGAN offer distinct advantages depending on the validation metric used.
Three types of synthetic time series, each matching the length of the original time series, were generated. The fBm dataset, produced solely through fractional Brownian motion, models long-term memory and self-similarity. The Adjusted fBm dataset, created by applying spectral decomposition to the fBm, captures the specific variance patterns of the original time series. The TGAN dataset, generated using a TGAN neural network, is a deep learning model designed to replicate the statistical properties of the original time series. The Area Under the Curve (AUC) and mean correlation values offer insights into the overall correlation strength between the synthetic and original time series.
Subsequently, for each of the three data generation methods, three sets of synthesized datasets were produced and compared to the original time series. For each iteration, cross-correlation and autocorrelation analyses were conducted to evaluate self-similarity and similarity to the historical data. Various hyperparameter configurations were tested, and the properties of the simulated synthetic data were analyzed. The configuration in Table 7 was identified as the most effective for generating synthetic data that closely mirrors the original series’ properties. The models successfully produced stationary data, as indicated by the ADF test, and captured long-term dependencies, as shown by the Hurst parameter. While the M-SEAS model effectively models short-term dependencies with behaviour similar to Gaussian noise, the Q-SEAS model captures strong long-term persistence, a requirement for modelling time series with memory effects. The higher training iterations in the Q-SEAS model are justified by the desire to capture the more persistent patterns suspected to exist in the Q-SEAS data compared to the M-SEAS data. Overall, the optimized configurations enable the generation of realistic synthetic data, supporting robust modelling and simulation of the original time series.
As discussed earlier, M-SEAS is expected to contain minimal additional information beyond what was captured in the trend and seasonal components. Table 8 presents the properties of the synthetic data generated for M-SEAS. The mean autocorrelation is highest for fBm and lowest for TGAN, indicating that the fBm model better preserves autoregressive patterns than TGAN. The AUC values support this trend, where fBm consistently outperforms the other two models. While all three models struggle to fully capture the characteristics of the original time series, as evidenced by the mean cross-correlation, the Adjusted fBm appears to perform slightly better based on its AUC value. Based on these findings, we conclude that for M-SEAS, the pure fBm model is most effective in preserving autocorrelation characteristics.
Figure 4 illustrates the learning process of the TGAN network. As observed, the discriminator and generator gradients oscillate inversely. As the networks continue to update, these oscillations diminish, indicating that the model converges toward an optimal state during later epochs. The generator loss in both cases is also lower than it started, which is another good sign that the model has improved its replicating capabilities. It still must be noted that the neural networks appear to struggle to achieve a balance and understand the properties of the data, a common challenge in training GANs, hinting that the model architecture can be further improved to achieve higher stability.
For the Q-SEAS series, the TGAN model again exhibited the lowest autocorrelation and cross-correlation values (as indicated in Table 9), showing a poor ability to capture the long-term properties of the original residuals. While fBm achieved stronger autocorrelation, its cross-correlation with the original data were inconsistent. In contrast, the Adjusted fBm demonstrated moderate autocorrelation and the strongest, most consistent cross-correlation with the original data. Overall, the Adjusted fBm model demonstrated the most balanced performance, effectively capturing short-term and long-term dependencies in the original time series.
Table 10 and Table 11 present the results of the LSTM regression network trained to evaluate the predictive capabilities of the generated time series for M-SEAS and Q-SEAS. A fourth synthetic time series, generated using a standard Brownian Motion (BM) process, was introduced as a baseline. The dataset was partitioned into 50% for training, 10% for validation, and 40% for testing, with each segment maintaining the original order of the data points. A 5-fold time-split cross-validation approach was applied for model training. The model architecture consists of two LSTM layers: the first with 50 neurons and the second with 30 neurons, both employing the Rectified Linear Unit (ReLU) transformation, followed by a final output perceptron. Dropout and regularization were implemented between layers, randomly omitting 20% of data points to mitigate overfitting.
For M-SEAS, the TGAN produced the second-highest loss and Mean Absolute Error (MAE) averaged across training folds, indicating inferior performance relative to fBm. In contrast, the fBm model demonstrated the best training performance. However, during testing, the fBm model exhibited signs of overfitting, implying that TGAN may generalize more effectively. Overall, the TGAN outperformed both the standard fBm and Adjusted fBm models in testing. In the case of M-SEAS, all synthetic time series outperformed BM in predicting the original series, indicating the presence of predictive power in the generated data.
The regression results for Q-SEAS (presented in Table 11) provide insights into the learning process and the predictive capabilities of the synthetic data. TGAN exhibited the lowest average MAE and loss across all but the fourth fold. On average, fBm achieved the second lowest MAE during training, while the Adjusted fBm showed mixed results during training. During the testing phase, TGAN had the lowest MAE, suggesting a superior ability to predict out-of-sample data. Conversely, Adjusted fBm exhibited the worst MAE, indicating model instability. Overall, in the case of Q-SEAS, all three synthetic data series outperformed the BM, indicating that the original time series exhibits latent structures and long-memory behaviour.
Given that the proposed validation framework involves training an LSTM model on synthetic time series and evaluating it on the original data, it cannot be directly compared to the ARIMAX model due to fundamental methodological differences in their structure and data inputs. Nonetheless, ARIMAX is included as an auxiliary reference to contextualize model performance. To complement this, a multiple linear regression (MLR) model was also applied to the same datasets, with results reported in Table 10 and Table 11. Notably, ARIMAX yields the lowest mean absolute error in both Table 10 (MAE = 8.509) and Table 11 (MAE = 8.567), with statistically significant results (p = 0.000), underscoring its high predictive accuracy. However, these outcomes are not intended as direct benchmarks for the LSTM-based architecture due to the fundamental differences in data handling and model structure. While ARIMAX is trained directly on the original series, the LSTM model operates on sequences derived from synthetic data. Therefore, ARIMAX serves strictly as a contextual anchor, illustrating the potential performance of classical models and emphasizing the need for benchmarking approaches that are methodologically aligned with the validation strategy employed.

5.5. Model Robustness and Validation

This subsection evaluates the robustness and generalization capacity of the proposed deep learning models, with a particular focus on mitigating overfitting risks. Given the complexity and flexibility of architectures such as TGAN and adjusted fBm, we implement a multi-layered validation strategy—including expanding window cross-validation, Bayesian hyperparameter optimization, and dropout regularization—to simulate realistic forecasting conditions and assess model stability across different temporal regimes.
Given the complexity and flexibility of hybrid deep learning architectures, a key methodological concern is ensuring that observed forecasting performance is not the result of model overfitting or data leakage. The risks associated with overfitting are particularly relevant in real-world applications, where models may learn patterns specific to a historical dataset but fail under new or irregular market conditions. In the context of carbon markets, this can lead to inaccurate risk assessments or trading strategies that no longer hold in practice. To address these risks, it is essential to implement rigorous validation techniques. This section outlines the validation strategies adopted to assess the robustness and generalization capacity of the proposed modelling framework, particularly the Time-series Generative Adversarial Network (TGAN) and adjusted fractional Brownian motion (fBm) models.
A multi-layered validation protocol was employed to simulate real-world forecasting conditions and evaluate model stability across different data regimes. Specifically, a time-series cross-validation procedure based on an expanding window approach was used to generate successive training and test sets without compromising temporal causality. This method is particularly well-suited for financial and energy time series, where standard K-fold cross-validation may introduce look-ahead bias.
Moreover, Bayesian optimization was applied to fine-tune hyperparameters for the TGAN’s ability to recreate data that preserve the LTD characteristics of the training data. The optimization criterion was based on the difference between the estimated Hurst parameter of the real residual data and the generated data. To avoid overfitting in deep learning components, techniques such as dropout and regularization were also applied.
The empirical robustness of the models is demonstrated in Section 5.4, which reports performance metrics across both in-sample and out-of-sample evaluation windows. The TGAN model shows consistently superior performance compared to fBm and Brownian motion baselines, especially on the Q-SEAS dataset. These findings demonstrate that the model does not merely overfit the training data. Instead, it effectively captures underlying structural patterns that generalize to future, previously unseen observations.
The validation strategies implemented confirm that the hybrid models are suitable for reproducing the long-term dependence patterns necessary for real-time forecasting in emerging carbon markets. This is particularly relevant in contexts such as the UK ETS, where price dynamics are affected by both structural policy shifts and short-term volatility. The ability of the models to generalize beyond the training regime suggests that they can be reliably integrated into forecasting tools or early-warning systems used by market analysts and regulatory bodies.
While the current robustness checks indicate a certain level of reliability within the UK ETS context, future work should extend these tests across multiple ETS regimes (e.g., EU ETS, California Cap-and-Trade) and under different market phases (e.g., during external shocks or regulatory transitions). Additionally, regime-switching validation frameworks or adversarial robustness tests could further reinforce the empirical stability of hybrid models in high-stakes policy applications.

6. Conclusions and Implications

This study integrates advanced computational methods, including deep learning and statistical modelling, to analyze and simulate the price behaviour of the UK Allowance Futures Contract (UKAFMc1). By applying Simple Linear Regression (SLR), Polynomial Regression, and Fourier Transform Regression, it effectively models the underlying trends and seasonality in carbon market prices using Sequential (SQ) and Mixed (MM) modelling approaches. While some models, including MM3, achieved high R2, the regression residuals retained possible unaccounted properties. The combination of polynomial regression with Fourier-based seasonal adjustments showed that fifth-degree polynomial models offered the best fit. Moreover, the 90-day seasonality better explained the periodic behaviour than the 30-day cycle. LOESS decomposition further highlighted a significant long-term shift in carbon prices, characterized by an initial growth followed by a general downward trend from early 2023 onward. The monthly decomposition (M-SEAS) effectively modelled short-term fluctuations, leaving mostly random noise in the residuals. In contrast, the quarterly decomposition (Q-SEAS) revealed long-term persistence and deeper underlying patterns, as confirmed by a high Hurst parameter, indicating that future models must account for long-term dependencies.
To further explore these dynamics, synthetic time series were generated using fractional Brownian motion (fBm), Adjusted fBm (enhanced with the autocovariance matrix), and Time-series Generative Adversarial Network (TGAN). Comparative analysis revealed that the Adjusted fBm consistently outperformed fBm and TGAN in preserving autocorrelation and cross-correlation properties. However, Adjusted fBm exhibited overfitting during testing, raising concerns about its generalization capabilities. Nevertheless, all synthetic models, including fBm, Adjusted fBm, and TGAN, outperformed random noise—the baseline Brownian motion—confirming the existence of self-similar patterns in UK Carbon Permits and supporting the first hypothesis. To contextualize the performance of the proposed models, we also included an ARIMAX model as a traditional reference while acknowledging the architectural and functional differences between approaches. This provides an additional benchmark, suggesting that synthetic data-driven and neural network-based methods can achieve competitive performance in carbon price forecasting. These findings suggest that the models effectively capture some inherent structures in the data, even if not perfectly. These results are consistent with previous research, reinforcing the presence of long-term memory characteristics in financial time series [76]. Unlike prior studies, which primarily focused on general asset markets, this research applies these methodologies to UK Carbon Permits, demonstrating their relevance in an emissions trading context. This domain-specific validation enhances the robustness of existing models while providing practical implications for policymakers and market participants.
The second hypothesis that concerns the models’ ability to capture long-term dependencies fully remains inconclusive due to TGAN’s poor correlation with the original data. However, TGAN demonstrated superior out-of-sample predictive performance, particularly for Q-SEAS. This outcome aligns with findings from [77]. Additionally, research on China’s Hubei and Guangdong carbon markets by [78] highlighted GANs’ ability to leverage multi-source information and extract time-series features, enhancing their performance in modelling carbon price fluctuations. Moreover, using TGAN for data augmentation in the EU Emissions Trading System (ETS) Phase 4 improved prediction accuracy by generating expanded datasets, which enhanced the training process [79]. Although TGAN exhibited poor correlation with the original data, its strong out-of-sample performance suggests potential for further improvement. Addressing these limitations requires refining the GAN architecture and hyperparameters to enhance model stability and better replicate the statistical properties of the original data.
This study offers key economic implications for carbon market participants, policymakers, and financial institutions. The observed long-range dependence in carbon prices suggests that price movements are not purely random but are influenced by historical price trends and regulatory interventions rather than purely current market conditions. This finding aligns with research that challenges the Efficient Market Hypothesis in carbon trading, as price signals may not always align with fundamental abatement costs. This directly affects market efficiency, as predictable structures may allow investors to develop data-driven trading strategies, which have implications for speculative trading and market stability. Since the UK ETS deviates from strong market efficiency, regulatory actions and market expectations of future interventions introduce distortions, potentially affecting investment decisions. To mitigate such risks, the superior predictive performance of models like TGAN and Adjusted fBm suggests strong potential for algorithmic trading applications, provided safeguards are in place to monitor model-driven market dynamics.
Based on the findings and model evaluations, the following time-bound policy recommendations are proposed to enhance the performance, transparency, and resilience of the UK ETS market. In the short term, it is recommended to establish a transparent allowance release calendar that aligns with emission projections and prevailing market conditions. This would help reduce market volatility and discourage speculative trading behaviour. A dynamic price collar—setting a flexible minimum and maximum price range—should be introduced and adjusted semi-annually to prevent excessive price fluctuations while maintaining the integrity of market-based price discovery mechanisms. Additionally, quarterly policy outlooks should be published, including projected intervention thresholds, emissions trends, and key market risk indicators, to enhance transparency and build market confidence.
Over the medium term, the reintroduction or adaptation of a UK-specific Market Stability Reserve (MSR) is recommended. This mechanism should be informed by long-memory indicators and regime-switch detection methods, such as those discussed in Section 5.2 and Section 5.4, to allow for dynamic management of certificate supply in response to market signals. Furthermore, regulatory bodies should pilot the adoption of AI-based monitoring tools, leveraging advanced deep learning models like TGAN and fractional Brownian motion (fBm) for early warning systems. These tools would be instrumental in detecting structural shifts or emerging instabilities in price behaviour.
Looking ahead, partial interoperability with the EU Emissions Trading System (EU ETS) should be pursued. This integration would enhance market liquidity and reduce arbitrage-induced volatility, as supported by the autocorrelation analyses presented in the study. Additionally, the incorporation of forecasting models with long-memory structures into regulatory planning and carbon budget simulations is essential. Such integration would ensure that short-term volatility management is effectively aligned with long-term decarbonization goals. The implementation of these policy measures would significantly enhance the operational efficiency and stability of the UK Emissions Trading Scheme (UK ETS), while promoting a forward-looking, data-informed regulatory approach to carbon market governance.
This study contributes to the understanding of carbon price dynamics by linking empirical evidence from the UK ETS with established economic theory on carbon pricing mechanisms. The findings highlight that cap-and-trade markets exhibit more complex price behaviours than traditional economic models predict, with long-memory effects and regulatory sensitivity playing key roles in price fluctuations. This research provides a stronger foundation for designing more effective carbon pricing policies, ensuring that market mechanisms remain efficient in driving emissions reductions and fostering low-carbon investments.

Author Contributions

Conceptualization, Z.D. and C.O.-S.; methodology, Z.D.; software, Z.D.; validation, C.O.-S. and D.B.-L.; formal analysis, Z.D.; investigation, C.O.-S.; data curation, D.B.-L.; writing—original draft preparation, Z.D.; writing—review and editing, C.O.-S.; visualization, D.B.-L.; supervision, D.B.-L.; project administration, C.O.-S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Lucian Blaga University of Sibiu under Grant LBUS-IRG-2024.

Data Availability Statement

Data are available on reasonable request.

Acknowledgments

This work was supported by the EU’s NextGenerationEU instrument through the National Recovery and Resilience Plan of Romania—Pillar III-C9-I8, managed by the Ministry of Research, Innovation and Digitalization, within the project entitled “Non-Gaussian self–similar processes: Enhancing mathematical tools and financial models for capturing complex market dynamics”, contract no. 760243/28.12.2023, code CF 194/31.07.2023.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
fBmFractional Brownian Motion
GANsGenerative Adversarial Networks
EU ETSEuropean Union Emissions Trading System
UK ETSUK Emissions Trading Scheme
UKAsUK allowances
UKAFMc1United Kingdom Emission Energy Allowances Future contracts
EMHEfficient Market Hypothesis
LSTMLong Short-Term Memory
LOESSLocally Estimated Scatterplot Smoothing
M-SEAS/Q-SEASmonthly/quarterly decompositions

References

  1. Su, C.-W.; Pang, L.-D.; Qin, M.; Lobonţ, O.-R.; Umar, M. The spillover effects among fossil fuel, renewables and carbon markets: Evidence under the dual dilemma of climate change and energy crises. Energy 2023, 274, 127304. [Google Scholar] [CrossRef]
  2. Li, C.; Chen, Z.; Hu, Y.; Cai, C.; Zuo, X.; Shang, G.; Lin, H. The energy conservation and emission reduction co-benefits of China’s emission trading system. Sci. Rep. 2023, 13, 13758. [Google Scholar] [CrossRef]
  3. Richardson-Barlow, C.; Pimm, A.J.; Taylor, P.G.; Gale, W.F. Policy and pricing barriers to steel industry decarbonisation: A UK case study. Energy Policy 2022, 168, 113100. [Google Scholar] [CrossRef]
  4. Flora, M.; Vargiolu, T. Price dynamics in the European Union Emissions Trading System and evaluation of its ability to boost emission-related investment decisions. Eur. J. Oper. Res. 2020, 280, 383–394. [Google Scholar] [CrossRef]
  5. Li, W.; Li, J.-P.; Su, C.-W. Are economic policy uncertainty and carbon futures price interacting? Evidence from the European Union. Heliyon 2023, 9, e21132. [Google Scholar] [CrossRef]
  6. Li, X.; Li, Z.; Su, C.-W.; Umar, M.; Shao, X. Exploring the asymmetric impact of economic policy uncertainty on China’s carbon emissions trading market price: Do different types of uncertainty matter? Technol. Forecast. Soc. Change 2022, 178, 121601. [Google Scholar] [CrossRef]
  7. Su, C.W.; Wei, S.; Wang, Y.; Tao, R. How does climate policy uncertainty affect the carbon market? Technol. Forecast. Soc. Change 2023, 200, 123155. [Google Scholar] [CrossRef]
  8. Li, M.-F.; Hu, H.; Zhao, L.-T. Key factors affecting carbon prices from a time-varying perspective. Environ. Sci. Pollut. Res. 2022, 29, 65144–65160. [Google Scholar] [CrossRef] [PubMed]
  9. Liu, S.; Zhang, Y.; Wang, J.; Feng, D. Fluctuations and Forecasting of Carbon Price Based on A Hybrid Ensemble Learning GARCH-LSTM-Based Approach: A Case of Five Carbon Trading Markets in China. Sustainability 2024, 16, 1588. [Google Scholar] [CrossRef]
  10. Yang, Y.; Guo, H.; Jin, Y.; Song, A. An Ensemble Prediction System Based on Artificial Neural Networks and Deep Learning Methods for Deterministic and Probabilistic Carbon Price Forecasting. Front. Environ. Sci. 2021, 9, 740093. [Google Scholar] [CrossRef]
  11. Lai, G.; Chang, W.C.; Yang, Y.; Liu, H. Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017. [Google Scholar]
  12. Marwala, T.; Fournier-Tombs, E.; Stinckwich, S. The Use of Synthetic Data to Train AI Models: Opportunities and Risks for Sustainable Development. arXiv 2023, arXiv:2309.00652. [Google Scholar] [CrossRef]
  13. Jones, L.B.; Burns, C.J. REACHing for divergence?—UK chemical regulation post-Brexit. Integr. Environ. Assess. Manag. 2024, 20, 1529–1538. [Google Scholar] [CrossRef] [PubMed]
  14. Nguyen, W. Investigating the Distributional Effect of the UK Carbon Price Floor. UCL J. Econ. 2023, 2, 4–11. [Google Scholar] [CrossRef]
  15. Fama, E.F. Efficient Capital Markets: A Review of Theory and Empirical Work. J. Financ. 1970, 25, 383–417. [Google Scholar] [CrossRef]
  16. Cleveland, R.B.; Cleveland, W.; McRae, J.; Terpenning, I. STL: A Seasonal-Trend Decomposition Procedure Based on Loess. J. Off. Stat. 1990, 6, 3–73. [Google Scholar]
  17. Lin, B.; Wesseh, P.K. On the economics of carbon pricing: Insights from econometric modeling with industry-level data. Energy Econ. 2020, 86, 104678. [Google Scholar] [CrossRef]
  18. Kohlscheen, E.; Moessner, R.; Takats, E. Effects of Carbon Pricing and Other Climate Policies on CO2 emissions. SSRN Electron. J. 2021. [Google Scholar] [CrossRef]
  19. Carbone, J.C.; Rivers, N.; Yamazaki, A.; Yonezawa, H. Comparing Applied General Equilibrium and Econometric Estimates of the Effect of an Environmental Policy Shock. J. Assoc. Environ. Resour. Econ. 2020, 7, 687–719. [Google Scholar] [CrossRef]
  20. Lee, C.; Lee, J.; Park, S. Forecasting the urbanization dynamics in the Seoul metropolitan area using a long short-term memory–based model. Environ. Plan. B Urban Anal. City Sci. 2023, 50, 453–468. [Google Scholar] [CrossRef]
  21. Cao, Y.; Zha, D.; Wang, Q.; Wen, L. Probabilistic carbon price prediction with quantile temporal convolutional network considering uncertain factors. J. Environ. Manag. 2023, 342, 118137. [Google Scholar] [CrossRef]
  22. Luo, J.; Hong, T.; Gao, Z.; Fang, S.-C. A robust support vector regression model for electric load forecasting. Int. J. Forecast. 2023, 39, 1005–1020. [Google Scholar] [CrossRef]
  23. Wang, H.; Xue, W.; Liu, Y.; Peng, J.; Jiang, H. Probabilistic wind power forecasting based on spiking neural network. Energy 2020, 196, 117072. [Google Scholar] [CrossRef]
  24. Wang, X.; Li, X.; Li, S. Point and interval forecasting system for crude oil price based on complete ensemble extreme-point symmetric mode decomposition with adaptive noise and intelligent optimization algorithm. Appl. Energy 2022, 328, 120194. [Google Scholar] [CrossRef]
  25. Frederiks, E.R.; Stenner, K.; Hobman, E.V.; Fischle, M. Evaluating energy behavior change programs using randomized controlled trials: Best practice guidelines for policymakers. Energy Res. Soc. Sci. 2016, 22, 147–164. [Google Scholar] [CrossRef]
  26. Vrolijk, K.; Sato, M. Quasi-Experimental Evidence on Carbon Pricing. World Bank Res. Obs. 2023, 38, 213–248. [Google Scholar] [CrossRef]
  27. Döbbeling-Hildebrandt, N.; Miersch, K.; Khanna, T.M.; Bachelet, M.; Bruns, S.B.; Callaghan, M.; Edenhofer, O.; Flachsland, C.; Forster, P.M.; Kalkuhl, M.; et al. Systematic review and meta-analysis of ex-post evaluations on the effectiveness of carbon pricing. Nat. Commun. 2024, 15, 4147. [Google Scholar] [CrossRef]
  28. Ariyawansa, T.; Amirthalingam, K. A Comparative Mathematical Study of the Relationship Between Marginal Social Cost and Pigouvian Tax in the Presence of Commodity and Wage Taxes: Putting Ramsey theorem into Practice. Vidyodaya J. Manag. 2022, 8, 47–65. [Google Scholar] [CrossRef]
  29. Hong, S.; Sim, S.-G. Inelastic Supply of Fossil Energy and Competing Environmental Regulatory Policies. Sustainability 2018, 10, 287. [Google Scholar] [CrossRef]
  30. Yao, Q.; Shao, L.; Yin, Z.; Wang, J.; Lan, Y. Strategies of property developers and governments under carbon tax and subsidies. Front. Environ. Sci. 2022, 10, 916352. [Google Scholar] [CrossRef]
  31. Wang, H.; Li, Y.; Bu, G. How carbon trading policy should be integrated with carbon tax policy—Laboratory evidence from a model of the current state of carbon pricing policy in China. Environ. Sci. Pollut. Res. 2022, 30, 23851–23869. [Google Scholar] [CrossRef]
  32. Haites, E. Carbon taxes and greenhouse gas emissions trading systems: What have we learned? Clim. Policy 2018, 18, 955–966. [Google Scholar] [CrossRef]
  33. Cronin, J.A.; Fullerton, D.; Sexton, S. Vertical and Horizontal Redistributions from a Carbon Tax and Rebate. J. Assoc. Environ. Resour. Econ. 2019, 6, S169–S208. [Google Scholar] [CrossRef]
  34. Nowlin, M.C.; Gupta, K.; Ripberger, J.T. Revenue use and public support for a carbon tax. Environ. Res. Lett. 2020, 15, 084032. [Google Scholar] [CrossRef]
  35. Zhao, A.; Song, X.; Li, J.; Yuan, Q.; Pei, Y.; Li, R.; Hitch, M. Effects of Carbon Tax on Urban Carbon Emission Reduction: Evidence in China Environmental Governance. Int. J. Environ. Res. Public Health 2023, 20, 2289. [Google Scholar] [CrossRef] [PubMed]
  36. Gao, X.; Zhang, Y. Feasibility Study of China’s Carbon Tax System under the Carbon Neutrality Target—Based on the CGE Model. Sustainability 2023, 15, 1026. [Google Scholar] [CrossRef]
  37. Qin, G.; Tao, F.; Li, L.; Chen, Z. Optimization of the simultaneous pickup and delivery vehicle routing problem based on carbon tax. Ind. Manag. Data Syst. 2019, 119, 2055–2071. [Google Scholar] [CrossRef]
  38. Liu, H. The Carbon Tax Effect on British Columbia Economy and Carbon Emission. BCP Bus. Manag. 2023, 40, 233–240. [Google Scholar] [CrossRef]
  39. Koch, N.; Fuss, S.; Grosjean, G.; Edenhofer, O. Causes of the EU ETS price drop: Recession, CDM, renewable policies or a bit of everything?—New evidence. Energy Policy 2014, 73, 676–685. [Google Scholar] [CrossRef]
  40. Zhang, T.; Zou, S. Carbon Market Evaluation Based on Random Walk Hypothesis in China. Math. Probl. Eng. 2022, 2022, 5726108. [Google Scholar] [CrossRef]
  41. Oprean, C.; Tanasescu, C. Applications of Chaos and Fractal Theory on Emerging Capital Markets. Int. J. Acad. Res. Bus. Soc. Sci. 2013, 3. [Google Scholar] [CrossRef]
  42. Camelia, O.; Cristina, T.; Amelia, B. A new proposal for efficiency quantification of capital markets in the context of complex non-linear dynamics and chaos. Econ. Res. Istraz. 2017, 30, 1669–1692. [Google Scholar] [CrossRef]
  43. Chevallier, J. Carbon Price Drivers. Int. J. Appl. Logist. 2013, 4, 1–7. [Google Scholar] [CrossRef]
  44. Wei, P.; Zhou, J.; Ren, X.; Huynh, L.D.T. Financialisation of the European Union Emissions Trading System and its influencing factors in quantiles. Int. J. Financ. Econ. 2025, 30, 925–940. [Google Scholar] [CrossRef]
  45. Lo, A.W. Adaptive Markets and the New World Order (corrected May 2012). Financ. Anal. J. 2012, 68, 18–29. [Google Scholar] [CrossRef]
  46. Perino, G. New EU ETS Phase 4 rules temporarily puncture waterbed. Nat. Clim. Change 2018, 8, 262–264. [Google Scholar] [CrossRef]
  47. Liu, R.; Jing, Z. The impacts of introducing carbon prices in the electricity market under China’s carbon neutrality goal. E3S Web Conf. 2021, 252, 01012. [Google Scholar] [CrossRef]
  48. de Oliveira, A.M.B.; Mandal, A.; Power, G.J. Impact of COVID-19 on Stock Indices Volatility: Long-Memory Persistence, Structural Breaks, or Both? Ann. Data Sci. 2024, 11, 619–646. [Google Scholar] [CrossRef]
  49. Brătian, V.; Acu, A.-M.; Oprean-Stan, C.; Dinga, E.; Ionescu, G.-M. Efficient or Fractal Market Hypothesis? A Stock Indexes Modelling Using Geometric Brownian Motion and Geometric Fractional Brownian Motion. Mathematics 2021, 9, 2983. [Google Scholar] [CrossRef]
  50. Li, F.; Zheng, X. Carbon Awareness and Return Co-movement. SSRN Electron. J. 2024. [Google Scholar] [CrossRef]
  51. Xu, L.; Deng, S.-J.; Thomas, V.M. Carbon emission permit price volatility reduction through financial options. Energy Econ. 2014, 53, 248–260. [Google Scholar] [CrossRef]
  52. Brophy, E.; Wang, Z.; She, Q.; Ward, T. Generative Adversarial Networks in Time Series: A Systematic Literature Review. ACM Comput. Surv. 2023, 55, 1–31. [Google Scholar] [CrossRef]
  53. Mandelbrot, B.B.; Wallis, J.R. Robustness of the rescaled range R/S in the measurement of noncyclic long run statistical dependence. Water Resour. Res. 1969, 5, 967–988. [Google Scholar] [CrossRef]
  54. Hurst, H.E. Long-Term Storage Capacity of Reservoirs. Trans. Am. Soc. Civ. Eng. 1951, 116, 770–808. [Google Scholar] [CrossRef]
  55. Weron, R.; Przybyłowicz, B. Hurst analysis of electricity price dynamics. Phys. A Stat. Mech. Its Appl. 2000, 283, 462–468. [Google Scholar] [CrossRef]
  56. Lo, A.W. Long-Term Memory in Stock Market Prices. Econometrica 1991, 59, 1279. [Google Scholar] [CrossRef]
  57. Davies, R.B.; Harte, D.S. Tests for Hurst effect. Biometrika 1987, 74, 95–101. [Google Scholar] [CrossRef]
  58. Gray, R.M.; Davisson, L.D. An Introduction to Statistical Signal Processing; Cambridge University Press (CUP): Cambridge, UK, 2004. [Google Scholar]
  59. Rubín, T.; Panaretos, V.M. Spectral Simulation of Functional Time Series. arXiv 2020, arXiv:2007.08458. [Google Scholar] [CrossRef]
  60. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 28th International Conference on Neural Information Processing Systems—Volume 2 (NIPS’14), Montreal, QC, Canada, 8–13 December 2014; MIT Press: Cambridge, MA, USA, 2014; pp. 2672–2680. [Google Scholar]
  61. Yoon, J.; Jarrett, D.; Van der Schaar, M. Time-series Generative Adversarial Networks. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, ON, Canada, 8–14 December 2019. [Google Scholar]
  62. Kwon, S.; Lee, Y. Can GANs Learn the Stylized Facts of Financial Time Series? In Proceedings of the 5th ACM International Conference on AI in Finance, Brooklyn, NY, USA, 14–17 November 2024; ACM: New York, NY, USA, 2024; pp. 126–133. [Google Scholar]
  63. Boursin, N.; Remlinger, C.; Mikael, J. Deep Generators on Commodity Markets Application to Deep Hedging. Risks 2022, 11, 7. [Google Scholar] [CrossRef]
  64. Jeon, J.; Kim, J.; Song, H.; Cho, S.; Park, N. GT-GAN: General Purpose Time Series Synthesis with Generative Adversarial Networks. Adv. Neural Inf. Process. Syst. 2022, 36999. Available online: https://proceedings.neurips.cc/paper_files/paper/2022/hash/f03ce573aa8bce26f77b76f1cb9ee979-Abstract-Conference.html (accessed on 9 January 2025).
  65. Chen, K.; Zhou, Y.; Dai, F. A LSTM-based method for stock returns prediction: A case study of China stock market. In Proceedings of the IEEE International Conference on Big Data (Big Data), Santa Clara, CA, USA, 29 October–1 November 2015; pp. 2823–2824. [Google Scholar]
  66. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  67. Pei, H.; Ren, K.; Yang, Y.; Liu, C.; Qin, T.; Li, D. Towards Generating Real-World Time Series Data. In Proceedings of the 2021 IEEE International Conference on Data Mining (ICDM), Auckland, New Zealand, 7–10 December 2021; pp. 469–478. [Google Scholar]
  68. Hyndman, R.J.; Athanasopoulos, G. Forecasting: Principles and Practice, 2nd ed.; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  69. Watson, F. Carbonwise. Available online: https://carbonwise.co/uk-emissions-aall-12-5-in-2023-work-moves-forward-on-ccp-labelled-carbon-credits/ (accessed on 24 February 2025).
  70. Department for Energy Security and Net Zero. Digest of UK Energy Statistics (DUKES). London; 2023. Available online: https://www.gov.uk/government/collections/digest-of-uk-energy-statistics-dukes (accessed on 3 January 2025).
  71. Dumont, V.; Ju, X.; Mueller, J. Hyperparameter Optimization of Generative Adversarial Network Models for High-Energy Physics Simulations. arXiv 2022, arXiv:2208.07715. [Google Scholar] [CrossRef]
  72. Pinheiro, D.N.; Gonzalez, J.C.; Corso, G.; Gebre, M.G.; da Costa, C.A.; Xavier-De-Souza, S.; Barros, T. Hyperparameter determination for GAN-based seismic interpolator with variable neighborhood search. Comput. Geosci. 2024, 192, 105689. [Google Scholar] [CrossRef]
  73. Chowdhury, A.A.; Das, A.; Hoque, K.K.S.; Karmaker, D. A Comparative Study of Hyperparameter Optimization Techniques for Deep Learning; Springer Nature: Singapore, 2022; pp. 509–521. [Google Scholar] [CrossRef]
  74. Vrskova, R.; Sykora, P.; Kamencay, P.; Hudec, R.; Radil, R. Hyperparameter Tuning of ConvLSTM Network Models. In Proceedings of the 2021 44th International Conference on Telecommunications and Signal Processing (TSP), Virtual, 26–28 July 2021; pp. 15–18. [Google Scholar]
  75. Wojciuk, M.; Swiderska-Chadaj, Z.; Siwek, K.; Gertych, A. Improving classification accuracy of fine-tuned CNN models: Impact of hyperparameter optimization. Heliyon 2024, 10, e26586. [Google Scholar] [CrossRef] [PubMed]
  76. Aslam, F.; Latif, S.; Ferreira, P. Investigating Long-Range Dependence of Emerging Asian Stock Markets Using Multifractal Detrended Fluctuation Analysis. Symmetry 2020, 12, 1157. [Google Scholar] [CrossRef]
  77. Lin, Z.; Jain, A.; Wang, C.; Fanti, G.; Sekar, V. Using GANs for Sharing Networked Time Series Data. In Proceedings of the ACM Internet Measurement Conference, Virtual, 27–29 October 2020; ACM: New York, NY, USA, 2020; pp. 464–483. [Google Scholar]
  78. Huang, Z.; Zhang, W. Forecasting carbon prices in China’s pilot carbon market: A multi-source information approach with conditional generative adversarial networks. J. Environ. Manag. 2024, 359, 120967. [Google Scholar] [CrossRef] [PubMed]
  79. Liu, D.; Chen, K.; Cai, Y.; Tang, Z. Interpretable EU ETS Phase 4 prices forecasting based on deep generative data augmentation approach. Finance Res. Lett. 2024, 61, 105038. [Google Scholar] [CrossRef]
Figure 1. Conceptual architecture of the Time-series Generative Adversarial Network (TGAN).
Figure 1. Conceptual architecture of the Time-series Generative Adversarial Network (TGAN).
Fractalfract 09 00350 g001
Figure 2. The price behaviour of UK Emission Trading Scheme (ETS) allowances (2022–2023). Source: Refinitiv Eikon.
Figure 2. The price behaviour of UK Emission Trading Scheme (ETS) allowances (2022–2023). Source: Refinitiv Eikon.
Fractalfract 09 00350 g002
Figure 3. Trend, seasonal, and residual components. Source: Authors’ representations based on the LOESS decomposed data.
Figure 3. Trend, seasonal, and residual components. Source: Authors’ representations based on the LOESS decomposed data.
Fractalfract 09 00350 g003
Figure 4. The learning process of the TGAN network. Source: Authors’ representations.
Figure 4. The learning process of the TGAN network. Source: Authors’ representations.
Fractalfract 09 00350 g004
Table 1. Comparative evaluation of modelling paradigms for carbon price forecasting.
Table 1. Comparative evaluation of modelling paradigms for carbon price forecasting.
Analytical CriterionTraditional Econometric ModelsComputable General Equilibrium (CGE) ModelsMachine/Deep Learning Models
Theoretical foundationStrong (economically grounded)Very strong (macroeconomic equilibrium theory)Limited (primarily data-driven)
InterpretabilityHigh (parameter-based, transparent)High (clearly structured assumptions)Low (black-box characteristics)
Predictive accuracyModerate (limited for non-linear or volatile series)Weak (low temporal granularity)High (strong in non-linear and high-dimensional data
Ability to capture long memory and seasonalityLimited (short autoregressive memory)AbsentExcellent (designed for long-range temporal dependencies)
Robustness and stabilityHigh (under classical assumptions)Sensitive to calibration and elasticitiesMedium–Low (overfitting, parameter sensitivity)
Data requirementsLow to moderateVery high (macro-aggregated data)High (large, high-frequency datasets needed)
Adaptability to emerging carbon marketsLow (requires stationarity, linearity)Very low (static, scenario-based)High (flexible to structural change and noise)
Policy relevanceHigh (interpretable, scenario-driven)High (macro policy analysis)Low (difficult to translate into actionable recommendations)
Key limitationsPoor handling of non-linearity and regime shiftsInflexible, overly abstractLack of transparency, risk of overfitting, training instability
Core strengthsTransparent, interpretable, easy to validateTheoretically rigorous, system-wide simulationsHigh-dimensional learning, adaptability, forecasting power
Table 2. UK ETS descriptive statistics (2022–2023).
Table 2. UK ETS descriptive statistics (2022–2023).
StatisticValueStatisticValue
Basic Statistics Temporal Statistics
Mean65.91Mean of First Difference−0.06
Median71.08Std Dev of First Difference1.77
Mode79.95Mean of 30-day Rolling Means66.44
Range64.59Std Dev of 30-day Rolling Means15.46
Variance270.01Mean of 90-day Rolling Means67.44
Standard Deviation16.43Std Dev of 90-day Rolling Means13.85
Interquartile Range (IQR)29.62Mean Percentage Change
Skewness−0.4830-day Windows−0.025
Kurtosis−1.0390-day Windows−0.066
Extremes
Minimum Value32.56
Maximum Value97.15
Source: Authors’ calculations.
Table 3. Statistics of linear regressions.
Table 3. Statistics of linear regressions.
SLR2nd°3rd°4th°5th°6th°
Akaike Information Criterion (AIC) 3083.2464753.8834692.9844659.2824636.0326628.589
Bayesian Information Criterion (BIC)3074.0744740.1244674.6404636.3514608.5156605.659
Shapiro–Wilk (SW) p-values0.0000.0000.0000.0000.0000.000
R20.7430.8510.8640.8710.877−0.945
Source: Authors’ calculations.
Table 4. Statistics of combined linear and Fourier Transformed Regressions.
Table 4. Statistics of combined linear and Fourier Transformed Regressions.
SQ1SQ2SQ3SQ4MM1MM2MM3
AIC3079.8083056.3492553.4832515.2835135.2695108.5234643.058
BIC3102.7393088.4522590.1722579.4895153.6145126.8684670.575
Fourier R20.0010.0380.0020.069---
R2----0.7430.7530.871
SW p-values0.0000.0000.0000.0000.0000.0000.000
Harmonic Order1124111
Polynomial DegreeSLRSLR55SLRSLR3
Period Length30903090309090
Source: Authors’ calculations.
Table 5. Time series statistics.
Table 5. Time series statistics.
Time SeriesADF Statisticp-ValueShapiro Statisticp-ValueHurst Parameter
Monthly Season−7.4770.0000.9970.2320.566
Quarterly Season−6.5030.0000.9890.0000.758
Source: Authors’ calculations.
Table 6. Ablation study.
Table 6. Ablation study.
IterationMethodDiscriminative (Absolute Accuracy)Predictive (MSE)
M–SEASGAN0.02036.755414
Q–SEAS0.04833.68814
M–SEASTGAN0.03721.23505
Q–SEAS0.09130.847832
M–SEASWTGAN0.04171.702863
Q–SEAS0.06371.78071
Table 7. Model configuration.
Table 7. Model configuration.
Properties/HyperparametersM-SEASQ-SEAS
Generator Learning Rate (LR)Adaptive Moment Estimation (ADAM) 10−3
Discriminator Learning Rate (LR)Adaptive Moment Estimation (ADAM) 10−4
Embeddings Learning Rate (LR)Adaptive Moment Estimation (ADAM) 10−4
Generator Structure1× Perceptron Layer—64 Neurons
1× LSTM Layer—64 Neurons
2× Dropout Layers—10%
Output Layer—Sigmoid Activation
Discriminator Structure1× LSTM Layer—64 Neurons
1× Perceptron Layer—64 Neurons
2× Dropout Layers—10%
Output Layer—Sigmoid Activation
Embedding Structure1× LSTM Layer—64 Neurons
1× Perceptron Layer—64 Neurons—Relu Activation
2× Dropout Layers—10%
Recovery Structure1× LSTM—64 Neurons
Lags20 Consecutive Lags50 Consecutive Lags
Noise Dimension55
Batches50100
Epochs10001400
Loss FunctionBinary Cross Entropy
Hurst Iteration 10.6030.721
Hurst Iteration 20.5740.782
Hurst Iteration 30.5770.822
Source: Authors’ calculations.
Table 8. M-SEAS synthetic data (20 lags).
Table 8. M-SEAS synthetic data (20 lags).
IterationMethodMean Correlation (Auto)AUC (Auto)Mean Correlation (Cross)AUC (Cross)
1TGAN0.0992.206−0.0112.223
20.0701.962−0.0150.992
30.0601.712−0.0121.316
1fBm0.97020.365−0.0190.764
20.90318.9690.0170.832
30.87718.422−0.0381.739
1Adjusted fBm 0.50710.643−0.0663.234
20.4229.9240.0264.668
30.35810.229−0.0235.270
Source: Authors’ calculations.
Table 9. Q-SEAS synthetic data (50 lags).
Table 9. Q-SEAS synthetic data (50 lags).
IterationMethodMean Correlation (Auto)AUC (Auto)Mean Correlation (Cross)AUC (Cross)
1TGAN0.0894.6310.0062.814
20.1055.520−0.0042.343
30.1567.958−0.0052.957
1fBm0.87244.494−0.0414.176
20.59930.5430.16216.422
30.73337.404−0.0103.379
1Adjusted fBm 0.65933.6030.0246.113
20.55728.4060.16517.352
30.49725.322−0.0739.789
Source: Authors’ calculations.
Table 10. M-SEAS regression results (20 lags).
Table 10. M-SEAS regression results (20 lags).
StepTGANfBmAdjusted fBmBM
LossMAELossMAELossMAELossMAE
F14508.03166.796779.59723.2893502.96854.3273574.50058.674
F2196.43011.28887.6227.774121.5739.298113.8458.870
F392.8657.41738.3854.98069.2596.29424.4493.976
F494.3777.77362.7075.98178.9926.83255.0995.920
F548.9015.12642.8385.01728.0424.07341.0084.987
Average988.12119.680202.2309.408760.16716.165735.55016.361
Test MAE-31.717-66.868-28.276-168.457
ARIMAX-8.509------
p-value-0.000-0.000-0.000-0.000
Source: Authors’ calculations.
Table 11. Q-SEAS regression results (50 lags).
Table 11. Q-SEAS regression results (50 lags).
StepTGANfBmAdjusted fBmBM
LossMAELossMAELossMAELossMAE
F14962.96270.3445332.58072.9291167.42928.5013129.71751.720
F25993.59877.2701425.07032.9941669.34629.9663641.66259.421
F36111.87878.065633.22921.1684506.06962.9055497.76270.583
F4283.43912.61239.7174.8924586.61866.0848370.82983.445
F563.3966.1916166.84178.2596123.02877.976858.11820.696
Average3483.05548.8972719.48742.0483610.49853.0864299.61857.173
Test MAE-38.395-42.002-72,305.797-113.539
ARIMAX-8.567------
p-value-0.000-0.000-0.000-0.000
Source: Authors’ calculations.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dinca, Z.; Oprean-Stan, C.; Balsalobre-Lorente, D. UK Carbon Price Dynamics: Long-Memory Effects and AI-Based Forecasting. Fractal Fract. 2025, 9, 350. https://doi.org/10.3390/fractalfract9060350

AMA Style

Dinca Z, Oprean-Stan C, Balsalobre-Lorente D. UK Carbon Price Dynamics: Long-Memory Effects and AI-Based Forecasting. Fractal and Fractional. 2025; 9(6):350. https://doi.org/10.3390/fractalfract9060350

Chicago/Turabian Style

Dinca, Zeno, Camelia Oprean-Stan, and Daniel Balsalobre-Lorente. 2025. "UK Carbon Price Dynamics: Long-Memory Effects and AI-Based Forecasting" Fractal and Fractional 9, no. 6: 350. https://doi.org/10.3390/fractalfract9060350

APA Style

Dinca, Z., Oprean-Stan, C., & Balsalobre-Lorente, D. (2025). UK Carbon Price Dynamics: Long-Memory Effects and AI-Based Forecasting. Fractal and Fractional, 9(6), 350. https://doi.org/10.3390/fractalfract9060350

Article Metrics

Back to TopTop