1. Introduction
The quantification of tail risk constitutes a fundamental challenge in modern statistical science, with profound implications for financial stability and climate resilience [
1,
2]. While extreme value theory has matured significantly since its inception in the work of Fisher and Tippett [
3], contemporary applications demand information-theoretic tools that transcend traditional moment-based analyses [
4,
5]. This imperative emerges from the limitations in existing risk measures, which were vividly exposed during the 2008 financial crisis [
6], recent climate catastrophes [
7], and the COVID-19 pandemic [
8,
9].
Traditional entropy metrics [
10,
11], though powerful for central tendency analysis, fail in tail regimes where standard distributional assumptions collapse. This failure concretely manifests in financial bubble detection latency [
12,
13,
14] and climate risk underestimation [
15,
16]. This paper addresses this dichotomy through three pivotal advances: (1) the development of
Tail-Weighted Entropy (TWE), whose measures are compatible with peaks-over-threshold frameworks, (2) the derivation of entropy decay laws linking information dynamics to tail index parameters, and (3) the construction of entropic hypothesis tests with exact asymptotic properties. The financial implications are immediate: the proposed framework detects speculative bubbles earlier than current moment-based approaches.
Foundational work in extreme value theory by Gnedenko [
17], Balkema and De Haan [
18], Pickands III [
19], and Leadbetter et al. [
20] established mathematical frameworks for tail analysis, later operationalized through peaks-over-threshold (POT) methodologies. Parallel developments in information theory by Shannon [
21], Kullback and Leibler [
22], and Jaynes [
23] created powerful tools for uncertainty quantification, although their initial applications focused on central tendencies [
10,
24]. The limitations of Shannon entropy in tail regimes were first identified by Jaynes [
25], who highlighted information loss in extreme value analysis. Modern risk management frameworks [
26] subsequently revealed the concrete consequences of this limitation through a systematic underestimation of the financial tail risks.
Recent advances follow two main trajectories: Chen and Singh [
27] developed climate-specific tail entropy measures, extended by Faranda et al. [
28], who incorporated dynamical systems theory for extreme weather attribution, and Ragone and Bouchet [
29] who applied information-theoretic rare event sampling in climate models. Concurrently, financial applications have seen innovations such as Billio et al. [
30]’s entropy-based systemic risk indicators, augmented by Simsekli et al. [
31] using neural networks to estimate tail entropy in high-dimensional markets and Ibragimov et al. [
32] establishing geometric properties of divergence measures under heavy-tailed distributions. However, these approaches remain disconnected from POT frameworks—climate applications often rely on block maxima methods [
15] while financial implementations use ad hoc thresholding [
33]—resulting in estimators that are sensitive to distributional misspecification. Bridging this gap, Ardakani [
8] recently proposed a POT-compatible entropy estimator but lacked the necessary theoretical links to extremal index estimation. This paper synthesizes these strands through measure-theoretic innovations that preserve asymptotic properties while enabling practical implementation.
The primary contributions of this paper are threefold. First, the TWE generalizes Shannon entropy to extreme regimes through conditional survival analysis. Second, the results establish asymptotic bounds for TWE under generalized Pareto distributions, revealing explicit dependence on the tail index. Third, an asymptotically normal test statistic is introduced to enable hypothesis testing for speculative bubbles. The empirical analysis demonstrates superior bubble detection by highlighting the average lead time compared to other methodologies. It also shows threshold stability with a higher consistency ratio compared to the Hill estimator. Applications to E-mini S&P 500, gold, crude oil, and Bitcoin futures markets establish entropy dynamics as fundamental market observables.
The remainder of this paper unfolds as follows.
Section 2 introduces a tail-weighted entropy measure, establishing concentration bounds under generalized Pareto distributions and explicit entropy–decay relationships via digamma functions.
Section 3 develops the dynamics of extremal entropy for explosive financial processes, deriving Kullback–Leibler (KL) divergence growth laws and entropy-based bubble detection thresholds.
Section 4 characterizes information dissipation through entropy production laws and provides exact critical values for phase transition detection using catastrophe theory.
Section 5 applies this framework across major asset classes, demonstrating improvements in early warning signals compared to moment-based approaches while validating threshold stability through Basel III-inspired robustness checks.
Section 6 and
Section 7 discuss the implications for risk management and regulatory surveillance and provide concluding remarks.
2. Extremal Entropy Measures
Early efforts to quantify tail uncertainty focused on moment-based approaches [
18], but these lack sensitivity to higher-order information structures. Jaynes [
25] was the first to highlight the information-theoretic limitations in extreme value analysis, while Chen and Singh [
27] developed tail-entropy measures for climate applications. However, existing formulations do not directly align with POT frameworks [
19]. This section introduces the tail-weighted entropy measure, which integrates with the POT methodology and establishes fundamental concentration bounds for this information-theoretic measure under generalized Pareto (GPD) assumptions. It then provides explicit connections between entropy decay rates and tail heaviness. The proposed TWE formulation addresses the compatibility issues in Chen and Singh [
27] while generalizing the Shannon entropy framework to extreme regimes. The concentration bounds improve upon the moment inequalities in Balkema and De Haan [
18] by incorporating information-theoretic constraints.
Definition 1. For a random variable X with absolutely continuous distribution function (CDF) F and probability density function (PDF) f, the Tail-Weighted Entropy beyond threshold τ iswhere is the survival function. This quantifies the information content in the conditional distribution . The TWE in Definition 1 generalizes classical Shannon entropy [
21], specifically focusing on tail regions and addressing the limitations identified by Jaynes [
25] in extreme value analysis. Unlike the standard entropy
, the TWE captures how information scales with extremeness, which is essential for risk quantification when tail dynamics dominate system behavior. Recently, Chen and Singh [
27] demonstrate similar tail-focused measures in climate extremes, but the proposed formulation offers direct compatibility with peaks-over-threshold frameworks.
Lemma 1. Let with shape parameter , scale , and support . The TWE satisfieswhere as . Furthermore, the entropy decay rate admits the asymptotic expansionwhere is the digamma function. Equality holds in the exponential tail limit . Proof. Let
where
. The conditional density is
The tail entropy becomes
First, compute
I. Let
. Then,
Substitute
,
Thus,
. For
,
The survival function satisfies
, so
Substitute into
,
From the regular variation of
,
. Using properties of the digamma function
,
where the equality follows from the asymptotic expansion of the incomplete gamma function. Combining terms yields the digamma-form expansion. The
case follows from
where
is Euler–Mascheroni. □
Lemma 1 establishes limits on tail entropy for heavy-tailed models, complementing moment-based analyses in Balkema and De Haan [
18]. The entropy decomposition reveals an explicit dependence on the tail index
through the digamma function
. Empirical studies of flood risks [
34] confirm that
leads to entropy values exceeding
by over 40%, aligning with the theoretical phase transition at
. This results bridge information theory and extreme value analysis by showing how entropy production accelerates nonlinearly with tail heaviness via
asymptotics. The digamma relationship explains the threshold effects observed in financial bubbles [
13] through entropy convexity in
, while matching the climate extremality patterns in Coles et al. [
15]. The TWE framework demonstrates that heavier tails (
) permit greater entropy production but reduce predictability—a critical insight for coherent risk measures [
35,
36,
37]. Specifically,
when
, confirming Chen and Singh [
27]’s empirical findings through POT theory.
3. Extreme Value Theory and Entropy
The intersection of extreme value theory and information theory has emerged as a frontier in financial econometrics, particularly for analyzing market instability. Bhaduri [
38] introduces entropy methods for bubble detection, leveraging Shannon entropy to quantify tail risk information. Diks and Panchenko [
39] also advance entropy-based causality testing, revealing nonlinear dependencies in financial markets. Traditional approaches relying on moment explosions [
12,
13,
14] and variance ratios [
40] often miss early warning signals due to their insensitivity to higher-order information structures. The recent literature in heavy-tailed time series provides theoretical foundations for modeling financial extremes but lacks explicit connections to information dynamics [
41]. For instance, Chavez-Demoulin and Davison [
42] formalize extremal dependence in multivariate settings, yet their framework does not incorporate entropy as a measure of tail uncertainty. Concurrently, Billio et al. [
30] develop entropy-based crisis indicators, although their framework remains limited to stationary regimes.
To address nonstationarity, Dimpfl and Peter [
43] and Ardakani [
36] introduce transfer entropy to capture asymmetric information flows in time-varying markets, while Ren and Sun [
44] propose wavelet entropy for detecting structural breaks in non-stationary financial series. Bridging these strands, Costa et al. [
45] establish multiscale entropy frameworks for analyzing complexity in non-equilibrium systems, offering insights applicable to explosive financial processes. This section synthesizes these studies by establishing entropy–decay laws for explosive financial processes, unifying insights from extreme value analysis, non-stationary information dynamics, and entropy-based tail risk measurement.
This section (1) establishes exact asymptotic laws for tail entropy decay in explosive AR processes, (2) quantifies information divergence growth during bubble episodes via KL divergence bounds, and (3) provides explicit phase transition thresholds between stable and bubble regimes. The results generalize the entropy inequalities of Cover [
10] to non-stationary heavy-tailed processes while improving the detection thresholds in Phillips et al. [
13] by incorporating information-theoretic criteria.
Theorem 1. Let follow an explosive autoregressive process with during bubble periods, where are i.i.d. innovations with regularly varying tails of index . Then, the following occur:
- 1.
The tail-weighted entropy grows logarithmically:where the convergence rate is determined by the tail index α. - 2.
The KL divergence between consecutive tail distributions decays polynomially:where depends on the process parameters.
Proof. Part 1. According to the Kesten–Goldie theorem [
46], the stationary distribution of
under
has regularly varying tails with index
, where
. When exceeding threshold
, the Pickands–Balkema–de Haan theorem [
18,
19] implies convergence to a generalized Pareto distribution
, where
.
From Lemma 1, the tail-weighted entropy is
As
,
, so
Normalizing by
,
Part 2. For large
, the explosive term dominates (
). The densities satisfy
in the tail. The KL divergence becomes
Substituting
, we obtain
For regularly varying
,
Thus,
Using
,
□
Theorem 1 establishes an information-theoretic foundation for analyzing financial bubbles, complementing traditional moment-based approaches [
13]. The entropy scaling law in Part 1 reveals how bubble dynamics fundamentally alter tail uncertainty quantification, while Part 2 characterizes the decay of distributional similarity in extreme regimes. The results advance beyond variance-ratio methods [
40] by incorporating the higher-order information structures critical for detecting non-linear bubble mechanisms [
35,
47].
Example 1. Consider a sequence of explosive AR(1) processes indexed by time step where innovations are scaled to produce a diffusion limit. Let with , establishing the volatility scaling. Under the Kesten–Goldie regularity condition [46], the tail index α satisfiesAs , Taylor expansion yields , making tails exponentially decaying rather than heavy. Applying Theorem 1, tail entropy growth becomes dominated by Gaussian-like decaymatching the entropy rate for geometric Brownian motion (GBM) [48]. KL divergence decay transitions from polynomial to exponentialconsistent with log-normal tail distinguishability [10]. This limit reveals three fundamental phase transitions. First, tail entropy scaling crosses over from (heavy-tailed) to (log-normal). Second, KL divergence decay accelerates from power law to Gaussian-type exponential rates. Finally, extreme value index collapses to 0, marking the disappearance of regular variation. This example demonstrates the framework’s universality across tail regimes, with Theorem 1 recovering classical diffusion entropy laws through The logarithmic entropy growth law establishes a universal scaling relationship for explosive processes. This can reveal that bubble formations fundamentally reconfigure tail information structures rather than merely amplifying volatility and explains why variance ratio tests [
47] exhibit detection latency—they monitor second-moment explosions while entropy captures the underlying information geometry shift. The polynomial KL divergence decay quantifies the regulatory implications: distinguishability between successive tail distributions diminishes as
increases, creating detection windows where
-estimates from Theorem 1 determine intervention deadlines. The example demonstrates the framework’s universality, showing how TWE transitions between heavy-tailed and log-normal regimes—a phase change undetectable by extremal dependence measures [
42]. By linking entropy production to the Kesten–Goldie tail index
, this result reconciles Diks and Panchenko [
39]’s causality entropy with Embrechts et al. [
41]’s extremal coupling models, providing microfoundations for multiscale crisis indicators [
45].
Practically, these results enable the real-time monitoring of systemic risk through dual-entropy tracking: rising
signals tail uncertainty proliferation, while decaying
indicates reducing distinguishability between normal and bubble regimes. For regulators, the explicit phase transition thresholds permit the data-driven calibration of circuit-breakers based on
-estimates rather than ad hoc volatility limits [
49]. Asset managers could leverage the TWE-KL divergence ratio to optimize tail hedge ratios, as
quantifies the marginal information cost of tail risk mitigation.
3.1. Optimal Threshold Selection
Threshold selection in extreme value analysis presents a dual challenge: maintaining fidelity to the generalized Pareto model while ensuring estimation stability. Traditional approaches emphasize moment-based criteria [
50] or likelihood maximization [
51], but these overlook the information-theoretic structure of tail dynamics. The framework proposed here addresses this by unifying distributional fidelity (via KL divergence) with entropy stability—an advancement in financial applications, where threshold sensitivity directly affects crisis prevention [
13].
Proposition 1. For price processes with empirical tail density and GPD approximation , the entropy-optimal threshold solveswhere is the admissible range and regulates variance penalization. Under a Lipschitz continuity of , the solution exists and is unique. Proof. Let
. Assuming
is Fréchet differentiable over
, the first variation is
From Lemma 1,
. The Euler–Lagrange equation becomes
Existence and uniqueness follow from
convexity in
(verified via positive-definite Hessian),
compactness (ensuring attainment), and
monotonicity in
(Lemma 1). □
Example 2. Revisit Example 1 with . The KL term becomeswhile the entropy variance scales as . Minimizing yields optimal thresholdThis explicitly links model parameters to threshold choice—heavier trading volumes () lower , while volatility () raises it. Compared to Zhang et al. [52]’s quantile-based rule , the proposed adapts to both drift and regulation intensity λ. Proposition 1 advances threshold selection by embedding information-theoretic stability directly into the optimization. The KL divergence enforces consistency with extreme value theory’s POT framework [
19], while the entropy variance term mitigates overfitting to local tail fluctuations—a key challenge in Thompson [
53]’s MLE approach. Example 2 demonstrates how market microstructure parameters (
) directly govern threshold choice, enabling the scenario-specific calibration absent in heuristic methods.
Practically, implementing Proposition 1 requires (1) estimating
via nonparametric entropy estimators [
54], (2) computing
using robust GPD fits [
52], and (3) tuning
via cross-validation on crisis periods. This operationalizes Basel III’s “sound estimation practices” requirement [
55] by providing auditable thresholds grounded in information theory. For bubble detection, setting
inversely to market volatility (higher
in turbulent periods) stabilizes threshold choices during crises—an improvement over static rules.
Remark 1. The variance term encodes regulatory risk appetite: prioritizes model fit (aggressive bubble signals), while emphasizes stability (reduced false positives). This implements the "precautionary principle" advocated in Ardakani [35]. By unifying information-theoretic fidelity with estimation stability, this optimal threshold framework addresses longstanding dilemmas in extreme value analysis. The explicit linkage between entropy variance and KL divergence (Proposition 1) provides justification for adaptive thresholding, while Example 2 demonstrates practical calibration to market conditions. This advances beyond Drees [
56]’s bias–variance trade-off by incorporating tail information dynamics and subsumes Thompson [
53]’s MLE approach as the
limit. For financial applications, the protocol enables dynamic threshold adjustment aligned with both market microstructure and regulatory constraints—a requirement for crisis prevention.
3.2. Entropic Hypothesis Testing for Tail Regimes
Modern financial surveillance requires hypothesis tests that detect regime shifts in tail behavior with both statistical power and economic interpretability. Traditional moment-based approaches [
13,
40] lack sensitivity to early-stage bubbles due to their reliance on variance explosions rather than information structure changes. While subsampling methods [
57] improve detection latency, they fail in the non-stationary environments common to financial crises [
2]. Recent advances in information-theoretic risk measures [
35,
58] motivate entropy-based testing, but require asymptotic justification for their regulatory adoption.
This section establishes an entropy-based hypothesis testing framework with three key advances: (a) asymptotic normality of tail entropy estimators under non-explosive regimes, (b) explicit influence function calculations for sensitivity analysis, and (c) composite hypothesis testing via the continuous mapping theorem. The proposed test generalizes moment-based approaches [
13] while avoiding subsampling artifacts [
57]. I derive closed-form variance expressions under exponential and Pareto innovations, enabling exact power calculations. Practical applications span Basel III stress testing, cryptocurrency surveillance [
59], and housing market risk monitoring [
60].
Theorem 2. Let follow with . For threshold satisfying , define with light-tailed (exponential decay) and with regularly varying (heavy tails). Under , the normalized entropy ratiowhere for . Reject at level α if . Proof. Under
, Theorem 3.1 of Embrechts et al. [
26] implies
. For exponential
,
The empirical entropy
uses a kernel density estimator
with bandwidth
. Define the entropy functional
According to Van der Vaart [
61]’s functional delta method, the influence function is
Under
,
converges to a normal distribution with variance
For exponential tails,
where
. Slutsky’s theorem yields the asymptotic normality. □
Example 3. For with , Theorem 1 obtainsThe test statistic becomesyielding consistent detection since slower than . Theorem 2 addresses three limitations of existing bubble tests. The entropy ratio diverges under (Example 3) unlike variance-based statistics. Asymptotic normality under enables exact critical values. Rejection thresholds map to entropy units (nats/bits).
Remark 2. The diverging threshold must satisfyto ensure consistency. Implement via the algorithm developed in Drees [56]. The entropic test supports (1) Basel III compliance by mapping
exceedances to capital surcharges, (2) the detection of stablecoin collapses via hourly
monitoring, and (3) the identification of regional bubbles through price entropy divergence. This enables regulators to phase out ad hoc volatility circuit breakers in favor of entropy-based triggers aligned with systemic risk accumulation [
2]. By establishing entropic hypothesis testing on asymptotic foundations, this section provides regulators with a tool for real-time bubble detection. The test’s dual interpretability—statistical significance and information-theoretic units—bridges the gap between econometric theory and prudent policy.
4. Information Dynamics
Recent advances in econophysics have revealed profound connections between information entropy and market instability, building on foundational work in nonequilibrium thermodynamics [
62,
63]. Traditional market efficiency theories [
64] establish price discovery mechanisms but lack tools to quantify information transfer during extreme events. The emergence of stochastic thermodynamics [
65] provided entropy production as a key measure of system irreversibility, later adapted to financial markets through thermodynamic analogies [
66]. Despite progress, gaps remain in modeling non-stationary bubble regimes. Current approaches either rely on heuristic thresholds [
67] or lack statistical rigor [
68]. This section addresses these limitations through exact entropy production laws with time-varying thresholds and measure-theoretic phase transition detection with Benjamini–Hochberg error control.
4.1. Entropic Information Flow
I first establish exact entropy production laws for non-stationary financial processes and quantify information dissipation through KL divergence-regularized Brownian dynamics. A result also provides closed-form solutions for geometric Brownian motion benchmark cases. These results generalize the thermodynamic inequalities to adaptive threshold regimes while encompassing the entropy evolution laws.
Theorem 3. For Itô process with , the entropy production rate satisfieswhere is the Kyle-Lambda [69] and . Proof. From the threshold-adapted Fokker–Planck equation,
Compute entropy production via Nelson’s derivative [
70],
Substituting
and applying Itô’s lemma to
-dynamics yields the three-component decomposition. The KL divergence emerges from the relative entropy rate, while threshold adaptation follows from the Leibniz rule. □
Example 4. Consider high-frequency trading (HFT) strategies with and . Theorem 3 shows the following:Monitoring detects manipulative order flows through entropy dissipation anomalies [71]. Theorem 3 establishes entropy production as a fundamental observable for analyzing market stability, synthesizing insights from stochastic thermodynamics [
62,
72] and information-based asset pricing [
69,
73]. The tripartite decomposition of entropy dynamics—information dissipation through KL divergence, exogenous shocks via Brownian drivers, and adaptive threshold effects—provides a thermodynamic lens for quantifying financial fragility. By linking the Kyle–Lambda (
) to entropy dissipation rates, I formalize how informational efficiency governs market disorder during crises, addressing the microstructure indeterminacy in Fama [
64]’s efficiency paradigm. Example 4 demonstrates this framework’s practical power: the entropy production signature of high-frequency strategies reveals manipulative order flows through anomalous dissipation rates
, empirically validating Cont [
71]’s latency arbitrage models. For regulators, monitoring
enables the real-time detection of predatory trading patterns that elude variance-based surveillance [
8,
40], while adaptive thresholds
implement dynamic risk measures consistent with recent robustness frameworks [
74]. These results shift crisis prevention from static volatility monitoring to entropy flow regulation—a paradigm where markets stabilize by controlling information dissipation rather than merely constraining price swings. Future extensions could integrate this entropy production metric into central bank stress-testing frameworks, creating early warning systems that preempt systemic collapses triggered by informational cascades [
75].
4.2. Phase Transition Detection
Detecting transitions in financial markets presents a fundamental challenge in systemic risk analysis, necessitating methods that are sensitive to nonlinear regime shifts. Catastrophe theory, first introduced to financial markets by Zeeman [
76], offers mathematical foundations for modeling discontinuous transitions, while modern entropy methods provide mechanisms to quantify market disorder.
The recent literature developed heuristic bubble indicators using machine learning [
67], but their black-box nature limits their regulatory adoption. Concurrently, the literature proposes physics-inspired early warning systems [
68], although these lack statistical error control. The results in this section bridge these approaches through entropy-based phase transition theory, extending the thermodynamic framework to adaptive financial regimes. This section (1) develops an entropy-based phase transition indicator with exact critical values, (2) establishes links between catastrophe theory and multiple testing procedures, and (3) provides the measure-theoretic foundation for bubble collapse detection.
Theorem 4. Let follow an explosive process with tail entropy and market efficiency parameter γ from Theorem 3. Define the Entropic Bubble Index (EBI) aswhere ψ is the digamma function. A systemic phase transition occurs whenCritical values satisfywhere is the Benjamini–Hochberg false discovery rate and γ is the Kyle–Lambda from Theorem 3. Proof. Construct the entropy potential
, where
is the autoregressive parameter from Theorem 1. Critical points satisfy
The Hessian condition for saddle-node bifurcation
Taylor expands
near
,
Set
for False Discovery Rate (FDR) control [
77], yielding
. For convexity,
□
Theorem 4 establishes the first entropy-based phase transition detection framework with exact critical values, addressing the heuristic calibration issues in Zhao and Sornette [
67]. The detection threshold
directly embeds Benjamini–Hochberg false discovery rate control [
77], while the convexity requirement
ties market efficiency (
) to crisis predictability through the Kyle–Lambda [
69].
This framework addresses three systemic challenges: (1) replacing arbitrary volatility triggers with entropy-based critical values, (2) formalizing Zeeman [
76]’ catastrophe theory through measure-theoretic bifurcation analysis, and (3) addressing Astill et al. [
57]’s critique via
-controlled detection latency. Practically, Federal Reserve systemic risk dashboards can integrate EBI thresholds for real-time monitoring. Furthermore, circuit breaker calibration transitions from fixed percentage rules to
-
phase diagrams dynamically adjusted via market efficiency parameter
.
In summary, this section establishes a framework for analyzing information dynamics in non-stationary financial markets using entropy methods. Theorem 3 quantifies entropy production as the sum of three components: information dissipation through KL divergence, exogenous shocks via Brownian drivers, and adaptive threshold effects. This decomposition generalizes stochastic thermodynamics to time-varying regimes while incorporating traditional market efficiency metrics. Theorem 4 further connects catastrophe theory and multiple testing procedures by deriving exact critical values for phase transition detection with Benjamini–Hochberg error control. The Entropic Bubble Index transforms heuristic indicators into statistically robust early warning signals. This framework advances several research strands. First, it extends thermodynamic inequalities to non-equilibrium financial systems via Theorem 3. Second, it formalizes entropy-based bubble detection through the critical values of Theorem 4 and quantifies microstructure fragility through HFT entropy signatures. Lastly, it incorporates adaptive thresholds into risk management.
5. Application in Financial Markets
This empirical study implements the proposed framework across major asset classes. While the existing literature establishes entropy’s theoretical utility in extreme value analysis [
27,
37], few studies operationalize these concepts for real-time market surveillance across heterogeneous instruments. This section addresses this gap by evaluating the TWE framework’s capacity to detect speculative bubbles and systemic risks through (1) an adaptive threshold optimization that jointly minimizes KL divergence and entropy variance, (2) entropy ratio statistics with exact asymptotic distributions, and (3) phase transition detection via thermodynamic critical values.
By applying these tools to equity, commodity, and cryptocurrency markets, I quantify how tail entropy decay mechanisms govern crisis formation. The results demonstrate that entropy-driven signals provide statistically and economically significant early warnings compared to traditional variance-ratio tests, while robustness checks validate the framework’s stability across threshold selection methods and finite-sample regimes. This empirical validation confirms the theoretical entropy scaling laws from
Section 3 and provides practitioners with a replicable protocol for integrating information-theoretic risk measures into prudential oversight systems [
35,
78].
This empirical study uses daily data on the E-mini S&P 500, gold, crude oil (WTI), and Bitcoin futures over the period from 1 January 2015, to 24 March 2025, with . The data are obtained as continuous front-month contracts, and Bitcoin spot prices serve as a proxy for cryptocurrency market behavior. All data are sourced from Bloomberg. Daily log returns are computed using the continuously compounded transformation to ensure time additivity and statistical tractability. Days with missing data across any series are excluded from the analysis to maintain a balanced panel.
Table 1 presents summary statistics for the daily log returns. For each asset, annualized mean returns are calculated by multiplying the average daily return by 252. Annualized standard deviations are derived from the daily standard deviation multiplied by the square root of 252. The annualized Sharpe ratio is defined as the ratio of the annualized mean to the annualized standard deviation. The S&P 500 exhibits the highest Sharpe ratio, indicating the most favorable risk-adjusted performance among the assets. Bitcoin and oil demonstrate the highest volatility, with extreme minimum and maximum returns. Bitcoin also shows strong negative skewness and excess kurtosis, which are consistent with its heavy-tailed return distribution. In contrast, gold returns are relatively stable, with lower volatility and skewness. These distributional features highlight the heterogeneity in risk–return profiles across traditional and alternative assets.
Then, the daily log returns are preprocessed to remove volatility clustering and potential seasonality. First, a GARCH(1,1) model is applied to each return series individually. The model is specified as
where
captures conditional heteroskedasticity and
. The standardized residuals
are used from the model as filtered returns to ensure homoskedasticity. This is important for inference in models sensitive to variance dynamics. To account for potential intra-year seasonality in volatility and return magnitudes, the filtered returns are standardized using monthly z-score normalization
where
and
are the mean and standard deviation of returns within each month
m. The resulting series is used in subsequent analyses to ensure comparability across assets and time.
5.1. Tail Entropy Dynamics
Table 2 reports the optimal thresholds
, tail-weighted entropy estimates
, and generalized Pareto shape parameters
for each asset in the preprocessed return series. Entropy is computed over exceedances
, where
is selected via the regularized optimization procedure described in Proposition 1.
quantifies the disorder among tail events, and its decay rate reflects the informational dynamics governing extremes.
The results suggest distinct entropy behaviors across assets. Gold exhibits the highest positive entropy (), suggesting relatively uniform tail dispersion. In contrast, the S&P 500 and Bitcoin display strongly negative entropy values, indicating sharper tail concentration and greater predictability in extreme episodes. Despite their different volatility regimes, all assets exhibit negative shape parameters consistent with bounded tails and subexponential decay. These findings support the entropy decay implications of Theorem 1 and underscore the heterogeneity in tail informativeness across asset classes.
5.2. Bubble Detection Analysis
The detection process is formalized through Algorithm 1. This integrates (i) adaptive threshold optimization (Proposition 1), (ii) entropy ratio monitoring (Theorem 2), and (iii) phase transition criteria (Theorem 4) into a computationally replicable workflow. The algorithm decouples model specification (Step 1) from real-time monitoring (Step 2) and regime classification (Step 3).
Algorithm 1 Entropic bubble detection framework |
Input: Return series , regulation intensity , FDR level |
Output: Bubble signals , Entropic Bubble Index - 1.
Threshold Optimization (Proposition 1):
- 1:
- 2:
Estimate via Probability Weighted Moments [ 79].
- 2.
Entropy Monitoring (Theorem 2):
- 1:
for each window do - 2:
- 3:
- 4:
end for
- 3.
Phase Transition Detection (Theorem 4):
- 1:
- 2:
- 3:
Return .
|
The entropic detection framework was implemented on the four major financial assets. The three-stage methodology operationalizes the theoretical results.
Table 3 reveals consistent early warnings from the entropic test versus the Phillips et al. [
13] (PSY) benchmark. S&P 500 detected the COVID crash days earlier (19 February vs. 23 March 2020), with stronger signal (
vs.
). Oil’s March 2020 collapse and gold’s Ukraine crisis response were detected days faster. Bitcoin flagged the 2021 speculative peak days prior (20 October vs. 10 November), showing extreme ER (
) from heavy-tailed entropy (
).
The superiority of the entropic framework arises from the entropy decay mechanism of Theorem 1. For Bitcoin (
), we observe
compared to PSY’s variance-based detection. Bubble formation is demonstrated through dual criteria: EBI crossing
(99% quantile) and entropy convexity
. During Bitcoin’s 2021 bubble, EBI breached
on October 20 when prices reached USD 64,382, while the PSY test triggered only after the November 10 peak (USD 68,982). The entropic signal provided a 21-day lead time with rising convexity (
), enabling preventive action before the 48.6% collapse to USD 35,104 by January 2022. The proposed framework’s average 23-day early warning capability across assets (vs. 9 days for PSY) translates to critical risk management advantages. For a USD 10B portfolio, USD 1.4B with PSY. Basel III countercyclical buffers could be activated before momentum traps form. High-frequency surveillance systems can incorporate entropy decay rates for circuit-breakers.
This empirical study confirms that tail entropy dynamics provide an informationally efficient channel for bubble detection, overcoming the latency and distributional assumptions of moment-based approaches. The methodology’s foundations in extreme value theory make it suited for modern heavy-tailed markets.
5.3. Robustness Checks
To validate the reliability of the entropic bubble detection framework, three critical robustness analyses were conducted: (i) alternative threshold selection methodologies, (ii) finite-sample properties through subsampling, and (iii) comparison with traditional moment-based approaches.
The regularized entropy threshold (Proposition 1) was compared against two established methods: MLE thresholding and the Hill estimator. MLE Thresholding maximizes the GPD likelihood [
51]
while the Hill estimator focuses on tail index stability [
80]
Table 4 compares the proposed entropy-regularized threshold selection against traditional methods across four key metrics. The Average Detection Lag measures the temporal delay in identifying bubble onsets, where the entropy method shows 34% faster detection than MLE approaches. This acceleration stems from Proposition 1’s dual optimization of KL divergence and entropy variance, enabling the earlier recognition of tail distribution shifts. The Threshold Stability Ratio quantifies threshold consistency over time, calculated as
. The 0.87 ratio versus 0.52 for the Hill estimator demonstrates superior robustness to market noise. This stability directly results from the entropy variance regularization term
in our loss function.
KL divergence values confirm the entropy method’s distributional fidelity, with 33% lower divergence than MLE thresholds (, paired t-test). This aligns with Theorem 2’s requirement for accurate GPD approximation in entropy ratio calculation. The Variance Ratio Test p-values further validate our thresholds’ homoskedastic properties (), satisfying the i.i.d. assumptions underlying extreme value theory.
The proposed method demonstrates superior performance in threshold stability (, Wilcoxon test) while preserving distributional fidelity, especially for heavy-tailed assets like Bitcoin, where Hill estimator thresholds differ by compared to .
Through block bootstrap subsampling (
B = 10,000 replications), we can assess
where
C depends on the tail entropy properties from Theorem 2. The results show that (a) test size remains below
for
, (b) power exceeds 90% for
across assets, and (c) entropic test dominates PSY in subsample agreement (
vs.
).
To evaluate operational effectiveness in crisis prevention,
Table 5 benchmarks the proposed framework against classical approaches: the Variance Ratio Test [
40], which detects heteroskedasticity bursts, and Tail Index Monitoring [
80], which tracks changes in the shape parameter. The Average Lead Time advantage of 23.1 days over PSY’s 8.9 days translates into critical risk management windows. For a USD 10B portfolio, this represents USD 2.1B compared to
$0.7B in preventable losses based on the dynamics of the 2021 Bitcoin crash. Crisis Recall rates reflect each method’s ability to detect true bubble events. The 92% recall rate versus 71% for PSY results from the sensitivity of entropy to tail information decay (Theorem 1), in contrast to variance-based methods that overlook early log-scale entropy changes. The differential in False Positive Rates (2.1% vs. 6.2% for PSY) stems from the method’s dual thresholding: statistical significance via
and economic significance through phase transition criteria (
). This differs from moment-based methods that tend to overfit volatility clusters instead of identifying genuine bubble formations.
The proposed framework achieves a 2.5× higher early warning efficiency (
, McNemar test) while maintaining low false positives. The entropy advantage stems from simultaneously capturing distributional shape (through
dynamics), information decay (via
-dependence), and regime transitions (using KL divergence). These robustness checks confirm the proposed method’s superiority in modern markets characterized by heavy tails and volatile extremes [
41]. The entropy framework’s mathematical coherence (Lemma 1) and adaptive thresholding provide unique advantages over both moment-based and alternative threshold selection approaches.
6. Discussion and Implications
For financial practitioners, the results offer three valuable insights. First, the entropy decay law provides a real-time diagnostic for tail risk accumulation, enabling portfolio managers to adjust hedging strategies before volatility spikes. The early warning window during Bitcoin’s 2021 crash could prevent significant losses on a portfolio. Second, the entropic phase transition criteria allow central banks to activate countercyclical buffers preemptively, addressing Basel III’s mandate for dynamic risk management. Third, the convex optimization framework for threshold selection addresses backtesting overfitting in expected shortfall models, enhancing Basel III compliance.
The findings challenge three conventions in financial econometrics. First, variance-based bubble tests are suboptimal in heavy-tailed markets—entropy’s sensitivity to distributional shape accounts for its higher crisis recall. Second, regulatory stress tests can incorporate entropy production metrics to quantify information dissipation during crises, advancing beyond static scenario analysis. Third, high-frequency trading surveillance can utilize the entropy–Brownian noise decomposition to distinguish speculative exuberance from fundamental shocks. The TWE measure bridges two recent research subjects: information-theoretic risk measures [
1] and non-stationary EVT [
41]. While Zhang et al. [
58] develop coherent entropy measures, the GPD concentration bounds developed in this paper address extremal compatibility. The entropy decay rate extends Phillips et al. [
13]’s moment explosion criteria to information dynamics, formalizing Billio et al. [
30]’s intuition about entropy as a crisis precursor.
Methodologically, the phase transition detection enhances Zhao and Sornette [
67]’s machine learning approach by providing exact critical values. The threshold optimization framework incorporates Zhang et al. [
52]’s data-driven method through entropy regularization, reducing overfitting in quantile estimation. These contributions align with four emerging trends: (1) thermodynamic finance [
66], where the entropy production laws quantify market irreversibility; (2) tail risk network analysis [
81], where TWE can measure systemic information contagion; (3) crypto-market surveillance [
59], where Bitcoin’s entropy dynamics reveal speculative cycles; and (4) climate–finance integration [
34], where the proposed GPD–entropy linkage enables compound risk assessment. Future research can extend TWE to multivariate extremes, exploring broader applications in statistical modeling and potentially unifying EVT and deep learning for high-dimensional risk management.