1. Introduction and Review of the Literature
In order to adequately apply methods of non-asymptotic expectation and allocation to fat-tailed-distributed outcomes, we must examine the history of all three of these facets. First, we will review the history of fat-tailed distributions in economic time series, leading to the proposal of the generalized hyperbolic distribution (GHD). Next, we will explore the history of mathematical expectation as well as optimal allocation methods, which we will apply to stable fat-tailed-distributed outcomes.
1.1. Fat-Tailed Distributions in Economic Time Series—History
While explicit mentions of fat-tailed distributions in the financial price context may have been limited prior to the cited references, observations and discussions regarding market fluctuation irregularities and extreme event occurrences existed in financial history.
Historical examples, such as the financial crises and market panics occurring throughout history, provide insights into understanding market dynamics and extreme event presence. Documented observations of events, such as the 17th century Tulip Mania and 18th century South Sea Bubble, offer glimpses into the irregular, non-normal behaviors exhibited in financial prices.
The field of speculative trading and risk management, with roots in ancient civilizations, involved assessing trading risks and managing uncertainties in commodity prices and exchange rates. While formal conceptualization of fat-tailed distributions may not have been explicitly articulated during these early periods, experiences and observations related to market irregularities and extreme price movements contributed to gradually understanding the complexities and non-standard behaviors present in financial markets and asset prices.
Exploring historical accounts of financial crises and the evolution of trading practices in ancient and medieval civilizations provides valuable insights into the early recognition of irregular market behaviors and consideration of extreme events in financial transactions and risk management.
One of the early explicit mentions of fat-tailed distributions in the financial context is found in the work of economist Alfred Marshall. His contributions to supply–demand theory and goods pricing included discussions on market price variability and economic outlier occurrences. While Marshall’s work focused on microeconomic principles, his insights into market fluctuation nature provided a preliminary understanding of heavy-tailed distribution’s potential presence in economic phenomena.
The field of actuarial science and insurance, predating modern finance theory, involved risk analysis and extreme event occurrence, laying foundations for understanding tail risk and accounting for rare but significant events in risk management.
While these earlier references did not explicitly focus on fat-tailed distributions in finance, they contributed to understanding variability and risk in economic/financial systems and provided insights leading to modern heavy-tailed distribution theories and finance applications.
Marshall’s 1890 “Principles of Economics” [
1] laid foundations for modern microeconomic theory and analysis. While focused on general economic principles, it included discussions on market dynamics, supply–demand, and price fluctuation factors.
Later, in his 1919 work “Industry and Trade” [
2], Marshall examined industrial organization dynamics and market functioning, providing insights into price variability and factors influencing economic fluctuations.
Marshall’s works provided insights into economic principles and market dynamics, laying groundwork for modern economic theory development. While not explicitly discussing finance-related fat-tailed distributions, they contributed to understanding market variability and economic analysis, providing context on risk and uncertainty ideas in economic systems.
Louis Bachelier’s 1900 thesis “Théorie de la spéculation” [
3] delved into mathematical stock price modeling and introduced random walks, playing a significant role in efficient market hypothesis development and understanding financial price movements. While not focused on fat-tailed distributions, Bachelier’s work is considered an early reference laying groundwork for understanding financial market stochastic processes.
The work of French mathematician Paul Lévy [
4,
5] made significant contributions to probability theory and stochastic processes, including those related to finance. Specifically, Lévy’s research on random variable addition and general stochastic processes greatly influenced modern probability theory development and understanding stochastic dynamics across fields. Lévy’s contributions to Brownian motion theory and introduction of Lévy processes provided groundwork for understanding asset price dynamics and modeling market fluctuations. Through foundational work on topics such as Brownian motion, random processes, and stable distributions, Lévy helped establish mathematical tools and probability frameworks later applied to quantitative finance and modeling.
While explicit mentions of finance-related fat-tailed distributions may have been limited up to this time, Bachelier and Lévy’s contributions provided important insights into the probabilistic nature of finance and asset price movement governing processes.
In Feller’s work [
6], we begin to find discussions on heavy-tailed distributions and their applications across fields. Feller’s work contributed to understanding heavy-tailed distribution.
Additionally, the notion of fat-tailed distributions in economics/finance was discussed by prominent statisticians and economists. Notably, Vilfredo Pareto [
7] introduced the Pareto distribution in the early 20th century, closely related to the stable Pareto distribution, though it may not have been directly applied to price data then.
The concept of stable Pareto distributions in the finance context was notably introduced by mathematician Benoit Mandelbrot in his influential fractal and finance modeling work. Mandelbrot extensively discussed heavy-tailed distribution’s presence and implications for finance modeling.
In particular, Mandelbrot [
8] discussed stable Pareto distributions as a potential model for certain financial data. This seminal paper laid foundations for his research on the financial market’s fractal nature and non-Gaussian model applications describing market dynamics.
Additionally, Mandelbrot [
9] expanded on these ideas, delving into stable distribution concepts and their significance in understanding market fluctuations and risk dynamics.
Differences between the Pareto distribution, originally described by Vilfredo Pareto, and the stable Pareto distribution, referred to by Benoit Mandelbrot in 1963, lie in their specific characteristics and properties.
The Pareto distribution, introduced by Vilfredo Pareto in the early 20th century, is a power-law probability distribution describing wealth and income distribution in society. It exhibits a heavy tail, indicating a small number of instances with significantly higher values than the majority. It follows a specific mathematical form, reflecting the “Pareto principle” or the “80-20 rule”, suggesting that roughly 80% of effects come from 20% of causes.
The stable Pareto distribution, referred to by Benoit Mandelbrot, represents an extension of the traditional Pareto distribution. Mandelbrot’s 1963 work marks a pivotal point in applying stable distributions, including stable Pareto, to financial data and market behavior modeling. It incorporates additional parameters, allowing greater flexibility in modeling extreme events and heavy-tailed phenomena across fields such as finance and economics. The stable Pareto distribution accounts for distribution shape and scale stability across scales, making it suitable for analyzing complex systems with long-term dependencies and extreme fluctuations.
While both distributions share heavy-tailed characteristics and implications for analyzing extremes and rarities, the stable Pareto distribution, as conceptualized by Mandelbrot, provided a more nuanced and adaptable framework for modeling complex phenomena, especially in finance and other dynamic systems where traditional distributions typically fell short in capturing real-world data intricacies.
1.1.1. Subsequent Historical Milestones in Heavy-Tailed Distributions Applied to Financial Time Series
With time, many other fat-tailed modeling techniques found their way into quantitative departments. We examine several historically influential contributions within the realm of heavy-tailed distributions in finance.
Reference [
10] made the area of non-Gaussian stable processes accessible to researchers, graduate students, and practitioners. This book presented the similarity between Gaussian and non-Gaussian stable multivariate distributions and introduced one-dimensional stable random variables. It also discussed the most basic sample path properties of stable processes, including sample boundedness and continuity.
The problem of parametrically specifying distributions suitable for asset-return models is reconsidered in [
11]. The authors described alternative distributions, showing how they could be estimated and applied to stock-index and exchange-rate data. The book also investigates the implications for options pricing, making it a valuable resource for understanding the application of stable Paretian models in finance.
Further along the line of essential study in heavy-tailed distributions in finance, Reference [
12] was a comprehensive resource that presented current research focusing on heavy-tailed distributions with specific applications in financial time series. The book covered methodological issues, including probabilistic, statistical, and econometric modeling under non-Gaussian assumptions. It also explored the applications of stable and other non-Gaussian models in finance and risk management, making it a valuable resource for finance and economics professors, researchers, and graduate students.
Finally, Reference [
13] provided an accessible introduction to stable distributions for financial modeling. This article highlighted the need for better models for financial returns, as the normal (or Gaussian) model did not capture the large fluctuations seen in real assets. The author introduced stable laws, a class of heavy-tailed probability distributions that could model large fluctuations and allow more general dependence structures, offering a more accurate model for financial returns, and even focused on pitfalls in parameter estimation of such forms.
1.1.2. The Generalized Hyperbolic Distribution (GHD)
The generalized hyperbolic distribution (GHD) was introduced in [
14], examining aeolian processes.
The GHD constitutes a flexible continuous probability law defined as the normal variance-mean mixture distribution, where the mixing distribution assumes the generalized inverse Gaussian form. Its probability density function can be expressed in terms of modified Bessel functions of the second kind, commonly denoted as BesselK functions in the literature. The specific functional form of the GHD density involves these BesselK functions along with model parameters governing aspects such as tail behavior, skewness, and kurtosis. While not possessing a simple closed-form analytical formula, the density function can be reliably evaluated through direct numerical calculation of the BesselK components for given parameter values. The mathematical tractability afforded by the ability to compute the GHD density and distribution function underpins its widespread use in applications spanning economics, finance, and the natural sciences.
Salient features of the GHD include closure under affine transformations and infinite divisibility. The latter property follows from its constructability as a normal variance-mean mixture based on the generalized inverse Gaussian law. As elucidated in [
15], infinite divisibility bears important connections to Lévy processes, whereby the distribution of a Lévy process at any temporal point manifests infinite divisibility. While many canonical infinitely divisible families (e.g., Poisson and Brownian motion) exhibit convolution-closure, whereby the process distribution remains within the same family at all times, Podgórski et al. showed the GHD does not universally possess this convolution-closure attribute.
Owing to its considerable generality, the GHD represents the overarching class for various pivotal distributions, including the Student’s t, Laplace, hyperbolic, normal-inverse Gaussian, and variance-gamma. Its semi-heavy tail properties, unlike the lighter tails of the normal distribution, enable modeling of far-field behaviors. Consequently, the GHD finds ubiquitous applications in economics, particularly in financial market modeling and risk management contexts, where its tail properties suit the modeling of asset returns and risk metrics, and by 1995, the GHD appeared in financial markets application in [
16]. They applied the hyperbolic subclass of the GHD to fit German financial data.
This work was later extended by Prause in 1999 [
17], who applied GHDs to model financial data on German stocks and American indices. Since then, the GHD has been widely used in finance and risk management to model a wide range of financial and economic data, including stock prices, exchange rates, and commodity prices. See [
18,
19,
20,
21,
22,
23,
24,
25] for applications of the GHD to economic and share price data.
The GHD has been shown to provide a more realistic description of asset returns than the other classical distributional models and has been used to estimate the risk of financial assets and to construct efficient portfolios in energy and stock markets.
The generalized hyperbolic distribution has emerged as an indispensable modeling tool in modern econometrics and mathematical finance due to its advantageous mathematical properties and empirical performance. Salient features underscoring its suitability for analyzing economic and asset price data include:
Substantial flexibility afforded by multiple shape parameters, enabling the GHD to accurately fit myriad empirically observed, non-normal behaviors in financial and economic data sets. Both leptokurtic and platykurtic distributions can be readily captured.
Mathematical tractability, with the probability density function, cumulative distribution function, and characteristic function expressible in closed analytical form. This facilitates rigorous mathematical analysis and inference.
Theoretical connections to fundamental economic concepts, such as utility maximization. Various special cases also share close relationships with other pivotal distributions, such as the variance-gamma distribution.
Empirical studies across disparate samples and time horizons consistently demonstrate a superior goodness-of-fit compared to normal and stable models when applied to asset returns, market indices, and other economic variables.
Ability to more accurately model tail risks and extreme events compared to normal models, enabling robust quantification of value-at-risk, expected shortfall, and other vital risk metrics.
Despite the lack of a simple analytical formula, the distribution function can be reliably evaluated through straightforward numerical methods, aiding practical implementation.
In summary, the mathematical tractability, theoretical foundations, empirical performance, and remarkable flexibility of the GHD render it effective for modeling the non-normal behaviors ubiquitous in economic and financial data sets.
Although we will be using the symmetric GHD for ease-of-use in demonstrating the techniques we will illuminate in the discussion herein (the techniques being applicable to any distributional form), and despite the GDH possessing numerous favorable characteristics, its flexibility is limited in comparison to other extensions of the normal distribution that provide improved adaptability.
1.1.3. Extensions of the Normal Distribution for Greater Flexibility (GHD)
The skew two-piece, skew-normal distribution is a generalization of the normal distribution with an additional shape parameter. This distribution can model both left-skewed (negatively skewed) and right-skewed (positively skewed) data, depending on the value of the skewness parameter [
26,
27,
28,
29,
30,
31,
32]. It becomes a normal distribution when the skewness is equal to one.
This is particularly useful in economic time series, where data often exhibit asymmetry and non-normality [
33]. The skew two-piece, skew-normal distribution becomes a normal distribution when the skewness is equal to one, making it a more general model that can capture a wider range of data patterns [
34]. Additionally, using a distribution that directly handles skewness can perform better than data transformation, which may have limitations, such as reduced information, difficulty in interpreting transformed variables, and no guarantee of joint normality [
33].
The skew two-piece, skew-normal distribution and the skew normal distribution are not the same, but they are related. Both distributions are generalizations of the normal distribution that can accommodate skewed data, but they do so in different ways.
The skew normal distribution is a type of distribution that introduces skewness into a normal distribution. It is characterized by an asymmetry in the probability distribution of a real-valued random variable about its mean [
35]. The skewness value can be positive, zero, negative, or undefined. For a unimodal distribution, negative skew commonly indicates that the tail is on the left side of the distribution, and positive skew indicates that the tail is on the right.
On the other hand, the skew two-piece, skew-normal distribution, also known as the Fernández-Steel Skew Normal distribution, is a more general model that introduces an additional shape parameter to the normal distribution. This distribution can model both left-skewed (negatively skewed) and right-skewed (positively skewed) data, depending on the value of the skewness parameter. It becomes a normal distribution when the skewness is equal to one. This distribution is particularly useful for modeling data that are concentrated in two directions in roughly equal proportions [
34].
In summary, while both distributions can model skewed data, the skew two-piece, skew-normal distribution provides a more flexible model that can capture a wider range of data patterns.
Overall, the skew two-piece, skew-normal distribution provides a more flexible and accurate model for capturing the specific characteristics of economic time series data [
36].
We now turn our attention to the odd log-logistic-type distributions.
The odd log-logistic normal distribution, the odd log-logistic geometric normal distribution, and the odd log-logistic skew normal distribution are all different, albeit related probability distributions.
The odd log-logistic normal distribution is a three-parameter model that is symmetric, platykurtic, leptokurtic, and may be unimodal or bimodal. It is a new distribution that has been studied for its various structural properties, including explicit expressions for the ordinary and incomplete moments, generating function, and mean deviations. It has been shown to provide more realistic fits than other special regression models in certain cases [
37].
The odd log-logistic geometric normal distribution is a new regression model based on the odd log-logistic geometric normal distribution for modeling asymmetric or bimodal data with support in ℝ. This distribution is designed to model data that are not symmetric and may exhibit bimodal characteristics [
38].
The odd log-logistic skew normal distribution, on the other hand, is a different distribution that combines the odd log-logistic and skew normal distributions. It is designed to model data that are not symmetric and may exhibit skewness characteristics. This distribution is particularly useful for modeling data that are concentrated in two directions in roughly equal proportions [
37].
In summary, the odd log-logistic normal distribution, the odd log-logistic geometric normal distribution, and the odd log-logistic skew normal distribution are all different distributions that have been developed to model specific characteristics of data, including symmetry, kurtosis, bimodality, and skewness. Each distribution is designed for different purposes and may be more or less suitable depending on the specific characteristics of the data being modeled.
1.2. Mathematical Expectation—History
The genesis of mathematical expectation lies in the mid-17th century study of the “problem of points”, which concerns the equitable division of stakes between players forced to conclude a game prematurely. This longstanding conundrum was posed to eminent French mathematician Blaise Pascal in 1654 by compatriot Chevalier de Méré, an amateur mathematician. Galvanized by the challenge, Pascal commenced correspondence with eminent jurist Pierre de Fermat, and the two deduced identical solutions rooted in the fundamental tenet that a future gain’s value must be proportional to its probability.
Pascal’s 1654 work [
39], compiled amidst his collaboration with Fermat, explores sundry mathematical concepts, including foundational probability theory and expectation calculation for games of chance. Through rigorous analysis of gambling scenarios, Pascal delved into mathematical quantification of expected values, establishing core principles that seeded probability theory’s growth and diverse applications. The “Pascal–Fermat correspondence” offers insights into the germination of mathematical expectation and its integral role in comprehending the outcomes of stochastic phenomena.
Contemporaneously in the mid-17th century, the Dutch mathematician, astronomer, and physicist Christiaan Huygens significantly advanced the field of probability and study of mathematical expectation. Building upon the contributions of Pascal and Fermat, Huygens’ 1657 work [
40] incisively examined numerous facets of probability theory, with particular emphasis on games of chance and the calculation of fair expectations in contexts of gambling. He conceived an innovative methodology for determining expected value, underscoring the importance of elucidating average outcomes and long-term behaviors of random processes.
Huygens’ mathematical expectation framework facilitated subsequent scholars in expanding on his ideas to construct a comprehensive system for disentangling stochastic phenomena, quantifying uncertainty, and illuminating the nature of randomness across diverse scientific realms.
Essentially, the calculation for mathematical expectation, or “expectation”, remained unchanged for centuries, until the work in [
41]. It was Vince who decided to look at a different means of calculating it to reflect an individual or group of individuals who are in an “existential” contest of finite length. Consider a young trader placed on an institution’s trading desk and given a relatively short time span to “prove himself” or be let go, or the resources used by a sports team in an “elimination-style” playoff contest, or such resources of a nation state involved in an existential war.
Such endeavors require a redefinition of the “expectation”—what the party “expects” to happen—and can readily be defined as that outcome, where half the outcomes are better or the same and half are worse or the same at the end of the specified period or number of trials.
Thus, the answer to this definition of “expectation” is no longer the classic, centuries-old definition of the probability weighted mean outcome, but instead is the mean sorted cumulative outcome over the specified number of trials.
For example, consider the prospect of a game where one might win 1 unit with a probability of 0.9, and lose −10 units with a probability of 0.1. The classical expectation is to lose −0.1 per trial.
Contrast this to the existential contest, which is only 1 trial long. Here, what one “expects” to happen is to win 1 unit. In fact, it can be shown that if one were to play and quit this game after 6 trials, one would “expect” to make 6 units (that is, half the outcomes of data distributed as such would show better or the same results, and half would show worse or the same results). After 7 trials, with these given parameters, however, the situation turns against the bettor.
Importantly, as demonstrated in [
41], this non-asymptotic expectation converges to the asymptotic, centuries-old “classic” expectation in the limit as the number of trials approaches infinity. Thus, the “classic” expectation can be viewed as the asymptotic manifestation of Vince’s more general non-asymptotic expression.
This redefined “expectation” then links to optimal resource allocations in existential contests. The non-asymptotic expectation provides a modern conceptualization of “expectation” tailored to modeling human behavior in high-stakes, finite-time scenarios and, in the limit as the number of trials becomes ever-large, converges on the classical definition of mathematical expectation.
1.3. Optimal Allocations—History
The concept of geometric mean maximization originates with Daniel Bernoulli, who made the first known reference to it in 1738. Prior to that time, there is no known mention in any language of even generalized optimal reinvestment strategies by merchants, traders, or in any of the developing parts of the earth. Evidently, no one formally codified the concept. If anyone contemplated it, they did not record it.
Bernoulli’s 1738 paper [
42] was originally published in Latin. A German translation appeared in 1896, and it was referenced in John Maynard Keynes’ work [
43]. In 1936, John Burr Williams’ paper [
44], pertaining to trading in cotton, posited that one should bet on a representative price. If profits and losses are reinvested, the method of calculating this price is to select the geometric mean of all possible prices.
Bernoulli’s 1738 paper was finally translated into English in
Econometrica in 1954. When game theory emerged in the 1950s, concepts were being widely examined by numerous economists, mathematicians, and academics. Against this fertile backdrop, John L. Kelly Jr. [
45] demonstrated that to achieve maximum wealth, a gambler should maximize the expected value of the logarithm of his capital. This is optimal because the logarithm is additive in repeated bets and satisfies the law of large numbers. In his paper, Kelly showed how Claude Shannon’s information theory [
46] could determine growth-optimal bet sizing for an informed gambler.
Maximizing the expected end wealth is known as the Kelly criterion. Whether Kelly knew it or not, the cognates to his paper are from Daniel Bernoulli. In all fairness, Bernoulli was likely not the originator either. Kelly’s paper presented this as a solution to a technological problem absent in Bernoulli’s day.
Kelly’s paper makes no reference to applying the methods to financial markets. The gambling community embraced the concepts but applying them to various capital market applications necessitated formulaic alterations, specifically scaling for the absolute value of the worst-case potential outcome, which become particularly important considering fat-tailed probabilities. In fairness, neither Kelly nor Shannon were professional market traders, and the work presented did not claim applicability to finance. However, in actual practice, scaling to worst-case outcomes is anything but trivial. It is this author’s contention that Kelly and Shannon (who signed off on it [
28]) would have caught this oversight (they would not have intentionally opted to make the application scope of this work knowingly less general) had the absolute value of worst-case not simply been unity, thus cloaking it. The necessary scaling was provided in [
47].
In subsequent decades after Reference [
45], many papers by numerous researchers expanded on geometric growth optimization strategies in capital market contexts, notably Bellman and Kalaba [
48], Breiman [
49], Latane [
50,
51], Tuttle [
51], Thorp [
52,
53], and others. Thorp, a colleague of Claude Shannon, developed a winning strategy for Blackjack using the Kelly criterion and presented closed-form formulas to determine the Kelly fraction [
52].
The idea of geometric mean maximization was also well-critiqued. Samuelson [
54,
55], Goldman [
56], Merton [
57], and others argued against universally accepting it as the investor criterion. Samuelson [
54] highlighted that the Kelly fraction is only optimal asymptotically, as the number of trials approaches infinity, not for finite trials, whereas, in fact, it would always be sub-optimal.
The formulation that yields the growth-optimal allocation to each component in a portfolio of multiple components for a finite number of trials is provided in [
41]. This approach incorporates the net present value of amounts wagered and any cash flows over the finite timespan, acknowledging that real-world outcomes typically do not manifest instantly.
Growth-optimality, however, is not always the desired criterion. The formulaic framework for determining it can be used to discern other “optimal” allocations. In Vince and Zhu’s work [
58], we find two fractional allocations less than the growth-optimal allocation for maximizing the various catalogued return-to-risk ratios. This was further expanded upon by de Prado, Vince, and Zhu [
59].
Finally, Reference [
60] provides the case of using the formula for geometric growth maximization as a “devil’s advocate” in contexts where the outcome from one period is a function of available resources from previous periods, and one wishes to diminish geometric growth. Examples include certain biological/epidemiological applications and “runaway” cost functions, such as national debts.
2. Characteristics of the Generalized Hyperbolic Distribution (GHD)—Distributional Form and Corresponding Parameters
As discussed, although the two-piece, skew-normal distributions and the odd log-logistic-type distributions offer greater flexibility than the GHD and, therefore, are generally considered to better represent financial time series data, we will be illustrating our methods on the symmetrical GHD for simplicity.
These methods can be applied to any distributional form. The focus herein is not on the specific distributional form employed to model financial time series, but on the techniques involved in discerning a sample set of probable outcomes based on the particular distribution of an existential contest of finite length, from which numerous time-dependent metrics can be determined.
In addition to having the characteristics of being closed under affine transformations and possessing an infinitely divisible property, which enables connections to Lévy processes, the GHD provides some practical implantation advantages to the study of economic data series. Among these, the GHD possesses:
A more general distribution class that includes Student’s t, Laplace, hyperbolic, normal-inverse Gaussian, and variance-gamma as special cases. Its mathematical tractability provides the ability to derive other distribution properties.
Semi-heavy tail behavior, which allows modeling data with extreme events and fat-tailed probabilities.
A density function, involving modified Bessel functions of the second kind (BesselK functions). Thus, although there is a lack of a closed-form density function, it can still be numerically evaluated in a straightforward manner.
We examine the derivation of the probability density function (PDF) of the symmetrical GHD, and then derive the first integral of the PDF with respect to the random variable to obtain the cumulative density function (CDF) of the symmetrical GHD.
A characteristic function is a mathematical tool that provides a way to completely describe the probability distribution of a random variable. It is defined as the Fourier transform of the PDF of the random variable and is particularly useful for dealing with linear functions of independent random variables. We, therefore, derive the PDF of the symmetrical GHD:
1. Starting with the characteristic function of the symmetrical GHD, as derived from [
61] as:
where:
α = concentration
δ = scale
μ = location
λ = shape
γ = skewness
2. Using the inversion formula, the PDF of the GHD can be obtained from the characteristic function:
n.b. In the characteristic function provided in the paper by Klebanov and Rachev (1996) [
15],
is a variable used to represent the argument of the characteristic function. It is not the same as
, which is a variable used to represent the value of the random variable being modeled by the GHD.
The characteristic function of a distribution is a complex-valued function that is defined for all values of its argument, . It is the Fourier transform of the probability density function (PDF) of the distribution, and it provides a way to calculate the moments of the distribution and other properties, such as its mean, variance, and skewness.
In the case of the GHD, the characteristic function is given by (1).
The argument
of the characteristic function is a complex number that can take on any value. It is not the same as the value of the random variable being modeled by the GHD, which is represented by the variable
. The relationship between the characteristic function and the PDF of the GHD is given by the Fourier inversion formula:
where
is the PDF of the GHD and
is the value of the random variable being modeled.
3. Substituting the expression for the characteristic function into the Fourier inversion formula:
4. Simplifying the expression inside the exponential:
5. Using the identity:
, to simplify the expression:
6. Substituting:
, to simplify the integral:
7. Finally, setting the skewness parameter
to zero to obtain the simplified PDF of the symmetric GHD:
The derivation of the CDF of the symmetrical GHD is as follows:
1. Integrating the PDF from negative infinity to
x to obtain the CDF:
2. Substituting the PDF into the integral and simplifying:
3. Substituting
and
to obtain:
4. With the identity:
, to obtain:
5. Substituting
and
to obtain:
6. Finally, simplifying the expression by noting that
and taking the limit as
to obtain:
where:
α = concentration
δ = scale
μ = location
λ = shape
= sign function, the sign of (x − µ), or simply .
Thus far, we have discussed the symmetrical GHD, where we are not accountable for skew (γ).
Implementing skew can become an unwieldy process in the GHD. Our focus is on implementation in this text, and the reader is reminded that although we present our work herein on the symmetrical GHD, it is applicable to all distributional forms, with the symmetrical GHD chosen so as to be current and relatively clear in terms of implementation of the techniques. It is the techniques that are important here—not so much the distributional form selected to demonstrate them. As such, the balance of the text will be using the symmetrical form of the GHD.
Parameter Description
The symmetric generalized hyperbolic distribution (GHD) is parameterized by four variables governing salient distributional properties:
The concentration parameter (α) modulates tail weight, with larger values engendering heavier tails and hence elevating the probability mass attributable to extreme deviations. α typically assumes values on the order of 0.1 to 10 for modeling economic phenomena.
The scale parameter (δ) controls the spread about the central tendency, with larger values yielding expanded variance and wider dispersion. Applied economic analysis often utilizes δ ranging from 0.1 to 10.
The location parameter (μ) shifts the distribution along the abscissa, with positive values effecting rightward translation. In economic contexts, μ commonly falls between −10 and 10.
The shape parameter (λ) influences peakedness and flatness, with larger values precipitating more acute modes and sharper central tendencies. λ on the order of 0.1 to 10 frequently appears in economic applications.
While these indicated ranges provide first approximations, the precise parametric tuning warrants meticulous examination of the data properties and phenomenon under investigation. Moreover, complex interdependencies between the four variables potentially exist, necessitating holistic assessment when calibrating the symmetric GHD for applied modeling tasks.
3. Parameter Estimation of the Symmetrical GHD
The task is now specified as follows. We have a set of empirical observations that represent the returns or percentage changes in an economic time series from one period to the next. This can be stock prices, economic data, the output of a trading technique (These are often expressed as d0/d1 − 1, where d0 represents the most recent data and d1 the data from the previous period. In the case of outputs of a trading approach, the divisor is problematic—the results must be expressed in log terms, in terms of a percentage change. Such results must be expressed, therefore, in terms of an amount put at risk to assume such results, which are generalized to the largest potential loss. The notion of “potential loss” is another Pandora’s box worthy of a separate discussion), or any other time series data that comport to an assumed probability density function whose parameters we wish to fit to this sample data.
In the context of finding the concentration, scale, location, and shape parameters that best fit the PDF(x) or the symmetrical GHD, we can use maximum likelihood estimation (MLE) to estimate these parameters.
Maximum likelihood estimation (MLE) stands as a widely used statistical technique for estimating probability distribution parameters based on observed data. This method involves determining parameter values that optimize the likelihood of observing the data within a given statistical model assumption. MLE can be viewed as a special instance of maximum a posteriori estimation (MAP) under the assumption of a uniform parameter prior distribution. Alternatively, it can be considered a variant of MAP that disregards the prior, resulting in an unregularized approach. The MLE estimator selects the parameter value that maximizes the probability (or probability density, in continuous cases) of observing the given data. When the parameter comprises multiple components, individual maximum likelihood estimators are defined for each component within the complete parameter’s MLE. MLE is recognized as a consistent estimator, meaning that under certain conditions and as the sample size increases, the estimates derived from MLE tend toward the true parameter values. Put differently, with larger sample sizes, the likelihood of obtaining accurate parameter estimates increases.
Mathematically, we can express MLE as follows:
1. Beginning with the likelihood function for the symmetrical GHD, which is the product of the PDF evaluated at each data point:
2. Taking the natural logarithm of the likelihood function to obtain the log-likelihood function:
3. Taking the partial derivative of the log-likelihood function with respect to each parameter and setting the result equal to zero:
4. Solving the resulting system of equations to obtain the maximum likelihood estimates of the parameters.
5. Checking the second-order conditions to ensure that the estimates are indeed maximum likelihood estimates.
6. Using the estimated parameters to obtain the best-fit symmetrical GHD.
In step 1, represents the vector of parameters to be estimated, which includes the concentration, scale, location, and shape parameters. In step 2, we take the natural logarithm of the likelihood function to simplify the calculations and avoid numerical issues. In step 3, we take the partial derivative of the log-likelihood function with respect to each parameter and set the result equal to zero to find the values of the parameters that maximize the likelihood function. In step 4, we solve the resulting system of equations to obtain the maximum likelihood estimates of the parameters. In step 5, we check the second-order conditions to ensure that the estimates are indeed maximum likelihood estimates. Finally, in step 6, we use the estimated parameters to obtain the best-fit symmetrical GHD.
We shall now apply the discussed method of fitting the symmetrical GHD parameters to a real-world, financial time series. We implemented the parameter fitting techniques to the determination of the values for the real-world data set of S&P500 Index percent daily price changes for the first six months of 2023.
By way of this example, we obtained a time-ordered vector of daily percentage changes in the S&P500 for the first six months of 2023. We proffer these results as follows in
Table 1, which we will use to fit our symmetrical GHD parameters to.
We performed the MLE best-fit analysis on our returns sample data. One of the benefits of the GHD, as mentioned earlier, is the flexibility afforded by multiple shape parameters. We do not necessarily have to fit per MLE or any other “best-fit” algorithm—we could fit by eye if we so desired, the GHD is that extensible. In our immediate case, we decided to fit but set the location parameter equal to the mean of our sample returns (0.0029048), and fit the other three parameters, arriving at a best-fit set of parameters in
Table 2.
This resulted in the following PDF and CDF, in
Figure 1a,b, respectively, of our sample data set.
For the shape of the PDF in the first 6 months of the 2023 daily price changes in the S&P500 index, the distribution is symmetrical solely because we used the symmetrical incarnation of the GHD. What has been fit here is the scale parameter and the shape parameter, reflected in
Figure 1 for both the PDF and CDF of our sample data set. In working with the GHD, it is often necessary to use the flexibility of the parameters, in addition to a best-fit algorithm such as MLE, to obtain a satisfactory fit, not only to the sample data, but also to what the implementor may be anticipating in the future. Best-fit routines such as MLE are, in actual practice, a starting point.
4. Discussion
This far, we have laid the structural foundation to how we will use a given probability distribution and its parameters, which are fitted to a sample set of data. The presented technique allows any distributional shape to be applied. We chose the symmetrical GHD for its ease-of-use. In actual practice, it will be necessary to account for skewness in the distribution of the sampled data (this has been intentionally avoided in this paper to maintain simplicity).
In actual implementation, we would use the full, non-symmetrical GHD. Deriving the PDF from the characteristic function of the GHD:
Starting with the characteristic function of the GHD:
where
is the location parameter,
is the concentration parameter,
is the scale parameter,
is the shape parameter,
is the skewness parameter, and
.
Using the inversion formula to obtain the PDF of the GHD:
Substituting the characteristic function into the inversion formula and simplifying:
1. Changing the variable
:
2. Using the identity:
, to simplify the integral:
This is the full PDF with the skewness parameter (γ) for the GHD.
The integral required to compute the CDF(x) of the GHD with skewness (γ) does not have a closed-form solution, so numerical methods must be used to compute the CDF(x). However, there are software packages, such as R 4.2.0, MATLAB R2021a 9.10, or SAS 9.4M8, that have built-in functions for computing the CDF(x) of the GHD. Once the CDF(x) is computed, the inverse CDF(x) can be obtained by solving the equation: for , where is a probability value between 0 and 1. This equation can be solved numerically using methods such as bisection, Newton–Raphson, or the secant method. Alternatively, software packages, such as R, MATLAB, or SAS, have built-in functions for computing the inverse CDF of the GHD.
Our focus is on finite sequences of outcomes from a stochastic process. Although in our discussion, we are making the assumption that these outcomes are independent and identically distributed (IID), we assume this for simplicity of argument only; absent this IID assumption, the probabilities need only be amended at each subsequent trial in the finite, N-length sequence of outcomes based on previous outcomes. We refer to these finite sequences as being “existential contests”, realizing they exist in the milieu of life itself, as part of a longer sequence of a “lifetime-long” sequence of stochastic outcomes, or in the milieu of many players confronted with a finite sequence of stochastic outcomes.
The gambler wandering through the casino does not see this sequence stop and start as it continues when he continues gambling.
These are “existential contests” because we are walling-off these principles of ultimate ergodicity. We are examining what the single, individual (in the context of one man, or one team, one nation state—all one “individual”) player whose existence hangs on the outcome of the N-length, finite sequence of outcomes—the milieu it resides in be damned. Examples of existential contests might include a young trader placed on an institution’s trading desk and given a relatively short time span to “prove himself” or be let go, or the resources used by a sports team in an “elimination-style” playoff contest, or such resources of a nation state involved in an existential war. Even our gambler wandering through a Las Vegas casino, placing the minimum sized bets, who, suddenly, decides to wager his life savings on numbers 1–24 at the roulette wheel before heading for the airport, is now confronted with an N = 1 existential contest.
The notion of the existential contest does not fly in the face of The Optional Stopping Theorem, which still applies in the limit as N gets ever-larger. This theorem proffers that, under certain conditions, the expected value of a martingale at a stopping time is equal to its initial value. A martingale is a stochastic process that satisfies a certain property related to conditional expectations. The Optional Stopping Theorem is an important result in probability theory and has many applications in finance, economics, and other fields. The theorem is based on the idea that, if a gambler plays a fair game and can quit whenever they like, then their fortune over time is a martingale, and the time at which they decide to quit (or go broke and are forced to quit) is a stopping time. The theorem says that the expected value of the gambler’s fortune at the stopping time is equal to their initial fortune. The Optional Stopping Theorem has many generalizations and extensions, and it is an active area of research in probability theory. The theorem becomes less applicable as N approaches 0 from the right.
Not all human endeavors are regarded with the same, isotropic importance by the individual as traditional analysis assumes—some are more important than others, and some are absolutely critical to the individual, deserving of analysis in their own, critical enclosure.
4.1. The Goal of Median-Sorted Outcome after N Trials—The Process of Determining the Representative “Expected” Set of Oucomes
Let us return to our gambler in Las Vegas, ready to head for the airport, who decides to wager his life savings—half of it on the roulette ball falling on the numbers 1–12, “First Twelve” on the board, and the other half on 13–24, “Second Twelve” on the board. The payoff is 2:1. If the ball falls in any slot numbered 1–24, he will be paid 2:1 on one half of his life savings, keeping the half of his life savings wagered, and losing the other half. In effect, he has risked his entire life savings to increase it by half should the ball fall 1–24.
Here, 24 of the 38 slots will produce this result for him, so there is 24/38, or approximately 63.2% chance, he will see a 50% increase in his life savings. On one, single play, what one expects is the median outcome, which in this case, given the 0.632 probability of winning, he “expects” to win (in a European casino, he would have a better probability, as there is only one green zero, and his probabilities, therefore, increase to 24/37, or approximately 64.9%).
This is a critical notion here—what one “expects” to occur is that outcome, over N trials, where half of the outcomes after N trials are better, and half are worse. This is the definition of “expectation” in an existential contest.
However, it is always the expectation, as demonstrated in [
41], that, as N increases, this definition of expectation converges on the classical one accepted since the seventeenth century, as the sum of the probability-weighted outcomes. This “classical” definition of expectation is the expectation in the limit as the number of trials increases. Over a finite sequence of trials, the expectation is the median-sorted outcome, which converges to the classic definition as the number of trials,
N, becomes ever greater. Thus, we can state, in all instances, the finite-sequence expectation is the actual expectation.
Examining this further, we find the divergence between the finite-sequence definition and the classical one is a function of skewness in the distribution of outcomes, whereas the finite-sequence definition (expectation as the median-sorted outcome) is always the correct “expectation” amount, regardless of N. The divergence between this result and the classical definition (the sum of the probability-weighted outcomes) is a function of skewness in the distribution of outcomes.
Most games of chance offer the player a probability < 0.5 of winning, with various payoffs depending on the player’s chances, even if the payoff is even money. However, in any game where the probability of profitability on a given play is <0.5, and the classical expectation is negative, so too will the finite-sequence expectation be negative. Convergence is rapidly attained and never in the player’s favor for any period of plays.
Earlier, we referred to a hypothetical proposition where one might win 1 unit with a probability of 0.9 and lose −10 units with a probability of 0.1. The classical expectation is to lose −0.1 per trial.
This is in contrast to the existential contest, which is only 1 trial long. Here, what one “expects” to happen is to win 1 unit. In fact, it can be shown that if one were to play and quit this game after 6 trials, one would “expect” to make 6 units (that is, half the outcomes of data distributed as such would show better or the same results, and half would show worse or the same results). After 7 trials, with these given parameters, however, the situation turns against the bettor.
We find that similar situations arise in capital markets. It is not at all uncommon for commodity traders to employ trend-following models. Such models suffer during periods of price congestion and generally profit handsomely when prices break out and enter a trend. It is not at all uncommon for such approaches to see only 15% to 30% winning trades and yet be profitable in the “long run.” Thus, an investor in such programs would be wise to have as large an “N” as possible. The early phase of being involved in such an endeavor is likely to be a losing one, yet something that would payoff well in the long run—as the finite period becomes ever longer, the finite expectation converges on the classical one.
As we are discussing finite-length outcomes—“existential contests”, we wish to:
Determine what these finite-length expectations are, given a probability distribution we believe the outcomes comport to, so as to base on our sample data and the fitted parameters of that distribution.
Determine a likely N-length sequence of outcomes for what we “expect”; that is, a likely set of N outcomes representative of the outcome stream that would see half of the outcome streams of N length be better than, and half less than, such a stream of expected outcomes. It is this “expected, representative stream” that we will use as input to other functions, such as determining growth-optimal allocations.
Regarding point #2, returning to our hypothetical game of winning one unit with p = 0.9, and losing 10 units otherwise, here, the classical expectation is to lose 0.1 per play, with the finite expectation a function of N, and positive for the first 6 plays (assuming IID).
If one were to apply, say, the Kelly criterion, which requires a positive classical expectation, one would wager nothing. However, this asymptotic assumption under the finite-period expectation is found to be precisely the opposite, which would have the player start out risking a “fraction” equal to 1.0.
4.2. Determining Median-Sorted Outcome after N Trials
4.2.1. Calculating the Inverse CDF
To determine the median-sorted outcome, we need to first construct the inverse CDF function. An inverse cumulative distribution function (inverse CDF) is a mathematical function that provides the value of the random variable for which a given probability has been reached. It is also referred to as the quantile function or the percent-point function. In simpler terms, the inverse CDF essentially allows you to determine the value at which a specified probability is attained in a given distribution.
To derive the inverse CDF of the symmetrical GHD, we first start with the CDF of the symmetrical GHD, which is given by:
where:
α = concentration
δ = scale
μ = location
λ = shape
= sign function, the sign of (x − μ), or simply .
Rearranging the equation to isolate the term inside the curly braces:
We raise both sides to the power of
:
Solving for
:
The inverse CDF function takes a probability value,
, as input and returns the corresponding value of
, such that:
Simplifying the expressions inside the parentheses:
Simplifying the expressions using the fact that
:
This is our inverse CDF for the symmetrical GHD.
4.2.2. Worst-Case Expected Outcome
Since we sought to construct a “representative” stream of outcomes, a stream of outcomes we would “expect” to be typical (half better, half worse) of what we would expect to happen over N trials, and given the critical nature of the “worst-case outcome” in [
41,
47,
58,
59,
60,
62] for which we wish to apply this expected stream of outcomes to, we thus wished to find that point on the distribution that we would expect be seen, typically, over N trials.
Using the inverse CDF, we can find this point, this “worst-case expected outcome,” as the value of the inverse CDF for the input:
We can use this to examine against the results of our representative stream of outcomes at the median-sorted outcome.
4.2.3. Calculating the Representative Stream of Outcomes from the Median-Sorted Outcome
We proceeded to discern the median-sorted outcome. We generated a large sequence of random variables (0…1) to process via our inverse CDF to yield the random outcomes from our distribution, which we took N at a time:
For each of N outcomes, we determined the corresponding probability by taking the value returned by the inverse CDF as input to the CDF. When using the inverse CDF function, the random number generated is not the probability of the value returned by the inverse CDF. Instead, the random number generated is used as the input to the inverse CDF function to obtain a value from the distribution. This value is a random variable that follows the distribution specified by the CDF. To determine the probability of this value, one would need to use the CDF function and evaluate it at the value returned by the inverse CDF function. The CDF function gives the probability that a random variable is less than or equal to a specific value. Therefore, if the CDF function is evaluated at the value returned by the inverse CDF function, the probability of that value will be obtained.
For our purposes, however, we took any of the probabilities that were >0.5 and took the absolute value of their difference to 0.5. We called this p`:
p` = p if p ≤ 0.5
p` = |1 − p| if p > 0.5
We took M sets of N outputs, corresponding to outputs drawn from our fitted distribution as well as the corresponding p` probabilities.
For each of these M sets of N outcome—p` combinations, we summed the outcomes, and took the product of all of the p` values.
We sorted these M sets based on their summed outcomes.
We then took the sum of these M product of p` values. We called this the “SumOfProbabilities”.
For each M set, we proceeded sequentially through them, where each M has a cumulative probability, specified as the cumulative probability of the previous M in the (outcome-sorted) sequence, plus the probability of that M divided by the SumOfProbabilities.
It is the final calculation, we sought that M whose final calculation was closest to 0.5. This is our median-sorted outcome, and the values for N that comprise it are our typical, median-sorted outcome, our expected, finite-sequence set of outcomes.
4.2.4. Example—Calculating the Representative Stream of Outcomes from the Median-Sorted Outcome
Here, we demonstrate the steps from
Section 4.2.3. We assumed a finite stream of 7 outcomes in length (since we used a distribution of daily S&P500 returns for the first half of 2023, we assume we are modeling an existential contest of 7 upcoming days in the S&P500). Thus, N = 7.
We calculated 1000 such sets of these 7 outcomes and, therefore, M = 1000, and we proceeded to create, for each M, 7 random values (0…1), representing the inputs to the inverse CDF, and the output for that N, then deriving the probability by taking the output for that N as the input to the CDF. This final probability number was processed into a p` value.
For each set of 7 probability–outcome pairs, we took the product of the
p` values and the sum of the outcomes (
Table 3).
We next calculated the cumulative sum of the probabilities for each M. This is simply the previous M in the sorted sequence, plus the p` product of the N p` valuess in the outcome—p` pairs, divided by the sum of such p` values over all M (“SumOfProbs”).
Lastly, as shown in
Table 4, we found the M in the sorted set of M sets that corresponds to a cumulative sum of probabilities value closest to 0.5. This is the median-sorted outcome. Its output values (o1…o7) represent the “typical” expected outcomes over N trials (
Table 4).
In this instance of 1000 samples of sequences of N outcomes per the parameters for our symmetrical GHD representation of S&P500 closing percentage changes, we found the sample set, the “typical” set of seven closing prices—what we would expect, where half the time we would see a better sequence, and half the time a worse one, as:
Of note, our expected worst-case outcome was calculated as −0.00642. As we obtained an outcome that is at least as negative as this, we are not deluding ourselves on what might be the worst-case outcome for the “typical”, i.e., the median outcome stream of 7 outputs.
This set of outcomes represents the expected—the typical set of outcomes from data distributed in the first half of 2023 for daily percentage changes in the S&P500.
Further, for 7 trials in this distribution of outcomes, we can expect to see the ∑o outcome of 0.016764 as our non-asymptotic expectation. Of note, we used a symmetrical distribution in this paper and example, and as mentioned earlier, the divergence between the actual, finite expectation and the classical, asymptotic expectation is a function of the skewness of the distribution, which was 0 in our example. Nevertheless, the concept remains, and the procedure illuminated this for more perverse and skewed distributions.
4.3. Optimal Allocations
4.3.1. Growth-Optimal Fraction
Earlier, mention was made of determining the growth-optimal fraction from a sample data set, and how such results might diverge from its non-asymptotic determination.
Individuals and individual entities enmeshed in existential conflicts may very well be interested in finite-length growth optimization.
Consider, for example, a proposition with p (0.5) of winning 2 units versus losing 1 unit. The traditional method (even by way of the Kelly criterion, as the absolute value of the largest loss here is 1 unit) would determine asymptotic growth maximized by risking 0.25 of the stake on each play. However, when examined under the lens of finite-length existential contests, terminal growth is maximized if the length of the contest is 1 play at a fraction of 1.0, and if N = 2 plays, at a fraction of 0.5 per play, ad infinitum, to the asymptotic 0.25 risk of stake as N grows ever larger.
In determining optimal fractions of finite-length existential contests, the input data (the values of −0.006670152, −0.00638097, −0.006615831, −0.005151115, 0.013180558, 0.01693477, and 0.011466985 in our example, above) are what ought to be used for determining the optimal fraction. The data determine the expectation in such finite-length trials; that is, the data are germane to also determining these other facets—the data one “expects” over those N trials—and seeking to maximize them.
4.3.2. Return/Risk Optimal Points
The amount to risk that is growth-optimal is return/risk-optimal only asymptotically. For finite-length streams—existential contests—the authors of [
58] demonstrated the two points on the spectrum of percentage risk that maximize the ratio of return to risk both as a tangent to the function on this spectrum as well as the inflection point “left” of the peak of the function.
Later, in [
59], this was extended to the determination of multiple, simultaneous contests (portfolios).
Particularly in determining these critical points of return/risk maximization, which are length-dependent, utilizing the input data that would be “expected” over such length of trials, as demonstrated herein, is more appropriate to the individual engaged in an existential contest.
4.3.3. Risk of Ruin/Drawdown Calculations
Similarly, the determination of the risk of ruin, defined as drawdown on cumulative equity of a given percent of capital from an initial starting point, as well as the risk of drawdown, defined as drawdown on cumulative equity of a given percent of capital from a high watermark point, are more accurately calculated for the course of existential contests using a sample of outcomes determined by this method, as opposed to the universe of potential outcomes, as dictated by the full distribution of outcomes. In [
63], it is demonstrated that these calculations depend on a specific-length horizon (and, in fact, demonstrate that risk of drawdown = 1.0, certainty, for any defined percentage for drawdown as the length of time approaches infinity). As these calculations are also time-specific, the accurate calculation requires the use of a realistic time-specific sample set, as provided herein.
4.3.4. Growth Diminishment in Existential Contests
In [
60], a technique was presented, whereby examining an isomorphic mapping of the fraction of risk to the cosine of the standard deviation in percentage inputs to their arithmetic average allows one to examine growth functions sans a “risk fraction”, and instead in terms of variance in outcomes. The presented technique examined geometric growth functions, where growth was a negative property. Such functions might be the cumulative debt of a nation, the growth of a pathogen in a population or within an individual organism, etc.
On the function plotted in terms of the fraction of a stake to risk versus growth, as there is a growth-optimal point, there is necessarily a point where the function begins decreasing and where growth becomes negative by risking too much (for example, repeatedly risking beyond 0.5 in the 2:−1 payoff game mentioned earlier with p (0.9), one is certain to go broke as they continue to risk 0.5 of their stake or more, even in this very favorable game).
The isomorphic mapping provided in this paper allows one to examine such traits of the growth function without regard to the notion of a percentage of stake to risk, but rather of tweaking the variance.
As most endeavors to break malevolent growth functions are inherently time-sensitive (i.e., of finite N—true “existential contests”), the notion of using the median-expected outcome stream over N trials is very germane to this exercise, as well.
5. Further Research
We have demonstrated how to fit an extensible, albeit symmetrical distribution to economic data series, and done so with techniques applicable to other distributional forms. As discussed, although the two-piece, skew-normal distributions and the odd log-logistic-type distributions offer greater flexibility than the GHD and, therefore, are generally considered to better represent financial time series data, we have illustrated our methods on the symmetrical GHD for simplicity, but these methods can be applied to any distributional form.
However, it is the very parameter of skewness—the differences between mean and median—that accounts for the difference between conclusions drawn for existential contests and the classical, asymptotic assessments. Generally speaking, the greater the skewness, the more likely these differences will be highlighted, and the more likely to see cases where, for example, the expectation may be positive for the existential contest, but negative asymptotically, and vice versa. Thus, this paper leaves wide open the ripened area of applying the techniques to data modeled per the two-piece, skew-normal distributions and the odd log-logistic-type distributions.
6. Conclusions
We were concerned with “individuals” (as one or a collection of individuals with a common goal) involved in finite-length sequences of stochastic outcomes—existential contests—where the distribution of such outcomes is characterized by fat tails.
Once fitted, we demonstrated the process to determine the “expected” terminal outcome over the finite-length sequence, and the “expected” or “typical” outcomes that characterize such. We have demonstrated how to obtain the expected set of discrete outcomes for the given form for a given length of an existential contest.
This is important in analyzing the expectation of finite sequences, which differ from the classical expectation of the sum of the probabilities times their outcomes, with the former being what is germane to the “individual” in an existential contest, whereby the former is always the correct expectation, one that approaches the classical one asymptotically.
Such “expectation,” what the “individual” can “expect”, as that which has half the outcomes better half worse, also yields a typical stream of outputs achieving this expected outcome.
It is this stream of outputs which, to the individual in the existential contest, are the important inputs, considering that they are what the individual “expects” for such finite-length calculations as growth-optional fractions, maximizing return to risk, calculating ruin or drawdown in the existential contest, or even for those whose charge is growth diminishment with a finite deadline, given that life itself, despite the ever-present asymptotes, is rife with existential contests.