1. Introduction
Risk measures play an essential role in both academic research and financial practice, as they provide a systematic way to assess the potential losses of a financial position. Their importance has been growing in the industry and academic research since the work of
Artzner et al. (
1999), which introduced an axiomatic framework for coherent risk measures. Subsequent studies, such as (
Föllmer and Schied 2011, chp. 4) or
Delbaen (
2000), have refined and extended these ideas.
In the monetary risk-measure framework, one models a financial position using a real-valued random variable
X on a probability space
. The number
is the discounted net worth of the position if scenario
occurs. A monetary risk measure,
, assigns a real number,
, to the outcome of a random variable. This real number represents the minimal amount of capital needed to make
X acceptable according to certain risk criteria. Desirable properties include monotonicity (increasing payoffs lowers the risk) and translation invariance (adding a sure amount of cash decreases the risk by the same amount), and when
is further assumed to be convex or coherent, it reflects the benefits of diversification. We refer to
Artzner et al. (
1999);
Delbaen (
2000);
Föllmer and Schied (
2011) for standard references on these terms.
A preferred way to price financial claims in financial mathematics (and hence risk measurement, when viewed from a market perspective) is to determine the price as the expectation of the discounted payoff under an equivalent (local) martingale measure. However, in general markets and for certain price processes, a (local) martingale measure need not exist. Instead, from the more general perspective of no-arbitrage (or no free lunch with vanishing risk, to be more precise), one can only ensure the existence of an equivalent
-martingale measure (E
MM). See, for instance,
Delbaen and Schachermayer (
1998);
Kallsen (
2004);
Sohns (
2025) for details on
-martingale arguments. In addition, in incomplete markets, there may be multiple such measures, and a natural question is how to select the “best” or “preferred” measure among the many.
One popular selection criterion in incomplete markets is to pick the measure that is “closest” to the real-world probability measure
in terms of the relative entropy (also known as the Kullback–Leibler divergence); see
Delbaen et al. (
2002);
Frittelli (
2000);
Fujiwara and Miyahara (
2003);
Miyahara (
2004). Minimizing relative entropy leads to the well-known minimal-entropy martingale measure, a construction that has proven valuable in option pricing and hedging and which is closely connected to maximizing the investor’s utility (see
Section 4).
Most of the literature focuses on the local martingale setting when studying minimal-entropy measures. The corresponding minimal-entropy
-martingale measure has not been studied, even though, in some markets, one must work with E
MMs. We close this gap by introducing and studying the minimal-entropy
-martingale measure:
where
denotes the set of E
MMs, and
is the relative entropy with respect to
. The associated minimal-entropy risk measure is then
This measure is an extension of the classical minimal-entropy martingale measure (since an ELMM is a special case of an EMM) but is strictly more general whenever no local martingale measure exists. Notably, while the minimal-entropy martingale measure has been studied for pricing and hedging, it has not been viewed or analyzed as a traditional risk measure in the classical sense. Nor has the -martingale version been examined at all. In this paper, we fill this gap by proving that the minimal-entropy -martingale measure leads to a coherent risk measure with desirable properties.
A different risk measure and one of the most popular ones is the entropic risk measure. Because of the similar name, one might suspect similar definitions of the entropic risk measure and the minimal-entropy methods. Nevertheless, the entropic risk measure is typically introduced via the exponential formula
which does not indicate any connection to entropy.
Its name stems from its robust representation,
showing that it penalizes deviations from
in proportion to the relative entropy. Although this measure is well understood as a convex, time-consistent risk measure, the label “entropic” might not be fully transparent when beginning from an exponential definition. Here, we show that one can equivalently start with a relative-entropy-based formulation, making the “entropic” nature more obvious. We then show that, with this alternative definition, one easily arrives at the same results and conclusions.
This measure is strictly convex (rather than linear) in the payoff, X. In contrast, the minimal-entropy risk measure only picks out a single measure—the one that eliminates arbitrage and is of the least relative entropy. As a result, it turns out to be coherent. Despite these structural differences, both measures share a fundamental entropy-based underpinning.
The foundations of entropic risk measures and minimal-entropy martingale measures (MEMM) date back over two decades, but the topic remains actively researched. Recent studies address both theory and applications.
Chong et al. (
2019) introduced a forward entropic risk measure based on ergodic backward-stochastic differential equations and studied its behavior at large maturities.
Pichler and Schlotter (
2020) analyzed entropy-based risk measures and, in particular, their convexity and dual representations.
Wang et al. (
2020) investigated the entropic measure transform and its applications in risk-sensitive decisions.
Dhaene et al. (
2015) studied MEMMs for markets that contain financial and actuarial risks.
Ishikawa and Robertson (
2020) examined optimal investment and pricing problems that include default risks through entropy-based techniques.
Doldi et al. (
2024) connected entropy martingale optimal transport to dualities in nonlinear pricing and hedging.
Y. Kabanov and Sonin (
2025) derived explicit characterizations of MEMMs for exponential Ornstein–Uhlenbeck volatility models.
McCloud (
2025) studied pricing in incomplete markets through an entropic risk measure that accounts for default risks. Finally,
Marthe et al. (
2025) used entropic risk measures for risk-sensitive planning within Markov decision processes.
The main contributions of this paper are as follows:
We introduce the minimal-entropy -martingale measure for general semimartingale models, which is new to the literature.
We prove that the induced minimal-entropy risk measure is coherent and extends the classical minimal-entropy martingale measure results.
We define the entropic risk measure via its robust representation and, therefore, provide an alternative approach that highlights its connection to entropy.
We demonstrate key properties including convexity, coherence, dynamic consistency, and optimal risk transfer for both measures, thereby revealing that minimal-entropy techniques are not only pricing tools but also valid risk measures in their own right.
We provide some estimates, comparing the two risk measures to their real-world expectations.
The paper is organized as follows.
Section 2 contains the precise definitions of both the minimal-entropy
-martingale measure (by minimizing relative entropy under the no-arbitrage condition) and the entropic risk measure (via a relative-entropy supremum), along with existence criteria. We establish their existence and compare their properties.
Section 3 focuses on the definition of monetary risk measures, convexity and coherence, proving that the minimal-entropy risk measure is coherent, while the entropic measure is convex. In
Section 4, we explore duality and highlight the deeper relationship between entropy-based valuations and risk measures and, in particular, show that our definition of the entropic risk measure is equivalent to the more common definition in the literature and we elaborate that the setup works for general processes and probability measures.
Section 5 provides dynamic versions, establishing time consistency. Finally,
Section 6 discusses optimal risk transfer and how each of these risk measures behaves in that context. In the
Appendix A, we repeat some well-known results that are used in this publication.
2. Definition and Existence
Let
be a probability space,
the space of bounded random variables, and
the set of probability measures on
. Furthermore, let
be a filtration with
, and let
S be a (potentially multi-dimensional) stochastic process adapted to
. We assume
is the discounted price process (for details on discounting, see
Sohns (
2023)).
In most models, at least one probability measure exists such that the discounted process,
, is a local martingale under this probability measure. However,
Delbaen and Schachermayer (
1998) showed that, in an arbitrage-free market (more precisely a market that satisfies no risk with vanishing risk), you can only assume that a probability measure exists, such that
is a
-martingale under this probability measure. Therefore, it makes sense to study
-martingales, equivalent
-martingale measures, and derived risk measures.
First, let us recall the following:
Definition 1. A one-dimensional semimartingale, S, is called a σ-martingale if there exists a sequence of predictable sets, , such that the following applies:
- (i)
for all n;
- (ii)
;
- (iii)
For each , the process is a uniformly integrable martingale.
Such a sequence, , is called a σ-localizing sequence. A d-dimensional semimartingale is called a σ-martingale if each of its components is a one-dimensional σ-martingale.
Definition 2. An equivalent σ-martingale measure (EσMM) is a probability measure with , such that is a σ-martingale under . The set of all EσMM is denoted as .
We also define a broader sets of absolutely continuous measures,
, under which
is a
-martingale:
Definition 3. The relative entropy of a probability measure, , with respect to is defined as The relative entropy provides a notion of distance between a probability measure, , and a reference probability measure, . Even though it can be interpreted as a distance, it is not a metric since neither the symmetry property nor the triangle inequality holds.
Beyond financial mathematics, relative entropy is a crucial concept in statistical physics and information theory, where it is referred to as the Kullback–Leibler divergence (
Shunsuke 1993). It also plays a fundamental role in large deviations theory, where it underpins results such as Sanov’s theorem (
Cover and Thomas 2012). A comprehensive summary of its applications can be found in
Cherny and Maslov (
2004).
Further illustrations of this definition are available in
Dacunha-Castelle and Duflo (
1986), including a demonstration of how relative entropy can be understood as a distance measure. In statistics, this interpretation is reinforced by results such as Stein’s Lemma, which can be found in (
Hesse 2003, Satz 10.4).
One of the central questions in mathematical finance is how to determine the fair price of a claim, X. If the payoff at time T can be replicated with trading in the underlying asset, S, then the initial cost of the replicating strategy equals the fair price; otherwise, an arbitrage opportunity would exist.
More concretely, suppose there is an admissible self-financing strategy,
, such that the stochastic integral
matches the terminal payoff
. Then the fair price is
. If there exists an equivalent probability measure,
, under which
S is a martingale, and if suitable integrability conditions hold, then
is also a martingale under
. In this case, the initial value equals the expected terminal value:
so the fair price is given via the expectation under
.
An equivalent true martingale measure may not always exist. In such cases, one often uses an equivalent local martingale measure. There are also models without any local martingale measure. However, if the market is free of arbitrage, the first fundamental theorem of asset pricing ensures the existence of at least one equivalent
-martingale measure (see
Delbaen and Schachermayer (
1998)), making the discounted price process a
-martingale.
In a complete market, this measure is unique, leading to a single fair price. In incomplete markets, there are multiple equivalent
-martingale measures. The range of fair values then lies between
Thus, the fair price must lie within this interval (
Sohns 2023). It is, therefore, reasonable to define the minimal-entropy risk measure as an optimization problem over all equivalent
-martingale measures.
Definition 4. - (a)
An equivalent σ-martingale measure is called the minimal-entropy σ-martingale measure if it minimizes the relative entropy (or Kullback–Leibler divergence) among all , i.e., The corresponding minimal-entropy risk measure is defined by - (b)
For and , the entropic risk is defined as
Remark 1. Equivalence of to ensures that the relative entropy remains finite since any singular part would make infinite on a positive- set. Also, if ignored events that deems possible, the entropic penalty term would lose its economic meaning by excluding potentially significant outcomes.
Remark 2. In the above definition of , the parameter can be viewed as a risk-aversion level or scaling factor, much like in exponential-utility frameworks where larger γ corresponds to greater risk aversion. Thus, can be seen as an “entropic” or “exponential-penalized” valuation: the higher the γ, the stronger the penalty on deviating from the reference measure .
Note that the existence of these risk measures is not immediately clear. In particular, the existence of a minimal-entropy
-martingale measure can still depend on the properties of
S. For instance, if the market is not arbitrage-free (e.g., the No Free Lunch With Vanishing Risk condition fails) and
S represents the price process of a tradable asset, then no equivalent
-martingale measure can exist. Even if the market does satisfy NFLVR, one may still require additional conditions on the price processes (e.g., local boundedness or boundedness from below) to ensure that there is some equivalent
-martingale measure carrying finite relative entropy. For further details, see
Delbaen and Schachermayer (
1994,
1998,
2006).
Example 1. One classic example of a complete market is given via the Black–Scholes setup Black and Scholes (1973). In this model, the asset price typically followsfor constants μ and , where W is a Brownian motion under the real-world probability. After discounting by the risk-free rate r, one finds a unique equivalent martingale measure, , that replaces μ with r in the drift term. The market is, therefore complete, and any bounded contingent claim can be perfectly replicated. Since there is exactly one equivalent martingale measure, pricing is straightforward: every claim has a unique fair value given by its discounted expectation under . In this case, the σ-martingale condition is also trivially satisfied by the unique risk-neutral measure; hence, minimizing any functional over the set of equivalent σ-martingale measures offers no ambiguity (there is only one element in that set). Consequently, the minimal-entropy measure (cf. Definition 4) coincides with the unique martingale measure, and the resulting minimal-entropy risk measure effectively matches the usual Black–Scholes pricing rule. By contrast, many models for asset prices are incomplete. As a specific example, consider the Heston stochastic volatility model (Heston 1993). There, one typically assumeswhere represents the instantaneous variance process, and are correlated Brownian motions under the real-world probability. Unlike Black–Scholes, this model does not admit a perfect hedge for arbitrary claims, so there can be multiple local (or σ-)martingale measures equivalent to the real-world probability. In other words, the market is incomplete, so one cannot replicate every payoff. Different choices of the market price of volatility risk and jump risk (if extended) can lead to families of risk-neutral measures, all of which ensure no arbitrage but yield different pricing implications for claims. In such an incomplete setting, the minimal-entropy σ-martingale measure (from Definition 4) provides a systematic way to single out a preferred measure among the many. By penalizing deviations from the real-world probability through the relative entropy functional, this approach selects the measure that is “closest” (in Kullback–Leibler divergence) to the original distribution. Hence, the resulting risk measure assigns a unique fair price (or risk assessment) to each claim. For more details, see Biagini et al. (2000); Boguslavskaya and Muravey (2016); Hull and White (1987); Pham (2001); Sircar and Zariphopoulou (2004); Sohns (2022); Wiggins (1987).
Unlike the minimal-entropy -martingale measure, the entropic risk always exists.
Theorem 1. The entropic risk always exists.
Proof. It suffices to show that is indeed a finite number.
Let
. It is straightforward to verify the following:
and
because
X is almost surely bounded.
Since
is increasing in
, we conclude for a fixed
:
□
While the minimal-entropy
-martingale measure is new in the literature, the notion of the minimal-entropy martingale measure has been extensively studied in the literature. The definition of the latter is analog to ours, but instead of equivalent
-martingale measures, one minimizes the entropy among all equivalent martingale measures (see, for example,
Grandits and Rheinländer (
2002);
Lee and Rheinländer (
2013);
Miyahara (
1999);
Rheinländer and Steiger (
2006);
Schweizer (
2010)). There are numerous criteria in the literature for the existence of the minimal-entropy martingale measure. Most of these criteria impose strong conditions on the stochastic process
. A prominent example is when
is a Lévy process (
Fujiwara 2004, Theorem 3.1). Another simple result is the existence of the minimal-entropy martingale measure if
is bounded as shown in (
Frittelli 2000, Theorem 2.1). Other results depend on the specific form of the Radon–Nikodym derivative (
Grandits and Rheinländer 2002) or are applicable only in discrete time settings (
Föllmer and Schied 2011, Corollary 3.27).
Things are a bit easier for the minimal-entropy -martingale measure. We now state and prove a theorem showing that, under a mild condition—namely, the existence of at least one -martingale measure with finite —one confirms the existence (and uniqueness) of a minimal-entropy -martingale measure. This holds even if is not locally bounded and even if no equivalent local martingale measures exist.
For the proof, we need some lemmas.
Lemma 1. Let be the set of probability measures on . We have the following:
- (a)
. Furthermore, if and only if .
- (b)
The mappingis convex and strictly convex on the set of probability measures that are absolutely continuous with respect to .
Proof. We have the following:
Hence, it is reasonable to look at the function
, which we define as a continuous extension of the function
to
. Because of
and Theorem A1, we have
for
, which proves the strict convexity of
.
It is easy to see that the function
takes its minimum at
. Hence, we have
, and thus,
- (a)
According to Equation (
2), we have, for
, the following:
Now, let
. We must show that the equality
follows. We have the following:
Hence, —almost everywhere, we have . Now, it follows that —almost surely. If is not absolutely continuous with respect to , the statement is obvious.
- (b)
The statement is obvious for probability measures that are not absolutely continuous with respect to
. So, we focus on the case where
. For
with
, we have the following:
Thus, strict convexity follows. □
Lemma 2. Suppose there exists a measure, , withThen, is the unique minimal-entropy σ-martingale measure. Proof. Define the set
By assumption,
is nonempty because
. Thus, we have
Next, we show that is convex and closed in total variation:
For any , their convex combination , , is also in . Indeed, convex combinations preserve both the finite entropy condition (due to convexity of relative entropy) and the -martingale property as -martingales form a vector space.
The set
can be expressed through linear constraints as
Sets defined via linear constraints of the form
are closed in total variation topology. Intersecting with the closed set of measures having finite entropy preserves this closedness.
By applying Csiszár’s Theorem A6, we conclude that the set
has a unique
I-projection
of
, which satisfies
By hypothesis,
already satisfies this minimality, so it must coincide with
. Thus, uniqueness follows directly from the strict convexity of relative entropy.
Finally, we verify that is equivalent to . Suppose this is not the case. Then, there exists a , which is equivalent to , as given by hypothesis. If is not equivalent to , its support is strictly smaller than that of , which contradicts minimality via Csiszár–Gibbs inequality (Theorem A7). Hence, .
Thus, , is equivalent to , and minimizes entropy. Therefore, it is the unique minimal-entropy -martingale measure. □
In the upcoming proof of Theorem 2 (the existence of the minimal-entropy -martingale measure), we construct a sequence in the class of absolutely continuous -martingale measures (each having finite entropy). We then take appropriate convex combinations (or a subsequence) whose Radon–Nikodým derivatives converge in . A priori, it is not obvious that the limit measure remains a -martingale measure. The following Lemma 3 guarantees exactly this: the limiting measure is still a -martingale measure.
Lemma 3. Let be a probability space, and let be probability measures on with for each n andSuppose X is a semimartingale that is a σ-martingale under each measure . Then, X is also a σ-martingale under . Proof. Since X is a -martingale under each measure , there exists, for every n, a suitable family of predictable sets making each localized piece of X a uniformly integrable martingale under . By reindexing or combining these families appropriately, we can select a single sequence of predictable sets for which is a uniformly integrable martingale simultaneously under every measure .
To show that
remains a uniformly integrable martingale under
, fix arbitrary times
and an arbitrary set,
. Define the bounded random variable
Since
is a martingale under each
, we immediately have
for every
n. Using the convergence of the Radon–Nikodým derivatives in
and boundedness of
Z, we get
This equality implies that the conditional expectation of , given under equals , establishing the martingale property for each localized piece. Since this argument applies to each predictable set, , it follows that X is a -martingale under . □
In a discrete-time market setting, the -martingale condition reduces to the usual martingale requirement on each time step. Then, the statement of the lemma is more transparent: if each makes X a martingale (so ), and if in , then X remains a martingale under . The “unified localizing sets” in discrete time simply become the entire time index set. The continuous-time version is conceptually the same but requires working with a single predictable sequence, , across all for the localization.
Theorem 2. Suppose there exists at least one measure, , with finite entropy, . Then, the unique minimal-entropy σ-martingale measure exists.
Proof. We define
By assumption, there exists a sequence,
, such that
is decreasing and satisfies
Because the sequence of the relative entropies is monotone, we obtain for the convex function
and, therefore, the sequence
is uniformly integrable according to Theorem A3. Per Theorem A4, there exists a subsequence,
, which converges weakly in
. However, according to Mazur’s lemma (Theorem A5), there exists a sequence,
, consisting of convex combinations
which converges in
to some
.
The measure
is uniquely defined if we interpret
as a Radon–Nikodym derivative. One can see directly that
. By assumption,
is a
-martingale with respect to all
(and hence also with reference to all
). Now, we want to show that it is also a
-martingale under
. To prove this, let
be a sequence of predictable sets, such that
is a martingale under a fixed
. For each
,
because
. Thus,
is also a martingale under
; hence,
is a
martingale under
.
Moreover, converges in to , and so, per Lemma 3 we have .
Because of the convexity of the relative entropy (Lemma 1), Equation (
3), and since
is a decreasing sequence, we have
Thus, by applying Fatou’s Lemma, we obtain
Since
is decreasing, it follows that
is already minimal. Now, all conditions of Lemma 2 are satisfied, so we conclude that
is equivalent to
, proving the existence of the minimal-entropy martingale measure.
Uniqueness follows from Lemma 1 and the strict convexity of . □
Remark 3. Theorem 2 shows that the “entropy-minimizing” approach extends smoothly to σ-martingale models: as soon as there is one finite-entropy σ-martingale measure, there must be a unique one of minimal entropy, even if is unbounded or not locally bounded. This measure can be used for entropy-based valuation, risk measurement, or other applications. The local martingale approach, in contrast, might be impossible if .
3. Convex and Coherent Risk Measures
In this section, we define the notions of monetary, convex, and coherent risk measures. We also highlight the relationships among translation invariance, monotonicity, convexity, positive homogeneity, and subadditivity.
Definition 5. Let be the space of linear bounded, real-valued random variables.
- (a)
A mapping, , is called a monetary risk measure if it satisfies for all
Monotonicity: if almost surely, then .
Translation invariance: for all , .
- (b)
The monetary risk measure ρ is called convex if it also satisfies
Convexity: for all and ,
- (c)
A convex risk measure, ρ, is called coherent if, in addition to monotonicity and translation invariance, it satisfies
Positive homogeneity: for all , Subadditivity: for all ,
Theorem 3. Let ρ be a monetary risk measure on a linear space of random variables .
- (a)
We say that ρ is normalized if . In particular, if ρ is positively homogeneous, then it follows that automatically.
- (b)
Suppose ρ satisfies translation invariance and monotonicity (as in Definition 5). Then, any two of the following three properties imply the remaining third:
Convexity;
Positive homogeneity;
Subadditivity.
Theorem 4. The minimal-entropy martingale measure and the entropic risk are both convex risk measures. The minimal-entropy martingale measure is also a coherent risk measure.
Proof. We start with the minimal-entropy martingale measure. It suffices to prove the coherence. The convexity follows from the fact that positive homogeneity and subadditivity imply convexity.
We now address the entropic risk.
Monotonicity. From
, almost surely,
follows for all
, and hence
for all
. Thus,
.
Cash translability. With
we obtain
.
And since, for general functions
and
, we have
the result follows.
□
In general, the entropic risk is not a coherent risk measure, even though it is additive for independent positions. However, there is also a coherent version of the entropic risk measure described in
Föllmer and Knispel (
2011), which is quite similar to the two risk measures we are examining here.
The coherence of a risk measure is particularly significant in the context of financial regulation. Regulators, such as the Basel Committee on Banking Supervision, emphasize capital requirements that are consistent with diversification benefits and penalize concentration risk (
Basel Committee on Banking Supervision 2010). In particular, the subadditivity property (cf. Definition 5 and the subsequent theorem establishing that
is coherent) ensures that merging two portfolios
and
never results in a higher combined capital charge than the sum of their individual charges. This requirement aligns with Basel III’s core principle that holistic risk management should not be discouraged by artificially additive capital rules.
When we specialize to the minimal-entropy risk measure
introduced in
Section 2 and shown in
Section 3 to be coherent, we see that it satisfies precisely these subadditivity and positive-homogeneity conditions. Concretely, if a bank adopts
to quantify its market or counterparty exposures, then merging business units or pooling risk factors would not inflate the aggregate capital requirement beyond the sum of stand-alone capital allocations. This feature could be advantageous in an internal models approach to Basel III capital calculation, provided that the bank can justify the assumptions on market completeness or the existence of
-martingale measures with finite entropy (see Theorem 2). Theoretically, replacing non-coherent measures (such as certain forms of value at risk) with a coherent measure like
better captures the true diversification effect.
Nonetheless, adopting a coherent measure in practice also raises considerations:
Model complexity. Calculating may require advanced numerical methods to identify or approximate the minimal-entropy -martingale measure.
Data intensity. Banks must maintain high-quality market data and stress scenarios to ensure robust parameter estimation under .
Regulatory validation. Any internal model, including entropy-based approaches, must pass supervisory review, which entails transparency on modeling assumptions and backtesting (
Basel Committee on Banking Supervision 2010).
Finally, to better understand these risk measures, we explore their relationship with the real-world probability measure. The following lemma will be helpful.
Lemma 4. Let be absolutely continuous σ–martingale measures, i.e., each of is in , and assume that and are both finite. Then, for any bounded random variable, , we have Proof. First, note that, for any probability measures
(absolutely continuous w.r.t.
P),
where
denotes the total variation distance:
Next, according to the triangle inequality in total variation,
Finally, we invoke the Pinsker inequality Theorem A8 and obtain
whenever
(resp.
) is finite. Hence,
Putting these pieces together yields
Therefore,
That completes the proof. □
Remark 4. Inequality (4) shows that two absolutely continuous σ-martingale measures and with small relative entropy yield similar expectations for any bounded payoff, X. In particular, a measure with small entropy is close to in total variation distance. Theorem 5. Let be bounded and . Then, the entropic risksatisfies the following two-sided bound: Proof. By taking the specific choice
in the definition of
, we see
since
. This proves the left inequality in (
5).
For any probability measure,
, with finite entropy
, we use the triangle bound from total variation Lemma 4 to obtain
Hence,
Therefore, for any such
,
Define
. Then, we want to maximize
over
. A short calculus argument shows that the maximum occurs at
and that the maximum value is
. Consequently,
Hence, for all
with
,
Taking the supremum over all
leads to
This completes the proof of (
5). □
Corollary 1. Let be the minimal-entropy σ-martingale measure for a (bounded) price process . Then,Furthermore, combining this estimate with the two-sided bound of Theorem 5, we have the following: Proof. Since
is defined as the supremum
it clearly holds that, for any
in that admissible set,
If
(the minimal-entropy
-martingale measure) is in
(i.e.,
and
), then simply take
in the supremum. This yields the lower bound (
7).
Finally, to deduce (
8), observe that
also holds by choosing
. Then, Theorem 5 provides the upper bound, so taking the maximum of these lower estimates proves (
8). □
Remark 5. The assumption that requires that the minimal-entropy measure be absolutely continuous with reference to with finite relative entropy. If, for instance, the market admits no arbitrage of the first kind (NA1), and there exists a finite-entropy EσMM, one can show . In that case, (7) provides a non-trivial lower estimate on the entropic risk regarding the minimal-entropy measure. Remark 6. When the risk-aversion parameter γ tends to zero, the penalty factor in the entropic risk measurebecomes very large. As a result, any measure with gets heavily penalized in the objective, so that the supremum is forced increasingly toward . Formally, it can be shown thatEconomically, this reflects a situation of ultra-strong aversion to deviating from the reference measure . In the limit , the only measure not incurring an infinite penalty is . Thus, the entropic risk measure collapses to , the plain (negative) expectation under the real-world probability . 4. Duality
The usual approach to defining the entropic risk measure is as follows:
and its robust representation, Equation (
1), is typically derived using dual representation theorems.
In this paper, we defined the entropic risk using its robust representation and now demonstrate that this definition is equivalent to the traditional one found in the literature. Since the convexity of Equation (
9) has not been explicitly shown here, we present an alternative proof.
Theorem 6. For and , we have the following: Proof. Define a new probability measure,
, as
Then,
Substituting this into the optimization problem yields
where the last equality follows from
and Lemma 1.
Taking the logarithm and dividing by yields the result. □
A detailed analysis of the dual problem for the minimal-entropy martingale measure is more complex but provides valuable insights into its economic interpretation as an equivalent local martingale measure. The dual problem, in essence, connects minimizing relative entropy to maximizing a specific utility function. Below, we elaborate briefly on this connection.
Let
U be a utility function of the form
We aim to maximize the expected utility of the terminal value of a wealth process by selecting an optimal strategy,
, from a set of admissible strategies,
. If the terminal value is represented as
, then the objective becomes
Under weak conditions,
Delbaen et al. (
2002) show
as well as
where
is the strategy that maximizes the utility in Equation (
10).
One key insight from the discussion above and, in particular, Theorem 6 is that the entropic risk measure can be derived by optimizing an exponential-based functional, namely
In principle, this representation is quite general, as it only requires that the exponential moment
be finite. Concretely, this exponential expectation is automatically well defined for
(bounded random variables). For unbounded positions, one needs
which is a standard integrability condition often seen in exponential utility frameworks (
Delbaen et al. 2002;
Föllmer and Schied 2011).
For many distributions commonly used in finance (e.g., lognormal, normal, or other exponential-family models), the condition (
11) is satisfied for all
. Thus, the entropic risk measure is well defined across a broad range of “light-tailed” or exponentially decaying distributions.
If X can exhibit extremely heavy tails (e.g., some Pareto-type or stable distributions), there may be values of for which . In that case, the entropic risk measure may not be well defined or may be only valid for smaller . In practical models, one imposes truncated tails or ensures that remains in a regime where is integrable.
From the dual perspective (cf. Equation (
1)), the supremum
remains finite under fairly general conditions. Essentially, if
X belongs to an appropriate Orlicz space or
space for which the exponential integrals exist, then the entropic risk measure is valid, regardless of the specific distribution family (
Detlefsen and Scandolo 2005;
Föllmer and Schied 2011).
Hence, the underlying probability measure is not required to come from a specific family (such as Gaussian or Lévy processes) for the exponential-based or entropy-based risk measure to work. The key assumption is that one can handle the necessary moment conditions (or integrability constraints) so that the exponentials are finite and the relative entropy functionals are well defined. Otherwise, if X exhibits tails so heavy that for every positive , the approach can fail to produce a finite value.
5. Dynamic Consistency
In this section, we present dynamic versions of the entropic risk measure and the risk measure associated with the minimal-entropy martingale measure and show that they are time-consistent. Dynamic consistency is particularly important when dealing with practical financial applications. For instance, dynamic risk measures based on entropy have been successfully applied to energy markets and stochastic volatility models, as discussed in
Swishchuk (
2007).
Definition 6. - (a)
A map, , is called a dynamic risk measure.
- (b)
A dynamic risk measure is called time-consistent if, for , - (c)
The dynamic entropic risk measure is defined as - (d)
The dynamic minimal-entropy risk measure is defined aswhere is the minimal-entropy martingale measure.
It is possible to allow to be an adapted process instead of a constant, resulting in a slightly more general definition of the dynamic entropic risk measure. However, in such cases, time consistency may not be guaranteed.
If, for
, we define the conditional relative entropy
as
The dynamic entropic risk measure can also be expressed in a robust representation involving the relative entropy
It is straightforward to verify that the two risk measures defined above satisfy the properties of dynamic risk measures.
Theorem 7. The dynamic entropic risk measure and the dynamic minimal-entropy risk measure are time-consistent.
Proof. We start with the entropic risk measure. We aim to show that
for
. Time consistency then follows from the intertemporal monotonicity theorem (
Föllmer and Penner 2006, Proposition 4.2).
Using the tower property of conditional expectation, we have the following:
For the minimal-entropy martingale measure, we proceed similarly:
This proves time consistency for both measures. □
In practice, many other risk measures are used. Therefore, it makes sense to compare the minimal-entropy risk measure with more commonly used approaches in the financial industry, such as value at risk (VaR) and conditional value at risk (CVaR, also called the expected shortfall). Below, we incorporate the standard definitions of VaR and CVaR, discuss their strengths and limitations, and highlight how they compare with the minimal-entropy framework, especially under incomplete markets and high volatility conditions.
The most used and well-known risk measure is value at risk (VaR). At a confidence level
, VaR typically answers the following:
“What is the smallest loss L so that X does not exceed −L with probability more than 1 − α?”
Formally, one writes
Due to its relative simplicity, VaR is still widely used by banks and regulatory bodies (
Basel Committee on Banking Supervision 2016). However, VaR is not in general subadditive, so it fails to be coherent (
Artzner et al. 1999) and does not quantify the magnitude of losses in the tail region beyond the threshold.
By comparison, the minimal-entropy risk measure
introduced in
Section 2 is coherent, thus reflecting diversification benefits more accurately. Moreover,
depends on underlying price dynamics and the selection of a minimal-entropy
-martingale measure, thereby embedding consistency with no-arbitrage principles in incomplete markets, while VaR typically lacks a direct link to hedging or replication arguments.
Alongside value at risk (VaR), conditional value at risk (CVaR) is likely the second most commonly used risk measure. Often referred to as the expected shortfall, CVaR at level
considers average tail losses above the VaR threshold. One standard formula is
Basel III and the Fundamental Review of the Trading Book increasingly advocate the use of CVaR for market risk calculations, recognizing its coherent properties and superior sensitivity to extreme events (
Basel Committee on Banking Supervision 2016). However, CVaR remains distribution-based, often sidelining explicit dynamic hedging or replication considerations. By contrast,
(the minimal-entropy measure) directly incorporates no-arbitrage dynamics and penalizes large deviations from the real-world measure via the Kullback–Leibler divergence. In highly volatile or incomplete markets,
thus captures both model risk (through entropy) and market risk (through
-martingale constraints) more naturally than CVaR can, though at the cost of greater computational and data requirements.
Beyond VaR and CVaR, the entropic VaR (
Ahmadi-Javid 2011) also introduces an entropy-based correction to the usual VaR framework, aiming for improved coherence-like properties. Meanwhile, expectile-type risk measures (
Bellini and Di Bernardino 2017;
Zaevski and Nedeltchev 2023) focus on generalized quantiles that can exhibit coherence under certain assumptions and capture the asymmetric tail risks. Nonetheless, these constructs typically remain tied to direct distributional assumptions, without necessarily unifying real-world dynamics and no-arbitrage arguments.
In
Section 3, we see that
fulfills the coherence axioms and, thanks to Theorem 2, is firmly rooted in a no-arbitrage framework for incomplete markets. By minimizing relative entropy with respect to
,
selects a risk-neutral (or
-martingale) measure that is closest to the real-world measure, thereby unifying pricing and risk. This unification becomes particularly relevant when local martingale measures may not exist.
In contrast to the above-mentioned advantages, there are also some drawbacks. Approximating or computing the minimal-entropy -martingale measure can be significantly more involved than applying VaR or other distribution-based measures. In volatile environments, calibrating Radon–Nikodým derivatives so that the entropy remains finite also demands extensive data and robust modeling. Moreover, although regulators increasingly tolerate advanced internal models (see Page 13), institutions must still validate the market incompleteness assumptions and explain the chosen measure to supervisors. Hence, some market participants prefer simpler, purely distribution-based schemes like CVaR or entropic VaR because they are more transparent and easier to backtest or justify to stakeholders.
Overall, while more sophisticated than VaR and even CVaR, the minimal-entropy risk measure furnishes a coherent and market-consistent framework that directly incorporates incomplete-market structures and penalizes divergence from real-world probabilistic views. Its principal obstacles involve higher computational costs and data intensity, potentially limiting immediate adoption in practice unless sufficiently robust calibration and infrastructure are available. Nonetheless, for investors, regulators, or institutions confronting model uncertainty and market incompleteness, the minimal-entropy methodology offers a powerful alternative that more fully respects no-arbitrage and coherent-risk requirements.
6. Optimal Risk Transfer
Optimal risk transfer is a classical topic in the study of risk measures and has been studied extensively in, e.g.,
Barrieu and El Karoui (
2005). In this section, we present a somewhat more general and extended view, highlighting in particular the property of
-dilated families of risk measures, the associated inf-convolution approach, and how these results can be restricted to the risk-neutral setting. We then provide a brief discussion of why no non-trivial
-dilation family can be constructed from the minimal-entropy coherent risk measure, whereas the standard entropic risk measure naturally fits into this framework. Finally, we show how restricting to (local) risk-neutral measures force the minimal-entropy measure to emerge as the unique solution to the risk-transfer problem.
Let
be any (convex or coherent) risk measure on
. The fundamental question in optimal risk sharing between two entities with risk measures
and
is to split a position,
X, into two parts
and
F. The goal is to minimize the combined risk
. This optimization problem is well expressed via the inf-convolution operator:
The optimal risk transfer is solved when the infimum in (
12) is achieved via some
. Well-known results (see
Barrieu and El Karoui (
2005)) show that certain risk measures, particularly those arising from
-dilated families, enjoy a simple solution: the optimal allocation is often a linear split of
X.
Definition 7 (
-dilated family).
A family of risk measures on is called γ-dilated if there exists a base risk measure ρ, such thatwhere is some index set. An immediate (but important) example is the entropic risk measure, which is
-dilated by construction:
if one sets
. By contrast, a coherent risk measure with positive homogeneity cannot usually form a non-trivial
-dilated family unless it is purely linear in
and thus collapses to the
-type functional (more details below).
Theorem 8. Let be a γ-dilated family of convex risk measures on . Then, the following statements hold:
- (a)
For any , we have the inf-convolution identity: - (b)
The optimal allocation solving is linear in X, specifically - (c)
Consequently, we also have - (d)
Moreover, if and are two convex risk measures that each can be embedded in a single γ-dilated family (i.e., they are both instances of for some base ρ), then, for any ,under a suitable identification. Hence, the γ-dilation structure is preserved under inf-convolution.
Sketch of Proof The key insight is that, for a
-dilated family, scaling
X by
and adjusting the risk measure by
yields a consistent re-parametrization of the same base functional
. Once this is established, standard inf-convolution arguments for the exponentially (or more generally,
-) tilted function imply the linear sharing rule (
14). For full details, see
Barrieu and El Karoui (
2005) or references therein. □
We now remark why a coherent risk measure—in particular, the minimal entropy risk measure
—cannot generally be part of a non-trivial
-dilated family. Recall that a coherent risk measure
is positively homogeneous, i.e.,
Combining this with the
-dilation requirement
quickly forces
for all
, unless
or
. Indeed, for coherent
, we would have
Hence,
cannot genuinely “depend” on
. In short, non-trivial
-dilated families (like the entropic class) are not coherent. This explains why the minimal-entropy risk measure,
, cannot appear in a standard
-dilated framework.