1. Introduction
The motivation for this study comes from an online query asking whether natural exponential families (NEFs) with unbounded support can have bounded variance functions (VFs) beyond the classical normal NEF. In response, we establish explicit sufficient conditions for the existence of such NEFs. These conditions reveal a broad class of previously undocumented non-elementary NEFs whose Laplace transforms (LTs) or VFs cannot be expressed in elementary (algebraic) form. Thanks to modern mathematical software, such as Maple and Mathematica, as well as open-source libraries in Python and R, this obstacle is no longer a practical limitation; symbolic or numerical evaluation of the relevant transforms has become routine. The normal NEF serves as a benchmark within the subclass studied in this context. Our partial solution deepens understanding of the interplay between support structure and variance behavior in NEFs, thereby addressing a gap in the current literature. Although our focus is theoretical, we expect the subclasses introduced in this paper to assist in modeling real-world data across diverse fields.
The VF is central to the theory of natural exponential families (NEFs). Although the normal NEF is the only known example with a constant VF, it is natural to ask whether other NEFs—particularly those with unbounded support on —can also have bounded VFs. This issue is important both theoretically and practically, as bounded VFs often lead to more stable modeling.
This paper establishes three complementary results—Theorem 1 and Propositions 1 and 2—that provide sufficient conditions under which an NEF supported on the entire real line can still have a bounded VF. Theorem 1 imposes a polynomial-growth constraint on the cumulant generating function, ensuring sufficient tail decay of the density. Proposition 1 relies on the Legendre duality: a uniform positive lower bound on the second derivative of the generating function guarantees global convexity and, consequently, a bounded VF. Proposition 2 builds on the standard normal distribution and constructs an explicit sequence of NEFs whose VFs remain bounded; for this sequence, both the LTs and the VFs can be written in closed form. Together, these results introduce a broad new class of NEF models, greatly expanding the range of distributions available for real-data analysis and providing strong alternatives to traditional families.
This paper is structured as follows.
Section 2 provides an overview of the essential background on NEFs.
Section 3 presents Theorem 1, along with Propositions 1 and 2, and explains how these results are related.
Section 4 provides several illustrative examples, while
Section 5 offers concluding remarks.
2. Preliminaries on NEFs
The preliminaries are mainly drawn from [
1,
2]. We provide only the brief background necessary for the developments that follow. Let
be a non-Dirac positive Radon measure on
,
S the support of
, and
C the convex hull of
S. LT of
is a mapping
defined by
Let
D denotes the effective domain of
, i.e.,
and assume that
Let
be the cumulant transform of
Then the NEF
F generated by
is defined by the family of probabilities
The measure
is called a generating measure of the NEF. It is not unique, as any other measure of the form
with real constants
a and
b, generates the same NEF (c.f., [
2]). The function
defined on
is real analytic on there, and its successive derivatives are the successive cumulants of
F. In particular,
and
, are the mean and variance of
F.
As we consider only cases where is absolutely continuous with respect to the Lebesgue measure on , we henceforth restrict our attention to generating measures of the form , where f is the respective Radon–Nikodym derivative.
The following are some notions and properties related to NEFs that will be used in the sequel.
Regular NEFs. If D is an open interval, F is called regular. Particularly, if or F is regular.
Mean value parameterization: The cumulant transform
k is strictly convex and real analytic on
and
is the mean function of
F. The open interval
is called the mean domain of
F. Since the map
is one-to-one, its inverse function
is well defined. Hence, the map
is one-to-one from
M onto
F and is called the mean value parameterization of
Variance Function (VF). The variance corresponding to the NEF (
1) is
The map
from
M into
is called the variance function (VF) of
F. In fact, a VF of an NEF
F is a pair
. It uniquely determines an NEF within the class of NEFs. For instance, the VF
with fixed
characterizes the normal NEF within the class of NEFs.
Steep NEFs. An NEF
is called steep iff
C. Also,
F is steep if its effective domain
D is an open interval (see [
3], Theorem 8.2), particularly,
F is steep if
or
Steep NEFs ensure, with probability one, the existence of the MLE for
m as the stationary point of the first derivative of the likelihood function.
3. Sufficient Conditions on the Boundedness of VFs
Finding necessary and sufficient conditions for this problem is extremely challenging. Even identifying necessary conditions that closely approach sufficiency is highly non-trivial. Undoubtedly, achieving such results would constitute a significant theoretical advance, but we leave this task for future research.
Nevertheless, sufficient conditions can be established, and this paper provides them in Theorem 1 and Propositions 1 and 2. Theorem 1 imposes a polynomial-growth bound on the generating measure of the NEF, ensuring rapid tail decay and a uniformly bounded VF. Proposition 1 is based on Legendre duality and requires a uniform positive lower bound on the second derivative of the generating measure, which likewise ensures a bounded VF. Proposition 2 begins with the standard normal distribution and constructs an explicit sequence of NEFs whose LTs and VFs remain bounded.
Most examples stemming from Theorem 1 or Proposition 1 lack closed-form expressions for their LTs or VFs, but modern mathematical software can handle them efficiently in practice. By contrast, every example associated with Proposition 2 admits explicit LTs and VFs.
Collectively, these results uncover an exceptionally broad class of NEFs. To our knowledge, none of the NEFs produced by these three results have appeared previously in the literature, thereby providing a rich collection of practical models that rival the normal and other classical distributions for real-data analysis. In this paper, we do not address applications of these families to real data; this will be pursued in future research.
Theorem 1. Let be a continuous function satisfying the following tail growth conditions:for some constants and Defineand let F be the NEF generated by Then, 1. , implying that the NEF generated by f is supported on
2. The LT is finite for all , i.e., the respective canonical parameter space is . Thus, F is steep and .
3. For , the VF of F is uniformly bounded by 1. (For the VF is not uniformly bounded as it grows without limit as θ or m tends to ∞ or ).
Proof. 1. Let be sufficiently large. Split the integral into three intervals of integration: and Since f is continuous on , then by the extreme value theorem, is bounded. On , , which integrates for . The left tail is similar to . Thus, . Hence,
2. For this part, we use Young’s inequality ([
4], p. 27), which states the following: Let
with
. Then, for all real
,
For this part, we split the implementation of Young’s inequality into the two parts of the tails:
(a) On the right tail (
,
) we take
and
. Set
Then
and
(b) On the left tail (
,
) we take
,
and
giving
Fix
and use the bounds derived from Young’s inequality to obtain
Exponentiating and integrating yield
The integral converges because
. The left tail for
already decays exponentially; the compact part
is finite. For
a symmetric bound with
q gives convergence. Hence
for all
. Thus,
F is regular and therefore steep with
.
3. For
,
we shall find a quadratic bound on the cumulant transform
For this, we first show that
Let
and split
Consider the central interval
and use the elementary inequality
with
and
, then
Insert this into the central integral in (
4), then
For the right and left tails, there exist constants
A and
B such that
Similarly, for
and
we get
Thus, for the right tail, we have
where
. The left tail is similar with a constant
. Let
then (
3) holds and thus
, where
We now show that the VF is bounded. Consider the second finite difference
The strict convexity of
k yields a lower bound:
. An upper bound of
gives
Hence,
Because
is real analytic,
exists everywhere and
This implies that
□
Remark 1. By Theorem 1 and all of the propositions below, is not necessarily a probability density function. To make it such, we need to divide it by , where is the normalizing constant, which may depend on the various parameters involved. In most of the following examples, these normalizing constants can be computed numerically or expressed in terms of certain transcendental functions.
In the following proposition, the tail condition required by Theorem 1 is satisfied. However, we prefer to present it separately as Proposition 1, since it assumes that is bounded below by some constant , which provides a stronger and more self-contained control over the variance function. The proof, based on Legendre duality, is both elegant and innovative, and we believe that its presentation is valuable in its own right. Moreover, the condition on is relatively easy to verify in concrete examples and facilitates the construction of NEFs that satisfy it. This property of will be illustrated in Examples 3–5 below.
The Legendre duality presents the cumulant transform
and the convex generating function
g as Legendre conjugates (see [
5], Chapter 26; [
3]). Concretely,
and, for each
, the maximizer
of the second expression satisfies
and
Proposition 1. Let be a twice continuously differentiable and convex function satisfying the following condition: There exists a constant such that for all . Define and F the NEF generated by f. Then,
1. , i.e., the support of F is .
2. The LT of f is finite for all , and thus F is regular and steep with .
where is the unique minimizer of
4. g satisfies Theorem 1.
Proof. 1. Since
, we can integrate this inequality:
Hence,
as
, and
as
.
So for large
,
for some constant
C. Therefore,
and this is integrable over
. Thus,
.
2. For showing
, we analyze
From above,
, so,
This is integrable over for all since the exponent is quadratic in x, and the quadratic term dominates. Therefore, for all , and the canonical parameter space .
3. We prove this part by using the Legendre duality. Let
. The unique minimizer
satisfies
, due to the strict convexity of
g. Since
g is convex and differentiable, the Legendre transform of
g is
So,
Hence,
and
Differentiate both sides of
using the chain rule:
4. We show that there exist constants
and
such that
For this, fix any point
. We can integrate this inequality because
. By integrating from
to
, we get
Now, integrating
from
to
x, then
Using the lower bound for
implies
Hence, as
, this inequality shows
for some constant
(since
,
, and
are fixed). So, for sufficiently large
x,
for some
. A symmetric argument applies for
, using integration from
x to
. Therefore,
satisfies the tail lower bounds required by Theorem 1, with
. □
The following proposition is based on the standard normal distribution and generates many examples that, in some cases, satisfy Theorem 1 while violating Proposition 1. The LTs and VFs for these examples can be expressed explicitly, and the corresponding NEFs may have broad practical applicability.
Proposition 2. Letand be the NEF generated by . Denote by and , respectively, the LT and VF corresponding to . Then, 1. The support and the canonical parameter space of is .
2. The LT of has the formwhere Here, is the double factorial, and is an even polynomial of degree whose all coefficients are strictly positive.
3. Fix then the VF satisfies 4. does satisfy the premises of Theorem 1 but not of Proposition 1.
5. The LT and VF of in (6) are given, respectively, byandwhere is given by (7). The special cases are given, respectively, by (a) Proof. 1. This part is trivial
2. Clearly,
where
To get the explicit form of
write
Taking expectations
Only even powers contribute due to the symmetry of
, and thus,
so that
is an even polynomial of degree
with strictly positive coefficients. Note, however, that
, where
denotes the Hermite polynomial.
3. To prove (
8), note that
The following steps carry out the proof of this part: (a) computing
at 0; (b) defining the tilted measure
; (c) showing that
for
; (d) showing the symmetry of
where
is defined in (
12); (e) justifying the log-convexity of
in
k; (f) showing the monotonicity of
(a) Evaluation at
: Compute
So,
(d) Since
it follows that
So, it suffices to prove that
for
.
(e) Let
and define
, then
We now analyze the sequence
to show that, for each fixed
To this end, we establish the following: strict log-convexity of the sequence
; convexity of
in
k; monotonicity of successive ratios, followed by the application of (
13). Recall that a sequence
of positive numbers is log-convex if
For the normal moments, this strict inequality follows from Turán’s inequality for Hermite polynomials (c.f., [
6,
7,
8]); hence, the sequence
is strictly log-convex. Setting
and taking logarithms in (
14) yields
That is, the discrete second difference of the function
is positive. In discrete calculus, this is precisely the definition of strict convexity. Moreover, for any positive sequence, strict log-convexity is equivalent to the property that
Applying this to
yields (
13), for every fixed
.
(f) Differentiate
to obtain
Hence, from the final inequality in (e), it follows that
All the preceding steps demonstrate that
and
For
we use the inequality
and then note that
once
. Hence,
Thus, the quadratic tail condition of Theorem 1 (with
) is indeed satisfied. However,
violates Proposition 1 as
Since
is not defined everywhere, we cannot even begin to verify a uniform bound for all
x.
5. A straightforward calculation yields these results. □
To illustrate the functional relationship between the variance and the canonical parameter
, as well as the upper bound of the variance function,
Figure 1 displays
,
and 3, corresponding to Equations (
9), (
10), and (
11), respectively.
4. Examples
None of the following examples has appeared in the literature in NEF form, likely because their Laplace transforms lack closed-form expressions (except for those related to Proposition 2). As noted in the Introduction, modern mathematical software can easily handle such cases numerically.
Collectively, these examples provide a rich class of NEFs with unbounded support and bounded variance functions, offering valuable alternatives to the normal NEF and other standard distributions in statistical modeling. We present only a small selection from a much larger set, omitting additional cases for brevity.
Example 1. The Generalized normal distributions (GNDs). Let andwhere the appropriate normalizing constant making f a density is . The class of generalized normal distributions (GNDs) has a long history, beginning with Russian researchers (see [9], and the references therein). A GND with provides a robust, flexible alternative to the normal distribution, making it suitable for data that exhibit sharp peaks and moderate outliers. It is especially useful in signal processing, robust regression, and sparsity-based inference. The density given in (15) is symmetric and unimodal, with heavier tails and a sharper peak than the normal distribution; these features make it leptokurtic. and well suited to settings where moderate outliers or pronounced central tendencies are expected. GNDs with appear as error distributions in robust estimation ([10]), Bayesian regression ([11]), signal and audio processing ([12]), and independent component analysis and sparse coding ([13]). These examples satisfy Theorem 1 but not Proposition 1—except when which corresponds to the normal model. Specifically, Theorem 1 requires for large , and for large , with constants and exponents . This condition yields a bounded variance for the associated NEF. For this case,which meets the first two conditions of Theorem 1 with and . For , Part 3 of Theorem 1 only hints that the VF is not uniformly bounded; indeed, it diverges as the canonical parameter or m, tends to ∞ or . Uniform boundedness of V requires so Proposition 1 applies only to the Gaussian case Specifically,For , this second derivative vanishes at the origin, whereas for , it diverges there. Consequently, Proposition 1 fails whenever In particular, (Details are omitted for brevity.)
Example 2. The extended generalized normal distributions (EGNDs). LetWe refer to the NEF generated by as the EGND NEF, as we intend to explore it further and apply it to real data sets. As in Example 1, all the assumptions of Theorem 1 are satisfied. Proposition 1 holds only when either p or q equals 2. For example, if , then with , in which case, Example 3. LetThis g satisfies both Theorem 1 and Proposition 1. Theorem 1 requires as and as , for some and . Here,and is bounded for all x. As , we have so for sufficiently large ,with . Hence, both Theorem 1 and Proposition 1 are satisfied, with The choice is is essential, as the function loses strict convexity when . Example 4. LetFirst, we show that is integrable over ; that is,For , consider the left tail:where the change of variables is used. A similar change of variables handles the right tail. Hence, f is integrable. Next, with a single minimum at so Consequently, both Theorem 1 and Proposition 1 hold.
Example 5. Differentiating twice giveswhere . Since , it follows that Thus, Proposition 1 is satisfied with Theorem 1 also holds: for large positive with and for large negative x, with Hence, the required growth conditions are met, and since everywhere, Proposition 1 is confirmed. 5. Concluding Remarks
NEFs supported on the entire real line and possessing bounded VFs are uncommon; historically, the normal family has been the primary example. This paper broadens that landscape by presenting three complementary sets of sufficient conditions that ensure unbounded support and bounded VFs.
Tail-growth control (Theorem 1). If the cumulant generator g dominates two even polynomials of degree , then is integrable, its Laplace transform is finite on , and the corresponding VF is bounded by 1.
Uniform strong convexity (Proposition 1). Whenever for all , Legendre duality yields for every mean m.
A closed-form Gaussian sequence (Proposition 2). Weighting the standard normal density by even powers of x produces a family of NEFs whose VFs satisfy and whose Laplace transforms remain elementary.
A set of twelve fully worked examples demonstrates how each route succeeds—or fails—in practice, highlighting the logical independence of the two analytic criteria. Although most examples lack closed-form Laplace transforms, modern symbolic and numerical software make them readily applicable. These results can generate almost infinitely many NEFs of this type, satisfying at least one of the three statements in Theorem 1 or Propositions 1 and 2.
For an empirical study, it is often advantageous to embed a given
g in a richer parametric family. Simple rescaling
or decompositions
preserve the hypotheses of Theorem 1 and Propositions 1 and 2, while endowing the resulting NEFs with additional shape control (see Example 2 for an illustration). These extra degrees of freedom, in conjunction with the canonical parameter, make the families more adaptable to real data and will be further explored in forthcoming applied studies.
The purpose of presenting so many examples is to underscore the virtually endless supply of exponential families supported on the real line that can be constructed—none of which were previously known. Moreover, each of these families can be extended to a multi-parameter family, as shown in Equations (
17) and (
18).
Looking ahead, two applied research projects are currently underway with several collaborators, aiming to apply the theoretical results of this paper to modeling real data in the various fields indicated below.
1. Model competition. The proposed NEFs, all defined on with bounded VFs, offer natural competitors to the normal, Laplace, logistic, hyperbolic, and secant distributions—as well as other classical distributions—in contexts where tail behavior or regularization is essential.
2. Statistical methodology. Investigating maximum likelihood and Bayesian inference for these families—particularly those lacking closed-form transforms—promises new tools for robust regression, signal processing, and econometrics.
3. Focus. In particular, we will focus on the NEFs generated by the generalized normal distributions (Example 1) and the extended generalized normal distributions (Example 2). We will explore their probabilistic properties and apply them to real data arising from signal and audio processing, independent component analysis, and sparse coding.