1. Introduction and Summary
This article proposes a family of generalized mutual information whose members are indexed by a positive integer n, with the nth member being the mutual information of nth order. The mutual information of the first order coincides with Shannon’s, which may or may not be finite. It is however established that each mutual information of an order greater than 1 is finitely defined for all distributions of two random elements on a joint countable alphabet, and that each and every member of the family enjoys several important utilities of a finite Shannon’s mutual information.
Let
Z be a random element on a countable alphabet
with an associated distribution
. Let the cardinality or support on
be denoted
, where
is the indicator function.
K is possibly finite or infinite. Let
denote the family of all distributions on
. Let
be a pair of random elements on a joint countable alphabet
with an associated joint probability distribution
, let the two marginal distributions be respectively denoted
and
. Let
denote the family of all distributions on
. Shannon [
1] offers two fundamental building blocks of information theory, Shannon’s entropy
, where the logarithm is 2-based; and mutual information
, where
,
and
are entropies respectively defined with the distributions
,
and
.
Mutual information plays a central role in the theory and the practice of modern data science for three basic reasons. First, the definition of
does not rely on any metrization on an alphabet, nor does it require the letters of the alphabet to be ordinal. This generality allows it to be defined and used in data spaces beyond the real coordinate space
, where random variables (as opposed to random elements) reside. Second, when
X and
Y are random variables assuming real values, that is, the joint alphabet is metrized,
captures linear as well as any non-linear stochastic association between
X and
Y. See Chapter 5 of [
2] for examples. Third, it offers a single-valued index measure for the stochastic association between two random elements, more specifically
for any probability distribution of
X and
Y on a joint alphabet and
if and only if
X and
Y are independent, under a wide class of general probability distributions.
However, mutual information
, in its current form, may not be finitely defined for joint distributions in a subclass of
, partially due to the fact that any or all of the three Shannon’s entropies in the linear combination may be unbounded. The said unboundedness prevents the potential utility of mutual information from being fully realized, and hence there is a deficiency of
, which leaves a void in
. (More detailed arguments are provided in
Section 2 below). This article introduces a family of generalized mutual information indexed by a positive integer
, denoted
, each of whose members,
, is referred to as the
nth order mutual information. All members of
are finitely defined for each and every
, except
, and all of them preserve the utilities of Shannon’s mutual information when it is finite.
The said deficiency of
is due to the fact that Shannon’s entropy may not be finite for “thick-tailed” distributions (with
decaying slowly in
k) in
. To address the deficiency of
, the issue of unboundedness of Shannon’s entropy on a subset of
must be addressed, through some generalization in one way or the other. The effort to generalize Shannon’s entropy has been long and extensive in the existing literature. The main perspective in the generalization in the existing literature is based on axiomatic characterization of Shannon’s entropy. Interested readers may refer to [
3,
4] for details and references therewithin. In a nutshell, with respect to the functional form,
, under certain desirable axioms, for example, [
5,
6],
is uniquely determined up to a multiplicative constant; if the strong additivity axiom is relaxed to be one of the weaker versions, say
-additivity or composability, then
may be of other forms, which give rise to Rényi’s entropy [
7], and the Tsallis entropy [
8]. However, all such generalization effort does not seem to lead to an information measure on a joint alphabet that would possess all the desirable properties of
, in particular
if and only if
X and
Y are independent, which is supported by an argument via the Kullback–Leibler divergence [
9].
Toward repairing the said deficiency of
, a new perspective of generalizing Shannon’s entropy is introduced in this article. In the new perspective, instead of searching for alternative forms of
in
under weaker axiomatic conditions, it is sought to apply Shannon’s entropy not to the original underlying distribution
but to distributions induced by
. One particular set of such induced distributions is a family, each of whose members is referred to as a conditional distribution of total collision (CDOTC) indexed by
. It is shown that Shannon’s entropy defined with every CDOTC induced by any
is bounded above, provided that
. The boundedness of the generalized entropy allows mutual information to be defined for any CDOTC of degree
for any
. The resulting mutual information is referred to as the
nth order mutual information index and is denoted
, which is shown to possess all the desired properties of
but with boundedness guaranteed. The main results are given and established in
Section 3 after several motivating arguments for the generalization of mutual information in
Section 2.
2. Generalization Motivated
To further motivate the generalization of mutual information in this article, let the definition of mutual information be considered in a broader perspective. Inherited from the Kullback–Leibler divergence, mutual information on a joint alphabet,
, is unbounded for a large subclass of distributions in
. Example 1 below demonstrates the existence of such a subclass of joint distributions.
Example 1. Letbe a probability distribution withfor every k but unbounded entropy. Letbe such thatfor allandfor all, henceand. Then.
One of the most attractive properties of mutual information is that mutual information
is finitely defined for all joint distributions such that
for all
and
and
if and only if the two random elements
X and
Y are independent. However, the utility of mutual information is beyond a mere indication of whether it is zero or not. The magnitude of mutual information is also of essential importance, although Shannon did not elaborate on that in his landmark paper [
1]. The said importance is perhaps best illustrated by the notion of the standardized mutual information defined as
and Theorem 1 below.
Remark 1. There are several variants of standardized mutual information proposed in the existing literature. Interested readers may refer to [10,11,12,13]. Not all variants of the standardized mutual information have the properties given in Theorem 1. A summary of standardized mutual information is found in Chapter 5 of [2]. However, before stating Theorem 1, Definition 1 below is needed.
Definition 1. Random elementsandare said to have a one-to-one correspondence, or to be one-to-one corresponded, under a joint probability distributionon, if:
for every i satisfying, there exists a unique j such that, and
for every j satisfying, there exists a unique i such that.
Theorem 1. Letbe a pair of random elements on alphabetwith joint distributionsuch that. Then:
,
if and only if X and Y are independent, and
if and only if X and Y are one-to-one corresponded.
A proof of Theorem 1 can be found on page 159 of [
2]. Theorem 1 essentially maps the independence of
X and
Y (the strongest form of unrelatedness) to
, one-to-one correspondence (the strongest form of relatedness) to
, and everything else in between. In so doing, the magnitude of mutual information is utilized in measuring the degree of dependence in pairs of random elements, which could lead to all sorts of practical tools for evaluating, ranking, and selecting variables in data space.
It is important to note that the condition of
is essential in Theorem 1, since obviously, without it,
may not be well defined. In fact, if
is not imposed, and even observing reasonable conventions such as
and
, the statements of Theorem 1 may not be true. To see this, consider the following constructed example.
Example 2. Letbe a probability distribution withfor every k but unbounded entropy. Letbe such thathenceand.
X and Y are obviously not independent, and It follows thatbut in this case. Therefore Part 2 of Theorem 1 fails.
Example 2 indicates that mutual information in its current form is deprived of the potential utility of Theorem 1 for a large class of joint distributions and therefore leaves much to be desired.
Another argument for the generalization of mutual information can be made in a statistical perspective. In practice, mutual information is often to be estimated from sample data. For statistical inference to be meaningful, the estimand
needs to exist, i.e.,
. More specifically, in testing the hypothesis of independence between
X and
Y,
, where
is the subclass of all joint distributions for independent
X and
Y on
, and
needs to be finitely defined in an open neighborhood of
in
, or else the logic framework of statistical inference is not well supported. Let
be the subclass of
such that
. In general, it can be shown that
is dense in
with respect to the
p-norm for
. Specifically, for any
, there exists a sequence of distributions
such that
. See Example 3 below.
Example 3. Letwherefor allsuch thatand. Obviously X and Y are independent under, that is, . Letbe constructed based onas follows.
Remove an arbitrarily small quantitywhereaway from each of the four positive probabilities inso each becomesfor all,
such thatand.
Extend the range oftoand,
and allocate the mass ε over the extended range according towhere c is such that.
Under the constructed,
for any,
X and Y are not independent, and the corresponding mutual information is However, noting that as,
and hence,
All things considered, it is therefore desirable to have a mutual information measure, say , or for that matter a family of mutual information measures indexed by a positive integer n, such that for all distributions in , and with an accordingly defined standardized mutual information measure such that the utility of Theorem 1 is preserved with in place of for all distributions in .
3. Main Results
Given
and
, consider the experiment of drawing an identically and independently distributed (
) sample of size
n. Let
denote the event that all observations of the sample take on a same letter in
, and let
be referred to as the event of total collision. The conditional probability, given
, that the total collision occurs at letter
is
It is clear that
is a probability distribution induced from
. For each
n,
of (
1) is the conditional distribution of total collision (CDOTC) with
n particles.
Remark 2. It is to be noted that, given a,
of (1) is a special member of the family of the escort distributions introduced by [14]. The escort distributions are a useful tool in thermodynamics. Interested readers may refer to [15] for a concise introduction. Lemma 1. For each n, , anduniquely determine each other.
Proof. Given
, by (
1),
is uniquely determined. Conversely, given
, for each
n and all
,
and therefore
The lemma follows. □
Lemma 2. For each n, , and for any, .
Proof. Write
. Noting
,
and therefore
for all
,
The lemma follows. □
On the joint alphabet
with distribution
, consider the associated CDOTC for an
n and all pairs
such that
and
,
Let
. It is to be noted that
. The two marginal distributions of (
3) are
and
, respectively, where
Lemma 3. if and only if.
Proof. For each positive integer
n, if
for all pairs
,
and
, then
where the two factors of the last expression above are respectively
and
,
,
, are letter values of the
n observations in the sample.
Conversely, if
where
depends only on
n and
i and
only depends on
n and
j, then by (
2),
The lemma immediately follows the factorization theorem. □
For each
, let
,
and
be Shannon’s entropies defined with the joint CDOTC,
as in (
3), and the marginal distributions
and
as in (
4) and (
5), respectively. Let
Theorem 2. For everyand any,
,
if and only X and Y are independent.
Proof. In Part 1, , since is a mutual information and by Lemma 2. Part 2 follows Lemma 3 and the fact that is a mutual information. □
Let
be referred to as the
nth order standardized mutual information, and write
. Let
be a pair of random elements on
according to the induced joint distribution
with index value
.
Lemma 4. X and Y have a one-to-one correspondence if and only if and have one.
Proof. If
X and
Y have a one-to-one correspondence, then for each
i, there is a unique
such that
and
for all other
j,
. By (
3),
and
for all other
j,
. That is,
and
have a one-to-one correspondence.
Conversely, if
and
have a one-to-one correspondence, then for each
i, there is a unique
such that
and
for all other
j,
. On the other hand, by (
2),
it follows that
and
for all other
j,
. That is,
X and
Y have a one-to-one correspondence. □
Corollary 1. For everyand any,
,
if and only if X and Y are independent, and
if and only if X and Y are one-to-one corresponded.
Proof. By Lemma 3, X and Y are independent if and only if and are. By Lemma 4, X and Y are one-to-one corresponded if and only if and are. The statement of Corollary 1 follows directly from Theorem 1. □
Theorem 2 and Corollary 1 together fill the void in left behind by .
4. Concluding Remarks
The main results of this article may be summarized as follows. A family of generalized mutual information indexed by a positive integer n is proposed. The member corresponding to is Shannon’s mutual information for a given joint distribution, . The other members of the family correspond to other integer values of n. They are also Shannon’s information defined, not with , but with induced distributions based on the given distribution . These induced distributions are called conditional distributions of total collision (CDOTC), which collectively is a special subset of a more general family called the escort distributions, which is often studied in extensive thermodynamics. The main motivation of the generalized mutual information is to resolve the issue of the fact that the standard mutual information is not finitely defined for all distributions of a countable joint alphabet , which leads to the issue of mutual information’s utility only realized on a fraction of .
On a more specific and finer level, the following facts are established.
There is a one-to-one correspondence between each CDOTC and the given distribution on a countable alphabet, and hence each CDOTC is a characteristic representation of the original distribution . One of the implications of this fact is that understanding the underlying is equivalent to understanding one of its CDOTC. It can be shown that the CDOTC with an order greater than 1 is much easier to estimate than with sparse data.
Each generalized mutual information is guaranteed to be finite. This result essentially guarantees the validity of statistically testing the null hypothesis of independence of two discrete random elements, as it guarantees the existence of (generalized) mutual information anywhere in the alternative space of dependent join distributions.
It is shown that a particular form of standardized mutual information , defined with any CDOTC of any order greater than 1, preserves the zero-to-one scale with independence on one end and total dependence on the other, which is enjoyed by Shannon’s entropy only when it is finite.
In short, the family of conditional distributions of total collision embeds the underlying probability distribution as a special member, and the family of generalized mutual information embeds Shannon’s mutual information as a special member. Consequently, the stochastic association on joint alphabets can be measured by not only one index but by a host of indices, which collectively offer a much extended space to study stochastic dependence in information theory.