Abstract
Mutual information is one of the essential building blocks of information theory. It is however only finitely defined for distributions in a subclass of the general class of all distributions on a joint alphabet. The unboundedness of mutual information prevents its potential utility from being extended to the general class. This is in fact a void in the foundation of information theory that needs to be filled. This article proposes a family of generalized mutual information whose members are indexed by a positive integer n, with the nth member being the mutual information of nth order. The mutual information of the first order coincides with Shannon’s, which may or may not be finite. It is however established (a) that each mutual information of an order greater than 1 is finitely defined for all distributions of two random elements on a joint countable alphabet, and (b) that each and every member of the family enjoys all the utilities of a finite Shannon’s mutual information.
Keywords:
mutual information; Shannon’s entropy; conditional distribution of total collision; generalized entropy; generalized mutual information MSC:
primary 60E10; secondary 94A15, 82B30
1. Introduction and Summary
This article proposes a family of generalized mutual information whose members are indexed by a positive integer n, with the nth member being the mutual information of nth order. The mutual information of the first order coincides with Shannon’s, which may or may not be finite. It is however established that each mutual information of an order greater than 1 is finitely defined for all distributions of two random elements on a joint countable alphabet, and that each and every member of the family enjoys several important utilities of a finite Shannon’s mutual information.
Let Z be a random element on a countable alphabet with an associated distribution . Let the cardinality or support on be denoted , where is the indicator function. K is possibly finite or infinite. Let denote the family of all distributions on . Let be a pair of random elements on a joint countable alphabet with an associated joint probability distribution , let the two marginal distributions be respectively denoted and . Let denote the family of all distributions on . Shannon [1] offers two fundamental building blocks of information theory, Shannon’s entropy , where the logarithm is 2-based; and mutual information , where , and are entropies respectively defined with the distributions , and .
Mutual information plays a central role in the theory and the practice of modern data science for three basic reasons. First, the definition of does not rely on any metrization on an alphabet, nor does it require the letters of the alphabet to be ordinal. This generality allows it to be defined and used in data spaces beyond the real coordinate space , where random variables (as opposed to random elements) reside. Second, when X and Y are random variables assuming real values, that is, the joint alphabet is metrized, captures linear as well as any non-linear stochastic association between X and Y. See Chapter 5 of [2] for examples. Third, it offers a single-valued index measure for the stochastic association between two random elements, more specifically for any probability distribution of X and Y on a joint alphabet and if and only if X and Y are independent, under a wide class of general probability distributions.
However, mutual information , in its current form, may not be finitely defined for joint distributions in a subclass of , partially due to the fact that any or all of the three Shannon’s entropies in the linear combination may be unbounded. The said unboundedness prevents the potential utility of mutual information from being fully realized, and hence there is a deficiency of , which leaves a void in . (More detailed arguments are provided in Section 2 below). This article introduces a family of generalized mutual information indexed by a positive integer , denoted , each of whose members, , is referred to as the nth order mutual information. All members of are finitely defined for each and every , except , and all of them preserve the utilities of Shannon’s mutual information when it is finite.
The said deficiency of is due to the fact that Shannon’s entropy may not be finite for “thick-tailed” distributions (with decaying slowly in k) in . To address the deficiency of , the issue of unboundedness of Shannon’s entropy on a subset of must be addressed, through some generalization in one way or the other. The effort to generalize Shannon’s entropy has been long and extensive in the existing literature. The main perspective in the generalization in the existing literature is based on axiomatic characterization of Shannon’s entropy. Interested readers may refer to [3,4] for details and references therewithin. In a nutshell, with respect to the functional form, , under certain desirable axioms, for example, [5,6], is uniquely determined up to a multiplicative constant; if the strong additivity axiom is relaxed to be one of the weaker versions, say -additivity or composability, then may be of other forms, which give rise to Rényi’s entropy [7], and the Tsallis entropy [8]. However, all such generalization effort does not seem to lead to an information measure on a joint alphabet that would possess all the desirable properties of , in particular if and only if X and Y are independent, which is supported by an argument via the Kullback–Leibler divergence [9].
Toward repairing the said deficiency of , a new perspective of generalizing Shannon’s entropy is introduced in this article. In the new perspective, instead of searching for alternative forms of in under weaker axiomatic conditions, it is sought to apply Shannon’s entropy not to the original underlying distribution but to distributions induced by . One particular set of such induced distributions is a family, each of whose members is referred to as a conditional distribution of total collision (CDOTC) indexed by . It is shown that Shannon’s entropy defined with every CDOTC induced by any is bounded above, provided that . The boundedness of the generalized entropy allows mutual information to be defined for any CDOTC of degree for any . The resulting mutual information is referred to as the nth order mutual information index and is denoted , which is shown to possess all the desired properties of but with boundedness guaranteed. The main results are given and established in Section 3 after several motivating arguments for the generalization of mutual information in Section 2.
2. Generalization Motivated
To further motivate the generalization of mutual information in this article, let the definition of mutual information be considered in a broader perspective. Inherited from the Kullback–Leibler divergence, mutual information on a joint alphabet, , is unbounded for a large subclass of distributions in . Example 1 below demonstrates the existence of such a subclass of joint distributions.
Example 1.
Letbe a probability distribution withfor every k but unbounded entropy. Letbe such thatfor allandfor all, henceand. Then.
One of the most attractive properties of mutual information is that mutual information is finitely defined for all joint distributions such that for all and and if and only if the two random elements X and Y are independent. However, the utility of mutual information is beyond a mere indication of whether it is zero or not. The magnitude of mutual information is also of essential importance, although Shannon did not elaborate on that in his landmark paper [1]. The said importance is perhaps best illustrated by the notion of the standardized mutual information defined as and Theorem 1 below.
Remark 1.
There are several variants of standardized mutual information proposed in the existing literature. Interested readers may refer to [10,11,12,13]. Not all variants of the standardized mutual information have the properties given in Theorem 1. A summary of standardized mutual information is found in Chapter 5 of [2].
However, before stating Theorem 1, Definition 1 below is needed.
Definition 1.
Random elementsandare said to have a one-to-one correspondence, or to be one-to-one corresponded, under a joint probability distributionon, if:
- for every i satisfying, there exists a unique j such that, and
- for every j satisfying, there exists a unique i such that.
Theorem 1.
Letbe a pair of random elements on alphabetwith joint distributionsuch that. Then:
- ,
- if and only if X and Y are independent, and
- if and only if X and Y are one-to-one corresponded.
A proof of Theorem 1 can be found on page 159 of [2]. Theorem 1 essentially maps the independence of X and Y (the strongest form of unrelatedness) to , one-to-one correspondence (the strongest form of relatedness) to , and everything else in between. In so doing, the magnitude of mutual information is utilized in measuring the degree of dependence in pairs of random elements, which could lead to all sorts of practical tools for evaluating, ranking, and selecting variables in data space.
It is important to note that the condition of is essential in Theorem 1, since obviously, without it, may not be well defined. In fact, if is not imposed, and even observing reasonable conventions such as and , the statements of Theorem 1 may not be true. To see this, consider the following constructed example.
Example 2.
Letbe a probability distribution withfor every k but unbounded entropy. Letbe such that
henceand. X and Y are obviously not independent, and
It follows thatbut in this case. Therefore Part 2 of Theorem 1 fails.
Example 2 indicates that mutual information in its current form is deprived of the potential utility of Theorem 1 for a large class of joint distributions and therefore leaves much to be desired.
Another argument for the generalization of mutual information can be made in a statistical perspective. In practice, mutual information is often to be estimated from sample data. For statistical inference to be meaningful, the estimand needs to exist, i.e., . More specifically, in testing the hypothesis of independence between X and Y, , where is the subclass of all joint distributions for independent X and Y on , and needs to be finitely defined in an open neighborhood of in , or else the logic framework of statistical inference is not well supported. Let be the subclass of such that . In general, it can be shown that is dense in with respect to the p-norm for . Specifically, for any , there exists a sequence of distributions such that . See Example 3 below.
Example 3.
Letwherefor allsuch thatand. Obviously X and Y are independent under, that is, . Letbe constructed based onas follows.
Remove an arbitrarily small quantitywhereaway from each of the four positive probabilities inso each becomesfor all, such thatand. Extend the range oftoand, and allocate the mass ε over the extended range according to
where c is such that. Under the constructed, for any, X and Y are not independent, and the corresponding mutual information is
However, noting that as, and hence,
All things considered, it is therefore desirable to have a mutual information measure, say , or for that matter a family of mutual information measures indexed by a positive integer n, such that for all distributions in , and with an accordingly defined standardized mutual information measure such that the utility of Theorem 1 is preserved with in place of for all distributions in .
3. Main Results
Given and , consider the experiment of drawing an identically and independently distributed () sample of size n. Let denote the event that all observations of the sample take on a same letter in , and let be referred to as the event of total collision. The conditional probability, given , that the total collision occurs at letter is
It is clear that is a probability distribution induced from . For each n, of (1) is the conditional distribution of total collision (CDOTC) with n particles.
Remark 2.
It is to be noted that, given a, of (1) is a special member of the family of the escort distributions introduced by [14]. The escort distributions are a useful tool in thermodynamics. Interested readers may refer to [15] for a concise introduction.
Lemma 1.
For each n, , anduniquely determine each other.
Proof.
The lemma follows. □
Lemma 2.
For each n, , and for any, .
Proof.
Write . Noting , and therefore for all ,
The lemma follows. □
On the joint alphabet with distribution , consider the associated CDOTC for an n and all pairs such that and ,
Let . It is to be noted that . The two marginal distributions of (3) are and , respectively, where
Lemma 3.
if and only if.
Proof.
For each positive integer n, if for all pairs , and , then
where the two factors of the last expression above are respectively and , , , are letter values of the n observations in the sample.
The lemma immediately follows the factorization theorem. □
For each , let , and be Shannon’s entropies defined with the joint CDOTC, as in (3), and the marginal distributions and as in (4) and (5), respectively. Let
Theorem 2.
For everyand any,
- ,
- if and only X and Y are independent.
Proof.
In Part 1, , since is a mutual information and by Lemma 2. Part 2 follows Lemma 3 and the fact that is a mutual information. □
Let
be referred to as the nth order standardized mutual information, and write . Let be a pair of random elements on according to the induced joint distribution with index value .
Lemma 4.
X and Y have a one-to-one correspondence if and only if and have one.
Proof.
If X and Y have a one-to-one correspondence, then for each i, there is a unique such that and for all other j, . By (3), and for all other j, . That is, and have a one-to-one correspondence.
Conversely, if and have a one-to-one correspondence, then for each i, there is a unique such that and for all other j, . On the other hand, by (2),
it follows that and for all other j, . That is, X and Y have a one-to-one correspondence. □
Corollary 1.
For everyand any,
- ,
- if and only if X and Y are independent, and
- if and only if X and Y are one-to-one corresponded.
Proof.
By Lemma 3, X and Y are independent if and only if and are. By Lemma 4, X and Y are one-to-one corresponded if and only if and are. The statement of Corollary 1 follows directly from Theorem 1. □
Theorem 2 and Corollary 1 together fill the void in left behind by .
4. Concluding Remarks
The main results of this article may be summarized as follows. A family of generalized mutual information indexed by a positive integer n is proposed. The member corresponding to is Shannon’s mutual information for a given joint distribution, . The other members of the family correspond to other integer values of n. They are also Shannon’s information defined, not with , but with induced distributions based on the given distribution . These induced distributions are called conditional distributions of total collision (CDOTC), which collectively is a special subset of a more general family called the escort distributions, which is often studied in extensive thermodynamics. The main motivation of the generalized mutual information is to resolve the issue of the fact that the standard mutual information is not finitely defined for all distributions of a countable joint alphabet , which leads to the issue of mutual information’s utility only realized on a fraction of .
On a more specific and finer level, the following facts are established.
- There is a one-to-one correspondence between each CDOTC and the given distribution on a countable alphabet, and hence each CDOTC is a characteristic representation of the original distribution . One of the implications of this fact is that understanding the underlying is equivalent to understanding one of its CDOTC. It can be shown that the CDOTC with an order greater than 1 is much easier to estimate than with sparse data.
- Each generalized mutual information is guaranteed to be finite. This result essentially guarantees the validity of statistically testing the null hypothesis of independence of two discrete random elements, as it guarantees the existence of (generalized) mutual information anywhere in the alternative space of dependent join distributions.
- It is shown that a particular form of standardized mutual information , defined with any CDOTC of any order greater than 1, preserves the zero-to-one scale with independence on one end and total dependence on the other, which is enjoyed by Shannon’s entropy only when it is finite.
In short, the family of conditional distributions of total collision embeds the underlying probability distribution as a special member, and the family of generalized mutual information embeds Shannon’s mutual information as a special member. Consequently, the stochastic association on joint alphabets can be measured by not only one index but by a host of indices, which collectively offer a much extended space to study stochastic dependence in information theory.
Funding
This research received no external funding
Conflicts of Interest
The author declares no conflict of interest.
References
- Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef]
- Zhang, Z. Statistical Implications of Turing’s Formula; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2017. [Google Scholar]
- Csiszár, I. Axiomatic characterizations of information measures. Entropy 2008, 10, 261–273. [Google Scholar] [CrossRef]
- Amigó, J.M.; Balogh, S.G.; Hernández, S. A brief review of generalized entropies. Entropy 2018, 20, 813. [Google Scholar] [CrossRef]
- Khinchin, A.I. Mathematical Foundations of Information Theory; Dover: New York, NY, USA, 1957. [Google Scholar]
- Chakrabarti, C.G.; Chakrabarty, I. Shannon entropy: Axiomatic characterization and application. Int. J. Math. Math. Sci. 2005, 17, 2847–2854. [Google Scholar] [CrossRef]
- Rényi, A. On measures of information and entropy. In Volume 1: Contributions to the Theory of Statistics, Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 20 June–30 July 1960; University of California Press: Berkeley, CA, USA, 1961; pp. 547–561. [Google Scholar]
- Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
- Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
- Kvalseth, T.O. Entropy and correlation: Some comments. IEEE Trans. Syst. Man Cybern. 1987, 17, 517–519. [Google Scholar] [CrossRef]
- Strehl, A.; Ghosh, J. Cluster ensembles—A knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 2002, 3, 583–617. [Google Scholar]
- Yao, Y.Y. Information-theoretical measures for knowledge discovery and data mining. In Entropy Measures, Maximum Entropy Principle and Emerging Applications, 1st ed.; Karmeshu, Ed.; Springer: Berlin, Germany, 2003; pp. 115–136. [Google Scholar]
- Vinh, N.X.; Epps, J.; Bailey, J. Information theoretical measures for clustering comparison: Variants, properties, normalization and correlation for chance. J. Mach. Learn. Res. 2018, 11, 2837–2854. [Google Scholar]
- Beck, C.; Schlögl, F. Thermodynamics of Chaotic Systems: An Introduction; Cambridge University Press: Cambridge, UK, 1993. [Google Scholar]
- Matsuzoe, H. A sequence of escort distributions and generalizations of expectations on q-exponential family. Entropy 2017, 19, 7. [Google Scholar] [CrossRef]
© 2020 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).