The Structure of the Class of Maximum Tsallis–havrda–chavát Entropy Copulas

A maximum entropy copula is the copula associated with the joint distribution, with prescribed marginal distributions on [0, 1], which maximizes the Tsallis–Havrda–Chavát entropy with q = 2. We find necessary and sufficient conditions for each maximum entropy copula to be a copula in the class introduced in Rodríguez-Lallena and Úbeda-Flores (2004), and we also show that each copula in that class is a maximum entropy copula.


Introduction
In thermodynamics, entropy is a measure of randomness or disorder, and in statistics, entropy is a measure of uncertainty in a probability distribution.During the past one and a half centuries, a number of such measures have been proposed.In terms of a continuous bivariate distribution with density f (x, y), three well known entropies are the following: B( f ) = − f (x, y) ln( f (x, y))dxdy [Boltzmann-Gibbs-Shannon (BGS)]; R q ( f ) = 1 1 − q log 2 f (x, y) q dxdy , q ≥ 0, q = 1 [Rényi]; T q ( f ) = 1 q − 1 1 − f (x, y) q dxdy , q ≥ 0, q = 1 [Tsallis-Havrda-Chavát (THC)].
These measures are related to one another.For example, the limit as q approaches 1 of the THC entropy (see [1]) is the BGS entropy; and the THC and Rényi entropies for a common value of q are monotone functions of each other.
One objective in the study of entropy is to identify the joint distributions that maximize entropy subject to given margins, and to identify the copulas associated with those distributions.In [2] Pougaza and Mohammad-Djafari use Lagrange multipliers to show that when the support of f (x, y) is I 2 , I = [0, 1], the joint probability density function that maximizes BGS entropy is f (x, y) = f 1 (x) f 2 (y), the product of the two marginal probability density functions.Thus, in the case of BGS entropy, every joint distribution maximizing BGS entropy has the same copula, namely the independence copula.
Pougaza and Mohammad-Djafari [2] also identified the joint distributions that maximize THC entropy in the case q = 2 (a case of interest since it coincides with Simpson's diversity index [3]).
In [4] the authors construct several families of copulas by considering the THC entropy with index q = 2 and particular expressions for the marginals.Our goal in this paper is to study the copulas associated with those joint distributions.We show that the family of maximum entropy copulas (in the THC q = 2 sense) coincides with the family of copulas studied previously by Rodríguez-Lallena and Úbeda-Flores [5].In practical terms, this result guarantees that researchers using maximum entropy copulas (in the THC q = 2 sense) can find in [5] all the relevant properties of the copulas in [2].
We proceed as follows.In the next section we review the maximization of THC entropy in the case q = 2. Since not every solution to the Lagrange multiplier optimization in [2] yields a proper bivariate distribution, in Section 3 we establish necessary and sufficient conditions for the solutions to be proper bivariate distributions and identify the associated copulas.We illustrate the results with several examples.

Preliminaries
Consider the bivariate probability density function f (x, y) with (x, y) ∈ I 2 , I=[0, 1], which maximizes the THC entropy T 2 ( f ), given by with the following constraints We use the notations T 2 ( f ) and T 2 (F) to refer to the THC ( [1]) entropy of the density f and the entropy of the cumulative distribution F, respectively.
We note that distributions with the same copula but different marginals (hence different joint densities) may have different entropies (1).For example, consider the following cases, all with the product copula and different margins: According to [2] for absolutely continuous distribution functions F 1 (x) and F 2 (y), the joint density and distribution functions f (x, y) and F(x, y) on I 2 with maximum entropy given by Equation ( 1), under the constraints ( 2 for appropriate functions F 1 and F 2 , and with a maximum entropy (ME) copula Not every choice of cumulative distribution functions F 1 and F 2 in (5) yields a bivariate distribution function F. For example, let F 1 (x) = x 3 and F 2 (y) = y 3 .Then f (x, y) = 3x 2 + 3y 2 − 1, which equals − 1 3 when x = y = 1 3 .Since the only maximum entropy copula for BGS entropy is the product copula, the entropy (1) identifies models of greater flexibility in terms of dependence structure than models derived from BGS entropy.

Results
We initiate this section with a result, showing that every ME copula (Equation ( 6)) has the form of a RU copula (Equation ( 9)).Since each inverse cumulative distribution function F −1 i is absolutely continuous and nondecreasing on its domain I, the inverse F i of F −1 i is an absolutely continuous distribution function on I.We find necessary and sufficient conditions on the functions F 1 and F 2 (in terms of the derivatives of their inverses) for the ME copula in Equation (6) to be a RU copula.These conditions also insure that the functions F 1 and F 2 yield a proper joint distribution function in (5).

Theorem 1.
i Every ME copula has the functional form of a RU copula.ii Let F −1 1 (u) and F −1 2 (v) be inverse distribution functions on I, and set A = {u ∈ I :

Proof.
i Equation ( 6) is equivalent to which has the form of a RU copula with
We now state and prove a converse of Theorem 1 by showing that every RU copula is a ME copula.Theorem 2. Let C(u, v) = uv + f (u)g(v) be a RU copula.Then there exists a constant k > 0 such that setting F −1 Proof.Let C(u, v) = uv + f (u)g(v) be a RU copula.We need to show that there exists a k > 0 such that −y , and the maximum entropy bivariate cumulative distribution function F(x, y) with margins As a point of comparison, in the independence case we have The joint distribution F given by Equation ( 5) for a = 0, and T 2 (F) = 0 for a = 0, where .
The next example is an application to real data.We use a Farlie-Gumbel-Morgenstern copula to model the dependence between two indicators, under this model we test the independence assumption between such indicators, since it is the solution indicated by the BGS entropy.The probability density function in the Farlie-Gumbel-Morgenstern copula case is c(u . The Farlie-Gumbel-Morgenstern copulas are solutions that maximize THC entropy in certain cases (see Example 4).Because we use a data set with moderate sample size, to test the independence, we apply the Full Bayesian Significance Test (FBST) [7], a Bayesian approach especially designed to test precise hypothesis.The aim of this application is to use the family of ME (or RU) copulas derived from THC entropy in Example 4 to test the hypothesis of the independence assumption.
Example 5.The database is composed by paired observations {(x i , y i )} n i=1 coming from the bivariate vector (X, Y).The observations are taken from each one of the 50 American states and the District of Columbia.The observations of X are "Gonorrhea Rates per 100,000 Population" collected for each of the n=51 units by the Centers for Disease Control and Prevention in 2011.The data can be obtained from Table 14 in [8].See also [9].The observations of Y are records of poverty "400% FPL and over" records the upper section of the distribution of the total population by Federal Poverty Level (FPL) for the 51 units, in 2011.The data can be obtained from the Kaiser Family Foundation [10].The hypothesis of independence in the case of Farlie-Gumbel-Morgenstern copula is translated to test θ = 0, which leads naturally to apply the procedure FBST as follows: H 0 : θ = 0 versus H 1 : θ = 0, compute Ev = 1 − T g(θ|D)dθ, where T = {θ : −1 ≤ θ ≤ 1 and g(θ|D) > g(0|D)} n , i = 1, . . ., n, with posterior density on θ given by g(θ|D), θ ∈ [−1, 1].In FBST, H 0 is rejected when Ev is small, for details about this methodology see [11].In order to explore some possibilities we compare the results using a prior density on θ proportional to ( 1+θ 2 ) a−1 ( 1−θ 2 ) b−1 with a and b hyperparameters of the distribution.We fixed the hyperparameters to consider five settings (i) uniform distribution on [−1, 1], (ii) prior density with a mode on θ = 0, (iii) prior density with a mode on θ = −0.5, (iv) prior density with a mode on θ = 0.5, (v) prior density with a mode near to θ = −1.This last setting is included in concordance with the Spearman's coefficient observed in the data, ρ s = −0.32113,that indicates a θ value near to −0.96.
According to our results (Table 1), we have evidence to reject the independence assumption under the Farlie-Gumbel-Morgenstern copula model, except in a setting as (iv), in which the prior distribution on θ and the likelihood of the data expose some conflict of information, allowing us to disregard this case.

Conclusions
In [2], Pougaza and Mohammad-Djafari studied bivariate distributions with maximum entropy (in the Tsallis-Havrda-Chavát sense with q = 2) for the purpose of generating new families of copulas, which we call ME copulas.In this paper we show that every ME copula coincides with a member of a family studied by Rodríguez-Lallena and Úbeda-Flores in [5], copulas we call RU copulas.Furthermore, we show that every RU copula is a ME copula, and provide illustrative examples.