Next Article in Journal
A Methodology to Extract Knowledge from Datasets Using ML
Previous Article in Journal
TANS: A Tolerance-Aware Neighborhood Search Method for Workflow Scheduling with Uncertainties in Cloud Manufacturing
Previous Article in Special Issue
Physics-Informed Neural Networks and Fourier Methods for the Generalized Korteweg–de Vries Equation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rank-Based Family of Probability Laws for Testing Homogeneity of Variable Grouping

1
Department of Mathematics, Nova School of Science and Technology and Nova Math, Universidade Nova de Lisboa, 2829-516 Caparica, Portugal
2
Department of Higher Mathematics, Don State Technical University, Gagarin Square 1, Rostov-on-Don 344000, Russia
3
Nova Math, Universidade Nova de Lisboa, 2829-516 Caparica, Portugal
4
Department of Mathematics, Center of Mathematics and Applications, University of Beira Interior, 6201-001 Covilhã, Portugal
5
Cocoa Research Institute of Ghana, New Tafo-Akim P.O. Box 8, Ghana
6
Center of Mathematics and Applications, University of Beira Interior, 6201-001 Covilhã, Portugal
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(11), 1805; https://doi.org/10.3390/math13111805
Submission received: 24 March 2025 / Revised: 25 April 2025 / Accepted: 15 May 2025 / Published: 28 May 2025
(This article belongs to the Special Issue Asymptotic Analysis and Applications)

Abstract

:
In order to test within-group homogeneity for numerical or ordinal variable groupings, we have introduced a family of discrete probability distributions, related to the Gini mean difference, that we now study in a deeper way. A member of such a family is the law of a statistic that operates on the ranks of the values of the random variables by considering the sums of the inter-subgroups ranks of the variable grouping. Being so, a law of the family depends on several parameters such as the cardinal of the group of variables, the number of subgroups of the grouping of variables, and the cardinals of the subgroups of the grouping. The exact distribution of a law of the family faces computational challenges even for moderate values of the cardinal of the whole set of variables. Motivated by this challenge, we show that an asymptotic result allowing approximate quantile values is not possible based on the hypothesis observed in particular cases. Consequently, we propose two methodologies to deal with finite approximations for large values of the parameters. We address, in some particular cases, the quality of the distributional approximation provided by a possible finite approximation. With the purpose of illustrating the usefulness of the grouping laws, we present an application to an example of within-group homogeneity grouping analysis to a grouping originated from a clustering technique applied to cocoa breeding experiment data. The analysis brings to light the homogeneity of production output variables in one specific type of soil.

1. Introduction

In this work, taking as a starting point a statistic introduced in [1], we formally define and study a family of discrete finite probability laws that are useful to test the within-group homogeneity of groupings—or clusters—of observations for a finite set of random variables.
We call an element of this family of probability laws a grouping distribution. A grouping distribution, of the family we are interested in, is related to the Gini’s mean difference (for information on this statistic, see [2,3,4,5]). It has also some affinity with the distribution used in the Kruskal–Wallis test (see [6]) but, while this test compares ranks across multiple groups but focuses on the between-group variance of ranks, the grouping distribution focuses on the within-group dispersion.
The most well known homogeneity test is based on the Chi-square statistic and is applied to samples of numerical variables to determine if the samples originate from the same distribution. In this sense, homogeneity is an intergroup property, that is, the groups are homogeneous to one another, or not. In the same vein, the work [7] studies a family of nonparametric tests to decide if two independent samples have a common distribution with the same density; the tests rest on a family of distances between densities. In this work, we consider “homogeneity” in the sense of the homogeneity of elements within the groups; the principle is that, using some criteria, we first partition the data in groups, and then we test if the partition criteria give rise to groups which are homogeneous by themselves, or not.
In our approach, a grouping distribution is applied to a finite set of ranks of observations of random variables and so, as a consequence, a fruitful study of the grouping statistic law seems to rely heavily on combinatorics. To the best of our knowledge, the distribution of a random variable with the grouping probability law can only be determined with computational help due to the large number of events.
Among the tools available for the study of probability laws, the transform methods such as Laplace, Fourier or z-transform, for instance, are usually effective. The study of the Fourier transform by means of signal theory does not appear to be useful since the patterns observed in the graphical representations of the Fourier transforms of the laws result from an artifact—the French battement—resulting from the interference of trigonometric waves with frequencies close to one another.
The idea of considering semigroups generated either by the values taken by the distributions of the probabilities of the values can be motivated by the description of sums of independent random variables having a common grouping law. We will deal with this idea in another article.
Let us briefly describe the contents and main contributions of this work.
  • In Section 2, we define the family of grouping distributions—which, in our opinion, should be denominated Mexia distributions, honoring the statistician extraordinaire João Tiago Mexia who proposed them for the first time—we detail some properties of the family and present tables of quantiles of the grouping distributions for moderate values of parameters computed from the full discrete distribution, and we state some remarks that highlight the motivation for the problems studied in the following.
  • In Section 3, we present some results on moment generating functions that may have an independent interest, namely, Theorem 3, showing, under some general hypothesis, that the Kolmogorov distance between probability distributions is bounded by the L 1 distance between the correspondent moment generating functions. We prove a negative asymptotic result based on observed properties of some fully computable grouping distributions and we propose two methodologies to build approximate laws for grouping distributions with large parameter values; these methodologies may be considered implementation open problems due to technical difficulties.
  • In Section 4, we present an application of a grouping distribution to assess the within-group homogeneity of clusters of observations from a cocoa breeding experiment.
    The clusters are obtained by the well-known KMeans clustering technique applied to the whole set of values of the variables. We show that the use of the grouping distribution allows to recover some important features of the observations, namely, the homogeneity of production variables in one type of soil that other statistical methods—such as ANOVA—do not bring to light.
The figures and the computations are performed using Wolfram’s Mathematica ™.

2. The Definition of the Grouping Probability Law

In this section, we define—in a constructive way, that is, by giving the values and the algorithm to obtain the correspondent probabilities—the family of the grouping probability laws.
Definition 1
(The family of grouping probability laws). The grouping test statistic law—denoted by G ( N , p , n 1 , , n p ) —is the probability law defined by the following:
1. 
The set of integers S = { 1 , 2 , , N } ;
2. 
The set S p of all partitions of S in p subsets { S 1 , , S p } such that
# S k = n k , k = 1 , , p   a n d   n 1 + + n p = N .
3. 
The values of the following function Q G , defined on S p :
Q G ( S 1 , , S p ) = k = 1 p i , j S k i j , { S 1 , , S p } S p ,
from which we can obtain the corresponding frequencies of those values.
Let us detail some important observations about this family of probability laws. Firstly, it is a useful family for a statistical diagnostic of lack of homogeneity in groups; an example of such a statistical application is presented in Section 4. Secondly, the actual computation of values and respective frequencies is a challenge even for moderate values of N, say, for example N = 27 ; in fact, for N = 27 , p = 2 , n 1 = 15 , and n 2 = 12 , we have that # S 2 = 17,383,860, a number that can bring a computational challenge. Let us look first at some aspects of the possible statistical applications.

2.1. Statistical Relevance of Grouping Distributions

We first observe an elementary fact that is relevant for the control of any algorithm computing the law of a grouping distribution in the sense that it allows to estimate some of the worst case scenarios.
Remark 1
(The largest value for the partitions). Since the number of ways of depositing N distinct objects into p distinct bins, with n 1 objects in the first bin, n 2 objects in the second bin, and so on, is given by the multinomial coefficient and since n 1 + + n p = N , we obtain that
# S p = N ! n 1 ! n p ! .
and so # S p takes one of its largest values whenever we have
n k N p , k = 1 , , p .
See Appendix A.4 for an idea of a proof.
Remark 2
(Notation and example). We write X G ( N , p , n 1 , , n p ) to denote a random variable X that follows the law G ( N , p , n 1 , , n p ) , that is, X takes the values taken by Q G , defined in Formula (1), with the corresponding frequencies or probabilities. Let us consider the law G ( 9 , 2 , 4 , 5 ) . The number of possible partitions is equal, according to Formula (2), to 126. But the set of values taken by Q G ( 9 , 2 , 4 , 5 ) has only 16 different values; these values, together with the corresponding probabilities or frequencies of occurrence, are displayed in Table 1.
Neither the values nor the frequencies of occurrence of these values, in Table 1, convey a discernible pattern of regularities allowing for an a priori description of values and corresponding probabilities for the law G ( 9 , 2 , 4 , 5 ) . The quest for regularities allowing such an a priori description is a very interesting theme of research not considered in this work.
Remark 3
(On the statistical use of the grouping law). Given a collection of N random variables with completely ordered values—for instance, numerical random variables—it may be advisable to consider the partition of the N random variables in p groups of these random variables; the reason may be multifold: common origins, common data collecting structure, common structural characteristics, or even as a result of some clustering process. In order to test the statistical significance of the within-group homogeneity of the grouping, we consider, for each group, the global ranks of each of the variables of the group; such an attribution of ranks is possible since the variables have completely ordered values. This grouping procedure defines an initial family { S 1 , , S p } of subsets of S = { 1 , 2 , , N } —such that # S i = n i —for which the function Q G ( S 1 , , S p ) has a determined value. The grouping will have statistical significance if the value of Q G ( S 1 , , S p ) is small enough with respect to all the other possible values of the function Q G taken on all possible partitions of S = { 1 , 2 , , N } in different p subsets { S 1 , , S p } such that # S i = # S i = n i . Once we have the law G ( N , p , n 1 , , n p ) we can define lower quantiles of this law and perform the adequate statistical test.
Let us look at two tables of quantiles q α Table 2 and Table 3—of the grouping distributions G ( N , 2 , r , N r ) , for N = 8 , , 19 and r = 2 , , N 2 , observing that due to the fact that there are only two subgroups, p = 2 , there is a symmetry, allowing us to reduce the number of cases presented. In this case, a statistical analysis focuses on testing the homogeneity of p = 2 groups.
With the quantile Table 2 and Table 3, the statistical test may be performed in the following way. Consider a set of values—corresponding to the observation of random variables—a set with a number of elements between 8 and 19, let us say, for example, N = 13 elements; this set can be decomposed into two groups corresponding to ( n 1 , n 2 ) such that n 1 + n 2 = N as follows: (2,11), (3,10), (4,9), (5,8) and (6,7). Suppose we have to test the homogeneity of a decomposition of the set of 13 values in two groups, for example, the decomposition (4,9), that is, in one subgroup with 4 values and the other subgroup with 9 values. In Table 3, we obtain the quantiles q 0.05 = 332 and q 0.1 = 356. We then order the 13 values and we consider the ranks of these values. Considering the subset of ranks S 1 corresponding to the subgroup with 4 values and the subset of ranks S 2 of the subgroup with 9 values, we can compute the value of the statistic Q G ( S 1 , S 2 ) defined in Formula (1) and compare it with a quantile, say q 0.05 or q 0.1 , for some instances. Since the more significant one is the within-group homogeneity of the grouping, the lesser one is the value of the statistic Q G ( S 1 , S 2 ) ; an observation of a statistic value smaller that one of the chosen quantiles will provide indication of the within-group homogeneity.

2.2. Computational Issues for the Groupings Statistic Family

We first observe that the number of ways to partition a set of N objects into p non-empty subsets without the restriction of having a determined number of elements in each subset as in Definition 1 and Remark 1—a number denominated Stirling number of the second kind, or Stirling partition number—is given by
S ( N , p ) = 1 p ! i = 0 p ( 1 ) p i p i i N = i = 0 p ( 1 ) p i i N ( p i ) ! i !
for N 1 and p N (see [8], p. 258) for substantial information on this subject). It is clear that for N 2 , we have S ( N , 2 ) = 2 N 1 1 , and so we have exponential growth, with N, of the number of evaluations, for ( S 1 , S 2 ) , of the statistic Q G ( S 1 , S 2 ) for the family of probability laws G ( N , 2 , r , N r ) when r varies.
Moreover, summing S ( N , p ) in Formula (3), over all possible p, we obtain the total number of partitions of N, that is, the Bell number, a number that grows super exponentially (see [8], pp. 374, 493, 603).
This observation suggest two approaches. A first approach is to have at our disposal a set of asymptotic results allowing to approach the laws G ( N , p , n 1 , , n p ) , for large values of N, more easily computable distributions. We present a negative result in Section 3 that leads to the proposal of two methodologies to obtain approximations for the grouping laws for large parameter values. A second approach consists of studying the structure of the distributions G ( N , p , n 1 , , n p ) in order to find regularities allowing for amenable computations; we defer to a posterior publication for some preliminary results with this approach, where we use elementary numerical semigroup facts to study convolutions of grouping laws.

3. On an Asymptotic Result for Some Grouping Distributions

Suppose that we have a random variable X with G ( N , p , n 1 , , n p ) grouping distribution taking the values α i with probabilities p i with i = 1 , , m . We know that the higher numbers of different configurations of groupings occur when n j N / p for all j = 1 , , p . In this case, it is to be expected that both m, that is, the number of different values taken by X—denoted by α i —and, simultaneously, these values α i taken by X, will all be large, while, concurrently, the values of the probabilities p i such that p 1 = P [ X = α i ] will be uniformly small. In the case of N and p being simultaneously large, the computational burden of computing the discrete finite valued distribution is excessive, and so it would be useful to have an asymptotic result for the X grouping distribution.
Let us firstly detail some of the challenges we face in the quest for an asymptotic result. We take as an example a random variable X with the grouping distribution G ( 23 , 2 , 11 , 12 ) ; we first consider the values taken by X and the respective probabilities of occurrence and, from those, we obtain the probability law of the corresponding standardized random variable Y from which we can derive the empirical distribution. The PDF of the empirical distribution of Y is depicted in blue color in Figure 1.
We also fitted the discrete data ( x , ln ( pdf Y ( x ) ) of the logarithm of pdf Y ( x ) , the PDF function of Y, by a function of the form f ( x ) = a + b ( x c ) + d ( x e ) 2 , thus obtaining, upon the exponentiation of f ( x ) and the normalization of e f ( x ) , a fitting by a normally distributed random variable of the standardized data of X. The PDF of the normal distribution fitting the PDF of the empirical distribution of Y is depicted in red color also in Figure 1. We observe that the PDF in blue is more akin to a trajectory of a stochastic process having as a mean the PDF in red. The lower irregularity in the left tail of the blue PDF is why, in our perspective, we can only expect an asymptotic normality result that can give acceptable results only for the lower quantiles of the law G ( N , p , n 1 , , n p ) .

3.1. Some Auxiliary Results on Moment Generating Functions

We first prove an auxiliary result—Theorem 3 ahead—that shows that if we have some control over the behavior of a sequence of moment generating functions, then we also have control over the sequence of the correspondent distribution functions. For that purpose, in the sequel, we consider the following setting of main hypothesis.
We have probability laws μ and ( ν n ) n 1 with respective densities f μ ( x ) and ( f ν n ( x ) ) n 1 and the correspondent moment generating functions (MGFs) defined by
M μ ( t ) = e t x f μ ( x ) d x , M ν n ( t ) = e t x f ν n ( x ) d x .
Let us suppose that the following holds:
  • The densities f μ ( x ) and f ν n ( x ) have exponential decay that is
    λ 1 , λ 2 > 0 , n 1 , | f μ ( x ) | λ 1 e λ 2 | x | , | f ν n ( x ) | λ 1 e λ 2 | x | .
    This is equivalent to the fact that the interior of the sets D μ : = { t R : M μ ( t ) < + } and D ν n : = { t R : M ν n ( t ) < + } —that is, respectively, D μ ˚ and D ν n ˚ —are non-empty sets (see Proposition 5.1 in [9]). The sets D μ and D ν are, in fact, intervals.
  • Also as a consequence, we have that M μ ( σ + i ω ) and M ν n ( σ + i ω ) are holomorphic (analytic) in a strip of the complex plane defined by σ D μ ˚ D ν n ˚ (see Proposition 5.3 in [9]).
  • For every σ + i ω in the complex plane strip defined by σ D μ ˚ D ν n ˚ , we have that
    lim n + M ν n ( σ + i ω ) = M μ ( σ + i ω ) .
  • There exists an integrable function G in the variable ω such that, for every σ + i ω in the complex plane strip defined by σ D μ ˚ D ν n ˚ , we have
    | M μ ( σ + i ω ) M ν n ( σ + i ω ) | G ( σ + i ω ) .
  • In order to address the inversion formula, we recall that, by the definition in Formula (4), we can extend the MGF to a strip t = σ i ω in the complex plane, with M μ ( t ) defined, for σ D μ ˚ D ν n ˚ , by
    M μ ( t ) = M μ ( σ i ω ) = e σ x f μ ( x ) e i ω x d x ,
    since in the right-hand side we always have an integrable function. Formula (6) shows that the extension of the MGF to the strip defined by S : = { σ D μ ˚ D ν n ˚ } can be considered a family of Fourier transforms, namely, the Fourier transforms of the family ( e σ x f μ ( x ) ) σ S . Now by the standard results on the inversion of the Fourier transform—or the characteristic function—we have, for instance, that if for all σ S the function M μ ( σ i ω ) is integrable, then by the inversion theorem (see [10], p. 185 or [11], p. 126), we obtain that almost surely in x,
    e σ x f μ ( x ) = 1 2 π e i ω x M μ ( σ i ω ) d ω ,
    that is,
    f μ ( x ) = 1 2 π e ( σ i ω ) x M μ ( σ i ω ) d ω = 1 2 π lim A + A A e ( σ i ω ) x M μ ( σ i ω ) d ω = : 1 2 i π σ i σ + i e z x M μ ( z ) d z ,
    with z = σ i ω S , where the right-hand expression is the usual one for the inversion of the Laplace transform (see [12], pp. 60–66), with the integral being taken in the principal value sense. We have similar expressions for all f ν n ( x ) . The inversion of the MGF in Formula (7) may hold with a more general hypothesis. It must be noticed that by reason of the Cauchy theorem on contour integrals, the integral in the right-hand side of Formula (7) does not depend on σ (see [13], p. 226).
We secondly recall two important results needed in the sequel allowing the determination of MGF. The first guarantees the existence of a limit distribution from the convergence of a sequence of moments (see [14], p. 145).
Theorem 1
(Weakly convergence from convergence of moments). Let ( ν n ) n 1 be a sequence of probability measures with all elements of the sequence admitting a finite sequence ( M n k ) n 1 , k 0 of raw moments such that, for all k 0 , there exists M k R such that
lim n + M n k = lim n + R x k d ν n ( x ) = M k ,
and moreover for some t > 0 ,
k = 0 + | M k | t k k ! < + .
Then, there exists a probability measure μ having as raw moments the sequence ( M k ) k 0 and such that the sequence ( ν n ) n 1 converges weakly to μ.
Proof. 
Due to the importance of this result, we present a proof that differs from the proof suggested in ([14], p. 145). The existence of μ having ( M k ) k 0 as its moment sequence is a consequence of the Hamburger’s moment theorem. In fact, we consider an arbitrary sequence of real numbers ( ξ n ) n 0 and we observe that for all n 0 ,
k = 0 , l = 0 n ξ k ξ l M k + l = k = 0 , l = 0 n ξ k ξ l lim n + R x k + l d ν n ( x ) = lim n + k = 0 , l = 0 n ξ k ξ l R x k + l d ν n ( x ) = = lim n + R k = 0 , l = 0 n ξ k ξ l x k + l d ν n ( x ) = lim n + R | k = 0 n ξ k x k | 2 d ν n ( x ) 0 ,
thus showing that the sequence ( M k ) k 0 is positive semidefinite. By Hamburger’s theorem, there exists a positive measure μ over the reals, having the sequence ( M k ) k 0 as its moment sequence (see [15], p. 63). Moreover, the measure μ is a probability measure since
M 0 = R d μ ( x ) = lim n + R d ν n ( x ) = 1 ,
using the fact that all measures ν n are probability measures. Now the condition in Formula (8) shows that M μ , that is, the MGF of the probability measure μ is defined in an interval around zero I with a non-empty interior. As a consequence, for t in a compact interval around zero with a non-empty interior J I , we have that
M μ ( t ) = k = 0 + M k t k k ! = k = 0 + R x k d μ ( x ) t k k ! = k = 0 + lim n + R x k d ν n ( x ) t k k ! = = lim n + k = 0 + R x k d ν n ( x ) t k k ! = lim n + k = 0 + M n k t k k ! = lim n + M ν n ( t ) ,
since the convergence of the series of moments defining M μ ( t ) is uniform in J by reason of the condition in Formula (8). Now, by a well-known result (see again [14], p. 145), since the sequence of moment generating functions ( M ν n ( t ) ) n 1 converges to M μ ( t ) for t J , which is a non-degenerate interval around zero, then the sequence ( ν n ) n 1 converges weakly to μ , and the theorem is proved. □
The next result, a companion of the previous one, gives Carleman’s condition for the unicity of the limit distribution (see [16], p. 123, for a proof).
Theorem 2
(The Carleman sufficient condition for the unicity of the solution of the moment problem). Under the conditions and notations of Theorem 1, if we have that
k = 0 + 1 M 2 k 1 2 k = + ,
then there is at most one probability measure admitting ( M k ) n ) k 0 as its moment sequence.
The next theorem considers the case of probability laws admitting densities with respect to the Lebesgue measure over R and gives a bound on the Kolmogorov distance in terms of an integral of the difference of the MGF.
Theorem 3
(Convergence on the Kolmogorov distance from the convergence of the moment generating functions). Suppose that the hypotheses above are verified. Then, for an arbitrary σ such that σ i ω S , there exists a constant C ( σ ) > 0 such that, with F μ ( x ) = μ ] , x ] and F ν n ( x ) = ν n ] , x ] ,
sup x R | F μ ( x ) F ν n ( x ) | + | f μ ( x ) f ν n ( x ) | d x C ( σ ) | M μ ( σ i ω ) M ν n ( σ i ω ) | d ω .
Formula (11) shows that the sequence of distribution functions associated with the sequence of probability measures ( ν n ) n 1 converges in the Kolmogorov distance to the distribution function of the probability measure μ. This happens provided that the sequence of the moment generating functions of the sequence of probability measures ( ν n ) n 1 converge almost surely to the moment generating function of the probability measure μ, and the difference of the MGF is bounded by an integrable function in the imaginary variable.
Proof. 
We take the inversion formula in (7), recalling that the definition is taken in the principal value sense with σ i ω S . We also recall that the inversion formula does not depend on σ . We then have, for x > 0 and σ > 0 ,
| f μ ( x ) f ν n ( x ) | | 1 2 π e ( σ i ω ) x M μ ( σ i ω ) M ν n ( σ i ω ) d ω | = = e σ | x | 2 π | e i ω x M μ ( σ i ω ) M ν n ( σ i ω ) d ω | .
And knowing that S is symmetric around zero, for x < 0 and taking κ = σ ,
| f μ ( x ) f ν n ( x ) | e κ x 2 π | e i ω x M μ ( σ i ω ) M ν n ( σ i ω ) d ω | = = e σ | x | 2 π | e i ω x M μ ( σ i ω ) M ν n ( σ i ω ) d ω | .
Since, by Formula (5), the MGF difference is bounded by the integrable function G, we have that for all x R
| f μ ( x ) f ν n ( x ) | e σ | x | 2 π | M μ ( σ i ω ) M ν n ( σ i ω ) | d ω < + ,
and so, for a constant C ( σ ) that depends on an arbitrary σ such that σ i ω S ,
+ | f μ ( x ) f ν n | d x + e σ | x | 2 π | e i ω x M μ ( σ i ω ) M ν n ( σ i ω ) d ω | d x C ( σ ) | M μ ( σ i ω ) M ν n ( σ i ω ) | d ω .
Now, since we have that
sup x R | F μ ( x ) F ν n ( x ) | = sup x R | μ ] , x ] ν n ] , x ] | = sup x R | x f μ ( u ) d u x f ν n ( u ) d u | sup x R x | f μ ( u ) d u f ν n ( u ) | d u = + | f μ ( u ) d u f ν n ( u ) | d u ,
by the hypotheses above and the dominated convergence Lebesgue theorem, the final statement in theorem is proved. □
Remark 4
(On the application of Theorem 3). We observe that, by an application of the Cauchy theorem, Formula (11) can be rewritten as
sup x R | F μ ( x ) F ν n ( x ) | C ( σ ) | M μ ( i ω ) M ν n ( i ω ) | d ω ,
that is, considering σ = 0 in the integral. Moreover, the bound in Formula (5) is verified if, for instance, there is c , d > 0 such that,
| M μ ( i ω ) M ν n ( i ω ) | d e c ( i ω ) 2 ,
and then the integral on the right-hand side of Formula (12) is finite. The bound in Formula (13) occurs for normal distributed random variables since, for instance, if X N ( 0 , σ ) , then M X ( t ) = e σ 2 t 2 / 2 .

3.2. A Negative Asymptotic Result on the Limit Law of Grouping Distributions

We next formulate a negative result that shows that, from a set of hypotheses that we can derive from the observation of the examples analyzed, there is not an asymptotic probability distribution for the grouping laws. Firstly as a motivation, we report some empirical observations that are relevant for the set of hypothesis underlying the stated result.
Example 1
(Observations of a particular behavior of X G ( N , p , n 1 , , n p ) ). For the case of X G ( N , 2 , n 1 , n 2 ) with N = 13 , 15 , 17 , 19 , 21 , 23 and the specific pairs ( n 1 , n 2 ) that give the maximum of the number of configurations for the partition in two set groups, we compute the ratios of the moments of the standardized random variables Y N :
E [ Y N n + 1 ] E [ Y N n ] f o r n = 2 , , 60 .
We fit logistic functions of the form given in Formula (14), to the cases of X G ( N , 2 , n 1 , n 2 ) for N = 13 , 15 , 17 , 19 , 21 , 23 ,
f L , a ( x ) = L 1 + e a x for a > 0 ,
and we obtain the results in Table 4 that show that L N / 2 as seen, also, in Figure 2.
The logistic fitting of the first absolute values of moment ratios for X G ( 23 , 2 , 11 , 12 ) —and the correspondent residuals of the fitting—can be observed, for instance, in Figure 3. The fittings—for small parameter values—give important information for the larger values of the moments since this information is extremely relevant for the the theorem on asymptotic the asymptotic behaviour.
As already pointed out, we observe that the variation of L with N represented in the Figure 2 corresponds to a linear fitting with L, verifying:
L N 2 .
The observations reported in Example 1 are reflected in Hypothesis B—by taking into consideration the results in Table 4—and also in Hypothesis C, both formulated in the following. We stress that Hypothesis D is also a reflection of observed facts although these fact are not explicitly reported in this paper. It should be interesting to have a formal proof for all possible cases of the law of X G ( N , p , n 1 , , n p ) .
Hypothesis 1
(Observed hypothesis on X G ( N , p , n 1 , , n p ) for large N and p). Consider ( Y m ) m 1 to be the sequence of standardized versions of ( X m ) m 1 , a sequence of random variables with grouping probability laws denoted by G ( N m , p m , n m , 1 , , n m , p m ) verifying the following hypothesis:
A. 
lim m + N m = + .
B. 
For all m 1 , there exists a sequence ( ϵ n m ) n 1 such that, with lim n + ϵ n m = 0 , for n 2 we have
E [ Y m n + 1 ] E [ Y m n ] = p m N m + ϵ n m .
C. 
For m 1 , there exists c ] 0 , 1 [ such that
lim n , m + p m N m + ϵ n m = c .
D. 
We finally suppose that for all m and n 3 ,
E [ Y m n ] = ( 1 ) n E [ Y m n ] .
  • These hypothesis are verified in the cases analyzed and referred to in Example 1.
If we assume the set of conditions in Hypothesis 1, we have the following negative result.
Theorem 4
(On the asymptotic behavior of X G ( N , p , n 1 , , n p ) for large N and p). If we have that
lim m + p m N m = c , k 1 | ϵ k m | p m / N m < + , d = lim m + k = 2 + 1 + ϵ k m p m / N m ,
then there is a unique generalized function λ c d ( x ) —that is not a probability measure—giving rise to a moment generating function defined by
M λ c d ( t ) : = λ c d ( x ) , e t x = 1 + d c 2 e c t 1 + c t + ( 1 d ) t 2 2 ,
for t in a non-empty interval centered at zero.
Proof. 
We will show that a contradiction comes from the hypothesis in Formula (16). Recall that since the random variables Y m are standardized and since we are interested in the asymptotic behavior, we consider that, by Hypothesis B, we have for n 2 and m 1 ,
| E [ Y m n + 1 ] | = k = 2 n p m N m + ϵ k m E [ Y m 2 ] = k = 2 n p m N m + ϵ k m = = p m N m n 1 k = 2 n 1 + ϵ k m p m / N m .
Now, using the hypothesis in Formula (16) and considering for d m > 0 a constant depending on m such that
d m = lim n + k = 2 n 1 + ϵ k m p m / N m < + ,
that is, originated from the product limit in Formula (18), we have that
n 2 , | E [ Y m n + 1 ] | d m c m n 1 , lim m + | E [ Y m n + 1 ] | = d c n 1 , lim m + E [ Y m n + 1 ] = d ( c ) n 1 ,
by the alternating signs of the moments in Formula (15). Since we have that, for all t R ,
d n = 3 + t n c n 2 n ! = d c 2 n = 3 + t n c n n ! = d c 2 ( e c t 1 c t ( c t ) 2 2 ) < + ,
by Theorem 1, there should exist a probability law G such that the sequence of laws given by ( G ( N m , p m , n m , 1 , , n m , p m ) ) m 1 converges weekly to the law G . Being so, if we have a random variable Y c with this law, that is Y c G , the raw moments of this random variable are given by
E Y c 0 = 1 , E Y c 1 = 0 , E Y c 2 = 1 , E Y c n = d ( c ) n 2 , n 3 .
Moreover, since the condition in Formula (10) of Theorem 2 is easily seen to be verified, there should exist only one probability distribution with this moment sequence. As a consequence, the MGF of the probability law G should be given by
M Y c ( t ) = 1 + t 2 2 + d n = 3 + ( c ) n 2 t n n ! = 1 + t 2 2 + d c 2 n = 3 + ( c t ) n n ! = 1 + d c 2 e c t 1 + c t + ( 1 d ) t 2 2 .
We are now going to show that for all c, the objects giving rise to the sequence of moments in Formula (19)—and the consequent moment generating function in Formula (17)—are not probability measures but, instead, distributions of finite order 1 and 2. These distributions of finite order 1 and 2 are Schwarz generalized functions given by derivatives of measures for which the MGF is also well defined (see again [13], p. 226). Consider δ a ( x ) to be the Dirac probability measure placed in a, and δ a ( x ) and δ a ( x ) as its first and second derivatives. We have the following two cases.
I 
If c 0 , we have that
λ c d ( x ) = 1 d c 2 δ 0 ( x ) + d c 2 δ c ( x ) d c δ 0 ( x ) + 1 d 2 δ 0 ( x ) ,
this case encompasses the extreme case c , d = 1 , where the number of groups converges to the number of elements to group, and so asymptotically, all the groups have only one element.
II 
As a consequence, if c = 0 , we have that
λ 0 ( x ) = δ 0 ( x ) + 1 2 δ 0 ( x ) ,
a case which encompasses the case where the number of groups stays fixed while the number of elements to group goes to infinity and so asymptotically at least one of the subgroups has an unbounded number of elements.
  • We observe that in both cases, the generalized function λ c d ( x ) presented—such that its MGF is given by Formula (17)—is unique. If c 0 , we want to recover Formula (17) by the rules of calculus of the generalized functions. We notice that λ c ( x ) , given by Formula (21), is a generalized function with compact support K = { 0 , c } and so, its definition domain is the test space of indefinitely differentiable functions over R . Recall that if the generalized function λ ( x ) with compact support K is given by a locally integrable function f with support K, we have that for every indefinitely differentiable function ϕ ,
λ ( x ) , ϕ ( x ) = K f ( x ) ϕ ( x ) d x .
We stress that the rules of calculus with general generalized functions may be seen as extensions of this particular case. We notice that the variable x corresponds to the integration variable. Since ϕ ( x ) = e t x is a indefinitely differentiable function of the variable x, with t being a parameter, we have that
M λ c d ( t ) = λ c d ( x ) , e t x = 1 d c 2 δ 0 ( x ) + d c 2 δ c ( x ) d c δ 0 ( x ) + 1 d 2 δ 0 ( x ) , e t x = = 1 d c 2 δ 0 ( x ) , e t x + d c 2 δ c ( x ) , e t x d c δ 0 ( x ) , e t x + 1 d 2 δ 0 ( x ) , e t x = = 1 d c 2 e t x | x = 0 + d c 2 e t x | x = c ( 1 ) d c δ 0 ( x ) , d d x e t x + ( 1 ) 2 1 d 2 δ 0 ( x ) , d 2 d x 2 e t x = = 1 d c 2 + d c 2 e c t + d c δ 0 ( x ) , t e t x + 1 d 2 δ 0 ( x ) , t 2 e t x = 1 + d e c t 1 + c t c 2 + 1 d 2 t 2 .
a result in accordance with Formula (20). If c = 0 , then Formula (17) is M λ 0 ( t ) = lim c 0 + M λ c ( t ) = 1 + t 2 / 2 , and we then have
M λ 0 ( t ) = λ 0 ( x ) , e t x = δ 0 ( x ) + 1 2 δ 0 ( x ) , e t x = δ 0 ( x ) , e t x + 1 2 δ 0 ( x ) , e t x = = 1 + 1 2 ( 1 ) 2 δ 0 ( x ) , ( e t x ) = 1 + t 2 2 .
We signal that the unicity is a consequence of a general property of the Laplace transforms of generalized functions (see [17], p. 80). We can conclude the following. Assuming Hypothesis 1 and, in particular, that lim m + p m / N m is equal to c, entails a contradiction between the consequence of Theorem 1 and the observations just made above. That is, that the object having as moments the expressions in Formula (19) is not a measure but instead a generalized function of order greater than one. □
Remark 5
(On an interpretation of the result in Theorem 4). We so have that although the set of conditions in Hypothesis 1 may be a reasonable extrapolation from the observations—in the cases where computations are feasible—these conditions entail a contradiction. As a consequence of this contradiction, these conditions are not tenable as a general hypothesis on grouping laws. A possible asymptotic result should rely on a different set of conditions.
Although, for the moment, we do not have an asymptotic result, we may nevertheless take advantage of approximate laws. The next remark shows several examples of such approximations.
Remark 6
(On the error committed by using the normal distribution quantiles). The use of the grouping distribution allows to test the significance of grouping by considering the lower quantiles of the distribution. In Table 5, we present the quantiles for X G ( N , 2 , n 1 , n 2 ) for N = 13 , 15 , 17 , 19 , 21 , 23 , the correspondent quantiles obtained by reverting the standard normal distribution with the mean and variance of the adequate G ( N , 2 , n 1 , n 2 ) , and the relative error committed by taking the normal quantiles instead of the proper distribution quantiles.
We observe that the error ϵ 0.05 for the quantile q N : 0.05 decreases with increasing N > 13 , as expected, while the error ϵ 0.1 for the quantile q N : 0.1 is always bounded by 0.2% and oscillates, which may be revealing a slower rate of convergence. A caveat for the use of the quantiles deduced from the normal approximation is that their computation requires the computation of the mean and standard deviation of the law G ( N , 2 , n 1 , n 2 ) and, for the moment, this can only be achieved exactly by the computation of the values and respective probabilities of the distribution.

3.3. Determining Approximate Laws for Grouping Distributions with Large Parameter Values

In the following, we take into consideration the approximate distributions presented in Remark 6 and a consequence of Theorem 4, namely, that it is irrelevant to assume lim m + c m = c having in view an asymptotic result. We will consider, for practical purposes, that we know a sequence of approximate values for the moments of X G ( N , p , n 1 , , n p ) for large N and p, and we will provide approximations for the correspondent laws. We stress that an adequate sequence of approximate values for the moments can be determined by a simulation procedure and, being so, does not require the determination of the law of the grouping distribution. Next, we briefly describe methodologies—see Methodologies 1 and 2—to achieve approximations to the unknown laws.
From now on, we consider the hypothesis formulated in Remark 1. And that for the random variable Y m G ( N m , p m , n m , 1 , , n m , p m ) , we have approximations to the raw moments of this random variable given by
E Y m 0 = 1 , E Y m 1 = 0 , E Y m 2 = 1 , n 2 , | E [ Y m n + 1 ] | d m n 1 c m n 1 ,
with lim n + d m n 1 = d ( c m ) > 0 . We now recall an important result, instrumental for a choice between one of two methodologies aiming at the determination of approximate laws for Y m G ( N m , p m , n m , 1 , , n m , p m ) for large N m . The result allows the possible determination of a finite support for a probability measure (see [15], p. 64).
Theorem 5
(A condition for a representing measure to have a finite support). Consider H ( n ) the Hankel matrix of order n 0 —presented in Formula (23)—of a positive semidefinite sequence ( M n ) n 0 and its determinant denoted by D e t [ H ( n ) ] :
H ( n ) : = M 0 M 1 M 2 M 3 M 4 M n M 1 M 2 M 3 M 4 M n M n + 1 M 2 M 3 M 4 M n M n + 1 M n + 2 M 3 M 4 M n M n + 1 M n + 2 M n + 3 M 4 M n M n + 1 M n + 2 M n + 3 M n + 4 M n M n + 1 M n + 2 M n + 3 M n + 4 M n M n + 1 M n + 2 M n + 3 M n + 4 M 2 n .
Then, the following properties are equivalent:
I 
The sequence ( M n ) n 0 is a moment sequence represented by a measure supported on n points.
II 
We have that
k n 1 , D e t [ H ( k ) ] > 0 ; k n , D e t [ H ( k ) ] = 0 .
Now, we consider ( E Y m n ) n 0 the sequence of raw moments of the random variable Y m . It is a positive semidefinite sequence as shown in Formula (9). In principle, we only know approximations to this sequence given by Formula (22) and we have approximate values to the corresponding Hankel matrices and to its determinants; for instance, we have that H ˜ m ( 6 ) , the approximate Hankel matrix of order 6 is given by
H ˜ m ( 6 ) 1 0 1 d m 1 ( c m ) d m 2 ( c m ) 2 d m 3 ( c m ) 3 0 1 d m 1 ( c m ) d m 2 ( c m ) 2 d m 3 ( c m ) 3 d m 4 ( c m ) 4 1 d m 1 ( c m ) d m 2 ( c m ) 2 d m 3 ( c m ) 3 d m 4 ( c m ) 4 d m 5 ( c m ) 5 d m 1 ( c m ) d m 2 ( c m ) 2 d m 3 ( c m ) 3 d m 4 ( c m ) 4 d m 5 ( c m ) 5 d m 6 ( c m ) 6 d m 2 ( c m ) 2 d m 3 ( c m ) 3 d m 4 ( c m ) 4 d m 5 ( c m ) 5 d m 6 ( c m ) 6 d m 7 ( c m ) 7 d m 3 ( c m ) 3 d m 4 ( c m ) 4 d m 5 ( c m ) 5 d m 6 ( c m ) 6 d m 7 ( c m ) 7 d m 8 ( c m ) 8 ,
with ( d m n ) n 1 being a sequence of variable correcting terms issued from Formula (18) and given by
d m n : = k = 2 n 1 + ϵ k m p m / N m .
We now propose a two-way methodology to fit a probability distribution to the approximate sequence of moments given in Formula (22) for large N and p in the grouping law G ( N , p , n 1 , , n p ) . The following methodological description may be taken as a formulation of an open technical implementation problem.
Methodology 1
(Fitting a discrete law with finite support to the approximated moment sequence). Suppose that for the sequence of the determinants of the approximated Hankel matrices ( H ˜ m ( n ) ) n 0 , we observe that the condition given in Formula (24) is verified for a given n m . We consider the problem of determining a discrete law determined by the values it takes, given by the vector A = ( α 1 , α 2 , , α n m ) t and the corresponding vector of probabilities P = ( p 1 , p 2 , , p n m ) t with moments given by the vector M = ( 1 , 0 , 1 , d m 1 ( c m ) , , d m n m 3 ( c m ) n m 3 ) t . This problem is tantamount to the problem of solving the matricial equation given by V ( A ) P = M , with V ( A ) being the Vandermonde matrix associated with the vector A, that is, the matricial equation given by
1 1 1 1 α 1 α 2 α 3 α n m α 1 2 α 2 2 α 3 2 α n m 2 α 1 3 α 2 3 α 3 3 α n m 3 α 1 4 α 2 4 α 3 4 α n m 4 α 1 n m 1 α 2 n m 1 α 3 n m 1 α n m n m 1 p 1 p 2 p 3 p 4 p 5 p n m = 1 0 1 d m 1 ( c m ) , d m 2 ( c m ) 2 , d m n m 3 ( c m ) n m 3
in the unknown vectors A and P, given the vector M. A natural assumption for the matrix problem in Formula (25) is that the components of the vector A—that is, the values taken by the random variable with the unknown law—are all distinct; this assumption entails that the determinant of the Vandermonde matrix V ( A ) is non-zero and so the matrix V ( A ) has an inverse.
It may happen that the condition in Formula (24) is not verified for the determinants of the approximated Hankel matrices ( H ˜ m ( n ) ) n 0 . In that case we have, as an alternative, to look for an absolutely continuous approximation law. This approach is detailed in Methodology 2. Again, we stress that the following methodological description may be taken as a formulation of an open technical implementation problem.
Methodology 2
(Fitting a continuous law to the approximated moment sequence). We now suppose that for the sequence of the determinants of the approximated Hankel matrices ( H ˜ m ( n ) ) n 0 , we observe that the condition given in Formula (24) is not verified. Then, having in view Proposition 3.11 in [15] (p. 65), it is natural to look for an absolutely continuous probability distribution. A first natural choice is a mixture with weights W = ( w 1 , w 2 , , w k ) t in the interval ] 0 , 1 [ of Gaussian laws N ( μ i , σ i ) for i = 1 , , k . We recall that the moments M i n of such a law are given by
M i n = k = 0 n / 2 n 2 k μ i n 2 k σ i 2 k ( 2 k ) ! 2 k k ! .
In order to determine the parameters of this mixture law, we can consider solving the matricial equation given by
1 1 1 1 M 1 1 M 2 1 M 3 1 M k 1 M 1 2 M 2 2 M 3 2 M k 2 M 1 3 M 2 3 M 3 3 M k 3 M 1 4 M 2 4 M 3 4 M k 4 M 1 k 1 M 2 k 1 M 3 k 1 M k k 1 w 1 w 2 w 3 w 4 w 5 w k = 1 0 1 d m 1 ( c m ) , d m 2 ( c m ) 2 , d m k 3 ( c m ) k 3
in the unknown vectors μ = ( μ 1 , μ 2 , , μ k ) t , σ = ( σ 1 , σ 2 , , σ k ) t and W, given the known vector of approximate moments in the right-hand side of the matricial equation. Let us suppose that ν ˜ k is an absolutely continuous probability law that is a solution to the matricial equation in Formula (26). This means that ( M ν ˜ k n ) 0 n k 1 , the sequence of the first k 1 moments of the mixture law ν ˜ k are approximately equal to the moments of the given vector M. Let us consider Formula (12) in order to estimate the Kolmogorov distance between F Y m , the distribution function of the random variable Y m , and F ν ˜ k the distribution function of the probability law ν ˜ k . We observe that since ν ˜ k is a finite mixture of Gaussian laws, the hypothesis in Formula (13) is verified. As a consequence, for all ϵ > 0 , there exist R ϵ 1 such that
sup x R | F Y m ( x ) F ν ˜ k ( x ) | C ( σ ) | M Y m ( i ω ) M ν ˜ k ( i ω ) | d ω = = C ( σ ) | ω | R ϵ | M Y m ( i ω ) M ν ˜ k ( i ω ) | d ω + | ω | > R ϵ | M Y m ( i ω ) M ν ˜ k ( i ω ) | d ω C ( σ ) 2 R ϵ sup | ω | R ϵ | M Y m ( i ω ) M ν ˜ k ( i ω ) | + ϵ .
We then have that, if as a consequence of the numerical method used to solve the matricial equation in Formula (26), we can guarantee that,
lim k + | M Y m ( i ω ) M ν ˜ k ( i ω ) | = 0 .
We will have that the sequence of laws ( ν ˜ k ) k 1 will approach G ( N m , p m , n m , 1 , , n m , p m ) —that is, the law of Y m —in the Kolmogorov distance.
Remark 7
(Further studies in asymptotic results). We intend to further develop Methodologies 1 and 2 in future work starting with a study of laws in Example 1. It is possible that an asymptotic result under a different set of more adequate hypotheses exists; if that is the case, it may be useful to use established techniques to an approach of the limit distribution by finite dimensional ones (see [18,19,20]).

4. An Application of the Grouping Statistic to a Clustering Analysis

We present in the following an application of the grouping statistic to cocoa plants data, from a agricultural station in Ghana, already analyzed in [1]. The data consist of 144 five-component vectors with the first component indicating the plant variety, the second component indicating the type of soil, and the remaining three components indicating the observed measurements of the variables plant height, stem diameter and dry matter; these variables allow the assessment of the performance of the plantation, linking it to the adequacy between soil and plant variety. The data of twelve plant varieties are in Table A1; the four soils description, allowing for the generic location of the soil in Ghana, are in Table A2. A summary of the main chemical characteristics of the four soils is given in Table A3.
Our purpose with the following analysis is to use an adequate grouping statistic to extract information from the data. In the previous approach, presented in [1], the grouping statistic was applied to a natural grouping of the varieties, taking into account some genetic ascent of the varieties. In this work, we first look for a grouping obtained by a clustering technique, namely, the KMeans clustering methodology, applied directly to the set of values of the three variables. The analysis method proposed goes as follows:
1.
We first determine the data grouping in three clusters using the KMeans clustering classification method. A good introduction to classification techniques can be read in [21]. The results of the classification are presented in Table 6, and a graphical representation of the 3-dimensional clusters of observations is presented in Figure 4.
2.
As a consequence of the classification, we then use the grouping statistics G ( 12 , 2 , 9 , 3 ) and G ( 13 , 3 , 9 , 3 , 1 ) to assess for each one of the four soils the quality of the grouping. The values obtained are given in Table 7.
3.
We then present some conclusions reached by the use of the statistics G ( 12 , 2 , 9 , 3 ) and G ( 13 , 3 , 9 , 3 , 1 ) to the results of the KMeans grouping.
Remark 8
(Analysis of the KMeans grouping in Table 6). It is relevant to notice that for all the three variables—plant height, stem diameter, and dry matter—the three clusters contain the same varieties. With the exception of Soil 2, the set of varieties V is always partitioned into two groups—each group in its own cluster—to wit, the set { V 2 , V 5 , V 16 } and the complement of this set with respect to V , that is, V { V 2 , V 5 , V 16 } . The exception of Soil 2 is that the set of varieties V has variety V 14 in a cluster that is empty in the case of the other soils. Since the grouping statistic is applied to ranks of the varieties, obtained from the sums of all the observations in each variety, we discard the observation variety V 14 in cluster 1 for Soil 2. For Soil 2, the alternative of considering, instead of G ( 12 , 2 , 9 , 3 ) , the grouping statistic G ( 13 , 3 , 9 , 3 , 1 ) —corresponding to having one more rank—does not change the final result since the rank value corresponding to the V 14 variable is always equal to 13.
Remark 9
(Analysis of the statistic results in Table 7). Let us consider a statistical test using the statistical distribution G ( 12 , 2 , 9 , 3 ) , where the null hypothesis is given as follows: the grouping produces inhomogeneous groups of varieties. We observe that the rejection region for this test consists of values of the statistic for sufficiently rank homogeneous groups. If the within-group ranks are homogeneous, the values of the statistic—given by Formula (1)—are smaller and so, given that the 5% quantile for the distribution G ( 12 , 2 , 9 , 3 ) is q G ( 12 , 2 , 9 , 3 ) : 0.05 = 296 , we reject the null hypothesis—that the group is not homogeneous—for all the three variables, that is, plant height, stem diameter and dry matter, in Soil 3. This allow us to conclude that Soil 3 provides homogeneous production characteristics as given by the three variables’ values. In order to have a more detailed look at the particular behavior of Soil 3, we present in Table 8 the ranks of the varieties in each cluster for all three variables.
Remark 10
(Analysis of the ranks of the varieties per clusters in Table 8). It is clear from the observation of the ranks (in red color) that the varieties in cluster 3 for Soil 3— V 2 , V 5 , V 16 —perform homogeneously worse than the remaining varieties of the set V since the ranks are, in general, larger, the exception being variety V 2 for the variable stem diameter. Varieties V 5 , V 16 have the second characteristic string CRG0314 in common—see Table A1 in the Appendix A—and so this may be an indication of the inadequacy of this ancestor characteristic of varieties V 5 , V 16 to Soil 3.

5. Discussion

As stated in the first line of the introduction, a particular example of grouping distribution was used in [1], initially to allow the application of a well-known statistical methodology analysis and secondly to extract within-group homogeneity information of the grouping. In the present work, we studied the whole family of grouping distributions. Having observed severe computational challenges for the computation of the laws of grouping distributions—for large parameter values—we explored the possibility of an asymptotic result that would allow the determination of an approximate distribution for large parameter values. In order to obtain such an asymptotic result, we assumed some hypothesis heuristically deduced from the observation of the law behavior for small parameter values. It came as a surprise that under such a hypothesis, the asymptotic limit is a generalized function and not a measure that may be considered a probability distribution. Left with the problem of obtaining approximate distributions for the laws of grouping distributions, in the case of large parameter values, we proposed two methodologies that must be further studied since they present difficult technical implementation problems. In the context of Big Data, the grouping of variables and the determination of potential within-group homogeneities may represent a useful tool.
In this work, we deepened the study of some grouping distributions that find an important usage in testing the internal homogeneity—the within-group homogeneity—of subgroups of a given group. An example of such usage using real data was, presented and the comparison between the use of grouping statistics and the use of classic ANOVA highlighted the power of the grouping distributions to bring forward new information from the data. We pointed out the computational challenges arising from large values of the cardinal of the values to be grouped and the cardinal of the set of subgroups to consider—the main parameters of the grouping distribution—and aimed at a better understanding of the properties of the grouping distributions for large parameter values by several ways. In one of those ways, we proved that there does not exist an asymptotic result, giving a limit probability law when the main parameters tend to infinity and under a set of hypothesis that were observed in a significant set of examples. To contravene this negative result, we proposed, as implementation open problems, the ulterior study of two methodologies aimed at providing distribution approximations for large values of the main parameters.

6. Conclusions and Future Work

In this work, we formally defined and studied a family of discrete probability distributions—already introduced in a previous applied work—that may be used to assess the within-group homogeneity of a partition of observations in groups according to some criteria; examples of such criteria are the variables having some genetic common ascendent, and the values of the variables being grouped according to some clustering method such as KMeans. For the smaller values of the parameters, for which the computational effort is manageable, we provided tables of quantiles allowing to perform a test; the proposed test has as a null hypothesis that the group is not homogeneous. We also further developed a moment generating function result with a twofold purpose: firstly to analyze the possibility of an asymptotic result for the grouping distributions for large values of the parameters under some natural hypothesis, and secondly to assess the quality of a continuous approximation of the grouping distribution by a mixture of normal distributions. We presented an analysis of cocoa breeding experiment data, using an adequate grouping distribution, and we were able to extract information on the homogeneous behavior of cocoa plants production variables in one specific type of soil. By further analyzing the behavior of the variables grouping, we were able to conclude on the possible inadequacy of two variables in that specific type of soil.
There are at least three directions of development in the study of the grouping distributions. The first avenue is the development of the two methodologies proposed to provide distribution approximations for large values of the main parameters; this development may require an ingenious use of some optimization techniques. The second avenue is the investigation of the properties of the numerical semigroups associated to the sequence of convolutions of the grouping laws. And the third avenue is the quest for asymptotic results under different sets of hypotheses.

Author Contributions

Conceptualization, M.L.E., N.P.K., C.N., K.O.-A. and P.P.M.; methodology, M.L.E.; software, M.L.E.; validation, M.L.E., N.P.K., C.N., K.O.-A. and P.P.M.; formal analysis, M.L.E., N.P.K., C.N., K.O.-A. and P.P.M.; investigation, M.L.E., N.P.K., C.N., K.O.-A. and P.P.M.; resources, M.L.E.; data curation, M.L.E.; writing—original draft preparation, M.L.E.; writing—review and editing, M.L.E., N.P.K., C.N., K.O.-A. and P.P.M.; visualization, M.L.E., N.P.K., C.N., K.O.-A. and P.P.M.; supervision, M.L.E.; project administration, M.L.E.; funding acquisition, M.L.E. and C.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by national ressources through the FCT—Fundação para a Ciência e a Tecnologia, I.P., under the scope of the project UIDB/00297/2020 (https://doi.org/10.54499/UIDB/00297/2020, accessed on Monday 30 September 2024) and project UIDP/00297/2020 (https://doi.org/10.54499/UIDP/00297/2020, accessed on Monday 30 September 2024) (Centre for Mathematics and Applications, University Nova de Lisboa) and also through project UIDB/00212/2020 (https://doi.org/10.54499/UIDB/00212/2020, accessed on Thursday 13 September 2025) and project UIDB/00212/2020 (https://doi.org/10.54499/UIDP/00212/2020, accessed on Thursday 13 September 2024). (Centre for Mathematics and Applications, Universidade da Beira Interior). We thank Jerome Agbesi Dogbatse of the Cocoa Research Institute of Ghana for allowing us to use the data in this study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MGFMoment Generating Function
ANOVAAnalysis of Variance
PDFProbability Density Function
KMeansClustering classification methodology

Appendix A. Data Characteristics

In this appendix, we detail some information available on the data and on the ANOVA analysis on the data.

Appendix A.1. On the Varieties of Plants of Cocoa in the Data

According to Table A1, there are 12 different varieties, 4 from the ascendant PA150 ( V 3 , V 4 , V 5 , V 11 ) , 5 from T63/967 ( V 12 , V 13 , V 14 , V 15 , V 16 ) and 3 other varieties with no common ascendant.
Table A1. List of varieties.
Table A1. List of varieties.
CodeVarietyCodeVarietyCodeVariety
V 1 T 85 / 799 × P A 7 / 808 V 2 T 60 / 887 × C R G 8914 V 3 P A 150 × E Q 3338
V 4 P A 150 × P A 88 V 5 P A 150 × C R G 0314 V 6 Standard Variety
V 11 P A 150 × C R G 3019 V 12 T 63 / 967 × I M C 60 V 13 T 63 / 967 × E Q 78
V 14 T 63 / 967 × C R G 2022 V 15 T 63 / 967 × C R G 9066 V 16 T 63 / 967 × C R G 0314

Appendix A.2. On the Soils of the Data

The surface soil (0–15 cm) of three soil series, namely, Soil 1, Soil 2, and Soil 3, was collected from secondary forests in the Western Region. Soil 4 is from the Eastern Region. This region has the recommended soil pH required for cocoa production and is considered to be one of the best cocoa-growing soils and so was also included in the study. The four surface soils were passed through a 2 mm mesh sieve and a subsample characterized through the initial laboratory analysis. A specification of the origin of the soils is given in Table A2.
Table A2. List of soils.
Table A2. List of soils.
NameRegion of Provenance in Ghana
Soil 1Forest Ochrosols-Oxysols intergrades (Saamang)
Soil 2Forest Oxysol (Samreboi)
Soil 3Forest Ochrosol (Boako)
Soil 4Forest Rubrisols-Ochrosol intergrade (Tafo)
In general, all forest Ochrosols are acidic in nature, but the management practices applied—e.g., fertilizer application—may change the levels of chemical properties measured. The chemical properties of the different soils are specified in Table A3.
Table A3. Chemical properties of different soils.
Table A3. Chemical properties of different soils.
Soil
Name
pH (a)
OC (b)
(%)
TN (c)
(%)
AP (d)
(mg/kg)
K (e)
(cmolc/kg)
Mg (f)
(cmolc/kg)
Ca (g)
(cmolc/kg)
Soil 14.9913.360.32717.5930.2630.557.98
Soil 24.2080.50.09520.7150.5693.1421.729
Soil 35.0722.340.30919.8490.3292.8616.309
Soil 45.6350.760.12323.8450.573.0096.473
Critical min5.63.50.09200.251.337.5
(a) pH—potential of hydrogen. (b) OC—Organic Carbon. (c) TN—Total Nitrogen. (d) AP—Available Phosphorus. (e) K—Exchangeable Potassium. (f) Mg—Exchangeable Magnesium. (g) Ca—Exchangeable Calcium.

Appendix A.3. ANOVA of the Data

We present the results of the ANOVA of the data with the purpose of establishing a contrast—in terms of the production performance—of the influence of the factors variety and soil on the three variables plant height, stem diameter, and dry matter. For the purpose of allowing a correct perspective on what follows, we must stress the distinct nature of the two statistical approaches presented: the grouping distribution application and the ANOVA analysis. The ANOVA analysis is focused on comparing the means of variables and so it gives information on differences; in contrast, the grouping statistics operates on ranks of the sums of all the values of the variables and gives information on the within-group homogeneity.
Remark A1
(Analysis of the results in Table A4). As usual, we have that the null hypothesis for each factor is that there is no significant difference between groups of that factor and so, a sufficiently small p-value—values coloured in red in the table—allows us to reject the null hypothesis. We may then conclude that the type of soil factor is always relevant for the production performance in the three variables, that the variety factor is only relevant for the plant height variable, and that the interaction in factors variety–soil is relevant for the dry matter variable. As a consequence, the results of the statistical clustering analysis concerning Soil 3 acquire relevance.
Table A4. ANOVA p-values for factors variety, soil, and interaction variety–soil.
Table A4. ANOVA p-values for factors variety, soil, and interaction variety–soil.
FactorsPlant HeightStem DiameterDry Matter
Variety 0.00482738 0.713182 0.992835
Soil 1.53828 × 10 20 0.00731414 1.37275 × 10 11
Variety–Soil 0.430225 0.717139 0.000468982
Remark A2
(Analysis of the results in Table A5). The results in this table should be compared with the results of Table 8 where we see that, in general, the ranks in Cluster 3 for varieties V 2 , V 5 , and V 16 are almost always the worst possible ranks obtained from the sum of all observations, the exception being for the variable stem diameter, a case for which variety V 2 has the first rank.
Nevertheless, we can observe that in more other cases, the averages for some varieties surpass the global averages. The values in the table show that for the variable plant height, the average in Soil 3 is larger than the global average for all three varieties; for the variable stem diameter, the average in Soil 3 is larger than the global average for varieties V 2 and V 16 ; and for the variable dry matter, the average in Soil 3 is larger than the global average for varieties V 2 and V 5 .
Moreover, we have that for the variable plant height, the average plant height in Soil 3 is larger than the global average of plant height in Soil 3 for the variety V 2 ; for the variable stem diameter, the average stem diameter in Soil 3 is larger that the global average of the stem diameter in Soil 3 also for the variety V 2 ; for the variable dry matter, the average dry matter in Soil 3 is larger than the global average of dry matter in Soil 3 for the varieties V 2 and V 5 .
These results dealing with averages complement the results based on the ranks of total outputs in each of the variables and varieties.
Table A5. Averages values for the variables plant height, stem diameter, and dry matter globally and in Soil 3.
Table A5. Averages values for the variables plant height, stem diameter, and dry matter globally and in Soil 3.
VarietyA (PH) (a)Plant Height (b)A (SD) (c)Stem Diameter (d)A (DM) (e)Dry Matter (f)
V 2 67.2778 70.0361 10.4125 11.1756 34.1715 41.4906
V 5 59.7215 65.5667 11.297 10.5994 32.4229 35.5894
V 16 64.3505 67.26928 10.2296 10.3962 31.4934 30.3542
A (PH)(a): Average plant height for each variety. Plant Height(b): The global average plant height in Soil 3 is 67.395. A (SD)(c): Average stem diameter for each variety. Stem Diameter(d): The global average stem diameter in Soil 3 is 10.765. A (DM)(e) Average dry matter for each variety. Dry Matter(f): The global average dry matter for Soil 3 is 34.4709.

Appendix A.4. A Proof of a Referred Result

In this appendix, we collect the proof of a result that was quoted in the main text and that may have some interest on its own. The following proof is well known and is presented only for the completeness of the work.
Theorem A1.
The multinomial coefficient N n 1 , n 2 , , n p is maximized when the values of n 1 , n 2 , , n p are as close to each other as possible.
Proof. 
The multinomial coefficient is defined by
N n 1 , n 2 , , n p = N ! n 1 ! n 2 ! n p ! ,
where n 1 + n 2 + + n p = N . The Stirling’s approximation formula for the integer factorials is the following for large n:
n ! 2 π n n e n .
Applying Stirling’s approximation to the multinomial coefficient gives
N n 1 , n 2 , , n p 2 π N N e N i = 1 p 2 π n i n i e n i .
Simplifying this expression, we obtain
N n 1 , n 2 , , n p N e N i = 1 p n i e n i · 2 π N i = 1 p 2 π n i = N N n 1 n 1 · n 2 n 2 n p n p · 1 e N p · 1 2 π p 1 · N n 1 · n 2 n p
We are now going to perform an analysis of the logarithmic terms. To maximize the multinomial coefficient, we maximize its logarithm:
log N n 1 , n 2 , , n p = log N ! i = 1 p log n i !
Using Stirling’s approximation for logarithms
log N ! N log N N , log n i ! n i log n i n i .
So,
log N n 1 , n 2 , , n p ( N log N N ) i = 1 p ( n i log n i n i ) = N log N N i = 1 p n i log n i + i = 1 p n i = N log N i = 1 p n i log n i .
Since i = 1 p n i = N , we have
log N n 1 , n 2 , , n p = N log N i = 1 p n i log n i .
We now use Jensen’s Inequality. The function f ( x ) = x log x is convex for x > 0 . By Jensen’s inequality, for n 1 , n 2 , , n p such that n 1 + n 2 + + n p = N :
i = 1 p n i N log n i log i = 1 p n i · n i N = log N p ,
therefore,
i = 1 p n i log n i N log N p .
This implies that:
N log N i = 1 p n i log n i N log N N log N p = N log N N log N + N log p = N log p .
Hence, the expression log N n 1 , n 2 , , n p is maximized when n i are as equal as possible. □

References

  1. Opoku-Ameyaw, K.; Nunes, C.; Esquível, M.L.; Mexia, J.T. CMMSE: A nonparametric test for grouping factor levels: An application to cocoa breeding experiments in acidic soils. J. Math. Chem. 2023, 61, 652–672. [Google Scholar] [CrossRef]
  2. Yitzhaki, S. Gini’s mean difference: A superior measure of variability for non-normal distributions. Metron 2003, 61, 285–316. [Google Scholar]
  3. Zenga, M.; Polisicchio, M.; Greselin, F. The variance of Gini’s mean difference and its estimators. Statistica 2004, 64, 455–475. [Google Scholar]
  4. Haye, R.L.; Zizler, P. The Gini mean difference and variance. Metron 2019, 77, 43–52. [Google Scholar] [CrossRef]
  5. Baz, J.; Pellerey, F.; Díaz, I.; Montes, S. Stochastic ordering of variability measure estimators. Statistics 2024, 58, 26–43. [Google Scholar] [CrossRef]
  6. Conover, W.J. Practical Nonparametric Statistics, 3rd ed.; John Wiley & Sons, Inc.: New York, NY, USA, 1999; pp. viii+584. [Google Scholar]
  7. Butucea, C.; Tribouley, K. Nonparametric homogeneity tests. J. Stat. Plan. Inference 2006, 136, 597–639. [Google Scholar] [CrossRef]
  8. Graham, R.L.; Knuth, D.E.; Patashnik, O. Concrete Mathematics: A Foundation for Computer Science, 2nd ed.; Addison-Wesley Publishing Company: Reading, MA, USA, 1994; pp. xiv+657. [Google Scholar]
  9. Esquível, M.L. Probability generating functions for discrete real-valued random variables. Teor. Veroyatn. Primen. 2007, 52, 129–149. [Google Scholar] [CrossRef]
  10. Rudin, W. Real and Complex Analysis, 3rd ed.; McGraw-Hill Book Co.: New York, NY, USA, 1987; pp. xiv+416. [Google Scholar]
  11. Katznelson, Y. An Introduction to Harmonic Analysis; Dover Publications, Inc.: New York, NY, USA, 1976; corrected edition; pp. xiv+264. [Google Scholar]
  12. Widder, D.V. The Laplace Transform; Princeton Mathematical Series; Princeton University Press: Princeton, NJ, USA, 1941; Volume 6, pp. x+406. [Google Scholar]
  13. Schwartz, L. Mathematics for the Physical Sciences; Dover Publications, Inc.: Mineola, NY, USA, 1966; p. 358. [Google Scholar]
  14. Kallenberg, O. Foundations of Modern Probability, 3rd ed.; Probability Theory and Stochastic Modelling; Springer: Cham, Switzerland, 2021; Volume 99, pp. xii+946. [Google Scholar] [CrossRef]
  15. Schmüdgen, K. The Moment Problem; Graduate Texts in Mathematics; Springer: Cham, Switzerland, 2017; Volume 277, pp. xii+535. [Google Scholar]
  16. Durrett, R. Probability—Theory and Examples, 5th ed.; Cambridge Series in Statistical and Probabilistic Mathematics; Cambridge University Press: Cambridge, UK, 2019; Volume 49, pp. xii+419. [Google Scholar] [CrossRef]
  17. Beffa, F. Weakly Nonlinear Systems—With Applications in Communications Systems; Understanding Complex Systems; Springer: Cham, Switzerland, 2024; pp. xiv+371. [Google Scholar] [CrossRef]
  18. Lindner, M. An introduction to the limit operator method. In Infinite Matrices and Their Finite Sections; Frontiers in Mathematics; Birkhäuser Verlag: Basel, Switzerland, 2006; pp. xv+191. [Google Scholar]
  19. Ran, A.C.M.; Serény, A. The finite section method for infinite Vandermonde matrices. Indag. Math. 2012, 23, 884–899. [Google Scholar] [CrossRef]
  20. Hernández-Pastora, J.L. On the solutions of infinite systems of linear equations. Gen. Relativ. Gravit. 2014, 46, 1622. [Google Scholar] [CrossRef]
  21. Scitovski, R.; Sabo, K.; Martínez-Álvarez, F.; Ungar, Š. Cluster Analysis and Applications; Springer: Cham, Switzerland, 2021. [Google Scholar] [CrossRef]
Figure 1. Empirical distribution PDF of standardized X G ( 23 , 2 , 11 , 12 ) (blue) and fitting of the standardized data of the law of X G ( 23 , 2 , 11 , 12 ) by a normal distribution.
Figure 1. Empirical distribution PDF of standardized X G ( 23 , 2 , 11 , 12 ) (blue) and fitting of the standardized data of the law of X G ( 23 , 2 , 11 , 12 ) by a normal distribution.
Mathematics 13 01805 g001
Figure 2. Linear fitting of the limit points of the logistic fittings for X G ( N , 2 , n 1 , n 2 ) considering N = 13 , 15 , 17 , 19 , 21 , 23 .
Figure 2. Linear fitting of the limit points of the logistic fittings for X G ( N , 2 , n 1 , n 2 ) considering N = 13 , 15 , 17 , 19 , 21 , 23 .
Mathematics 13 01805 g002
Figure 3. Logistic fitting of the first absolute values of moment ratios for X G ( 23 , 2 , 11 , 12 ) (left) and correspondent residuals (right).
Figure 3. Logistic fitting of the first absolute values of moment ratios for X G ( 23 , 2 , 11 , 12 ) (left) and correspondent residuals (right).
Mathematics 13 01805 g003
Figure 4. Graphical representation of the 3 KMeans defined clusters in the three variables plant height, stem diameter and dry matter.
Figure 4. Graphical representation of the 3 KMeans defined clusters in the three variables plant height, stem diameter and dry matter.
Mathematics 13 01805 g004
Table 1. Values and corresponding probabilities for G ( 9 , 2 , 4 , 5 ) .
Table 1. Values and corresponding probabilities for G ( 9 , 2 , 4 , 5 ) .
( 60 , 1 63 ) ( 74 , 1 63 ) ( 84 , 2 63 ) ( 90 , 1 21 ) ( 92 , 2 63 ) ( 94 , 1 63 ) ( 100 , 1 14 ) ( 102 , 2 63 )
( 106 , 2 21 ) ( 108 , 10 63 ) ( 110 , 1 21 ) ( 112 , 5 126 ) ( 114 , 10 63 ) ( 116 , 1 6 ) ( 118 , 4 63 ) ( 120 , 1 126 )
Table 2. Quantiles q α = q 0.05 and q β = q 0.1 for groupings in sets having as elements N = 14 , 15 , , 19 .
Table 2. Quantiles q α = q 0.05 and q β = q 0.1 for groupings in sets having as elements N = 14 , 15 , , 19 .
14 q α , q β 15 q α , q β 16 q α , q β 17 q α , q β 18 q α , q β 19 q α , q β
(2,12)598(2,13)778(2,14)964(2,15)1178(2,16)1422(2,17)1698
618 778 984 1200 1450 1728
(3,11)512(3,12)670(3,13)832(3,14)1022(3,15)1260(3,16)1512
528 678 852 1046 1284 1540
(4,10)446(4,11)580(4,12)724(4,13)910(4,14)1110(4,15)1350
466 598 752 934 1146 1380
(5,10)400(5,10)514(5,11)652(5,12)814(5,13)1004(5,14)1220
416 536 676 842 1036 1256
(6,8)374(6,9)476(6,10)600(6,11)750(6,12)922(6,13)1118
394 502 628 780 954 1156
(7,7)364(7,8)456(7,9)568(7,10)706(7,11)864(7,12)1044
384 482 596 734 896 1084
(8,8)560(8,9)684(8,10)830(8,11)998
588 714 862 1036
(9,9)816(9,10)974
852 1012
Table 3. Quantiles q α = q 0.05 and q β = q 0.1 for groupings in sets having as elements N = 8 , 9 , , 13 .
Table 3. Quantiles q α = q 0.05 and q β = q 0.1 for groupings in sets having as elements N = 8 , 9 , , 13 .
8 q α , q β 9 q α , q β 10 q α , q β 11 q α , q β 12 q α , q β 13 q α , q β
(2,6)72(2,7)114(2,8)186(2,9)260(2,10)352(2,11)464
84 128 186 274 368 482
(3,5)60(3,6)102(3,7)148(3,8)208(3,9)296(3,10)392
68 102 156 222 304 410
(4,4)52(4,5)84(4,6)126(4,7)180(4,8)248(4,9)332
60 90 130 188 264 356
(5,5)120(5,6)160(5,7)228(5,8)304
128 182 240 320
(6,6)216(6,7)286
236 302
Table 4. Limit points for moment ratio logistic fittings of X G ( N , 2 , n 1 , n 2 ) for N = 13 , 15 , 17 , 19 , 21 , 23 .
Table 4. Limit points for moment ratio logistic fittings of X G ( N , 2 , n 1 , n 2 ) for N = 13 , 15 , 17 , 19 , 21 , 23 .
G ( 13 , 2 , 6 , 7 ) : L 6.34319 G ( 15 , 2 , 7 , 8 ) : L 7.45872 G ( 17 , 2 , 8 , 9 ) : L 8.57213
G ( 19 , 2 , 9 , 10 ) : L 9.68292 G ( 21 , 2 , 10 , 11 ) : L 10.79090 G ( 23 , 2 , 11 , 12 ) : L 11.89560
Table 5. Quantiles q G : 0.05 and q G : 0.1 for G ( N , 2 , n 1 , n 2 ) for N = 13 , 15 , 17 , 19 , 21 , 23 , quantiles q N : 0.05 and q N : 0.1 from the normal distribution and corresponding relative errors ϵ 0.05 and ϵ 0.1 .
Table 5. Quantiles q G : 0.05 and q G : 0.1 for G ( N , 2 , n 1 , n 2 ) for N = 13 , 15 , 17 , 19 , 21 , 23 , quantiles q N : 0.05 and q N : 0.1 from the normal distribution and corresponding relative errors ϵ 0.05 and ϵ 0.1 .
G = G ( N , 2 , n 1 , n 2 ) q G : 0.05 q N : 0.05 ( ϵ 0.05 ) q G : 0.1 q N : 0.1 ( ϵ 0.1 )
G ( 23 , 2 , 11 , 12 ) 17801808.76 (1.59%)18381836.86 (0.062%)
G ( 21 , 2 , 10 , 11 ) 13381360.82 (1.68%)13861384.2 (0.13%)
G ( 19 , 2 , 9 , 10 ) 974993.586 (1.97%)10121012.67 (0.066%)
G ( 17 , 2 , 8 , 9 ) 684699.057 (2.15%)714714.285 (0.04%)
G ( 15 , 2 , 7 , 8 ) 456469.233 (2.82%)482481.035 (0.2%)
G ( 13 , 2 , 6 , 7 ) 286296.114 (1.59%)302304.924 (0.062%)
Table 6. Subsets of varieties present in each cluster according to the 3 variables and 4 soils.
Table 6. Subsets of varieties present in each cluster according to the 3 variables and 4 soils.
Types of SoilPlant HeightStem DiameterDry Matter
Soil 1: cluster varieties (a) 1 : V { V 2 , V 5 , V 16 } 1 : V { V 2 , V 5 , V 16 } 1 : V { V 2 , V 5 , V 16 }
2 : 2 : 2 :
3 : { V 2 , V 5 , V 16 } 3 : { V 2 , V 5 , V 16 } 3 : { V 2 , V 5 , V 16 }
Soil 2: cluster varieties 1 : { V 14 }  (b) 1 : { V 14 }  (b) 1 : { V 14 }  (b)
2 : V { V 2 , V 5 , V 16 } 2 : V { V 2 , V 5 , V 16 } 2 : V { V 2 , V 5 , V 16 }
3 : { V 2 , V 5 , V 16 } 3 : { V 2 , V 5 , V 16 } 3 : { V 2 , V 5 , V 16 }
Soil 3: cluster varieties 1 : 1 : 1 :
2 : V { V 2 , V 5 , V 16 } 2 : V { V 2 , V 5 , V 16 } 2 : V { V 2 , V 5 , V 16 }
3 : { V 2 , V 5 , V 16 } 3 : { V 2 , V 5 , V 16 } 3 : { V 2 , V 5 , V 16 }
Soil 4: cluster varieties 1 : V { V 2 , V 5 , V 16 } 1 : V { V 2 , V 5 , V 16 } 1 : V { V 2 , V 5 , V 16 }
2 : 2 : 2 :
3 : { V 2 , V 5 , V 16 } 3 : { V 2 , V 5 , V 16 } 3 : { V 2 , V 5 , V 16 }
(a) The complete set of plant varieties is V = { V 1 , V 2 , V 3 , V 4 , V 5 , V 6 , V 11 , V 12 , V 13 , V 14 , V 15 , V 16 } . (b) One observation of variety V 14 appears in cluster 2 ; for Soil 2, we only consider two groups of varieties.
Table 7. Values of the grouping statistic according to the 3 variables and 4 soils.
Table 7. Values of the grouping statistic according to the 3 variables and 4 soils.
Types of SoilPlant HeightStem DiameterDry Matter
Soil 1360324344
Soil 2332344332
Soil 3248 (a)284 (a)268 (a)
Soil 4320312352
(a) The 5% quantiles for the relevant distributions are: q G ( 12 , 2 , 9 , 3 ) : 0.05 = 296 = q G ( 13 , 3 , 9 , 3 , 1 ) : 0.05 .
Table 8. Subsets of ranks of varieties present in each cluster according to the 3 variables and 4 soils.
Table 8. Subsets of ranks of varieties present in each cluster according to the 3 variables and 4 soils.
Types of SoilPlant HeightStem DiameterDry Matter
Soil 1: cluster ranks (a) 1 : R { 2 , 7 , 8 } 1 : V { 8 , 9 , 11 } 1 : V { 2 , 4 , 10 }
2 : 2 : 2 :
3 : { 2 , 7 , 8 } 3 : { 8 , 9 , 11 } 3 : { 2 , 4 , 10 }
Soil 2: cluster ranks (b) 1 : { 13 }  (c) 1 : { 13 }  (c) 1 : { 13 }  (c)
2 : R { 1 , 5 , 11 } 2 : R { 1 , 5 , 8 } 2 : R { 3 , 10 , 11 }
3 : { 1 , 5 , 11 } 3 : { 1 , 5 , 8 } 3 : { 3 , 10 , 11 }
Soil 3: cluster ranks (a) 1 : 1 : 1 :
2 : R { 10 , 11 , 12 } 2 : R { 1 , 11 , 12 } 2 : R { 9 , 11 , 12 }
3 : {10, 11, 12}3 : {1, 11, 12}3 : {9, 11, 12}
Soil 4: cluster ranks (a) 1 : R { 1 , 8 , 12 } 1 : R { 1 , 10 , 11 } 1 : R { 3 , 5 , 11 }
2 : 2 : 2 :
3 : { 1 , 8 , 12 } 3 : { 1 , 10 , 11 } 3 : { 3 , 5 , 11 }
(a) The complete set of possible ranks for plant varieties is R = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 11 , 12 } . (b) The complete set of possible ranks for plant varieties is R = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 11 , 12 , 13 } . (c) The rank for the observation of variety V 14 that appears in cluster 1 ; for Soil 2, we only consider two groups of varieties.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Esquível, M.L.; Krasii, N.P.; Nunes, C.; Opoku-Ameyaw, K.; Mota, P.P. Rank-Based Family of Probability Laws for Testing Homogeneity of Variable Grouping. Mathematics 2025, 13, 1805. https://doi.org/10.3390/math13111805

AMA Style

Esquível ML, Krasii NP, Nunes C, Opoku-Ameyaw K, Mota PP. Rank-Based Family of Probability Laws for Testing Homogeneity of Variable Grouping. Mathematics. 2025; 13(11):1805. https://doi.org/10.3390/math13111805

Chicago/Turabian Style

Esquível, Manuel L., Nadezhda P. Krasii, Célia Nunes, Kwaku Opoku-Ameyaw, and Pedro P. Mota. 2025. "Rank-Based Family of Probability Laws for Testing Homogeneity of Variable Grouping" Mathematics 13, no. 11: 1805. https://doi.org/10.3390/math13111805

APA Style

Esquível, M. L., Krasii, N. P., Nunes, C., Opoku-Ameyaw, K., & Mota, P. P. (2025). Rank-Based Family of Probability Laws for Testing Homogeneity of Variable Grouping. Mathematics, 13(11), 1805. https://doi.org/10.3390/math13111805

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop