1. Introduction
Order statistics have found important applications in many diverse areas, including life testing and reliability, robust inference, statistical quality control, filtering theory, and signal and image processing (see, e.g., [
1]). The properties of order statistics based on independent and identically distributed random variables and, to a large extent, on independent but non-identically distributed samples are well-established and have been extensively studied (see, e.g., [
2,
3]). Several classical results obtained for independent observations have been extended to dependent samples with prescribed joint distributions or partially known moments (see, e.g., [
4,
5,
6,
7,
8,
9,
10,
11]).
Research on the extremal properties of order statistic distributions under dependence uncertainty has primarily focused on cases where no restrictions are imposed on the interdependence of observations. In these frameworks, either the one-dimensional marginals are assumed to be known [
12,
13,
14,
15,
16,
17], or the distributions of the maxima (or minima) of all
k-tuples are assumed to be identical and predefined (see [
18,
19]). Few studies have investigated the optimal estimation of distribution functions of order statistics from samples of dependent random variables with partially known dependence structures. Kemperman’s [
20] analysis of
k-independent, identically distributed observations yielded the first significant result in this field, providing a general approach for deriving pointwise sharp bounds for the distribution functions of order statistics and establishing the best-possible upper bounds for single-order statistics from pairwise independent observations. Extensions of Kemperman’s results to piecewise uniform marginal copulas and linear combinations of distribution functions of single-order statistics were presented in [
21]. Mallows [
22] made another significant contribution by considering three-element, two-independent samples with uniform marginals and constructing explicit extremal distributions that maximize the distribution function of the minimum. Furthermore, Okolewski [
23] analyzed the extremal properties of order statistic distributions for dependent samples with partially specified multivariate marginals when the marginal copula diagonals up to a certain dimension are known.
The present study is concerned with determining pointwise sharp bounds for linear combinations of joint distribution functions and joint reliability functions of selected order statistics derived from identically distributed k-independent random variables. The problem is reformulated as a moment problem and solved using a geometric approach. The bounds established are significant because, in particular, they hold under minimal assumptions-specifically, for arbitrarily dependent random variables, without requiring knowledge of their joint dependence structure. In reliability theory, these results enable the derivation of best-possible lower and upper bounds for the reliability functions of semicoherent systems with shared, exchangeable components when the dependence structure among component lifetimes is completely unspecified or exhibits k-independence (see Remark 3). Additionally, the bounds provide precise expectation ranges for functions of order statistics when the underlying random variables take values in a finite set. In the case of possibly dependent random variables, these results may be applicable in the context of largest-claims (LC) reinsurance, where conservative bounds are essential for risk assessment and pricing under unknown dependence structures (see Remark 2).
This paper is organized as follows:
Section 2 derives sharp distribution bounds for order statistics and illustrates their applications.
Section 3 presents explicit examples that demonstrate the theoretical results.
2. Distribution Bounds for Order Statistics
Suppose we have
n random variables
, and order statistics
based on them. Let
be distinct real numbers. The number
q may be greater than, equal to, or less than
n. We consider multivariate marginal distribution functions
of several order statistics
for some
, evaluated at points belonging to the fixed finite set
It suffices to consider only strictly increasing arguments
, since otherwise redundant terms can be removed.
More generally, we study arbitrary linear combinations
involving multivariate marginal distribution functions of order statistics evaluated at the elements of the finite set
. Here,
. By setting some coefficients
, various reductions are possible. The value
is uniquely determined when the distribution of the vector
is fully known. When only partial information on
is available,
takes values within a certain range.
In statistical theory, it is well known that for a set of three or more random variables to be mutually independent, pairwise independence is necessary but not sufficient. A classic example of three discrete variables that are pairwise independent but not mutually independent, frequently cited in the statistical literature, is due to S. Bernstein (see Cramér [
24]). For an overview of generalizations of this example to cases with more than three variables and continuous distributions, see [
25]. In this work, the author constructs an infinite sequence of random variables such that the components of any proper subset are independent if and only if the size of the subset is less than or equal to a fixed positive integer
k. Such random variables are called
k-independent.
k-independence is a powerful tool because it provides a balance between full independence and arbitrary dependence. It is often used to reduce the amount of randomness required in probabilistic algorithms, for example, when weaker sources of randomness are sufficient for analyses that use Chernoff–Hoeffding bounds under limited independence, such as the analysis of randomized algorithms for random sampling (see [
26]).
We are interested in determining sharp lower and upper bounds for
over the set of all possible distributions of
k-independent vectors
having the same one-dimensional marginal distribution
F. Here,
is
k-independent if every
k-tuple
is independent (for
), while in the case
, the components may be arbitrarily dependent. Our approach transforms this problem into a moment problem based on Kemperman’s characterization of
k-independence [
20] [Theorem 1]: A random vector
has the same distribution as the vector
, where
associated to some
k-independent
, if and only if
takes values in the set
and satisfies the moment conditions
where
if
s is true and
otherwise,
is the set of non-negative integers,
denotes the falling factorial, indices satisfy
, and
with probabilities
, assuming
and
.
Throughout, we assume that
to avoid trivial cases.
Our main result states that the bounds on
can be expressed as values of certain known functions, which depend solely on the coefficients
evaluated at points determined exclusively by the values
In essence, these bounds take the same form as those for linear combinations of distribution functions of individual order statistics from possibly dependent observations, as established in [
16].
Before formulating the result, we introduce some notation and terminology. Since
, it suffices to require condition (
3) only for
(cf. Kemperman [
20] [Remark]). Denote by
the
elements of the set
, which consists of all orders not exceeding
k of factorial moments of the random vector
taking values in the set
Define the vector function
where
for
. Each component
represents a possible value of the product
for a vector
from
. Accordingly, the coordinates of
enumerate the possible values of such products for all vectors in
.
Let
be the compact and convex set of all possible moments
, where the distribution of
varies over all probability distributions on
. Clearly, we have
where
and
denotes the convex hull of the set
A.
We now define two functions necessary to state our results. For any function
, define
to be, respectively, the greatest convex function such that
and the smallest concave function such that
Theorem 1. Let , , and be fixed integers. Let and be fixed real numbers, where , , and .
(i) If is a vector of k-independent random variables with a common distribution function F satisfying condition (5), then the following bounds hold:wherefor andwithin which and (ii) Moreover, for any distribution function F that satisfies condition (5), there exist k-independent random variables with common distribution function F such that equality is attained in the first (resp. second) inequality (7). Proof. From Kemperman’s [
20] [Theorem 1, Remark] characterization of
k-independence, it follows that a random vector
has the same distribution as the random vector
, associated as in (
2) to some
k-independent random vector
with marginals
F, if and only if
takes values in
and satisfies the condition
with
,
as in (
4), and
. Observe that condition (
10) can be rewritten in the form
with
as in (
9) and
.
For any given function
, minimizing (resp. maximizing) the moment
is equivalent to minimizing (resp. maximizing)
, subject to the constraints that
takes values in
and satisfies the moment condition (
11). Since
if and only if
and
, by (
1) and (
2), one has
Hence, determining bounds on
reduces to minimizing (or maximizing)
subject to
and the moment constraint (
11).
We solve the latter problem using a geometric approach. Note that
, where
is given by (
8). Recall that
is the compact set of all possible moment points
where the distribution of
varies over all distributions on
.
Consider the auxiliary function
Since
is the compact set of all possible moment points
, where the random vector
takes its values in
the line segment
represents the range of possible moment points
where the distribution of
varies over all distributions on
satisfying the condition
The lower (resp. upper) endpoint belongs to the lower (resp. upper) envelope of
, defined by
(resp.
for
, providing the sharp lower (resp. upper) bound on
(cf. [
27]).
This completes the proof of (i). Statement (ii) follows directly from [
20] [Theorem 1]. □
Remark 1.
(i) The bounds for multivariate marginal distribution functions of order statistics provided by Theorem 1 are new even in the case that is, when the ’s are arbitrarily dependent identically distributed random variables.
(ii) Fix and let , and in (7). Then we recover Kemperman’s [20] bounds for single-order statistics from pairwise independent observations. If we fix and and take and in (7), we obtain more explicit expressions for the bounds [21] Equation (28) for linear combinations of distribution functions of single-order statistics from k-independent observations. (iii) From the proof of Theorem 1 it follows that the range of possible values of over all possible distributions of k-independent vectors with one-dimensional marginals F is equal to the interval where
Remark 2. The pointwise sharp distribution bounds (7) do not generally yield sharp expectation bounds (cf. [20]). However, the bounds obtained in this way may be accurate in certain particular cases. Indeed, suppose that and the k-independent ’s have a common piecewise constant distribution function F with jumps at points For example, for any fixed and a function , we havewhich can be expressed aswhereThus, applying (7) yields two-sided attainable bounds for , which are expressed in terms of the points and the values of the distribution function Sharp expectation bounds for L-statistics from arbitrarily dependent observations (i.e., for ) were presented in [16]. Note that for , the expectation of represents the net premium for the largest-claims (LC) reinsurance of the two largest claims in an individual model of homogeneous, and in particular arbitrarily dependent, risks (cf. [28]). In the context of LC reinsurance, conservative bounds are essential for risk assessment and pricing under unknown dependence structures. Our next objective is to establish an analogue of Theorem 1 for linear combinations of joint reliability functions associated with selected order statistics, defined by
where
are fixed real numbers, and
Theorem 2. Under the assumptions and notation of Theorem 1, the following statements hold.
(i) Let be a vector of k-independent random variables with a common distribution function F fulfilling condition (5). Thenwhere (ii) Moreover, for any distribution function F satisfying condition (5), there exist k-independent random variables with distribution function F for which the lower (resp. upper) bound in (13) is attained. Proof. Combining (
12) with the fact that
if and only if
, we obtain
An analysis analogous to that in the proof of Theorem 1 shows that determining bounds on
is equivalent to minimizing (maximizing) the quantity
where the vector
takes values in
and satisfies condition (
11). The proof is completed by adapting the second part of the proof of Theorem 1 with
replaced by
and
replaced by
□
Remark 3. If the ’s are not only k-independent but also non-negative, exchangeable and have no ties, then (13) provides sharp bounds for the joint reliability function of any pair of semi-coherent systems based on common components. Specifically,where is a probability matrix of order n depending solely on the system structure, known as the structure signature of the system (cf. [29]). To see this, observe thatwhere , andSharp bounds on the lifetime distributions and expectations of a single system with arbitrarily dependent exchangeable component lifetimes were derived in [30]. Remark 4. Theorems 1 and 2 reduce the problem of determining distribution bounds for order statistics from dependent observations to finding the convex hull of a finite set of points in , where . Computing the convex hull is a fundamental step in many practical problems, including statistical tasks such as robust estimation, isotonic regression, and clustering (see [31]). Even in moderate dimensions, such as 10 or 20, convex hull computation can be challenging. Knowledge of specific properties of the convex hull in a given problem, for example, the presence of symmetry, can be helpful (see [32]). A comprehensive overview of algorithms and methods for determining the convex hull of a finite set of points in is provided in [33], along with a detailed discussion of their respective advantages, disadvantages, and recommended applications.