1. Introduction
Copula models provide a flexible framework for modeling multivariate dependence by separating marginal distributions from the dependence structure, as guaranteed by Sklar’s theorem [
1,
2]. In many applications, however, the strength and form of dependence vary with observable covariates such as time, spatial location, or environmental conditions. Conditional copula models accommodate this heterogeneity by allowing the copula parameter to vary smoothly with covariates [
3,
4,
5].
Early work in this area includes the semiparametric conditional copula model of [
6], who proposed local pseudo-likelihood estimation with local polynomial approximation and established consistency, asymptotic normality, and bandwidth selection procedures. A related local likelihood framework for parametric conditional copulas was developed by [
3], who derived pointwise bias and variance expressions, introduced cross-validated copula selection, and constructed confidence intervals for covariate-dependent dependence parameters. Fully nonparametric approaches were studied by [
7], who analyzed the asymptotic properties of conditional copula estimators and associated dependence measures.
Let
denote a vector of covariates and
pseudo-observations obtained from continuous conditional marginals. A conditional copula model assumes that
where
is an unknown parameter function taking values in a parameter space,
.
The central inferential problem is the uniform estimation of the covariate-dependent copula parameter
. Local likelihood methods are particularly well suited for this task: rather than modeling
directly, one locally approximates a suitably transformed calibration function by a polynomial obtained from a Taylor expansion around each evaluation point and maximizes a kernel-weighted copula log-likelihood [
3,
8]. This approach, introduced by [
3], has been further developed in [
4]. Throughout this literature, including the present paper, the marginal distributions are assumed to be known, so that the pseudo-observations
U are treated as directly observed.
The assumption of known margins is standard in theoretical analyses of conditional copula models and is appropriate in several practically relevant settings, including cases where marginal models are specified a priori, where margins can be estimated at a faster parametric rate and treated as known in a second stage, or where inference focuses primarily on covariate-dependent dependence. Under these conditions, the copula likelihood provides a valid basis for inference on
, and marginal estimation error can be neglected at the level of first-order asymptotics. In empirical applications where marginal distributions depend on covariates, pseudo-observations that are approximately
conditional on
can be constructed by first removing the systematic effect of
from each margin and then applying an empirical CDF (rank) transformation to the adjusted observations, following [
3,
9,
10]. This two-step adjustment ensures compatibility with the copula modeling framework while preserving the focus on covariate-dependent dependence.
From a decision-theoretic perspective, the model with unknown margins strictly contains the known-margin experiment studied in this paper as a submodel. Consequently, the corresponding minimax risk in the larger experiment cannot be smaller than that in the restricted setting analyzed here. The lower bound established in the main result of this work (Theorem 1), therefore, remains valid under unknown margins, since any estimator in the larger experiment must, in particular, operate over the known-margin submodel.
The minimax lower bound proved below is information-theoretic and does not depend on the construction of a particular estimator. The additional nuisance estimation of the marginal distributions may affect constants or higher-order terms in the uniform rate, but it cannot invalidate the lower bound itself. When the margins are estimated at a
-consistent rate, standard two-step semiparametric arguments (see, e.g., [
11]) suggest that their impact should be negligible relative to the slower nonparametric rate
. A fully uniform-in-
y treatment of the joint margin–copula experiment is beyond the scope of the present work.
Related work has addressed the testing and estimation of covariate effects in conditional copula models. For example, [
12] proposed fully nonparametric tests of the simplifying assumption that covariates affect dependence only through the margins, while [
10] developed score-based tests for parametric specifications of covariate effects. Earlier work by [
13] established consistency and weak convergence of copula estimators under the simplifying assumption, and extensions to multivariate and functional covariates were studied in [
14].
From a theoretical perspective, early analyses of local likelihood methods focused on pointwise bias, variance, and asymptotic normality at fixed covariate values [
3]. However, uniform guarantees are essential in practice, as they underpin numerical stability, bandwidth selection based on global criteria, and simultaneous inference over covariate regions.
Recent work has developed a uniform asymptotic theory for kernel-weighted local likelihood estimation of covariate-dependent copula parameters, establishing uniform convergence of the local log-likelihood and its derivatives and deducing uniform consistency of the estimated parameter curve and dependence measures, such as Kendall’s
[
4]. The corresponding uniform upper bound for the estimation error
, together with the required regularity conditions, is established in the companion paper [
4]. Under standard smoothness and design conditions, the local likelihood estimator
satisfies [
4]
uniformly over compact
, where the logarithmic factor arises from entropy bounds for kernel-indexed function classes [
4,
15].
While such uniform upper bounds provide strong guarantees, they do not establish optimality. Minimax theory characterizes the intrinsic difficulty of an estimation problem by identifying the best achievable worst-case risk over a function class. Establishing minimax lower bounds is, therefore, essential to determine whether existing procedures attain optimal rates and to identify unavoidable sources of estimation error.
The goal of this paper is to derive minimax lower bounds for the uniform estimation of the calibration function in the conditional copula model (
1), under the known-margin framework described above. Working over Hölder classes of order
, we establish a lower bound for the minimax sup-norm risk over
. The resulting rate coincides with the classical minimax rate for the sup-norm estimation of a smooth regression function, showing that the nonlinear copula likelihood does not alter the fundamental global difficulty of the problem.
The proof is based on a localized packing construction combined with an information-theoretic testing argument of Fano–Le Cam type. The control of the Kullback–Leibler divergences between the induced joint distributions is achieved via second-order expansions of the conditional copula likelihood under mild curvature conditions. The argument is self-contained and relies on standard tools from empirical process theory and minimax analysis [
15,
16,
17]. To the best of our knowledge, this is the first minimax lower bound for uniform estimation in conditional copula models.
The remainder of the paper is organized as follows.
Section 2 introduces the conditional copula model and reviews the local likelihood framework that motivates our minimax analysis.
Section 3 defines the smoothness class, loss function, and standing regularity assumptions.
Section 4 presents the main result: a minimax lower bound for uniform estimation over compact covariate regions.
Section 5 reports a simulation study illustrating the finite-sample performance of local polynomial likelihood estimators under Clayton and Gumbel copulas and compares empirical uniform errors with the minimax benchmark rate.
Section 6 concludes with a discussion of the main implications and possible extensions.
Appendix A contains the complete proof of the minimax lower bound, including all auxiliary lemmas and the Fano–Le Cam testing argument with Kullback–Leibler control derived from likelihood curvature. The subsequent
Appendix B verifies Assumption (A3) by establishing quadratic mean differentiability and uniform Fisher information bounds for the bivariate Clayton and Gumbel copula families used in the simulation study.
5. Simulation Study
We illustrate the finite-sample behavior of the local likelihood estimator from
Section 2 in a setting aligned with the smoothness regime of the minimax analysis. We consider the bivariate case (
) with a single covariate (
) and employ a local linear fit (
), corresponding to a calibration function that is
Hölder smooth.
Bandwidth choice plays a central role both in practice and in theory. In estimation, the local likelihood estimator is implemented with a smoothing bandwidth,
, selected by leave-one-out cross-validated local likelihood (LOO-CVL), as defined in (
6). In contrast, the minimax lower-bound proof (see
Appendix A) introduces a theoretical localization scale through the support radius of the bump functions, tuned at the canonical rate
to balance the separation between alternative parameter functions and the control of the Kullback–Leibler divergence. Motivated by this construction, we define the minimax benchmark localization scale
which represents the minimax-optimal spatial resolution for uniform estimation.
The reader should note that is not a data-adaptive kernel smoothing bandwidth chosen for estimation but, rather, a theoretical localization scale arising from the minimax lower-bound argument. Its comparison with the data-driven LOO-CVL bandwidth is nevertheless informative, since both quantities govern the effective spatial resolution at which the local likelihood estimator can reliably detect variation in . Examining whether tracks this canonical rate (up to constants), therefore, provides an operational link between the minimax benchmark and a standard bandwidth selection rule used in applications.
In each replication, covariate values are generated as , a truncated normal distribution with mean 0, variance 4, and support restricted to . Estimation and evaluation are restricted to the interior region and carried out on an equally spaced grid of size to mitigate boundary effects.
Conditional on each observed covariate value,
y, we generate a single pseudo-observation,
, from a conditional copula model with parameter
. We consider both Clayton and Gumbel copula families, with smooth, non-constant calibration functions given by
These specifications induce moderate covariate-driven variation in dependence while remaining within standard parameter ranges for each family.
Estimation is performed using the kernel-weighted local likelihood procedure described in
Section 2, with the Epanechnikov kernel. For each bandwidth choice, the local likelihood is maximized at each grid point to obtain
, using an unconstrained parameterization and an inverse link mapping to ensure admissibility. We compare the localization scale
with the data-driven leave-one-out cross-validated bandwidth
defined in (
6).
The experiment is repeated over
independent replications for each sample size
. Performance is evaluated on the grid over
using both the sup-norm loss and a discrete
loss,
where
denotes the equally spaced evaluation grid on
. Thus,
is the root mean squared pointwise error across the grid points, i.e., a Riemann-sum approximation to an integrated squared error on
under the uniform measure on the grid.
To connect the numerical results to the theoretical analysis, we report the minimax benchmark rate from Theorem 1, specialized to and , and we compare the observed sup-norm errors to in rate plots.
Since minimax rates are defined only up to unknown positive multiplicative constants, agreement in slopes on a log–log scale (rather than absolute vertical alignment) is the appropriate diagnostic when comparing empirical errors with .
Figure 1 shows curve recovery at
for both copula families. The blue curve denotes the true calibration function
, while the orange and green curves represent the Monte Carlo mean of
obtained under the theoretical localization scale
and the data-driven LOO-CVL bandwidth
, respectively. For a fixed
N,
is deterministic, whereas
is selected in each replication. Both estimators closely track the smooth structure of
over the interior region
.
The shaded region corresponds to a pointwise 95% confidence band constructed only for the estimator using
. Following [
3], the asymptotic variance of the local polynomial likelihood estimator is
, where
with
and
denote the kernel moment matrices corresponding to the local linear fit. The resulting confidence interval is
. The band is shown for diagnostic purposes and is not intended for uniform inference.
Table 1 reports the localization scale
and the Monte Carlo mean of the LOO-CVL bandwidth,
, together with its Monte Carlo standard deviation. As implied by its definition,
is deterministic for each
N and decreases monotonically as
N increases. In contrast, the data-driven
is systematically larger than
for both copula families, reflecting the additional smoothing preferred by the finite-sample LOO-CVL objective.
Figure 2 visualizes the theoretical localization scale
and the Monte Carlo mean LOO-CVL bandwidth
across
for both copula families.
Table 2 and
Table 3 report finite-sample performance based on
Monte Carlo replications for
. For each sample size
N, the columns
and
denote the Monte Carlo mean of the sup-norm loss
evaluated on
when the estimator is computed using the benchmark localization scale
and the data-driven LOO-CVL bandwidth
, respectively. The corresponding columns
and
report the Monte Carlo standard deviations of these losses across replications. Analogously,
and
denote the Monte Carlo mean of the discrete
loss over
under the two smoothing choices, with
and
giving the associated Monte Carlo standard deviations.
Across both copula families, mean errors decrease monotonically with N under both and , and the associated Monte Carlo standard deviations also decline, reflecting improved stability in larger samples.
For the Clayton copula, yields uniformly smaller mean errors than the theoretical localization scale under both loss metrics at all reported N. The same pattern holds for the Gumbel copula: the LOO-CVL choice produces smaller mean sup-norm and errors at every sample size, although the magnitude of improvement is moderate. Overall, the errors decrease monotonically with N in both families, and the LOO-CVL rule typically achieves modest gains in accuracy without inflating variability.
The decay of estimation error with
N is summarized in
Figure 3,
Figure 4 and
Figure 5. In
Figure 3 and
Figure 4, the vertical axis is log-scaled to highlight rate behavior, while the horizontal axis displays the raw sample size
N. For both copulas and both loss metrics, the Monte Carlo mean errors decrease steadily with
N. As reflected in
Table 2 and
Table 3, the LOO-CVL choice
yields uniformly smaller mean errors than the theoretical localization scale
at all reported sample sizes for both Clayton and Gumbel.
Figure 5 compares the Monte Carlo mean sup norm error under the LOO CVL bandwidth to the minimax benchmark rate
. In both copula families, the empirical error decays at a rate comparable to
across the considered sample sizes. Because minimax rates are defined up to multiplicative constants, agreement is assessed through scaling behavior, rather than vertical alignment. Finite sample constants and lower-order logarithmic factors can generate visible vertical separation, so agreement is assessed through slope, that is, the rate of decay, rather than the exact vertical coincidence. The observed slopes, therefore, support agreement with the predicted minimax scaling, while the remaining gap reflects finite sample effects.
We emphasize that minimax lower bounds describe worst-case asymptotic scaling over the smoothness class and do not determine the finite sample ordering of specific smoothing choices. Both and operate at the same asymptotic rate, while the observed differences reflect finite sample constants and bias variance tradeoffs.