Abstract
In this study, we look at the wavelet basis for the nonparametric estimation of density and regression functions for continuous functional stationary processes in Hilbert space. The mean integrated squared error for a small subset is established. We employ a martingale approach to obtain the asymptotic properties of these wavelet estimators. These findings are established under rather broad assumptions. All we assume about the data is that they are ergodic, but beyond that, we make no assumptions. In this paper, the mean integrated squared error findings in the independence or mixing setting were generalized to the ergodic setting. The theoretical results presented in this study are (or will be) valuable resources for various cutting-edge functional data analysis applications. Applications include conditional distribution, conditional quantile, entropy, and curve discrimination.
Keywords:
multivariate regression estimation; multivariate density estimation; stationarity; ergodicity; rates of strong convergence; wavelet-based estimators; martingale differences; continuous time series MSC:
62G07; 62G08; 62G05; 62G20; 62H05; 60G42; 60G46
1. Introduction and Motivations
In recent years, the statistical literature has become increasingly interested in statistical issues pertaining to the study of functional random variables or variables with values in an infinite-dimensional space. The availability of data measured on ever-finer temporal/spatial grids, such as in meteorology, medicine, and satellite images, is driving the expansion of this research topic; statistical modeling of these data as random functions uncovered numerous complex theoretical and numerical research challenges. The reader may consult the monographs to summarize functional data analysis’s theoretical and practical aspects. [1] for linear models for random variables with values in a Hilbert space, [2] for scalar-on-function and function-on-function linear models, functional principal component analysis, and parametric discriminant analysis. [3], on the other hand, concentrated on nonparametric methods, particularly kernel-type estimation for scalar-on-function nonlinear regression models. Such tools were extended to classification and discrimination analysis. [4] discussed the application of several interesting statistical concepts to the functional data framework, including goodness-of-fit tests, portmanteau tests, and change point problems. [5] was interested in analyzing variance for functional data, whereas [6] was more concerned with regression analysis for gaussian processes. Recent studies and surveys on functional data modeling and analysis can be found in [7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22].
The subject of estimating conditional models has been extensively studied in the statistical literature, utilizing several estimation techniques, the most prevalent of which is the conventional kernel method. This is because these methods have broad applications and play an important role in statistical inference. However, such methods may have certain drawbacks when predicting compactly supported or discontinuous curves at boundary locations. The prevalence of alternative wavelet approaches can be attributed to the adaptability of these methods to discontinuities in the curve that is to be approximated. In application, the wavelet method offers a straightforward estimating algorithm to both implement and compute. Refs. [23,24,25,26] provided more information on wavelet theory. We quote the work of [27], which discusses the properties of wavelet approximation and analyzes the application of wavelets in various curve estimation issues in depth. In [28], numerous applications of wavelet theory are described for the independent unidimensional case, with an emphasis on calculating the integrated squared derivative density function. The findings of [28] were later extended by [29] to estimate the density derivatives for negatively and positively related sequences, respectively. In [30], wavelet estimators for partial derivatives of a multivariate probability density function were constructed, and convergence rates for an independence scenario could be determined. In detail, [31] discusses the estimation of partial derivatives of a multivariate probability density function in the presence of additive noise. We could consult [32,33] for the most recent information on this subject. Using the independent and identically distributed paradigm, [34] examined density and regression estimation issues unique to functional data. Ref. [34] proposed a novel adaptive method based on the term-by-term selection of wavelet coefficient estimators and wavelet bases for Hilbert spaces of functions.
The primary purpose of this paper is to provide further context for the earlier discussion of stationary ergodic processes. The lack of research on the general dependence framework for wavelet analysis prompted us to conduct the present study. Several arguments for contemplating an ergodic dependency structure in that data as opposed to a mixing structure are presented in [35,36,37,38,39,40,41,42], where further information on the notion of the ergodic property of processes and examples of such processes are provided. In [43], one of the arguments used to justify the ergodic setting is that establishing ergodic characteristics rather than the mixing condition can be significantly simpler for some classes of processes. Therefore, the ergodicity hypothesis provides the ideal framework for examining data series created by chaotic noise.
In the body of statistical research that has been published, estimation issues based on discretely sampled observations as well as continuous time have been investigated. In finance, even if the underlying process is continuous across time, only its values at a finite number of points are known. The first one is more important from the standpoint of financial econometrics (see [44,45]). On the other hand, the technical development of statistical inference using a continuous time record of observations is far easier to accomplish than an inference based on discretely observed data. It makes it possible to advance further in the model’s statistical analysis and find solutions to various concerns concerning discretely observed diffusions that had not been resolved until this point. Further, remember that the discrete-time model hits its limits when the discretization step equals zero, which is when the continuous-time model becomes the limit (see [46]). As a result, the asymptotic behavior of estimate processes in actuality may be close to the asymptotic behavior established theoretically for continuous-time observations if the available data are “dense enough” relative to the observation period. The knowledge of the most accurate estimate derived from continuous-time data is also of practical significance. Ref. [47] investigated the challenge of estimating an invariant density function for a continuous-time Markov process and established the mean-square consistency of kernel-type estimators. Ref. [48] demonstrated the homogeneity and consistency of these estimators. Ref. [48] showed that a family of smoothed estimators, including kernel-type estimators, are asymptotically normal with a rate of convergence for continuous-time processes. This finding is unexpected because the invariant density estimators operate differently in this instance than in discrete-time processes, where the convergence rate is often smaller than . Numerous physical phenomena, such as seismic waves, the Earth’s magnetic field, and isotherms and isobars in meteorology, are functional variables observable in continuous time, as is evident.
In [49], we have considered the wavelet basis for the nonparametric estimation of density and regression functions for continuous functional stationary processes in Hilbert space. We have characterized the mean integrated square errors. This research aims to provide the first complete theoretical rationale for wavelet-based functional density and regression function estimation for continuous stationary processes by extending our earlier work [49] to the continuous setting. The employment of wavelet estimators in functional ergodic data frames and the resulting difficulty of establishing the mean integrated square error over suitable decomposition spaces is, to the best of our knowledge, still an unresolved topic in the literature. By merging different martingale theory techniques utilized in the mathematical construction of the proofs, we aim to contribute to the literature by addressing this gap. In the independent or mixing setting, the tools employed for regression estimation change significantly in the ergodic setting. Nevertheless, we shall see that more than merely mixing preexisting ideas and outcomes is needed to address the issue. In order to deal with wavelet estimators in an ergodic setting, one must resort to complex mathematical derivations.
The paper’s structure is as follows: The multiresolution analysis is described in Section 2. The principal density estimate results are presented in Section 3. Section 4 summarizes the principal outcomes for regression estimation. Section 5 provides some examples of potential applications. There are some concluding observations in Section 6. All proof is compiled in Section 7.
2. Multiresolution Analysis
Following [34,49,50], we will now introduce some basic notations for defining wavelet bases for Hilbert spaces of functions with a few modifications to accommodate our context. In this study, nonlinear, thresholded, wavelet-based estimators are examined. Beginning with a description of the fundamental theory of wavelet approaches, we then introduce nonlinear wavelet-based estimators. The interested reader should consult [23,24], see also [51,52] and the references therein, despite the fact that the wavelet is based on a separable Hilbert space of real or complex-valued functions on a complete separable metric space. Let represent the separable Hilbert space of real-valued functions defined on a separable complete metric space . Given that is separable; it possesses an orthonormal basis
where is a countable index set. The space is equipped with an inner product and a norm . Consider the sequence of subsets an increasing sequence of finite subsets of such that
The subset denotes the orthogonal complement of in , i.e.,
Choose, for any , , and , , such that the following matrices
fulfill one of the two following conditions, for instance, see [34,50] and the references therein.
- (A.1)
- and where and for and are positive constants.
- (A.2)
- and where and for and are positive constants.
Condition (A.1) implies that
which means that all the columns of and are not zero vectors. As for (A.2), it gives
saying that all the rows of and are not zero vectors. For any , we set
where
The following collection serves as the orthonormal basis for (see Theorem 2 of [50]):
For further details, see [34,50,53]. Hence, we infer that for any , we have
where
Adding two additional assumptions to the orthonormal basis :
- (E.1)
- There exists a constant such that, for any integer , we have
- (i)
- (ii)
- (E.2)
- There exists a constant such that, for any integer , we have
Remark 1.
Clearly, the assumption (E.1) is satisfied under assumption (A.1) whenever ; we may also consult [50], Example 2 and its applications for further details. [50,53] have provided three examples satisfying condition (E.2) taking
see also [53] Theorem 3.2. Moreover, [34] has used both assumptions in the independent and identically distributed functional data.
Besov Space
Over the years, numerous statisticians have pondered the following question: given an estimating method and a required estimation rate for a certain loss function, what is the largest space over which this rate can be accomplished, for example, see [27,54] and the references therein. We are interested in estimating methods based on thresholding methods given by wavelet bases in a natural situation. It is common knowledge that wavelet bases serve to characterize the smoothness of spaces such as the Hölder spaces , Sobolev spaces , and Besov spaces for a range of indices s that depend on both the smoothness properties that are dependent on both the smoothness of and its dual function , for instance, see [54] for more detail and examples, at this point, we may consult [55]. The following statistical definition is used in approximation theory for the study of nonlinear procedures, such as thresholding and greedy algorithms; see [27,52,54,56].
Definition 1
(Besov space). Let . We say that the function , defined by statement (7), belongs to the Besov space if and only if:
Definition 2
(Weak Besov space). Let . We say that the function , defined by statement (7), belongs to the weak Besov space if and only if:
3. The Density Estimation
Let denote a sequence of strictly stationary ergodic pairs of random elements, where takes values in a complete separable metric space of Hilbert space associated with the corresponding Borel -algebra and is a real or complex-valued variable. Let denote the probability measure induced by on . Assume that there exists -finite measure on the measurable space in such a way that is dominated by . The Radon–Nikodym theorem guarantees the existence of a measurable function that is nonnegative in such a way that
In this framework, we intend to estimate on the basis of n observed functional data . We assume that , where is a separable Hilbert space of real or complex-valued functions defined on and square integrable with respect to the -finite measure . In this research, we are especially interested in the wavelet estimation processes created in the 1990s, see Meyer’s work for the functional data of a Hilbert space and, more specifically, the nonlinear estimators. The majority of this model’s approaches involve introducing kernel estimator techniques to estimate the functional component of the model, see [57]. Let be the sample’s common density function , which is assumed to be
- (F.1)
- is a known constant such that
Density Function Estimator
From now, suppose that the density function is a separable Hilbert space. Then satisfies the wavelet representation (7). Assume that we observe a sequence of copies of that is supposed to be functional stationary and ergodic with with the density function . We examine density estimation utilizing wavelet bases for Hilbert spaces of functions of [50]. We consider the estimated coefficients and given, respectively, by (14) and (15), for any . Here the resolution level at the rate specified below. Since we assume that and have a compact support so that the summations in (7) are finite for each fixed (note that in this case, the support of and is a monotonically increasing function of their degree of differentiability [24]). We focus our attention on the nonlinear estimators (13), which will be studied in the mean integrated squared error over adapted decomposition spaces, in a similar way as in [34] in the setting of the independent and identically distributed functional processes, in particular, we refer to [49]. The density wavelet hard thresholding estimator is defined for all , by
where
Here denotes a large-enough constant, and is the integer satisfying
See [36,38] for further details on the multivariate case.
Comments on the Method of Estimation
Our estimation method is divided into three steps:
- 1.
- 2.
- The greatest is selected by applying hard thresholding;
- 3.
- Reconstruction of the selected elements of the initial wavelet basis.
It is essential to highlight that our choice is the universal threshold and the definition of is based on theoretical considerations. The estimator under consideration is not dependent on the smoothness of ; see [53] for more details in the case of the linear wavelet estimator of . Moreover, for additional information on the case of and more standard nonparametric models, refer to [27,58]. Notation is necessary for stating the results. Throughout the paper, we shall denote by the open set of the Borel algebra . For any and small real, we define
and
as the distribution function and the conditional distribution function, given the field , respectively. Before presenting our results, we present supplementary notation and our hypotheses. For the remainder of the paper, for a positive real , we will denote by
Let be the field generated by
Let be the field generated by
and
The following assumptions are required for the entire paper.
- (C.0)
- There is a nonnegative measurable function such thatWe may refer to [3,35,59] for further details.
- (C.1)
- For anyAt this point, we may refer to [60] for further details.
Comments on hypotheses. Approximating the integral by its Riemann’s sum, we have
By the fact that process is stationary and ergodic, in a similar way as in [61] (see, Lemma 4 and Corollary 1 together with their proofs), one may establish that the sequence
of random functions is stationary and ergodic. Indeed, it suffices to substitute the conditional densities in the work of Delecroix with and the density by function . Refer to [35,62] and the references therein.
Theorem 1.
there is a constant such that
for a large-enough T.
The following upper bound finding is an immediate consequence: if for , then there is a constant in such a way that
This convergence rate is close to optimal in the “standard” minimax configuration (see, [27]). Moreover, on the application of [52] (Theorem 3.2), one obtains that is the “maxiset” associated with at the convergence rate of , i.e.,
4. The Regression Estimation
Let be a measurable function. The regression function is defined by
where is a random variable independent of with . We assume that , where is a separable Hilbert space of real or complex-valued functions defined on and a square-integrable with respect to the -finite measure . We shall assume that there is a known constant and in such a way that
In this framework, we redefine the probability measure in (11) and assume that is a nonnegative measurable known function.
- (M.1)
- We shall assume that there is a known constant in such a way that
- (M.2)
- We shall suppose that there is a known constant in such a way that
Regression Function Estimator
In this framework, our goal is to estimate based on observed functional data . The kernel estimator for the regression function of functional data has been suggested by [59]
This estimator is similar to the one proposed by [59] in the discrete framework. By using the idea of [58], we propose the hard wavelet thresholding estimator , for all , by
where
where is the integer satisfying
and is a large-enough constant. The multivariate case for discrete- and continuous-time linear wavelet estimators was examined by [36,38].
Theorem 2.
Under the assumptions (E.1), (E.2) (M.1), (M.2), (C.0), and (C.1), combined with condition (24), for any , , there is a constant such that
for T large enough.
Assume that and fulfill (M.1) and, for any
where is in (Definition 1) with and is in (Definition 2) with , then there is a constant in such a way that
for T large enough. Again, note that for , if
then there is a constant in such a way that
This convergence rate is close to optimal in the “standard” minimax framework (see, [27]) up to an extra logarithmic term. According to our knowledge, Theorem 2 is the first one to examine an adaptive wavelet-based estimator for functional data in the context of nonparametric regression for an ergodic process. Ref. [49] studied the same estimator in a discrete time setting.
From the fact that the coefficients defined in Equations (22) and (23) depend on the unknown function , it is possible to use
The wavelet hard thresholding estimator is defined by , for all , by
Keep in mind the following elementary observation
We can infer, from the last equation, the following
Notice that
This gives
Under the conditions of Theorems 1 and 2, one can find a positive constant such that
A combination of Theorem 1 with Theorem 2 gives the following corollary.
Corollary 1.
Under the hypotheses of Theorems 1 and 2, there exists a constant in such a way that
for T large enough.
Remark 2.
Let be a stationary sequence. Consider the backward field and the forward field . The sequence is strongly mixing if
The sequence is ergodic if
where τ is the shift transformation or time-evolution. The labeling of strong mixing in the preceding definition is more rigorous than what is commonly referred to as strong mixing (when using the terminology of measure-preserving dynamical systems), meaning that it satisfies the following conditions
for any two measurable sets , see, for instance, [63]. Therefore, substantial mixing requires ergodicity, although the opposite is not necessarily true (see, for instance, Remark 2.6 on page 50 about Proposition 2.8 on page 51 in [64]).
Remark 3.
Ref. [43] provided an example of an ergodic but non-mixing process in their discussion, which can be summarized as follows: Let be a strictly stationary process such that is a Poisson process with parameter , where is the -field generated by . Assume
and is a given function. This process is not mixing in general (see Remark 3 of [65]). It is known that any sequence of independent and identically distributed random variables is ergodic. Hence, according to Proposition 2.10 in [64], it is easy to see that with
for some Borel-measurable function . Ref. [66] has constructed an example of a non-mixing ergodic continuous-time process. It is well known that the fractional Brownian motion with parameter has strictly stationary increments. Otherwise, the fractional Gaussian noise, defined for every by
is a strictly stationary-centered, long memory process when (for instance, see [67] p. 55, [68] p. 17), hence the condition of strong mixing is not satisfied. Let be a strictly stationary-centered Gaussian process with correlation function
Relying on the work of [69], Lemma 4.2, it follows that the process is ergodic whenever
which is the case for the process .
Remark 4.
In continuous time, sampling is frequently used to obtain data. Several discretization strategies, including deterministic and random sampling, have been proposed in the literature. The interested reader is referred to [70,71,72,73,74]. To simplify the concept, we consider the density estimator of based on and its sampled discrete sequence . The estimator of the sampled density is
As stated in [70], we can only recollect two designs: irregular sampling and random sampling.
- Deterministic sampling.Consider the situation where is deterministically irregularly spaced withfor some . the σ-field generated by . Clearly, in a family of increasing σ-fields.
- Random sampling. Assume that the instants in the interval form a sequence of uniform random variables independent of the process . Defineas the corresponding order statistics. Observe that are the observation points for the procedure. Clearly, all of the distances between these sites are positive. As a consequence, taking the σ-field generated by , it follows that is a sequence of increasing σ-fields.
As established in [75], the penalization approach for the choice of the mesh δ of the observations provides an optimal rate of convergence; we leave this subject open for future investigation within the context of ergodic processes.
Remark 5.
In a previous publication [36], we tackled the nonparametric estimation of the density and regression function in a finite-dimensional setting using an orthonormal wavelet basis. Our findings differ significantly from those presented in this publication. In [36], we demonstrated the strong uniform consistency characteristics of these estimators over compact subsets of under a general ergodic condition on the underlying processes. In addition, we demonstrate the asymptotic normality of wavelet-based estimators. The Burkholder–Rosenthal inequality, a more complex technique than the exponential inequality used in the previous publication, was the central concept of this study. Significantly, the current study studies the mean integrated square error over compact subsets, which is fundamentally different from the conclusions of the prior publication.
5. Applications
5.1. The Conditional Distribution
Our findings can be utilized to examine the conditional distribution for . To be more precise, let . The wavelet’s hard thresholding estimator of , is defined, for all , by
where
A direct consequence of Theorem 2 is
5.2. The Conditional Quantile
Remark that whenever is strictly increasing and continuous in a neighborhood of , the function has a unique quantile of order at a point , which is . In this situation
which can be estimated by . Consequently, the fact
and is continuous and strictly increasing, we then have
implying that,
In addition, suppose that, for fixed is differentiable at with
where is a real number, and is uniformly continuous for all . From Taylor’s expansion of the function in the neighborhood of , we have
where is a point between and . Using the fact that converges a.s. toward as T goes to infinity, combined with the uniform continuity of , allows us to write that
By the fact that is uniformly bounded from below, we can then claim that
Remark 6.
(Expectile regression). For , the choice given by leads to quantities called expectiles by [76]. Expectiles, as defined by [76], may be introduced either as a generalization of the mean or as an alternative to quantiles. Indeed, classical regression provides us with a high sensitivity to extreme values, allowing for more reactive risk management. Quantile regression, on the other hand, provides the ability to acquire exhaustive information on the effect of the explanatory variable on the response variable by examining its conditional distribution; refer to [77,78,79] for further details on expectiles in functional data settings.
Remark 7.
(Conditional winsorized mean). As in [80], if we consider if , , or , then will be the conditional winsorized mean. Notably, this parameter was not considered in the literature on nonparametric functional data analysis involving wavelet estimators. Our paper offers asymptotic results for the conditional winsorized mean when the covariates are functions.
5.3. Shannon’s Entropy
The differential (or Shannon) entropy of is defined to be
whenever this integral is meaningful. We will apply the convention since as . The notion of differential entropy was originally introduced in [81]. Since this epoch, the concept of entropy has been the topic of extensive theoretical and applied research. We are referencing [82] (Chapter 8) for a thorough overview of differential entropy and its applications and mathematical characteristics. Entropy concepts and principles play a fundamental role in many applications, such as quantization theory [83], statistical decision theory [84], and contingency table analysis [85]. Ref. [86] introduced the concept of convergence in entropy and showed that the latter convergence concept implies convergence in . This property indicates that entropy is a useful concept for measuring “closeness in distribution”, and also heuristically justifies the usage of sample entropy as a test statistic when designing entropy-based tests of goodness-of-fit. This line of research has been pursued by [87,88,89] (including the references therein). The idea here is that many families of distributions are characterized by maximization of entropy subject to constraints (see [90,91]). Given in Equation (13), we estimate using representation (35), by setting
where
and is a sequence of positive constants. To prove the strong consistency of , we shall consider another, but more appropriate and more computationally convenient, centering factor than the expectation , which is delicate to handle. This is given by
We first decompose , as in [92,93,94,95], into the sum of two components, by writing
We remark that for all ,
Thus, for any , we obtain
Making use of Theorem 1, gives
for positive constant C. Note that this is the first estimation result for the entropy of functional continuous time series processes. A similar idea can be used to estimate other density functions, such as the extropy; refer to [96] for a definition.
5.4. The Curve Discrimination
The curve discrimination problem can be stated as follows. Let denote a sample of curves, and each of them is known to pertain to one among groups . We denote the group of the curve by . Suppose that each pair of variables has the same distribution as the pair . Given a new curve , the question is to determine its class membership. To do so, we will estimate the conditional probability for every , as follows:
Following the proposal made in [97,98] permitting the estimation of these probabilities by
where
As remarked by [97,98], for each we make use of the notation
then we can write
An application of Theorem 2 is
6. Concluding Remarks
In this study, we investigated the nonparametric estimation of density and regression function using continuous functional stationary processes and wavelet bases for Hilbert spaces of functions. We described the mean square error integrated over compact subsets. The martingale method was used to determine the asymptotic properties of these estimators, which differs significantly from the mixing and independent settings. Ergodicity is the assumption of the process’s dependence. Extending nonparametric functional concepts to local stationary processes is a relatively new area of research. It would be interesting to extend our work to the case of the functional local stationary process, but this would require nontrivial mathematics and is well outside the scope of this paper. The exact logarithm rates of convergence depend on the smoothness parameters of functions and defined in the space . These results are typical of all nonparametric estimations and are consistent with the vast majority of academic references. Although is unknown, the smoothness metrics assumed here are more nuanced than the differentiability criterion of integer orders required for convolution kernel methods. To make the estimator more practical, techniques have been developed to select the optimal adaptive value of , such that the most commonly used methods of “Stein,” the “rule of thumb,” and “cross-validation”, we refer to [99] and [27] for details on an asymptotically optimal empirical bandwidth selection rule. It would be worthwhile to investigate this subject in the future.
7. Proofs
This section is dedicated to proving our findings. The previously presented notations are used again in the sequel. In this research, we need an upper bound inequality for partial sums of unbounded martingale differences to derive the asymptotic results for the estimations of the density and regression functions based on strictly stationary and ergodic functional data. Here and in the following, “C” denotes a positive constant that may vary from line to line. The following lemmas express this inequality.
Lemma 1.
(Burkholder–Rosenthal inequality) Following Notation1in [100].
Let be a stationary martingale adapted to the filtration , define as the sequence of martingale differences adapted to and
then for any positive integer n,
where, as usual, the norm .
Lemma 2
([101]). Let denote a sequence of martingale differences in such a way that
then, for all and all n large enough, we obtain
The subsequent lemmas characterize the asymptotic behavior of estimators and .
Lemma 3.
For each and each , under conditions (C.0), (C.1), (F.1), and (E.1)(i), there is a constant in such a way that
Lemma 4.
For each and each , and under conditions (C.0), (C.1), (F.1), (E.1), and (E.2), and condition (16), there is a constant in such a way that
Lemma 5.
For each and each , for sufficiently large and under conditions (C.0), (C.1), (F.1), (E.1), and (E.2), and assumption (16) there is a constant in such a way that
Proof of Theorem 1.
Keep in mind that the proof of Theorem 1 is a straightforward application is a direct application of [52] (Theorem 3.1), using Lemmas 3–5 alongside , , . We modified and expanded the approach used to prove Theorem 3.1 of [34], and to include the stationary ergodic process. □
Proof of Theorem 3.
Consider the subsequent decomposition
where
Under the assumptions (C.0) and (C.2), we have
We readily obtain
implying that
Therefore, we infer that
We now consider the term . We infer
where
Notice that, with respect to the sequence of fields , is a sequence of martingale differences. It is evident, by Lemma 1, to infer that
where
Observe that for all , the -field represents the trivial one. On the one hand, by using the classical decomposition in connection with the fact that is the trivial field, we have
Remark that, under the conditions (F.1) and (E.1)(i) and the fact that is an orthonormal basis of H, we obtain
where is a positive constant,
therefore,
Furthermore, we examine the second term of decomposition (49), and remark that
using the notable identity combined with Jensen’s inequality, we obtain
observe that, using (50), we have
Hence,
Therefore, there exists a constant , such that
Hence the proof is complete. □
Proof of Lemma 4.
Consider the following decomposition
where
Remark that, under conditions (F.1) and (E.1)(i) and making use of the fact that is an orthonormal basis of H, and reasoning in a similar way as in Equation (47), we infer that
This, in turn, implies that
Therefore, we obtain
Hence, we readily infer
where
Remark that, with respect to the sequence of fields , is a sequence of martingale differences, an application of the Burkholder–Rosenthal inequality (see Lemma 1), gives
Consider the first term of Equation (59). We recall that for all , the -field represents the trivial one. Applying Jensen’s inequality, we have
Using the Minkowski inequality, we obtain
This gives that
By reasoning in a similar way as in Equation (50) and making use of conditions (F.1) and (E.1)(i), for all , we infer that
where C denotes a positive constant. In addition, by the Cauchy–Schwarz inequality in combination with conditions (E.1)(ii), (E.2) and condition (16), we infer
We then obtain
Recall that , we deduce that
We now consider the upper bound of in Equation (59). Remark that
Observe that for all for all ; we have . Making use of Jensen and Minkowski’s inequality, it follows that
For all , using the stationarity of the process , we have
Remark, under conditions (F.1), (E.1)(i), (C.0), and (C.1) and statement (61), that
It follows that
We conclude that
This implies that there exists a constant , such that
The proof is achieved. □
Proof of Lemma 5.
Consider the previous decomposition in Lemma 4, to write that
where
Making use of Equation (57), we achieve the desired result for the term
We now remark
The application of Lemma 2 implies that
Let . Then, for all where n is sufficiently large, we have
where
By choosing such that we have
The proof of (45) is achieved. □
Recall that
where
Lemma 6.
For each and each and under conditions (E.1)(i), (M.1), (M.2), (C.0), and (C.1), there is a constant in such a way that
Lemma 7.
For each and each , and under conditions (E.1), (E.2) (M.1), (M.2), (C.0), and (C.1), in combination with condition (24), there is a constant in such a way that
Lemma 8.
For each and each , for sufficiently large, (E.1), (E.2) (M.1), (M.2), (C.0), and (C.1), in combination with condition (24), there is a constant in such a way that
Proof of Lemma 2.
Remark that the proof of Theorem 2 is a consequence of [52] (Theorem 3.1) with , , in connection with Lemmas 6–8. We generalized the method of the proof in [34] (Theorem 4.1). □
Proof of Lemma 6.
Consider the subsequent decomposition
where
For all ; we have and from the independence between and , we have
Observe that
This implies that
Using conditions (M.1), (M.2), (C.0), and (C.1), we obtain
We readily obtain that
implying that
Therefore, we infer that
Remark that, with respect to the sequence of fields , is a sequence of martingale differences. It is obvious, going with the proof of Equation (54), to observe that
where
On the one hand, using Jensen and Minkowski’s inequalities combined, we obtain
It follows, from assumptions (M.1) and (M.2), that
combined with the independence between and , . Remark that, under condition (E.1)(i) and using the fact that is an orthonormal basis of H, we infer
Therefore, we have
Furthermore, we examine the second term of decomposition (79) and continue as per the proof of (53) and considering (81), one obtains
therefore, combining Equations (82) and (83), we obtain
Hence, there exists a positive constant C such that
Thus, the proof of Lemma 6 is completed. □
Proof of Lemma 7.
Consider the subsequent decomposition
where
Remark that, under conditions (M.1), (M.2), and (C.1), Equation (76), and using the fact that is an orthonormal basis of H, by reasoning as in Equation (77), we obtain
giving that
Therefore, we have
Hence we obtain
where
Notice that is a sequence of martingale differences with respect to the sequence of fields , applying the Burkholder–Rosenthal inequality (see Lemma 1), we obtain
Consider the first term of the Equation (90). Applying Jensen and Minkowski’s inequalities, we have
By proceeding as in Equation (81) and under the same conditions (M.1), (M.2), and (E.1)(i), for all , we obtain
where C is a positive constant. Moreover, by the Cauchy–Schwarz inequality in connection with conditions (E.1)(ii) and (E.2) and assumption (24), we have
It follows that
Let us now examine the upper bound of of Equation (59). Remark that
Observe that for all for all ; we have
Making use of Jensen and Minkowski’s inequality in the same manner as in Equation (91), it follows
From the independence between and , for all , we have
For all , using the stationarity of the process , we have
We deduce
where
It follows
Hence there exists a constant , such that
The proof is achieved. □
Proof of Lemma 8.
Statement (88) achieves the desired result for the term
We consider the next decomposition
where
and
and represents a constant that will be selected later. Remark that
where
First, we aim to bound the term of Equation (103). The Markov inequality and the Cauchy–Schwarz inequality yield
Observe that
Using Equation (92) combined with an elementary Gaussian inequality and taking to have
We obtain
where Now, we are going to look into determining an upper bound for from decomposition (103). We begin by confirming Lemma’s 2 conditions. Assume that conditions (M.1) and (M.2) are fulfilled in connection with Equation (93), we infer
which implies
where , let , then, for all with a sufficiently large T, we have
where
Taking such that we have
The proof of Equation (74) is achieved. □
8. Besov Spaces
In terms of wavelet coefficients, in [23], the Besov space was described as follows: is a real-valued smoothness parameter of , then equivalent to
- (B.1)
- (B.2)
with
and
and classical sup-norm modification for . The Besov spaces are useful for describing the smoothness properties of functional estimation and approximation theory. They contain well-known statistical research spaces, such as the Hilbert–Sobolev space , Holder space for , and others. We refer readers to [58] for additional descriptions of Besov space and its merits in approximation theory and statistics. Reasoning as in [102], suppose that Let
For , set
For set
For and define
For put with an integer and Define to be the space of functions in such that for all The norm is defined by
Author Contributions
S.D. and S.B.: conceptualization, methodology, investigation, writing—original draft, and writing—review and Editing. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
Not applicable.
Acknowledgments
The authors would like to thank the Special Issue Editor of the Special Issue on “Functional Data Analysis: Theory and Applications to Different Scenarios”, Mustapha Rachdi, for the invitation. The authors are very grateful to the three referees whose comments helped in improving the quality of this work.
Conflicts of Interest
The authors declare that they have no competing interests.
References
- Bosq, D. Linear processes in function spaces. In Lecture Notes in Statistics; Springer: New York, NY, USA, 2000; Volume 149, pp. xiv+283. [Google Scholar] [CrossRef]
- Ramsay, J.O.; Silverman, B.W. Functional Data Analysis, 2nd ed.; Springer Series in Statistics; Springer: New York, NY, USA, 2005; pp. xx+426. [Google Scholar]
- Ferraty, F.; Vieu, P. Nonparametric Functional Data Analysis; Springer Series in Statistics; Springer: New York, NY, USA, 2006; pp. xx+258. [Google Scholar]
- Horváth, L.; Kokoszka, P. Inference for Functional Data with Applications; Springer Series in Statistics; Springer: New York, NY, USA, 2012; pp. xiv+422. [Google Scholar] [CrossRef]
- Zhang, J.T. Analysis of variance for functional data. In Monographs on Statistics and Applied Probability; CRC Press: Boca Raton, FL, USA, 2014; Volume 127, pp. xxiv+386. [Google Scholar]
- Shi, J.Q.; Choi, T. Gaussian Process Regression Analysis for Functional Data; CRC Press: Boca Raton, FL, USA, 2011; pp. xx+196. [Google Scholar]
- Geenens, G. Curse of dimensionality and related issues in nonparametric functional regression. Stat. Surv. 2011, 5, 30–43. [Google Scholar] [CrossRef]
- Cuevas, A. A partial overview of the theory of statistics with functional data. J. Statist. Plann. Inference 2014, 147, 1–23. [Google Scholar] [CrossRef]
- Shang, H.L. A survey of functional principal component analysis. AStA Adv. Stat. Anal. 2014, 98, 121–142. [Google Scholar] [CrossRef]
- Horváth, L.; Rice, G. An introduction to functional data analysis and a principal component approach for testing the equality of mean curves. Rev. Mat. Complut. 2015, 28, 505–548. [Google Scholar] [CrossRef]
- Müller, H.G. Peter Hall, functional data analysis and random objects. Ann. Statist. 2016, 44, 1867–1887. [Google Scholar] [CrossRef]
- Nagy, S. An overview of consistency results for depth functionals. In Functional Statistics and Related Fields; Springer: Cham, Switzerland, 2017; pp. 189–196. [Google Scholar]
- Vieu, P. On dimension reduction models for functional data. Statist. Probab. Lett. 2018, 136, 134–138. [Google Scholar] [CrossRef]
- Aneiros, G.; Cao, R.; Fraiman, R.; Genest, C.; Vieu, P. Recent advances in functional data analysis and high-dimensional statistics. J. Multivariate Anal. 2019, 170, 3–9. [Google Scholar] [CrossRef]
- Ling, N.; Vieu, P. Nonparametric modelling for functional data: Selected survey and tracks for future. Statistics 2018, 52, 934–949. [Google Scholar] [CrossRef]
- Goia, A.; Vieu, P. An introduction to recent advances in high/infinite dimensional statistics [Editorial]. J. Multivariate Anal. 2016, 146, 1–6. [Google Scholar] [CrossRef]
- Almanjahie, I.M.; Bouzebda, S.; Chikr Elmezouar, Z.; Laksaci, A. The functional kNN estimator of the conditional expectile: Uniform consistency in number of neighbors. Stat. Risk Model 2022, 38, 47–63. [Google Scholar] [CrossRef]
- Almanjahie, I.M.; Bouzebda, S.; Kaid, Z.; Laksaci, A. Nonparametric estimation of expectile regression in functional dependent data. J. Nonparametr. Stat. 2022, 34, 250–281. [Google Scholar] [CrossRef]
- Bouzebda, S.; Nezzal, A. Uniform consistency and uniform in number of neighbors consistency for nonparametric regression estimates and conditional U-statistics involving functional data. Jpn. J. Stat. Data Sci. 2022, 5, 431–533. [Google Scholar] [CrossRef]
- Bouzebda, S.; Nemouchi, B. Weak-convergence of empirical conditional processes and conditional U-processes involving functional mixing data. Stat. Inference Stoch. Process. 2022, 1–56. [Google Scholar] [CrossRef]
- Bouzebda, S.; Nemouchi, B. Uniform consistency and uniform in bandwidth consistency for nonparametric regression estimates and conditional U-statistics involving functional data. J. Nonparametr. Stat. 2020, 32, 452–509. [Google Scholar] [CrossRef]
- Almanjahie, I.M.; Kaid, Z.; Laksaci, A.; Rachdi, M. Estimating the Conditional Density in Scalar-On-Function Regression Structure: K-N-N Local Linear Approach. Mathematics 2022, 10, 902. [Google Scholar] [CrossRef]
- Meyer, Y. Wavelets and Operators; Cambridge University Press: Cambridge, UK, 1993; pp. 35–58. [Google Scholar] [CrossRef]
- Daubechies, I. Ten lectures on wavelets. In CBMS-NSF Regional Conference Series in Applied Mathematics; Society for Industrial and Applied Mathematics (SIAM): Philadelphia, PA, USA, 1992; Volume 61, pp. xx+357. [Google Scholar] [CrossRef]
- Mallat, S. A Wavelet Tour of Signal Processing, 3rd ed.; Elsevier/Academic Press: Amsterdam, The Netherlands, 2009; pp. xxii+805. [Google Scholar]
- Vidakovic, B. Statistical Modeling by Wavelets; Wiley Series in Probability and Statistics: Applied Probability and Statistics; John Wiley & Sons Inc.: New York, NY, USA, 1999; pp. xiv+382. [Google Scholar] [CrossRef]
- Härdle, W.; Kerkyacharian, G.; Picard, D.; Tsybakov, A. Wavelets, approximation, and statistical applications. In Lecture Notes in Statistics; Springer: New York, NY, USA, 1998; Volume 129, pp. xviii+265. [Google Scholar] [CrossRef]
- Rao, B.L.S.P. Nonparametric estimation of the derivatives of a density by the method of wavelets. Bull. Inform. Cybernet. 1996, 28, 91–100. [Google Scholar] [CrossRef]
- Chaubey, Y.P.; Doosti, H.; Prakasa Rao, B.L.S. Wavelet based estimation of the derivatives of a density for a negatively associated process. J. Stat. Theory Pract. 2008, 2, 453–463. [Google Scholar] [CrossRef]
- Rao, B.L.S.P. Nonparametric Estimation of Partial Derivatives of a Multivariate Probability Density by the Method of Wavelets. In Asymptotics in Statistics and Probability, Papers in Honor of George Gregory Roussas; De Gruyter: Berlin, Germany, 2018; pp. 321–330. [Google Scholar] [CrossRef]
- Prakasa Rao, B.L.S. Wavelet estimation for derivative of a density in the presence of additive noise. Braz. J. Probab. Stat. 2018, 32, 834–850. [Google Scholar] [CrossRef]
- Allaoui, S.; Bouzebda, S.; Chesneau, C.; Liu, J. Uniform almost sure convergence and asymptotic distribution of the wavelet-based estimators of partial derivatives of multivariate density function under weak dependence. J. Nonparametr. Stat. 2021, 33, 170–196. [Google Scholar] [CrossRef]
- Allaoui, S.; Bouzebda, S.; Liu, J. Multivariate wavelet estimators for weakly dependent processes: Strong consistency rate. Comm. Statist. Theory Methods 2022, 1–34. [Google Scholar] [CrossRef]
- Chesneau, C.; Kachour, M.; Maillot, B. Nonparametric estimation for functional data by wavelet thresholding. REVSTAT 2013, 11, 211–230. [Google Scholar]
- Laib, N.; Louani, D. Nonparametric kernel regression estimation for functional stationary ergodic data: Asymptotic properties. J. Multivar. Anal. 2010, 101, 2266–2281. [Google Scholar] [CrossRef]
- Bouzebda, S.; Didi, S.; El Hajj, L. Multivariate wavelet density and regression estimators for stationary and ergodic continuous time processes: Asymptotic results. Math. Methods Statist. 2015, 24, 163–199. [Google Scholar] [CrossRef]
- Bouzebda, S.; Didi, S. Additive regression model for stationary and ergodic continuous time processes. Comm. Statist. Theory Methods 2017, 46, 2454–2493. [Google Scholar] [CrossRef]
- Bouzebda, S.; Didi, S. Multivariate wavelet density and regression estimators for stationary and ergodic discrete time processes: Asymptotic results. Comm. Statist. Theory Methods 2017, 46, 1367–1406. [Google Scholar] [CrossRef]
- Bouzebda, S.; Didi, S. Some asymptotic properties of kernel regression estimators of the mode for stationary and ergodic continuous time processes. Rev. Mat. Complut. 2021, 34, 811–852. [Google Scholar] [CrossRef]
- Bouzebda, S.; Didi, S. Some results about kernel estimators for function derivatives based on stationary and ergodic continuous time processes with applications. Comm. Statist. Theory Methods 2022, 51, 3886–3933. [Google Scholar] [CrossRef]
- Bouzebda, S.; Chaouch, M. Uniform limit theorems for a class of conditional Z-estimators when covariates are functions. J. Multivar. Anal. 2022, 189, 104872. [Google Scholar] [CrossRef]
- Bouzebda, S.; Chaouch, M.; Didi, S. Asymptotics for function derivatives estimators based on stationary and ergodic discrete time processes. Ann. Inst. Statist. Math. 2022, 74, 1–35. [Google Scholar] [CrossRef]
- Leucht, A.; Neumann, M.H. Degenerate U- and V-statistics under ergodicity: Asymptotics, bootstrap and applications in statistics. Ann. Inst. Statist. Math. 2013, 65, 349–386. [Google Scholar] [CrossRef]
- Aït-Sahalia, Y. Maximum likelihood estimation of discretely sampled diffusions: A closed-form approximation approach. Econometrica 2002, 70, 223–262. [Google Scholar] [CrossRef]
- Jiang, G.J.; Knight, J.L. A nonparametric approach to the estimation of diffusion processes, with an application to a short-term interest rate model. Econom. Theory 1997, 13, 615–645. [Google Scholar] [CrossRef]
- Kutoyants, Y.A. Efficient density estimation for ergodic diffusion processes. Stat. Inference Stoch. Process. 1998, 1, 131–155. [Google Scholar] [CrossRef]
- Banon, G. Nonparametric identification for diffusion processes. SIAM J. Control Optim. 1978, 16, 380–395. [Google Scholar] [CrossRef]
- Castellana, J.V.; Leadbetter, M.R. On smoothed probability density estimation for stationary processes. Stoch. Process. Appl. 1986, 21, 179–193. [Google Scholar] [CrossRef]
- Didi, S.; Alharby, A.; Bouzebda, S. Wavelet Density and Regression Estimators for Functional Stationary and Ergodic Data: Discrete Time. Mathematics 2022, 10, 3433. [Google Scholar] [CrossRef]
- Goh, S.S. Wavelet bases for Hilbert spaces of functions. Complex Var. Elliptic Equ. 2007, 52, 245–260. [Google Scholar] [CrossRef]
- Kerkyacharian, G.; Picard, D. Density estimation in Besov spaces. Statist. Probab. Lett. 1992, 13, 15–24. [Google Scholar] [CrossRef]
- Kerkyacharian, G.; Picard, D. Thresholding algorithms, maxisets and well-concentrated bases. Test 2000, 9, 283–344. [Google Scholar] [CrossRef]
- Prakasa Rao, B.L.S. Nonparametric density estimation for functional data via wavelets. Commun. Stat. Theory Methods 2010, 39, 1608–1618. [Google Scholar] [CrossRef]
- Cohen, A.; DeVore, R.; Kerkyacharian, G.; Picard, D. Maximal spaces with given rate of convergence for thresholding algorithms. Appl. Comput. Harmon. Anal. 2001, 11, 167–191. [Google Scholar] [CrossRef]
- DeVore, R.A. Nonlinear approximation. In Acta Numerica, 1998; Cambridge University Press: Cambridge, UK, 1998; Volume 7, pp. 51–150. [Google Scholar] [CrossRef]
- Autin, F. Point de vue Maxiset en Estimation non Paramétrique. Ph.D. Thesis, Université Paris, Paris, France, 2004. [Google Scholar]
- Gasser, T.; Hall, P.; Presnell, B. Nonparametric estimation of the mode of a distribution of random curves. J. R. Stat. Soc. Ser. B Stat. Methodol. 1998, 60, 681–691. [Google Scholar] [CrossRef]
- Donoho, D.L.; Johnstone, I.M.; Kerkyacharian, G.; Picard, D. Density estimation by wavelet thresholding. Ann. Statist. 1996, 24, 508–539. [Google Scholar] [CrossRef]
- Ferraty, F.; Vieu, P. Nonparametric models for functional data, with application in regression, time-series prediction and curve discrimination. J. Nonparametr. Stat. 2004, 16, 111–125, The International Conference on Recent Trends and Directions in Nonparametric Statistics. [Google Scholar] [CrossRef]
- Peškir, G. The uniform mean-square ergodic theorem for wide sense stationary processes. Stoch. Anal. Appl. 1998, 16, 697–720. [Google Scholar] [CrossRef]
- Delecroix, M. Sur l’Estimation et la Prévision non Paramétrique des Processus Ergodiques Cycles économiques et Taches Solaires. Ph.D. Thesis, Université Lille, Lille, France, 1987. [Google Scholar]
- Didi, S.; Louani, D. Asymptotic results for the regression function estimate on continuous time stationary and ergodic data. Stat. Risk Model. 2014, 31, 129–150. [Google Scholar] [CrossRef]
- Rosenblatt, M. Uniform ergodicity and strong mixing. Z. Wahrscheinlichkeitstheorie und Verw. Gebiete 1972, 24, 79–84. [Google Scholar] [CrossRef]
- Bradley, R.C. Introduction to Strong Mixing Conditions; Kendrick Press: Heber City, UT, USA, 2007; Volume 3, pp. xii+597. [Google Scholar]
- Neumann, M.H. Absolute regularity and ergodicity of Poisson count processes. Bernoulli 2011, 17, 1268–1284. [Google Scholar] [CrossRef]
- Didi, S. Quelques propriétés Asymptotiques en Estimation non Paramétrique de Fonctionnelles de Processus Stationnaires en Temps Continu. Ph.D. Thesis, Université Pierre et Marie Curie, Paris, France, 2014. [Google Scholar]
- Beran, J. Statistics for Long-Memory Processes. In Monographs on Statistics and Applied Probability; Chapman and Hall: New York, NY, USA, 1994; Volume 61, pp. x+315. [Google Scholar]
- Lu, Z. Analysis of Stationary and Non-Stationary Long Memory Processes: Estimation, Applications and Forecast. 2009. Available online: https://tel.archives-ouvertes.fr/tel-00422376/document (accessed on 10 October 2022).
- Maslowski, B.; Pospíšil, J. Ergodicity and parameter estimates for infinite-dimensional fractional Ornstein-Uhlenbeck process. Appl. Math. Optim. 2008, 57, 401–429. [Google Scholar] [CrossRef]
- Masry, E. Probability density estimation from sampled data. IEEE Trans. Inform. Theory 1983, 29, 696–709. [Google Scholar] [CrossRef]
- Prakasa Rao, B.L.S. Nonparametric density estimation for stochastic processes from sampled data. Publ. Inst. Statist. Univ. Paris 1990, 35, 51–83. [Google Scholar]
- Prakasa Rao, B.L.S. Statistical inference for diffusion type processes. In Kendall’s Library of Statistics; Edward Arnold: London, UK; Oxford University Press: New York, NY, USA, 1999; Volume 8, pp. xvi+349. [Google Scholar]
- Bosq, D. Nonparametric Statistics for Stochastic Processes, 2nd ed.; Springer: New York, NY, USA, 1998; Volume 110, pp. xvi+210. [Google Scholar] [CrossRef]
- Blanke, D.; Pumo, B. Optimal sampling for density estimation in continuous time. J. Time Ser. Anal. 2003, 24, 1–23. [Google Scholar] [CrossRef]
- Comte, F.; Merlevède, F. Adaptive estimation of the stationary density of discrete and continuous time mixing processes. ESAIM Probab. Statist. 2002, 6, 211–238. [Google Scholar] [CrossRef]
- Newey, W.K.; Powell, J.L. Asymmetric least squares estimation and testing. Econometrica 1987, 55, 819–847. [Google Scholar] [CrossRef]
- Mohammedi, M.; Bouzebda, S.; Laksaci, A. On the nonparametric estimation of the functional expectile regression. C. R. Math. Acad. Sci. Paris 2020, 358, 267–272. [Google Scholar] [CrossRef]
- Mohammedi, M.; Bouzebda, S.; Laksaci, A. The consistency and asymptotic normality of the kernel type expectile regression estimator for functional data. J. Multivar. Anal. 2021, 181, 104673. [Google Scholar] [CrossRef]
- Litimein, O.; Laksaci, A.; Mechab, B.; Bouzebda, S. Local linear estimate of the functional expectile regression. Statist. Probab. Lett. 2023, 192, 109682. [Google Scholar] [CrossRef]
- Huber, P.J. Robust estimation of a location parameter. Ann. Math. Statist. 1964, 35, 73–101. [Google Scholar] [CrossRef]
- Shannon, C.E. A mathematical theory of communication. Bell System Tech. J. 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef]
- Cover, T.M.; Thomas, J.A. Elements of Information Theory; Wiley Series in Telecommunications; John Wiley & Sons Inc.: New York, NY, USA, 1991; pp. xxiv+542. [Google Scholar] [CrossRef]
- Rényi, A. On the dimension and entropy of probability distributions. Acta Math. Acad. Sci. Hungar. 1959, 10, 193–215. [Google Scholar] [CrossRef]
- Kullback, S. Information Theory and Statistics; John Wiley & Sons, Inc: New York, NY, USA; Chapman & Hall, Ltd.: London, UK, 1959; pp. xvii+395. [Google Scholar]
- Gokhale, D.V.; Kullback, S. The information in contingency tables. In Statistics: Textbooks and Monographs; Marcel Dekker, Inc.: New York, NY, USA, 1978; Volume 23, pp. x+365. [Google Scholar]
- Csiszár, I. Informationstheoretische Konvergenzbegriffe im Raum der Wahrscheinlichkeitsverteilungen. Magy. Tud. Akad. Mat. Kutató Int. Közl. 1962, 7, 137–158. [Google Scholar]
- Vasicek, O. A test for normality based on sample entropy. J. Roy. Statist. Soc. Ser. B 1976, 38, 54–59. [Google Scholar] [CrossRef]
- Dudewicz, E.J.; van der Meulen, E.C. Entropy-based tests of uniformity. J. Amer. Statist. Assoc. 1981, 76, 967–974. [Google Scholar] [CrossRef]
- Ebrahimi, N.; Habibullah, M.; Soofi, E. Testing exponentiality based on Kullback-Leibler information. J. Roy. Statist. Soc. Ser. B 1992, 54, 739–748. [Google Scholar] [CrossRef]
- Jaynes, E.T. Information theory and statistical mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar] [CrossRef]
- Verdugo Lazo, A.C.G.; Rathie, P.N. On the entropy of continuous probability distributions. IEEE Trans. Inform. Theory 1978, 24, 120–122. [Google Scholar] [CrossRef]
- Bouzebda, S.; Elhattab, I. A strong consistency of a nonparametric estimate of entropy under random censorship. C. R. Math. Acad. Sci. Paris 2009, 347, 821–826. [Google Scholar] [CrossRef]
- Bouzebda, S.; Elhattab, I. Uniform in bandwidth consistency of the kernel-type estimator of the Shannon’s entropy. C. R. Math. Acad. Sci. Paris 2010, 348, 317–321. [Google Scholar] [CrossRef]
- Bouzebda, S.; Elhattab, I.; Nemouchi, B. On the uniform-in-bandwidth consistency of the general conditional U-statistics based on the copula representation. J. Nonparametr. Stat. 2021, 33, 321–358. [Google Scholar] [CrossRef]
- Bouzebda, S.; El-hadjali, T. Uniform convergence rate of the kernel regression estimator adaptive to intrinsic dimension in presence of censored data. J. Nonparametr. Stat. 2020, 32, 864–914. [Google Scholar] [CrossRef]
- Lad, F.; Sanfilippo, G.; Agrò, G. Extropy: Complementary dual of entropy. Statist. Sci. 2015, 30, 40–58. [Google Scholar] [CrossRef]
- Ferraty, F.; Vieu, P. The functional nonparametric model and application to spectrometric data. Comput. Statist. 2002, 17, 545–564. [Google Scholar] [CrossRef]
- Ouassou, I.; Rachdi, M. Regression operator estimation by delta-sequences method for functional data and its applications. AStA Adv. Stat. Anal. 2012, 96, 451–465. [Google Scholar] [CrossRef]
- Hall, P.; Penev, S. Cross-validation for choosing resolution level for nonlinear wavelet curve estimators. Bernoulli 2001, 7, 317–341. [Google Scholar] [CrossRef]
- Burkholder, D.L. Distribution function inequalities for martingales. Ann. Probab. 1973, 1, 19–42. [Google Scholar] [CrossRef]
- de la Peña, V.H.; Giné, E. Decoupling; Probability and its Applications (New York); Springer: New York, NY, USA, 1999; pp. xvi+392. [Google Scholar] [CrossRef]
- Masry, E. Wavelet-based estimation of multivariate regression functions in Besov spaces. J. Nonparametr. Statist. 2000, 12, 283–308. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).