Next Article in Journal
A Novel Hyper-Heuristic Algorithm for Bayesian Network Structure Learning Based on Feature Selection
Previous Article in Journal
On Subprojectivity of Goldie Torsion Modules
Previous Article in Special Issue
Modified Bimodal Exponential Distribution with Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modal Regression Estimation by Local Linear Approach in High-Dimensional Data Case

1
Department of Mathematical Sciences, College of Science, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
Department of Mathematics, College of Science, King Khalid University, Abha 62223, Saudi Arabia
*
Author to whom correspondence should be addressed.
Axioms 2025, 14(7), 537; https://doi.org/10.3390/axioms14070537
Submission received: 24 May 2025 / Revised: 30 June 2025 / Accepted: 11 July 2025 / Published: 16 July 2025
(This article belongs to the Special Issue Advances in Statistical Simulation and Computing)

Abstract

This paper introduces a new nonparametric estimator for detecting the conditional mode in the functional input variable setting. The estimator integrates a local linear approach with an L 1 -robust algorithm and treats the modal regression as the minimizer of the quantile derivative. As an asymptotic result, we derive the theoretical properties of the estimator by analyzing its convergence rate under the almost complete consistency framework. The result is stated under standard conditions, characterizing both the functional structure of the data and the local linear approximation properties of the model. Moreover, the expression of the convergence rate retains the usual form of the stochastic convergence rate in functional statistics. Simulations and real-data applications demonstrate the algorithm’s effectiveness, showing its advantage over existing methods in high-dimensional prediction tasks.

1. Introduction

In the last decade, interest in the statistical analysis of functional data has grown significantly, driven by its wide range of applications in various domains. The foundational work in this field was first established by [1,2], as pioneer works, and [3] or [4] for recent results and references. Within this field, nonparametric statistics have proven to be highly effective in modeling functional data, particularly when it comes to predicting using conditional mode and conditional quantiles. However, most of the existing nonparametric functional quantiles or mode estimators heavily depend on traditional kernel methods. In this paper, we improve the accuracy and robustness of the previous estimators using the L 1 -local linear estimation (LLE) method.
Modal and quantile regression are popular tools in nonparametric statistics for making predictions. They present an alternative approach to traditional regression, which is based on conditional expectation. Unlike the conventional approaches, the proposed predictors often impart deeper insights, especially when the data have asymmetric or multimodal conditional densities, or when the white noise shows a heavy-tailed distribution (refer to [5] for the limitation of the conditional expectation). The latter demonstrated that the conditional mode outperforms the conditional mean. Based on this, Ref. [6] created a prediction model that uses the derivative of the conditional density for multivariate data. Theoretical progress was made by [7], who established the asymptotic distribution of modal regression estimators for independent data. Ref. [8] extended the last result to the dependent cases. Further advancements in the conditional mode prediction are comprehensively covered in [9,10]. While the cited works concern the kernel estimator of the conditional mode (CM) in finite-dimensional cases, the Nadaraya–Watson (NW) estimator was extended for functional data by [11]. They estimated this regression model by maximizing the conditional density and proved its almost complete consistency. This result has been generalized to the functional time series case by [12]. The asymptotic distribution of the NW estimator of the conditional mode was explored for the independent case in [13] and for the strongly mixing case by [14]. There were also significant theoretical advancements made by [15], who determined L p -convergence rates for NW-based functional CM estimators. For an alternative functional time series, Ref. [16] established stochastic consistency for ergodic functional time series data containing missing at random. Further theoretical and practical contributions in the field of functional CM estimation can be found in [17] and the references therein.
On the other hand, quantile regression has gained increasing interest in nonparametric data analysis. It is worth noting that the quantile regression function is also one of the most well-studied models in nonparametric vector statistics. Earlier foundational works in multivariate statistics go back to [18], and were popularized by [19,20]. Once again in functional statistics, quantile regression has received considerable attention. This high level of flexibility justifies the approach, as it can be treated as a linear, semiparametric, or nonparametric model. The functional linear modeling of the conditional quantile was considered by [21,22], which provides comprehensive overviews of the latest developments. While the semiparametric estimation was developed by [23]. Concerning nonparametric estimation, quantile regression is estimated using the functional version of the NW estimator by [11]. They proved the Borel–Cantelli of the constructed estimator. Ref. [24] has stated the moment integrability of the functional NW estimator. We return to [25,26] for recent results.
The main novelty of this work is the estimation by the LLE method. This approach provides an important advantage compared to the NW estimator. In particular, it permits increasing the efficiency of the estimation by reducing bias. This statement has been observed by [27] in the multivariate case and expanded into functional statistics by [28]. From a bibliographical perspective, the LLE technique has been thoroughly explored in functional statistics by many authors. For example, Ref. [29] demonstrated L 2 consistency of the LLE estimator when estimating the regression operator in Hilbert spaces. Ref. [30] proposes a more general estimator allowing coverage of a wider range of functional covariates, while [31] developed an alternative estimator based on the inverse local covariance operator. Ref. [32] introduced functional LLE-fitting for nonparametric conditional models, establishing pointwise and uniform almost-complete (a.co.) consistency for the conditional density mode. The first results on the LLE-estimation of the quantile function were stated by [33]. Their estimation is obtained by inverting the cumulative distribution function (CDF) estimator introduced by [32]. An alternative was recently considered by [34] based on the M-estimation of the quantile regression.
While most current methods focus on estimating the conditional mode by optimizing a kernel-based conditional density estimator, this paper introduces a new LLE approach based on quantile regression. Unlike traditional methods, our estimator accumulates the robustness of L 1 -regression and the performance of the LLE-estimation, which enhances its resistance to outliers and extends its accuracy as a predictor or classifier in functional statistics. Theoretically, we demonstrate the almost complete convergence of the estimator under standard conditions, with precision in the convergence rate. This new estimator’s benefits are also emphasized by comparing it to the standard one using a simulated and real data example. Our empirical analysis shows that the new estimator not only outperforms the accuracy of the previous estimators but also remains robust across a wide range of scenarios. Additionally, to demonstrate its practical value, we apply our method to predict the chemical composition of cocoa using spectrometry data. Typically, we have focused on the prediction of fat and moisture content using near-infrared spectroscopy data as functional regressors. It is well documented that near-infrared spectroscopy is a powerful, non-destructive tool for rapidly quantifying cocoa quality, allowing real-time quality control in production. Thus, combining this spectrometry processing with the proposed functional model can help chocolate producers to maintain good flavor and texture in their products, ensuring quality at every step.
The paper is organized as follows. In the next Section, we introduce the conditional mode based on the L 1 rule and explain its LLE estimator. Section 3 covers the principal assumptions and the main theoretical findings. In Section 4 We outline the principal features of our approach and provide a comparative analysis with existing methods. The finite-sample performance of our proposed estimator is examined in Section 5, using both simulation studies and real-world data applications. The detailed proofs of the intermediate results are presented in the Appendix A.

2. The L 1 Conditional Mode and Its Local Linear Estimator

Let { ( F i , R i ) } i = 1 n be an independent and identically distributed (i.i.d.) sample from the random pair ( F , R ) , where F is functional input variable taking values in a semi-metric space ( H , m ) (m being the semi-metric) and R is the output real variable. Next, fix F H and consider a neighborhood N F of F . We assume the existence of the regular conditional distribution function D ( · | F ) of R given F = F , which is strictly increasing and has a continuous density d ( · | F ) with respect to Lebesgue measure on R . The standard conditional mode is usually defined as the maximizer of this conditional density over a compact set S:
M R ( F ) = arg max y S d ( y | F ) .
However, in order to introduce a more robust definition of the conditional mode, we reformulate this definition and we relate the modal regression to the quantile derivatives. Indeed, for q D 1 ( S | F ) = [ a F , b F ] ( 0 , 1 ) , let Q R q ( F ) denote the conditional quantile of order q given F = F . Following similar ideas to [35] in the linear case, we assume Q R · ( F ) is continuously differentiable ( C 1 ) with D ( Q R q ( F ) | F ) > 0 , yielding:
Q R q ( F ) = Q R q ( F ) q = 1 D ( Q R q ( F ) | F ) .
It follows that the conditional mode can be expressed as:
M R ( F ) = Q R q M R ( F ) , where q M R = arg min q [ a F , b F ] Q R q ( F ) .
Recall that the theoretical conditional quantile Q R q ( F ) is robustly defined as M-regression with respect to the following score function
min t E L q ( R t ) F = F ,
where L q ( y ) = y ( q 1 I { y < 0 } ) is the quantile loss function. Thus, the LLE estimator of the conditional mode is strongly linked to the LLE estimator of the quantile regression. For this last point, we adopt the fast LLE approach of [30], approximating Q R q ( · ) in a neighborhood N F by:
Q R q ( Z ) A + B m ( F , Z ) Z N F ,
and estimate the coefficients ( A , B ) as solutions of
min ( A , B ) R 2 i = 1 n L q ( R i A B m ( F i , F ) ) F m ( F , F i ) f n ,
where F is a kernel function, f n is a bandwidth sequence. The conditional quantile estimator is then Q R ^ q ( F ) = a ^ . Consequently, the LLE estimator of the conditional mode is
M R ^ ( F ) = Q R ^ ( q ^ M R | F ) where q ^ M R = arg min q [ a F , b F ] Q R ^ ( q | F ) ,
with Q R ^ being the estimator of the derivative of the conditional quantile defined by
Q R ^ ( q | F ) = Q R ^ ( q + b n | F ) Q R ^ ( q b n | F ) 2 b n ,
where b n is a positive bandwidth-like sequence.
We point out that the functional NW estimator studied by [17] is a particular case of this study; it corresponds to B = 0 . Next, let us clarify that the existence and uniqueness of M R ( F ) are guaranteed by the continuity and strict monotonicity of D ( · | F ) , while M R ^ ( F ) may not be unique. In what follows, M R ^ ( F ) refers to any minimizer in (6). Furthermore, the primary contribution of this work is to establish asymptotic properties of M R ^ ( F ) when F takes values in the semi-metric space F N . To our knowledge, this is the first study to employ this estimation strategy for the conditional mode, even in multivariate settings. Note that the finite-dimensional case ( F N = R p ) emerges as a special case, highlighting the generality of our approach.

3. Main Results

All along the paper, when no confusion is possible, we will denote by C or C some strictly positive generic constants, and we set B ( F , f ) = F F : m ( F , F ) < f . Now, we list some required conditions that are necessary in deriving the almost complete convergence of M R ^ ( F ) of M R ( F ) .
(B1)
For any r > 0 , α F ( f ) : = α F ( f , f ) > 0 and there exists a function β F ( · ) such that:
for all t ( 1 , 1 ) , lim f 0 α F ( t f n , f n ) α F ( f n ) = β F ( t )
(B2)
The functions Q R ( · | F ) is of class C 3 ( [ a F , b F ] ) and D ( · | F ) satisfies the following Lipschitz condition:
for all ( F 1 , F 2 ) N F , | D ( t | F 1 ) D ( t | F 2 ) | C d b ( F 1 , F 2 ) for some b > 0 ,
where N F denotes a neighborhood of F .
(B3)
The kernel F is a positive and differentiable function which is supported within ( 1 , 1 ) , and such that:
F ( 1 ) 1 1 F ( t ) β F ( t ) d t F ( 1 ) 1 1 ( t F ( t ) ) β F ( t ) d t F ( 1 ) 1 1 ( t F ( t ) ) β F ( t ) d t F ( 1 ) 1 1 ( t 2 F ( t ) ) β F ( t ) d t
is a positive definite matrix.
(B4)
The bandwidth f n satisfies:
lim n log n n b n α F ( f n ) = 0
The assumed conditions are not very restrictive. They allow us to analyze various aspects of our subject, such as model structure, data framework, and convergence rates. Considering the complexity of our proposed local linear algorithm and the strength of the Borel–Cantelli (BC) consistency, these assumptions offer a good compromise between generality and analytical feasibility. Specifically, each element of the subject is explored through a distinct assumption. For instance, (B1) is particularly essential in NFDA. Specifically, Condition (B1) is indispensable in NFDA, and the function α F ( . ) can be explicitly given for several continuous processes (see [11]). While the regularity conditions are necessary to explore the functional space of the nonparametric path under study, this assumption significantly impacts the bias term in the convergence rate of M R ^ ( F ) . Finally, conditions (B3)–(B4) relate to the kernel F and the bandwidths f n and b n , which cover the technical aspects of the estimator M R ^ ( F ) . These conditions are less restrictive than the usual technical requirements in conditional mode estimation. Unlike the conditional density approach, which relies on two kernels, our method requires only one. It is important to recall that these conditions are necessary to determine the convergence rate of the kernel estimator under the Borel–Cantelli mode. Typically, a weaker form of consistency—distinct from BC consistency—can be established for the estimator under more restrictive assumptions. The latter is obtained by evaluating the expectation and variance of the estimator. Although this probability consistency is weaker than Borel–Cantelli consistency, it remains sufficient to justify the estimator’s practical applicability. Nevertheless, because this paper aims to demonstrate both the theoretical generality of the model and its practical viability, we have chosen to emphasize the stronger consistency result, which implies the weaker.
The following theorem establishes the almost-complete convergence (cf. [2] for details), (a.co.), of M R ^ ( F ) . This kind of convergence implies both almost sure convergence and convergence in probability.
Theorem 1.
Under assumptions (B1)–(B4) and if inf q ( 0 , 1 ) 3 Q R q ( F ) q > 0 we have:
M R ^ ( F ) M R ( F ) = O ( f n b / 2 ) + O ( b n 1 / 2 ) + O log n n b n 2 α F ( f n ) 1 4 a . c o .
Proof of Theorem 1.
The proof is based on some standard analytical arguments. Indeed, we write
M R ^ ( F ) M R ( F ) = Q R ^ q M R ^ ( F ) Q R q M R ^ ( F ) + Q R q M R ^ ( F ) Q R q M R ( F )
Trivially,
M R ^ ( F ) M R ( F ) sup q [ a F , b F ] | Q R ^ ( q | F ) t ( q | F ) | + Q R q M R ^ ( F ) Q R q M R ( F )
By Taylor expansion,
Q R q M R ^ ( F ) Q R q M R ( F ) = ( q M R ^ q M R ) Q R q M R * ( F ) , q M R * being   in ( q M R ^ , q M R ) .
Since q M R is minimizer of Q R · ( F ) ,
Q R q M R ^ ( F ) Q R q M R ( F ) = ( q M R ^ q M R ) 2 Q R q M R * * ( F ) , where q M R * * ( q M R ^ , q M R ) .
In addition,
Q R q M R ^ ( F ) Q R q M R ( F ) = Q R q M R ^ ( F ) Q R ^ q M R ^ ( F ) + Q R ^ q M R ^ ( F ) Q R q M R ( F ) | Q R q M R ^ ( F ) Q R ^ q M R ^ ( F ) | + | min q Q R ^ q ( F ) min q Q R q ( F ) | 2 sup q [ a F , b F ] | Q R ^ q ( F ) Q R q ( F ) |
Thus,
| M R ^ ( F ) M R ( F ) | C sup q [ a F , b F ] | Q R ^ q ( F ) Q R q ( F ) | + sup q [ a F , b F ] | Q R ^ q ( F ) Q R q ( F ) | .
Now, for a large enough n
Q R ^ q ( F ) Q R q ( F ) = Q R ^ q + b n ( F ) Q R q + b n ( F ) + Q R q b n ( F ) Q R ^ q b n ( F ) 2 b n + Q R q + b n ( F ) Q R q ( F ) + Q R q ( F ) Q R q b n ( F ) 2 b n Q R q ( F ) 2 b n C b n 1 sup q ( a F b n , b F + b n ) | Q R ^ q ( F ) Q R q ( F ) | + O ( b n )
Thus, it suffices to evaluate, for a large enough n
sup q [ 0 , 1 ] | Q R ^ q ( F ) Q R q ( F ) | .
Thus, Theorem 1 is a consequence of the following proposition. □
Proposition 1.
Under assumptions of Theorem 1, we have
sup q ( 0 , 1 ) | Q R ^ q ( F ) Q R q ( F ) | = O ( f n b ) + O a . c o . log n n α F ( f n ) 1 2
The proof of this proposition is based on the Bahadur representation of the conditional quantile. The latter is a consequence of the following lemmas.
Lemma 1.
Let Π n be a sequence of decreasing real random functions and Υ n be a random real sequence such that
Υ n = o a . c o . ( 1 ) a n d sup | ν | M | Π n ( ν ) + λ ν Υ n | = o a . c o . ( 1 ) f o r   c e r t a i n   c o n s t a n t s λ , M > 0 .
Then, for any real sequence ν n such that Π n ( ν n ) = o a . c o . ( 1 ) , we have:
n = 1 I P | ν n | M < .

4. On the Potential Impact of the Contribution

  • Comparison with existing approaches
    In an earlier contribution [36], we have investigated robust mode estimation under a functional single-index structure, employing the local constant method. Alternatively, in this study, we examine local linear estimation for the same model under a general functional framework. Firstly, observe that the topic of the present contribution can be viewed as a generalization of [36], in the sense that the local constant constitutes a particular case of the local linear approach, and the functional single index structure is a special case of the general functional structure. Moreover, it is well documented (see, for instance, [27]) that the local linear method has many advantages over the Nadaraya–Watson (the local constant method). Particularly, the local linear estimation is usually used to reduce the bias and the boundary effect of the Nadaraya–Watson method. So, the use of the local linear method instead of the standard kernel method substantially improves the prediction accuracy. On the other hand, it is widely recognized that the single-index model reduces analytical complexity by projecting functional covariates onto a univariate index. Thus, it transforms the functional data analysis problem into a one-dimensional real data issue. This oversimplification is unusual in practice, as it ignores potentially influential higher-dimensional interactions. Our generalized framework of this contribution circumvents this limitation. From a practical point of view, the kernel estimator assumes that the nonparametric model is flat within the neighborhood of the location point, which leads to suboptimal prediction. In contrast, the local linear method assumes that the model has a linear approximation in the neighborhood of the location point, which is more realistic and improves the prediction results.
  • On the bias reduction
    As mentioned in the previous comment, the behavior of the bias term is one of the main reasons for adopting the local linear method. Although the asymptotic behavior of the bias term is usually linked to the smoothness assumptions of the nonparametric model, this term can be significantly improved in the local linear approach. This beneficial characteristic is related to the weighting functions in the local linear estimator (see [27] in the non-functional case). A similar statement can be deduced in the functional case. Specifically, under standard conditions, the local linear approach offers a better bias term compared to the Nadaraya–Watson [36]. Indeed, this improvement is also due to the specific weighting functions implemented in Q R ^ q ( F ) . It is clear that the leading term of the Bahadur representation of Q R ^ q ( F ) such that
    I E D i ( A 3 F i A 2 f n 1 D i F i ) = 0
    where
    F i = F f n 1 m ( F , F i ) , and D i = m ( F i , F )
    and
    A j = α F 1 ( f n ) f n 1 j I E [ D i j 1 F i ] ; j = 2 , 3 .
    Using the same analytical arguments as in [37], we prove that the first part of the bias term can be reduced to o ( f n b / 2 ) . In conclusion, although the local linear and the Nadaraya–Watson (NW) estimators share similar asymptotic properties, the local linearity of the model and the weighting functions of the LLE method allow us to improve the bias term in certain situations.
  • On the applicability of the estimator
    Of course, the simple use of the estimator greatly depends on choosing its parameters easily. Since the estimator is derived from quantile regression, there are multiple cross-validation methods available for selecting the parameters, particularly the bandwidth parameters associated with the functional component. First, for a given subset H n of real positive numbers, we consider the cross-validation criterion used by [38].
    arg min f n H n 1 n i = 1 n Q p L ( F i ) Q ^ p i ( F i ) 2 ,
    where Q p L ( X i ) denotes a local empirical quantile and Q ^ p i is the conditional quantile estimator at F i after excluding this observation from the sample. Secondly, we apply the cross-validation approach that was utilized by [39]
    arg min f n H n 1 n i = 1 n L p R i Q ^ p i ( F i ) .
    Additionally, various other methods can be utilized, such as the least squares cross-validation technique suggested by [6]:
    arg min f n H n 1 n i = 1 n R i M R ^ ( F i ) 2 .
    Of course, the diversity of selection methods makes the estimator easier to implement in practice.

5. Computational Part

5.1. Simulation Study

In this section, we demonstrate the practical effectiveness of our theoretical development through some simulation experiments. Our principal goal is to show the easy insertion of the estimator in practice and to emphasize clear benefits of the proposed estimator over its competitors, such as the kernel conditional mode and the conditional density-based LLE-mode. In order to cover many practical situations, we generate data from different scenarios. Indeed, we consider two different nonparametric functional models that are:
heteroscedastic ( Het . ) Model : R i = 4 0 1 exp ( 1 3 + F i 3 ( t ) ) d t + cos 3 + F i 3 ( t ) ϵ i , homoscedastic ( Hom . ) Model : R i = 5 0 1 log 2 + F i 2 ( t ) 3 + F i 3 ( t ) d t + ϵ i ,
where ϵ i and F i are independent. The functional input variable is also generated in two ways: smoothly and roughly. The smooth way is generated according to the following formula:
Smooth curves F i ( t ) = a i cos ( 4 ( b i t π ) ) + b i
while the rough way is drawn from the following equation:
rough curves F i ( t ) = a i cos ( 4 ( b i t π ) ) + b i + η i , t
For both cases, b i is distributed in N ( 0 , 0.7 ) ) and a i is generated according to a N ( 3 , 4 ) distribution, while the random variables η i , t ) are drawn from N ( 0 , 2 ) . All the input curves F i ’s are discretized on the same grid represented by 100 equispaced measurements taken in the interval ( 0 , 1 ) . The resulting functional variables are shown in Figure 1 and Figure 2.
Obviously, choosing various curve forms is essential for evaluating how the data’s functional structure affects the accuracy of estimates. Typically, the spline metric is used for modeling smooth curves, while the PCA metric is used for rough curves.
This strategy of data sampling is intentional. It allows us to test the robustness of different models and assess how the data functionality affects the estimation accuracy. We also evaluate the nonparametric model using three distributions: Normal, Weibull, and Laplace. These were chosen because they remain consistent under translation and exhibit different levels of heavy-tailed behavior, which helps us to measure how well the estimators handle outliers. In mathematical terms, the two competitors’ modal regression estimators are defined by:
The   NW   estimator : M R ˜ ( F ) = arg max y i = 1 n F f n 1 m ( F , F i ) F b n 1 ( y R i ) i = 1 n b n F f n 1 m ( F , F i ) ,
and   the   LLEestimator : M R ¯ ( F ) = arg max y i , j = 1 n W i j F b n 1 ( y R j ) i , j = 1 n b n W i j ,
where W i j ( x ) = m ( F i , F , ) m ( F i , F , ) m ( F j , F , ) F ( f n 1 m ( F , F i ) F ( f n 1 m ( F , F j ) .
Of course, the behavior of the three estimators, M R ^ , M R ˜ , and M R ¯ , as well as their straightforward insertion, is influenced by the choice of parameters, metric m, kernel F , and smoothing parameters ( f n , b n ) . Evidently, the selection of the metric m is intrinsically linked to the smoothness of the functional curves. Smoother curves are better analyzed by the B-spline metric, which accommodates their regularity, while rougher curves require the PCA-metric to handle their more irregular structure. We refer to [2] for more details on the definitions of these metrics. The choice of kernel is motivated by technical considerations. In order to incorporate the assumption (B3), we have simulated using the quadratic kernel in ( 1 , 1 ) . For the smoothing parameters ( f n , b n ) , we have used the cross-validation rule based on the mean square error defined by
M S E ( M R ¨ ) = 1 n i = 1 n R i M R ¨ ( F i ) 2 ,
where M R ¨ represents either M R ^ , M R ¯ , or M R ˜ . It follows that
( f o p t , b o p t ) = arg min f n H n , b n B n M S E ( M R ¨ )
H n (respectively, B n ) is the set of positive real numbers, f n , such that the ball centered at F with radius f n (respectively, the interval [ y b n , y + b n ] ) contains exactly k neighbors of (respectively, y). Typically, the number k is selected from the subset { 5 , 10 , 20 , , 0.5 n } . For our goal, we generate n = 200 observations of ( F i , R i ) , and we examine the robustness and the accuracy of different estimators using the MSE defined in (14). The obtained results are reported in the following table.
The performance of these estimators is strongly influenced by various components of the study, particularly the nature of the functional data and the nonparametric approach. Nevertheless, combining L 1 -techniques with the LLE-smoothing method proves highly advantageous in terms of robustness and accuracy. As evidenced by the MSE results in Table 1, the variability in M R ^ is lower than that of the other estimators. This lower variability demonstrates that the L 1 -LLE estimator for the conditional mode offers greater robustness and improved accuracy.

5.2. Real Data Application

Spectrometry data analysis is a powerful tool for generating high-frequency functional data. In this study, we focus on food quality as a critical real-world issue. Clearly, ensuring high-quality food is a top priority for both consumers and producers. Traditionally, quality control relies on chemical testing as an effective but destructive, time-consuming, and expensive approach. Instead, we provide a smart solution that combines spectrometry techniques with advanced statistical tools. It permits us to create an efficient, cost-effective, and non-invasive approach to ensure food quality. Specifically, we demonstrate how our new predictor can accurately predict fat and moisture contents in chocolate. Therefore, we evaluate our approach against leading alternatives, including the NW estimator, LLE-conditional expectation, and density-based LLE-mode. This computational study allows us to demonstrate the superiority and practical advantages of M R ^ in robustness as well as in accuracy. We compare these three approaches by testing their predictive performance using the near-infrared spectroscopy curve as functional input F, with the fat (or moisture) quantity as response variable R. The near-infrared data set available on the website (https://data.mendeley.com/datasets/7734j4fd98/1, (accessed on 8 May 2025)). This issue has been considered by many authors using different algorithms (see [40] for a list of references). Formally, in this real data example, we assume that the spectrum curve gives the absorbance of light across wavelengths ranging from 1000 to 2500 nm. These curves are visualized in Figure 3.
These curves represent the absorbance spectrum (NIR) in the wavelength range from 1000 to 2500 nm for a total of 72 bulk intact cocoa bean samples. Each bulk amounted to 50 g of intact beans. Recall that the choice of the response variable R is motivated by the fact that the fat or moisture quantities are determinants of cocoa quality. The quality of cocoa beans is the cornerstone of exceptional chocolate production, profoundly influencing the sensory, nutritional, and commercial value of the final product. Thus, precise determination of cocoa contents is essential for compliance with chocolate quality standards, economic valuation, and research in agriculture and nutrition. Consequently, optimizing fat and moisture percentage is primordial to delivering high-quality chocolate products that meet both consumer and industry demands. Now, to put our predictors into action, we need to carefully select the principal parameters that drive our estimator’s performance. This step is crucial to ensuring the potential feasibility of the four predictors M R ^ , M R ˜ , M R ¯ , and
C E ¯ = i , j = 1 n W i j R j i , j = 1 n b n W i j ,
Based on our earlier discussion, the kernel selection is a direct consequence of assumption (B3), ensuring theoretical consistency. Meanwhile, the metric m plays a crucial role in the quantification of the smoothness of the functional curves F i , allowing us to better analyze their behavior. Clearly, the curve’s shape in Figure 3 shows that the metric of L 2 between the first derivatives of the curves (the curves being regular) is more suitable for this data. Once again, the smoothing parameters ( f n , b n ) according the same strategy as in (14). This latest approach has proven highly effective in our predictive modeling, highlighting the importance of leave-one-out cross-validation data-driven algorithms in functional data analysis. Figure 4 and Figure 5 summarize our statistical predictions for fat and moisture for the 70 observations, comparing each approach’s performance by plotting predicted values against the true values.
We use (MSE) error to assess model performance for the predictiom results. It is defined by
M S E ( θ ¨ ) = 1 70 i = 1 70 R i M R ¨ ( F i ) 2 ,
where θ ¨ can represent either M R ^ , M R ¯ , M R ˜ or C E ¯ . Clearly, the predictor M R ^ outperforms the other predictors in precision. The latter predicts (fat, moisture) with MSE equal, respectively, to ( 0.97 , 1.06 ) . It is the best one compared to the results obtained by the predictors M R ¯ , M R ˜ , and C E ¯ . Typically, we record for both components (fat, moisture) ( 1.70 , 2 , 89 ) for M R ¯ , ( 4.89 , 7.48 ) for M R ˜ and ( 3.45 , 6.22 ) for C E ¯ .

Author Contributions

The authors contributed approximately equally to this work. Formal analysis, F.A.A.; Validation, M.B.A.; Writing—review & editing, A.L. and Z.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Princess Nourah bint Abdulrahman University Researchers; Supporting Project number (PNURSP2025R515); Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia; and the Deanship of Scientific Research and Graduate Studies at King Khalid University for funding this work through Small Research Project under grant number RGP1/41/46.

Data Availability Statement

The data used in this study are available through the link https://data.mendeley.com/datasets/7734j4fd98/1 (accessed on 8 May 2025).

Acknowledgments

The authors would like to thank the Editor, the Associate-Editor and two anonymous reviewers for their valuable comments and suggestions which improved substantially the quality of an earlier version of this paper. The authors thank and extend their appreciation to the funders of this work: This work was supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2025R515), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia, and the Deanship of Scientific Research and Graduate Studies at King Khalid University for funding this work through Small Research Project under grant number RGP1/41/46.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proofs of Intermediate Results

Proof of Proposition 1.
We apply Lemma 1 on
ν n = A ^ A f n ( B ^ B )
X n ( ν ) = 1 n α F ( F ˜ n ) i = 1 n ζ q ( ν ) 1 f n 1 D i F i , for ν = c d
and
Υ n = X n ( ν 0 ) with ν 0 = 0 0
where
ζ q ( ν ) = q 1 I R i ( c + A ) + ( f n 1 d + B ) D i , F i = F f n 1 m ( F , F i ) , and D i = m ( F i , F ) .
So, the main result is a consequence of the following lemmas. □
Lemma A1.
Under assumptions (B1)–(B5), we have
Υ n = O ( f n min ( k 1 , k 2 ) ) + O a . c o . log n n α F ( f n ) 1 / 2 .
Proof of Lemma A1.
Firstly, we set
Z i 1 = q 1 I [ R i A + B D i ] F i I E q 1 I [ R i A + B D i ] F i
and
Z i 2 = q 1 I [ R i A + B D i ] D i F i I E q 1 I [ R i A + B D i ] F i .
Then
Υ n I E [ Υ n ] = Υ n 1 Υ n 2
where
Υ n 1 = 1 n α F ( F ˜ n ) i = 1 n Z i 1 Υ n 2 = 1 n h α F ( F ˜ n ) i = 1 n Z i 2 .
Obviously
| Z i 1 | C and | Z i 2 | C h .
Moreover,
I E Z i 1 2 C α F ( F ˜ n ) and I E Z i 2 2 C f n 2 α F ( F ˜ n ) .
It follows that
Υ n 1 I E Υ n 1 = O a . c o . log n n α F ( F ˜ n )
and
Υ n 2 I E Υ n 2 = O a . c o . log n n α F ( F ˜ n ) .
Next,
I E Υ n 1 = 1 α F ( F ˜ n ) I E q 1 I [ Y 1 A + B D 1 ] F 1 1 α F ( f n ) I E D ( Q R q ( F ) | F ) D ( ( A + B D 1 ) F 1 | F ) = O ( f n min ( k 1 , k 2 ) ) .
Similarly,
I E Υ n 2 = 1 h α F ( F ˜ n ) I E q 1 I [ Y 1 A + B D 1 ] D 1 F 1 1 h α F ( f n ) I E D ( Q R q ( F ) | F ) ) D ( ( A + B D 1 ) D 1 F 1 | F ) = O ( f n min ( k 1 , k 2 ) ) .
Therefore,
Υ n = O ( f n min ( k 1 , k 2 ) ) + O a . c o . log n n α F ( f n ) 1 / 2
Lemma A2.
Under assumptions (B1)–(B5), we have
sup ν M X n ( ν ) + λ 0 D ν Υ n = o a . c o . ( 1 )
with
D = F ( 1 ) 1 1 K ( t ) β F ( t ) d t F ( 1 ) 1 1 ( t F ( t ) ) β F ( t ) d t F ( 1 ) 1 1 ( t F ( t ) ) β F ( t ) d t F ( 1 ) 1 1 ( t 2 F ( t ) ) β F ( t ) d t
and
λ 0 = d ( Q R q ( F ) | F )
Proof of Lemma A2.
We prove
sup ν M X n ( ν ) Υ n I E X n ( ν ) Υ n = O a . c o . log n n α F ( F ˜ n )
and
sup ν M I E X n ( ν ) Υ n + d ( Q R q ( F ) | F ) D ν = O ( f n min ( k 1 , k 2 ) ) .
Next, use the compactness of the ball B ( 0 , M ) in I R 2 and we write
B ( 0 , M ) j = 1 d n B ( ν j , l n ) , ν j = c j d j and l n = d n 1 = 1 / n .
Taking j ( ν ) = arg min j | ν ν j | and we use the fact that
sup ν M X n ( ν ) Υ n I E X n ( ν ) Υ n sup ν M X n ( ν ) X n ( ν j ) + sup ν M X n ( ν j ) Υ n I E X n ( ν j ) Υ n + sup ν M I E X n ( ν ) X n ( ν j ) .
Since | 1 I [ Y < a ] 1 I [ Y < b ] | 1 I | Y b | | a b | then
sup ν M X n ( ν ) X n ( ν j ) 1 n α F ( F ˜ n ) i Z i
where
Ω i = sup ν M 1 I | R i ( c j + A ) ( f n 1 d j + d ) D i | C l n 1 f n 1 D i F i .
Clearly
| Ω i | C , I E [ Ω i ] = O ( l n α F ( F ˜ n ) ) and I E [ Ω i 2 ] = O ( l n α F ( F ˜ n ) ) .
By using the fact that
l n = o log n n α F ( F ˜ n ) 1 / 2
we get
sup ν M X n ( ν ) X n ( ν j ) = O a . c o . log n n α F ( F ˜ n ) 1 / 2 .
For the last term,
sup ν M I E X n ( ν ) X n ( ν j ) 1 α F ( F ˜ n ) I E [ Ω 1 ] C l n
then, by (A3) we obtain
sup ν M I E X n ( ν ) X n ( ν j ) = o a . c o . log n n α F ( F ˜ n ) 1 / 2 .
Now
sup ν M X n ( ν j ) Υ n I E X n ( ν j ) Υ n .
For this
X n ( ν j ) Υ n I E X n ( ν j ) Υ n = Θ n 1 ( ν j ) Θ n 2 ( ν j )
where
Θ n 1 ( ν j ) = 1 n α F ( f n ) i = 1 n Λ i 1 Θ n 2 ( ν j ) = 1 n f n α F ( f n ) i = 1 n Λ i 2
with
Λ i 1 = ζ q ( ν j ) ζ q ( ν 0 ) F i I E ζ q ( ν j ) ζ q ( ν 0 ) F i
and
Λ i 2 = ζ q ( ν j ) ζ q ( ν 0 ) D i F i I E ζ q ( ν j ) ζ q ( ν 0 ) D i F i .
clearly
| Λ i 1 | C and | Λ i 2 | C f n .
Moreover,
I E Λ i 1 2 C α F ( f n ) and I E Λ i 2 2 C f n 2 α F ( f n ) .
Therefore, there exits η > 0 such that
n I P sup ν M X n ( ν j ) Υ n I E X n ( ν j ) Υ n η log n n α F ( f n ) n d n max j I P X n ( ν j ) Υ n I E X n ( ν j ) Υ n η log n n α F ( f n ) <
which complete the proof of (A1).
Concerning (A2) we write
X n ( ν ) Υ n = Θ n 1 ( ν ) Θ n 2 ( ν )
where
Θ n 1 ( ν ) = 1 n α F ( f n ) i = 1 n ζ q ( ν ) ζ q ( ν 0 ) F i
and
Θ n 2 ( ν ) = 1 n h α F ( f n ) i = 1 n ζ q ( ν ) ζ q ( ν 0 ) D i F i
On the other hand, we have
I E Θ n 1 ( ν ) = 1 α F ( f n ) I E 1 I Y 1 ( c + A ) + ( f n 1 d + B ) D 1 1 I Y 1 A + B D 1 F i = 1 α F ( f n ) I E D ( ( ( c + A ) + ( f n 1 d + B ) D 1 ) | F 1 ) D ( ( A + B D 1 ) | F 1 ) F 1 = 1 α F ( f n ) I E D ( ( c + A ) + ( f n 1 d + B ) D 1 | F ) D ( A + B D 1 ) | F F 1 + O ( f n k 1 ) = 1 α F ( f n ) I E d ( A + B D 1 ) | F 1 , f n 1 D 1 ν F 1 + O ( f n k 1 ) + o ν = d ( Q R q ( F ) | F ) 1 α F ( f n ) I E F 1 , f n 1 I E [ D 1 F 1 ] ν + O ( f n min ( k 1 , k 2 ) ) + o ν
Similarly,
I E Θ n 2 ( ν ) = d ( Q R q ( F ) | F ) 1 h α F ( f n ) I E D 1 F 1 , f n 1 I E [ D 1 2 F 1 ] ν + O ( f n min ( k 1 , k 2 ) ) + o ν .
Thus
I E X n ( ν ) Υ n = d ( Q R q ( F ) | F ) 1 α F ( f n ) I E F i I E F i f n 1 D i I E F i f n 1 D i I E f n 2 D i 2 F i ν + O ( f n min ( k 1 , k 2 ) ) + o ν .
Obviously
f n a I E [ D i a F i b ] = α F ( f n ) F b ( 1 ) 1 1 ( u a F b ( u ) ) D ( u ) d u + o ( α F ( f n ) ) .
Hence,
sup ν M I E X n ( ν ) Υ n + d ( Q R q ( F ) | F ) D ν + o ( ν ) = O ( f n min ( k 1 , k 2 ) )
implying the result (A2). □
Lemma A3.
Under assumptions (B1)–(B5), we have
sup | ν | M sup q [ 0 , 1 ] | Π n ( ν , q ) I E Π n ( ν , q ) | = O a . c o . log n n α F ( f n ) 1 / 2 .
where
Π n ( ν , q ) = 1 n α F ( F ˜ n ) i = 1 n ζ q ( ν ) 1 f n 1 D i F i , f o r ν = c d
Proof of Lemma A3.
The compactness of [ 0 , 1 ] implies
[ 0 , 1 ] k = 1 d n [ q k l n , q k + l n ] , f o r q k [ 0 , 1 ] .
Next, for all q [ 0 , 1 ] we put F q = arg min k | q q k | and we evaluate the term as function of ν and q. We have,
sup | ν | M sup q [ 0 , 1 ] | Π n ( ν , q ) I E Π n ( ν , q ) | sup | ν | M sup q [ 0 , 1 ] | Π n ( ν , q ) Π n ( ν j ( ν ) , q ) | + sup | ν | M sup q [ 0 , 1 ] | Π n ( ν j ( ν ) , q ) Π n ( ν j ( ν ) , q F q ) | + sup | ν | M sup q [ 0 , 1 ] | Π n ( ν j ( ν ) , q F q ) I E [ Π n ( ν j ( ν ) , q F q ) ] | + sup | ν | M sup q [ 0 , 1 ] | I E [ Π n ( ν j ( ν ) , q F q ) ] I E [ Π n ( ν , q F q ) ] + sup | ν | M sup q [ 0 , 1 ] | I E [ Π n ( ν , q F q ) ] I E [ Π n ( ν , q ) ] .
Firstly, we write
sup | ν | M sup q [ 0 , 1 ] | Π n ( ν , q ) Π n ( ν j ( ν ) , q ) | 1 n I E [ F 1 ] i Ω i 0
with
Ω i 0 = sup | ν | M sup q [ 0 , 1 ] 1 I | R i ν j ( ν ) Q R ( q | F ) | C l n F i .
As
| Ω i 0 | C , I E [ Ω i 0 ] = O ( l n α F ( f n ) ) and I E [ Ω i 0 2 ] = O ( l n α F ( f n ) )
we get
sup | ν | M sup q [ 0 , 1 ] | Π n ( ν , q ) Π n ( ν j ( ν ) , q ) = O a . c o . log n n α F ( f n ) 1 / 2
and
sup | ν | M sup q [ 0 , 1 ] | I E Π n ( ν , q ) Π n ( ν j ( ν ) , q ) = o log n n α F ( f n ) 1 / 2 .
Similarly,
sup | ν | M sup q [ 0 , 1 ] | Π n ( ν j ( ν ) , q ) Π n ( ν j ( ν ) , q F q ) | 1 n I E [ F 1 ] i Ω i 1
with
Ω i 1 = sup | ν | M sup q [ 0 , 1 ] 1 I | R i ν j ( ν ) t ( q | F ) | C l n + 1 I | R i t ( q F q | F ) | C l n F i .
Since
| Ω i 1 | C , I E [ Ω i 1 ] = O ( l n α F ( f n ) ) and I E [ Ω i 0 1 ] = O ( l n α F ( f n ) )
we obtain
sup | ν | M sup q [ 0 , 1 ] | Π n ( ν , q ) Π n ( ν j ( ν ) , q ) = O a . c o . log n n α F ( f n ) 1 / 2
and
sup | ν | M sup q [ 0 , 1 ] | I E Π n ( ν , q ) Π n ( ν j ( ν ) , q ) = o log n n α F ( f n ) 1 / 2 .
By the same manner
Π n ( ν j , q k ) I E [ Π n ( ν j , q k ) ] = 1 n I E [ F 1 ] i = 1 n Λ i
where
Λ i = 1 I R i t ( q k | F ) 1 I R i ν j + t ( q k | F ) F i I E 1 I R i t ( q k | F ) 1 I R i ν j + t ( q k | F ) F i .
By Bernstein inequality
| Λ i | C and I E Λ i 2 C α F ( f n )
to show that there exits η > 0 such that
n I P sup | ν | M sup q [ 0 , 1 ] | Π n ( ν j ( ν ) , q F q ) I E [ Π n ( ν j ( ν ) , q F q ) ] | η log n n α F ( f n ) n d n 2 max j max k I P | Π n ( ν j , q k ) I E [ Π n ( ν j , q k ) ] | η log n n α F ( f n ) < .
Therfore,
sup q [ 0 , 1 ] sup | ν | M | Π n ( ν , q ) Υ n I E Π n ( ν , q ) Υ n | = O a . c o . log n n α F ( f n )
Obserrve that by the same techniques we prove that
sup q [ 0 , 1 ] | Υ n I E Υ n | = O a . c o . log n n α F ( f n )

References

  1. Ramsay, J.O.; Silverman, B.W. Functional Data Analysis, 2nd ed.; Springer: New York, NY, USA, 2005. [Google Scholar]
  2. Ferraty, F.; Vieu, P. Nonparametric Functional Data Analysis. Theory and Practice; Springer Series in Statistics; Springer: New York, NY, USA, 2006. [Google Scholar]
  3. Xu, Y. Functional Data Analysis. In Springer Handbook of Engineering Statistics; Springer: London, UK, 2023; pp. 67–85. [Google Scholar]
  4. Ahmed, M.S.; Frévent, C.; Génin, M. Spatial Scan Statistics for Functional Data. In Handbook of Scan Statistics; Springer: New York, NY, USA, 2024; pp. 629–645. [Google Scholar]
  5. Collomb, G.; Härdle, W.; Hassani, S. A note on prediction via estimation of the conditional mode function. J. Stat. Plan. Inference 1986, 15, 227–236. [Google Scholar] [CrossRef]
  6. Quintela-Del-Rio, A.; Vieu, P. A nonparametric conditional mode estimate. J. Nonparametr. Stat. 1997, 8, 253–266. [Google Scholar] [CrossRef]
  7. Ioannides, D.; Matzner-Løber, E. A note on asymptotic normality of convergent estimates of the conditional mode with errors-in-variables. J. Nonparametr. Stat. 2004, 16, 515–524. [Google Scholar] [CrossRef]
  8. Louani, D.; Ould-Saïd, E.L.I.A.S. Asymptotic normality of kernel estimators of the conditional mode under strong mixing hypothesis. J. Nonparametr. Stat. 1999, 11, 413–442. [Google Scholar] [CrossRef]
  9. Bouzebda, S.; Khardani, S.; Slaoui, Y. Asymptotic normality of the regression mode in the nonparametric random design model for censored data. Commun. Stat. Theory Methods 2023, 52, 7069–7093. [Google Scholar] [CrossRef]
  10. Bouzebda, S.; Didi, S. Some results about kernel estimators for function derivatives based on stationary and ergodic continuous time processes with applications. Communications in Statistics. Theory Methods 2022, 51, 3886–3933. [Google Scholar] [CrossRef]
  11. Ferraty, F.; Laksaci, A.; Vieu, P. Estimating some characteristics of the conditional distribution in nonparametric functional models. Stat. Inference Stoch. Process. 2006, 9, 47–76. [Google Scholar] [CrossRef]
  12. Bouzebda, S.; Chaouch, M.; Laïb, N. Limiting law results for a class of conditional mode estimates for functional stationary ergodic data. Math. Methods Stat. 2016, 25, 1066–5307. [Google Scholar] [CrossRef]
  13. Ezzahrioui, M.H.; Ould-Saïd, E. Asymptotic normality of a nonparametric estimator of the conditional mode function for functional data. J. Nonparametr. Stat. 2008, 20, 3–18. [Google Scholar] [CrossRef]
  14. Ezzahrioui, M.H.; Saïd, E.O. Some asymptotic results of a non-parametric conditional mode estimator for functional time-series data. Stat. Neerl. 2010, 64, 171–201. [Google Scholar] [CrossRef]
  15. Dabo-Niang, S.; Kaid, Z.; Laksaci, A. Asymptotic properties of the kernel estimate of spatial conditional mode when the regressor is functional. AStA Adv. Stat. Anal. 2015, 99, 131–160. [Google Scholar] [CrossRef]
  16. Bouanani, O.; Laksaci, A.; Rachdi, M.; Rahmani, S. Asymptotic normality of some conditional nonparametric functional parameters in high-dimensional statistics. Behaviormetrika 2019, 46, 199–233. [Google Scholar] [CrossRef]
  17. Azzi, A.; Belguerna, A.; Laksaci, A.; Rachdi, M. The scalar-on-function modal regression for functional time series data. J. Nonparametr. Stat. 2024, 36, 503–526. [Google Scholar] [CrossRef]
  18. Stone, C.J. Consistent nonparametric regression. Ann. Stat. 1977, 5, 595–620. [Google Scholar] [CrossRef]
  19. Koenker, R.; Zhao, Q. Conditional quantile estimation and inference for ARCH models. Econom. Theory 1996, 12, 793–813. [Google Scholar] [CrossRef]
  20. Hallin, M.; Lu, Z.; Yu, K. Local linear spatial quantile regression. Bernoulli 2009, 15, 659–686. [Google Scholar] [CrossRef]
  21. Cardot, H.; Crambes, C.; Sarda, P. Quantile regression when the covariates are functions. J. Nonparametr. Stat. 2005, 17, 841–856. [Google Scholar] [CrossRef]
  22. Wang, H.; Ma, Y. Optimal subsampling for quantile regression in big data. Biometrika 2021, 108, 99–112. [Google Scholar] [CrossRef]
  23. Jiang, Z.; Huang, Z. Single-index partially functional linear quantile regression. Commun. Stat.-Theory Methods 2024, 53, 1838–1850. [Google Scholar] [CrossRef]
  24. Dabo-Niang, S.; Kaid, Z.; Laksaci, A. Spatial conditional quantile regression: Weak consistency of a kernel estimate. Rev. Roum. Math. Pures Appl. 2012, 57, 311–339. [Google Scholar]
  25. Chowdhury, J.; Chaudhuri, P. Nonparametric depth and quantile regression for functional data. Bernoulli 2019, 25, 395–423. [Google Scholar] [CrossRef]
  26. Mutis, M.; Beyaztas, U.; Karaman, F.; Lin Shang, H. On function-on-function linear quantile regression. J. Appl. Stat. 2025, 52, 814–840. [Google Scholar] [CrossRef] [PubMed]
  27. Fan, J. Local Polynomial Modelling and Its Applications: Monographs on Statistics and Applied Probability 66; Routledge: Abingdon-on-Thames, UK, 2018. [Google Scholar]
  28. Rachdi, M.; Laksaci, A.; Demongeot, J.; Abdali, A.; Madani, F. Theoretical and practical aspects of the quadratic error in the local linear estimation of the conditional density for functional data. Comput. Stat. Data Anal. 2014, 73, 53–68. [Google Scholar] [CrossRef]
  29. Baíllo, A.; Grané, A. Local linear regression for functional predictor and scalar response. J. Multivar. Anal. 2009, 100, 102–111. [Google Scholar] [CrossRef]
  30. Barrientos-Marin, J.; Ferraty, F.; Vieu, P. Locally modelled regression and functional data. J. Nonparametr. Stat. 2010, 22, 617–632. [Google Scholar] [CrossRef]
  31. Berlinet, A.; Elamine, A.; Mas, A. Local linear regression for functional data. Ann. Inst. Stat. Math. 2011, 63, 1047–1075. [Google Scholar] [CrossRef]
  32. Demongeot, J.; Laksaci, A.; Madani, F.; Rachdi, M. Functional data: Local linear estimation of the conditional density and its application. Statistics 2013, 47, 26–44. [Google Scholar] [CrossRef]
  33. Messaci, F.; Nemouchi, N.; Ouassou, I.; Rachdi, M. Local polynomial modelling of the conditional quantile for functional data. Stat. Methods Appl. 2015, 24, 597–622. [Google Scholar] [CrossRef]
  34. Laksaci, A.; Ould Saïd, E.; Rachdi, M. Uniform consistency in number of neighbors of the k NN estimator of the conditional quantile model. Metrika 2021, 84, 895–911. [Google Scholar] [CrossRef]
  35. Ota, H.; Kato, K.; Hara, S. Quantile regression approach to conditional mode estimation. Electron. J. Stat. 2019, 13, 3120–3160. [Google Scholar] [CrossRef]
  36. Almulhim, F.A.; Alamari, M.B.; Bouzebda, S.; Kaid, Z.; Laksaci, A. Robust Estimation of L 1-Modal Regression Under Functional Single-Index Models for Practical Applications. Mathematics 2025, 13, 602. [Google Scholar] [CrossRef]
  37. Belarbi, F.; Chemikh, S.; Laksaci, A. Local linear estimate of the nonparametric robust regression in functional data. Stat. Probab. Lett. 2018, 134, 128–133. [Google Scholar] [CrossRef]
  38. Iglesias Pérez, M.D.C. Estimación de la función de distribución condicional en presencia de censura y truncamiento: Una aplicación al estudio de la mortalidad en pacientes diabéticos. Estad. Esp. 2003, 45, 275–302. [Google Scholar]
  39. Yuan, M. GACV for quantile smoothing splines. Comput. Stat. Data Anal. 2006, 50, 813–829. [Google Scholar] [CrossRef]
  40. Panchbhai, K.G.; Lanjewar, M.G. Portable system for cocoa bean quality assessment using multi-output learning and augmentation. Food Control 2025, 174, 111234. [Google Scholar] [CrossRef]
Figure 1. Smooth curves F i ( t ) for 100 equispaced measurements t in ( 0 , 1 ) , as displayed on the X-axis.
Figure 1. Smooth curves F i ( t ) for 100 equispaced measurements t in ( 0 , 1 ) , as displayed on the X-axis.
Axioms 14 00537 g001
Figure 2. Rough curves F i ( t ) for 100 equispaced measurements t in ( 0 , 1 ) , as displayed on the X-axis.
Figure 2. Rough curves F i ( t ) for 100 equispaced measurements t in ( 0 , 1 ) , as displayed on the X-axis.
Axioms 14 00537 g002
Figure 3. NIR curves.
Figure 3. NIR curves.
Axioms 14 00537 g003
Figure 4. Moisture prediction.
Figure 4. Moisture prediction.
Axioms 14 00537 g004
Figure 5. Fat prediction.
Figure 5. Fat prediction.
Axioms 14 00537 g005
Table 1. M S E for different estimators in different scenarios.
Table 1. M S E for different estimators in different scenarios.
Cond. Dist.ModelCURVES MR ^ MR ¯ MR ˜
NormalHom.modelSmooth0.02580.08970.1672
Rough0.10670.46760.5679
NormalHet.modelSmooth0.26720.74261.0578
Rough0.36171.27251.8811
LaplaceHom. modelSmooth0.42570.61500.6317
Rough0.98041.09221.1788
LaplaceHet. modelSmooth0.89671.68241.7088
Rough0.91762.45212.6588
WeibullHom. modelSmooth0.51791.50051.5446
Rough0.83992.48732.7098
WeibullHet. modelSmooth0.78401.62531.4102
Rough0.97053.35674.3456
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Almulhim, F.A.; Alamari, M.B.; Laksaci, A.; Kaid, Z. Modal Regression Estimation by Local Linear Approach in High-Dimensional Data Case. Axioms 2025, 14, 537. https://doi.org/10.3390/axioms14070537

AMA Style

Almulhim FA, Alamari MB, Laksaci A, Kaid Z. Modal Regression Estimation by Local Linear Approach in High-Dimensional Data Case. Axioms. 2025; 14(7):537. https://doi.org/10.3390/axioms14070537

Chicago/Turabian Style

Almulhim, Fatimah A., Mohammed B. Alamari, Ali Laksaci, and Zoulikha Kaid. 2025. "Modal Regression Estimation by Local Linear Approach in High-Dimensional Data Case" Axioms 14, no. 7: 537. https://doi.org/10.3390/axioms14070537

APA Style

Almulhim, F. A., Alamari, M. B., Laksaci, A., & Kaid, Z. (2025). Modal Regression Estimation by Local Linear Approach in High-Dimensional Data Case. Axioms, 14(7), 537. https://doi.org/10.3390/axioms14070537

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop