Abstract
Allowing for misspecification in the linear conditional quantile function, this paper provides a new interpretation and the semiparametric efficiency bound for the quantile regression parameter in Koenker and Bassett (1978). The first result on interpretation shows that under a mean-squared loss function, the probability limit of the Koenker–Bassett estimator minimizes a weighted distribution approximation error, defined as , i.e., the deviation of the conditional distribution function, evaluated at the linear quantile approximation, from the quantile level. The second result implies that the Koenker–Bassett estimator semiparametrically efficiently estimates the quantile regression parameter that produces parsimonious descriptive statistics for the conditional distribution. Therefore, quantile regression shares the attractive features of ordinary least squares: interpretability and semiparametric efficiency under misspecification.
Keywords:
semiparametric efficiency bounds; misspecification; conditional quantile function; conditional distribution function; best linear approximation JEL classifications:
C14; C21
1. Introduction
This paper revisits the approximation properties of the linear quantile regression under misspecification ([1,2,3]). The quantile regression estimator, introduced by the seminal paper of Koenker and Bassett [4], offers parsimonious summary statistics for the conditional quantile function and is computationally tractable. Since the development of the estimator, researchers have frequently used quantile regression, in conjunction with ordinary least squares regression, to analyse how the outcome variable responds to the explanatory variables. For example, to model wage structure in labour economics, Angrist, Chernozhukov, and Fernández-Val [1] study returns to education at different points in the wage distribution and changes in inequality over time. A thorough review of recent developments in quantile regression can be found in [5]. The object of interest of this paper is the quantile regression (QR) parameter that is the probability limit of the Koenker–Bassett estimator without assuming the true conditional quantile function to be linear. Two results are presented: a new interpretation and the semiparametric efficiency bound for the QR parameter.
The topic of interest is the conditional distribution function (CDF) of a continuous response variable Y given the regressor vector X, denoted as . An alternative for the CDF is the conditional quantile function (CQF) of Y given X, defined as for any quantile index . Assuming integrability, the CQF minimizes the check loss function
where is the set of measurable functions of X, is known as the check function and is the indicator function. A linear approximation to the CQF is provided by the QR parameter , which solves the population minimization problem
assuming the integrability and uniqueness of the solution, and d is the dimension of X. The QR parameter provides a simple summary statistic for the CQF. The QR estimator introduced in [4] is the sample analogue
for the random sample on the random variables . By the equivalent first-order condition, this estimator is also the generalized method of moments (GMM) estimator based on the unconditional moment restriction ([6,7])
This paper focuses on the population QR parameter defined by (1) or equivalently (3).
If the CQF is modelled to be linear in the covariates or , the coefficient satisfies the conditional moment restriction
almost surely. In the theoretical and applied econometrics literature, this linear QR model is often assumed to be correctly specified. Nevertheless, a well-known crossing problem arises: the CQF for different quantiles may cross at some values of X, except when is the same for all τ. A logical monotone requirement is violated for or its estimator to be weakly increasing in the probability index τ given X. The crossing problem for estimation could be treated by rearranging the estimator (for example, see [8] and the references therein.1). However, the crossing problem remains for the population CQF, suggesting that the linear QR model (4) is inherently misspecified. That is, there is no satisfying the conditional moment (4) almost surely. Therefore, the parameter of interest in this paper is the QR parameter defined by (1) or (3) without the linear CQF assumption in (4). We can view as the pseudo-true value of the linear QR model under misspecification. As the Koenker–Bassett QR estimator is widely used, it is important to understand the approximation nature of the estimand.
For the mean regression counterpart, ordinary least squares (OLS) consistently estimates the linear conditional expectation and minimizes mean-squared error loss for fitting the conditional expectation under misspecification. Chamberlain [9] proves the semiparametric efficiency of the OLS estimator, which provides additional justification for the widespread use of OLS. The attractive features of OLS, interpretability and semiparametric efficiency, under misspecification, motivate my investigation of parallel properties in QR. I study how this QR parameter approximates the CQF and the CDF and calculate its semiparametric efficiency bound.
The first contribution of this paper is on how minimizes the distribution approximation error, defined by , under a mean-squared loss function. The first-order condition (3) can be understood as the orthogonality condition of the covariates X and the distribution approximation error in the projection model. I show that the QR parameter minimizes the mean-squared distribution approximation error, inversely weighted by the conditional density function . Angrist, Chernozhukov, and Fernández-Val [1] (henceforth ACF) show that minimizes the mean-squared quantile specification error, defined by , using a weight primarily determined by the conditional density. ACF’s study, as well as my own results, suggests that QR approximates the CQF more accurately at points with more observations, but the corresponding CDF evaluated at the approximated point is more distant from the targeted quantile level τ. This trade-off is controlled by the conditional density, which is distinct from OLS approximating the conditional mean, because the distribution and quantile functions are generally nonlinear operators. This observation is novel and increases the understanding of how the QR summarizes the outcome distribution. A numerical example in Figure 1 in Section 4 illustrates this finding.
The second result is the semiparametric efficiency bound of the . Chamberlain’s results in [9] on the mean regression based on differentiable moment restrictions cannot be applied to semiparametric efficiency for QR, due to the lack of moment function differentiability in (3). Although Ai and Chen [10] provide general results for sequential moment restrictions containing unknown functions, which could cover the quantile regression setting, I calculate the efficiency bound accommodating regularity conditions specifically for the QR parameter using the method of Severini and Tripathi [11]. It follows that the misspecification-robust asymptotic variance of the QR estimator in (2) attains this bound, which means no regular2 estimator for (3) has smaller asymptotic variance than . This result might be expected for an M-estimator, but, to my knowledge, the QR application has not been demonstrated and discussed rigorously in any publication. Furthermore, I calculate the efficiency bounds for jointly estimating QR parameters at a finite number of quantiles for both linear projection (3) and linear QR (4) models. Employing the widely-used method of Newey [12], Newey and Powell [13] find the semiparametric efficiency bound for of the correctly-specified linear CQF in (4). Note that the efficiency bounds for (3) do not imply the bounds for (4); nor does the converse hold.
In Section 2, I discuss the interpretation of the misspecified QR model in terms of approximating the CDF and the CQF. The theorems for the semiparametric efficiency bounds are in Section 3. In Section 4, I discuss the parallel properties of QR and OLS. The paper is concluded by a review of some existing efficient estimators for linear projection model (3) and linear QR model (4).
2. Interpreting QR under Misspecification
Let Y be a continuous response variable and X be a regressor vector. The quantile-specific residual is defined as the distance between the response variable and the CQF, with the conditional density at or at for any . This is a semiparametric problem in the sense that the distribution functions of and X, as well as the CQF, are unspecified and unrestricted other than by the following assumptions, which are standard in QR models. I assume the following regularity conditions, based on the conditions of Theorem 3 in ACF.
- (R1)
- are independent and identically distributed on the probability space for each n;
- (R2)
- the conditional density exists and is bounded and uniformly continuous in y, uniformly in x over the support of X;
- (R3)
- is positive definite for all , where is uniquely defined in (1);
- (R4)
- for some ;
- (R5)
- to be bounded away from zero.
The parameter of interest is equivalent to solving
by applying the law of iterated expectations on Equation (3). Equation (5) states that X is orthogonal to the distribution approximation error . The following theorem interprets QR via a weighed mean-squared error loss function on the distribution approximation error.
Theorem 1.
Assume (R1)–(R5). Then, solves the equation
Furthermore, if is positive definite at , then is the unique solution to this problem (6).
Proof of Theorem 1.
The objective function in (6) is finite by the assumptions. Any fixed point would solve the first-order condition, . By the law of iterated expectations, (3) implies the above first-order condition. Therefore, solves (6). When the second-order condition holds, i.e., is positive definite at , solves (6) uniquely. ☐
Theorem 1 states that is the unique fixed point to an iterated minimum distance approximation, with a weight of a function of X only. The mean-squared loss makes it clear how the linear function matches the CDF to the targeted probability of interest. The loss function puts more weight on points where the conditional density is small. As a result, the distribution approximation error is smaller at points with smaller conditional density.
Now, I discuss the approximation nature of QR based on the distributional approximation error and quantile specification error. ACF interpret QR as the minimizer of the weighted mean-squared error loss function for quantile specification error, defined as the deviation between the approximation point and the true CQF ,3
ACF define in (8) to be the importance weights that are the averages of the response variable over a line connecting the approximation point and the true CQF. ACF note that the regressors contribute disproportionately to the QR estimate and the primary determinant of the importance weight is the conditional density.
Moreover, the first-order condition implied by (7) is a weighted orthogonal condition of the quantile specification error. A Taylor expansion provides intuition to connect the distribution approximation error and the quantile specification error: by . This observation implies the quantile specification error is smaller at points where the conditional density is larger. On the other hand, the distribution approximation error is larger at points with larger . Comparing with the OLS, where the mean operator is linear, the CDF and its inverse operator, the CQF, are generally nonlinear. The distribution approximation error can be interpreted as the distance after a nonlinear transformation by the CDF, . A Taylor expansion linearizes the distribution function to the quantile specification error multiplied by the conditional density function. The conditional density plays a crucial role on weighting the distribution approximation error and the quantile specification error. The above discussion provides additional insights to how the QR parameter approximates the CQF and fits the CDF to the targeted quantile level.
Remark 1
(Mean-squared loss under misspecification). The linear function is the best linear approximation under the check loss function in (1). While is the least absolute derivations estimation, the QR parameter for is the best linear predictor for a response variable under the asymmetric loss function in (1). ACF note that the prediction under the asymmetric check loss function is often not the object of interest in empirical work, with the exception of the forecasting literature, for example [15]. For the mean regression counterpart, OLS consistently estimates the linear conditional expectation and minimizes mean-squared error loss for fitting the conditional expectation under misspecification. The robust nature of OLS also motivates research on misspecification in panel data models. For example, Galvao and Kato [16] investigate linear panel data models under misspecification. The pseudo-true value of the fixed effect estimator provides the best partial linear approximation to the conditional mean given the explanatory variables and the unobservable individual effect.4
3. The Semiparametric Efficiency Bounds
Section 3.1 presents the semiparametric efficiency bound for the unconditional moment restriction (3). Section 3.2 discusses the existing results on the semiparametric efficiency bound for the conditional moment restriction (4).
3.1. QR under Misspecification
I calculate the semiparametric efficiency bound for the unconditional moment restriction (3) by the approach of Severini and Tripathi [11].
In general, the semiparametrically-efficient joint asymptotic covariance of the estimators for is , for any , , for a finite integer .
Proof of Theorem 2.
See the Appendix. ☐
My proof accommodates the regularity assumptions for quantile regression and modifies Section 9 of [11]. For example, the covariate X can contain discrete components, by constructing two tangent spaces for the conditional density of Y given X and the marginal density of X, respectively. In the efficiency bound, is obtained by assuming the interchangeability of integration and differentiation for the nonsmooth check function.5
The method in [11] has been used in the monotone binary model in [18], Lewbel [19] latent variable model in [20] and the partial linear single index model in [21], for example. I work in the Hilbert space of tangent vectors of the square-root density functions and using the Riesz–Fréchet representation theorem. Another equivalent approach in [12] works in a Hilbert space of random variables and uses the projection on the linear space spanned by the scores from the one-dimensional subproblems to find the efficient influence function. The efficiency bound is then the second moment of the efficient influence function, . Newey’s efficient influence function is the score function evaluated at the unique representers by the Riesz–Fréchet theorem used in [11]; a more detailed comparison of these two approaches is discussed in [11].
ACF show that the QR process is asymptotically mean-zero Gaussian with the covariance function for any , which is the semiparametric efficiency bound in Theorem 2. This asymptotic covariance under misspecification for a single quantile, , has been presented in [2] and [3]. Hahn [3] further shows the QR estimator is well approximated by the bootstrap distribution, even when the linear quantile restriction is misspecified. An alternative estimator for the misspecification-robust asymptotic covariance matrix of is the nonparametric kernel method in ACF.
3.2. QR for Linear Specification
Assuming the linear QR model in (4) is correctly specified, i.e., almost surely, the asymptotic covariance for the QR process derived by ACF is simplified to , where for any . The asymptotic covariance for a single quantile τ, first derived by Powell [7], is widely used for inference in most empirical studies, which implicitly assume correct specification.
The semiparametric efficiency bound for the correctly-specified quantile regression (4) is , where a.s. and is assumed to be finite and nonsingular. This is first calculated in [13] using the method developed in [12]. If, in addition, the conditional density function of given X is independent of X, i.e., , and , the semiparametric efficiency bound becomes . This asymptotic covariance is attained by , first shown in [4]. This has an interesting resemblance to the fact that the OLS estimator is semiparametrically efficient in a homoskedastic regression model, i.e., , , and .
I further show, in general, that the semiparametrically-efficient joint asymptotic covariance of the estimators for is
for any , , for any finite integer . The regularity conditions imposed, (R1), (R2) and (R4), are weaker than the assumptions in [13]; for example, they assume is absolutely continuous in ε, which implies uniform continuity in (R2). See the Appendix for the detailed proof for (9).
4. Discussion and Conclusions
Misspecification is a generic phenomenon; especially in quantile regression (QR), the true conditional quantile function (CQF) might be nonlinear or different functions of the covariates at different quantiles. Table 1 summarizes the parallel properties of QR and OLS. Under misspecification, the pseudo-true OLS coefficient can be interpreted as the best linear predictor of the conditional mean function, , in the sense that the coefficient minimizes the mean-squared error of the linear approximation to the conditional mean. The approximation properties of OLS have been well studied (see, for example, [22]). With respect to the QR counterpart, I present the inverse density-weighted mean-squared error loss function based on the distribution approximation error . This result complements the interpretation based on the quantile specification error in [1]. My results imply that the Koenker–Bassett estimator is semiparametrically efficient for misspecified linear projection models and correctly specified linear quantile regression models when does not depend on X. Alternatively, the smoothed empirical likelihood estimator using the unconditional moment restriction in [23] has the same asymptotic distribution as the Koenker-Bassett estimator and, hence, attains the efficiency bound.
Table 1.
Summary properties of OLS and quantile regression (QR).
| OLS | QR | |
|---|---|---|
| Linear Projection Model | ||
| objective minimized | ||
| (interpretation) | ||
| unconditional moment | ||
| (interpretation) | ||
| efficient estimators | ||
| (Koenker–Bassett) | ||
| (OLS) | ||
| asymptotic covariance | ||
| efficiency bounds | Chamberlain (1987) [9] | Theorem 2 |
| Linear Regression Model | ||
| conditional moment | ||
| or | ||
| efficiency bounds | Chamberlain (1987) [9] | Newey and Powell (1990) [13] |
| homoscedasticity-type | ||
| condition | ||
| efficient estimators | OLS | Koenker–Bassett |
and where ; The feasible generalized least squares estimator is semiparametrically efficient, for example.
Under the linear quantile regression model, the Koenker–Bassett estimator consistently estimates the true , although it is not semiparametrically efficient given heteroskedasticity. Researchers have proposed many efficient estimators for the correctly-specified linear quantile regression parameter, for example the one-step score estimator in [13], the smoothed conditional empirical likelihood estimator in [24] and the sieve minimum distance (SMD) estimator in [25,26]. However, for all of these estimators, the pseudo-true values under misspecification are different, and their interpretations have not been thoroughly studied. Therefore, the semiparametric efficiency bounds of these pseudo-true values are also different. For example, an unweighted SMD estimator converges to a pseudo-true value that minimizes .6 The first-order condition is , which is the unconditional moment used in [13] for the semiparametrically efficient GMM estimator under correct specification. The conditional density weight is similar to the generalized least squares in the mean regression in that it uses a weight function of the conditional variance to construct an efficient estimator.
It is interesting to note that the pseudo-true value of the SMD estimator minimizes . The distribution approximation error is weighted evenly over the support of X for , in contrast to the QR parameter, which is weighted more at points with smaller conditional density in Theorem 1. Therefore, the SMD estimator might have more desirable and reasonable approximation properties than QR. Nevertheless, the SMD estimator is computationally more demanding than the Koenker–Bassett estimator. A numerical example in Figure 1 illustrates how the Koenker and Bassett (KB) and SMD estimators approximate the CQF and the CDF.
Figure 1.
This numerical example is constructed by , and . Therefore, , and . Set for the median. The red solid line is for the QR parameter defined in (3) and estimated by the Koenker-Bassett (KB) estimator. The blue dashed line is the approximation by the SMD estimator minimizing . The approximations are and . The left panel shows the linear approximations , and the true CQF. The green circles are 300 random draws from the DGP. The right panel shows the corresponding CDFs and . For smaller x where the conditional density is larger, the quantile specification error of SMD is smaller than that of KB in the left panel. For the distribution approximation error in the right panel, SMD weights more evenly over the support of X, while KB has smaller distribution approximation error at larger x with smaller density.
This discussion leads to open-ended questions: What is an appropriate linear approximation or a meaningful summary statistic for the nonlinear CQF? How should economists measure the marginal effect of the covariates on the CQF? An approach that circumvents this problem is measuring the average marginal response of the covariates on the CQF directly. The average quantile derivative, defined as where is a weight function, offers such a succinct summary statistic ([27]). Sasaki [28] investigates the question that quantile regressions may misspecify true structural functions. He provides a causal interpretation of the derivative of the CQF, which identifies a weighted average of heterogeneous structural partial effects among the subpopulation of individuals at the conditional quantile of interest. Sasaki’s work adds economic content to this misspecified question. This paper complements the prior literature on understanding how the QR statistically summarizes the outcome distribution.
Acknowledgments
I am grateful to Bruce Hansen and Jack Porter for invaluable guidance and comments. I also thank Yuya Sasaki for helpful discussion. I would also like to thank three anonymous referees for their suggestions. All remaining errors are mine.
Conflicts of Interest
The author declares no conflict of interest.
Appendix
Proof of Theorem 3.
I implement the main results in [11]. I start with the definitions and construct the Hilbert space. The unknown probability density or mass function of the random vector with respect to the measure P, the products of the Lebesgue measure and ,7 is written as . The functionals and belong to the following sets defined by the regularity conditions
Let .
Definition 1.
A vector is said to be tangent to at if it is the slope of at , i.e., .
Definition 2.
The tangent space to at the true value , denoted as is the smallest linear space, which is closed under the -norm and contains all tangent to at .
Severini and Tripahi [11] show that the tangent space is the product of and , where
The pseudo-true model is the unconditional moment restriction in (3). Here, is the expectation with respect to the true density functions , and denotes the pseudo-true for notational simplicity. The objective is to estimate the efficiency bound for estimating . Equivalently, I can instead look at the efficiency bound for estimating the functional for any arbitrary vector . Severini and Tripahi [11] parameterize and as a one-dimensional subproblem. For some , let be a curve from into , which passes through at . That is, estimating at the true parameter is equivalent to estimating . The likelihood of estimating t using a single observation is given by . Therefore, the score function for estimating is
Then, the Fisher information at can be written as
where the third equality is because , and denotes integrals with respect to the distribution of X. Therefore, the Fisher information inner product and the corresponding norm are defined as
for any , which is a closed subset of . Hence, I have constructed the Hilbert space .
Now, I am ready to derive the efficiency bounds. It is known that the information inequality holds for all regular estimators, i.e., the asymptotic covariance of the estimator . The semiparametric bound can be interpreted as the supremum of the asymptotic covariance over the parametric submodels. By [11], the lower bound is
The third equality is the norm of the linear functional , the path-wise derivative of η (p. 105 in [29]). The forth equality is from the Riesz–Fréchet theorem: there exists a unique for the continuous linear functional on the Hilbert space , such that for all , i.e.,
Therefore, to find the lower bound by (10), I need to find , which is known as the representer of the continuous linear functionals .
The submodel is also required to satisfy the unconditional moment restriction, . For any , . I simultaneously estimate , a -dimensional vector, so the unconditional moment restriction is
Taking the derivative with respect to t evaluated at ,8
where the second term uses the fact that if , then . Note that , which is assumed to be positive definite by (R3), so exists. Define
so exists. Then
I confirm that is a continuous linear functional on , so η is indeed path-wise differentiable. Using (12) to replace in the left-hand side of (11), I can find the representer for as
because . It can be easily checked that . For notational ease, denote and , . Then, the lower bound for regular -consistent estimators of is
where
by the law of iterated expectations, and so, .
Therefore, the lower bound for estimating is . The asymptotic covariance of the estimators for and cannot be smaller than .
☐
Remark 2.
Consider the efficiency bound for estimating one single quantile using Newey’s approach in [12]. Severini and Triphathi [11] claim that the efficient influence function for is . Then, the efficient influence function for is , where the efficient score . Newey shows that the semiparametric bound is .
Proof of the Semiparametric Efficiency Bound for the Linear QR Model (4).
Under the linear specification, . The random vectors satisfy the conditional moment restriction , i.e., , where the joint distribution of is . The Hilbert space and are defined in the proof of Theorem 2. Consider any , . The parameterized submodel also has to satisfy the moment condition
Remark 3.
Consider the efficiency bound for estimating one single quantile by Newey’s approach in [12]. Severini and Triphathi [11] claim that Newey’s efficient influence function for is . Then, the efficient influence function for is , where the efficient score . Newey shows the semiparametric bound is .
References
- J. Angrist, V. Chernozhukov, and I. Fernández-Val. “Quantile regression under misspecification, with an application to the U.S. wage structure.” Econometrica 74 (2006): 539–563. [Google Scholar]
- T.-H. Kim, and H. White. “Estimation, inference, and specification testing for possibly misspecified quantile regression.” Adv. Econom. 17 (2003): 107–132. [Google Scholar]
- J. Hahn. “Bayesian bootstrap of the quantile regression estimator: A large sample study.” Int. Econ. Rev. 38 (1997): 795–808. [Google Scholar] [CrossRef]
- R. Koenker, and G. Bassett. “Regression quantile.” Econometrica 46 (1978): 33–50. [Google Scholar] [CrossRef]
- R. Koenker. Quantile Regression. Econometric Society Monographs. New York, NY, USA: Cambridge University Press, 2005. [Google Scholar]
- J. Powell. “Least absolute deviations estimation for the censored regression model.” J. Econom. 25 (1984): 303–325. [Google Scholar] [CrossRef]
- J. Powell. “Censored regression quantiles.” J. Econom. 32 (1986): 143–155. [Google Scholar] [CrossRef]
- V. Chernozhukov, I. Fernández-Val, and A. Galichon. “Quantile and probability curves without crossing.” Econometrica 78 (2010): 1093–1125. [Google Scholar]
- G. Chamberlain. “Asymptotic efficiency in estimation with conditional moment restrictions.” J. Econom. 34 (1987): 305–334. [Google Scholar] [CrossRef]
- C. Ai, and X. Chen. “The semiparametric efficiency bound for models of sequential moment restrictions containing unknown functions.” J. Econom. 170 (2012): 442–457. [Google Scholar] [CrossRef]
- T.A. Severini, and G. Tripathi. “A simplified approach to computing efficiency bounds in semiparametric models.” J. Econom. 102 (2001): 23–66. [Google Scholar] [CrossRef]
- W.K. Newey. “Semiparametric efficiency bounds.” J. Appl. Econom. 5 (1990): 99–135. [Google Scholar] [CrossRef]
- W.K. Newey, and J.L. Powell. “Efficient estimation of linear and type 1 censored regression models under conditional quantile restrictions.” Econom. Theory 6 (1990): 295–317. [Google Scholar] [CrossRef]
- R. Koenker, S. Leorato, and F. Peracchi. Distributional vs. Quantile Regression. Ceis Working Paper No. 300. 2013. Available online: http://papers.ssrn.com/sol3/papers.cfm?abstract_id=2368737 (accessed on 20 November 2015).
- I. Komunjer. “Quasi-maximum likelihood estimation for conditional quantiles.” J. Econom. 128 (2005): 137–164. [Google Scholar] [CrossRef]
- A.F. Galvao, and K. Kato. “Estimation and inference for linear panel data models under misspecification when both n and t are large.” J. Bus. Econ. Stat. 32 (2014): 285–309. [Google Scholar] [CrossRef]
- Y. Lee. “Bias in dynamic panel models under time series misspecification.” J. Econom. 169 (2012): 54–60. [Google Scholar] [CrossRef]
- T. Magnac, and E. Maurin. “Identification and information in monotone binary models.” J. Econom. 139 (2007): 76–104. [Google Scholar] [CrossRef]
- A. Lewbel. “Semiparametric latent variable model estimation with endogenous or mismeasured regressors.” Econometrica 66 (1998): 105–122. [Google Scholar] [CrossRef]
- D.T. Jacho-Chávez. “Efficiency bounds for semiparametric estimation of inverse conditional-density-weighted functions.” Econom. Theory 25 (2009): 847–855. [Google Scholar] [CrossRef]
- T. Chen, and T. Parker. “Semiparametric efficiency for partially linear single-index regression models.” J. Multivar. Anal. 130 (2014): 376–386. [Google Scholar] [CrossRef]
- H. White. “Using least squares to approximate unknown regression functions.” Int. Econ. Rev. 21 (1980): 149–170. [Google Scholar] [CrossRef]
- Y.-J. Whang. “Smoothed empirical likelihood methods for quantile regression models.” Econom. Theory 22 (2006): 173–205. [Google Scholar] [CrossRef]
- T. Otsu. “Conditional empirical likelihood estimation and inference for quantile regression models.” J. Econom. 142 (2008): 508–538. [Google Scholar] [CrossRef]
- X. Chen, and D. Pouzo. “Efficient estimation of semiparametric conditional moment models with possibly nonsmooth residuals.” J. Econom. 152 (2009): 46–60. [Google Scholar] [CrossRef]
- X. Chen, and D. Pouzo. “Estimation of nonparametric conditional moment models with possibly nonsmooth generalized residuals.” Econometrica 80 (2012): 277–321. [Google Scholar]
- P. Chaudhuri, K. Doksum, and A. Samarov. “On average derivative quantile regression.” Ann. Stat. 25 (1997): 715–744. [Google Scholar] [CrossRef]
- Y. Sasaki. “What do quantile regressions identify for general structural functions? ” Econom. Theory 31 (2015): 1102–1116. [Google Scholar] [CrossRef]
- D.G. Luenberger. Optimization by Vector Space Methods. New York, NY, USA: Wiley, 2005. [Google Scholar]
- 1.Chernozhukov, Fernández-Val, and Galichon [8] rearrange an estimator to be monotonic. The original estimator can be computationally tractable. The rearranged monotonic estimated CDF is . The rearranged quantile estimation is .
- 2.See [12] for the definition of regular estimators.
- 3.For estimation, [14] studies different approaches based on the distribution regression and quantile regression.
- 4.Galvao and Kato [16] show that misspecification affects the bias correction and convergence rate of the estimator and provide a misspecification-robust inference procedure. In panel models under time series misspecification, Lee [17] proposes bias reduction methods for the incidental parameter.
- 5.Severini and Tripathi construct the tangent space for the continuous and bounded joint density in Section 9 of [11]. Additionally, they define J on the derivative of the moment restriction.
- 6.The conditional moment restriction in (4) can be expressed as . In [26], an unweighted penalized sieve minimum distance estimator minimizes a possibly penalized consistent estimate of the minimum distance criterion, .
- 7. may not be a Lebesgue measure, since I allow discrete components in the covariates X.
- 8.The interchange of differentiation and integration is allowed, assumed throughout [11], by the smoothness of in by the construction of regular parametric submodels; see [12] for details.
© 2015 by the author; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons by Attribution (CC-BY) license ( http://creativecommons.org/licenses/by/4.0/).