Abstract
Minimum Renyi’s pseudodistance estimators (MRPEs) enjoy good robustness properties without a significant loss of efficiency in general statistical models, and, in particular, for linear regression models (LRMs). In this line, Castilla et al. considered robust Wald-type test statistics in LRMs based on these MRPEs. In this paper, we extend the theory of MRPEs to Generalized Linear Models (GLMs) using independent and nonidentically distributed observations (INIDO). We derive asymptotic properties of the proposed estimators and analyze their influence function to asses their robustness properties. Additionally, we define robust Wald-type test statistics for testing linear hypothesis and theoretically study their asymptotic distribution, as well as their influence function. The performance of the proposed MRPEs and Wald-type test statistics are empirically examined for the Poisson Regression models through a simulation study, focusing on their robustness properties. We finally test the proposed methods in a real dataset related to the treatment of epilepsy, illustrating the superior performance of the robust MRPEs as well as Wald-type tests.
Keywords:
generalized linear model; independent and nonidentically distributed observations; minimum Rényi’s pseudodistance estimators; robust Wald-type test statistics for GLMs; influence function for GLMs; poisson regression model MSC:
62F35; 62J12
1. Introduction
Generalized linear models (GLMs) were first introduced by Nelder and Wedderburn [] and later expanded upon by McCullagh and Nelder []. The GLMs represent a natural extension of the standard linear regression models, which enclose a large variety of response variable distributions, including distributions of count, binary, or positive values. Let be independent response variables. The classical GLM assumes that the density function of each random variable belongs to the exponential family, having the form
for where the functions and are known. Therefore, the observations are independent but not identically distributed, depending on a location parameter and a nuisance parameter Further, we denote by the expectation of the random variable and we assume that there exists a monotone differentiable function, so called link function g, verifying
with the regression parameter vector. The -vector of explanatory variables, is assumed to be nonrandom, i.e., the design matrix is fixed. Correspondingly, the location parameter depends on the explanatory variables the density function given in (1) can be written as empathizing its dependence of and .
The maximum likelihood estimator (MLE) and the quasilikelihood estimators were well studied for the GLMs, and it is well known that they are asymptotically efficient but lack robustness in the presence of outliers, which can result in a significant estimation bias. Jaenada and Pardo [] revised the different robust estimators in the statistical literature and studied the lack of robustness of the MLE as well. Among others, Stefanski et al. [] studied optimally bounded score functions for the GLM and generalized the results obtained by Krasker and Welsch [] for classical LRMs. Künsch et al. [] introduced the so-called conditionally unbiased bounded-influence estimate, and Morgenthaler [], Cantoni and Ronchetti [], Bianco and Yohai [], Croux and Hesbroeck [], Bianco et al. [], and Valdora and Yohai [] continued the development of robust estimators for the GLMs based on general M-estimators. Later, Ghosh and Basu [] proposed robust estimators for the GLM, based on the density power divergence (DPD) introduced in Basu et al. [].
There are not many papers considering robust tests for GLMs. In this sense, Basu et al. [] considered robust Wald-type tests based on the minimum DPD estimator, but assuming random explanatory variables for the GLM. The main purpose of this paper is to introduce new robust Wald-type tests based on the MRPE under fixed (not random) explanatory variables.
Broniatowski et al. [] presented robust estimators for the parameters of the linear regression model (LRM) with random explanatory variables and Castilla et al. [] considered Wald-type test statistics, based on MRPE, for the LRM. Toma and Leoni–Aubin [] defined new robustness and efficient measures based on the RP and Toma et al. [] considered the MRPE for general parametric models, and constructed a model selection criterion for regression models. The term “Rényi pseudodistance” (RP) was adopted in Broniatowski et al. [] because of its similarity with the Rényi divergence (Rényi []), although this family of divergences was considered previously in Jones et al. []. Fujisawa and S. Eguchi [] used the RP under the name of -cross entropy, introduced robust estimators obtained by minimizing the empirical estimate of the -cross entropy (or the -divergence associated to the -cross entropy) and studied their properties. Further, Hirose and Masuda [] considered the likelihood function to find robust estimation. Using the -divergence, Kawashima and Fujisawa [,] presented robust estimators for sparse regression and sparse GLMs with random covariates. The robustness of all the previous estimators is based on density power weight, which gives a small weight to outliers observations. This idea was also developed by Basu et al. [] for the minimum DPD estimator and was considered some years ago by Windham []. More concretely, Basu et al. [] considered the density power function multiplied by the score function.
The outline of the paper is as follows: in Section 2, some results in relation to the MRPEs for GLMs, previously obtained in Jaenada and Pardo [], are presented. Section 3 introduces and studies Wald-type tests based on the MRPE for testing linear null hypothesis for the GLMs. In Section 4, the influence function of the MRPE as well as the influence functions of the Wald-type tests are derived. Finally, we empirically examine the performance of the proposed robust estimators and Wald-type test statistics for the Poisson regression model through a simulation study in Section 5, and we illustrate its applicability with real data sets for binomial and Poisson regression.
2. Asymptotic Distribution of the MRPEs for the GLMs
In this Section, we revise some of the results presented in Jaenada and Pardo [] in relation to the MRPE. Let be INIDO random variables with density functions with respect to some common dominating measure, respectively. The true densities are modeled by the density functions given in (1), belonging to the exponential family. Such densities are denoted by highlighting its dependence on the regression vector the nuisance parameter and the observation , In the following, we assume that the explanatory variables are fixed, and therefore the response variables verify the INIDO set up studied in Castilla et al. [].
For each of the response variables , the RP between the theoretical density function belonging to the exponential family, and the true density underlying the data, can be defined, for as
where
does not depend on
We consider a random sample of independent but nonhomogeneous observations of the response variables with fixed predictors Since only one observation of each variable is available, a natural estimate of its true density is the degenerate distribution at the the observation Consequently, in the following we denote the density function of the degenerate variable at the point Then, substituting in (2) the theoretical and empirical densities, yields to the loss
If we consider the limit when tends to zero we get
Last expression coincides with the Kullback–Leibler divergence, except for the constant More details about Kullback–Leiber divergence can be seen in Pardo [].
Based on the previous idea, we shall define an objective function averaging the RP between all the the RPs. Since minimizing in is equivalent to maximizing we define a loss function averaging those quantities as
Based on (5), we can define the MRPE of the unknown parameter , by
with defined in (5)
at The MRPE coincides with the MLE at , and therefore the proposed family can be considered a natural extension of the classical MLE.
Now, since the MRPE is defined as a maximum, it must annul the first derivatives of the loss function given in (5). The estimating equations of the parameters and are given by
For the first equation, we have
The previous partial derivatives can be simplified as
and
See Ghosh and Basu [] for more details. Now using the simplified expressions, we can write the estimating equation for as
being
and
Subsequently, the estimating equation for is given by
and thus, the estimating equation for is given by
being
and
Under some regularity conditions, Castilla et al. [] established the consistency and asymptotic normality of the MRPEs under the INIDO setup. Before stating the consistence and asymptotic distribution of the MRPEs for the GLM, let us introduce some useful notation. We define
for all and
Theorem 1.
Let be a random sample from the GLM defined in (1). The MRPE is consistent and its asymptotic distribution is given by
where denotes the design matrix, is the k-dimensional identity matrix and the matrices and are defined by
with
and
Proof.
The consistency is proved for general statistical models in Castilla et al. [] and the asymptotic distribution of the MRPEs for GLM is derived in Jaenada and Pardo []. □
3. Wald Type Tests for the GLMs
In this section, we define Wald-type tests for linear null hypothesis of the form
being a full rank matrix and
a r-dimensional vector . If the nuisance parameter is known, as with logistic and Poisson regression, the matrix Additionally, choosing
gives rise to a null hypothesis defined by a linear combination of the regression coefficients, , with known or unknown. Further, the simple null hypothesis is a particular case when choosing as the identity matrix of rank k,
with
In the following we assume that there exist a matrix verifying
Definition 1.
Let be the MRPE of for the GLM. The Wald-type tests, based on the MRPE, for testing (11) are defined by
The following theorem presents the asymptotic distribution of the Wald-type test statistics,
Theorem 2.
The Wald-type test follows asymptotically, under the null hypothesis presented in (11), a chi-square distribution with degrees of freedom equal to the dimension of the vector in (12)
Under the null hypothesis given in (11) the asymptotic distribution of the Wald-type test statistics is a chi-square distribution with r degrees of freedom.
Proof.
We know that
Therefore,
Now, the result follows taking into account that is a consistent estimator of □
Based on the previous convergence, the null hypothesis in (11) is rejected, if
being the percentile of a chi-square distribution with r degrees of freedom.
Finally, let be a parameter point verifying i.e., is not on the null hypothesis. The next result establishes that the Wald-type tests given in (14) are consistent (see Fraser []).
Theorem 3.
Let be a parameter point verifying Then the Wald-type tests given in (14) are consistent, i.e.,
Proof.
See Appendix A. □
Remark 1.
In the proof of the previous Theorem was established the approximate power function of the Wald-type tests defined in (13),
where
and
From the above expression, the necessary sample size n for the Wald-type tests to have a predetermined power, is given by , with
being
and the integer part.
In accordance with Maronna et al. [], the breakdown point of the estimators of a parameter is the largest amount of contamination that the data may contain such that still gives enough information about The derivation of a general breakdown points it is in general not easy, so it may deserve a separate paper where it may be jointly considered the replacement finite-sample breakdown point introduced by Donoho and Huber []. Although breakdown point is an important theoretical concept in robust statistics, perhaps is more useful the definition of breakdown point associated to a finite sample: replacement finite-sample break down point. More details can be seen in Section 3.2.5 of Maronna et al. [].
4. Influence Function
We derive in this section the IF of the MRPEs of the parameters and Wald-type statistics based on these MRPEs, The influence function (IF) of an estimator quantifies the impact of an infinitesimal perturbation in the true distribution of the data on the asymptotic value of the resulting parameter estimate (in terms of the corresponding statistical functional). An estimator is said to be robust if its IF is bounded. If we denote the true distributions underlying the data, the functional and associated to the MRPE for the parameters is such that
The IF of a estimator is defined as the limiting standardized bias due to infinitesimal contamination. That is, given a contaminated distribution at the point , with the degenerate distribution at , the IF of the estimator in terms of its associated functional is computed as
In the following, let us denote where and are the functionals associated the parameters and , respectively. Then, they must satisfy the estimating equations of the MRPE given by
where the quantities and are defined in Section 2. Now, evaluating the previous equation at the contaminated distribution , implicitly differentiating the estimating equations in and evaluating them at we can obtain the expression of the IF for the GLM.
We first derive the expression IF of MRPEs at the direction. For this purpose, we consider the contaminated distributions
with Here, only the -th component of the vector of distributions is contaminated. If the true density function of each variable belongs to the exponential model, we have that
Accordingly, we define
the MRPE when the true distribution underlying the data is Based on Remark 5.2 in Castilla et al. [] the IF of the MRPE at the direction with the point of contamination is given by
In a similar manner, the IF in all directions (i.e., all components of the vector of distributions are contaminated) has the following expression
with the point of contamination. We next derive the expression of the IF for the Wald-type tests presented in Section 3. The statistical functional associated with the Wald-type tests for the linear null hypothesis (11) at the distributions , ignoring the constant is given by
Again, evaluating the Wald-type test functionals at the contaminated distribution and implicitly differentiating the expression, we can get the expression of it IF. In particular, the IF of the Wald-type test statistics at the -th direction and the contamination point is given by
Evaluating the previous expression at the null hypothesis, the IF becomes identically zero,
Therefore, it is necessary to consider the second order IF of the proposed Wald-type tests. Twice differentiating in , we get
Finally, the second order IF of the Wald-type tests in all directions is given by
To asses the robustness of the MRPEs and Wald-type test statistics we must discuss the boundedness of the corresponding IF. The boundedness of the second order IF of the Wald-type test statistics is determined by the boundedness of the IF of the MRPEs. Further, the matrix is assumed to be bounded, so the robustness of the estimators only depend on the second factor of the IF. Most standard GLMs enjoy such properties for positives values of , but the influence function is unbounded at corresponding with the MLE. As an illustrative example, Figure 1 plots the IF of the MRPEs for the Poisson regression model with different values of at one direction. The model is fitted with only one covariate, the parameter is known for Poisson regression () and the true regression vector is fixed As shown, the IF of the MRPEs with positives values of are bounded, whereas the IF of the MLE is not, indicating it lack of robustness.

Figure 1.
IF of MRPEs with (left) and (right) of Poisson regression model.
5. Numerical Analysis: Poisson Regression Model
We illustrate the proposed robust method for the Poisson regression model. As pointed out in Section 1 the Poisson regression model belongs to the GLM with known shape parameter location parameter and known functions and . Since the nuisance parameter is known, for the seek of simplicity in the following we only use In Poisson regression, the mean of the response variable is linked to the linear predictor through the natural logarithm, i.e., Thus, we can apply the previous proposed method to estimate the vector of regression parameters with objective function given in Equation (5).
The results provided are computed in the software R. The minimization of the objective function is performed using the implemented optim() function, which applies the Nelder–Mead iterative algorithm (Nelder and Mead []). Nelder–Mead optimization algorithm is robust although relatively slow. The corresponding objective function given in (5) is highly nonlinear and requires the evaluation of nontrivial quantities. Further, the computation of the Wald-type test statistics defined in (13) requires to evaluate the covariance matrix of the MRPEs, involving nontrivial integrals. Some simplified expressions of the main quantities defined throughout the paper for the Poisson regression model, such as or are given in the Appendix B. There is no closed expression for these quantities, and they need to be approximated numerically. Since the minimization is iteratively performed, computing such expressions at each step of the algorithm and for each observation may entail an increased computational burden. Nonetheless, the complexity is not significant for low-dimensional data. On the other hand, the optimum in (5) need not to be uniquely defined, since the objective function may have several local minima. Then, the choice of the initial value of the iterative algorithm is crucial. Ideally, a good initial point should be consistent and robust. In our results the MLE is used as initial estimate for the algorithm.
We analyze the performance of the proposed methods in Poisson regression through a simulation study. We asses the behavior of the MRPE under the sparse Poisson regression model with covariates but only 3 significant variables. We set the 12-dimensional regression parameter and we generate the explanatory variables, , from the standard uniform distribution with variance-covariance matrix having Toeplitz structure, with the -th element being . The response variables are generated from the Poisson regression model with mean To evaluate the robustness of the proposed estimators, we contaminate the responses using a perturbed distribution of the form where b is a realization of a Bernoulli variable with parameter so called the contamination level. That is, the distribution of the contaminated responses lies in a small neighbourhood of the assumed model. We repeat the process for each value of .
Figure 2 presents the mean squared error of the estimate (MSE), (left) and the MSE on the prediction (right) against contamination level on data for different values of and . The sample size is fixed at and the MSE on the prediction is calculated using new observations following the true model. As shown, greater values of correspond to more robust estimators, revealing the role of the tuning parameter on the robustness gain. Most strikingly, the MSE grows linearly for the MLE, while the proposed estimators manage to maintain a low error in all contaminated scenarios.

Figure 2.
Mean Squared Error (MSE) on estimation (left) and prediction (right) against contamination level on data.
Furthermore, it is to be expected that the error of the estimate decreases with larger samples sizes. In this regard, Figure 3 shows the MSE for different values of and against the sample size in the absence of contamination (left) and under of contamination. Our proposed estimators are more robust than the classical MLE with almost all contaminated scenarios, since the MSE committed is lower for all positives values of than for (corresponding to the MLE), except for too small sample sizes. Conversely, the MLE is, as expected, the most efficient estimator in absence of contamination, closely to our proposed estimators with , highlighting the importance of in controlling the trade-off between efficiency and robustness. In this regard, values of about perform the best taking into account the low loss of efficiency and the gain in robustness. Finally, note that small sample sizes adversely affect to greater values of .

Figure 3.
MSE in estimation of in absence of contamination (left) and under of contamination level in data (right) with different values of against sample size for Poisson regression model.
On the other hand, one could be interested on testing the significance of the selected variables. For this purpose, we simplify the true model and we examine the performance of the proposed Wald-type test statistics under different true coefficients values. In particular, let us consider a Poisson regression model with only two covariates, generated from the uniform distribution as before, and the linear null hypothesis
That is, we are interested in assessing the significance of the second variable. The sample size if fixed at and the true value of the component of the regression vector is set We study the power of the tests under increasing signal of the second parameter and increasing contamination level. Here, the model is contaminated by perturbing the true distribution with where is the mean of the Poisson variable in the absence of contamination, is the contaminated mean, with and b is a realization of a Bernoulli variable with probability of success Table 1 presents the rejection rate of the Wald-type test statistics for different true values of under different contaminated scenarios. As expected, stronger signals produce higher power for all Wald-type test. Moreover, the power of the Wald-type test statistics based on the MLE decreases when increasing the contamination, whereas the power of the statistics based on the MRPEs with positives values of keeps sufficiently high. Then, our proposed robust estimators are able to detect the significance of the variable even in heavily contaminated scenarios.

Table 1.
Rejection rate of Wald-type test statistics based on MRPEs with different true values of and contamination levels.
6. Real Data Applications
6.1. Example I: Poisson Regression Regression
We finally apply our proposed estimators in a real dataset arising from Crohn’s disease. The data were first studied in Lô and Ronchetti [] to asses the adverse events of a drug. The clinical study included 117 patients affected by the disease, for whom information was recorded for 7 explanatory variables: BMI (body mass index), HEIGHT, COUNTRY (one of the two countries where the patient lives), SEX, AGE, WEIGHT, and TREAT (the drug taken by the patient in factor form: placebo, Dose 1, Dose 2), in addition to the response variable AE (number of adverse events). Lô and Ronchetti [] considered a Poisson regression model for the Crohn data and determined that only variables Dose 1, BMI, HEIGHT, SEX, AGE, and COUNTRY may be essentially significant. Further, they flagged observations 23rd, 49th, and 51st to be highly influential on the classical analysis. Table 2 presents the estimated coefficient of the explanatory variable when fitting the Poisson regression model. Robust methods suggest higher coefficients for the variables BMI and AGE, whereas fewer values for the coefficients of the categorical variables COUNTRY, SEX, Dose 1.

Table 2.
Estimated coefficients for Crohn’s disease data for different values of with original data and clean data (after removing influential observations).
Following the discussion in Lô and Ronchetti [], classical tests may not select variable AGE to be significant. Then, we propose testing the significance of that variable using Wald-type test statics based on different values . Table 3 shows the p-values of the corresponding tests with null hypothesis : AGE = 0, with the original data and after removing the outlying observations.

Table 3.
p-values of test with null hypothesis : AGE = 0 with original and clean data (after removing influential observations).
The MLE rejects the significance of the variable AGE when the original data are used, whereas the Wald-type test statistics with positives values of indicate strong evidence against the null hypothesis. In contrast, if the influential observations are removed, all Wald-type test statistics agree in the significance of the variable. This example illustrates the robustness of the proposed statistics.
6.2. Example II: Binomial Regression
We finally illustrate the applicability of the MRPE for robust inference in the binomial regression model. We examine the damaged carrots dataset, first studied in Phelps [] and later discussed by Cantoni and Ronchetti [] and Ghosh and Basu [] to illustrate robust procedures for binomial regression. The data contain 24 samples, among which the 14th observation was flagged as an outlier in the y-space but not a leverage point. The data are issued from a soil experiment and give the proportion of carrots showing insect damage in a trial with three blocks and eight dose levels of insecticide. The explanatory variables are the logarithm transform of the dose (Logdose) and two dummy variables for Blocks 1 and 2.
Binomial regression is a natural extension of the logistic regression when the response variable Y does not follow a Bernoulli distribution but a Binomial distribution counting the number of successes in a series of m independent Bernoulli trials. Binomial regression model belongs to the GLM with known shape parameter location parameter and functions and . The mean of the response variable is then linked to the linear predictor through the logit function, i.e.,
Table 4 presents the estimated coefficients of the regression vector for the carrots data using the MLE and robust MRPEs when the model is fitted with the original data and the model fitted without the outlying observation. The results provided are computed in the same manner as in Section 5, adapting the corresponding quantities in Equation (5) for the binomial model. All integrals involved were numerically approximated, and the MLE is used as initial estimate for the optimization algorithm. The influence of observation 14 stands out when using the MLE; the estimated coefficients are remarkably different when fitting the model with and without observation 14. In contrast, all methods estimate similar coefficients after removing the outlying observation, coinciding with the robust estimates for moderately high values of the tuning parameter .

Table 4.
Estimated coefficients for damaged carrots data for different values of with original data and clean data (after outliers removal).
7. Conclussions
In this paper, we presented the MRPE and Wald-type test statistics for GLMs. The proposed MRPEs and statistics have appealing robustness properties where the data are contaminated due to outliers or leverage points. MRPEs are consistent and asymptotically normal and represent an attractive alternative to the classical nonrobust methods. Additionally, robust Wald-type test statistics, based on the MRPEs, were developed. Through the study of the IFs and the development of an extensive simulation study, we proved their robustness from a theoretical and practical point of view, respectively. In particular, we illustrated the superior performance of the MRPEs and the corresponding Wald-type tests for the Poisson regression model.
Author Contributions
Conceptualization, M.J. and L.P.; methodology, M.J. and L.P.; software, M.J. and L.P.; validation, M.J. and L.P.; formal analysis, M.J. and L.P.; investigation, M.J. and L.P.; resources, M.J. and L.P.; data curation, M.J. and L.P.; writing—original draft preparation, M.J. and L.P.; writing—review and editing, M.J. and L.P.; visualization, M.J. and L.P.; supervision, M.J. and L.P.; project administration, M.J. and L.P.; funding acquisition, M.J. and L.P. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Spanish Grants PGC2018-095194-B-100 (L. Pardo and M. Jaenada) and FPU/018240 (M. Jaenada).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The real datasets are publicly available on the R package robustbase in CRAN under the names of CrohnD (Poisson regression example) and carrots (binomial regression example).
Acknowledgments
We are very grateful to the referees and associate editor for their helpful comments and suggestions. This research is supported by the Spanish Grants PGC2018-095194-B-100 (L. Pardo and M. Jaenada) and FPU/018240 (M. Jaenada). M.Jaenada and L. Pardo are members of the Instituto de Matematica Interdisciplinar, Complutense University of Madrid.
Conflicts of Interest
The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.
Abbreviations
The following abbreviations are used in this manuscript:
DPD | Density Power Divergence |
IF | Influence Function |
GLM | Genelarized Linear Model |
LRM | Linear Regression Model |
MLE | Maximum Likelihood Estimator |
MRPE | Minimum Rényi Pseudodistance Estimator |
RP | Rényi Pseudodistance |
Appendix A. Proof of Theorem 3
Let us define
so the Wald-type test statistic is such that
We know that and therefore and have the same asymptotic distribution. A first order Taylor expansion of at around gives,
Based on the asymptotic distribution of we have
therefore
have asymptotically the same distribution, i.e.,
Now, we shall denote,
Then, we have,
where represents the distribution function of a standard normal distribution evaluated at Finally,
Appendix B. Poisson Regression Model
We derive here some explicit expression for the particular case of the Poisson regression. Following the discussion in Section 5, we denote here since the nuisance parameter is known, The Poisson distribution with parameter is given by
Differentiating its logarithm with respect to the regression vector, we get
so we can write
Further, we have that
so the estimating equations of the Poisson regression model are given by
For we have
so the estimating equations are given by
yielding to the maximum likelihood estimating equations.
On the other hand, the asymptotic distribution of is given by
being
with
and
Finally
References
- Nelder, J.A.; Wedderburn, R.W.M. Generalized linear models. J. R. Stat. Soc. 1972, 135, 370–384. [Google Scholar] [CrossRef]
- McCullagh, P.; Nelder, J.A. Generalized Linear Models. In Monographs on Statistics and Applied Probability; Chapman and Hall: London, UK, 1983. [Google Scholar]
- Jaenada, M.; Pardo, L. The minimum Renyi’s Pseudodistances estimators for Generalized Linear Models. In Data Analysis and Related Applications: Theory and Practice; Proceeding of the ASMDA; Wiley: Athens, Greece, 2021. [Google Scholar]
- Stefanski, L.A.; Carroll, R.J.; Ruppert, D. Optimally bounded score functions for generalized linear models with applications to logistic regression. Biometrika 1986, 73, 413–424. [Google Scholar] [CrossRef]
- Krasker, W.S.; Welsch, R.E. Efficient bounded-influence regression estimation. J. Am. Stat. Assoc. 1982, 77, 595–604. [Google Scholar] [CrossRef]
- Künsch, H.R.; Stefanski, L.A.; Carroll, R.J. Conditionally unbiased bounded-influence estimation in general regression models, with applications to generalized linear models. J. Am. Stat. Assoc. 1989, 84, 460–466. [Google Scholar]
- Morgenthaler, S. Least-absolute-deviations fits for generalized linear models. Biometrika 1992, 79, 747–754. [Google Scholar] [CrossRef]
- Cantoni, E.; Ronchetti, E. Robust inference for generalized linear models. J. Am. Stat. Assoc. 2001, 96, 1022–1030. [Google Scholar] [CrossRef]
- Bianco, A.M.; Yohai, V.J. Robust estimation in the logistic regression model. In Robust Statistics, Data Analysis, and Computer Intensive Methods; Springer: New York, NY, USA, 1996; pp. 17–34. [Google Scholar]
- Croux, C.; Haesbroeck, G. Implementing the Bianco and Yohai estimator for logistic regression. Comput. Stat. Data Anal. 2003, 44, 273–295. [Google Scholar] [CrossRef]
- Bianco, A.M.; Boent, G.; Rodrigues, I.M. Robust tests in generalized linear models with missing responses. Comput. Stat. Data Anal. 2013, 65, 80–97. [Google Scholar] [CrossRef]
- Valdora, M.; Yohai, V.J. Robust estimators for generalized linear models. J. Stat. Plan. Inference 2014, 146, 31–48. [Google Scholar] [CrossRef]
- Ghosh, A.; Basu, A. Robust estimation in generalized linear models: The density power divergence approach. Test 2016, 25, 269–290. [Google Scholar] [CrossRef]
- Basu, A.; Harris, I.R.; Hjort, N.L.; Jones, M.C. Robust and efficient estimation by minimising a density power divergence. Biometrika 1998, 85, 549–559. [Google Scholar] [CrossRef]
- Basu, A.; Ghosh, A.; Mandal, A.; Martin, N.; Pardo, L. Robust Wald-type tests in GLM with random design based on minimum density power divergence estimators. Stat. Method Appl. 2021, 3, 933–1005. [Google Scholar] [CrossRef]
- Broniatowski, M.; Toma, A.; Vajda, I. Decomposable pseudodistances and applications in statistical estimation. J. Stat. Plan. Inference 2012, 142, 2574–2585. [Google Scholar] [CrossRef][Green Version]
- Castilla, E.; Martín, N.; Muñoz, S.; Pardo, L. Robust Wald-type tests based on Minimum Rényi Pseudodistance Estimators for the Multiple Regression Model. J. Stat. Comput. Simul. 2020, 14, 2592–2613. [Google Scholar] [CrossRef]
- Toma, A.; Leoni-Aubin, S. Optimal robust M-estimators using Rényi pseudodistances. J. Multivar. Anal. 2013, 115, 259–273. [Google Scholar] [CrossRef]
- Toma, A.; Karagrigoriou, A.; Trentou, P. Robust model selection criteria based on pseudodistances. Entropy 2020, 22, 304. [Google Scholar] [CrossRef] [PubMed]
- Rényi, A. On measures of entropy and information. In Proceeding of the 4th Symposium on Probability and Statistics; University of California Press: Berkely, CA, USA, 1961; pp. 547–561. [Google Scholar]
- Jones, M.C.; Hjort, N.L.; Harris, I.R.; Basu, A. A comparison of related density-based minimum divergence estimators. Biometrika 2001, 88, 865–873. [Google Scholar] [CrossRef]
- Fujisawa, H.; Eguchi, S. Robust parameter estimation with a small bias against heavy contamination. J. Multivar. Anal. 2008, 99, 2053–2081. [Google Scholar] [CrossRef]
- Hirose, K.; Masuda, H. Robust relative error estimation. Entropy 2018, 20, 632. [Google Scholar] [CrossRef]
- Kawashima, T.; Fujisawa, H. Robust and sparse regression via γ-divergence. Entropy 2017, 19, 608. [Google Scholar] [CrossRef]
- Kawashima, T.; Fujisawa, H. Robust and sparse regression in generalized linear model by stochastic optimization. Jpn. J. Stat. Data Sci. 2019, 2, 465–489. [Google Scholar] [CrossRef]
- Windham, M.P. Robustifying model fitting. J. R. Stat. Soc. Ser. B 1995, 57, 599–609. [Google Scholar] [CrossRef]
- Castilla, E.; Jaenada, M.; Pardo, L. Estimation and testing on independent not identically distributed observations based on Rényi’s pseudodistances. arXiv 2021, arXiv:2102.12282. [Google Scholar]
- Pardo, L. Statistical Inference Based on Divergence Measures; Chapman and Hall/CRC: Boca Raton, FL, USA, 2018. [Google Scholar]
- Fraser, D.A.S. Non parametric Methods in Statistics; John Wiley & Sons: New York, NY, USA, 1957. [Google Scholar]
- Maronna, R.A.; Martin, R.D.; Yohai, V.J. Robust Statistics Theory and Methods; John Wiley & Sons. Inc.: Hoboken, NJ, USA, 2006. [Google Scholar]
- Donoho, D.L.; Huber, P.J. The notion of breakdown point. In A Festschrift for Erich L. Lehmann; CRC Press: Boca Raton, FL, USA, 1983. [Google Scholar]
- Nelder, J.A.; Mead, R. A simplex method for function minimization. Comput. J. 1965, 7, 308–313. [Google Scholar] [CrossRef]
- Lô, S.N.; Ronchetti, E. Robust and accurate inference for generalized linear models. J. Multivar. Anal. 2009, 100, 2126–2136. [Google Scholar] [CrossRef]
- Phelps, K. Use of the Complementary Log-Log Function to Describe Dose Response Relationships in Insecticide Evaluation Field Trials. In Lecture Notes in Statistics, No. 14.: Proceedings of the International Conference on Generalized Linear Models; Gilchrist, R., Ed.; Springer: Berlin, Germany, 1982. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).