SUMMARY
Robust statistical methods, such as M-estimators, are needed for nonlinear regression models because of the presence of outliers/influential observations and heteroscedasticity. Outliers and influential observations are commonly observed in many applications, especially in toxicology and agricultural experiments. For example, dose response studies, which are routinely conducted in toxicology and agriculture, sometimes result in potential outliers, especially in the high dose groups. This is because response to high doses often varies among experimental units (e.g., animals). Consequently, this may result in outliers (i.e., very low values) in that group. Unlike the linear models, in nonlinear models the outliers not only impact the point estimates of the model parameters but can also severely impact the estimate of the information matrix. Note that, the information matrix in a nonlinear model is a function of the model parameters. This is not the case in linear models. In addition to outliers, heteroscedasticity is a major concern when dealing with nonlinear models. Ignoring heteroscedasticity may lead to inaccurate coverage probabilities and Type I error rates. Robustness to outliers/influential observations and to heteroscedasticity is even more important when dealing with thousands of nonlinear regression models in quantitative high throughput screening assays. Recently, these issues have been studied very extensively in the literature (references are provided in this paper), where the proposed estimator is robust to outliers/influential observations as well as to heteroscedasticity. The focus of this paper is to provide the theoretical underpinnings of robust procedures developed recently.
Keywords: asymptotic linearity, heteroscedasticity, M-estimation procedure, nonlinear regression model
1 Introduction
Nonlinear regression models are widely used in many applications such as in dose response studies conducted in agricultural sciences, toxicology and other biological sciences. One of the reasons to consider nonlinear regression models is that in many applications it is not reasonable to assume that the rate of change in mean response E(y), where y is the response variable of interest, is constant with respect to the explanatory variables such as dose (x). For example, in toxicology it is common to assume, due to saturation in the mean response, the rate of change in E(y) may asymptotically go to zero as x → ∞. By nonlinear regression models we mean that the parameters of the model appear nonlinearly. Alternatively, one could consider models where the parameters appear linearly but the model is nonlinear in x, such as high order polynomials in x. An advantage of such models is that the parameters arise in the model linearly and hence are easy to estimate and one can even perform exact statistical inferences (under suitable assumptions). However, disadvantages outweigh the advantages of using such polynomial regression models. Often, unless very high degree polynomial is used, it is not always feasible to capture the underlying scientific phenomena using a polynomial curve. As the degree of the polynomial increases the power of statistical inference will decrease. Lastly, in toxicology and in other sciences, researchers are often interested in estimating dose corresponding to desired response, such as ED50. If one were to use standard polynomial regression models then such parameters are often nonlinear functions of the model parameters (e.g. ratio of regression parameters) which can potentially lead to Fieller’s problem which may not always have a satisfactory solution.
For the above reasons nonlinear regression models, which are not necessarily polynomials in dose, are widely used in applications [1, 7–10, 14–16]. More precisely, a model such as the following is widely used
| (1) |
where yi = (yi1,…, yi,ni)T are ni × 1 response vectors, xi are m × 1 vectors of known regression constants (e.g., dose), θ = (θ1, …, θp)T is a vector of unknown parameter, f is a pre-specified nonlinear function of θ and εi are independent and identically distributed random vectors with N(0, I). The total sample size n is given by .
Throughout this paper we assume that σi ≡ σ (zi,τ) where zi are q × 1 vectors of known constants (e.g., dose), τ = (τ1, …, τq)T is a vector of unknown parameters and τ is a pre-specified function of τ. Thus in the case of homoscedasticity σ (zi, τ) is constant for all zi and is not constant when the data are heteroscedastic. In toxicological studies, the variability in response y may sometimes depend upon the mean response, which is likely to be monotonic in dose. In practice, however, one never knows a priori if the data are homoscedastic or heteroscedastic. Consequently, one never knows, a priori, whether to use ordinary least squares (OLS) based methodology or to use iterated weighted least squares (IWLS) based methodology. Adding to the complexity, it is not uncommon to find outliers and/or influential observations in these data. It is well known that the efficiency and the accuracy of OLS and IWLS depend upon the underlying variance structure [7–10] and potential outliers and/or influential observations in the data. For these reasons, [9, 10] developed a preliminary test (PT) based methodology that is robust to the underlying variance structure and For the method to be robust against potential outliers and/or influential observations, they used M-estimation based methodology. PTE chooses either the ordinary M-estimator (OME) or the weighted M-estimator (WME) based on the results of a preliminary test for heteroscedasticity.
In Section 2 we define three robust estimators (OME, WME and PTE) and their asymptotic properties. In Section 3 we illustrate the methodology described in [9, 10] using a data set obtained from NIEHS quantitative high throughput screening (qHTS) assays. The asymptotic theory developed and used in [9, 10] require the asymptotic linearity of WME, which is demonstrated in Section 4 of this paper. For clarity of exposition, we relegated all regularity conditions and lemmas to the Appendix of the paper.
2 Estimators
In this section we define various estimators for estimating parameters in the model (1) and their asymptotic covariance matrices. Note that some of the definitions of the matrices and related quantities are provided in the Appendix.
The ordinary M-estimator (OME) for θ is defined as the solution of the following minimization problem:
where h is a suitable function such as Huber-score function. Asymptotic covariance matrix of OME is expressed as follows:
where
, γ4 = Eψ′ (σ(z,τ)ε), fθ(xi,θ) = (∂/∂θ) f (xi,θ), ψ(u) = (∂/∂u)h2(u), and ψ′(u) = (∂/∂u)ψ (u).
In [9], we defined WME of the unknown parameter vector, (θT,τT)T as follows:
| (2) |
The main theory in this paper together with the methodology described in [9] legitimizes the following asymptotic distribution of WME:
where
στ(zi, τ) = (∂/∂τ)σ(zi, τ), and k(zi, τ) = 1/σ(zi, τ).
To keep the model parsimonious, as in [9, 10], throughout this paper we model the standard deviation using the following log-linear model:
We take zi ≡ (1, xi)T where xi is the ith dose. We derive the preliminary test estimator (PTE) as was done in [9, 10] as follows. We first fit the nonlinear model (1) under homoscedasticity using OME. Let ε̃i = yi − f(xi, θ̃). Using these residuals we fit the following log-linear model using standard OLS methodology to estimate τ0 and τ1:
The PTE is then defined as
where Tn = τ̂1n/ √Var (τ̂1n, τ̂1n is the least squares estimator of τ1, tα,n−2 is the critical value of the t-distribution with n − 2 degrees of freedom having probability 1 − α and α is the significance level of the preliminary test.
The following asymptotic covariance matrix for PTE is derived in [9]:
where Ft is the cdf of the t-distribution with n − 2 degrees of freedom. Note that the asymptotic covariance matrix of PTE is a weighted average of the corresponding matrices for OME and WME.
Remark 2.1. From [9, 10] we now have the following confidence intervals:
- The 100(1 − α)% confidence interval for parameters θi, i = 1, …, p, of a nonlinear model (1) using PTE method can be computed using the following formula [10]:
The confidence limits for PTE use the maximum of the standard errors of OME and WME in their construction rather than a standard error based on the weighted average mentioned earlier. PTE can be expressed as a weighted average of OME and WME which both are asymptotically normally distributed. Therefore, the asymptotic distribution of PTE is a mixture of normal distributions and hence the usual confidence limits using a critical value from a t-distribution and the standard error of PTE may not be appropriate. That is the reason why we used the maximum of the standard errors of OME and WME, which is derived in [10].
3 Illustration
For illustration purposes we consider dose-response data for two compounds from the US National Toxicology Program (NTP) library of 1408 compounds that were evaluated using the quantitative high throughput screening (qHTS) assay. The raw data are provided in the Appendix and they are plotted against dose in Figures 1 (a) and (b), respectively. A visual inspection of Figure 1 (a) suggests that the data are perhaps homoscedastic, whereas Figure 1 (b) suggests potential heteroscedasticity. Indeed the Bartlett’s test for homogeneity yields p-values of 0.817 and 0.066 for Examples 1 and 2, respectively. As commonly done by toxicologists, we fitted the following nonlinear function, called the Hill function, for the two data sets:
Figure 1.
US NTP qHTS data sets; the corresponding fitted curves using OME (solid line) and WME (dashed line) methods.
We computed OME, WME and PTE (described in [9]) using Huber function with γ = 1.5. Note that PTE chooses either OME or WME based on the result of the preliminary test [9] and hence the fitted curve using PTE is the same as one of the two fitted curves using OME and WME. We computed the 95% confidence intervals for OME, WME and PTE using the formulas provided in the previous section. Results are summarized in Table 1. As expected, the performance of OME and WME depends upon whether the data are homoscedastic or heteroscedastic. For homoscedastic data even though the point estimates of OME and WME are quite similar, the standard errors of WME are almost two times larger than those of OME. On the other hand, for heteroscedastic data OME for θ2 and the standard errors of OME for θ2 and θ3 are very large. Consequently, the confidence intervals are also very highly affected by the choice of the method. The PTE provides the desired compromise between the two methods.
Table 1.
Estimate, standard error and 95% confidence interval for parameters of the models for US NTP qHTS data using OME, WME and PTE methods.
| Parameter | Method | Estimate | SE | 95% Con dence Interval | |||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Example 1 | θ0 | OME | −36.7 | 4.7 | (−47.7, −25.8) | ||||||
| WME | −37.7 | 8.0 | (−56.5, −18.9) | ||||||||
| PTE | −36.7 | 6.6 | (−55.6, −17.9) | ||||||||
| θ1 | OME | 34.5 | 4.9 | (23.1, 45.9) | |||||||
| WME | 35.3 | 8.3 | (16.0, 54.6) | ||||||||
| PTE | 34.5 | 6.8 | (15.2, 53.8) | ||||||||
| θ2 | OME | 3.1 | 1.3 | (0.1, 6.1) | |||||||
| WME | 2.6 | 1.4 | (−0.6, 5.9) | ||||||||
| PTE | 3.1 | 1.3 | (−0.2, 6.3) | ||||||||
| θ3 | OME | 28.9 | 5.7 | (15.7, 42.2) | |||||||
| WME | 29.9 | 9.1 | (8.5, 51.2) | ||||||||
| PTE | 28.9 | 7.6 | (7.5, 50.3) | ||||||||
| Example 2 | θ0 | OME | −45.7 | 2.8 | (−52.3, −39.1) | ||||||
| WME | −45.1 | 6.1 | (−59.3, −30.9) | ||||||||
| PTE | −45.1 | 6.0 | (−59.3, −30.9) | ||||||||
| θ1 | OME | 46.7 | 3.2 | (39.4, 54.1) | |||||||
| WME | 46.4 | 6.2 | (31.9, 61.0) | ||||||||
| PTE | 46.4 | 6.2 | (31.9, 61.0) | ||||||||
| θ2 | OME | 8.4 | 48.1 | (−103.9, 120.8) | |||||||
| WME | 4.2 | 2.2 | (−1.0, 9.4) | ||||||||
| PTE | 4.2 | 9.0 | (−108.4, 116.8) | ||||||||
| θ3 | OME | 18.9 | 8.9 | (−2.0, 39.8) | |||||||
| WME | 17.9 | 2.4 | (12.3, 23.4) | ||||||||
| PTE | 17.9 | 2.8 | (−3.0, 38.8) | ||||||||
4 Asymptotic linearity of weighted M-estimator (WME)
In this section we investigate the asymptotic linearity of WME (2) for the heteroscedastic nonlinear regression model (1). M-estimation methods have been well studied in the literature over the several decades for linear models (cf. [3, 4, 12]). Huber [2] proposed the M-estimator of the regression parameters in the univariate linear model and showed that under certain regularity conditions the M-estimator is consistent and asymptotically normal. The asymptotic theory for the M-estimator has been extensively studied (see, e.g., [2, 6, 17, 24]); for multivariate linear models one may refer to [5, 11, 13, 22, 23] among others. In the context of nonlinear regression models, Sanhueza [18] and Sanhueza and Sen [19, 20] proposed M-estimators and studied their asymptotic properties. More recently, Sanhueza et al. [21] extended these methods to nonlinear models for repeated measures data.
Asymptotic linearity of M-estimators is the first and important step to derive the asymptotic normality of the M-estimators in nonlinear regression models ([18–21]). Because we defined an M-estimator of not only the regression parameter θ, but also the variance parameter τ for heteroscedastic nonlinear regression model, the derivation of asymptotic linearity of the M-estimator is nontrivial.
The estimating equation for the minimization in (2) is given by
| (3) |
where
| (4) |
We explain the derivation of the uniform asymptotic linearity of WME very briey as follows: First we apply the Taylor’s expansion to λ(xi,yij, θ, τ) and decompose the first order term to several terms. After that we take expectation on the terms which are going to zero by the regularity conditions.
Then under the regularity conditions and lemmas provided in the Appendix, we now state and prove the uniform asymptotic linearity of WME defined in (2).
Theorem 4.1. Under the regularity conditions (A), (B) and (C) described in the Appendix of this paper, for any given C1, C2 > 0,
| (5) |
as n → ∞, where λ(xi, yi, θ, τ) was defined in (4).
Proof. We consider the lth element of the vector λ(xi, yi, θ, τ) denoted for
Using the first order term in the Taylor’s expansion, we have for 0 < u, υ < 1,
where for j = 1,… p,
and for j = 1, …, q,
Then, we have
-
for l = 1,…, p,Using Lemma 6.1 through 6.4 provided in the Appendix we deduce that for l = 1, …, p
(6) -
for l = p + 1, …, p + q,
5 Concluding remarks
We describe in this paper three recently published robust estimators for nonlinear regression models. All three estimators are robust to outliers and influential observations. OME is developed for homoscedastic data while WME is for heteroscedastic data. PTE is robust to error variance structure by choosing either OME or WME based on the result of a preliminary test for heteroscedasticity. PTE is an important estimator because a priori one may not know whether data are homoscedastic or heteroscedastic. Especially when there are number of data sets as in the case of qHTS assay PTE may be useful because of its robustness to outliers/influential observations as well as to error variance structure.
We illustrate these estimators by considering two real data sets from qHTS assay. One data set is potentially homoscedastic and the other is heteroscedastic. We fitted the Hill function for the data sets using the three estimators and observed that the performance and confidence intervals are affected by the choice of method.
We derive the uniform asymptotic linearity of WME which is theoretically important for developing the asymptotic normality of WME and the asymptotic covariance matrix of PTE. Because WME is estimating regression parameters as well as variance parameters simultaneously the derivation is quite complicated. However, under the suitably established regularity conditions and lemmas we prove the asymptotic linearity.
Optimal designs in the context of linear and nonlinear models are well studied in the literature. Yet, as observed in [10], very little has been done in the context of qHTS assays where a researcher is not only interested in performing thousands of statistical tests but is also interested in estimating parameters of the nonlinear model corresponding to each selected compound. Often it is a challenge to estimate the parameters of the nonlinear model precisely.
To illustrate this point, we considered dose-response data for a compound (Example 3 in Appendix) from NTP’s 1408 library of compounds. We fitted the Hill model for these data (see Figure 2) using both OME and WME. Visually, both OME as well as WME seem to fit the data well (especially OME), however, as seen from Table 2, the standard errors of all estimates appear to be so large that all confidence intervals contain zero. There could be several reasons for this to happen. For example, the total sample size may not be large enough for the amount of variability present in the data (especially at the high dose). A second reason could be dose-spacing. Although the chemical is tested at a large number of low doses, the chemical was not tested at higher doses containing the point of inflection. In fact there is no dose between 45 and 90 (µM) where the curve begins to drop. As consequence, it is not surprising that OME does not provide precise estimates for θ2, the slope parameter (standard error of 409.9), or θ3, the ED50 (standard error of 264.1). As noted in [10], with the advent of qHTS assays, optimal design for nonlinear models is an important area of future research.
Figure 2.
A US NTP qHTS data set; the corresponding fitted curves using OME (solid line) and WME (dashed line) methods.
Table 2.
Estimate, standard Error and 95% confidence interval for parameters of the models for US NTP qHTS data using OME, WME and PTE methods.
| Parameter | Method | Estimate | SE | 95% Con dence Interval |
|---|---|---|---|---|
| θ0 | OME | −76.1 | 77.1 | (−256.0, 103.7) |
| WME | −78.0 | 83.3 | (−272.8, 116.7) | |
| PTE | −78.0 | 82.0 | (−272.8, 116.7) | |
| θ1 | OME | 72.9 | 77.3 | (−107.5, 253.3) |
| WME | 74.8 | 83.5 | (−120.6, 270.2) | |
| PTE | 74.8 | 82.3 | (−120.6, 270.2) | |
| θ2 | OME | 9.7 | 409.9 | (−946.9, 966.4) |
| WME | 4.2 | 7.1 | (−12.3, 20.7) | |
| PTE | 4.2 | 187.3 | (−954.7, 963.0) | |
| θ3 | OME | 51.6 | 264.1 | (−564.6, 667.9) |
| WME | 63.4 | 56.2 | (−68.0, 194.7) | |
| PTE | 63.4 | 130.5 | (−554.3, 681.0) |
Table 3.
Data corresponding to three examples described in this paper. In each example we provide responses corresponding to the three replicates for each dose.
| Dose (µM) | Example 1 | Example 2 | Example 3 | ||||||
|---|---|---|---|---|---|---|---|---|---|
| 0.0005874 | 7.327 | −6.043 | −0.428 | −1.733 | −5.901 | 8.032 | −0.428 | −1.733 | −5.901 |
| 0.002937 | 3.110 | −2.340 | 3.103 | −6.491 | 1.196 | −1.665 | 3.103 | −6.491 | 1.196 |
| 0.01468 | −11.981 | −4.885 | −0.566 | −9.442 | −11.521 | 5.464 | −0.566 | −9.442 | −11.521 |
| 0.03283 | −4.475 | −5.747 | −1.660 | −5.877 | −0.634 | 9.618 | −1.660 | −5.877 | −0.634 |
| 0.07341 | −4.922 | −4.183 | −0.981 | −1.986 | −5.771 | −0.509 | −0.981 | −1.986 | −5.771 |
| 0.1642 | 6.090 | −10.287 | −1.919 | −10.696 | −0.011 | 7.133 | −1.919 | −10.696 | −0.011 |
| 0.3670 | −11.767 | −5.345 | −7.134 | −8.302 | −1.625 | −12.242 | −7.134 | −8.302 | −1.625 |
| 0.8207 | −3.393 | −2.099 | −4.490 | −4.888 | 0.283 | −0.770 | −4.490 | −4.888 | 0.283 |
| 1.835 | −6.812 | −5.808 | −9.666 | −4.959 | 1.045 | 8.512 | −9.666 | −4.959 | 1.045 |
| 4.103 | −0.600 | −5.707 | −4.203 | −3.499 | 5.638 | −5.215 | −4.203 | −3.499 | 5.638 |
| 9.175 | 4.313 | −8.110 | 0.388 | −5.977 | −2.008 | −9.651 | 0.388 | −5.977 | −2.008 |
| 20.52 | −10.838 | −22.657 | 3.820 | −3.014 | −8.214 | −39.837 | 3.820 | −3.014 | −8.214 |
| 45.87 | −34.218 | −30.895 | −23.929 | −24.587 | −8.603 | −44.239 | −23.929 | −24.587 | −8.603 |
| 91.74 | −30.911 | −44.233 | −36.776 | −76.046 | −83.195 | −23.337 | −36.776 | −76.046 | −83.195 |
Acknowledgement
This research was supported, in part, by the Intramural Research Program of the NIH, National Institute of Environmental Health Sciences [Z01 ES101744-04]. We thank Drs Kissling and Umbach for several helpful comments that led to substantial improvement in the content and presentation of the manuscript.
6 Appendix
6.1 The data
We provide the three data sets used in the Illustration.
6.2 Regularity conditions and proofs
We shall make the following sets of regularity assumptions concerning (A) the score function ψ, (B) the function f, and (C) the function σ. These are a generalization of the assumption established by [18].
[A1]:
ψ is a nonconstant, odd function which is absolutely continuous and differentiable with respect to θ.
[A2]: Let ε = {y − f(x, θ)}/σ(z, τ),
-
Eψ(ε) = 0; ; E{ψ(ε)ε} = γ1(≠ 0);
-
E|ψ′(ε)|1+δ < ∞, E|ψ′(ε)ε|1+δ < ∞, E|ψ′(ε)ε2|1+δ < ∞ for some 0 < δ ≤ 1, and
Eψ′(ε) = γ2 (≠ 0); E{ψ′(ε)ε} = 0; E{ψ′(ε)ε2} = γ3 (≠ 0);
Eψ′(σ(z, τ)ε) = γ4 (≠ 0)
[A3]: Let ε(θ, τ) = {y − f(x, θ)}/σ(z, τ),
[B1]:
f(x, θ) is continuous and twice differentiable with respect to θ ∈ ℜp.
[B2]:
limn→∞ n−1Γ1n(θ, τ) = Γ1(θ, τ), where
[B3]: For j, l = 1, …, p
limδ→0 sup‖Δ‖≤δ |(∂/∂θj) f (x, θ + Δ) (∂/∂θl) f (x, θ + Δ) − (∂/∂θj) f (x, θ) (∂/∂θl) f (x, θ)| = 0
limδ→0 sup‖Δ‖≤δ |(∂2/∂θj∂θl) f (x, θ + Δ) − (∂2/∂θj∂θl) f (x, θ)| = 0
[C1]:
σ(z, τ) is continuous and twice differentiable with respect to τ ∈ ℜq.
[C2]:
limn→∞ n−1Γ2n(θ, τ) = Γ2(θ, τ), where
and Στ (zi, τ) = (∂2/∂τ∂τT)σ(zi, τ).
[C3]: For j, l = 1, …, q
limδ→0 sup‖Δ‖≤δ |(∂/∂τj)σ(z, τ + Δ) (∂/∂τl)σ(z, τ + Δ) − (∂/∂τj)σ(z, τ) (∂/∂τl)σ(z, τ)| = 0
limδ→0 sup‖Δ‖≤δ |(∂2/∂τj∂τl)σ(z, τ + Δ) − (∂2/∂τj∂τl)σ(z, τ)| = 0
In order to prove the uniform asymptotic linearity of WME (θ̂T, τ̂T)T in (3), we need to consider the following series of lemmas. When proving the main theorem we decompose the first order term of the Taylor’s expansion to λ(xi, yi, θ, τ), into several sub-terms which are partial derivatives of λ(xi, yi, θ, τ) with respect to θ and τ. By these lemmas we put bounds on quantities about such partial derivatives which are used in the main theorem about the uniform asymptotic linearity of WME.
Lemma 6.1. Let the regularity conditions in Section 6.2 hold and let λl(xi, yi, θ, τ) be the lth element of the vector λ(xi, yi, θ, τ), defined in (4), for l = 1, …, p+q. Then for l = 1, …, p
| (8) |
where
| (9) |
and fθl(xi, θ) = (∂/∂θl) f (xi, θ).
Proof. By the definition of derivative, we may write for j, l = 1, …, p
| (10) |
Then,
and
![]() |
Then by taking expectations on both sides we get
![]() |
Thus, by conditions [A3] (i),(iii), and [B3] (i),(ii), we have that:
and
Also,
Therefore, we have (8).
Lemma 6.2. Let the regularity conditions in Section 6.2 hold and let λl(xi, yi, θ, τ) be the lth element of the vector λ(xi, yi, θ, τ), defined in (4), for l = 1, …, p+q. Then for l = 1, …, p
| (11) |
where λl(xi, yi, θ, τ) is defined in (9).
Proof. From (10), we have
which by using the Markov WLLN and conditions [A2] (i)–(ii) yields
and
Therefore, we have (11).
Lemma 6.3. Let the regularity conditions in Section 6.2 hold and let λl(xi, yi, θ, τ) be the lth element of the vector λ(xi, yi, θ, τ), defined in (4), for l = 1, …, p+q. Then for l = 1, …, p
| (12) |
where λl(xi, yi, θ, τ) is defined in (9).
Proof. By the definition of derivative, we may write for l = 1, …, p, j = 1, …, q,
| (13) |
where στj(zi, τ) = (∂/∂τj)σ(zi, τ). Then,
and
![]() |
Then by taking expectations on both sides we get
![]() |
Then, by conditions [A3] (i),(iv), we have that
and
Also,
Therefore, we have (12).
Lemma 6.4. Let the regularity conditions in Section 6.2 hold and let λl(xi, yi, θ, τ) be the lth element of the vector λ(xi, yi, θ, τ), defined in (4), for l = 1, …, p+q. Then for l = 1, …, p
| (14) |
where λl(xi, yi, θ, τ) is defined in (9).
Proof. From (13), we have
which by using the Markov’s Weak Law of Large Numbers (WLLN) and conditions [A2] (i)–(ii) yields
and
Therefore, we have (14).
Lemma 6.5. Let the regularity conditions in Section 6.2 hold and let λl(xi, yi, θ, τ) be the lth element of the vector λ(xi, yi, θ, τ), defined in (4), for l = 1, …, p+q. Then for l = p + 1, …, p + q
| (15) |
where
| (16) |
Proof. By the definition of derivative, we may write for j = 1, …, p, l = p+1, …, p+q,
| (17) |
Then since (17) is the same as (13), we can follow the proof of Lemma 6.4 to obtain the result of (15).
Lemma 6.6. Let the regularity conditions in Section 6.2 hold and let λl(xi, yi, θ, τ) be the lth element of the vector λ(xi, yi, θ, τ), defined in (4), for l = 1, …, p+q. Then for l = p + 1, …, p + q
where λl(xi, yi, θ, τ) is defined in (16).
Proof. We can also follow the proof of Lemma 6.4.
Lemma 6.7. Let the regularity conditions in Section 6.2 hold and let λl(xi, yi, θ, τ) be the lth element of the vector λ(xi, yi, θ, τ), defined in (4), for l = 1, …, p+q. Then for l = p + 1, …, p + q
| (18) |
where λl(xi, yi, θ, τ) is defined in (16).
Proof. By the definition of derivative, we may write for l = p + 1, …, p + q, j = 1, …, q,
| (19) |
Then,
and
![]() |
Then by taking expectations on both sides we get
![]() |
Then, by conditions [A3] (ii),(v), [C3] (i),(ii), we have that
and
Also,
Thus we have (18).
Lemma 6.8. Let the regularity conditions in Section 6.2 hold and let λl(xi, yi, θ, τ) be the lth element of the vector λ(xi, yi, θ, τ), defined in (4), for l = 1, …, p+q. Then for l = p + 1, …, p + q
| (20) |
where λl(xi, yi, θ, τ) is defined in (16).
Proof. From (19), we have
which by using the Markov WLLN and conditions [A2] (i)–(ii) yields
and
Therefore, we have the result in (20).
Contributor Information
Changwon Lim, Email: clim2@luc.edu.
Pranab K. Sen, Email: pksen@bios.unc.edu.
References
- 1.Arunachalam R, Balakrishnan V. Statistical Modeling for Wheat (Triticum Aestivum) Crop Production. International Journal of Statistics and Applications. 2012;2:40–46. [Google Scholar]
- 2.Huber PJ. Robust regression: asymptotics, conjectures and Monte-Carlo. Ann. Statist. 1973;1:799–821. [Google Scholar]
- 3.Huber PJ. Robust Statistics. New York: John Wiley & Sons; 1981. [Google Scholar]
- 4.Jureĉková J, Sen PK. Robust Statistical Procedures, Asymptotics and Interrelations. New York: John Wiley & Sons; 1996. [Google Scholar]
- 5.Kent JT, Tyler DE. Regularity and uniqueness for constrained M-estimates and redescending M-estimates. Ann. Statist. 2001;29(1):252–265. [Google Scholar]
- 6.Klein R, Yohai VJ. Asymptotic behavior of iterative M-estimators for the linear model. Comm. Statist. 1981;A10:2373–2388. [Google Scholar]
- 7.Lim C, Sen PK, Peddada SD. Statistical Inference in Nonlinear Regression under Heteroscedasticity. Sankhya Ser. B. 2011a;72:202–218. [Google Scholar]
- 8.Lim C, Sen PK, Peddada SD. M-estimation Methods in Heteroscedastic Nonlinear Regression Models. In: Bhattacharjee M, Dhar S, Subramanian S, editors. in Recent Advances in Biostatistics: False Discovery, Survival Analysis and Related Topics - Volume 4. Series in Biostatistics. World Scientific Press; 2011b. pp. 169–190. [Google Scholar]
- 9.Lim C, Sen PK, Peddada SD. Accounting for Uncertainty in Heteroscedasticity in Nonlinear Regression. J. Stat. Planning and Inference. 2012a;142:1047–1062. doi: 10.1016/j.jspi.2011.11.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Lim C, Sen PK, Peddada SD. Robust Analysis of High Throughput Screening (HTS) Assay Data. Technometrics. 2012b doi: 10.1080/00401706.2012.749166. in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Maronna RA. Robust M-estimators of multivariate location and scatter. Ann. Statist. 1976;4:163–169. [Google Scholar]
- 12.Maronna RA, Martin DR, Yohai VJ. Robust Statistics: Theory and Methods. New York: John Wiley & Sons; 2006. [Google Scholar]
- 13.Maronna RA, Yohai VJ. Robust low-rank approximation of data matrices with elementwise contamination. Technometrics. 2008;50(3):295–304. [Google Scholar]
- 14.Prajneshu K, Das PK. Modelling wheat (Triticum aestivum) production at state level in post-green revolution era. Ind. Jour. Agril. Sci. 1998;68:23–26. [Google Scholar]
- 15.Prajneshu K, Sharma VK. A nonlinear statistical model for adoption of high yielding varieties. Ind. J. Appl. Stat. 1992;1:12–18. [Google Scholar]
- 16.Ratkowsky DA. Handbook of Nonlinear Regression Models. New York, USA: Marcel Dekker; 1990. [Google Scholar]
- 17.Relles D. PhD dissertation. Yale University; 1968. Robust regression by modified least squares. [Google Scholar]
- 18.Sanhueza AI. Ph.D. thesis. University of North Carolina at Chapel Hill; 2000. Robust M-procedures in Nonlinear Regression Models. [Google Scholar]
- 19.Sanhueza AI, Sen PK. M-methods in generalized nonlinear models. In: Charalambides ChA, Koutras MV, Balakrishnan N., editors. Probability and Statistical Models with Applications. New York: Chapman & Hall/CRC; 2001. pp. 359–375. [Google Scholar]
- 20.Sanhueza AI, Sen PK. Robust M-procedures in univariate nonlinear regression models. Brazilian Journal of Probability and Statistics. 2004;18:183–200. [Google Scholar]
- 21.Sanhueza AI, Sen PK, Leiva V. A robust procedure in nonlinear for repeated measurements. Communications in Statistics - Theory and Methods. 2009;38:138–155. [Google Scholar]
- 22.Singer JM, Sen PK. M-methods in multivariate linear models. J. Multivar. Anal. 1985;17:168–184. [Google Scholar]
- 23.Tyler DE. High breakdown point multivariate M-estimation. Estadística. 2002;54:213–247. [Google Scholar]
- 24.Yohai VJ, Maronna RA. Asymptotic behaviour of M-estimators for the linear model. Ann. Statist. 1979;7:258–268. [Google Scholar]








