Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2012 Jul 6.
Published in final edited form as: Electron J Stat. 2011;5:1424–1449. doi: 10.1214/11-EJS647

Estimation via corrected scores in general semiparametric regression models with error-prone covariates

Arnab Maity 1,*, Tatiyana V Apanasovich 2,
PMCID: PMC3390987  NIHMSID: NIHMS335629  PMID: 22773940

Abstract

This paper considers the problem of estimation in a general semiparametric regression model when error-prone covariates are modeled parametrically while covariates measured without error are modeled nonparametrically. To account for the effects of measurement error, we apply a correction to a criterion function. The specific form of the correction proposed allows Monte Carlo simulations in problems for which the direct calculation of a corrected criterion is difficult. Therefore, in contrast to methods that require solving integral equations of possibly multiple dimensions, as in the case of multiple error-prone covariates, we propose methodology which offers a simple implementation. The resulting methods are functional, they make no assumptions about the distribution of the mismeasured covariates. We utilize profile kernel and backfitting estimation methods and derive the asymptotic distribution of the resulting estimators. Through numerical studies we demonstrate the applicability of proposed methods to Poisson, logistic and multivariate Gaussian partially linear models. We show that the performance of our methods is similar to a computationally demanding alternative. Finally, we demonstrate the practical value of our methods when applied to Nevada Test Site (NTS) Thyroid Disease Study data.

Keywords: Generalized estimating equations, generalized linear mixed models, kernel method, measurement error, Monte Carlo Corrected Score, semiparametric regression

1. Introduction

Regression models with measurement errors arise frequently in practice and have attracted much attention in the statistical literature. Semiparametric regression models with errors in covariates have been considered by several authors in the attempt to develop measurement error calibration techniques when the errors are in the linear part of linear regression ([9]) or generalized linear regression ([10]) models. [30] used a method of moments and deconvolution to construct the calibration for the case of partially linear models when the mismeasured covariate appears in parametric and nonparametric parts. However all of the above methodologies take advantage of the fact that unknown parameters in a parametric part enter the model through a linear combination with error prone covariates. We consider a general semiparametric regression problem where parameters can enter the model through any known function of covariates. Recently, for a general semiparametric regression problem we proposed to utilize a popular alternative to regression-calibration, a simulation-extrapolation method (SIMEX, Apanasovich et al. [2]). We considered the situations where the mis-measured variable is modeled purely parametrically, purely nonparametrically, or when the mismeasured variable has components that are modeled both parametrically and nonparametrically. Even though SIMEX is a general-purpose, widely applicable method for correcting parameter estimates for the biases induced by measurement error in covariates, it suffers from relying on a rather heuristic extrapolation step ([4]).

Ma and Carroll ([15]), building upon work of [29], developed a functional methodology for a general semiparametric measurement error regression framework. They require a specification of a parametric distribution for error-prone covariates given all other covariates. Nevertheless the method is general, the estimators are still consistent and asymptotically normally distributed even when that distribution is misspecified. However, the implementation of Ma and Carroll's method ([15]) requires solving integral equations, which can be quite computationally expensive. More importantly, this procedure is computationally infeasible when the error prone covariates are multivariate, e.g., in repeated measures or longitudinal data settings. Moreover, the methodology is efficient only when the posited parametric distribution for error-prone covariates given all other covariates is correctly specified, which is not a practical assumption for many applications.

In this paper, we develop an alternative functional methodology where almost no assumptions are made about the distribution of an error prone covariate. We consider a classical additive measurement error model, where covariate X is unobserved and W is observed instead, such that W = X + U. The measurement error U is independent of any other observed variables and has a normal distribution with zero mean and variance Σu,0. The normal distribution is often used in the literature and is not too restrictive. There are many other ways to model the relationship between X and W, which give additivity in some functional scale, e.g. logarithmic. See [6] for more details on transformations of X.

Our method is based upon the idea of Monte Carlo corrected scores (Novick and Stefanski [18]), where Monte Carlo simulations help to determine the corrected score for a large class of models. To our knowledge, this is the first attempt to introduce these ideas into semiparametric framework. While our method uses complex-value arithmetic, it is relatively easy to implement in many standard software packages. Specifically, its theory falls into the framework of standard estimation methods for criterion functions in semiparametric problems (Lin and Carroll [12]), thus we are taking advantage of some well established results. Hence, the core of our method's implementation is the standard semiparametric estimation technique adjusted for complex valued covariates and performed with a relatively small number of Monte Carlo runs.

Examples of widely used regression models for which our methods apply include: linear model ([9]), Poisson regression model ([3, 18]), Gamma regression model ([21, 1, 17]). In our paper we devote attention to each of these models except Gamma. We use univariate and multivariate partially linear models to illustrate the relationship between Monte Carlo corrected score method and other methods which exist in the literature. Specifically, in the univariate case, we show that as the number of Monte Carlo iterations goes to infinity, the estimators converge to that of [9]; and in the case of multivariate partially linear models, to that of [12]. Moreover, when X is Gaussian, our univariate estimator is efficient (Ma and Carroll [15]). Further, we demonstrate through simulations that our multivariate estimator performs similar to the efficient one with the reasonable number of Monte Carlo iterations even though its limit is not an efficient estimator ([12]). Poisson, logistic and multivariate Gaussian regression models are studied via simulations to demonstrate the ease of implementation and generality of proposed methods.

Note that logistic regression model is an exception to our method and we only present heuristic arguments for this model. The problem lies in the fact that the logistic distribution function is not an entire function, which is an essential theoretical condition for our method to be applicable. However, Novick and Stefanski ([18]) noted that for measurement error variance of the magnitudes commonly encountered in applications one can still apply corrected score based methods, with only a minor bias (see p. 479 of Novick and Stefanski [18]). Results from numerical studies show that our method also performs reasonably well for the logistic case when the measurement error is moderate.

An outline of this paper is as follows. In Section 2, we review the estimation in a general semiparametric regression where there is no measurement error as studied in Lin and Carroll ([12]); and corrected score method proposed by [17] and Novick and Stefanski ([18]). Then we introduce Monte Carlo corrected score method to the estimation in a general semiparametric regression with error prone covariates and study the asymptotic behavior of the proposed estimators. Among other results we offer asymptotic standard errors accounting for the uncertainty due to estimation of measurement error covariance matrix.

In Section 3, we focus on special cases of univariate and multivariate partially linear models. We demonstrate that as the size of Monte Carlo correction sample increases, our estimators converge to the ones mentioned in the literature ([9, 12]). Moreover, we show that when the error prone covariate is Gaussian, our estimator is efficient in the univariate case and performs similar to the efficient estimator in the multivariate case.

In Section 4, we present a simulation study using several semiparametric models to illustrate the performance of our method. We start with partially linear Poisson model and show that our method produces only a small bias and appropriate coverage probability. We also use the simulation scenario of Ma and Carroll ([15]) applied to the logistic partially linear model with a quadratic effect of X. In this case we note that our method, despite lacking a theoretical foundation for the logistic regression, produces even slightly smaller mean squared errors than theirs, while being computationally far less challenging. Moreover, when we triple the variance of the measurement error used in the scenario, we still get relatively small errors in the estimators. Last, we report results of simulations in the case of multivariate partially linear model with multivariate measurement errors. Such a model would be computationally challenging for the competing methodology (Ma and Carroll [15]), while our methods offer ease of implementation and satisfactory performance.

In Section 5.1, we apply our method to Nevada Test Site (NTS) Thyroid Disease Study data and report the results. We also present a simulation study where the measurement errors in the covariates are comparable to those in the real NTS data example. We show that for the assumed amount of uncertainty in radiation exposure, our method performs reasonably well.

Finally, Section 6 gives a few brief concluding remarks. All technical details are collected in an appendix.

2. Methodology

We describe a general semiparametric regression when there is no measurement error as studied in Lin and Carroll ([12]) first. Assume that data (Yi, Xi, Zi), i = 1, . . . , n are independent replications of a (py + px + 1)-dimensional random vector (Y, X, Z). Let B denote the parameter of interest and θ(·) be an infinitely dimensional nuisance parameter with true values of B0 and θ0(·) respectively. Let L{Y,X,B0,θ0(Z)} be a criterion function in the sense that E[LB{Y,X,B0,θ0(Z)}Z]=0 and E[Lθ{Y,X,B0,θ0(Z)}Z]=0, where here and in what follows, we use subscripts B and θ to denote the partial derivatives with respect to B and θ respectively. Suppose K(·) is a symmetric density function with variance 1 and define Kh(v) = h−1K(v/h), where h is the bandwidth. Let Gi(z) = {1, (Zi – z)/h}T. Given a fixed value of B=B, the modified kernel estimate of θ^(z,B) is a solution of the local-linear estimating equations

n1i=1nKh(Ziz)Gi(z)Lθ{Yi,Xi,B,(α0,α1)Gi(z)}=0 (2.1)

for α0, calling it α^0. To estimate B, Lin and Carroll ([12]) proposed profile and backfitting methods. The profile kernel estimator for B maximizes i=1nL{Yi,Xi,B,θ^(Zi,B)} in B which is equivalent to solving the score equation

n1i=1n[LB{Yi,Xi,B,θ^(Zi,B)}+Lθ{Yi,Xi,B,θ^(Zi,B)}θ^B(Zi,B)]=0. (2.2)

Maximization of the profile likelihood requires calculating θ^B(Zi,B)=θ^(Zi,B)B, which can be computed by numerical differentiation. In some cases where the profile kernel methods may be difficult to implement numerically, a backfitting algorithm can be used instead. Suppose that the current estimate is B, the updated backfitting estimate then maximizes B in the function i=1nL{Yi,Xi,B,θ^(Zi,B)} or is the solution of

n1i=1nLB{Yi,Xi,B,θ^(Zi,B)}=0.

The second step in backfitting iterations is to find a solution of the local linear estimating equations (2.1) using updated estimate of B. Lin and Carroll ([12]) showed that profiling and backfitting are asymptotically equivalent, however to obtain n-consistent estimator of B, unlike profiling, undersmoothing of the nonparametric function is required by the backfitting method.

In this paper, we consider the case where covariate X is unobserved and instead W is observed such that (possibly after transformation)

W=X+U,

where U is independent of any other observed variables and follows a Normal distribution with mean 0 and covariance matrix Σu,0. Measurement error induces bias in estimating equations (2.1) and (2.2), which results in biased estimators of the parameters. Thus, the purpose of the current study is to modify (2.1) and (2.2) and obtain unbiased estimating equations corrected for measurement error, as we describe next.

2.1. Monte-Carlo corrected score estimation

A function L~{Y,W,B0,θ0(Z)} is a criterion function if

E[L~{Y,W,B,θ(Z)}Y,X,Z]=L{Y,X,B,θ(Z)}

for any B, θ(Z) from the parameter space ([17]). Novick and Stefanski ([18]), considering parametric models, proposed a general method to construct corrected score functions based on the results from complex analysis. Let W~ be a complex random vector

W~=W+ιV,

where ι=1 and V is a normal random vector with mean 0 and covariance matrix Σu,0. [26] showed that if f(·) is an entire function then, under integrability conditions

E{f(W~)X}=E[Re{f(W~)}X]=f(X),

where Re(·) denotes the real part of its argument.

Assume that L() is an entire function of its second argument. We define the corrected criterion function as

L~{Y,W,B0,θ0(Z)}=E(Re[L{Y,W~,B0,θ0(Z)}]Y,W,Z).

However, the required conditional expectation is not always easy to obtain analytically. Novick and Stefanski ([18]) proposed to use Monte Carlo integration to approximate the conditional expectation in the parametric models they considered. We introduce Monte Carlo correction into our semiparametric models so that Monte Carlo corrected criterion function becomes

R()=M1m=1MRe[L{Y,W~m,B0,θ0(Z)}],

where here and in what follows (·) denotes a real argument {Y,W,V¯,B0,θ0(Z)}, where V¯=(V1,,VM). Here M is the Monte Carlo correction sample size. We will suppress the dependence on M in R() for notational convenience. Note that R() is a real valued function of real arguments and is a criterion function in the sense we discussed in the beginning of the section. Therefore, analogous to (2.1), we propose to estimate θ0(z) by solving

n1i=1nKh(Ziz)Gi(z)Rθ{Yi,Wi,V¯i,B,(α0,α1)Gi(z)}=0 (2.3)

for α0, setting θ^(z,B)=α^0 at some fixed B.

There are two methods to estimate B:

  1. The profile kernel estimator B^pf maximizes n1i=1nR{Yi,Wi,V¯i,B,θ^(Zi,B)} in B solving
    n1i=1n[RB{Yi,Wi,V¯i,B,θ^(Zi,B)}+Rθ{Yi,Wi,V¯i,B,θ^(Zi,B)}×θ^B(Zi,B)]=0. (2.4)
  2. The backfitting kernel estimator, B^bf is obtained at convergence of the following iterations. Based on the current estimate, B^cur, solve for B
    n1i=1nRB{Yi,Wi,V¯i,B,θ^(Zi,B^cur)}=0
    call B^new, and then solve the local linear estimating equation (2.3) using B^new.

2.2. Asymptotic properties

In this section, we derive the asymptotic properties of our method in the case where the measurement error covariance matrix Σu,0 is known, see Section 2.4 for the case where it is estimated. The results given in Lin and Carroll ([12]) for the profiling and backfitting methods are true for any criterion function as long as various conditions are satisfied: these conditions translate into A1-A4, given in the Appendix. We assume that conditional expectations of W, V¯ given Y,Z,X and the first and second order partial derivatives of L with respect to B and θ exist and are interchangeable. Define

θB(z,B0)=E{RθB()Z=z}E{Rθθ()Z=z}=E[LθB{Y,X,B0,θ0(Z)}Z=z]E[Lθθ{Y,X,B0,θ0(Z)}Z=z];Ω(z)=fZ(z)E{Rθθ()Z=z}=fZ(z)E[Lθθ{Y,X,B0,θ0(Z)}Z=z];V=E{RBB()+RBθ()θBT(Z,B0)}=E[LBB{Y,X,B0,θ0(Z)}+LBθ{Y,X,B0,θ(Z)}θBT(Z,B0)].

Then the following result is a direct consequence of the main results of Lin and Carroll ([12]).

Theorem 2.1. Assume that (Yi, Zi, Wi), i = 1, . . . , n are independent and identically distributed; and B^pf and θ^() are estimates obtained by using (2.3) and (2.4). Suppose further that the bandwidth h ∝ nc with 1/5 ≤ c ≤ 1/3. Let θ(2)(z) be the second derivative of θ(z) and ϕ2=z2K(z)dz. Then, for the nonparametric part

θ^(z,B^pf)θ0(z)=(h22)ϕ2θ0(2)(z)n1i=1nKh(Ziz)Riθ()Ω(z)θB(z,B0)TV1n1i=1n{RiB()+Riθ()θB(Zi,B0)}+op(n12);

and for the parametric part

n12(B^pfB0)=V1n12i=1n{RiB()+Riθ()θB(Zi,B0)}+op(1)Normal(0,V1FVT),

where F=var[RB()+Rθ()θB(Z,B0)].

Theorem 2.2. Make the same assumptions as in Theorem 2.1 except that nh4 → 0. Then the backfitting estimator B^bf has the same limiting distribution as does the profile estimator B^pf.

Remark 2.1. One can show that

F=F1+F2+F3,

where F is F1 in the absence of measurement error,

F1=var[LB{Y,Z,X,B0,θ0(Z)}+Lθ{Y,Z,X,B0,θ0(Z)}θB(Z,B0)],

F2 is the additional variation due to the use of corrected scores

F2=E(var[L~B{Y,Z,W,B0,θ0(Z)}+L~θ{Y,Z,W,B0,θ0(Z)}θB(Z,B0)Y,Z,X])

and F3 is the additional variation due to the use of Monte Carlo method

F3=E(var[RB()+Rθ()θB(Z,B0)Y,Z,W])=O(M1).

Remark 2.2. Estimation of the asymptotic variance of B^pf or B^bf is a straight-forward exercise. To construct such the estimates, all the expectations in the definitions of V and F are replaced by sums and all the regression replaced by kernel estimates.

2.3. Multivariate measurement error models

Consider longitudinal or repeated measures data, where for each subject we observe L responses Y = (Y1, . . . , YL); and predictors X = (X1, . . . , XL) and Z = (Z1, . . . , ZL), where each Zj is scalar. The underlying loglikelihood function is taken to be of the form L{Y,X,B0,θ0(Z1),,θ0(ZL)}. Here the key feature is that the nonparametric component θ0(·) is evaluated multiple times per individual. For notational convenience, we use boldface to denote the multivariate version of the corresponding observations, and let θ¯0(Z)={θ0(Z1),,θ0(ZL)}. In the measurement error settings under consideration, instead of observing Xj, we observe Wj = Xj + Uj, j = 1, . . . , L and assume that vec(U), where “vec” is a vector form of a matrix, has a Normal distribution with mean zero and covariance matrix Σu,0. Let W~m=W+ιVm for m = 1, . . . , M, where vec(Vm) has a Normal distribution with mean zero and covariance matrix Σu,0. Then Monte Carlo corrected criterion function is given by

M1m=1MRe[L{Y,W~m,B,θ¯(Z)}],

and our asymptotic results apply.

2.4. Estimation of error covariance matrix

We now consider the case in which the measurement error covariance matrix Σu,0 is estimated from replicated measurements. Suppose we observe R ≥ 2 replicates {Wi(r)}r=1R, where Wi(r) = Xi + Ui(r) and Ui(r) has Normal(0, Σu,0). Then a root-n consistent estimate Σ^u of Σu,0 is the sample covariance matrix of the terms Wi(r)s is n1i=1nSi where

Si=r=1R{Wi(r)W¯i.}{Wi(r)W¯i.}T(R1),andW¯i.=R1r=1RWi(r).

Let γ = vech(Σu), where “vech” is the vector half, i.e., the vector of the unique elements of Σu. Then we have that γ^γ0=n1ivech(SiΣu,0). Since V can be written as Σu12e where e comes from a standard Normal distribution, we can redefine the score equations (2.4) as

n1i=1n[RB{Yi,W¯i(r),Σu12e¯i,B,θ^(Zi,B,Σu)}+Rθ{Yi,W¯i(r),Σu12e¯i,B,θ^(Zi,B)}×θ^B(Zi,B,Σu)]=0.

Let subscript γ denote a partial derivative with respect to γ. Then following Section 4 of Lin and Carroll ([12]), we have the following asymptotic expansion for the profile estimator:

n12(B^pfB0)=V1[n12i=1n{RiB()+Riθ()θB(Zi,B0,Σu,0)}+VBγn12(γ^γ)]+op(1)=V1(n12i=1n[RiB()+Riθ()θB(Zi,B0,Σu,0)+VBγ{vech(SiΣu,0)}])+op(1),

where VBγ=E{RiBγ()+θB(Zi,B0,Σu,0)RiθγT()}. The covariance matrix of n12(B^pfB0) follows from the above expressions and its estimator can be constructed as described in Remark 2.2.

3. Special case: Partially linear model

Two regression examples, the Univariate and Multivariate Partially Linear Models, are considered to illustrate the relationship between proposed Monte Carlo corrected score method and other methods that exist in the literature. Poisson and Logistic Partially Linear Models are also studied via simulations in the next section to demonstrate the general applicability of proposed methods.

3.1. Univariate partially linear model

Estimation in the partially linear model with error prone covariates are described in [9]. In this section we derive the asymptotic distribution of our estimates explicitly and compare our estimates to that of [9].

Consider the model

Yi=XiTγ0+θ0(Zi)+i,i=1,,n,

where ε has a Normal distribution with mean zero and variance σ02. Let β = (γT, σ2)T and choose the loglikelihood to be our criterion function

L{Y,X,θ(Z),β}=log(σ2)2(2σ2)1{YXTγθ(Z)}2.

Then, the corrected loglikelihood as discussed in the previous section is

R{Y,W,V~,θ(Z),β}=log(σ2)2M1m=1MRe[(2σ2)1{Y(W+ιVm)Tγθ(Z)}2]=log(σ2)2(2σ2)1[{YWTγθ(Z)}2γTM1m=1MVmVmTγ].

Also, define

Γ=E[{XE(XZ)}(UTγ0)2{XE(XZ)}T]+E(UUT2)+E{(UUTΣu,0)γ0γ0T(UUTΣu,0)T};S=cov{XE(XZ)};τ2=E{(UTγ0)2(σ02+γ0TΣu,0γ0)}2;C=Σu,0γ0+[E{U(UTγ0)3}Σu,0γ0γ0TΣu,0γ0](2σ02).

Then we have the following result:

Theorem 3.1. Let γ^ and σ^2 denote the estimates based on our method. Then jointly,

n12(γ^γ0σ^2σ02)Normal{0,[S1ΓST+R112σ2S1C+R12τ2+R22]},

where R11, R12, R22 → 0 as M → ∞.

See Appendix B for the proof and the expressions of R11, R12 and R22.

Remark 3.1. It is important to note that as Rijs vanish, the limiting distribution of our estimators becomes the same as of estimators in [9]. Ma and Carroll ([15]) showed that the estimator in [9] is exactly the same as theirs when posited p(X), p(W|X) and p(Y|X, Z) are all normal, proving that our estimator is also efficient when M = ∞.

3.2. Multivariate partially linear model

We illustrate our approach on the following multivariate partially linear measurement error model discussed in Lin and Carroll ([12])

Yij=XijTβ0+θ0(Zij)+eij, (3.1)

for i = 1, . . . , n and j = 1, . . . , L, where ei = (ei1, . . . , eiL)T has a Normal distribution with mean zero and covariance matrix Σε,0. Let B=(β,Σ) be the parameter of interest. Then the criterion function ignoring the measurement errors is given by

L{Y,X,B,θ(Z~)}=(12)log{det(Σ1)}(12){YXβθ(Z)}TΣ1{YXβθ(Z)}.

The Monte-Carlo Corrected Scores criterion function is given by

R()=M1m=1MRe[L{Y,W~m,B,θ(Z)}]=(12)log{det(Σ1)}(12){YWβθ(Z)}TΣ1{YWβθ(Z)}+(12)βT(M1m=1MVmTΣ1Vm)β.

The backfitting algorithm is easy to apply in this case. Given the current estimates, B^cur=(B^cur,Σ^,cur), the new estimates are given by

β^new=[n1i=1n{WiTΣ^,cur1WiM1m=1M(VimTΣ^,cur1Vim)}]1×n1i=1nWiTΣ^,cur1{Yiθ^(Zi,B^cur)};

and

Σ^,new=n1i=1n[{YiWiβ^curθ^(Zi,B^cur)}×{YiWiβ^curθ^(Zi,B^cur)}TM1m=1M(Vimβ^curβ^curTVimT)].

Profile pseudolikelihood estimates are also easily constructed. Let S be a smoother matrix as in [11] and define Y=(Y11,,Ynm)T and T=(W1T,,WnT)T. Let T=(IS)T,Y=(IS)Y, Y=(IS)Y and Σ~=InΣ. Then for given Σε the profile estimate of β is given by

β^pf={TTΣ~1Ti(M1mVimTΣ1Vim)}1TTΣ~1Y.

A simple estimate of Σε,0 can be constructed by first forming the working independence estimate of β0, then applying the above equation to obtain Σ^,new and iterating the steps until convergence.

Remark 3.2. Estimation of the error covariance matrix Σu,0 and its impact on limiting distribution theory for estimation of B0 is described in Section 2.4.

Remark 3.3. As M → ∞, our estimators converges to those given in Lin and Carroll ([12]), see Appendix C for a sketch of proof.

3.3. Performance with respect to efficient methods

We show that under the assumption that X is generated from a Gaussian distribution, our method using different M < ∞, as well as Lin and Carroll's procedure ([12]), which is equivalent to our method with M = ∞, performs very similar to the semiparametrically efficient method.

Suppose X is from a Normal distribution with mean μx,0 and covariance matrix Σx,0; and for simplicity of notation we let β0 be a scalar. Then the criterion function becomes

LG()=(12){log(J(μx,Σx,Σ))+[YQ{W,β,θ(Z),μx,Σx}]T×{J(μx,Σx,Σ)}1[YQ{W,β,θ(Z),μx,Σx}]+log(Σx+Σu)(12)(Wμx)T(Σx+Σu)1(Wμx)},

where

Q{W,β,θ(Z),μx,Σx}=βμx+θ(Z)+βΣx(Σx+Σu)1(Wμx);J(β,Σx,Σ)=Σ+β2Σx(Σx+Σu)1Σu.

By the results of Lin and Carroll ([12]), the estimators based on LG() are semiparametrically efficient.

We compared our method with the optimal one (see discussion above) via a simulation study under the following scenario. We set L = 3 and let β0 = 0.7, θ0(z) = 0.5 cos(2z) – 1. We chose μx,0 = (−1, −1, −1)T, Σε,0 = I3 + 0.3(J3I3), Σx,0 = I3, and Σu,0 = 0.3I3 + 0.2J3, where Jk denotes the k × k matrix with all the elements equal to one. We generated Z from a Uniform on [0, π] distribution.

Under this setup, we generated 1, 000 data sets following the model given by (3.1) with n = 200. We used Epanechnikov kernel with the bandwidth estimated as σ^zn13, where σ^z is the sample standard deviation of Z. Using each data set we constructed backfitting estimator of β0 using our method with different values of M ranging from 100 to 500; Lin and Carroll method ([12]), which is ours when M = ∞; and semiparametrically efficient method (using LG()). Root mean squared error (RMSE) of β^ does not differ much between M = 100 and M = ∞ (as in Lin and Carroll [12]) and is equal to 0.1432. Hence we show that our methods do not require many Monte Carlo runs. In fact in our numerical studies we find M = 150 satisfactory. RMSE for the semiparametrically efficient method is 0.1421. This indicates that the efficiency of our method is very close to the optimal (0.77 percent loss).

4. Simulation study

4.1. Partially linear poisson regression model

We study the performance of our method via a simulation study. We considered the partially linear Poisson regression model where the response Y is Poisson distributed with mean λ(X, Z) = XTβ0+θ0(Z). The true variable X = (X1, X2) was generated from a bivariate standard Gaussian distribution, Z was generated from a Uniform on [0, π] distribution. Error prone covariate was generated as W = X + U where U followed a bivariate normal distribution with mean zero and a known covariance matrix Σu,0. We set Σu,0 = I2 and B0=(β1,0,β2,0)=(0.2,0.2) and used two different functions: (1) θ0(z) = 5 – 0.5 cos(z) and (2) θ0(z) = 5 – 0.5 cos(2z). We generated 1, 000 samples of size n = 1, 000 and used M = 500 as Monte Carlo correction sample size.

We employed Epanechnikov kernel to estimate the nonparametric function. We used the globally fixed bandwidth hn=κσ^zn13, where σ^z is the estimated standard deviation of Z and κ is some selected positive number. We report the results for κ = 1. Similar results were obtained for other values of κ ranging from 0.5 to 2. We used backfitting to estimate B0.

The results are displayed in Table 1. It is clear that our method produces only a small bias and favorable coverage probability.

Table 1.

Results of the simulations using Poisson regression model. In nonparametric part: (1) is θ0(z) = 5 – 0.5 cos(z) and (2) is θ0(z) = 5 – 0.5 cos(2z). Reported are the mean, empirical standard errors (e.s.e.), root mean squared error (RMSE) and empirical coverage of 95% confidence intervals of β1,0 and β2,0 based on 1000 simulated data sets each with a sample size n = 1000

θ0(z) estimation of β1,0 = 0.2 estimation of β2,0 = 0.2

mean e.s.e. RMSE 95% mean e.s.e. RMSE 95%
(1) 0.203 0.039 0.039 0.951 0.199 0.038 0.038 0.952
(2) 0.204 0.040 0.041 0.954 0.202 0.041 0.041 0.955

4.2. Partially linear logistic model

We borrowed a simulation scenario applied to a logistic regression model from Ma and Carroll ([15]). As in their paper, we considered the model logit{pr(Y = 1|X, Z)} = β1,0X + β2,0X2 + θ0(Z), where W = X + U and U is from Normal distribution with mean 0 and variance σu,02 with σu,02 known. We set σu,02=0.16, B0=(β1,0,β2,0)=(0.7,0.7) and used two different functions: (1) θ0(z) = 0.5 cos(z) – 1 and (2) θ0(z) = 0.5 cos(2z) – 1. The covariates X and Z were generated from Normal distribution with mean −1 and variance 1, and Uniform on [0, π], respectively. We used the sample size of n = 500 and M = 150 Monte Carlo correction sample size. Backfitting with Epanechnikov kernel was used to estimate the model components. For the sake of comparison, we used the global bandwidth hn=σ^zn13 as in Ma and Carroll ([15]).

Technically, the logistic regression setup as described above does not fall into our framework as the logistic distribution function is not entire in the complex plane. However, Novick and Stefanski ([18]) pointed out that for small measurement error variance one can still apply corrected score based methods, with only a minor bias. Specifically, Novick and Stefanski ([18]) followed the same paradigm to construct likelihood based corrected score, call it ΨM(Y, W) and showed that Ψ*(Y, X) = E[limM→∞ ΨM(Y, W)|Y, X] is not the same as the true likelihood score function. However, they argued that the differences between the components of Ψ*(Y, X) and the true likelihood score function are small for measurement error variances of the magnitudes commonly encountered in applications (see p. 479 of Novick and Stefanski [18]).

The results of the simulation study are displayed in Table 2. It is evident that our method is comparable in both cases to that of Ma and Carroll ([15]) in terms of mean squared error and coverage probability, albeit with the small bias expected from the fact that the logistic function is not entire on the complex plane. It is clear from this numerical example that even though technically our method is not applicable in logistic regression, it performs quite well and is very close to that of Ma and Carroll ([15]).

Table 2.

Results of the simulations using logit model. In nonparametric part: (1) is θ0(z) = 0.5 cos(z) – 1 and (2) is θ0(z) = 0.5 cos(2z) – 1. Reported are the mean, empirical standard errors (e.s.e.), root mean squared error (RMSE) and empirical coverage of 95% confidence intervals of β1,0 and β2,0 for two values of σu,02, based on 1000 simulated data sets each with a sample size n = 500. Our method is coded as MA in the column “Me” and the results from Ma and Carroll ([15]) are coded as MC. Ma and Carroll [15] did not consider the case of σu,02=0.5

θ0(z) σu,02 Me estimation of β1,0 = 0.7 estimation of β2,0 = 0.7

mean e.s.e. RMSE 95% mean e.s.e. RMSE 95%
(1) 0.16 MA 0.638 0.261 0.268 0.942 0.653 0.149 0.156 0.940
0.16 MC 0.720 0.277 0.278 0.947 0.726 0.156 0.158 0.939
0.50 MA 0.621 0.272 0.283 0.948 0.633 0.161 0.175 0.943
(2) 0.16 MA 0.615 0.238 0.253 0.943 0.639 0.135 0.148 0.942
0.16 MC 0.727 0.276 0.277 0.951 0.728 0.155 0.158 0.940
0.50 MA 0.600 0.250 0.269 0.942 0.618 0.155 0.175 0.945

The simulation was repeated for a much larger measurement error variance, σu,02=0.5 versus σu,02=0.16. The results are shown in Table 2. Again, our results indicate only a small bias and favorable coverage probability. Ma and Carroll ([15]) did not report results for this situation so it is not possible to compare our method with theirs.

4.3. Multivariate partially linear model

In this section we consider the partially linear Gaussian regression model with repeated measures. We generated data according to the model

Yij=Xij1β1,0+Xij2β2,0+θ0(Zij)+eij,

for i = 1, . . . , n and j = 1, . . . , L, where ei = (ei1, . . . , eiL)T has a Normal distribution with mean zero and covariance matrix Σε,0. We set L = 3 and let β1,0 = β2,0 = 0.7. The covariates Xijl were each generated from a Normal with mean −1 and variance 1 distribution and Z was generated from a Uniform on [0, π] distribution. The error prone variable was generated as Wijl = Xijl + Uijl, where (Uij1, Uij2, Uij3) follows a normal distribution with mean zero and covariance matrix Σu,0 = 0.3I3 + 0.2J3. Here Jk denotes the k × k matrix with all the elements equal to one.

We used two different functions: (1) θ0(z) = 0.5 cos(z) – 1 and (2) θ0(z) = 0.5 cos(2z) – 1, and considered two different covariance structures for Σε,0: (1) compound symmetry with common marginal variance 1 and correlation 0.3 and (2) AR(1) structure with autocorrelation parameter 0.4.

We generated 1, 000 samples of size n = 200 for each scenarion, and used M = 150 as Monte Carlo correction score sample size. The Gaussian kernel was used to estimate the nonparametric function and backfitting was used to estimate the parameters. We used the globally fixed bandwidth hn=κσ^zn13, where σ^z is the estimated standard deviation of Z and κ is some selected positive number. We report the results for κ = 1. Similar results were obtained for other values of κ ranging from 0.5 to 2.

The results are displayed in Table 3. It is evident that our method produces only a small bias and favorable coverage probability for both cases: compound symmetry and AR(1) error structures.

Table 3.

Results of the simulations using multivariate Gaussian measurement error. In nonparametric part: (1) is θ0(z) = 0.5 cos(z) – 1 and (2) is θ0(z) = 0.5 cos(2z) – 1. Reported are the mean, empirical standard errors (e.s.e.), root mean squared error (RMSE) and empirical coverage of 95% confidence intervals of β1,0 and β2,0 based on 1,000 simulated data sets each with a sample size n = 200 for different error correlation structures

Compound symmetry
θ0(z) estimation of β1,0 = 0.7 estimation of β2,0 = 0.7
mean e.s.e. RMSE 95% mean e.s.e. RMSE 95%
(1) 0.687 0.137 0.138 0.945 0.691 0.144 0.145 0.947
(2) 0.685 0.143 0.144 0.949 0.689 0.152 0.152 0.947
AR(1)
θ0(z) estimation of β1,0 = 0.7 estimation of β2,0 = 0.7
mean e.s.e. RMSE 95% mean e.s.e. RMSE 95%
(1) 0.687 0.137 0.138 0.948 0.690 0.144 0.145 0.949
(2) 0.686 0.143 0.143 0.946 0.689 0.152 0.153 0.946

5. Application

5.1. Nevada test site thyroiditis data example

In this section we apply our method to the Nevada test site (NTS) thyroid study data. The study was conducted in 1980's by the University of Utah. The original study is described in [27, 7] and [24]. The main idea of the study was to relate the incidence of thyroid related diseases to the exposure of radiation to the thyroid. In this study, 2, 491 individuals, who were exposed to radiation as children, were tested for thyroid disease. The primary radiation exposure to the thyroid glands of these children came from the ingestion of milk and vegetables contaminated with radioactive isotopes of iodine. Recently, the dosimetry for the study was redone ([25]), and the study results were reported in [14].

Due to the fact that the actual radiation doses in foods or in the thyroid gland of the individuals are not available, the estimated radiation doses are well known to be contaminated with measurement errors. Many authors have studied and described measurement error properties and analysis in this context ([20, 23, 16, 28, 13, 19, 22, 8]). A common approach is to build a large dosimetry model that attempts to convert the known data about above-ground nuclear testing to the radiation actually absorbed into the thyroid. Dosimetry calculations for individual subjects were based upon several variables, such as, age at exposure, gender, residence history, whether as a child the individual was breast-fed, and a diet questionnaire filled out by the parent focusing on milk consumption and vegetables. The data were then put into a complex model and for each individual, the point estimate of thyroid dose (the arithmetic mean of a lognormal distribution of dose estimates) and an associated error term (the geometric standard deviation) for the measurement error were reported.

It is typical to assume that radiation doses are estimated with a combination of Berkson and a classical measurement error ([20]). In the log-scale, true log-dose T is related to observed or calculated log-dose W by a latent intermediate X via

T=X+Uberk;W=X+Uclass,

where Uberk and Uclass are the Berkson uncertainty and the classical uncertainty, respectively, with corresponding variances σu,berk,02 and σu,class,02 depending on the individual. It is typical to assume that the errors Uberk have Gaussian distributions. In the NTS study, the total uncertainty σu,berk,02+σu,class,02 is known but not the relative contributions. We will let 50% of the total uncertainty be classical in our study.

We take the incidence of thyroiditis (inflammation of the thyroid gland), Y, as the response variable. If the latent, X, could be observed then typically the total mean dose, E(TX)=exp(X+σu,berk,022) would be the main predictor. In addition, we consider S, the sex of the patient and Z, age at exposure (standardized to have mean zero and variance 1), which are measured without measurement error. We include Z, age at exposure, nonparametrically into so called excess relative risk model

pr(Y=1X,S,Z)=H[β0S+log{1+γ0exp(X+σu,berk,022)}+θ0(Z)], (5.1)

where H(·) is the logistic distribution function, θ0(·) is an unknown function and γ0 is called the excess relative risk.

We employed our method discussed in Section 2 to the model given by (5.1). Specifically, we used backfitting with Epanechnikov kernel and bandwidth equal to 1.5 (similar results were obtained for 1.0 and 2.0). We used M = 100 as the Monte Carlo correction sample size. We compared our method to the naive method, when one ignores the measurement error of both types. The estimated effect of gender, β^11.75 for both the naive and Monte Carlo corrected scores method. This can be explained by the fact that gender and radiation dose for an individual are essentially independent and hence the effect of gender is not affected by measurement error in radiation dose.

The estimated value of the relative risk parameter was 8.54 for the naive method and 17.19 for the proposed method. The effect of age, Z, is displayed in Figure 1 for both the naive and MCCS procedures. It is evident from the results that because of the difference in the estimate of the excess relative risk γ, there is a noticeable difference in the estimated age effect when measurement error is taken into account.

Fig 1.

Fig 1

Estimated age effect in the Nevada Test Site thyroiditis data. Solid line: the proposed estimate. Dashed line: the naive estimate ignoring the presence of measurement error.

Remark 5.1. As noted in Section 4, the logistic regression setup does not fall into our framework as the logistic distribution function is not entire in the complex plane. To observe the performance of semiparametric-Monte Carlo corrected scores in this example, we compared our results to the well known SIMEX procedure ([5, 26]). Please refer to Apanasovich et. al. ([2]) for details. In short, we modeled the age effect parametrically by a quadratic polynomial and used a quartic extrapolant for SIMEX. The estimated value of the excess relative risk parameter was 15.92. We can see that in this case the SIMEX estimate is close to what proposed method produces.

5.2. Simulation study mimicking the data example

To further assess the performance of our method when applied to the data example, we set up a simulation study where the measurement error present in the variables is of the same magnitude as the measurement error from the data example. Specifically, we generate responses using the model

pr(Y=1X,S,Z)=H[β0S+log{1+γ0exp(X+σu,berk,022)}+θ0(Z)],

where we generate S from a Binary distribution with rate 0.5, X from a Normal distribution with mean zero and variance 4; and Z from a Uniform distribution within [−3, 2] interval. Following the data example, we set true β0 = 1.7 and γ0 = 17.2 and take θ0(z) to be the assumed estimate of a true function. We generate 500 samples of size n = 2500 and, as in the example, use M = 100 as Monte Carlo corrected score runs.

As in the data example, the measurement error is Gaussian, a combination of Berkson and classical and we take the variance of classical uncertainty to be 50% of the total uncertainty present in the NTS data set.

The average of estimates for β0 and γ0 over the 500 generated data sets are 1.72 and 17.79, respectively, with empirical standard errors 0.18 and 4.69. It is evident that the proposed procedure performs quite well, giving low bias and moderate variability for the estimates when the magnitude of measurement errors is similar to that of the original NTS data.

6. Discussion

We consider the problem of estimation in a general semiparametric regression model when error-prone covariates are modeled parametrically while covariates measured without error are modeled nonparametrically. We propose to utilize ideas of corrected score methodology introduced by [18] for a purely parametric framework. Monte Carlo corrected scores method uses simulations and allows for simple implementation of the corrected score methodology in problems for which direct calculation of corrected scores is difficult. Its implementation is straightforward in any programming language that allows for complex-value arithmetic.

Following the common practice in a general semiparametric regression, our method is based upon profiling and backfitting for a criterion function. Our theoretical developments seem to be novel even though are supported by previously developed ideas. We demonstrate that our methods are general and include the existing in a literature methods for univariate and multivariate partially linear models as special cases. Moreover, we show that in a partially linear model and logistic model with quadratic effects of X, our method produces mean square errors similar to ones from the semiparametric efficient method. It should be noted that the theory of our method fails for logistic model, however it performs well even in presence of large measurement errors.

Here we have focused on the case where the covariate modeled nonparametrically is univariate. However, the idea of building a semiparametric criterion function using Monte-Carlo corrected scores can be applied to more general problems, e.g., additive models.

Acknowledgments

This work forms part of the first author's Ph.D. dissertation at Texas A&M University.

Appendix A: Conditions and assumptions

Regularity conditions

We require the following conditions.

  • A1. Distribution law of Z is absolutely continuous and has compact support Z, its density fZ(·) is differentiable on Z, the derivative is continuous and infzZfZ(z)>0. Moreover supzZθ0(z)M<. X also has a compact support, X.

  • A2. Mixed partial derivatives r+tBrθtL~(Y,W,B,θ), 0 ≤ r, t ≤ 4, r + t ≤ 4, exist for almost all (Y, W) and E{r+tBrθtL(Y,X,B,θ)2} are bounded.

  • A3. The smallest and the largest eigenvalues of matrix V are bounded away from zero and infinity. Moreover, G(Z) =
    E{(BT,θT)TL(Y,X,B0,θ)θ=θ0(Z)(BT,θT)L(Y,X,B0,θ)θ=θ0(Z)Z}
    possesses a continuous derivative and infzZG(z)>0.
  • A4. k+lzkBlθ0(z,B), 0 ≤ k + 1 ≤ 3, exist and continuous for almost all z and B; and k+lzkBlθ0(z,B)<.

Appendix B: Proof of Theorem 3.1

We start by noting that the loglikelihood is given by

L()=log(σ2)2{YXTγθ(Z)}2(2σ2).

Define ε* = YWTγθ(Z). Then by definition, we have

R()=log(σ2)2(2γTM1m=1MVmVmTγ)(2σ2).

Note that the parameter of interest is β = (γT, σ2)T.

Direct calculations yield

Rβ()=((W+M1ΣmVmVmTγ)σ21(2σ2)+[2γTM1ΣmVmVmTγ](2σ4));Rθ()=σ2;Rββ()=[(WWT+1MΣmVmVmT)σ2[W+1MΣmVmVmTγ]σ4[W+1MΣmVmVmTγ]σ412σ4[2γT1MΣmVmVmTγ]σ6];Rβθ()=(Wσ2σ4);Rθθ()=1σ2.

Using these, we obtain that

θβ()=E{Rβθ()Z}E{Rθθ()Z}={E(WZ)0};E{Rββ()}=[E(XXT)σ2001(2σ4)]E{Rβθ()θβ()T}=[(σ2)1E{WE(WZ)T}000]=[E{XE(XZ)T}σ2000].

Hence, using the definition that S=var{XE(XZ)}, we derive that

V=E{Rββ()+Rβθ()θβ()T}=[Sσ2001(2σ4).]

Also, let K=Rβ+Rθθβ. Then we have from Theorem 2.1 that

n12(β^β0)Normal(0,V1FVT),

where F=var(K).

To complete the proof, we need to derive the asymptotic covariance matrix V1FVT. First we note that

K=(1σ2)({WE(WZ)}+M1m=1MVmVmTγ12+[2γTM1m=1MVmVmTγ](2σ2))=(1σ2)(K1K2).

Hence we have

var(K)=(1σ4)(var(K1)F12F12Tvar(K2)),

where F12=E(K1K2)E(K1)E(K2). Now we derive,

var(K1)=var[{XE(XZ)+U}(UTγ)+M1m=1MVmVmTγ]=var[{XE(XZ)}(UTγ)+UUUTγ+M1m=1MVmVmTγ]=var[{XE(XZ)}(UTγ)]+var(U)+var{(UUTΣu)γ}+var{M1m=1M(VmVmTΣu)γ}=var[{XE(XZ)}(UTγ)]+var(U)+var{(UUTΣu)γ}+M1var{(VmVmTΣu)γ}=Γ+M1var{(VmVmTΣu)γ},

and

(4σ4)var(K2)=var(2)+var(γTM1m=1MVmVmTγ)=E{(UTγ)4}(σ2+γTΣuγ)2+M2m=1Mvar{γT(VmVmTΣu)γ}=E{(UTγ)2(σ2+γTΣuγ)}2+M1var{γT(VmVmTΣu)γ}=τ2+M1var{γT(VmVmTΣu)γ}.

Finally, we derive

E(K1)=E[{WE(WZ)}+M1m=1MVmVmTγ]=Σuγ+Σuγ=0;E(K2)=E(12+[2γTM1m=1MVmVmTγ](2σ2))=0;E(K1K2)=E([{WE(WZ)}+M1m=1MVmVmTγ]×[12+[2γTM1m=1MVmVmTγ](2σ2)]).

Note that ε* = εUTγ is independent of X and Z. Hence using W = X + U, we observe

E({WE(WZ)}3)=3σ2ΣuγE{U(UTγ)3};E[{WE(WZ)}γTM1m=1MVmVmTγ]=ΣuγγTΣuγ;E[M1m=1MVmVmTγ2]=Σuγσ2+ΣuγγTΣuγ;E[{M1m=1MVmVmTγ}{γTM1l=1MVlVlTγ}]=E[M2m=1Ml=1MVmVmTγγTVlVlTγ]=E[M2m=1MVmVmTγγTVmVmTγ]+E[M2m=1Mlm=1MVlVlTγγTVlVlTγ]=M1E{Vm(VmTγ)3}+M1(M1)ΣuγγTΣuγ.

Hence we see that

E(K1K2)=Σuγ23Σuγ2{E{U(UTγ)3ΣuγγTΣuγ}(2σ2)(12)Σuγ+(12)Σuγ+ΣuγγTΣuγ(2σ2)M1E{Vm(VmTγ)3}(2σ2)M1(M1)ΣuγγTΣuγ(2σ2)=Σuγ[E{U(UTγ)3}ΣuγγTΣuγ](2σ2)M1E{Vm(VmTγ)3}(2σ2)+{1M1(M1)}ΣuγγTΣuγ(2σ2).

Finally, we obtain that the asymptotic covariance matrix is given by

V1FVT=[S1ΓST+R112σ2S1C+R12τ2+R22],

where

R11=M1S1cov{(VmVmTΣu)γ}ST;R12=2σ2S1[M1E{Vm(VmTγ)3}{1M1(M1)}ΣuγγTΣuγ];R22=M1var{γT(VmVmTΣu)γ}.

Now the result follows from Theorem 2.1.

Appendix C: Proof of Remark 3.3

Recall that, for a fixed M, Monte Carlo corrected criterion function is given by

RM()=M1m=1MRe[L{Y,W~m,B,θ(Z)}]=(12)log{det(Σ1)}(12){YWβθ(Z)}TΣ1{YWβθ(Z)}+(12)βT(M1m=1MVmTΣ1Vm)β.

We observe that as M → ∞, RM()=R()+Op(M12), where

R()=[log{det(Σ1)}{YTβθ(Z)}TΣ1{YTβθ(Z)}+βTE(VmTΣ1Vm)β]2.

Note that R() is exactly the same criterion function as in Lin and Carroll[12] (see their equation (22), p. 81). Hence as M → ∞, the estimates based on R() are given as follows: given the current estimates, B^cur=(β^cur,Σ^,cur), the new estimates are given by

β^new=[n1i=1n{WiTΣ^,cur1WiE(VmTΣ^,cur1Vm)}]1×n1i=1nWiTΣ^,cur1{Yiθ^(Zi,B^cur)};Σ^,new=n1i=1n[{YiWiβ^curθ^(Zi,B^cur)}{YiWiβ^curθ^(Zi,B^cur)}TE(Vmβ^curβ^curTVmT)].

Again, as M → ∞, we note that the estimating equation for β and Σε are same as Lin and Carroll [12] (see their equation (23), p. 81). Hence the result.

Contributor Information

Arnab Maity, Department of Statistics, North Carolina State University, Raleigh, North Carolina 27695, U.S.A. amaity@ncsu.edu.

Tatiyana V. Apanasovich, Department of Biostatistics, Thomas Jefferson University, Philadelphia, Pennsylvania 19118, U.S.A. tatiyana.apanasovich@jefferson.edu.

References

  • 1.Al-Abood AM, Young DH. The power of approximate tests for the regression coefficients in a gamma regression model. IEEE Transactions On Reliability. 1986;R-35:216–220. [Google Scholar]
  • 2.Apanasovich TV, Carroll RJ, Maity A. SIMEX and standard error estimation in semiparametric measurement error models. Electronic Journal of Statistics. 2009;3:318–348. doi: 10.1214/08-EJS341. MR2497157. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Cameron AC, Trivedi PK. Regression analysis of count data. Cambridge University Press; Cambridge: 1998. MR1648274. [Google Scholar]
  • 4.Carroll RJ, Ruppert D, Crainiceanu C, Stefanski LA. Measurement Error in Nonlinear Models: A Modern Perspective. Second Edition. CRC Press; London: 2006. MR2243417. [Google Scholar]
  • 5.Cook JR, Stefanski LA. Simulation-extrapolation estimation in parametric measurement error models. Journal of the American Statistical Association. 1994;89:1314–1328. [Google Scholar]
  • 6.Eckert RS, Carroll RJ, Wang N. Transformations to additivity in measurement error models. Biometrics. 1997;53:262–272. MR1450184. [PubMed] [Google Scholar]
  • 7.Kerber RL, Till JE, Simon SL, Lyon JL, Thomas DC, Preston-Martin S, Rollison ML, Lloyd RD, Stevens W. A cohort study of thyroid disease in relation to fallout from nuclear weapons testing. Journal of the American Medical Association. 1993;270:2076–2083. [PubMed] [Google Scholar]
  • 8.Li Y, Guolo A, Owen Hoffman F, Carroll RJ. Shared Uncertainty in Measurement Error Problems, with Application to Nevada Test Site Fallout Data. Biometrics. 2007;63:1226–36. doi: 10.1111/j.1541-0420.2007.00810.x. MR2414601. [DOI] [PubMed] [Google Scholar]
  • 9.Liang H, Haerdle W, Carroll RJ. Estimation in a partially linear error-in-variables model. Annals of Statistics. 1999;27:1519–1535. MR1742498. [Google Scholar]
  • 10.Liang H, Ren HB. Generalized partially linear measurement error models. Journal of Computational and Graphical Statistics. 2005;14:237–250. MR2137900. [Google Scholar]
  • 11.Lin X, Wang N, Welsh A, Carroll RJ. Equivalent kernels of smoothing splines in nonparametric regression for clustered data. Biometrika. 2004;91:177–193. MR2050468. [Google Scholar]
  • 12.Lin X, Carroll RJ. Semiparametric estimation in general repeated measures problems. Journal of the Royal Statistical Society, Series B. 2006;68:69–88. MR2212575. [Google Scholar]
  • 13.Lubin JH, Schafer DW, Ron E, Stovall M, Carroll RJ. A reanalysis of thyroid neoplasms in the Israeli tinea capitis study accounting for dose uncertainties. Radiation Research. 2004;161:359–368. doi: 10.1667/rr3135. [DOI] [PubMed] [Google Scholar]
  • 14.Lyon JL, Alder SC, Stone MB, Scholl A, Reading JC, Holubkov R, Sheng X, White GL, Hegmann KT, Anspaugh L, Hoffman FO, Simon SL, Thomas B, Carroll RJ, Meikle AW. Thyroid disease associated with exposure to the Nevada Test Site radiation: a reevaluation based on corrected dosimetry and examination data. Epidemiology. 2006;17:604–614. doi: 10.1097/01.ede.0000240540.79983.7f. [DOI] [PubMed] [Google Scholar]
  • 15.Ma Y, Carroll RJ. Locally efficient estimators for semi-parametric models with measurement error. Journal of the American Statistical Association. 2006;101:1465–1474. MR2279472. [Google Scholar]
  • 16.Mallick B, Hoffman FO, Carroll RJ. Semiparametric regression modeling with mixtures of Berkson and classical error, with application to fallout from the Nevada Test Site. Biometrics. 2002;58:13–20. doi: 10.1111/j.0006-341x.2002.00013.x. MR1891038. [DOI] [PubMed] [Google Scholar]
  • 17.Nakamura T. Corrected score functions for error-in-variable models: methodology and application to generalized linear models. Biometrika. 1990;77:127–137. MR1049414. [Google Scholar]
  • 18.Novick JS, Stefanski LA. Corrected score estimation via complex variable simulation extrapolation. Journal of the American Statistical Association. 2002;97:472–481. MR1941464. [Google Scholar]
  • 19.Pierce DA, Kellerer A. Adjusting for covariate errors with nonparametric assessment of the true covariate distribution. Biometrika. 2004;91:863–876. MR2126038. [Google Scholar]
  • 20.Reeves GK, Cox DR, Darby SC, Whitley E. Some aspects of measurement error in explanatory variables for continuous and binary regression models. Statistics in Medicine. 1998;17:2157–2177. doi: 10.1002/(sici)1097-0258(19981015)17:19<2157::aid-sim916>3.0.co;2-f. [DOI] [PubMed] [Google Scholar]
  • 21.Singpurwalla ND. A problem in accelerated Life testing. Journal of the American Statistical Association. 1971;66:841–845. [Google Scholar]
  • 22.Schafer DW, Gilbert ES. Some statistical implications of does uncertainty in radiation dose-response analyses. Radiation Research. 2006;166:303–312. doi: 10.1667/RR3358.1. [DOI] [PubMed] [Google Scholar]
  • 23.Schafer DW, Lubin JH, Ron E, Stovall M, Carroll RJ. Thyroid cancer following scalp irradiation: a reanalysis accounting for uncertainty in dosimetry. Biometrics. 2001;57:689–697. MR1859805. [PubMed] [Google Scholar]
  • 24.Simon SL, Till JE, Lloyd RD, Kerber RL, Thomas DC, Preston–Martin S, Lyon JL, Stevens W. The Utah Leukemia case–control study: dosimetry methodology and results. Health Physics. 1995;68:460–471. doi: 10.1097/00004032-199504000-00003. [DOI] [PubMed] [Google Scholar]
  • 25.Simon SL, Anspaugh LR, Hoffman FO, et al. 2004 update of dosimetry for the Utah Thyroid Cohort Study. Radiation Research. 2006;165:208–222. doi: 10.1667/rr3483.1. [DOI] [PubMed] [Google Scholar]
  • 26.Stefanski LA, Cook JR. Simulation-Extrapolation: the measurement error jackknife. Journal of the American Statistical Association. 1995;90:1247–1256. MR1379467. [Google Scholar]
  • 27.Stevens W, Till JE, Thomas DC, et al. Assessment of leukemia and thyroid disease in relation to fallout in Utah: report of a cohort study of thyroid disease and radioactive fallout from the Nevada test site. University of Utah; 1992. [Google Scholar]
  • 28.Stram DO, Kopecky KJ. Power and uncertainty analysis of epidemiological studies of radiation-related disease risk in which dose estimates are based on a complex dosimetry system: some observations. Radiation Research. 2003;160:408–417. doi: 10.1667/3046. [DOI] [PubMed] [Google Scholar]
  • 29.Tsiatis AA, Ma Y. Locally efficient semiparametric estimators for functional measurement error models. Biometrika. 2004;91:835–848. MR2126036. [Google Scholar]
  • 30.Zhu L, Cui H. A semiparametric regression model with errors in variables. Scan. J. Statist. 2003;30:429–442. MR1983135. [Google Scholar]

RESOURCES