Summary
We study the heteroscedastic partially linear single-index model with an unspecified error variance function, which allows for high dimensional covariates in both the linear and the single-index components of the mean function. We propose a class of consistent estimators of the parameters by using a proper weighting strategy. An interesting finding is that the linearity condition which is widely assumed in the dimension reduction literature is not necessary for methodological or theoretical development: it contributes only to the simplification of non-optimal consistent estimation. We also find that the performance of the usual weighted least square type of estimators deteriorates when the non-parametric component is badly estimated. However, estimators in our family automatically provide protection against such deterioration, in that the consistency can be achieved even if the baseline non-parametric function is completely misspecified. We further show that the most efficient estimator is a member of this family and can be easily obtained by using non-parametric estimation. Properties of the estimators proposed are presented through theoretical illustration and numerical simulations. An example on gender discrimination is used to demonstrate and to compare the practical performance of the estimators.
Keywords: Dimension reduction, Double robustness, Linearity condition, Semiparametric efficiency bound, Single index model
1. Introduction
We study the partially linear single-index model
(1) |
where and are random variables with possibly high dimensions dγ and dβ respectively. We do not assume any parametric form for the conditional distribution of ε. In particular, we allow ε to be heteroscedastic. For identifiability, we assume that the first component of β is 1, and we use z−1 to denote the last dβ − 1 components of z. Let θ = (γT, βT)T and d = dβ + dγ − 1. Without loss of generality, we assume that E(zi) = 0 in what follows. The interest is usually in estimating θ, whereas all the other unknown components of model (1), such as the unspecified function g(·) and the unknown error distribution, are termed nuisance parameters.
Model (1) was proposed by Carroll et al. (1997). It is a natural extension of the partially linear model (Engle et al., 1986) in the situation when multiple covariates need to be included non-parametrically. It can also be viewed as an extension of the single-index model or as a special case of the multi-index model. See Xia et al. (1999) and Xia (2008) and the references therein for a discussion on the links. The parametric component xTγ provides a simple summary of covariate effects which are of the main scientific interest, the index zTβ enables us to simplify the treatment of the multiple auxiliary variables which may have high dimension and the smooth baseline component g (·) enriches model flexibility.
Estimation of model (1) has been studied in Carroll et al. (1997), Yu and Ruppert (2002) and Xia and Härdle (2006), in which consistently estimating g(·) is mandatory during the estimation process to ensure consistency of the estimator of θ. An intriguing finding here is that the consistency of the parameter estimator for θ can be achieved without estimating g(·). In addition, although Carroll et al. (1997), Yu and Ruppert (2002) and Xia and Härdle (2006) did not explicitly make an equal variance assumption for ɛ, they did not account for the heteroscedasticity of model (1) either. Consequently, none of these estimators are efficient. One might think that an inverse-to-variance weighting scheme can lead to efficiency. However, we discover that it is not always so. In fact, a formal study on the efficient estimator for model (1) has only been conducted in the homoscedastic error case (Wang et al., 2010), where an efficient estimator is constructed when the covariates are assumed elliptical. In this paper we construct an estimator that reaches the optimal semiparametric efficiency bound under heteroscedasticity and without distributional assumptions on the covariates.
When dβ is potentially large, a routinely assumed condition in the dimension reduction literature (Li, 1991) is the linearity condition
(2) |
Here and . Wang et al. (2010) also assumed this condition to facilitate their investigation on the consistency of the estimators in the homoscedastic partially linear single-index model. We find that this condition is unnecessary in our methodological development and it does not contribute to an improved efficiency of estimation. The only utility of condition (2) is in computation, in that it simplifies the construction of a non-optimal consistent estimator. We emphasize here that the linearity condition (2) does not lead to any simplification of the efficient estimator.
In summary, we propose a general class of estimating equations to estimate the parameters β and γ in model (1) with a heteroscedastic error. The estimating equations have several robustness properties, in the sense that they allow us to obtain a consistent estimator for β and γ even when several components, including g(·), are not consistently estimated or are misspecified. Consistent estimation of β and γ without consistently estimating g(·), whether with a homoscedastic or heteroscedastic error, has not been discovered before in the literature. In addition, the consistency holds without any distributional assumption on the covariates. In particular, the linearity condition (2) is not required although it can simplify some of the computations when it is true. Relaxing the linearity condition under this framework has also never been achieved before as far as we are aware. Note that here all the robustness properties hold in the sense that consistency of and is preserved under misspecification of several components of the model. If practical interest is in estimating some of these model components, then we can still estimate them. If we estimate all the components involved in the estimating equations consistently, then the resulting estimator for β and γ is guaranteed to be efficient. In other words, the optimal member of the general class requires consistent estimation of the various components, including g(·) and g′(·). To the best of our knowledge, this is the first time that efficient estimation in model (1), with possibly high dimensional covariates, has been proposed.
The rest of the paper is organized as follows. In Section 2, we examine a class of weighted estimators and propose a general class of consistent estimators for heteroscedastic model (1) that has a double-robustness property. We identify a member of this class that is easy to compute under condition (2). The efficient estimator is derived in Section 3 where a simple algorithm is provided and g(·) is also estimated as a by-product. The performance of the estimators proposed is illustrated in simulation studies in Section 4 and in real data in Section 5. We discuss some extensions of our method in Section 6.
2. A general class of consistent estimators
Our goal in this section is to present a general class of consistent estimators for θ that has a double-robustness property. Double robustness has also been discovered in other contexts. See, for example, Robins and Rotnitzky (2001) and Tan (2010). We explore this double-robustness property and derive a simplified estimation procedure in the presence of the linearity condition (2). We also reveal a limitation of the standard weighting method in handling the heteroscedastic error variance and hence provide an explanation why heteroscedastic errors are usually not accounted for even if it is acknowledged.
To ease the presentation of consistency considerations among different estimators, we present each estimator as a solution to an estimating equation. The consistency of the estimator is studied through the consistency of the estimating equation. We consider the following class of estimating equations:
(3) |
where f(zTβ) is a given function of zTβ that may or may not equal denotes a function of zTβ which may or may not be the true and a(·, ·) is an arbitrary function of x and z. It is easily verified that, at the true parameter values of β and γ, equation (3) has mean 0. We can view equation (3) as a sample version of the population mean. We assume that, when n → ∞, equation (3) has a unique solution in the neighbourhood of the true parameter values of β and γ. Because is the ith error ɛi, we can rewrite equation (3) as
It is now easy to recognize that the above estimating equation has a double-robustness property. On the one hand, if f(·) = g(·), then equation (3) is consistent whether or not . In an extreme case, we can simply set . This choice combined with the choice or yields
or
which are the most familiar forms of estimator. On the other hand, if , then equation (3) is also consistent whether or not f(·) = g(·). In an extreme case, we can set . Making the natural choice of , we obtain a consistent estimator from the estimating equation
Of course, the two choices of g(·) or E(·|zTβ) can also be combined as we now illustrate. For example, we can directly set g(·) = 0 in one part of the estimating equations and still retain the two different choices of either g(·) or E(·|zTβ) in the other part of the estimating equations. Specifically, we can consider estimating equations of the form
(4) |
In this construction, the first dγ-dimensional equation corresponds to choosing the first dγ components of a(xi, zi) in equation (3) to be xi, and the second dβ-dimensional equation corresponds to choosing f = 0 and the second dβ components of a(xi, zi) in equation (3) to be zi and . The construction (4) has a fully fixed form in the second dβ equations, while still leaving the flexibility of using either f(·) = g(·) or in the first dγ equations. The above estimating equation (4) still offers consistent estimators for both γ and β.
Next we examine the role that the linearity condition (2) plays in the estimating equation. We use the estimating equation (4) as an illustrative example. When condition (2) is true, expression (4) becomes
(5) |
We shall see that expression (5) leads to the simplest consistent estimator for model (1).
To solve equations (5), we rewrite the second equation as
The quantity in the curly brackets of this equation is a scalar. Hence we can ‘profile out’ β by setting
Here, Ê(·) and are the sample mean and covariance estimates, and the constant c(γ) ≠ 0 is a normalizing constant to ensure that the first component of is 1. Specifically, c(γ) is the inverse of the first component of . Two possibilities are available in obtaining corresponding to the two choices of specifying g(·) or E(x|zTβ) If we decide to specify E(x|zTβ), we can set f(zTβ)= 0 and plug in to solve for γ from
If we decide to specify g(·), we can set Ẽ(xi|ziTβ)= 0 and solve for γ from
If we do not have sufficient information or are reluctant to specify any one of the two functions, we can estimate one of them or both by using, for example, the standard kernel non-parametric regression or local linear approximation, provided that sufficient smoothness or regularity conditions hold.
To take advantage of the linearity condition (2), we have chosen to set in the second of equations (5) because of its simplicity. In general, f(·) can be any function of . However, f(·) needs to satisfy
(6) |
otherwise degeneration will occur and the second of equations (5) (corresponding to a special f(·) = 0) will not yield a solution. We give a detailed account of this phenomenon in the on-line supplementary information for this paper.
The estimating equation class (3) does not take into account the error variance form. A natural correction for this is to include inverse variance weights to improve the efficiency of estimation. Denote wi = w(xi, zi) = var (ɛi|xi, zi)−1. The weighted estimating equation class is then
(7) |
However, class (7) lacks some of the nice properties of the original family (3). It can be easily verified that, even if wi can be consistently estimated or is completely known, the consistency of and now always requires a consistent estimation of g(·). Thus, not only the double-robustness property is lost, but also the simplification that is contributed by the linearity condition (2) becomes limited. Similar controversy is discovered for the partially linear model in Ma et al. (2006), and we suspect that this is why a standard inverse variance weighting scheme is not pursued in the partially linear single-index model in the literature, even when there is heteroscedasticity.
3. Efficient estimator via proper weighting
To improve the estimation variance through incorporating the weights wi while retaining the nice properties of expression (3), we perform a slight modification of expression (7) and propose
(8) |
For simplicity in expression (8), we directly set , which is the simplest form to use in practice. Here b(·) can be an arbitrary function and is included here for increased flexibility.
Similarly to expression (3), expression (8) provides a consistent estimator under several misspecifications. For example, we can use f(·) that is not a consistent estimate of g(·); as long as Ẽ(·|zTβ) = E(·|zTβ), expression (8) yields consistent estimators for β and γ. We can also choose not to estimate E(·|zTβ) carefully, by using some parametric model or even simply setting it to zero; as long as f(·) = g(·), expression (8) is still valid. Finally, the weights wi can also be misspecified; as long as one of g(·) and E(·|zTβ) is correctly specified or both are consistently estimated, we still have consistency for β and γ. If we adopt a weight w that depends on x and z only through zTβ, expression (8) becomes a member of class (7). This indicates that the weighting scheme in expression (7) is valid only in some limited situations.
In addition to the robustness property, expression (8) also provides an efficient estimating equation when we have the correct weights, use a consistent estimator of g as the f-function, use a consistent estimator of g′ as the b-function and at the same time estimate the conditional expectation E(·|zTβ) consistently. To see this, we first derive the semiparametric efficient score function for model (1). The semiparametric efficient score is defined as the projection of the score vector onto the orthogonal complement of the nuisance tangent space; for details, see Bickel et al. (1993) and Tsiatis (2006).
Proposition 1
Assume that the conditional probability density function of ɛ given (x, z), pɛ. (ɛ|x, z), is differentiable with respect to ɛ and that 0 < E(ɛ2|x,z) < ∞ almost everywhere. Denote w = w(x, z) = E(ɛ2|x, z)−1. The semiparametric efficient score is
The expression for Seff suggests that, by carefully choosing w,f and b in expression (8), we should obtain a semiparametric efficient as the solution to expression (8). The proof of proposition 1 is given in the on-line supplementary document.
Obtaining w or the error variance var. (ɛ|x, z) through a non-parametric regression of the residuals on the covariates (x, z) is itself a high dimensional problem which suffers from the curse of dimensionality. To focus our presentation on the main concepts, we assume that there is a low dimensional variable ξ = ξ(x, z) such that var. (ɛ|x, z) = var (ɛ|ξ) and ξ has a known form. For example, ξ can be zTβ, indicating that the error variance depends on the covariates through zTβ only. ξ can also be xTγ, indicating that the error variance depends on the covariates through xTγ only. It can also be a combination of these two or can have any other form. In practice, a reasonable approximation of ξ can be obtained via standard procedures to model error variance, using the residuals from an initial estimation of the model. We shall concentrate on the case where ξ is univariate. It is worth noting that this assumption is often made to simplify the estimation of weights. The assumption can be weakened to include intermediate multivariate models as components, e.g. with additive structures, so that the univariate convergence rates remain achievable and that the variance structure is still very flexible.
If we use non-parametrically estimated g, g′, w, E(w|zTβ), E(wx|zTβ) and E(wz−1|zTβ), the estimator that is obtained from expression (8) can be written as the solution to
(9) |
In expression (9), the quantities with circumflexes can be estimated via kernel estimation. To be precise, let K be a kernel function and Kh(·) = h−1K(·/h). Recall also the notation for w(xi,zi) and ξi = ξ(xi, zi) defined in Section 3. For bandwidths h1, h2 and h3, we set
We remark that we use the same bandwidth in the last three quantities, because a same bandwidth already suffices to guarantee the efficiency of the estimate that is obtained from expression (9); see the proof of proposition 2. The algorithm to solve expression (9) is as follows.
Step 1: we use the estimation procedure that was described in Section 2 to obtain an initial estimator .
Step 2: obtain ĝ and ĝ′ by using non-parametric regression of on described above. Denote .
Step 3: obtain the residual . Estimate σ2(x,z) according to the variance model by using the data {ξ(xi, zi), ei2} and the initial parameter values if necessary. Set .
Step 4: obtain and via non-parametric regression by using the data . Set .
Step 5: obtain via non-parametric regression by using the data . Set .
- Step 6: update the estimate for γ from
where the Ê are simply the sample averages. Solve for β from
Here ĝ(·) and ĝ′(·) are treated as known functions and hence do not need to be updated, and are the previously obtained values.
In this algorithm, each time that a kernel estimation procedure needs to be performed, if a new bandwidth needs to be selected, we can simply use the traditional cross-validation procedure. In our experience, the performance of the estimation is very insensitive to the choice of bandwidth, which agrees with the common observation in many semiparametric problems and is also explained by the wide range of possible choices of bandwidth in the following regularity condition 6. In particular, we do not need to undersmooth anywhere. This is a by-product of the double-robustness and will be justified in proposition 2. It is a property that is also shared by the profile likelihood method (Lin and Carroll, 2006). We also emphasize that the only estimating equation that needs to be solved is in step 6. Since this step does not involve any non-parametric procedures, it is a simple parametric estimating equation and can be solved via for example the Newton–Raphson procedure. In finite sample situations, it can happen that no roots are available for an estimating equation. In this case, instead of solving for zero, one can aim at minimizing the square of the estimating equation. One may note that, although the linearity condition (2) can simplify the initial estimation , it will not contribute to the subsequent operations at all. The way in which the weights wi enter expression (9) requires a different type of linearity condition to achieve simplification, namely a weighted linearity condition of the form
(10) |
We can easily see that, when w is a function of zTβ or when w is uncorrelated with z conditional on zTβ, equation (10) is equivalent to condition (2). This indicates that equation (10) is not much stronger than the usual linearity condition (2) and may hold in various situations. If we are willing to make this assumption, then the second equation in expression (9) is simplified and it suffices to solve
(11) |
Therefore, the above algorithm can be simplified through skipping step 5, and replacing the last estimating equation in step 6 with equation (11).
We now list a set of regularity conditions that are sufficient for our main proposition to hold. These conditions are not the weakest possible, but they facilitate our technical derivations.
Condition 1. The second moments of all covariates are finite, i.e. and . There is a positive constant δ such that E(ɛ4+δ|x, z) < ∞.
Condition 2. There are v(·), ξ = ξ(x, z) and positive constants c1 and c2, such that E(ɛ2|x, z) = v(ξ), 0 < c1 < v(·) < c2 < ∞, and E(Xi2|ξ) < ∞.
Condition 3. The functions E(x|zTβ), E(z|zTβ), E(w|zTβ), E(wz|zTβ) and E(wx|zTβ) are twice continuously differentiable with finite derivatives. As a function of (x, z), ξ(x, z) is three times continuously differentiable with finite derivatives. The functions g(zTβ) and v(ξ) are four times continuously differentiable with finite derivatives.
Condition 4. Assume that the random variables ξ and zTβ have densities fξ(ξ) and such that fξ(ξ) and fztβ(ztβ) are twice continuously differentiable with finite derivatives, satisfying 0 < inf fξ(ξ) ≤ sup fξ(ξ) < ∞ and 0 < inf fzTβ(zTβ) ≤ sup fzTβ(zTβ) < ∞.
Condition 5. The kernel function K is symmetric, and its derivative K′ is continuous in compact support [−1, 1].
Condition 6. The bandwidths hi that are used in the kernel estimators satisfy log2(n)/(nhi) →0 for i = 1,2,3. In addition, and and log4(n)/(nh1hi) → 0 for i = 2,3, and h2 = O(n−1/5) and h3 = O(n−1/5).
Proposition 2
Assume that and solve expression (9) or expression (9) with the second equation replaced by equation (11). Then, under the above regularity conditions, and reach the optimal semiparametric efficiency bound. In particular, when n → ∞,
in distribution, where
and
The proof of proposition 2 is given in the on-line supplementary document. The above results suggest that, as long as regularity conditions 1–6 hold, the non-parametric estimator does not cause any loss of efficiency. In other words, is asymptotically equivalent to the solution to expression (9) with known g, g′, w, E(w|zTβ), E(wx|zTβ) and E(wz−1|zTβ); hence the optimal semiparametric efficiency is practically achieved. We can also see that the presence of the linearity condition (2) or the corresponding weighted form (10) does not improve efficiency. Hence the benefit of these conditions is merely computational, in that, if condition (2) or (10) holds, then the corresponding non-parametric regression of z or w or wz on zTβ can be avoided. In addition, the classical linearity condition (2) alone cannot contribute to computational simplification of the efficient estimator. Therefore, although the linearity condition is routinely assumed in the dimension reduction literature, it is not necessary in our case. In the special case when the error is homoscedastic and hence w is a constant, our resulting efficient estimation variance is the same as that given in Carroll et al. (1997). However, even in the homoscedastic error case, the result in proposition 2 is stronger than the efficient result in Carroll et al. (1997). This is because their efficient result was established only for the normal error case, whereas our result is not restricted to any special distributional form.
4. Simulations
In this section, we conduct simulation studies to evaluate the performance of various estimation procedures. We generate X1 from a uniform distribution U(0, 1), X2 from a binomial distribution with success probability 0.5 and X3 from a Poisson distribution with parameter 2. We generate other covariates as follows.
In case 1, we generate (X4,…, Xdγ, Z1,…, Zdβ)T from a multivariate normal distribution with mean 0 and variance–covariance matrix (σij)(d−2) × (d−2) where σij = 0.5|i−j| and d = dγ + dβ − 1. In this case, the covariates z = (Z1,…, Zdβ)T satisfy the linearity condition (2).
In case 2, we replace Z5 in case 1 by so that the linearity condition (2) is violated. All other covariates remain unchanged.
With the covariate vectors x = (X1,…, Xdγ)T and z = (Z1,…, Zdβ)T, we generate Y from a normal distribution with mean xTγ + exp(zTβ) and variance function |zTβ| where γ = (2,1,−1, 0:5,0/T and β = (4,1,1,1,1)T/√20. When dγ > 5 or dβ > 5, the additional components of γ or β are 0.
We carry out a group of simulations to examine the finite sample performance of the estimators corresponding to the following estimating equations (a)–(e):
(a) |
(b) |
(c) |
(d) |
(e) |
We choose these estimating equations on the basis of the following considerations. The estimator that is obtained from (a) corresponds to the ordinary least squares estimator in the classical linear model, which is asymptotically biased without any condition; estimators obtained from (b) are consistent only when condition (2) is satisfied. Owing to the double-robustness property, estimator (c) is always consistent. Similarly, estimator (d) offers consistent estimation, and the resulting estimator is efficient if condition (2) is satisfied; estimator (e) offers consistent and efficient estimates even when condition (2) is violated. In addition, for estimators (b)–(e), we also experimented with their corresponding oracle version, where ĝ and ĝ′ are replaced by the true functions g and g′. In all the non-parametric regression procedures, the Epanechnikov kernel function is used, and the bandwidth is always set to be , where σ = 1:26 is the standard deviation of the regression covariate when z satisfies the linearity condition (2), and 2.2152 is the adjustment from the normal kernel to the Epanechnikov kernel. We used an undersmoothed bandwidth rate of n−1/3 to ensure the consistency of the estimating equations (c) and (d). Although we could have implemented an optimal bandwidth for both (b) and (e), we decided not to do so for fair comparison. Similarly to the optimal bandwidth, this undersmoothed bandwidth still suffices to ensure the consistency of (b) and the efficiency of estimator (e).
We now evaluate the accuracy of estimation of β and γ on the basis of 1000 data replications. The bias (‘bias’) and the standard deviation (‘std’) of the estimates of γ and a subset of β are summarized in Table 1 for case 1 where condition (2) is satisfied and (n, dγ, dβ) = (250, 5, 5). We can clearly see that estimator (a) has very large biases, whereas the biases for all other situations are much smaller. In terms of the estimation variability, (d) and (e) produce estimates with smaller variance than all other estimating equations, which also complies with our expectation since (d) and (e) are efficient. In addition, we can see that the estimating equations (b)–(e) offer competitive or sometimes even better estimators than their corresponding oracle versions which use the true link function g and its first derivative g′. This is in line with our theoretical finding in proposition 2, which indicates that non-parametric smoothing does not affect the first-order asymptotic property of the estimator.
Table 1.
Estimator | Parameter | γ1 (%) | γ2 (%) | γ3 (%) | γ4 (%) | γ5 (%) | β1(%) | β2(%) | β5(%) |
---|---|---|---|---|---|---|---|---|---|
(a) | bias | 189.43 | 65.28 | 31.49 | −0.12 | 0.23 | −2.82 | 0.33 | −0.54 |
std | 54.11 | 42.83 | 12.98 | 26.31 | 31.38 | 5.39 | 12.13 | 9.93 | |
mse | 38811.01 | 6095.64 | 1160.10 | 692.10 | 984.88 | 37.03 | 147.34 | 98.84 | |
(b) | bias | 1.48 | 0.07 | −0.15 | −0.01 | 2.64 | −2.81 | 0.39 | −0.41 |
std | 26.03 | 14.69 | 5.22 | 8.47 | 11.22 | 5.31 | 11.81 | 9.80 | |
mse | 679.84 | 215.90 | 27.26 | 71.68 | 132.84 | 36.07 | 139.51 | 96.15 | |
Oracle | bias | 1.50 | 0.10 | −0.13 | −0.00 | 2.65 | −2.81 | 0.39 | −0.41 |
std | 26.23 | 14.81 | 5.23 | 8.52 | 11.27 | 5.31 | 11.81 | 9.80 | |
mse | 690.25 | 219.26 | 27.32 | 72.58 | 133.94 | 36.06 | 139.60 | 96.28 | |
(c) | bias | 1.65 | 0.08 | −0.16 | −0.15 | −0.20 | −0.13 | −0.60 | −0.44 |
std | 22.20 | 13.33 | 4.68 | 7.75 | 8.93 | 1.94 | 5.07 | 4.20 | |
mse | 495.49 | 177.79 | 21.89 | 60.09 | 79.71 | 3.76 | 26.05 | 17.80 | |
Oracle | bias | 1.57 | 0.07 | −0.15 | −0.16 | −0.18 | −0.15 | −0.54 | −0.41 |
std | 22.11 | 13.28 | 4.67 | 7.76 | 8.88 | 2.02 | 5.05 | 4.30 | |
mse | 491.48 | 176.30 | 21.88 | 60.24 | 78.83 | 4.10 | 25.75 | 18.66 | |
(d) | bias | 1.74 | 0.11 | −0.17 | −0.20 | 0.38 | −0.30 | −0.05 | −0.05 |
std | 21.66 | 12.96 | 4.55 | 7.64 | 8.28 | 1.29 | 3.80 | 3.23 | |
mse | 472.34 | 167.91 | 20.72 | 58.35 | 68.64 | 1.74 | 14.43 | 10.46 | |
Oracle | bias | 1.15 | 0.09 | −0.04 | −0.08 | 3.62 | −4.49 | 0.28 | −1.16 |
std | 22.51 | 12.90 | 4.65 | 7.62 | 10.00 | 7.15 | 15.66 | 12.91 | |
mse | 508.23 | 166.41 | 21.59 | 58.02 | 113.19 | 71.27 | 245.36 | 168.01 | |
(e) | bias | 1.71 | 0.11 | −0.17 | −0.19 | 0.36 | −0.29 | −0.10 | −0.05 |
std | 21.61 | 12.95 | 4.54 | 7.65 | 8.29 | 1.31 | 3.86 | 3.28 | |
mse | 469.85 | 167.77 | 20.68 | 58.54 | 68.85 | 1.79 | 14.87 | 10.75 | |
Oracle | bias | 1.34 | 0.07 | −0.09 | −0.22 | 0.33 | −0.28 | −0.02 | −0.06 |
std | 19.09 | 11.33 | 3.96 | 6.72 | 7.38 | 1.34 | 3.82 | 3.33 | |
mse | 366.31 | 128.47 | 15.67 | 45.25 | 54.62 | 1.86 | 14.61 | 11.12 |
The parallel simulation results are reported in Table 2 for case 2 where condition (2) is violated and (n, dγ, dβ) = (250,5,5). Again, estimator (a) has a non-ignorable bias. Since condition (2) is violated, estimator (b) also shows a large bias, and we can see that estimator (c) outperforms (b), and (e) is superior to (d) in terms of both the bias and the variance. This is especially clear in estimating β. This indicates that the linearity condition (2) needs to be verified. Blindly using a non-verified linearity condition will lead to biased results. The double-robustness form helps to achieve small bias without all the non-parametric regressions. It also yields superior finite sample performance when all the non-parametric regressions are performed.
Table 2.
Estimator | Parameter | γ1 (%) | γ2 (%) | γ3 (%) | γ4 (%) | γ5 (%) | β1(%) | β2(%) | β5(%) |
---|---|---|---|---|---|---|---|---|---|
(a) | bias | 190.81 | 65.45 | 31.65 | −0.90 | 1.19 | −16.89 | −3.37 | 24.65 |
std | 75.56 | 50.20 | 15.79 | 34.89 | 39.33 | 13.17 | 11.26 | 14.94 | |
mse | 42118.11 | 6804.15 | 1250.82 | 1218.09 | 1548.57 | 458.83 | 138.20 | 831.06 | |
(b) | bias | 1.04 | 0.25 | 0.12 | −0.85 | 27.03 | −19.49 | −2.89 | 26.63 |
std | 37.22 | 21.01 | 7.90 | 12.96 | 21.30 | 14.36 | 11.34 | 15.32 | |
mse | 1386.34 | 441.56 | 62.46 | 168.78 | 1184.42 | 586.24 | 136.86 | 943.95 | |
Oracle | bias | 1.76 | 0.40 | 0.28 | −0.98 | 27.42 | −19.53 | −2.90 | 26.64 |
std | 39.78 | 23.71 | 9.17 | 14.03 | 22.32 | 14.41 | 11.34 | 15.33 | |
mse | 1585.19 | 562.29 | 84.12 | 197.75 | 1249.73 | 588.90 | 137.07 | 944.92 | |
(c) | bias | 1.85 | 0.38 | −0.09 | 0.06 | −0.21 | −0.24 | −0.41 | −0.23 |
std | 21.68 | 13.14 | 4.62 | 7.82 | 8.96 | 1.93 | 5.18 | 4.07 | |
mse | 473.36 | 172.93 | 21.38 | 61.21 | 80.37 | 3.76 | 27.03 | 16.61 | |
Oracle | bias | 1.91 | 0.34 | −0.09 | 0.02 | −0.14 | −0.25 | −0.35 | −0.27 |
std | 21.48 | 13.10 | 4.59 | 7.76 | 8.86 | 1.96 | 5.16 | 4.09 | |
mse | 464.99 | 171.68 | 21.09 | 60.27 | 78.56 | 3.92 | 26.75 | 16.81 | |
(d) | bias | 1.66 | 0.26 | −0.14 | −0.11 | 0.47 | −0.31 | −0.13 | 0.28 |
std | 20.96 | 12.52 | 4.48 | 7.59 | 8.24 | 1.27 | 3.93 | 2.83 | |
mse | 441.97 | 156.89 | 20.08 | 57.57 | 68.05 | 1.70 | 15.45 | 8.10 | |
Oracle | bias | 2.01 | −0.03 | 0.15 | −0.46 | 13.60 | −11.98 | −2.10 | 15.46 |
std | 23.22 | 13.80 | 5.06 | 8.58 | 11.25 | 8.57 | 14.14 | 10.28 | |
mse | 542.97 | 190.36 | 25.66 | 73.90 | 311.56 | 216.89 | 204.47 | 344.57 | |
(e) | bias | 1.64 | 0.25 | −0.15 | −0.10 | 0.36 | −0.26 | −0.18 | −0.22 |
std | 20.97 | 12.52 | 4.48 | 7.61 | 8.20 | 1.31 | 3.97 | 2.99 | |
mse | 442.47 | 156.76 | 20.12 | 57.90 | 67.44 | 1.78 | 15.76 | 8.96 | |
Oracle | bias | 1.53 | 0.10 | −0.07 | −0.25 | 0.55 | −0.35 | −0.13 | −0.10 |
std | 18.83 | 11.22 | 4.03 | 7.15 | 8.42 | 2.14 | 4.55 | 3.23 | |
mse | 356.86 | 125.96 | 16.27 | 51.13 | 71.19 | 4.71 | 20.67 | 10.47 |
To study the performance of the estimators when the covariates are of high dimension, we further increased the dimension of z to 10 and the bias and the standard deviation in estimating γ and β are summarized in Tables 3 and 4 for (n, dγ, dβ) = (250, 5, 10). These results convey similar messages.
Table 3.
Estimator | Parameter | γ1 (%) | γ2 (%) | γ3 (%) | γ4 (%) | γ5 (%) | β1(%) | β2(%) | β5(%) |
---|---|---|---|---|---|---|---|---|---|
(a) | bias | 189.56 | 63.33 | 31.82 | 0.72 | 0.39 | −5.66 | −0.59 | −1.23 |
std | 56.03 | 40.14 | 13.50 | 26.16 | 29.27 | 5.91 | 12.23 | 11.33 | |
mse | 39073.67 | 5622.59 | 1194.71 | 684.93 | 857.03 | 66.97 | 149.94 | 129.78 | |
(b) | bias | −0.45 | 0.29 | 0.15 | 0.00 | 4.17 | −5.67 | −0.62 | −1.10 |
std | 27.34 | 16.19 | 5.67 | 9.17 | 11.64 | 5.81 | 12.04 | 11.02 | |
mse | 747.82 | 262.35 | 32.12 | 84.06 | 152.93 | 65.84 | 145.44 | 122.57 | |
Oracle | bias | −0.43 | 0.39 | 0.17 | −0.03 | 4.20 | −5.67 | −0.62 | −1.10 |
std | 27.44 | 16.46 | 5.71 | 9.23 | 11.82 | 5.82 | 12.04 | 11.03 | |
mse | 753.23 | 271.04 | 32.67 | 85.12 | 157.34 | 66.00 | 145.28 | 122.84 | |
(c) | bias | 0.00 | 0.23 | −0.05 | −0.33 | −0.74 | −0.76 | −0.55 | −0.40 |
std | 23.67 | 13.46 | 4.78 | 7.45 | 9.18 | 2.10 | 5.29 | 4.86 | |
mse | 560.16 | 181.34 | 22.88 | 55.67 | 84.88 | 4.98 | 28.31 | 23.80 | |
Oracle | bias | −0.13 | 0.22 | −0.06 | −0.30 | −0.80 | −0.76 | −0.57 | −0.45 |
std | 23.47 | 13.38 | 4.76 | 7.50 | 9.23 | 2.10 | 5.28 | 4.87 | |
mse | 550.80 | 179.12 | 22.63 | 56.29 | 85.85 | 5.00 | 28.24 | 23.94 | |
(d) | bias | 0.10 | 0.36 | −0.08 | −0.13 | −0.35 | −0.64 | −0.31 | −0.08 |
std | 22.34 | 12.59 | 4.52 | 7.08 | 8.23 | 1.45 | 4.15 | 3.90 | |
mse | 499.22 | 158.59 | 20.44 | 50.20 | 67.78 | 2.52 | 17.35 | 15.21 | |
Oracle | bias | 0.13 | 0.31 | 0.01 | 0.20 | 7.38 | −9.17 | −1.79 | −2.52 |
std | 24.53 | 14.68 | 5.16 | 8.14 | 10.77 | 7.73 | 15.29 | 14.04 | |
mse | 601.92 | 215.64 | 26.61 | 66.31 | 170.42 | 143.85 | 236.98 | 203.48 | |
(e) | bias | 0.18 | 0.34 | −0.09 | −0.13 | −0.35 | −0.65 | −0.34 | −0.09 |
std | 22.35 | 12.63 | 4.50 | 7.06 | 8.18 | 1.47 | 4.24 | 3.93 | |
mse | 499.37 | 159.56 | 20.29 | 49.87 | 67.06 | 2.60 | 18.10 | 15.42 | |
Oracle | bias | 0.23 | 0.30 | −0.09 | −0.10 | −0.22 | −0.69 | −0.13 | −0.10 |
std | 19.45 | 11.12 | 3.91 | 6.28 | 7.34 | 1.53 | 4.25 | 3.72 | |
mse | 378.44 | 123.85 | 15.31 | 39.40 | 53.89 | 2.83 | 18.06 | 13.84 |
Table 4.
Estimator | Parameter | γ1 (%) | γ2 (%) | γ3 (%) | γ4 (%) | γ5 (%) | β1(%) | β2(%) | β5(%) |
---|---|---|---|---|---|---|---|---|---|
(a) | bias | 187.83 | 64.28 | 32.53 | 0.69 | 0.34 | −18.25 | −3.89 | 23.04 |
std | 73.24 | 53.39 | 17.42 | 35.13 | 38.64 | 12.46 | 11.53 | 14.24 | |
mse | 40645.36 | 6983.39 | 1361.69 | 1234.67 | 1493.53 | 488.42 | 148.09 | 733.60 | |
(b) | bias | −0.83 | 0.53 | 0.05 | 0.21 | 26.34 | −20.78 | −3.60 | 25.01 |
std | 38.21 | 22.40 | 7.73 | 12.35 | 19.77 | 13.44 | 11.54 | 14.63 | |
mse | 1460.43 | 502.20 | 59.76 | 152.60 | 1084.49 | 612.39 | 146.11 | 839.40 | |
Oracle | bias | −0.03 | 0.50 | 0.15 | 0.41 | 26.56 | −20.83 | −3.58 | 25.04 |
std | 42.51 | 23.54 | 8.39 | 13.14 | 20.75 | 13.46 | 11.56 | 14.64 | |
mse | 1806.81 | 554.54 | 70.48 | 172.89 | 1135.78 | 615.05 | 146.46 | 841.32 | |
(c) | bias | 0.48 | 0.17 | −0.09 | −0.20 | −0.80 | −0.90 | −0.60 | −0.14 |
std | 23.57 | 13.10 | 4.83 | 7.34 | 9.25 | 2.15 | 5.28 | 4.42 | |
mse | 555.96 | 171.68 | 23.34 | 53.92 | 86.17 | 5.44 | 28.25 | 19.59 | |
Oracle | bias | 0.57 | 0.22 | −0.13 | −0.21 | −0.75 | −0.92 | −0.63 | −0.04 |
std | 23.49 | 13.07 | 4.85 | 7.29 | 9.17 | 2.19 | 5.29 | 4.55 | |
mse | 551.97 | 170.79 | 23.53 | 53.21 | 84.69 | 5.66 | 28.40 | 20.75 | |
(d) | bias | 0.25 | 0.49 | −0.06 | −0.08 | −0.16 | −0.71 | −0.27 | 0.27 |
std | 21.68 | 12.44 | 4.45 | 6.97 | 8.13 | 1.39 | 4.36 | 2.92 | |
mse | 470.15 | 155.02 | 19.81 | 48.61 | 66.06 | 2.43 | 19.08 | 8.61 | |
Oracle | bias | −0.38 | 0.33 | 0.11 | 0.28 | 16.23 | −15.95 | −3.47 | 14.28 |
std | 25.71 | 15.44 | 5.41 | 8.69 | 11.96 | 9.41 | 14.33 | 10.15 | |
mse | 661.40 | 238.64 | 29.26 | 75.62 | 406.37 | 342.74 | 217.35 | 306.85 | |
(e) | bias | 0.31 | 0.44 | −0.06 | −0.10 | −0.22 | −0.67 | −0.22 | −0.34 |
std | 21.79 | 12.43 | 4.43 | 6.98 | 8.11 | 1.41 | 4.42 | 3.09 | |
mse | 475.10 | 154.77 | 19.61 | 48.68 | 65.89 | 2.43 | 19.55 | 9.66 | |
Oracle | bias | 0.03 | 0.19 | −0.05 | −0.10 | −0.11 | −0.72 | −0.15 | −0.24 |
std | 19.24 | 11.12 | 3.87 | 6.28 | 7.37 | 1.47 | 4.34 | 3.39 | |
mse | 370.22 | 123.67 | 14.99 | 39.46 | 54.35 | 2.69 | 18.87 | 11.58 |
Since estimator (e) provides an efficient estimator regardless of whether the linearity condition (2) is satisfied or not or whether the error is homoscedastic or heteroscedastic, it is not surprising to note that estimator (e) always offers estimates with the smallest Monte Carlo standard errors among all estimators. Table 5 further reports the average estimated standard deviation and the Monte Carlo standard deviation, as well as the empirical coverage probabilities of the 95% confidence intervals for from the 1000 simulations. As we can see, the estimated standard deviations match the Monte Carlo counterparts reasonably well, and the coverage probabilities are close to the nominal level.
Table 5.
Case | Parameter | γ1 | γ2 | γ3 | γ4 | γ5 | |
---|---|---|---|---|---|---|---|
(n, d, dβ) = (250, 5, 5) | |||||||
1 |
|
0.2091 | 0.1206 | 0.0426 | 0.0696 | 0.0755 | |
sd | 0.2161 | 0.1295 | 0.0454 | 0.0765 | 0.0829 | ||
CI | 94.1% | 93.4% | 94.5% | 92.9% | 92.7% | ||
2 |
|
0.2045 | 0.1180 | 0.0417 | 0.0681 | 0.0743 | |
sd | 0.2097 | 0.1252 | 0.0448 | 0.0761 | 0.0820 | ||
CI | 95.4% | 93.6% | 93.8% | 92.6% | 92.8% | ||
(n, dγ, dβ) = (250, 5, 10) | |||||||
1 |
|
0.2045 | 0.1178 | 0.0417 | 0.0682 | 0.0740 | |
sd | 0.2235 | 0.1263 | 0.0450 | 0.0706 | 0.0818 | ||
CI | 93.2% | 94.4% | 93.8% | 95.1% | 92.0% | ||
2 |
|
0.2000 | 0.1153 | 0.0408 | 0.0667 | 0.0727 | |
sd | 0.2179 | 0.1243 | 0.0443 | 0.0698 | 0.0811 | ||
CI | 93.4% | 94.0% | 93.2% | 94.4% | 92.1% |
5. Example
The Fifth National Bank of Springfield facedagender discrimination suit in which it was charged with paying substantially lower salaries to its female employees than to its male employees. The bank’s employee database (based on 1995 data) is listed in Albright et al. (1999), with only the bank’s name being changed. For each of its 208 employees the data set includes each employee’s annual salary Y and gender X1.
The average salary for the male employees (X1 = 1) is 45.505 (in thousands of dollars) and the average for the females (X1 = 0) is 37.262, yielding a p-value of less than 10−6 from a two-sample t-test. However, a naive comparison of the average salaries of males and females may not be suitable because there are many confounding factors that may affect salary. To make a faithful and complete comparison, we must account for these confounding factors. In this data set, three categorical and three continuous confounders are collected. The categorical confounders include a binary variable X2 indicating whether the employee’s job is computer related or not, a five-level categorical variable (X3,1,…, X3,4) representing the employee’s educational level and a six-level categorical variable denoting the employee’s current job level (X4,1,…, X4,5). The three continuous confounders are respectively working experience at the current bank (Z1, measured by the year when an employee was hired), the employee’s age (Z2) and experience at another bank before working at the Fifth National Bank (Z3, measured by the number of years at another bank). There was an obvious outlier in the data set, which was removed from our subsequent analysis. The continuous confounders Zi were standardized marginally to have mean 0 and variance 1.
To account for the effect of the confounders, two models arise naturally: the linear model
(12) |
and the partial linear single-index model
(13) |
We require β = (β1, β2, β3)T in model (13) to have unit length and positive sign for the first non-zero element to ensure identifiability. We applied ordinary least squares to model (12), which corresponds to the estimating equation (a) in Section 4. In addition, we applied estimating equations (b), (c) and (e) to estimate the parameters in model (13). We reiterate here that estimating equation (b) requires condition (2) for consistency of estimation. Both (c) and (e) yield consistent estimators without condition (2), but (e) is more efficient if the data exhibit heteroscedasticity.
The estimated coefficients and their associated standard errors for models (12) and (13) are summarized in Table 6. The ordinary least squares estimate of γ1 in the linear model (12) has a p-value of nearly 0.10, indicating no significant gender effect on salary. The semiparametric inference procedures (b), (c) and (e) suggest a similar conclusion, which is contrary to the two-sample t-test result.
Table 6.
Parameter | Results for model (12), | Results for model (13) and the following estimators: | ||||||
---|---|---|---|---|---|---|---|---|
estimator (a) |
|
|||||||
|
Estimator (b) | Estimator (c) | Estimator (e) | |||||
Coefficient | Standard deviation |
|
|
|
||||
Coefficient | Standard deviation | Coefficient | Standard deviation | Coefficient | Standard deviation | |||
γ1 | 1.4273 | 0.8504 | 1.2381 | 1.3163 | −0.1572 | 1.5864 | 0.6726 | 1.5409 |
γ2 | 4.5738 | 1.2258 | 5.0360 | 2.2537 | 4.4178 | 2.3679 | 3.4229 | 2.1513 |
γ3,1 | −3.8173 | 1.3532 | −0.5834 | 2.2137 | 1.4330 | 2.7118 | −1.7172 | 2.4421 |
γ3,2 | −4.1015 | 1.3018 | −1.1123 | 2.1733 | 0.1020 | 2.5062 | −2.9037 | 2.5332 |
γ3,3 | −2.7977 | 0.9749 | −1.7869 | 1.7624 | −0.4008 | 1.8991 | −2.2040 | 1.7955 |
γ3,4 | −3.1605 | 1.8026 | −0.5708 | 3.3598 | −0.9586 | 3.5150 | −2.1893 | 3.2231 |
γ4,1 | −26.1685 | 2.3418 | −24.1351 | 2.0737 | −30.2798 | 3.4069 | −24.9195 | 2.9111 |
γ4,2 | −24.5755 | 2.2937 | −22.6682 | 1.9986 | −28.3341 | 3.2557 | −23.1889 | 2.8851 |
γ4,3 | −21.3149 | 2.2023 | −18.7183 | 1.8050 | −23.7748 | 3.1097 | −19.2959 | 2.7476 |
γ4,4 | −17.9505 | 2.0632 | −14.0674 | 1.9868 | −18.5167 | 3.0344 | −14.9075 | 2.7432 |
γ4,5 | −13.0403 | 1.9952 | −8.2875 | 2.0998 | −11.4228 | 3.1984 | −9.0453 | 2.6962 |
β1 | 3.9500 | 0.5570 | 0.7787 | 19.6816 | 0.2329 | 0.1963 | 0.5546 | 0.0220 |
β2 | −0.1526 | 0.4909 | −0.6020 | 15.3779 | 0.4854 | 0.3464 | 0.4647 | 0.0268 |
β3 | 0.5729 | 0.3664 | 0.1765 | 4.4941 | −0.8427 | 0.6024 | 0.6902 | 0.0218 |
R2 | 0.8342 | 0.8569 | 0.8446 | 0.9194 |
Because model (12) is nested within model (13), we first examine whether the link function g(·) in model (13) is linear. The estimated link function along with its pointwise confidence band is shown in Fig. 1(a), where the bandwidth is determined by leave-one-out cross-validation. We can see that there are quite a few data points in the upper right-hand part of Fig. 1(a) which exhibit a clear non-linear pattern. A closer inspection of these data points reveals that these mainly correspond to the executives who earned handsome salaries at the Fifth National Bank. The linear model (12) fails to capture this feature, indicating that it is insufficient to describe the data. Therefore, the results of semiparametric inference procedures obtained from estimating equations (b), (c) and (e) is more trustworthy.
From Table 6, we can see that both estimator (b) and estimator (c) indicate that none of the covariates in z has any effect on the employee’s salary, which contradicts our common sense because we would expect the working experience, for example, to be an important factor for salary. We suspect that estimator (b) failed to produce a consistent estimator, which would be so if the linearity condition (2) was not satisfied. To verify this conjecture, we examine whether z satisfies the linearity condition (2). The Q–Q-plot of zTβ that is presented in Fig. 1(b) indicates that the linearity condition (2) is probably violated because the distribution of zTβ deviates from a normal population significantly.
We next compare (c) with (e). We note that (c) differs from (e) in that (e) accounts for the heteroscedasticity. We calculated the residuals of model (13) to examine the existence of heteroscedasticity. Fig. 1(c) reports the scatter plot of the squared residuals versus the component zTβ. The full curve in Fig. 1(c) is the fitted variance function w(·)−1 obtained from standard non-parametric smoothing, where the bandwidth is determined by leave-one-out cross-validation. We can clearly see from Fig. 1(c) that the data points were heteroscedastic, which was ignored by estimator (c). Thus, estimator (e) offers a more efficient estimate than (c). It also complies with the results in Table 6 because the standard errors of the estimates obtained from estimator (e) are smaller than those obtained from (c).
The last row of Table 6 summarizes the R2-values for all procedures. We can see that estimator (e) has the largest R2-value, which suggests the best performance among all estimation procedures. Fig. 1(d) shows a nice fit of the predicted salaries from estimator (e) and the observed salaries of all employees.
Considering all the above discussion, we believe that the results of estimator (e) based on model (13) are convincing. In fact, Fan and Peng (2004) analysed the same data by using another semiparametric model and reached the same conclusion that there is no significant evidence supporting gender discrimination in salary.
6. Conclusion
We have studied the partially linear single-index model with possibly high dimensional covariates, which can be viewed as a generalization of the model in Carroll et al. (1997), with the normality and homoscedasticity assumptions relaxed. We proposed a class of estimators that do not rely on the profile likelihood method. In addition, estimators in this class have a double-robustness property, which enables us to misspecify some of the non-parametric components. We also derived a computationally simple estimation procedure that achieves semiparametric efficiency. This further generalizes the efficiency result of Carroll et al. (1997). The presence of the popular linearity condition further simplifies the computation. However, we have shown that this additional condition does not contribute to the theoretical property of the estimator. In other words, the benefit of the linearity condition is purely computational. This does not come as a surprise since it is a very weak condition that always holds in the asymptotic sense (Hall and Li, 1993).
To preserve the usual optimal bandwidths hi = O(n−1/5) for i = 1,2,3, we have intentionally avoided the undersmoothing condition . A more careful but tedious analysis than that in the on-line supplementary document shows that, to guarantee the root n consistency of the resulting estimates, the valid bandwidth range is between and . Hence, if one feels comfortable in using an undersmoothing bandwidth in estimating β and γ, then the computation of consistent estimates can be relaxed because of the double robustness. Specifically, if we use a bandwidth between and , then we can opt to estimate either the g-function or the appropriate conditional expectation E(·|zTβ). This is computationally beneficial when we are not willing to propose a model for either of the two quantities, in that it is not necessary to estimate both. If g or E(·|zTβ) is of interest, one can always perform an additional standard non-parametric regression after obtaining the β- and γ-estimates.
Equivalent arguments and conclusions hold when model (1) is generalized to the scenario in which the partially linear function is replaced by an arbitrary semiparametric function . The estimator will have the form
where . The estimator remains consistent and efficient provided that and are properly estimated non-parametrically.
We did not expand our model to allow common components in x and z, which actually leads to very different results. This will be investigated in our subsequent work, and readers are referred to Xia et al. (1999) for related discussion and results. Finally, when zTβ is not sufficient to describe the regression, the partially linear additive model or partially linear multi-index model represent common strategies to overcome the curse of dimensionality. Recent work in this area includes Xia (2008) and Li et al. (2011).
Supplementary Material
Acknowledgments
Ma’s work is supported by grants from the National Science Foundation (DMS-0906341 and DMS-1206693) and a grant from the National Institute of Neurological Disorders and Stroke (R01 NS073671).Zhu’s work is supported by the National Natural Science Foundation of China (grant 11071077).
Footnotes
Supporting information
Additional ‘supporting information’ may be found in the on-line version of this article:
‘Supplement materials to Doubly robust and efficient estimators for heteroscedastic partially linear single-index model allowing high-dimensional covariates’.
Please note: Wiley–Blackwell are not responsible for the content or functionality of any supporting materials supplied by the authors. Any queries (other than missing material) should be directed to the author for correspondence for the article.
Contributor Information
Yanyuan Ma, Texas A&M University, College Station, USA.
Liping Zhu, Shanghai University of Finance and Economics, People’s Republic of China.
References
- Albright SC, Winston WL, Zappe CJ. Data Analysis and Decision Making with Microsoft Excel. Pacific Grove; Duxbury: 1999. [Google Scholar]
- Bickel P, Klaassen CAJ, Ritov Y, Wellner JA. Efficient and Adaptive Inference in Semiparametric Models. Baltimore: Johns Hopkins University Press; 1993. [Google Scholar]
- Carroll RJ, Fan J, Gijbels I, Wand MP. Generalized partially linear single-index models. J Am Statist Ass. 1997;92:477–489. [Google Scholar]
- Engle RF, Granger CWJ, Rice J, Weiss A. Semiparametric estimates of the relation between weather and electricity sales. J Am Statist Ass. 1986;81:310–320. [Google Scholar]
- Fan J, Peng H. Nonconcave penalized likelihood with a diverging number of parameters. Ann Statist. 2004;32:928–961. [Google Scholar]
- Hall P, Li KC. On almost linearity of low dimensional projections from high dimensional data. Ann Statist. 1993;21:867–889. [Google Scholar]
- Li KC. Sliced inverse regression for dimension reduction (with discussion) J Am Statist Ass. 1991;86:316–342. [Google Scholar]
- Li L, Zhu L, Zhu L. Inference on the primary parameter of interest with the aid of dimension reduction estimation. J R Statist Soc B. 2011;73:59–80. [Google Scholar]
- Lin X, Carroll RJ. Semiparametric estimation in general repeated measures problems. J R Statist Soc B. 2006;68:69–88. [Google Scholar]
- Ma Y, Chiou J, Wang N. Semiparametric estimator in partially linear models. Biometrika. 2006;93:75–84. [Google Scholar]
- Robins JM, Rotnitzky A. Discussion on “Celebrating the new millennium” by Bickel, P. J. and Kwon, J. Statist Sin. 2001;11:920–926. [Google Scholar]
- Tan Z. Nonparametric likelihood and doubly robust estimating equations for marginal and nested structural models. Can J Statist. 2010;38:609–632. [Google Scholar]
- Tsiatis AA. Semiparametric Theory and Missing Data. New York: Springer; 2006. [Google Scholar]
- Wang J-L, Xue L, Zhu L, Chong YS. Estimation for a partial-linear single-index model. Ann Statist. 2010;38:246–274. [Google Scholar]
- Xia Y. A multiple-index model and dimension reduction. J Am Statist Ass. 2008;103:1631–1640. [Google Scholar]
- Xia Y, Härdle W. Semi-parametric estimation of partially linear single-index models. J Multiv Anal. 2006;97:1162–1184. [Google Scholar]
- Xia Y, Tong H, Li WK. On extended partially linear single-index models. Biometrika. 1999;86:831–842. [Google Scholar]
- Yu Y, Ruppert D. Penalized spline estimation for partially linear single-index models. J Am Statist Ass. 2002;97:1042–1054. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.