Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2015 Mar 1.
Published in final edited form as: J Agric Biol Environ Stat. 2014 Mar 1;19(1):57–81. doi: 10.1007/s13253-013-0157-7

Nonlinear Varying Coefficient Models with Applications to Studying Photosynthesis

Esra Kürüm 1, Runze Li 2, Yang Wang 3, Damla ŞEntürk 4
PMCID: PMC4070621  NIHMSID: NIHMS591009  PMID: 24976756

Abstract

Motivated by a study on factors affecting the level of photosynthetic activity in a natural ecosystem, we propose nonlinear varying coefficient models, in which the relationship between the predictors and the response variable is allowed to be nonlinear. One-step local linear estimators are developed for the nonlinear varying coefficient models and their asymptotic normality is established leading to point-wise asymptotic confidence bands for the coefficient functions. Two-step local linear estimators are also proposed for cases where the varying coefficient functions admit different degrees of smoothness; bootstrap confidence intervals are utilized for inference based on the two-step estimators. We further propose a generalized F test to study whether the coefficient functions vary over a covariate. We illustrate the proposed methodology via an application to an ecology data set and study the finite sample performance by Monte Carlo simulation studies.

Keywords: Generalized F test, Local linear regression, Nonlinear regression model, Varying coefficient models

1 INTRODUCTION

This work was motivated by an empirical study in the field of ecology. It is known that the rate of photosynthesis in an ecosystem is affected by sunlight intensity. The level of photosynthetic activity in a natural ecosystem is measured by the Net Ecosystem Exchange of CO2 (NEE), since leaves absorb carbon dioxide (CO2) during photosynthesis. NEE depends on the amount of Photosynthetically Active Radiation (PAR) available to an ecosystem. It is believed that the relationship between NEE and PAR is nonlinear (Monteith 1972) and follows the model,

NEE=β1β2PARPAR+β3+ε, (1.1)

where ε is the random error with mean zero, and β1; β2; β3 are the unknown parameters. The coefficient β1 represents the dark respiration rate, that is the rate of respiration in plants where carbon dioxide is released without the aid of sun light, and β2 represents the light-saturated net photosynthetic rate. The light saturation point refers to the amount of light that the chloroplast cannot absorb. At this point, photosynthesis still occurs, but the amount of light is more than the number of pigments in chlorophyll cells that are available for absorption. Finally, β2/β3 is the apparent quantum yield. The quantum yield of photosynthesis is a measure of efficiency which refers to the amount of the products of photosynthesis, such as O2, per unit light input (Garrett and Grisham 2005). When the quantum yield measured is based only on the incident light, that is the light falling on a surface, and does not take into account the scattered light, it is referred as the apparent quantum yield or lower limits of true quantum yield.

Model (1.1) was obtained in a laboratory where climate variables such as temperature and moisture variability can be controlled. However, when the temperature (T) cannot be controlled, the parameters β1, β2, and β3 are likely to depend on temperature. To demonstrate that the relationship between NEE and PAR depends on temperature, we fit a 2-dimensional kernel regression of NEE (μmol m−2s−1) on PAR (μmol m−2s−1) and temperature (°C). Figure 1(a) gives the contour plot of the regression fit, where three slices of the estimated regression fit are given at three quartiles of the observed temperature values in Figure 1(b). In Figure 1(a), there seems to be a weak temperature effect on the association of NEE and PAR for low values of PAR; however, as PAR increases, the non-parallel pattern demonstrates that the effect of temperature on this relation is getting stronger. The profiles in Figure 1(b) strengthen this observation: with large PAR values, the estimates seem to differ for different temperature values. Hence, we consider the following model allowing the nonlinear regression relation between NEE and PAR to change with temperature:

NEE=β1(T)β2(T)PARPAR+β3(T)+ε. (1.2)

A detailed analysis of the ecological data set will be given in Section 4.

Figure 1.

Figure 1

(a) Contour plot of the 2-dimensional kernel regression fit of NEE on PAR and TEMP, (b) Slices from the kernel regression fit at three quartiles of the observed temperature values.

Powerful modeling techniques have been proposed for dealing with ‘curse of dimensionality’ while modeling high dimensional data. Examples include additive models (Breiman and Friedman 1985; Hastie and Tibshirani 1990), low-dimensional interaction models (Friedman 1991; Gu and Wahba 1993; Stone, Hansen, Kooperberg, and Truong 1997), multiple-index models (Härdle and Stoker 1989; Li 1991), and partially linear models (Wahba 1984; Green and Silverman 1994). With the aim of increasing the flexibility of linear regression models and reducing the modeling bias, varying coefficient models

Y=XTβ(U)+ε, (1.3)

was first introduced by Cleveland, Grosse, and Shyu (1992) and became popular in the statistical literature due to the work by Hastie and Tibshirani (1993). In (1.3), Y denotes the response variable, X the vector of predictors, and U a scalar predictor, such as temperature. The varying coefficient models explore the dynamic features of the data by allowing the model coefficients to vary over a covariate U, such as temperature. At a fixed U = u value, the regression coefficients in a varying coefficient model can be interpreted in a similar way to a linear regression model.

Model (1.2) motivates us to consider a new class of models, that are a natural extension of varying coefficient models, namely, nonlinear varying coefficient models. We define the nonlinear varying coefficient model as

Y=f{X,β(U)}+ε, (1.4)

where f (·, ·) is a pre-specified function and β(U) consists of unknown coefficient functions. The error term satisfies E(ε|X;U) = 0. Similar to the interpretation of a varying coefficient model, at a fixed value U = u, the nonlinear varying coefficient model can be interpreted as a nonlinear regression model.

Estimation in varying coefficient models has been studied extensively. Earlier work on estimation of the varying coefficient models includes Wu, Chiang, and Hoover (1998), Hoover, Rice, Wu, and Yang (1998), Fan and Zhang (1999) and Kauermann and Tutz (1999). Fan and Zhang (2008) provides an excellent review of the literature on estimation and inference procedures for varying coefficient models. In this paper, we propose one-step and two-step local linear estimators for nonlinear varying coefficient models using local linear regression techniques (Fan and Gijbels 1996). While the one-step estimators estimate the varying coefficient functions using a single bandwidth, the two-step estimators adopt to cases where the varying coefficient functions admit different degrees of smoothness. Since there is no closed form solution for the local parameters, we propose an iterative linear regression algorithm to search for the local solutions. Thus, the proposed estimation procedures can be easily carried out. We establish the asymptotic normality of the proposed one-step estimators and derive consistent estimators for their asymptotic standard error, leading to asymptotic point-wise confidence intervals. Bootstrap percentile confidence intervals are utilized for inference based on the two-step estimators.

It is of interest to test whether the proposed nonlinear varying coefficient model reduces to a regular nonlinear regression model, which amounts to testing whether the varying coefficient functions in (1.4) are in fact varying over the covariate U. In this case the null hypothesis is parametric, while the alternative is nonparametric. For this hypothesis testing problem, we develop a generalized likelihood ratio test where the proposed test statistic follows a χ2 distribution and it inherits Wilks phenomenon (Fan, Zhang, and Zhang 2001), i.e. the asymptotic null distribution of the test statistic is independent of the nuisance parameters β(U) and the density function of U. Based on the Wilks phenomenon, we utilize a bootstrap procedure to estimate the degrees of freedom of the χ2 distribution of the proposed test statistic (Cai, Fan, and Li 2000). We apply the proposed methodology for a detailed empirical analysis of the aforementioned ecological data set. The efficacy of the proposed algorithms is studied via Monte Carlo simulations.

The paper is organized as follows. In Section 2, we develop one-step and two-step estimation procedures and propose a generalized F test for the nonlinear varying coefficient models. Asymptotic properties of the proposed one-step estimation procedure are also given in this section. Monte Carlo simulation studies are presented in Section 3. A detailed empirical analysis of the ecology data set is given in Section 4. Regularity conditions and technical proofs are given in the supplementary material.

2 STATISTICAL INFERENCE PROCEDURES

Suppose that {Ui,Xi,Yi}, i = 1,2, … ,n is an independent and identically distributed sample from the nonlinear varying coefficient model

Yi=f{Xi,β(Ui)}+εi,

where E(εi|Xi,Ui) = 0 and β(Ui) = {β1(Ui),β2(Ui); … ;βp(Ui)}T consists of p unknown nonparametric coefficient functions. At every fixed U = u value, a nonlinear varying coefficient model implies a nonlinear regression model. Hence, its identifiability requires the identifiability of the nonlinear models holding at each fixed U = u, i.e. that there exists no nonzero parameter β0(u) ≠ β(u) such that f(x;β0(u))= f(x;β(u)) for all x (Seber and Wild 2003).

2.1 ONE-STEP ESTIMATION

One approach for estimating the coefficient functions βj(·), j = 1, … , p, is to employ local linear fitting techniques. In what follows, we will use h1 to denote the bandwidth choice used in obtaining the one-step estimator and use h0 and h2 to denote the initial and final bandwidths used in obtaining the two-step estimators. In order to estimate the nonparametric coefficient functions at a fixed point u0, we locally approximate the functions in the neighborhood of a fixed point u0 by the Taylor expansion:

βj(u)βj(u0)+βj(u0)(uu0)aj+bj(uu0),

for j = 1,…,p. Let a = (a1,…,ap)T and b = (b1,…,bp)T. We minimize the local least squares,

(a,b)=i=1n[Yif{Xi,a+b(Uiu0)}]2Kh1(Uiu0), (2.1)

with respect to the local parameters (a,b), where Kh1(t)=h11K(th1) is the scaled kernel function K(·) with bandwidth h1. The estimators of the nonlinear regression coefficients can be given as β^(u0)=a^.

Note that f (X,β) is pre-specified, but may be a nonlinear function of X and β. Thus, it is typical that there is no closed form for a^ and b^. The Newton-Raphson algorithm may be employed to search the solution. However, based on our limited experience, the Hessian matrix of l(a,b) may not be positive definite during the iteration of the Newton-Raphson algorithm. This poses a challenge in finding the minimizer of the objective function. Hence, we propose the following iterative linear regression algorithm to obtain the values that minimize the local least squares function (2.1). During the course of the iteration, suppose we have the value (a0,b0) of (a,b) at the current step. We approximate f{Xi,a+b(Uu0)} in the neighborhood of (a0,b0) using the Taylor expansion,

f{Xi,a+b(Uiu0)}f{Xi,a0+b0(Uiu0)}+{(aa0)+(bb0)(Uiu0)}Tf{Xi,a0+b0(Uiu0)},

where f’ (Xi,β) is the p×1 vector equal to ∂f(Xi,β)/β. Denote the values of a and b at the kth iteration as a(k) and b(k). Let

Fk=[fT{X1,a(k)+b(k)(U1u0)}(U1u0)fT{X1,a(k)+b(k)(U1u0)}fT{Xn,a(k)+b(k)(Unu0)}(Unu0)fT{Xn,a(k)+b(k)(Unu0)}]n×2p,Yi,k=Yif{Xi,a(k)+b(k)(Uiu0)}+{a(k)+b(k)(Uiu0)}Tf{Xi,a(k)+b(k)(Uiu0)},

and Yk = (Y1,k, … ,Yn,k)T. Then, we update (a,b) according to

(a(k+1)b(k+1))=(FkTWFk)1FkTWYk,

where W= diag (Kh1(U1u0), …,Kh1(Unu0)). The solution of this iterative linear regression algorithm will satisfy l(a,b) = 0, and the estimators will be given by β^(u0)=a^ and β^(u0)=b^. The proposed algorithm is indeed the Fisher scoring algorithm if we further assume that the error follows a normal distribution. Hence, our algorithm shares the convergence properties of the Fisher scoring algorithm. The asymptotic rate of convergence of the scoring algorithm should approach that of the Newton method as the sample size goes to infinity (Osborne 1992). In other words, the scoring algorithm approaches a second order convergence rate as the sample size increases. We demonstrate that the proposed algorithm converges quickly in simulation studies.

A question that arises in the practical implementation of the proposed procedure is the bandwidth selection. We suggest using a multi-fold cross-validation technique (Geisser 1975). Specifically, we propose minimizing the following cross validation score:

CV(h)=jYjY^j2, (2.2)

where Y^j denotes the fitted value with jth group of subjects excluded. The bandwidth that minimizes (2.2) is chosen.

2.2 TWO-STEP ESTIMATION

As a second estimation proposal, we extend the two-step local linear estimation procedure of Fan and Zhang (1999) proposed for varying coefficient models to nonlinear varying coefficient models. While the proposed one-step estimation is easier to implement, the two-step estimators have been reported to lead to efficiency gains when the varying coefficient functions admit different degrees of smoothness and hence a single bandwidth is not optimal in estimating all varying coefficient functions. The proposed two-step estimation procedure is sketched as follows. In the first step, we obtain initial estimators β~j(Ui),j=1,,p, by minimizing (2.1) with a small bandwidth h0. These estimators would have a smaller bias, but a larger variance. As Fan and Zhang (1999) pointed out the two-step estimator is not sensitive to the initial bandwidth h0 choice, since the two-step estimator is optimal for a wide range of bandwidths. We suggest using a multifold cross-validation (2.2) or generalized cross-validation to obtain the bandwidth h^ and then use h0=0.5h^ as the initial bandwidth (Fan and Zhang 1999). In the second step, we obtain the two-step estimator of βj(·), j = 1, … , p, by replacing all the varying coefficient functions except βj(·) with their corresponding initial estimators. In this step, we reduce the variance by using further smoothing where we recommend using a multi-fold cross-validation (2.2) in choosing the optimal bandwidth. To obtain the two-step estimator of βj(·) at a fixed point u0, we start by locally approximating the function βj(·) in the neighborhood of a fixed point u0 as follows:

βj(u)βj(u0)+βj(u0)(uu0)aj+bj(uu0).

This leads to minimizing the following local least squares in order to obtain the two-step estimator of βj(u0):

(a,b)=i=1n[Yif{Xi,aj+bj(Uiu0)}]2Kh2(Uiu0),

with respect to (aj,bj), where f*(· , ·) is the function f(· , ·) with all coefficients except βj(u0) substituted with their corresponding initial estimators, and Kh2(t)=h21K(th2) is the scaled kernel function K(·) with bandwidth h2. The final estimator of βj(u0) is a^(u0).

In the first step of the proposed estimation, the iterative algorithm described in Section 2.1 is employed and the initial estimators β~j(Ui) for i = 1, … , n and j = 1, … , p, are obtained. For minimization of the local likelihood in the second step, we propose an iterative linear regression algorithm similar to the one given in Section 2.1. We denote the values of a* and b* at the kth iteration as a*(k) and b*(k), respectively, and f*’(Xi,βj) is the derivative of f*(Xi;βj) with respect to βj, that is ∂f*(Xi,βj)=∂βj. Let

Fk=[fT{X1,aj(k)+bj(k)(U1u0)}(U1u0)fT{X1,aj(k)+bj(k)(U1u0)}fT{Xn,aj(k)+bj(k)(Unu0)}(Unu0)fT{Xn,aj(k)+bj(k)(Unu0)}]n×2,Yi,k=Yif{Xi,aj(k)+bj(k)(Uiu0)}+{aj(k)+bj(k)(Uiu0)}Tf{Xi,aj(k)+bj(k)(Uiu0)},

and Yk = (Y1,k, … ,Yn,k)T. Then, we update aj=(aj,bj)T according to

aj(k+1)=(FkTWFk)1FkTWYk,

where W= diag (Kh2(U1u0); …;Kh2(Unu0)). The solution of this iterative linear regression algorithm will satisfy l*(a*;b*) = 0, and the two-step estimator of βj(u0) is given by a^j.

2.3 ASYMPTOTIC PROPERTIES OF THE ONE-STEP ESTIMATOR

We now derive the asymptotic distributions of our one-step estimators a^ and b^. Define θ(u0)=(a1, … ,ap,b1; … ,bp)T and θ^(u0)=(a^T,b^T)T. Further define μk = ∫tkK(t)dt, vk = ∫tkK2(t)dt and H = diag (1,h1) ⊗ Ip with ⊗ denoting the Kronecker product and Ip the p×p identity matrix. Let c(u) denote the marginal density of U, and let

Γ1(u0)=E(f{X;β(u0)}[f{X;β(u0)}]TU=u0)p×p,Γ2(u0)=E(σ2(u0,X)f{X;β(u0)}[f{X;β(u0)}]TU=u0)p×p.

The following theorem establishes the asymptotic normality of the one-step estimators of the regression coefficient functions β(u).

Theorem 1

Under the regularity conditions (A)—(G) in the supplementary material, we have the following result for θ^(u0):

nh1[H{θ^(u0)θ(u0)}h122(μ2μ12){(μ22μ1μ3)β(u0)(μ3μ1μ2)β(u0)}+op(h12)]DN(0,Δ1ΛΔ1)

as n→∞, where

Δ=c(u0)(1μ1μ1μ2)Γ1(u0)andΛ=c(u0)(ν0ν1ν1ν2)Γ2(u0).

If K(·) is symmetric, we obtain the following simplification:

nh1{a^(u0)β(u0)h12μ22β(u0)+op(h12)}DN(0,Σ(u0))

as n→∞, where

Σ(u0)=ν0Γ11(u0)Γ2(u0)Γ11(u0)c(u0).

Theorem 2

Under the regularity conditions given in the supplementary material, it holds that

H(FTWF)1FTWQWF(FTWF)1HPΔ1ΛΔ1

as n→∞, where F = Fk, Q=diag(e12,en2) with ei=Yif{Xi,β^(Ui)}.

The proofs of Theorems 1 and 2 are given in the supplementary material. Theorem 1 shows that when K(·) is symmetric, the asymptotic bias of a^(u) is the same as that in standard varying coefficient models (Fan and Zhang 2008). Theorem 1 also establishes the asymptotic variance of a^(u); however, we need a consistent estimator of this asymptotic variance to obtain confidence intervals for β(u). To this end, we use Theorem 2 to derive a consistent estimator of the asymptotic covariance matrix of θ^(u0) given in Theorem 1,

cov^{θ^(u0)}=(FTWF)1FTWQWF(FTWF)1. (2.3)

The accuracy of this estimator will be tested in our simulation studies. Thus, the (1–α) 100% asymptotic pointwise confidence intervals for θ(u0) can be given as β^j(u0)±Cα2[var^{β^j(u0)}]12, where Cα2 is the 100×(1−α=2)-percentile of N(0,1).

2.4 GENERALIZED F-TEST

It is of scientific interest to test whether the coefficient functions in (1.2) really change over the temperature T. In general, we may formulate this problem as a hypothesis test to study whether the coefficients of the nonlinear varying coefficient model (1.4) actually vary over the covariate U as follows:

H0:βj(U)βj0vsH1:βj(U)βj0forj=1,,p,

where βj0’s are unknown constants. This hypothesis testing is challenging because the corresponding null hypothesis is parametric, whereas the alternative is nonparametric. Thus, the parameter space under the null hypothesis is finite dimensional, while it is infinite dimensional under the alternative hypothesis. Intuitively, we would compare the residual sum of squares under H0 and under H1. Let β^() and β^0 be the estimators of β under H1 and H0 respectively. Denote by RSS(H1)=i=1n[Yif{Xi,β^(Ui)}]2 and RSS(H0)=i=1n{Yif(Xi,β^0)}2, the residual sum of squares under H0 and H1, respectively. Define

F0=RSS(H0)RSS(H1)n1RSS(H1).

Note that F0 is similar to the F-test for linear regression models. However, the parameter space under H1 is infinite dimensional. Recall that an F-test for linear hypothesis in linear regression models is equivalent to the likelihood ratio test under the normality assumption. Specifically, assume that εi are iid N(0,σ2); then, the local log-likelihood function conditional on {Ui,Xi} equals

{β(Ui)}=n2log(2πσ2)12σ2i=1n[Yif{Xi,β(Ui)}]2.

Then the likelihood ratio test is defined by

T0=n(H1)n(H0)=n2log{RSS(H0)RSS(H1)}, (2.4)

where ln(H0) and ln(H1) are the maximum likelihood functions under the null and alternative hypotheses, respectively, and (2.4) is a monotone function of the ratio RSS(H0)=RSS(H1). Noting that log(1+x) ≈ x for small x by the Taylor expansion, we obtain

T012RSS(H0)RSS(H1)n1RSS(H1)=12F0

under H0. Therefore, motivated by the generalized likelihood ratio test proposed in Fan, Zhang, and Zhang (2001), we define the generalized F-test statistic to be

F=rK2F0,

where rK = {K(0)–0:5 ∫ K2(u)du}/[{K(u)–0.5K *K(u)}du], K *K denotes the convolution of K, and K(·) is the kernel function used to estimate the regression coefficients. The value of rK is 1.2000, 2.1153, 2.3061, 2.3797, and 2.5375 for uniform, epanechnikov, biweight, triweight, and gaussian kernel, respectively. Note here that the generalized F test does not rely on the normality assumption of the error distribution in general.

Intuitively, under the alternative hypothesis, RSS(H0) would be larger than RSS(H1) which would make the test statistic F large and we would reject the null hypothesis. However, under H0, RSS(H0) will be close to RSS(H1); hence, F would be small and we would be in favor of H0. Fan, Zhang, and Zhang (2001) showed that the generalized likelihood ratio test statistic F has a χ2 limiting distribution for a variety of models including Gaussian white noise models, nonparametric regression models, varying coefficient models and generalized varying coefficient models. However, the null limiting distribution given in Fan, Zhang, and Zhang (2001) does not keep the type I error rate well. As demonstrated by Cai, Fan, and Li (2000), a bootstrap procedure may provide a better estimate of the null distribution with moderate sample size than the asymptotic distribution. Thus, we propose to use a nonparametric bootstrap method to estimate the null distribution. In the proposed nonparametric bootstrap procedure, we first estimate the model coefficients under the null and alternative hypotheses using the original data. Let β~(Ui) and β^(Ui) denote the estimates of the model coefficients under the null and the alternative hypotheses, respectively, from which the test statistic F is computed. We then obtain residuals ei=Yif{Xi,β~(Ui)}. Next, we obtain each bootstrap sample by generating data (Yi,Xi,Ui) using the model

Yi=f{Xi,β~(Ui)}+ei,

where ei denotes the residuals sampled from {e1, … ,en} with replacement. The distribution of the test statistics F* computed from the bootstrap samples estimate the null distribution of F. We study the properties of the proposed generalized F-test by Monte Carlo simulations in Section 3. Note that in simulation studies and data application, we employ the one-step estimator to estimate the null distribution of the test statistics F, but the two-step estimator can also be used and the procedure would be the same.

3 SIMULATION STUDIES

In this section, we study the performance of the proposed one-step and two-step estimators via Monte Carlo simulation studies. All simulations were conducted using R. The Epanechnikov kernel, K(t) = 0.75(1−t2)+, is used in our simulations. We evaluate the performance of the proposed procedures using root average squared error (RASE):

RASEj=(1ngridk=1ngrid[βj(uk)β^j(uk)max{βj(u)}min{βj(u)}]2)12,

where {uk,k = 1, … ,ngrid} is an equidistant set of grid points over the range of U with ngrid = 200 and j = 1, … , p. We further study the accuracy of the proposed estimator of the covariance matrix from Section 2.3.

Example 1

In this example, data are generated from the following nonlinear varying coefficient model:

Yi=exp{β1(Ui)+Xiβ2(Ui)}+εi,

where β1(Ui) = sin(πUi) and β2(Ui) = sin{4π(Ui–1/8)}. The predictor and error variables, Xi and εi, are generated from two independent N(0,1), while Ui are generated from U[0,1], the uniform distribution on [0,1]. We consider sample sizes n = 250,500 and 1000 and 500 Monte Carlo runs for each sample size.

In our simulation, we generate several pilot simulation data sets, and use a cross-validation bandwidth selector to get an overall picture about the optimal bandwidth. To save computing time, we fix the bandwidth to be close to the optimal ones from the pilot simulation data sets. Specifically, we set the bandwidth to be h1 = 0.10, 0.075 and 0.06 for n = 250, 500 and 1000, respectively, for the one-step procedure. We use h0 = h1/2 for the initial bandwidth of the two-step estimator and h2 = (0.2,0.1), (0.15,0.075), (0.12,0.06) (corresponding to that used for (β1 and β2)) for the final bandwidth of the two-step estimator for n = 250, 500 and 1000, respectively. It is of interest to examine the performance of the proposed estimators with a wide range of the bandwidth. Thus, for each sample size, we also set the bandwidth to be h1/2, 2h1, h2/2 and 2h2, corresponding to undersmoothing and oversmoothing simulations.

The sample mean and sample standard deviation of the RASE values for the one-step and two-step procedures from the 500 replications are summarized in Tables 1 and 2, respectively. Figure 2 depicts the one-step estimates of the parameter functions and 95% asymptotic point-wise confidence intervals based on a typical sample which has the median RASE value over the 500 Monte Carlo runs. The estimates target the underlying true coefficient functions in which the proposed confidence intervals get narrower with increasing sample size. In this simulation, β1(·) and β2(·) admit different degrees of smoothness and the one-step estimate for β1(·) is undersmoothed (Figure 2) . Hence, we employ the two-step procedure. Figure 3 displays the two-step estimates of β1(·) for each sample size based on a sample with the median RASE value over 500 Monte Carlo runs along with 95% percentile bootstrap confidence intervals based on 500 bootstrap samples. According to Figure 3, the two-step estimates are smoother and closer to the underlying true coefficient function than the one-step estimates. The two-step estimates of β2(·) are close to the one-step estimates and hence are omitted. Since the two varying coefficient functions admit different degrees of smoothness in this simulation, the two-step estimator leads to efficiency gains across all sample sizes, as demonstrated with smaller RASE values in Table 2. Even though the two-step procedure outperforms the one-step procedure in this simulation, it requires 3 times more computation time than the one-step procedure for n = 500 implemented with CPU 2.3 GHz Intel Core i5.

Table 1.

RASE for One-step Estimates in Example 1

h β1(u) β2(u)

n = 250 n = 500 n = 1000 n = 250 n = 500 n = 1000
12h1 .224(.112) .160(.048) .116(.033) .117(.058) .076(.029) .049(.018)
h 1 .154(.054) .114(.038) .081(.022) .083(.030) .053(.019) .035(.009)
2h1 .198(.125) .132(.053) .095(.028) .147(.030) .088(.017) .059(.009)

Table 2.

RASE for Two-step Estimates in Example 1

h β1(u) β2(u)

n = 250 n = 500 n = 1000 n = 250 n = 500 n = 1000
12h2 .194(.162) .128(.084) .086(.032) .106(.061) .066(.026) .043(.013)
h 2 .125(.057) .095(.038) .064(.021) .081(.038) .050(.020) .033(.010)
2h2 .191(.128) .121(.089) .075(.035) .143(.033) .082(.019) .052(.009)

Figure 2.

Figure 2

One-step estimates for the varying coefficient functions (dashed) overlaying the true functions (solid) along with 95% pointwise asymptotic confidence intervals (dotted) in (a) and (b) for n = 250, (c) and (d) for n = 500, (e) and (f) for n = 1000.

Figure 3.

Figure 3

Two-step estimates of the varying coefficient function β1(·) (dash-dotted) overlaying the true functions (solid) and 95% percentile bootstrap confidence intervals based on 500 bootstrap samples (dotted) in (a) for n = 250, (b) for n = 500, and (c) for n = 1000.

To examine the bias of the proposed one-step estimator, we also summarize the empirical bias of β^j(u). Table 3 reports the empirical bias at u = 0.5. The empirical bias of β^j(u) at other locations is similar, and is not reported here to save space. We also test the accuracy of the proposed standard error formula in (2.3). The standard deviation, denoted by SD in Table 3, of 500 β^j(u), based on 500 simulations, can be viewed as the true standard error. The sample average and the sample standard deviation of the 500 estimated standard errors of β^j(u), denoted by SE and SDse in Table 3, respectively, summarize the overall performance of the standard error formula (2.3). Table 3 presents the results at u = 0.5. Table 3 shows that the proposed covariance estimator slightly underestimates the true one, but the difference between SD and SE is less than twice of the SDse. We conclude that the proposed covariance estimator works reasonably well (Cai et al. 2000; Fan and Li 2004). Results for other locations are similar, and are not presented here to save space.

Table 3.

Bias and Standard Errors for One-step Estimates in Example 1 (h = h1)

n β1(u) with u = 0:5 β2(u) with u = 0:5

Bias SD SE(SDse) Bias SD SE(SDse)
250 .081 .097 .074(.015) .125 .112 .064(.024)
500 .067 .082 .065(.020) .081 .081 .053(.021)
1000 .046 .056 .048(.014) .052 .050 .036(.013)

We next investigate the behavior of the generalized F test and its power. It is expected that the null distribution approximately follows a χ2-distribution. Furthermore, the limiting χ2 distribution does not depend on the specific null values considered. This property has been called the Wilks phenomenon in Fan, Zhang, and Zhang (2001). To illustrate the Wilks phenomenon, we consider

H0:βj(U)=γj0vsH1:βj(U)γj0,j=1,2, (3.1)

with five different sets for (γ10, γ20): (μβ1 , μβ2) and four combinations of (μβ1 ±2std(β1),μβ2 ±2std(β2)), where μβj = E{βj(U)} and std(βj)=var{βj(U)} for j = 1, 2.

A nonparametric bootstrap procedure with resampling from residuals is applied for n = 500 to estimate the null distribution of the test statistic F at different null values. After one sample is generated from the above simulation set-up, the residuals ei = Yif(Xi, γ0) are obtained where γ0 contains the null values. Next, 500 bootstrap samples (Yi,Xi,Ui) of size n = 500 are generated under the null according to

Yi=f(Xi,γ0)+ei,

where ei is randomly drawn from {e1, ⋯ ,en} with replacement. The test statistic is computed for each bootstrap sample. The estimated density of the test statistic F for 5 different null values is plotted in Figure 4(a) along with the density of the chi-square distribution. The degrees of freedom for the null distribution of the test statistic F is chosen to be close to the sample mean of bootstrap test statistic values across different null values, which is approximately 31.2. The plotted densities in Figure 4(a) are very close to each other and to the chi-square density, which confirms that the null distribution of the test statistic follows a χ2 distribution with degrees of freedom 31.2 and that it is independent of the γj0 null values. Hence, the Wilks phenomenon holds for the nonlinear varying coefficient models. We also study the power of the proposed hypothesis testing procedure. Consider the hypotheses given in (3.1). The power is evaluated at a sequence of alternatives indexed by δ,

H1:βj(U,δ)=(1δ)γj0+δβj0(U),

where j = 1,2 and δ ∈ [0,0.4]. Hence, larger values of δ correspond to further deviation from the null hypothesis. Figure 4(b) depicts the five power functions at five different significance levels: 0.5, 0.25, 0.10, 0.05 and 0.01. The estimated power at δ = 0 for the five significance levels are 0.498, 0.242, 0.074, 0.048 and 0.012 based on 500 replications at n = 500. This implies that the test statistic keeps type I error rate very well. As expected, the power functions increase rapidly for increasing δ values.

Figure 4.

Figure 4

(a) Plots for estimated densities of the test statistic F from the proposed generalized F-test using 5 different null values. The solid line in (a) is the density function of a χ2 distribution with 31.2 degrees of freedom; (b) is the estimated power functions.

Example 2

In this example, we generate data from

Yi=β1(Ui)=β2(Ui)XiXi+β3(Ui)+εi,

which is the same model as that was fit to the ecological data. In our simulation, εi ~ N(0,0.1); the Xi follows a truncated normal distribution with mean and variance equal to the corresponding values of PAR variable in the ecological data and lies between the minimum value of PAR variable and the maximum value of PAR variable; and the Ui follows the uniform distribution over [11.5;24.7], which is the range for temperature in the ecological data. Furthermore, we set β1(Ui) = 0.5exp(0.5Ui), β2(Ui) = 23+sin(0.25Ui) and β3(Ui)=300+2(Ui–10)2 so that their ranges are similar to those of estimated parameters in Section 4. We consider three sample sizes, n = 500, 1000 and 2000. We take the bandwidth to be h1 =5.3, 4.6 and 4.0 for n =500, 1000 and 2000, respectively, for the one-step procedure. We use h0 = h1/2 for the initial bandwidth of the two-step estimator and h2 = (5.3,5.3,5.3), (4.6,4.6,4.6), (4.0,4.0,4.0) (corresponding to that used for (β1, β2, β3)) for the final bandwidth of the two-step estimator for n = 250, 500 and 1000, respectively. For each sample size, we also set the bandwidth to be h1/2, 2h1, h2/2 and 2h2, corresponding to undersmoothing and oversmoothing simulations for each sample size. We conduct 500 simulations for each case.

The sample mean and standard deviation of the 500 RASE values for the one-step and two-step procedures are reported in Tables 4 and 5, respectively. Since β1(·), β2(·) and β3(·) admit similar degrees of smoothness, the one-step and two-step procedures perform similarly as expected and the RASEs decrease as the sample size increases. Figure 5 depicts the one-step estimates of the parameter functions and their 95% point-wise asymptotic confidence intervals based on a typical sample which has the median RASE value over the 500 Monte Carlo runs. The two-step estimates are similar and are not displayed here. The estimated coefficient functions are close to the underlying true coefficient functions. In the second simulation, the two-step procedure requires 6 times more computation time than the one-step procedure for n = 1000 implemented with CPU 2.3 GHz Intel Core i5.

Table 4.

RASE for One-step Estimates in Example 2

h β 1 β 2 β 3

n = 500 n = 1000 n = 2000 n = 500 n = 1000 n = 2000 n = 500 n = 1000 n = 2000
12h1 .051(.019) .035(.012) .026(.007) .172(.060) .119(.038) .091(.025) .046(.019) .034(.012) .025(.008)
h 1 .046(.019) .031(.012) .023(.008) .143(.056) .102(.038) .077(.026) .041(.019) .031(.013) .022(.009)
2h1 .105 (.027) .078(.016) .056(.011) .211(.059) .159(.041) .119(.028) .110(.020) .083(.015) .061(.011)

Table 5.

RASE for Two-step Estimates in Example 2

h β 1 β 2 β 3

n = 500 n = 1000 n = 2000 n = 500 n = 1000 n = 2000 n = 500 n = 1000 n = 2000
12h2 .044(.020) .032(.012) .023(.008) .159(.058) .111(.037) .080(.024) .041(.019) .031(.013) .023(.009)
h 2 .041(.019) .029(.012) .021(.008) .142(.052) .101(.030) .077(.018) .040(.018) .030(.013) .022(.009)
2h2 .046(.015) .037(.010) .031(.007) .210(.039) .158(.017) .118 (.009) .052(.012) .042(.009) .035(.006)

Figure 5.

Figure 5

One-step estimates for the parameter functions and their confidence intervals. The dashed lines are the estimated parameter functions, the dotted lines are the corresponding 95% confidence intervals, and the solid lines are the true parameter functions. (a), (b), and (c) are for n=500, (d), (e), and (f) for n=1000, (g), (h), and (i) for n = 2000.

Table 6 depicts the empirical bias, the SD, the SE and the SDse of β^i(u) at u = 18. Note that the range of β3(·) is much larger than those for β1(·) and β2(·). Thus, the bias and SD for β3(U) at u = 18 is much larger than those for the other two coefficients. The overall pattern in Tables 4 and 6 is similar to that in Tables 1 and 3. The RASE and bias decrease as the sample size increases. The SE is slightly less than the SD, but their difference is still less than two times the SDse. This implies that the proposed procedures work reasonably well.

Table 6.

Bias and Standard Errors for One-step Estimates in Example 2 (h = h1)

n β1(u) with u = 18 β2(u) with u = 18 β3(u) with u = 18

Bias SD SE(SDse) Bias SD SE(SDse) Bias SD SE(SDse)
500 .132 .145 .115(.024) .126 .116 .106(.010) 9.265 10.633 9.586(1.204)
1000 .088 .095 .087(.014) .094 .081 .079(.005) 7.091 7.522 7.212(.668)
2000 .073 .075 .066(.008) .078 .065 .060(.003) 5.369 5.872 5.441(.394)

Similar to Example 1, we illustrate the Wilks phenomenon by considering

H0:βj(U)=γj0vsH1:βj(U)γj0,j=1,2,3,

with seven different sets for (γ10, γ20, γ30): (ββ1 , μβ2, μβ3), (μβ1±2std(β1),μβ2 , μβ3), (μβ1 , μβ2±2std(β2),μβ3), and (μβ1 ,μβ2 ,μβ3 ±2std(β3)), where μβj = E{βj(U)} and std(βj)=var{βj(U)} for j = 1,2,3. Figure 6(a) depicts the kernel density estimates of the null distributions with these seven different sets for (γ10, γ20, γ30) based on 500 bootstrap samples. The seven densities of the null distributions are very close to each other. This implies the null distribution is not sensitive to the value of (γ10, γ20, γ30). Figure 6(a) shows that the null density is very close to a χ2 distribution with 12.1 degrees of freedom. Similar to the first simulation study, the value of the degrees of freedom is obtained by computing the mean of the test statistic values based on 500 bootstrap samples. This implies that the Wilks phenomenon holds for the nonlinear varying coefficient models.

Figure 6.

Figure 6

(a) Plots for estimated densities of the test statistic F from the proposed generalized F-test using 7 different null values. In (a), the solid curve is the density of a χ2 distribution with 12.1 degrees of freedom. (b) Estimated power functions.

Figure 6(b) depicts the power functions at levels 0.5, 0.25, 0.10, 0.05 and 0.01 under the following alternative hypothesis:

H1:βj(U,δ)=(1δ)γj0+δβj0(U),

where γj0 = E{βj(U)} for j = 1,2,3 and δ ∈ [0,0.015]. The estimated power at δ = 0 for the five significance levels are 0.450, 0.234, 0.090, 0.048 and 0.014 based on 500 Monte Carlo simulations when n = 500. Again, the test procedure keeps type I error rate very well for this model, and the power functions increase rapidly for increasing δ values.

4 APPLICATION TO THE ECOLOGICAL DATA

In this section, we apply the methodologies proposed in Section 2 for the ecological data set introduced in Section 1. The data set was collected by the AmeriFlux network during summer growing seasons (from June 1 to August 31) of 1993-1995. This study is a part of the U.S. Climate Change Research Initiative conducted by the North American Carbon Program (NACP) which is a component of U.S. Interagency Carbon Cycle Science Program. The objective of AmeriFlux in this project is to help NACP in modeling the causes of variation in the net exchange of CO2 leading to general principles that can be applied more broadly. Let Y denote the response variable NEE, X denote PAR and U denote temperature. We fit the following nonlinear varying coefficient model

Y=β1(U)β2(U)XX+β3(U)+ε,

using the iterative algorithm proposed in Section 2.1. Since photosynthetic activity is low at extreme temperatures, we analyze a data subset consisting of observations with temperature values between 11.5 and 24.7. The total sample size of this data subset is n=4636. The bandwidth was chosen via a multi-fold cross-validation. Specifically, we minimize the cross-validation score

CV(h)=jYjY^j2,

where Y^j denotes the fitted value with jth group of 20 subjects excluded. This results in a bandwidth of h = 3.0.

Figures 7(a), (b), and (c) depict the estimated regression coefficients along with the proposed 95% point-wise asymptotic confidence intervals. From Figures 7(a), (b), and (c), we can see that the estimated coefficient functions vary with temperature; as will be described below the proposed F-test rejects the null hypothesis of a regular nonlinear model in favor of the nonlinear varying coefficient model with a p-value of 0.00036. In our data, we have mixed species plants in a deciduous broadleaf forest. Figure 7(a) shows that these plants have higher dark respiration activity at higher temperatures (T ≥ 16°C ) than lower temperatures (T < 16°C). According to Figure 7(b), at their light-saturation point, the species in our sample achieve the maximum net photosynthetic rate β2(·) of 36.5 around 12.5°C, beyond which the net photosynthetic rate experiences a period of decrease and stabilizes. In other words, at their light-saturation point, maximum photosynthetic capacity occurs at lower temperatures and the lowest activity is observed around 16°C. Similarly β3(·) achieves its maximum value shortly after 12°C, and beyond which point it declines and stabilizes. Figure 7(d) shows the estimated apparent quantum yield β^2()β^3() along with the 95% point-wise asymptotic confidence intervals. The apparent quantum yield increases with temperature until 19.5°C and stabilizes for higher temperatures. Note that the proposed inference assumes i.i.d observations and hence confidence intervals may be narrower than those that would take into account the correlation structure over time in the data set. Extension of the proposed inference framework for correlated data is of interest.

Figure 7.

Figure 7

(a), (b), and (c) One-step estimates for the varying coefficient functions (dashed) of the nonlinear varying coefficient model fit to the ecological data along with 95% point-wise asymptotic confidence intervals (dotted). The solid lines are the estimate of βj under the null hypothesis. (d) Estimated apparent quantum yield (solid) along with 95% point-wise asymptotic confidence intervals (dotted).

This application illustrates the advantages of the proposed nonlinear varying coefficient models outlined in the introduction. First, the dimension is reduced by regression NEE on only PAR in a local temperature window rather than considering a two-dimensional nonparametric regression of NEE on both PAR and temperature. Second, the proposed model reduces the modeling bias by allowing the regression coefficients to vary over temperature; the nonlinear varying coefficient model fit leads to smaller multifold cross validation error compared to the regular nonlinear regression model fit. Finally, as illustrated by the above interpretations, the fits from the nonlinear varying coefficient model can be interpreted as describing photosynthesis via separate nonlinear regression models for each fixed temperature value. This parallels the interpretation that a varying coefficient model can be viewed as a linear regression model at a fixed U = u.

For comparison, we also obtained the two-step estimates proposed in Section 2.2. The initial bandwidth is chosen to be h0 = 1.5, which is half of the bandwidth obtained by using cross-validation in the one-step procedure. After obtaining the initial estimators, we apply a cross-validation bandwidth selector for each coefficient separately, and obtain the optimal bandwidths h2 = 2.0, 2.5 and 3.0 for β1(·), β2(·), and β3(·), respectively. Figure 8 shows the two-step estimates of regression coefficients along with 95% percentile bootstrap confidence intervals based on 500 bootstrap samples. To have a sensible comparison between the two-step procedure and the one-step procedure, we rerun the one-step procedure with bandwidths 2.0, 2.5 and 3.0. These estimates are also shown in Figure 8. The two-step estimates are very close to the one-step estimates; however, they are smoother. To compare the two-methods, we also computed the ratio of residual sum of squares, that is RSS(one_step)= RSS(two_step), and found this ratio to be very close to 1. The two-step procedure requires 8 times more computation time than the one-step procedure implemented with CPU 2.3 GHz Intel Core i5.

Figure 8.

Figure 8

Two-step estimates for the varying coefficient functions (solid) of the nonlinear varying coefficient model fit to the ecological data along with one-step estimates (dashed) and 95% percentile bootstrap confidence intervals based on 500 bootstrap samples (dotted).

We apply the testing procedure proposed in Section 2.4 to test whether the regression coefficients are actually varying over temperature. In order to do so, we fit the nonlinear model (1.1) to the data and use the resulting estimates (β^1,β^2,β^3) = (4.809,31.940,562.677) as the null value for testing whether the functions are constant over temperature. Figure 9 depicts the estimated density function of the null distribution of the test statistic F, overlaying a chi-square density. The degrees of freedom of the chi-square density is chosen close to the sample mean of test statistic values obtained from 500 bootstrap samples, which is approximately 18.5. The figure is consistent with the Wilks phenomenon stating that under the null hypothesis the chisquare distribution is a good approximation for the asymptotic behavior of F. We reject the null hypothesis at the significance level α = 0.05 in favor of the newly proposed nonlinear varying coefficient model with a test statistic of 46.209 and a p-value of 0.00036. In other words, the proposed nonlinear varying coefficient model fits the data better than the nonlinear regression model (1.1).

Figure 9.

Figure 9

Estimated density of the test statistic F (dashed) based on 500 bootstrap samples, overlaying the density of a χ2 distribution with 18.5 degrees of freedom (solid).

5 DISCUSSION

Motivated by an empirical analysis of a real data set, we proposed nonlinear varying coefficients models. We further developed two estimation procedures based on local linear regression techniques and studied their finite sample properties. Similar to estimation in regular varying coefficient models, the two-step estimators for nonlinear varying coefficient models also lead to efficiency gains when the varying coefficient functions admit different degrees of smoothness, and are less computationally efficient than the one-step estimators. Asymptotic properties of the one-step estimator is established and those for the two-step estimator requires further research. Alternative estimation approaches for nonlinear varying coefficient models may be developed via basis expansions such as smoothing splines (Hastie and Tibshirani 1993). Although splines have many advantages, some challenges such as the difficulty in choosing all the smoothing parameters simultaneously and their computational intensity have been reported (Fan and Zhang 1999).

The generalized F test is proposed for testing whether the nonlinear varying coefficient model reduces to a nonlinear regression model, i.e. whether all the varying coefficient functions are constant over U. For assessing features of single varying coefficient functions, we proposed asymptotic and bootstrap confidence intervals. Developing valid hypothesis testing procedures for specific features of single varying coefficient functions would also be of interest and requires further research.

The proposal is based on a univariate index covariate U. Estimation in varying coefficient models in higher index dimensions quickly runs into issues of ‘curse of dimensionality’. Proposals for incorporating multi-dimensional indices include Xia and Li (1999); Fan, Yao, and Cai (2003) where authors consider a linear combination of a multidimensional U vector to create to a one-dimensional index. Similar extensions can be considered in the case of nonlinear varying coefficient models.

As extensions of generalized varying coefficient models and model (1.4), it may be of interest to consider a new class of following models:

E(YX,U)=g[f{X,β(U)}], (5.1)

where f(· , ·) is pre-specified, while g(·) is unspecified, nonparametric smooth function. This class of models include (1.4) as a special case by taking the link function g(·) to be the identical link. With a pre-specified link function, model (5.1) also includes generalized varying coefficient models as special cases by taking f(X,β) = XT β(U). Fan, Yao and Cai (2003) proposed adaptive varying coefficient models, in which the link function f(·) is an unknown nonparametric smooth function, and f(X,β) = XT β(U). The authors proposed an approach to estimating the unknown link function and β(·) iteratively.

In practice, data may be collected over time. Thus, the data would be correlated. The proposed estimation procedures are directly applicable for correlated data. It will involve theoretical techniques in time series to establish the sampling properties of the proposed procedure for correlated data. Thus, it may be an interesting future research topic to establish theoretical property of our proposed procedure under the setting of dependent data.

Supplementary Material

supple

Acknowledgments

His research was supported by National Institute on Drug Abuse grant P50-DA10075, National Cancer Institute (NCI) grant R01 CA168676, and National Natural Science Foundation of China grant 11028103, Her research was supported by National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK) grant R01 DK092232. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIDA, NCI, NIDDK or the NIH.

Contributor Information

Esra Kürüm, Department of Statistics, Istanbul Medeniyet University, Istanbul, Turkey. (esra.kurum@medeniyet.edu.tr)..

Runze Li, Department of Statistics and The Methodology Center, The Pennsylvania State University, University Park, PA 16802-2111, USA (rzli@psu.edu).

Yang Wang, Credit & Counterparty Risk Management, McLean, VA 22102, USA (yangwang@freddiemac.com)..

Damla ŞEntürk, Department of Biostatistics, University of California, Los Angeles, CA 90095, USA (dsenturk@ucla.edu)..

References

  1. Breiman L, Friedman JH. Estimating optimal transformations for multiple regression and correlation. Journal of the American Statistical Association. 1985;80:580–619. [Google Scholar]
  2. Cai Z, Fan J, Li R. Efficient estimation and inferences for varying-coefficient models. Journal of the American Statistical Association. 2000;95:888–902. [Google Scholar]
  3. Cleveland W, Grosse E, Shyu W. Local regression models. In: Chambers JM, Hastie TJ, editors. Statistical Models in S. Wadsworth & Brooks/Cole; Pacific Grove, CA: 1992. pp. 309–376. [Google Scholar]
  4. Fan J, Gijbels I. Local polynomial modelling and its applications. Chapman and Hall; London: 1996. [Google Scholar]
  5. Fan J, Li R. New estimation and model selection procedures for semiparametric modeling in longitudinal data analysis. Journal of the American Statistical Association. 2004;99:710–723. URL http://www.jstor.org/stable/27590442. [Google Scholar]
  6. Fan J, Yao Q, Cai Z. Adaptive varying-coefficient linear models. (Series B).Journal of the Royal Statistical Society. 2003;65:57–80. URL http://dx.doi.org/10.1111/1467-9868.00372. [Google Scholar]
  7. Fan J, Zhang C, Zhang J. Generalized likelihood ratio statistics and wilks phenomenon. Annals of Statistics. 2001;29:153–193. [Google Scholar]
  8. Fan J, Zhang W. Statistical estimation in varying coefficient models. Annals of Statistics. 1999;27:1491–1518. [Google Scholar]
  9. — Statistical methods with varying coefficient models. Statistics and Its Interface. 2008;1:179–195. doi: 10.4310/sii.2008.v1.n1.a15. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Friedman JH. Multivariate adaptive regression splines. Annals of Statistics. 1991;19:1–67. [Google Scholar]
  11. Garrett RH, Grisham CM. Biochemistry. 3rd ed Thomson Learning; Belmont, CA: 2005. [Google Scholar]
  12. Geisser S. The predictive sample reuse method with applications. Journal of the American Statistical Association. 1975;70:320–328. [Google Scholar]
  13. Green PJ, Silverman BW. Nonparametric regression and generalized linear models: a roughness penalty approach. Chapman and Hall; London: 1994. [Google Scholar]
  14. Gu C, Wahba G. Smoothing spline anova with component-wise bayesian “confidence intervals. Journal of Computational and Graphical Statistics. 1993;2:97–117. [Google Scholar]
  15. Härdle W, Stoker TM. Investigating smooth multiple regression by the method of average derivatives. Journal of the American Statistical Association. 1989;84:986–995. [Google Scholar]
  16. Hastie T, Tibshirani R. Generalized additive models. Chapman and Hall; New York: 1990. [DOI] [PubMed] [Google Scholar]
  17. — Varying-coefficient models. (Series B).Journal of the Royal Statistical Society. 1993;55:757–796. [Google Scholar]
  18. Hoover DR, Rice JA, Wu CO, Yang L-P. Nonparametric smoothing estimates of time-varying coefficient models with longitudinal data. Biometrika. 1998;85:809–822. [Google Scholar]
  19. Kauermann G, Tutz G. On model diagnostics using varying coefficient models. Biometrika. 1999;86:119–128. [Google Scholar]
  20. Li K-C. Sliced inverse regression for dimension reduction. Journal of the American Statistical Association. 1991;86:316–327. [Google Scholar]
  21. Monteith JL. Solar radiation and productivity in tropical ecosystems. Journal of Applied Ecology. 1972;9:747–766. [Google Scholar]
  22. Osborne MR. Fisher’s method of scoring. International Statistical Review/Revue Internationale de Statistique. 1992;60:99–117. [Google Scholar]
  23. Seber GAF, Wild CJ. Nonlinear regression. Wiley-Interscience; 2003. [Google Scholar]
  24. Stone CJ, Hansen MH, Kooperberg C, Truong YK. Polynomial splines and their tensor products in extended linear modeling: 1994 wald memorial lecture. Annals of Statistics. 1997;25:1371–1470. [Google Scholar]
  25. Wahba G. Partial spline models for the semiparametric estimation of functions of several variables. Statistical Analysis of Time Series, Proceedings of the Japan U.S. Joint Seminar; Tokyo. Institute of Statistical Mathematics; 1984. pp. 319–329. [Google Scholar]
  26. Wu CO, Chiang C-T, Hoover DR. Asymptotic confidence regions for kernel smoothing of a varying-coefficient model with longitudinal data. Journal of the American Statistical Association. 1998;93:1388–1402. [Google Scholar]
  27. Xia Y, Li WK. On single-index coefficient regression models. Journal of the American Statistical Association. 1999;94:1275–1285. URL http://www.jstor.org/stable/2669941. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

supple

RESOURCES