Skip to main content
Springer logoLink to Springer
. 2018 Sep 10;2018(1):231. doi: 10.1186/s13660-018-1832-6

The asymptotic normality of internal estimator for nonparametric regression

Penghua Li 1, Xiaoqin Li 2,, Liping Chen 3
PMCID: PMC6132389  PMID: 30839660

Abstract

In this paper, we aim to study the asymptotic properties of internal estimator of nonparametric regression with independent and dependent data. Under some weak conditions, we present some results on asymptotic normality of the estimator. Our results extend some corresponding ones.

Keywords: Asymptotic normality, Nonparametric regression, Internal estimator, Dependent data

Introduction

In this paper, we consider the nonparametric regression model

Yi=m(Xi)+Ui,1in,n1,

where (Xi,Yi)Rd×R, d1, and Ui are random variables satisfying E(Ui|Xi)=0, 1in, n1. So we have

E(Yi|Xi=x)=m(x),i1.

Let K(x) be a kernel function. Define Kh(x)=hdK(x/h), where h=hn is a sequence of positive bandwidths tending to zero as n. Kernel-type estimators of the regression function are widely used in various situations because of their flexibility and efficiency in the dependent and independent data. For the independent data, Nadaraya [1] and Watson [2] gave the most popular nonparametric estimator of the unknown function m(x) named the Nadaraya–Watson estimator mˆNW(x):

mˆNW(x)=i=1nYiKh(xXi)i=1nKh(xXi). 1.1

Jones et al. [3] considered various versions of kernel-type regression estimators such as the Nadaraya–Watson estimator (1.1) and the local linear estimator. They also investigated the internal estimator

mˆn(x)=1ni=1nYiKh(xXi)f(Xi) 1.2

for a known density f(). Here the factor 1f(Xi) is internal to the summation, whereas the estimator mˆNW(x) has the factor 1fˆ(x)=1n1i=1nKh(xXi) externally to the summation.

The internal estimator was first proposed by Mack and Müller [4]. Jones et al. [3] studied various kernel-type regression estimators, including the introduced internal estimator (1.2). Linton and Nielsen [5] introduced an integration method based on direct integration of initial pilot estimator (1.2). Linton and Jacho-Chávez [6] studied the other internal estimator

m˜n(x)=1ni=1nYiKh(xXi)fˆ(Xi), 1.3

where fˆ(Xi)=1nj=1nLb(XiXj) and Lb()=L(/b)/bd. Here L() is a kernel function, b is the bandwidth, and the density f() is unknown. Under the independent data, Linton and Jacho-Chávez [6] obtained the asymptotic normality of the internal estimator m˜n(x) in (1.3). Shen and Xie [7] obtained the complete convergence and uniform complete convergence of internal estimator mˆn(x) in (1.2) under the geometrical α-mixing (or strong mixing) data. Li et al. [8] weakened the conditions of Shen and Xie [7] and obtained the convergence rate and uniform convergence rate for the estimator mˆn(x) in probability.

As far as we know, there are no results on asymptotic normality of the internal estimator mˆn(x). Similarly to Linton and Jacho-Chávez [6], we investigate the asymptotic normality of the internal estimator mˆn(x) with independent data and φ-mixing data, respectively. Asymptotic normality results are presented in Sect. 3.

Denote Fnm=σ(Xi,nim) and define the coefficients

φ(n)=supm1supAF1m,BFm+n,P(A)0|P(B|A)P(B)|.

If φ(n)0 as n, then {Xn}n1 is said to be a φ-mixing sequence.

The concept of φ-mixing is introduced by Dobrushin [9], and many properties of φ-mixing are presented in Chap. 4 of Billingsley [10]. If the coefficient of the process is geometrically decreasing, then the autoregressive moving average (ARMA) process can construct a geometric φ-mixing sequence. Györfi et al. [11, 12] gave more examples and applications to nonparametric estimation. We can also refer to Fan and Yao [13] and Bosq and Blanke [14] for the works on nonparametric regression under independent and dependent data.

Regarding notation, for x=(x1,,xd)Rd, set x=max(|x1|,,|xd|). Throughout the paper, c,c1,c2,c3,,d,B0,B1 denote some positive constants not depending on n, which may be different in various places, x denotes the largest integer not exceeding x, → means to take the limit as n, and cndn means that cndn1, D means the convergence in distribution, and X=DY means that random variables X and Y have the same distribution. A sequence {Xi,i1} is said to be second-order stationary if (X1,X1+k)=D(Xi,Xi+k) for i1, k1.

Some assumptions

In this section, we list some assumptions.

Assumption 2.1

There exist two positive constants K¯>0 and μ>0 such that

supuRd|K(u)|K¯andRd|K(u)|du=μ. 2.1

Assumption 2.2

Let Sf denote the compact support of known density f() of X1. For xSf, the function m(x) is twice differentiable, and there exists a positive constant b such that

|2m(x)xixj|b,i,j=1,2,,d.

The kernel density function is symmetric and satisfies

Rd|vi||vj|K(v)dv<,i,j=1,2,,d.

Assumption 2.3

We assume the data observed {(Xi,Yi),i1} is an independent and identically distributed stochastic sequence with values in Rd×R. The known density f() of X1 is upon its compact support Sf and such that infxSff(x)>0. For 0<δ1, we suppose that

E|Y1|2+δ< 2.2

and

supxSfE(|Y1|2+δ|X1=x)f(x)B0<. 2.3

Assumption 2.3

We assume that the data observed {(Xi,Yi),i1} is a second-order stationary stochastic sequence with values in Rd×R. The sequence {(Xi,Yi),i1} is also assumed to be φ-mixing with n=1φ1/2(n)<. The known density f() of X1 is upon its compact support Sf and such that infxSff(x)>0. Let (2.2) and (2.3) be fulfilled. Moreover, for all j1, we have

supx1Sf,xj+1SfE(|Y1Yj+1||X1=x1,Xj+1=xj+1)fj(x1,xj+1)B1<, 2.4

where fj(x1,xj+1) denotes the joint density of (X1,Xj+1).

Remark 2.1

Assumption 2.1 is a usual condition on the kernel function, and Assumption 2.2 is used to get the convergence rate of |Emˆn(x)m(x)|. Assumptions 2.3 and 2.3 are the conditions of independent and dependent data {(Xi,Yi),i1}, respectively. Similarly to Hansen [15], conditions (2.2) and (2.3) are used to control the tail behavior of the conditional expectation E(|Y1|2+δ|X1=x), and (2.4) is used to estimate the covariance Cov(Y1,Yj+1).

Asymptotic normality of internal estimator mˆn(x) with independent and dependent data

In this section, we show some results on asymptotic normality of the internal estimator of a nonparametric regression model with independent and dependent data. Theorem 3.1 is for independent data, and Theorem 3.2 is for φ-mixing data.

Theorem 3.1

Let Assumptions 2.12.3 hold, and let limuudK(u)=0. Suppose that E(Y12|X1=x)f(x) is positive and continuous at point xSf. If 0<hd0, nhd, and nhd+40 as n, then

nhd[mˆn(x)m(x)]DN(0,σ2(x)), 3.1

where σ2(x)=E(Y12|X1=x)f(x)RdK2(u)du.

Theorem 3.2

Let the conditions of Theorem 3.1 be fulfilled, where Assumption 2.3 is replaced by Assumption 2.3. Then (3.1) holds.

Remark 3.1

The choice of a positive bandwidth h is easy to design. For example, with d1, if h=nβ and β(1d+4,1d), then the conditions 0<hd0, nhd, and nhd+40 are satisfied as n.

Conclusion

Linton and Jacho-Chávez [6] obtained some asymptotic normality results of the internal estimator m˜n(x) under independent data. Comparing Theorem 1 and Corollary 1 of Linton and Jacho-Chávez [6], our asymptotic normality results on the internal estimator mˆn(x) in Theorems 3.1 and 3.2 are relatively simple. Meanwhile, we use the method of Bernstein’s big-block and small-block and the inequalities of φ-mixing random variables to investigate the asymptotic normality of the internal estimator mˆn(x) for m(x), and we also obtain the asymptotic normality result of (3.1). Obviously, α-mixing is weaker than φ-mixing, but some moment inequalities of α-mixing are more complicated than those of φ-mixing [16, 17]. For simplicity, we study the asymptotic normality of internal estimator mˆn(x) under φ-mixing and obtain the asymptotic normality result of Theorem 3.2.

Some lemmas and the proofs of main results

Lemma 5.1

(Liptser and Shiryayev [18], Theorem 9 in Sect. 5)

Let (ξnk,Hkn)k1 be martingale differences (i.e. H0n={,Ω}, HknHk+1n, ξnk is an Hkn-measurable random variable, E(ξnk|Hk1n)=0 a.s., for all k1 and n1) with Eξnk2< for all k1 and n1. Let (γn)n1 be a sequence of Markov times with respect to (Hkn)k0, taking values in the set {0,1,2,}. If

k=1γnE(ξnk2I(|ξnk|>δ)|Hk1n)P0,δ(0,1],k=1γnE(ξnk2|Hk1n)Pσ2,

then

k=1γnξnkDN(0,σ2).

Lemma 5.2

(Billingsley [10], Lemma 1)

If ξ is measurable with respect to Mk and η is measurable with respect to Mk+n (n0), then

E|ξ|r<,E|η|s<,r,s>1,r1+s1=1,

implies

|E(ξη)E(ξ)E(η)|2φ1r(n)(E|ξ|r)1r(E|η|s)1s.

Lemma 5.3

(Yang [16], Lemma 2)

Let p2, and let {Xn}n1 be a φ-mixing sequence with n=1φ1/2(n)<. If EXn=0 and E|Xn|p< for all n1, then

E|i=1nXi|pC(i=1nE|Xi|p+(i=1nEXi2)p/2),

where C is a positive constant depending only on φ().

Lemma 5.4

(Fan and Yao [13], Proposition 2.6)

Let Fij and α() be the same as in (2.57) of Fan and Yao [13]. Let ξ1,ξ2,,ξk be complex-valued random variables measurable with respect to the σ-algebras Fi1j1,,Fikjk, respectively. Suppose il+1jln for l=1,,k1 and jlil and P(|ξl|1)=1 for l=1,2,,k. Then

|E(ξ1ξk)E(ξ1)E(ξk)|16(k1)α(n).

Proof of Theorem 3.1

It is easy to see that

nhd(mˆn(x)m(x))=nhd([mˆn(x)Emˆn(x)]+[Emˆn(x)m(x)]). 5.1

Combining Assumption 2.2 with the proof of Lemma 2 of Shen and Xie [7], we obtain that

|Emˆn(x)m(x)|=O(h2),xSf.

Then, it follows from nhd+40 that

nhd[Emˆn(x)m(x)]=O(nhd+4)0,xSf. 5.2

For xSf, let Zi:=hdYiKh(xXi)f(Xi), 1in. Denote

nhd[mˆn(x)Emˆn(x)]=1ni=1nhd[YiKh(xXi)f(Xi)EYiKh(xXi)f(Xi)]=1ni=1n(ZiEZi). 5.3

To prove (3.1), we apply (5.1)–(5.3) and have to show that

nhd[mˆn(x)Emˆn(x)]=1ni=1n(ZiEZi)DN(0,σ2(x)), 5.4

where σ2(x) is defined by (3.1).

Combining the independent and identically distributed stochastic sequence of {(Xi,Yi),i1} with Lemma 5.1, to prove (5.4), we have to show that

1ni=1nE(ZiEZi)2=Var(Z1)σ2(x) 5.5

and, for all λ(0,1],

1ni=1nE((ZiEZi)2I(|ZiEZi|n>λ))0. 5.6

Obviously, for any 1r2+δ (0<δ1), by (2.1) and (2.3) we have

hd(r1)E|Kh(xX1)Y1f(X1)|r=hd(r1)E(|Kh(xX1)|rfr(X1)E(|Y1|r|X1))=Sf|K(xuh)|rE(|Y1|r|X1=u)1hdf(u)fr(u)duSf|K(xuh)|r(E(|Y1|2+δ|X1=u)f(u))r2+δ1hd1f(3+δ)r2+δ1(u)du(B0)r2+δK¯r1μ(infxSff(x))(3+δ)r2+δ1:=μ¯(r)<. 5.7

By (5.7) with r=1 this yields

(EZ1)2=hd(EKh(xX1)Y1f(X1))2chd0. 5.8

Define

g(x)={E(Y12|X1=x)f(x)if xSf,0otherwise.

In view of condition (2.3), we have

Rdg(x)dx=SfE(Y12|X1=x)f(x)dx=SfE(Y12|X1=x)f22+δ(x)f4+δ2+δ(x)dxSf(E(|Y1|2+δ|X1=x))22+δf22+δ(x)f4+δ2+δ(x)dxB022+δ(infxSff(x))6+2δ2+δRdf(x)dx=B022+δ(infxSff(x))6+2δ2+δ<.

So we have g(x)L1. Since that E(Y12|X1=x)f(x) is positive and continuous at a point xSf and limuudK(u)=0, we obtain by Bochner lemma [14] that

E(Z12)=hdE(Kh(xX1)Y1f(X1))2=SfK2(xuh)E(Y12|X1=u)1hd1f(u)du=RdK2(xuh)1hdg(u)duE(Y12|X1=x)f(x)RdK2(u)du. 5.9

Then, it follows from (5.8) and (5.9) that, for xSf,

Var(Z1)E(Y12|X1=x)f(x)RdK2(u)du=σ2(x), 5.10

which implies (5.6). Meanwhile, for some δ(0,1] and any λ(0,1], by Cr inequality and (5.7) we get that

1ni=1nE((ZiEZi)2I(|ZiEZi|n>λ))=E((Z1EZ1)2I(|Z1EZ1|n>λ))1nδ2λδE|Z1EZ1|2+δc1nδ2λδE|Z1|2+δc2(nhd)δ20, 5.11

since nhd. Thus, (5.6) follows from (5.11). Consequently, the proof of the theorem is completed. □

Proof of Theorem 3.2

We use the same notation as in the proof of Theorem 3.1. Under the conditions of Theorem 3.2, by (5.1), (5.2), and (5.3), to prove (3.1), we need to show that

nhd[mˆn(x)Emˆn(x)]=1ni=1n(ZiEZi)DN(0,σ2(x)), 5.12

where σ2(x) is defined by (3.1). By the second-order stationarity, {(Xi,Yi),i1} are identically distributed. Then, for 1in, we have by (5.8) and (5.9) that

Var(ZiEZi)=Var(Z1)E(Y12|X1=x)f(x)RdK2(u)du=σ2(x). 5.13

For j1, in view of (2.4), we have

E|Kh(xX1)Kh(xXj+1)Y1Yj+1f(X1)f(Xj+1)|=E(|Kh(xX1)Kh(xXj+1)|f(X1)f(Xj+1)E(|Y1Yj+1||X1,Xj+1))=SfSf|K(xu1h)K(xuj+1h)|E(|Y1Yj+1||X1=u1,Xj=uj+1)×1h2d1f(u1)f(uj+1)fj(u1,uj+1)du1duj+1B1(infxSff(x))2RdRd|K(xu1h)K(xuj+1h)|1h2ddu1duj+1B1μ2(infxSff(x))2<. 5.14

So it follows from (5.7) and (5.14) that

|Cov(Z1,Zj)|E|Z1Zj|+(E|Z1|)2c1hd,j>1. 5.15

Obviously, by the stationarity we establish that

1nVar(i=1n(ZiEZi))=1nVar(i=1nZi)=Var(Z1)+2n1i<jnCov(Zi,Zj)=Var(Z1)+2n{[1i<jn1jirn+1i<jnji>rn]Cov(Zi,Zj)}. 5.16

For hd, we can choose rn satisfying that rn and hdrn0 as n. So, by (5.15),

2n1i<jn1jirn|Cov(Zi,Zj)|chdrn0. 5.17

By Lemma 5.2 with s=r=2, the condition n=1φ1/2(n)<, and (5.9), we can show that

2n1i<jnji>rn|Cov(Zi,Zj)|c1n1i<jnji>rnφ1/2(ji)c2k>rnφ1/2(k)0. 5.18

Therefore, by (5.13), (5.16), (5.17), and (5.18), we get that

1nVar(i=1nZi)=σ2(x)(1+o(1)).

Next, we employ Bernstein’s big-block and small-block procedure (see Fan and Yao [13] and Masry [19]). Partition the set {1,2,,n} into 2kn+1 subsets with large block of size μ=μn and small block of size ν=νn and set

k=kn=nμn+νn. 5.19

Define μ=μn=nhd and ν=νn=nhd. So we have by hd0 and nhd that

μn,νn,μnn0,νnn0,νnμn0,kn=O(nhd). 5.20

Define ηj, ξj, and ζj as follows:

ηj:=i=j(μ+ν)+1j(μ+ν)+μ(ZiEZi),0jk1, 5.21
ξj:=i=j(μ+ν)+μ+1(j+1)(μ+ν)(ZiEZi),0jk1, 5.22
ζk:=i=k(μ+ν)+1n(ZiEZi). 5.23

In view of

Sn:=i=1n(ZiEZi)=j=0k1ηj+j=0k1ξj+ζk:=Sn+Sn′′+Sn′′′, 5.24

we have to show that

1nE(Sn′′)20,1nE(Sn′′′)20, 5.25
|E(exp(itn1/2Sn))j=0k1E(exp(itn1/2ηj))|0, 5.26
1nj=0k1E(ηj2)σ2(x), 5.27
1nj=0k1E(ηj2I(|ηj|>εσ(x)n))0,ε>0. 5.28

Relation (5.25) implies that Sn′′n and Sn′′′n are asymptotically negligible, (5.26) shows that the summands {ηj} in Sn are asymptotically independent, and (5.27)–(5.28) are the standard Lindeberg–Feller conditions for the asymptotic normality of Sn under independence.

First, we prove (5.25). By (5.22) and (5.24) we have

E(Sn′′)2=Var(j=0k1ξj)=j=0k1Var(ξj)+20i<jk1Cov(ξi,ξj):=F1+F2. 5.29

By the stationarity and (5.10), similarly to the proof of (5.17) and (5.18), for 0jk1, we have

Var(ξj)=νnVar(Z1)+21i<jνnCov(Zi,Zj)=νnσ2(x)(1+o(1)). 5.30

Thus it follows from (5.19) and (5.20) that

F1=knνnσ2(x)(1+o(1))nνnμn+νnnνnμn=o(n). 5.31

We consider the term F2 in (5.29). With λj=j(μn+νn)+μn,

F2=20i<jk1l1=1νnl2=1νnCov(Zλi+l1,Zλj+l2),

but since ij, |λiλj+l1l2|μn for 0i<jk1, 1l1νn, and 1l2νn, similarly to the proof of (5.18), it follows that

|F2|21i<jnjiμn|Cov(Zi,Zj)|=o(n). 5.32

Hence by (5.29), (5.31), and (5.32) we have

1nE(Sn′′)20.

By (5.13), (5.20), and (5.23), similarly to the proofs of (5.17) and (5.18), we can find that

1nE(Sn′′′)21n(nkn(μn+νn))Var(Z1)+2n1i<jnkn(μn+νn)|Cov(Zi,Zj)|Cμn+νnnσ2(x)+o(1)0.

Thus

1nSn=1n(Sn+Sn′′+Sn′′′)=1nSn+op(1). 5.33

Second, it is easy to see that φ1/2(n)=o(1n) by φ(n)0 and n=1φ1/2(n)<. Note that ηa is Miaja-measurable with ia=a(μ+ν)+1 and ja=a(μ+ν)+μ. Since φ-mixing random variables are strong mixing random variables and α(n)φ(n), letting Vj=exp(itn1/2ηj), by Lemma 5.4 we have

|E(exp(itn1/2Sn))j=0k1E(exp(itn1/2ηj))|cknφ(νn+1)cnμn+νn1νn2cnhd0

by (5.19), (5.20), and the conditions hn0 and nhd as n.

Third, we show (5.27), where ηj is defined in (5.21). By the stationarity and (5.30) with μn replacing νn, we have

E(ηj2)=Var(ηj)=Var(η0)=μnσ2(x)(1+o(1)),0jk1, 5.34

so that

1nj=0kn1E(ηj2)=knμnnσ2(x)(1+o(1))σ2(x),

since knμn/n1.

Fourth, it is time to establish (5.28). Obviously, by (5.7) we obtain that

EZi2=EZ12c1andE|Zi|2+δ=E|Z1|2+δc2(hd)δ2,1in.

We can see that 1hdμnchdnhd=c(nhd)120, since nhd as n. Therefore, by Lemma 5.3 with n=1φ1/2(n)< we have that

E|i=1μn(ZiEZi)|2+δc1(i=1μnE|Zi|2+δ+(i=1μnEZi2)2+δ2)c2(μn1(hd)δ2+μn1+δ2)c3μn1+δ2. 5.35

Then, for all ε>0, by (5.34) and (5.35) it is easy to see that

E(η02I(|η0|εσ(x)n1/2))1(εσ(x)n1/2)δE|η0|2+δI(|η0|εσ(x)n1/2)1(εσ(x)n1/2)δE|η0|2+δc1μn1+δ2nδ2.

Similarly, for 0jk1, we get that

E(ηj2I(|ηj|εσ(x)n1/2))c2μn1+δ2nδ2.

Therefore, since 0<δ1 and nhd, we obtain that, for all ε>0,

1nj=0k1E(ηj2I(|ηj|εσ(x)n1/2))ckμn1+δ2n1+δ2cμn1+δ2nμn+νnn1+δ2cμnδ2nδ2=c(nhdn)δ2=c(nhd)δ40.

Therefore, (5.26), (5.27), and (5.28) hold for Sn, so that

1nSnDN(0,σ2(x)). 5.36

Consequently, (5.12) follows from (5.33) and (5.36). Finally, by (5.1), (5.2), and (5.12) we obtain (3.1). The proof of theorem is completed. □

Authors’ contributions

All authors contributed equally to the writing of this paper. All authors read and approved the final manuscript.

Funding

This work is supported by National Natural Science Foundation of China (11501005, 11701004, 61403115), Common Key Technology Innovation Special of Key Industries (cstc2017zdcy-zdyf0252), Artificial Intelligence Technology Innovation Significant Theme Special Project (cstc2017rgzn-zdyf0073, cstc2017rgzn-zdyf0033), Natural Science Foundation of Chongqing (cstc2018jcyjA0607), Natural Science Foundation of Anhui (1808085QA03, 1808085QF212, 1808085QA17) and Provincial Natural Science Research Project of Anhui Colleges (KJ2016A027, KJ2017A027).

Competing interests

The authors declare that they have no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Penghua Li, Email: lipenghua88@163.com.

Xiaoqin Li, Email: lixiaoqin1983@163.com.

Liping Chen, Email: lip_chenhut@126.com.

References

  • 1.Nadaraya E.A. On estimating regression. Theory Probab. Appl. 1964;9:141–142. doi: 10.1137/1109020. [DOI] [Google Scholar]
  • 2.Watson G.S. Smooth regression analysis. Sankhya, Ser. A. 1964;26:359–372. [Google Scholar]
  • 3.Jones M.C., Davies S.J., Park B.U. Versions of kernel-type regression estimators. J. Am. Stat. Assoc. 1994;89:825–832. doi: 10.1080/01621459.1994.10476816. [DOI] [Google Scholar]
  • 4.Mack Y.P., Müller H.G. Derivative estimation in nonparametric regression with random predictor variable. Sankhya. 1989;51:59–72. [Google Scholar]
  • 5.Linton O., Nielsen J. A kernel method of estimating structured nonparametric regression based on marginal integration. Biometrika. 1995;82:93–100. doi: 10.1093/biomet/82.1.93. [DOI] [Google Scholar]
  • 6.Linton O., Jacho-Chávez D. On internally corrected and symmetrized kernel estimators for nonparametric regression. Test. 2010;19:166–186. doi: 10.1007/s11749-009-0145-y. [DOI] [Google Scholar]
  • 7.Shen J., Xie Y. Strong consistency of the internal estimator of nonparametric regression with dependent data. Stat. Probab. Lett. 2013;83:1915–1925. doi: 10.1016/j.spl.2013.04.027. [DOI] [Google Scholar]
  • 8.Li X.Q., Yang W.Z., Hu S.H. Uniform convergence of estimator for nonparametric regression with dependent data. J. Inequal. Appl. 2016;2016:142. doi: 10.1186/s13660-016-1087-z. [DOI] [Google Scholar]
  • 9.Dobrushin R.L. The central limit theorem for non-stationary Markov chain. Theory Probab. Appl. 1956;1:72–88. [Google Scholar]
  • 10.Billingsley P. Convergence of Probability Measures. New York: Wiley; 1968. [Google Scholar]
  • 11.Györfi L., Härdle W., Sarda P., Vieu P. Nonparametric Curve Estimation from Time Series. Berlin: Springer; 1989. [Google Scholar]
  • 12.Györfi L., Kohler M., Krzyżak A., Walk H. A Distribution-Free Theory of Nonparametric Regression. New York: Springer; 2002. [Google Scholar]
  • 13.Fan J.Q., Yao Q.W. Nonlinear Time Series: Nonparametric and Parametric Methods. New York: Springer; 2003. [Google Scholar]
  • 14.Bosq D., Blanke D. Inference and Prediction in Large Dimensions. Chichester: Wiley; 2007. [Google Scholar]
  • 15.Hansen B.E. Uniform convergence rates for kernel estimation with dependent data. Econom. Theory. 2008;24:726–748. doi: 10.1017/S0266466608080304. [DOI] [Google Scholar]
  • 16.Yang S.C. Almost sure convergence of weighted sums of mixing sequences. J. Syst. Sci. Math. Sci. 1995;15:254–265. [Google Scholar]
  • 17.Yang S.C. Maximal moment inequality for partial sums of strong mixing sequences and application. Acta Math. Sin. Engl. Ser. 2007;23:1013–1024. doi: 10.1007/s10114-005-0841-9. [DOI] [Google Scholar]
  • 18.Liptser R.S., Shiryayev A.N. Theory of Martingales. Dordrecht: Kluwer Academic; 1989. [Google Scholar]
  • 19.Masry E.X. Nonparametric regression estimation for dependent functional data: asymptotic normality. Stoch. Process. Appl. 1989;115:155–177. doi: 10.1016/j.spa.2004.07.006. [DOI] [Google Scholar]

Articles from Journal of Inequalities and Applications are provided here courtesy of Springer

RESOURCES