Abstract
In this paper, we aim to study the asymptotic properties of internal estimator of nonparametric regression with independent and dependent data. Under some weak conditions, we present some results on asymptotic normality of the estimator. Our results extend some corresponding ones.
Keywords: Asymptotic normality, Nonparametric regression, Internal estimator, Dependent data
Introduction
In this paper, we consider the nonparametric regression model
where , , and are random variables satisfying , , . So we have
Let be a kernel function. Define , where is a sequence of positive bandwidths tending to zero as . Kernel-type estimators of the regression function are widely used in various situations because of their flexibility and efficiency in the dependent and independent data. For the independent data, Nadaraya [1] and Watson [2] gave the most popular nonparametric estimator of the unknown function named the Nadaraya–Watson estimator :
| 1.1 |
Jones et al. [3] considered various versions of kernel-type regression estimators such as the Nadaraya–Watson estimator (1.1) and the local linear estimator. They also investigated the internal estimator
| 1.2 |
for a known density . Here the factor is internal to the summation, whereas the estimator has the factor externally to the summation.
The internal estimator was first proposed by Mack and Müller [4]. Jones et al. [3] studied various kernel-type regression estimators, including the introduced internal estimator (1.2). Linton and Nielsen [5] introduced an integration method based on direct integration of initial pilot estimator (1.2). Linton and Jacho-Chávez [6] studied the other internal estimator
| 1.3 |
where and . Here is a kernel function, b is the bandwidth, and the density is unknown. Under the independent data, Linton and Jacho-Chávez [6] obtained the asymptotic normality of the internal estimator in (1.3). Shen and Xie [7] obtained the complete convergence and uniform complete convergence of internal estimator in (1.2) under the geometrical α-mixing (or strong mixing) data. Li et al. [8] weakened the conditions of Shen and Xie [7] and obtained the convergence rate and uniform convergence rate for the estimator in probability.
As far as we know, there are no results on asymptotic normality of the internal estimator . Similarly to Linton and Jacho-Chávez [6], we investigate the asymptotic normality of the internal estimator with independent data and φ-mixing data, respectively. Asymptotic normality results are presented in Sect. 3.
Denote and define the coefficients
If as , then is said to be a φ-mixing sequence.
The concept of φ-mixing is introduced by Dobrushin [9], and many properties of φ-mixing are presented in Chap. 4 of Billingsley [10]. If the coefficient of the process is geometrically decreasing, then the autoregressive moving average (ARMA) process can construct a geometric φ-mixing sequence. Györfi et al. [11, 12] gave more examples and applications to nonparametric estimation. We can also refer to Fan and Yao [13] and Bosq and Blanke [14] for the works on nonparametric regression under independent and dependent data.
Regarding notation, for , set . Throughout the paper, denote some positive constants not depending on n, which may be different in various places, denotes the largest integer not exceeding x, → means to take the limit as , and means that , means the convergence in distribution, and means that random variables X and Y have the same distribution. A sequence is said to be second-order stationary if for , .
Some assumptions
In this section, we list some assumptions.
Assumption 2.1
There exist two positive constants and such that
| 2.1 |
Assumption 2.2
Let denote the compact support of known density of . For , the function is twice differentiable, and there exists a positive constant b such that
The kernel density function is symmetric and satisfies
Assumption 2.3
We assume the data observed is an independent and identically distributed stochastic sequence with values in . The known density of is upon its compact support and such that . For , we suppose that
| 2.2 |
and
| 2.3 |
Assumption 2.3∗
We assume that the data observed is a second-order stationary stochastic sequence with values in . The sequence is also assumed to be φ-mixing with . The known density of is upon its compact support and such that . Let (2.2) and (2.3) be fulfilled. Moreover, for all , we have
| 2.4 |
where denotes the joint density of .
Remark 2.1
Assumption 2.1 is a usual condition on the kernel function, and Assumption 2.2 is used to get the convergence rate of . Assumptions 2.3 and 2.3∗ are the conditions of independent and dependent data , respectively. Similarly to Hansen [15], conditions (2.2) and (2.3) are used to control the tail behavior of the conditional expectation , and (2.4) is used to estimate the covariance .
Asymptotic normality of internal estimator with independent and dependent data
In this section, we show some results on asymptotic normality of the internal estimator of a nonparametric regression model with independent and dependent data. Theorem 3.1 is for independent data, and Theorem 3.2 is for φ-mixing data.
Theorem 3.1
Let Assumptions 2.1–2.3 hold, and let . Suppose that is positive and continuous at point . If , , and as , then
| 3.1 |
where .
Theorem 3.2
Let the conditions of Theorem 3.1 be fulfilled, where Assumption 2.3 is replaced by Assumption 2.3∗. Then (3.1) holds.
Remark 3.1
The choice of a positive bandwidth h is easy to design. For example, with , if and , then the conditions , , and are satisfied as .
Conclusion
Linton and Jacho-Chávez [6] obtained some asymptotic normality results of the internal estimator under independent data. Comparing Theorem 1 and Corollary 1 of Linton and Jacho-Chávez [6], our asymptotic normality results on the internal estimator in Theorems 3.1 and 3.2 are relatively simple. Meanwhile, we use the method of Bernstein’s big-block and small-block and the inequalities of φ-mixing random variables to investigate the asymptotic normality of the internal estimator for , and we also obtain the asymptotic normality result of (3.1). Obviously, α-mixing is weaker than φ-mixing, but some moment inequalities of α-mixing are more complicated than those of φ-mixing [16, 17]. For simplicity, we study the asymptotic normality of internal estimator under φ-mixing and obtain the asymptotic normality result of Theorem 3.2.
Some lemmas and the proofs of main results
Lemma 5.1
(Liptser and Shiryayev [18], Theorem 9 in Sect. 5)
Let be martingale differences (i.e. , , is an -measurable random variable, a.s., for all and ) with for all and . Let be a sequence of Markov times with respect to , taking values in the set . If
then
Lemma 5.2
(Billingsley [10], Lemma 1)
If ξ is measurable with respect to and η is measurable with respect to (), then
implies
Lemma 5.3
(Yang [16], Lemma 2)
Let , and let be a φ-mixing sequence with . If and for all , then
where C is a positive constant depending only on .
Lemma 5.4
(Fan and Yao [13], Proposition 2.6)
Let and be the same as in (2.57) of Fan and Yao [13]. Let be complex-valued random variables measurable with respect to the σ-algebras , respectively. Suppose for and and for . Then
Proof of Theorem 3.1
It is easy to see that
| 5.1 |
Combining Assumption 2.2 with the proof of Lemma 2 of Shen and Xie [7], we obtain that
Then, it follows from that
| 5.2 |
For , let , . Denote
| 5.3 |
To prove (3.1), we apply (5.1)–(5.3) and have to show that
| 5.4 |
where is defined by (3.1).
Combining the independent and identically distributed stochastic sequence of with Lemma 5.1, to prove (5.4), we have to show that
| 5.5 |
and, for all ,
| 5.6 |
Obviously, for any (), by (2.1) and (2.3) we have
| 5.7 |
By (5.7) with this yields
| 5.8 |
Define
In view of condition (2.3), we have
So we have . Since that is positive and continuous at a point and , we obtain by Bochner lemma [14] that
| 5.9 |
Then, it follows from (5.8) and (5.9) that, for ,
| 5.10 |
which implies (5.6). Meanwhile, for some and any , by inequality and (5.7) we get that
| 5.11 |
since . Thus, (5.6) follows from (5.11). Consequently, the proof of the theorem is completed. □
Proof of Theorem 3.2
We use the same notation as in the proof of Theorem 3.1. Under the conditions of Theorem 3.2, by (5.1), (5.2), and (5.3), to prove (3.1), we need to show that
| 5.12 |
where is defined by (3.1). By the second-order stationarity, are identically distributed. Then, for , we have by (5.8) and (5.9) that
| 5.13 |
For , in view of (2.4), we have
| 5.14 |
So it follows from (5.7) and (5.14) that
| 5.15 |
Obviously, by the stationarity we establish that
| 5.16 |
For , we can choose satisfying that and as . So, by (5.15),
| 5.17 |
By Lemma 5.2 with , the condition , and (5.9), we can show that
| 5.18 |
Therefore, by (5.13), (5.16), (5.17), and (5.18), we get that
Next, we employ Bernstein’s big-block and small-block procedure (see Fan and Yao [13] and Masry [19]). Partition the set into subsets with large block of size and small block of size and set
| 5.19 |
Define and . So we have by and that
| 5.20 |
Define , , and as follows:
| 5.21 |
| 5.22 |
| 5.23 |
In view of
| 5.24 |
we have to show that
| 5.25 |
| 5.26 |
| 5.27 |
| 5.28 |
Relation (5.25) implies that and are asymptotically negligible, (5.26) shows that the summands in are asymptotically independent, and (5.27)–(5.28) are the standard Lindeberg–Feller conditions for the asymptotic normality of under independence.
First, we prove (5.25). By (5.22) and (5.24) we have
| 5.29 |
By the stationarity and (5.10), similarly to the proof of (5.17) and (5.18), for , we have
| 5.30 |
Thus it follows from (5.19) and (5.20) that
| 5.31 |
We consider the term in (5.29). With ,
but since , for , , and , similarly to the proof of (5.18), it follows that
| 5.32 |
Hence by (5.29), (5.31), and (5.32) we have
By (5.13), (5.20), and (5.23), similarly to the proofs of (5.17) and (5.18), we can find that
Thus
| 5.33 |
Second, it is easy to see that by and . Note that is -measurable with and . Since φ-mixing random variables are strong mixing random variables and , letting , by Lemma 5.4 we have
by (5.19), (5.20), and the conditions and as .
Third, we show (5.27), where is defined in (5.21). By the stationarity and (5.30) with replacing , we have
| 5.34 |
so that
since .
Fourth, it is time to establish (5.28). Obviously, by (5.7) we obtain that
We can see that , since as . Therefore, by Lemma 5.3 with we have that
| 5.35 |
Then, for all , by (5.34) and (5.35) it is easy to see that
Similarly, for , we get that
Therefore, since and , we obtain that, for all ,
Therefore, (5.26), (5.27), and (5.28) hold for , so that
| 5.36 |
Consequently, (5.12) follows from (5.33) and (5.36). Finally, by (5.1), (5.2), and (5.12) we obtain (3.1). The proof of theorem is completed. □
Authors’ contributions
All authors contributed equally to the writing of this paper. All authors read and approved the final manuscript.
Funding
This work is supported by National Natural Science Foundation of China (11501005, 11701004, 61403115), Common Key Technology Innovation Special of Key Industries (cstc2017zdcy-zdyf0252), Artificial Intelligence Technology Innovation Significant Theme Special Project (cstc2017rgzn-zdyf0073, cstc2017rgzn-zdyf0033), Natural Science Foundation of Chongqing (cstc2018jcyjA0607), Natural Science Foundation of Anhui (1808085QA03, 1808085QF212, 1808085QA17) and Provincial Natural Science Research Project of Anhui Colleges (KJ2016A027, KJ2017A027).
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Penghua Li, Email: lipenghua88@163.com.
Xiaoqin Li, Email: lixiaoqin1983@163.com.
Liping Chen, Email: lip_chenhut@126.com.
References
- 1.Nadaraya E.A. On estimating regression. Theory Probab. Appl. 1964;9:141–142. doi: 10.1137/1109020. [DOI] [Google Scholar]
- 2.Watson G.S. Smooth regression analysis. Sankhya, Ser. A. 1964;26:359–372. [Google Scholar]
- 3.Jones M.C., Davies S.J., Park B.U. Versions of kernel-type regression estimators. J. Am. Stat. Assoc. 1994;89:825–832. doi: 10.1080/01621459.1994.10476816. [DOI] [Google Scholar]
- 4.Mack Y.P., Müller H.G. Derivative estimation in nonparametric regression with random predictor variable. Sankhya. 1989;51:59–72. [Google Scholar]
- 5.Linton O., Nielsen J. A kernel method of estimating structured nonparametric regression based on marginal integration. Biometrika. 1995;82:93–100. doi: 10.1093/biomet/82.1.93. [DOI] [Google Scholar]
- 6.Linton O., Jacho-Chávez D. On internally corrected and symmetrized kernel estimators for nonparametric regression. Test. 2010;19:166–186. doi: 10.1007/s11749-009-0145-y. [DOI] [Google Scholar]
- 7.Shen J., Xie Y. Strong consistency of the internal estimator of nonparametric regression with dependent data. Stat. Probab. Lett. 2013;83:1915–1925. doi: 10.1016/j.spl.2013.04.027. [DOI] [Google Scholar]
- 8.Li X.Q., Yang W.Z., Hu S.H. Uniform convergence of estimator for nonparametric regression with dependent data. J. Inequal. Appl. 2016;2016:142. doi: 10.1186/s13660-016-1087-z. [DOI] [Google Scholar]
- 9.Dobrushin R.L. The central limit theorem for non-stationary Markov chain. Theory Probab. Appl. 1956;1:72–88. [Google Scholar]
- 10.Billingsley P. Convergence of Probability Measures. New York: Wiley; 1968. [Google Scholar]
- 11.Györfi L., Härdle W., Sarda P., Vieu P. Nonparametric Curve Estimation from Time Series. Berlin: Springer; 1989. [Google Scholar]
- 12.Györfi L., Kohler M., Krzyżak A., Walk H. A Distribution-Free Theory of Nonparametric Regression. New York: Springer; 2002. [Google Scholar]
- 13.Fan J.Q., Yao Q.W. Nonlinear Time Series: Nonparametric and Parametric Methods. New York: Springer; 2003. [Google Scholar]
- 14.Bosq D., Blanke D. Inference and Prediction in Large Dimensions. Chichester: Wiley; 2007. [Google Scholar]
- 15.Hansen B.E. Uniform convergence rates for kernel estimation with dependent data. Econom. Theory. 2008;24:726–748. doi: 10.1017/S0266466608080304. [DOI] [Google Scholar]
- 16.Yang S.C. Almost sure convergence of weighted sums of mixing sequences. J. Syst. Sci. Math. Sci. 1995;15:254–265. [Google Scholar]
- 17.Yang S.C. Maximal moment inequality for partial sums of strong mixing sequences and application. Acta Math. Sin. Engl. Ser. 2007;23:1013–1024. doi: 10.1007/s10114-005-0841-9. [DOI] [Google Scholar]
- 18.Liptser R.S., Shiryayev A.N. Theory of Martingales. Dordrecht: Kluwer Academic; 1989. [Google Scholar]
- 19.Masry E.X. Nonparametric regression estimation for dependent functional data: asymptotic normality. Stoch. Process. Appl. 1989;115:155–177. doi: 10.1016/j.spa.2004.07.006. [DOI] [Google Scholar]
