Skip to main content
Springer logoLink to Springer
. 2018 May 22;2018(1):123. doi: 10.1186/s13660-018-1715-x

Bahadur representations of M-estimators and their applications in general linear models

Hongchang Hu 1,
PMCID: PMC5978921  PMID: 30137866

Abstract

Consider the linear regression model

yi=xiTβ+ei,i=1,2,,n,

where ei=g(,εi1,εi) are general dependence errors. The Bahadur representations of M-estimators of the parameter β are given, by which asymptotically the theory of M-estimation in linear regression models is unified. As applications, the normal distributions and the rates of strong convergence are investigated, while {εi,iZ} are m-dependent, and the martingale difference and (ε,ψ)-weakly dependent.

Keywords: Linear regression models, M-estimate, Bahadur representation, Normal distribution, Rate of strong convergence

Introduction

Consider the following linear regression model:

yi=xiTβ+ei,i=1,2,,n, 1.1

where β=(β1,,βp)TRp is an unknown parametric vector, xiT denotes the ith row of an n×p design matrix X, and {ei} are stationary dependence errors with a common distribution.

An M-estimate of β is defined as any value of β minimizing

i=1nρ(yixiTβ) 1.2

for a suitable choice of the function ρ, or any solution for β of the estimating equation

i=1nψ(yixiTβ)xi=0 1.3

for a suitable choice of ψ.

There is a body of statistical literature dealing with linear regression models with independent and identically distributed (i.i.d.) random errors, see e.g. Babu [1], Bai et al. [2], Chen [7], Chen and Zhao [8], He and Shao [24], Gervini and Yohai [23], Huber and Ronchetti [28], Xiong and Joseph [50], Salibian-Barrera et al. [44]. Recently, linear regression models with serially correlated errors have attracted increasing attention from statisticians; see, for example, Li [33], Wu [49], Maller [38], Pere [41], Hu [25, 26]. Over the last 40 years, M-estimators in linear regression models have been investigated by many authors. Let {ηi} be i.i.d. random variables. Koul [30] discussed the asymptotic behavior of a class of M-estimators in the model (1.1) with long range dependence errors ei=G(ηi). Wu [49] and Zhou and Shao [52] discussed the model (1.1) with ei=G(,ηi1,ηi) and derived strong Bahadur representations of M-estimators and a central limit theorem. Zhou and Wu [53] considered the model (1.1) with ei=j=0ajηij, and obtained some asymptotic results including consistency of robust estimates. Fan et al. [20] investigated the model (1.1) with the errors ei=f(ei1)+ηi and established the moderate deviations and strong Bahadur representations for M-estimators. Wu [47] discussed strong consistency of an M-estimator in the model (1.1) for negatively associated samples. Fan [19] considered the model (1.1) with φ-mixing errors, and the moderate deviations for the M-estimators. In addition, Berlinet et al. [4], Boente and Fraiman [5], Chen et al. [6], Cheng et al. [9], Gannaz [22], Lô and Ronchetti [37], Valdora and Yohai [45] and Yang [51] have also studied some asymptotic properties of M-estimators in nonlinear models. However, no people have investigated a unified the theory of M-estimation in linear regression models with more general errors.

In this paper, we assume that

ei=g(,εi1,εi), 1.4

where g() is a measurable function such that ei is a proper random variable, and {εi,iZ} (where Z is the set of integers) are very general random variables, including m-dependent, martingale difference, (ε,ψ)-weakly dependent, and so on.

We try to investigate the unified the theory of M-estimation in the linear regression model. In the article, we use the idea of Wu [49] to study the Bahadur representative of M-estimator, and we extend some results to general errors. The paper is organized as follows. In Sect. 2, the weak and strong linear representation of an M-estimate of the vector regression parameter β in the model (1.1) are presented. Section 3 contains some applications of our results, including the m-dependent, (ε,ψ)-weakly dependent, martingale difference. In Sect. 4, proofs of the main results are given.

Main results

In the section, we investigate the weak and strong linear representation of an M-estimate of the vector regression parameter β in the model (1.1). Without loss of generality, we assume that the true parameter β=0. We start with some notation and assumptions.

For a vector v=(v1,,vp), let |v|=(i=1pvi2)12. A random vector V is said to be in Lq,q>0, if E(|V|q)<. Let Vq=E(|V|q)1q, V=V2, Σn=i=1nxixiT=XTX and assume that Σn is positive definite for large enough n. Let xin=Σn12xi,βn=Σn12β. Then the model (1.1) can be written as

yi=xinTβn+ei,i=1,2,,n, 2.1

with i=1nxinxinT=Ip, where Ip is an identity matrix of order p. Assume that ρ has derivative ψ. For l0 and a function f, write fCl if f has derivatives up to lth order and f(l) is continuous. Define the function

ψk(t;Fi)=E(ψ(ek+t)|Fi),ψk(t;Fi)=E(ψ(ek+t)|Fi),k0, 2.2

where Fi=(,ε1,ε0,ε1,,εi1,εi),Fi=(,ε1,ε0,ε1,,εi1,εi), let εi be an i.i.d. copy of εi, and ek=g(Fk).

Throughout the paper, we use the following assumptions.

  1. ρ() is a convex function, Eψ(ei)=0,0<Eψ2(ei).

  2. φ(t)Eψ(ei+t) has a strictly positive derivative at t=0.

  3. m(t)E|ψ(ei+t)ψ(ei)|2 is continuous at t=0.

  4. rnmax1in|xin|=max1in(xiTΣn1xi)12=o(1).

  5. There exists a δ0>0 such that
    Lisup|s|,|t|δ0,st|ψi+1(s;Fi)ψi+1(t;Fi)||st|L1. 2.3
  6. Let ψi(;Fi1)Cl,l0. For some δ0>0,max1insup|δ|δ0ψi(l)(δ;Fi1)< and
    i=0sup|δ|<δ0E(ψi(l)(δ;Fi1)|F0)E(ψi(l)(δ;Fi1)|F0)<. 2.4
  7. i=0sup|δ|<δ0E(ψi(l)(δ;Fi1)|F0)E(ψi(l)(δ;Fi1)|F1)<, 2.5
    i=0sup|δ|<δ0|Eψ(l)(ei+δ)Eψ(l)(ei+δ)|<. 2.6

Remark 1

Conditions (A1)–(A5) and (A6) are imposed in the M-estimation considering the theory of linear regression models with dependent errors (Wu [49]; Zhou and Shao [52]). Condition (2.6) is similar to (7) of Wu [49]. E(ψi(l)(δ;Fi1)|F0)E(ψi(l)(δ;Fi1)|F0) measures the difference of the contribution of ε0 and its copy ε0 in predicting ψ(ei+δ). However, E(ψi(l)(δ;Fi1)|F0)E(ψi(l)(δ;Fi1)|F1) measures the contribution of ε0 in predicting ψ(ei+δ) under the given copy of ε0: ε0.

If {εi} are i.i.d., then (A6) and (A7) hold. For the other settings, (A6) and (A7) are very easily satisfied. The following proposition provides some sufficient conditions for (A6) and (A7).

Proposition 2.1

Let Fi(u|F0)=P(eiu|F0) and fi(u|F0) be the conditional distribution and density function of ei at u given F0, respectively. Let fi(u) and fi(u) be the density function of ei and ei, respectively.

  1. Let fi(|Fi)Cl,l0, ω(i)=Rfi(u|F0)fi(u|F0)ψ(u;δ0)du and ψ(u;δ0)=|ψ(u+δ0)|+|ψ(uδ0)|. If i=1ω(i)<, then (A6) holds.

  2. Let
    ω(i)=Rfi(l)(u|F0)fi(l)(u|F0)ψ(u;δ0)du
    and ω˜(i)=R|fi(u)fi(u)|ψ(l)(u;δ0)du. If i=1ω(i)< and i=1ω˜(i)<, then assumption (A7) holds.

Proof

(1) By the conditions of (1), we have

i=1sup|δ|δ0E(ψi(l)(δ;Fi1)|F0)E(ψi(l)(δ;Fi1)|F0)=i=1sup|δ|δ0Rψ(l)(u+δ)[fi(u|F0)fi(u|F0)]du+sup|δ|δ0Rψ(l)(u+δ)[f0(u|F1)f0(u|F1)]dui=1Rfi(l)(u|F0)fi(l)(u|F0)ψ(u+δ0)du=i=1ω(i)<. 2.7

Namely (A6) holds.

(2) (A7) follows from

i=1sup|δ|δ0E(ψi(l)(δ;Fi1)|F0)E(ψi(l)(δ;Fi1)|F1)=i=1sup|δ|δ0Rψ(l)(u+δ)[fi(l)(u|F0)fi(l)(u|F1)]du+sup|δ|δ0Rψ(l)(u+δ)[f0(u|F1)f0(u|F1)]dui=1Rfi(l)(u|F0)fi(l)(u|F1)ψ(u+δ0)du=i=1ω(i)<

and

i=1sup|δ|δ0|E(ψ(l)(ei+δ)E(ψ(l)(ei+δ)|=i=1sup|δ|δ0|Rψ(l)(u+δ)(fi(u)fi(u))|i=1R|fi(u)fi(u)|ψ(l)(u;δ0)du=i=1ω˜(i)<.

Hence, the proposition is proved. □

Define the M-processes

Kn(βn)=Ωn(βn)E(Ωn(βn)),K˜n(β)=Ω˜n(β)E(Ω˜n(β)),

where

Ωn(βn)=i=1nψ(eixinTβn)xin,Ω˜n(β)=i=1nψ(eixiTβ)xi.

Theorem 2.1

Let {δn,nN} be a sequence of positive numbers such that δn and δnrn0. If (A1)–(A5), and (A6) and (A7) with l=0,1,,p hold, then

sup|β|δn|Kn(βn)Kn(0)|=Op(τn(δn)logn+δni=1n|xin|4), 2.8

where

τn(δ)=i=1n|xin|2(m2(|xin|δ)+m2(|xin|δ)),δ>0.

Corollary 2.1

Assume that (A1)–(A5), and (A6) and (A7) with l=0,1,,p hold. If φ(t)=tφ(0)+O(t2) as t0, Ω(βˆn)=Op(rn), then, for |βˆn|δn,

φ(0)βˆni=1nψ(ei)xin=Op(τn(δn)logn+δn2rn). 2.9

Moreover, if, as t0, m(t)=O(|t|λ) for some λ>0, then

φ(0)βˆni=1nψ(ei)xin=Op(i=1n|xin|2+2λlogn+rn). 2.10

Remark 2

If {ei} i.i.d., then |βˆn|δn follows from (3.2) of Rao and Zhao [42]. If {εi} i.i.d., then |βˆn|δn follows from Theorem 1 of Wu [49] and Zhou and Shao [52]. If ei=f(ei1)+εi, where the function f:R×RR satisfies some condition and {εi} i.i.d., then |βˆn|δn follows from Theorem 2.2 of Fan et al. [20]. If {εi} NA, then |βˆn|δn follows from Theorem 1 of Wu [47]. Therefore the condition |βˆn|δn is not strong. In the paper, we do not discuss it.

Theorem 2.2

Assume that (A1)–(A3), (A5), and (A6) and (A7) with l=0,1,,p hold. Let λn be the minimum eigenvalue of Σn, bn=n12(logn)3/2(loglogn)1/2+υ, υ>0, n˜=2logn/log2 and q>32. If lim infnλn/n>0,i=1n|xi|2=O(n) and r˜n=max1in|xi|=O(n1/2(logn)2), then

sup|β|bn|K˜n(β)K˜n(0)|=Oa.s.(Ln˜+Bn˜),

where Ln=τ˜n(2bn)(logn)q, Bn˜=bn(i=1n|xi|4)1/2(logn)3/2(loglogn)(1+υ)/2 and

τ˜n(δ)=i=1n|xi|2(m2(|xi|δ)+m2(|xi|δ)),δ>0.

Corollary 2.2

Assume that φ(t)=tφ(0)+O(t2) and m(t)=O(t) as t0, and Ω˜n=Oa.s.(r˜n). Under the conditions of Theorem 2.2, we have:

  1. β˜n=Oa.s.(bn);

  2. φ(0)Σnβ˜ni=1nψ(ei)xi=Oa.s.(Ln˜+Bn˜+bn2i=1n|xi|3+r˜n),

where β˜n is the minimizer of (1.2).

Remark 3

From the above results, we easily obtain the corresponding conclusions of Wu [49].

From the corollary below, we only derive convergence rates of β˜n. However, it is to be regretted that we cannot give laws of the iterated logarithm n1/2(loglogn)1/2, which is still an open problem.

Corollary 2.3

Under the conditions of Corollary 2.2, we have

Σnβ˜n=Oa.s.(max{(n1/2(logn)3/2(loglogn)1/2+υ),(n1/2(logn)1/4+q(loglogn)1/4+υ/2),(i=1nψ(ei)xi)}).

Proof

Note that n˜=2logn/log2=O(n) and m(t)=O(t) as t0; we have

Ln˜=τ˜n(2bn)(logn)q=O(i=1n|xi|2|xi|bn)(logn)q=O(nn1/2(logn)2n1/2(logn)3/2(loglogn)1/2+υ)(logn)q=O(n1/2(logn)1/4+q(loglogn)1/4+υ/2),Bn˜=O(n1/2(logn)3/2(loglogn)1/2+υ(nr˜n2)1/2(logn)3/2(loglogn)(1+υ)/2)=O(n1/2(logn)(loglogn)1+3υ/2)

and

bn2i=1n|xi|3=O(n1(logn)3(loglogn)1+2υnn1/2(logn)2)=O(n1/2(logn)(loglogn)1+2υ).

By Corollary 2.2, we have

φ(0)Σnβ˜n=i=1nψ(ei)xi+Oa.s.(n1/2(logn)1/4+q(loglogn)1/4+υ/2) 2.11

and

Σnβ˜n=Oa.s.(nbn)=Oa.s.(n1/2(logn)3/2(loglogn)1/2+υ). 2.12

Thus the conclusion follows from (2.11) and (2.12). □

Applications

In the following three subsections, we shall investigate some applications of our results. In Sect. 3.1, we consider that εi is a m-dependent random variable sequence. We shall investigate that {εi} are (ε,ψ)-weakly dependent in Sect. 3.2, and martingale difference errors {εi} in Sect. 3.3.

m-dependent process

In the subsection, we shall firstly show that the m-dependent sequence satisfies conditions (A6) and (A7) and secondly obtain the asymptotic normal distribution and strong convergence rates for M-estimators of the parameter. Koul [30] discussed the asymptotic behavior of a class of M-estimators in the model (1.1) with long range dependence errors ei=g(εi), where εi i.i.d. Here we assume that εi is a m-dependent sequence, of which the definition was given by Example 2.8.1 in Lehmann [32]. For m-dependent sequences or processes, there are some results (e.g., see Hu et al. [27], Romano and Wolf [43] and Valk [46]).

Proposition 3.1

Let εi in (1.4) be a m-dependent sequence. Then (A6) and (A7) hold.

Proof

Note that εi is a m-dependent sequence, we have

i=1sup|δ|δ0E(ψi(l)(δ;Fi1)|F0)E(ψi(l)(δ;Fi1)|F0)=i=1sup|δ|δ0E(ψ(l)(ei+δ)|F0)E(ψ(l)(ei+δ)|F0)+sup|δ|δ0E(ψ(l)(ei+δ)|F1)E(ψ(l)(ei+δ)|F1)=i=1sup|δ|δ0E(ψ(l)(ei+δ))E(ψ(l)(ei+δ))=0< 3.1

and

i=1sup|δ|δ0E(ψi(l)(δ;Fi1)|F0)E(ψi(l)(δ;Fi1)|F1)=i=1sup|δ|δ0E(ψ(l)(ei+δ)|F0)E(ψ(l)(ei+δ)|F1)+sup|δ|δ0E(ψ(l)(e0+δ)|F1)E(ψ(l)(e0+δ)|F1)=i=1sup|δ|δ0Eψ(l)(e0+δ)Eψ(l)(ei+δ)=0<. 3.2

Therefore, (A6) and (A7) follow from (3.1), (3.2) and Eψ(l)(ei+δ)=Eψ(l)(ei+δ). □

Corollary 3.1

Assume that (A1)–(A5) hold. If φ(t)=tφ(0)+O(t2) and m(t)=O(|t|λ) for some λ>0 as t0, Ω(βˆn)=0 and 0<σψ2=E[ψ(ei)]2<, then

n1/2βˆn/([φ(0)]1σψ)N(0,Ip),n.

In order to prove Corollary 3.1, we give the following lemmas.

Lemma 3.1

(Lehmann [32])

Let {ξi,i1} be a stationary m-dependent sequence of random variables with Eξi=0 and 0<σ2=Var(ξi)<, and Tn=i=1nξi. Then

n1/2Tn/τN(0,1),

where τ2=limnVar(n1/2Tn)=σ2+2i=2m+1Cov(ξ1,ξi).

Using the argument of Lemma 3.1, we easily obtain the following result. Here we omit the proof.

Lemma 3.2

Let {ξi,i1} be a stationary m-dependent sequence of random variables with Eξi=0 and 0<σi2=Var(ξi)<, and Tn=i=1nξi. Then

n1/2Tn/τN(0,1),

where τ2=Var(n1/2Tn)=n1i=1nσi2+2n1i=2m+1(ni)Cov(ξ1,ξi).

Proof of Corollary 3.1

By (2.10), we have

n1/2βˆn=n1/2[φ(0)]1i=1nψ(ei)xin+Op(n1/2rnλlogn). 3.3

Since {ξi,i1} is a stationary m-dependent sequence, so is {[φ(0)]1ψ(ei)xin,i1}. Let uRp, |u|=1. Then E(uT[φ(0)]1ψ(ei)xin)=0 and

σi2=E(uT[φ(0)]1ψ(ei)xin)2=[φ(0)]2uTxinxinTuE[ψ(ei)]2.

Therefore, by rn=o(1) and 0<σψ2=E[ψ(ei)]2<, we have

τ2=n1i=1n[φ(0)]2uTxinxinTuE[ψ(ei)]2+2n1i=2m+1(ni)Cov(uT[φ(0)]1ψ(e1)x1n,uT[φ(0)]1ψ(ei)xin)=[φ(0)]2n1{i=1nE[ψ(ei)]2+2i=2m+1(ni)uTxinxinTuCov(ψ(e1),ψ(ei))}[φ(0)]2σψ2. 3.4

Thus the corollary follows from Lemma 3.2, (3.3) and (3.4). □

Corollary 3.2

Assume that (A1)–(A5) hold. If φ(t)=tφ(0)+O(t2) and m(t)=O(t) as t0, and Ω˜n(β˜n)=Oa.s.(r˜n),0<σψ2=E[ψ(ei)]2<, then

β˜n=Oa.s.(n1/2(logn)3/2(loglogn)1/2+υ).

Proof

The corollary follows from Proposition 3.1 and Corollary 2.2. □

(ε,ψ)-weakly dependent process

In the subsection, we assume that {εi} are (ε,ψ)-weakly dependent (Doukhan and Louhichi [14] and Dedecker et al. [11]) random variables. In 1999, Doukhan and Louhichi proposed a new idea of (ε,ψ)-weakly dependence which focuses on covariance rather than the total variation distance between joint distributions and the product of the corresponding marginal. It has been shown that this concept is more general than mixing and includes, under natural conditions on the process parameters, essentially all classes of processes of interest in statistics. Therefore, many researchers are interested in the (ε,ψ)-weakly dependent and related possesses, and one obtained lots of sharp results. For example, Doukhan and Louhichi [14], Dedecker and Doukhan [10], Dedecker and Prieur [12], Doukhan and Neumann [16], Doukhan and Wintenberger [17], Bardet et al. [3], Doukhan and Wintenberger [18], Doukhan et al. [13]. However, a few people (only Hwang and Shin [29], Nze et al. [40]) investigated regression models with (ε,ψ)-weakly dependent errors. Nobody has investigated a robust estimate for the regression model with (ε,ψ)-weakly dependent errors. To give the definition of the (ε,ψ)-weakly dependence, let us consider a process ξ={ξn,nZ} with values in a Banach space (E,). For h:EuR, uN, we define the Lipschitz modulus of h,

Liph=supyx|h(y)h(x)|/yx1, 3.5

where we have the l1-norm, i.e., (y1,y2,,yu)1=i=1u|yi|.

Definition 1

(Doukhan and Louhich [14])

A process ξ={ξn,nZ} with values in Rd is called a (ε,ψ)-weakly dependent process if, for some classes of functions Eu,EvR,Fu,Gv:

ε(r)=supu,vsups1s2su,t1t2tv,r=t1susupfFu,gGv|Cov(f(ξs1,ξs2,,ξsu),g(ξt1,ξt2,,ξtv))|Ψ(f,g)0

as r.

According to the definition, mixing sequences (α,ρ,β,φ-mixing), associated sequences (positively or negatively associated), Gaussian sequences, Bernoulli shifts and Markovian models or time series bootstrap processes with discrete innovations are (ε,ψ)-weakly dependent (Doukhan et al. [15]).

From now on, assume that the classes of functions contain functions bounded by 1. Distinct functions Ψ yield η,θ,κ and a λ weak dependence of the coefficients as follows (Doukhan et al. [15]):

Ψ(f,g)={uLipf+vLipgthen denote ε(r)=η(r),vLipgthen denote ε(r)=θ(r),uvLipfLipgthen denote ε(r)=κ(r),uLipf+vLipg+uvLipfLipgthen denote ε(r)=λ(r),uLipf+vLipg+uvLipfLipg+u+vthen denote ε(r)=ω(r). 3.6

In Corollary 3.3, we only consider λ and η-weakly dependence. Let {εi} be λ or η-weakly dependent, and assume that g satisfies: for each sZ, if x,yRZ satisfy xi=yi for each index is

|g(x)g(y)|bs(supis|xi|l1)|xsys|. 3.7

Lemma 3.3

(Dedecker et al. [11])

Assume that g satisfies the condition (3.7) with l0 and some sequence bs0 such that s|s|bs<. Assume that E|ε0|m< with lm<m for some m>2. Then:

  1. If the process {εi,iZ} is λ-weakly dependent with coefficients λε(r), then en is λ-weakly dependent with coefficients
    λe(k)=cinfr[k/2](irbi)[(2r+1)2λε(k2r)m1lm1+l]. 3.8
  2. If the process {εi,iZ} is η-weakly dependent with coefficients ηε(r), then en is η-weakly dependent and there exists a constant c>0 such that
    ηe(k)=cinfr[k/2](irbi)[(2r+1)1+1m1ηε(k2r)m2m1].

Lemma 3.4

(Bardet et al. [3])

Let {ξn,nZ} be a sequence of Rk-valued random variables. Assume that there exists some constant C>0 such that max1ikξipC,p1. Let h be a function from Rk to R such that h(0)=0 and for x,yRk, there exist a in [1,p] and c>0 such that

|h(x)h(y)|c|xy|(1+|x|α1+|y|α1). 3.9

Now we define the sequence {ζn,nZ} by ζn=h(ξn). Then:

(1) If the process {ξi,iZ} is λ-weakly dependent with coefficients λξ(r), then {ζn,nZ} is also with coefficients

λζ(r)=O(λξpap+a2(r)). 3.10

(2) If the process {ξi,iZ} is ζ-weakly dependent with coefficients ηξ(r), so is {ζn,nZ} with coefficients ηζ(r)=O(ηξpap1(r)).

Lemma 3.5

(Dedecker et al. [11])

Let {ξi,iZ} be a centered and stationary real-valued sequence with E|ξ0|2+ς<, ς>0, σ2=kZCov(ξ0,ξk) and Sn=i=1nξi. If λξ(r)=O(rλ) for λ>4+2/ς, then n1/2SnN(0,σ2) as n.

Corollary 3.3

Let {εi} be λ-weakly dependent with coefficients λε(r)=O(exp(rλ)) for some λ>0, and bi=O(exp(ib)) for some b>0. Assume that ψ(0)=0, and, for x,yR, there exists a constant c>0 such that

|ψ(x)ψ(y)|c|xy|. 3.11

Under the conditions of Corollary 2.1, we have

φ(0)n1/2TnN(0,Σ)as n, 3.12

where Σ=i=1nx1nCov(ψ(e1),ψ(ei))xinT.

Proof

Note that {εi} is λ-weakly dependent. By Lemma 3.3, we find that {ei} is λ-weakly dependent with coefficients

λe(r)=O(r2exp(λrb(m1l)b(m1+l+2α(m1l)))),α>0 3.13

from (3.8) and Proposition 3.1 in Chap. 3 (Dedecker et al. [11]).

Let uRp, |u|=1, and ζi=h(ei)=uψ(0)xin=0. Then h(0)=0=uψ(0)xin=0. Choose p=2,a=1, in (3.9), and by (3.11), we have

|h(x)h(y)|=|xin||ψ(x)ψ(y)|c|xy| 3.14

for x,yR and c>0. Therefore, by Lemma 3.4, {ζi,iN} is λ-weakly dependent with coefficients

λζ(r)=O(rnuvλepap+a2(r))=O(rnuvλe(r)). 3.15

By Corollary 2.1, we have

φ(0)n1/2βˆn=n1/2i=1nψ(ei)xin+op(1). 3.16

By (3.13) and (3.15), there exist b>0,a>0,l0 and m>lm for some m>2 such that

λζ(r)=O(rnuvr2exp(λrb(m1l)b(m1+l)+2α(m1l)))=O(rλ) 3.17

for enough large r and λ>4+2/ς with ς>0.

By Lemma 3.5 and (3.16)–(3.17), we have

φ(0)n1/2uTnN(0,σ2),

where σ2=i=1nuTx1nCov(ψ(e1),ψ(ei))xinTu. Using the Cramer device, we complete the proof of Corollary 3.3. □

Lemma 3.6

(Dedecker et al. [11])

Suppose that {ξi,1in} are stationary real-valued random variables with Eξi=0 and P(|ξi|M<)=1 for all i=1,2,,n. Let Ψ:N2N be one of the following functions:

Ψ(u,v)=2v,Ψ(u,v)=u+v,Ψ(u,v)=uv,Ψ(u,v)=α(u+v)+(1α)uv, 3.18

for some 0<α<1. We assume that there exist constants K,L1,L2<,μ0 and a nonincreasing sequence of real coefficients {ρ(n),n0} such that, for all u-tuples (s1,,su) and all v-tuples (t1,,tv) with 1s1sut1tvn, the following inequality is fulfilled:

|Cov(ξs1,,ξsu;ξt1,,ξtv)|K2Mu+v2Ψ(u,v)ρ(t1su), 3.19

where

s=0(s+1)kρ(s)L1L2k(k!)μ,k0. 3.20

Let Sn=i=1nξi and σn2=Var(i=1nξi). If σ2=limnσn2/n>0, then

lim supn|Sn|σ(2nloglogn)1/21. 3.21

Corollary 3.4

Let {εi} be η-weakly dependent with coefficients ηε(r)=O(exp(rη)) for some η>0, and bi=O(exp(ib)) for some b>0. Assume that ψ(0)=0 and (3.11) hold. Under the conditions of Corollary 2.2 with r˜n=O(n1/2(logn)2) replaced by 0<min1in|xij|<max1in|xij|<, and 0<σψ2=Eψ2(ei)<, we have:

  1. for 3/2<q7/4,Σnβ˜n=Oa.s.(nbn)=Oa.s.(n1/2(logn)3/2(loglogn)1/2+υ);

  2. for q7/4,Σnβ˜n=Oa.s.(n1/2(logn)1/4+q(loglogn)1/4+2/υ).

Proof

Let ξi=ψ(ei)xij,j=1,,p. Then for μn as n

P{|ψ(ei)xij|>μn}E|ψ(ei)xij|2μn2=σψ2max1in|xij2|μn20. 3.22

Therefore, there exists some 0<M< such that

P{|ψ(ei)xij|M}=1. 3.23

Similar to the proofs of (3.13) and (3.15), we easily obtain

ηζ(r)=O(r˜nuvηepap+a2(r))=O(r˜nuvηe(r)), 3.24

where

ηe(r)=O(rm1lm1exp(ηrb(m2)b(m1)+2η(m2))). 3.25

By (3.24) and (3.25), we have

|Cov(ξs1,,ξsu;ξt1,,ξtv)|(u+v)ηζ(r)(u+v)r˜nuvηe(r)(u+v)r˜nuvrm1lm1exp(ηrb(m2)b(m1)+2η(m2)). 3.26

Let Ψ(u,v)=u+v,K2=rnuvM1(u+v2) and

ρ(s)=rm1lm1exp(ηrb(m2)b(m1)+2η(m2)). 3.27

Thus (3.19) holds. Since limsln(s+1)/s=0, there exist b>0,η>0,l0 and m>lm for some m>2 and m>2 such that

exp(ηsb(m2)b(m1)+2η(m2))(s+1)(2+k),k0. 3.28

Thus

s=0(s+1)kρ(s)s=0(s+1)k+m1lm1exp(ηsb(m2)b(m1)+2η(m2))s=0(s+1)kρ(s)s=0(s+1)2+m1lm1<, 3.29
σ2=limn{n1i=1nEψ2(ei)xij2+n1i,k=1;iknxijxkjCov(ψ(ei),ψ(ek))}σ2=limn{n1i=1nEψ2(ei)xij2+n12O(xij2)i=1n1(ni)Cov(ψ(e1),ψ(ei+1))}σ2=σψ2x¯j>0. 3.30

By Lemma 3.6 and Corollary 2.3, we have

Σnβ˜n=Oa.s.((2nloglogn)1/2)+Oa.s.(n1/2(logn)1/4+q(loglogn)1/4+υ/2)=Oa.s.(n1/2(logn)1/4+q(loglogn)1/4+υ/2). 3.31

Therefore, by Corollary 2.3, (3.23) and (3.31), we complete the proof of Corollary 3.4. □

Linear martingale difference processes

In the subsection, we will investigate martingale difference errors {εi}. We shall provide some sufficient conditions for (A6) and (A7) and give the central limit theorem and strong convergence rates.

Let {εi} be a martingale difference sequence, and aj be real numbers such that ei=j=0ajεij exists. It is well known that the theory of martingales provides a natural unified method for dealing with limit theorems. Under its influence, there is great interest in the martingale difference. Liang and Jing [34] were concerned with the partial linear model under the linear com of martingale differences and obtained asymptotic normality of the least squares estimator of the parameter. Nelson [39] has given conditions for the pointwise consistency of weighted least squares estimators from multivariate regression models with martingale difference errors. Lai [31] investigated stochastic regression models with martingale difference sequence errors and obtained strong consistency and asymptotic normality of the least squares estimate of the parameter.

Let Fε be the distribution function of ε0 and let fε be its density.

Proposition 3.2

Suppose that Eε0=0,ε0L4/(2γ), κγ=Rψ2(u)ωγ(du)<,1<γ<2 and k=0pR|fε(k)(v)|2ωγ(dv)<, where ωγ(dv)=(1+|v|)γ. If j=0|aj|<, then i=0ω(i)<,i=0ω¯(i)< and i=0ω˜(i)<.

Proof

Let Zn=j=0ajεnj,Zn=Znanε0anε, and

Rn=R[fε(tUn)fε(tUna0εn)]2ωγ(dt), 3.32

where Un=Znanε0. By the Schwartz inequality, we have

ω2(n)=(Rfε(tZn)fε(tZn)(1+|t|)γψ(t;ε0)(1+|t|)γdt)2Rψ2(t;ε0)ωγ(dt)Rfε(tZn)fε(tZn)2ωγ(dt)=κγRfε(tZn)fε(tZn)2ωγ(dt)CE(Rn). 3.33

Note that

fε(tUn)fε(tUnanε0)=0anε0fε(tUnv)dv 3.34

and

R[fε(tu)]2ωγ(dt)=(1+|u|)γR[fε(v)]2(1+|u|)γ(1+|u+v|)γdv(1+|u|)γR[fε(v)]2ωγ(dv)I1(1+|u|)γ. 3.35

Let Ik=R[fε(k)(v)]2ωγ(dv). By the Schwartz inequality, we have

RnR|0anε012dv0anε0[fε(tUnv)]2dv|ωγ(dt)|anε0|0anε0I1(1+|Un+v|)γdv|anε0|2[(1+|Un|)γ+(1+|Un+anε0|)γ]C|anε0|2[(1+|Un|)γ+|anε0|γ]. 3.36

By supjEεj2< and Chatterji’s inequality (Lin and Bai [35]), we have

EUn2jn,j=1aj2Eεnj2j=0aj2. 3.37

By (3.33)–(3.37) and the Schwartz inequality, we have

E(Rn)CE{|anε0|2+|anε0|2+γ+|anε0|2|Un|γ}C{|1+|an|γ+E[|ε0|2|Un|γ]}Can2{|1+|an|γ+(E|Un|2)γ/2}Can2{|1+|an|γ+(j=0aj2)γ/2}. 3.38

Note that j=0|aj|< implies j=0aj2< and j=0|aj|1+γ/2<, and by (3.33) and (3.39), we have

i=0ω(i)n=0max(|an|,|an|1+γ/2)<. 3.39

The general case k1 similarly follows. Similar to the proof of (3.39), we easily prove the other results. □

From Propositions 2.1 and 3.2, (A6) and (A7) hold. Hence, we can obtain the following two corollaries from Corollaries 2.1 and 2.2. In order to prove the following two corollaries, we first give some lemmas.

Lemma 3.7

(Liptser and Shiryayev [36])

Let ξ=(ξk)<k< be a strictly stationary sequence on a probability space (Ω,F,P), and G be a σ-algebra of invariant sets of the sequence ξ and Fk=σ(,ξk1,ξk). For a certain p2, let E|ξ0|p< and k1γk(p)<, where γk(p)={E|E(ξk|F0)|pp1}p1p. Then

Zn=1nk=1nξkdZ(stably),

where the random variable Z has the characteristic function Eexp(12λ2σ2), and σ2=E(ξ02|G)+2k1E(ξ0ξk|G).

Corollary 3.5

Assume that (A1)–(A5) hold, φ(t)=tφ(0)+O(t2) and m(t)=O(|t|λ) for some λ>0 as t0, Ω(βˆn)=Op(rn). Under the conditions of Proposition 3.2, E|ψ(ek)|pp1<,p2 and k=1n|xkn|<, we have

n1/2βˆndZ(stably), 3.40

where the random variable Z has the characteristic function Eexp(12λ2σ2), and σ2=(φ(0))2x1nTx1nE(ψ2(e1)|G)+2(φ(0))2x1nTk2xknE(ψ(e1)ψ(ek)|G).

Proof

By Proposition 2.1, Proposition 3.2 and Corollary 2.1, we have

n1/2βnˆ=n1/2(φ(0))1i=1nψ(ei)xin+Op(n1/2(rnλlog1/2n+rn))=n1/2(φ(0))1i=1nψ(ei)xin+op(1). 3.41

By E|ψ(ek)|pp1< and k=1n|xkn|<, we have

γk(p)={E|E(ψ(ek)xkn|F0)|pp1}p1pγk(p){E[E(|ψ(ek)xkn|pp1|F0)]}p1pγk(p)={E|ψ(ek)xkn|pp1}p1p}C|xkn|, 3.42
k1γk(p)=k=1n|xkn|< 3.43

and

σ2=(φ(0))2E((ψ(e1)x1n)2|G)+2(φ(0))2k2E(ψ(e1)x1nψ(ek)xkn|G)=(φ(0))2x1n2E(ψ2(e1)|G)+2(φ(0))2x1nk2xknE(ψ(e1)ψ(ek)|G).

By Proposition 2.1, Proposition 3.2 and Corollary 2.2, we easily obtain the following result. Here we omit the proof. □

Corollary 3.6

Assume that (A1)–(A5) hold, φ(t)=tφ(0)+O(t2) and m(t)=O(t) as t0, Ω˜n(β˜n)=Oa.s.(r˜n). Under the conditions of Proposition 3.2, we have

β˜n=Oa.s(n1/2(logn)3/2(loglogn)1/2+υ),υ>0.

Proofs of the main results

For the proofs of Theorem 2.1 and Theorem 2.2, we need some lemmas as follows.

Lemma 4.1

(Freedman [21])

Let τ be a stopping time, and K a positive real number. Suppose that P{|ξi|K,iτ}=1, where {ξi} are measurable random variables and E(ξi|Fi1)=0. Then, for all positive real numbers a and b,

P{i=1nξia and Tnb, for some nτ}((bKa+b)Ka+beKa)K2exp(a22(Ka+b)).

Lemma 4.2

Let

Mn(βn)=i=1n{ψ(eixinTβn)E(ψ(eixinTβn)|Fi1)}xin. 4.1

Assume that (A5) and (A6) hold. Then

sup|βn|δn|Mn(βn)Mn(0)|=Op(τn(δn)logn+n3). 4.2

Proof

Note that p=i=1nxinTxin(max1in|xin|)2n=nrn2, and δnrn0, we have δn=o(n1/2). For any positive sequence μn, let

ϕn=2μnτn(δn)logn,tn=μnτn(δn)/logμn,un=tn2,ηi(βn)=(ψ(eixinTβn)ψ(ei))xin,Tn=max1insup|βn|δn|ηi(βn)|

and

Un=i=1nE{[ψ(ei+|xin|δn)ψ(ei|xin|δn)]2|Fi1}|xin|2.

By the monotonicity of ψ and δ0, we have

sup|βn|δ|ηi(βn)||xin|sup|βn|δ|ψ(eixinTβn)ψ(ei)||xin|max{ψ(ei|xin|δ)ψ(ei),ψ(ei+|xin|δ)ψ(ei)}|xin|{ψ(ei+|xin|δ)ψ(ei|xin|δ)}. 4.3

By (4.3), the cr-inequality and (A3), we have

E(sup|βn|δn|ηi(βn)|2)E{|xin|[ψ(ei+|xin|δ)ψ(ei|xin|δ)]}22|xin|2{E[ψ(ei+|xin|δ)ψ(ei)]2+E[ψ(ei|xin|δ)ψ(ei)]2}=2|xin|2[m2(|xin|δ)+m2(|xin|δ)].

Thus

E(Tn2)=E(max1insup|βn|δn|ηi(βn)|2)i=1nE(sup|βn|δn|ηi(βn)|2)2i=1n|xin|2[m2(|xin|δn)+m2(|xin|δn)]=2τn(δn). 4.4

By the Chebyshev inequality,

P(|Tn|tn)E(Tn2)/tn22τn(δn)/tn2=2log2μn/μn20. 4.5

Similarly,

P(|Un|tn)E(Un)/un=O((logμn/μn)2)0. 4.6

Let xin=(xi1n,,xipn)T=(xi1,,xip)T,Dx(i)=(2×1xi101,,2×1xip01)Πp,Πp={1,1}p. For dΠp,j=1,2,,p, define

Mn,j,d(βn)=i=1n[ψ(eixinTβn)E(ψ(eixinTβn)|Fi1)]xij1Dx(i)=d. 4.7

Since Mn(βn)=dΠp(Mn,1,d(βn),,Mn,p,d(βn))T, it suffices to prove that Lemma 4.2 holds with Mn(βn) replaced by (Mn,j,d(βn).

Let |βn|δn,ηi,j,d(βn)=(ψ(eixinTβn)ψ(ei))xij1Dx(i)=d and

Bn(βn)=i=1nE(ηi,j,d(βn)1|ηi,j,d(βn)|>tn|Fi1). 4.8

Note that

untnϕn=tnϕn=μnτn(δn)/logμn2μnτn(δn)logn=12lognlogμn0. 4.9

By (4.9), for large enough n, we have

P(|Bn(βn)|ϕn,Unun)=P(|i=1nE(ηi,j,d(βn)1|ηi,j,d(βn)|>tn|Fi1)|ϕn,Unun)P(tn1i=1nE(ηi,j,d(βn)1|ηi,j,d(βn)|>tn|Fi1)ϕn,Unun)P(tn1Unϕn,Unun)=P(tnϕnUnun)=0. 4.10

Let the projections Pk()=E(|Fk)E(|Fk1). Since

E{Pi(ηi,j,d(βn)1|ηi,j,d(βn)|tn)|Fi1}=E{[E(ηi,j,d(βn)1|ηi,j,d(βn)|tn|Fi)E(ηi,j,d(βn)1|ηi,j,d(βn)|tn|Fi1)]|Fi1}=E(ηi,j,d(βn)1|ηi,j,d(βn)|tn|Fi1)E(ηi,j,d(βn)1|ηi,j,d(βn)|tn|Fi1)=0. 4.11

Note that {Pi(ηi,j,d(βn)1|ηi,j,d(βn)|tn)} are bound martingale differences. By Lemma 4.1 and (4.10), for |βn|tn, we have

P{|Mn,j,d(βn)Mn,j,d(0)|2ϕn,Tntn,Unun}P{|i=1nPi(ηi,j,d(βn)1|ηi,j,d(βn)|tn)|ϕn,Tntn,Unun}+P{|i=1nPi(ηi,j,d(βn)1|ηi,j,d(βn)|>tn)|ϕn,Tntn,Unun}Cexp(ϕn24tnϕn+2un)+P(|Bn(βn)|ϕn,Unun)=O(exp(ϕn24tnϕn+2un)). 4.12

Let l=n8 and Kl={(k1/l,,kp/l):kiZ,|ki|n9}. Then #Kl=(2n9+1)p, where the symbol # denotes the number of elements of the set Kl. It is easy to show

tnϕnlogn=o(ϕn2)andunlogn=o(ϕn2). 4.13

By (4.12) and (4.13), for ς>1, we have

P{supβnKl|Mn,j,d(βn)Mn,j,d(0)|2ϕn,Tntn,Unun}#KlP{|Mn,j,d(βn)Mn,j,d(0)|2ϕn,Tntn,Unun}Cn9pexp(ϕn24tnϕn+2un)=Cn9pexp(logn4tnϕnlogn/ϕn2+2unlogn/ϕn2)=Cn9pexp(logno(1))=o(nςp). 4.14

By (4.5), (4.6) and (4.14), we have

P{supβnKl|Mn,j,d(βn)Mn,j,d(0)|2ϕn}0,n. 4.15

For a, let al,1=al=al/l and al,1=al=al/l. For a vector βn=(β1n,,βpn)T, let βnl,d=(β1nl,d1,,βpnl,dp).

By (A5), for |s|,|t|rnδn and large n, we have

|E{[ψ(eit)ψ(eis)]|Fi1}|Li1|st|.

Let Vn=i=1nLi1. By condition (A5), the Markov inequality and LiL1, we have

P(Vnn4)EVn/n4=i=1nELi1/n4Cn3. 4.16

Note that |βnβnl,d|Cl1, which implies max1in|xinT(βnβnl,d)|=o(l1). Thus

sup|βn|δn|i=1nE{[ηi(βnl,d)ηi(βn)]|Fi1}xin|sup|βn|δni=1n|E((ψ(eixinTβnl,d)ψ(ei))(ψ(eixinTβn)ψ(ei))|Fi1)xin|=sup|βn|δni=1n|E(ψ(eixinTβnl,d)ψ(eixinTβn)|Fi1)xin|sup|βn|δni=1n|xin|Li1|xinT(βnl,dβn)|Cl1Vn. 4.17

Without loss of generality, assume that j=1 in the following proof.

Let d=(1,1,1,,1). Then βnl,d=(β1nl,β2nl,β3nl,,βpnl) and βnl,d=(β1nl,β2nl,β3nl,,βpnl). Since ψ is nondecreasing,

ηi,1,d(βnl,d)ηi,1,d(βn)ηi,1,d(βnl,d).

Note that

ηi,1,d(βnl,d)E[ηi,1,d(βnl,d)|Fi1]+E[ηi,1,d(βnl,d)|Fi1]E[ηi,1,d(βn)|Fi1]ηi,1,d(βn)E[ηi,1,d(βn)|Fi1]ηi,1,d(βnl,d)E[ηi,1,d(βnl,d)|Fi1]+E[ηi,1,d(βnl,d)|Fi1]E[ηi,1,d(βn)|Fi1].

Namely

i=1n{ηi,1,d(βnl,d)E[ηi,1,d(βnl,d)|Fi1]+E[(ηi,1,d(βnl,d)ηi,1,d(βn))|Fi1]}xi11Dx(i)=di=1n{ηi,1,d(βn)E[ηi,1,d(βn)|Fi1]}xi11Dx(i)=di=1n{ηi,1,d(βnl,d)E[ηi,1,d(βnl,d)|Fi1]+E[(ηi,1,d(βnl,d)ηi,1,d(βn))|Fi1]}xi11Dx(i)=d.

Therefore

Mn,1,d(βnl,d)Mn,1,d(0)+i=1nE{[ηi,1,d(βnl,d)ηi,1,d(βn)]|Fi1}xi11Dx(i)=dMn,1,d(βn)Mn,1,d(0)Mn,1,d(βnl,d)Mn,1,d(0)+i=1nE{[ηi,1,d(βnl,d)ηi,1,d(βn)]|Fi1}xi11Dx(i)=d. 4.18

By (4.17) and (4.18), we have

Mn,1,d(βnl,d)Mn,1,d(0)Cl1VnMn,1,d(βn)Mn,1,d(0)Mn,1,d(βnl,d)Mn,1,d(0)+Cl1Vn. 4.19

Note that l1Vn=Op(n8n4)=Op(n4), (4.2) immediately follows from (4.15) and (4.19). □

Lemma 4.3

Assume that the processes Xt=g(Ft)L2. Let gn(F0)=E(g(Fn)|F0),n0. Then

gn(F0)gn(F0)g(Fn)g(Fn),P0Xngn(F0)gn(F0)+R, 4.20

where R=E[gn(F0)|F1]E[gn(F0)|F0].

Proof

Since

E{[g(Fn)g(Fn)]|(F1,ε0,ε0)}=E[g(Fn)|(F1,ε0)]E[g(Fn)|(F1,ε0)]=gn(F0)gn(F0),

we have

E|E{[g(Fn)g(Fn)]|(F1,ε0,ε0)}|2=E|gn(F0)gn(F0)|2. 4.21

By the Jensen inequality, we have

E|E{[g(Fn)g(Fn)]|(F1,ε0,ε0)}|2E{E[|g(Fn)g(Fn)|2|(F1,ε0,ε0)]}=E|g(Fn)g(Fn)|2. 4.22

By (4.21) and (4.22), we have

E|gn(F0)gn(F0)|2E|g(Fn)g(Fn)|2.

That is,

gn(F0)gn(F0)g(Fn)g(Fn). 4.23

Note that

E[gn(F0)|F1]=E[E(g(Fn)|F0)|F1]=E(gn(F0)|F1) 4.24

and

E[gn(F0)|F1]=E[gn(F0)|F0]+E[gn(F0)|F1]E[gn(F0)|F0]. 4.25

By (4.24), (4.25) and the Jensen inequality, we have

P0Xn=E(g(Fn)|F0)E(g(Fn)|F1)=E[gn(F0)|F0]E[gn(F0)|F1]=E[gn(F0)|F0]E[gn(F0)|F0]E[gn(F0)|F1]+E[gn(F0)|F0]E[gn(F0)|F0]E[gn(F0)|F0]+E[gn(F0)|F1]E[gn(F0)|F0]gn(F0)gn(F0)+R. 4.26

 □

Remark 4

If {εi} i.i.d., then R=0. In this case, the above lemma becomes Theorem 1 of Wu [48].

Lemma 4.4

Let {δn,nN} be a sequence of positive numbers such that δn and δnrn0. If (A6)–(A7) hold, then

sup|βn|δn|Nn(βn)Nn(0)|=O(i=1n|xin|4δn), 4.27

where

Nn(βn)=i=1n{ψi(xinTβn;Fi1)φ(xinTβn)}xin.

Proof

Let I={n1,,nq}{1,2,,p} be a nonempty set and 1n1<<nq, and uI=(u111I,,up1pI), with vector u=(u1,,up). Write

0βn,IqNn(uI)uIduI=0βn,m10βn,mqqNn(uI)um1umqdum1dumq,wi=xinxim1ximq.

In the following, we will prove that

|qNn(uI)uI|=|i=1n{ψi(q)(xinTuI;Fi1)φ(q)(xinTuI)}wi|=O(i=1n|xin|2+2q) 4.28

uniformly over |u|pδn.

In fact, let

Tn=i=1n{ψi(q)(xinTuI;Fi1)φ(q)(xinTuI)}wi

and

Jk=i=1nPik{ψi(q)(xinTuI;Fi1)φ(q)(xinTuI)}wi.

Then Tn=k=0Jk, and Jk are martingale differences. By the orthogonality of martingale differences and the stationarity of {ei}, and Lemma 4.3, we have

Jk2=i=1nPik{ψi(q)(xinTuI;Fi1)φ(q)(xinTuI)}wi2=i=1n|wi|2P0{ψk(q)(xknTuI;Fk1)φ(q)(xknTuI)}2. 4.29

By Lemma 4.3, ψi(;Fi1)Cl,l0 and the cr-inequality, for k0, we have

P0{ψk(q)(xknTuI;Fk1)φ(q)(xknTuI)}2E{[ψk(q)(xknTuI;Fk1)φ(q)(xknTuI)]|F0}E{[ψk(q)(xknTuI;Fk1)φ(q)(xknTuI)]|F0}2+Rk22E{ψk(q)(xknTuI;Fk1)|F0}E{ψk(q)(xknTuI;Fk1)|F0}2+2E{φ(q)(xknTuI)|F0}E{φ(q)(xknTuI)|F0}2+Rk22E{ψk(q)(xknTuI;Fk1)|F0}E{ψk(q)(xknTuI;Fk1)|F0}2+2|Eψ(q)(ekxknTuI)Eψ(q)(ekxknTuI)|2+Rk2, 4.30

where

Rk2=E{[ψk(q)(xknTuI;Fk1)φ(q)(xknTuI)]|F1}E{[ψk(q)(xknTuI;Fk1)φ(q)(xknTuI)]|F0}2.

Note that Eψ(q)(ei+δ)=dqEψ(ei+t)dtq|t=δ, we have

Rk2E[ψk(q)(xknTuI;Fk1)|F1]E[ψk(q)(xknTuI;Fk1)|F0]2+E[φ(q)(xknTuI)|F1]E[φ(q)(xknTuI)|F0]2=E[ψk(q)(xknTuI;Fk1)|F1]E[ψk(q)(xknTuI;Fk1)|F0]2+Eψk(q)(ekxknTuI)Eψk(q)(ekxknTuI)2=E[ψk(q)(xknTuI;Fk1)|F1]E[ψk(q)(xknTuI;Fk1)|F0]2. 4.31

By the conditions (A6), (A7) and (4.29)–(4.31), we have

Tn=i=1n|wi|2k=0Jk=O(i=1n|wi|2)=O(i=1n|xin|2+2q).

Let |u|pδn. By max1in|xinu|pδnrn0. Note that δn and δnrn0. By (4.28), we have

sup|βn|δn0βn,I|pNn(uI)uI|duIδnδnδnδn|pNn(uI)uI|duIδnδnδnδnpNn(uI)uIduI=O(δnqi=1n|xin|2+2q)=O(δni=1n|xin|4). 4.32

Since

Nn(βn)Nn(0)=I{1,2,,p}0βn,I|I|Nn(uI)uIduI, 4.33

the result (4.27) follows from (4.32) and (4.33). □

Lemma 4.5

Let πi,i1 be a sequence of bounded positive numbers, and let there exist a constant c01 such that max1i2dπic0min1i2dπi holds for all large n. And let ωd=2c0π2d and q>3/2. Assume that (A5) and r˜n=O(n) hold. Then as d

sup|βn|ωdmaxn<2d|M˜n(β)M˜n(0)|=Op(τ˜2d(ωd)dq+25d/2),

where M˜n(β)=i=1n{ψ(eixiTβ)E(ψ(eixiTβ)|Fi1)}xi.

Proof

Let

μn=(logn)q1,ϕ˜n=2μ2dτ˜2d(ωd)log(2d),t˜2d=μ2dτ˜2d(ωd)/logμ2d,μ˜2d=t˜2d2,η˜i(β)=(ψ(eixiTβ)ψ(ei))xi,T˜2d=max1i2dsup|βn|ωd|η˜i(β)|

and

U˜2d=i=12dE{[ψ(ei+|xi|ωd)ψ(ei|xi|ωd)]2|Fi1}|xi|2.

Since q>3/2 and 2(q1)>1,d=2(μ2d1logμ2d)2<. By the argument of Lemma 4.2 and the Borel–Cantelli lemma, we have

P(T˜2dt˜2d,i.o.)=0andP(U˜2du˜2d,i.o.)=0. 4.34

Similar to the proof of (4.12), we have

P{maxk2d|M˜k,j,d(β)Mk,j,d(0)|2ϕ˜2d,T˜2dt˜2d,U˜2du˜2d}=O(exp(ϕ˜2d24t˜2dϕ˜2d+2u˜2d)). 4.35

Let l=n8d and Kl={(k1/l,,kp/l):kiZ,|ki|n9d}. Then #Kl=(2n9d+1)p. By (4.34) and (4.35), for ς>1, we have

P{supβKl|M˜k,j,d(β)Mk,j,d(0)|2ϕ˜2d,T˜2dt˜2d,U˜2du˜2d}=O(nςdp). 4.36

Therefore,

P{supβKl|M˜k,j,d(β)Mk,j,d(0)|2ϕ˜2d,i.o.}0,n. 4.37

Since r˜n=O(n) and max1i2d|xiT(ββl,d)|=O(22dl1),Cl1V in (4.17) can be replaced by Cl122dV, and the lemma follows from P(V2d25d,i.o.)=0. □

Lemma 4.6

Let πi,i1 be a sequence of bounded positive numbers, and let there exist a constant c01 such that max1i2dπc0min1i2dπi and πn=o(n1/2(logn)2) hold for all large n. And let ωd=2c0π2d. Assume that (A6), (A7) and r˜n=O(n(logn)2) hold. Then

sup|β|πn|N˜n(β)N˜n(0)|=O(i1n|xi|4πn), 4.38

and, as d, for any υ>0,

sup|β|ωdmaxn<2d|N˜n(β)N˜n(0)|2=oa.s.(i=02d|xi|4ωd2d5(logd)1+υ), 4.39

where N˜n(β)=i=1n{ψ1(xiTβ;Fi1)φ(xiTβ)}xn.

Proof

Let Qn,j(β)=i=1nψ1(xiTβ;Fi1)xij,ijp, and

Sn(β)=Qn,j(β)Qn,j(0). 4.40

Note that

πnr˜n=o(n1/2(logn)2)O(n(logn)2)=o(1). 4.41

It is easy to see that the argument in the proof of Lemma 4.4 implies that there exists a positive constant C< such that

E{|sup|β|ωd|Sn(β)Sn(β)||2}Cq=1pωd2qi=n+1n|xi|2+2q 4.42

holds uniformly over 1n<n2d. Therefore (4.38) holds.

Let Λ=r=0dμr1, where

μr={m=12drsup|β|ωd|S2rm(β)S2r(m1)(β)|2}1/2. 4.43

For a positive integer k2d, write its dyadic expansion k=2r1++2rj, where 0rj<<r1d, and k(i)=2r1++2ri. By the Schwartz inequality, we have

sup|β|ωd|Sk(β)|2{i=1jsup|β|ωd|Sk(i)(β)Sk(i1)(β)|}2={i1jμri1/2μri1/2sup|β|ωd|Sk(i)(β)Sk(i1)(β)|}2i1jμri1i1jμrisup|β|ωd|Sk(i)(β)Sk(i1)(β)|2Λi1jμrim=12dηsup|β|ωd|S2ηm(β)S2η(m1)(β)|2Λr=0dμrm=12drsup|β|ωd|S2rm(β)S2r(m1)(β)|2. 4.44

Thus

maxn2dsup|β|ωd|Sn(β)|=n=12dsup|β|ωd|Sn(β)|n=12dsup|β|ωd|Sn(β)|n=12d{E|sup|β|ωd|Sn(β)||2}1/2r=0d{E|Λr=0dμrm=12dr|sup|β|ωd|S2rm(β)S2r(m1)(β)||2|}1/2r=0d{Λr=0dμrm=12drsup|β|ωd|S2rm(β)S2r(m1)(β)|2}1/2r=0d{Λr=0dμrμR2}1/2=r=0d{Λr=0dμr1}1/2=dΛ. 4.45

Since υ>0 and ωd2qi=12d|xi|2+2q=O(ωd2i=12d|xi|4), (4.42) implies that

d=2maxn2dsup|β|ωd|Sn(β)|2ωd2i=12d|xi|4d5(logd)1+υ=d=2O(d2(d+1)2)d5(logd)1+υ<. 4.46

By the Borel–Cantelli lemma, (4.39) follows from (4.46). □

Lemma 4.7

Under the conditions of Theorem 2.2, we have:

  1. sup|β|bn|K˜n(β)K˜n(0)|=Oa.s.(Ln˜+Bn˜);

  2. for and υ>0,K˜n(0)=Oa.s.(hn), where hn=n1/2(logn)3/2(loglogn)1/2+υ/4.

Proof

Observe that K˜n(β)=M˜n(β)+N˜n(β). Since n5/2=o(Bn˜), (1) follows from Lemma 4.5 and 4.6. □

As with the argument in (4.29), we have K˜n(0)=O(n).

Proof of Theorem 2.1

Observe that

Kn(βn)=i=1nψ(eixinTβn)xinE(i=1nψ(eixinTβn)xin)=i=1n{ψ(eixinTβn)E(ψ(eixinTβn)|Fi1)}xin+i=1n{E(ψ(eixinTβn)|Fi1)Eψ(eixinTβn)}xin=Mn(βn)+Nn(βn). 4.47

By (4.47), Lemma 4.2 and Lemma 4.4, we have

sup|βn|δn|Kn(βn)Kn(0)|sup|βn|δn|Mn(βn)Mn(0)|+sup|βn|δn|Nn(βn)Nn(0)|=Op(τn(δn)logn+n3)+O(i=1n|xin|4δn)=Op(τn(δn)logn+δni=1n|xin|4). 4.48

This completes the proof of Theorem 2.1. □

Proof of Corollary 2.1

Take an arbitrary sequence δn, which satisfies the assumption of Theorem 2.1. Note that

Kn(0)=i=1nψ(ei)xinE(i=1nψ(ei)xin)=i=1nψ(ei)xin 4.49

and

Kn(βˆn)=i=1nψ(eixinTβˆn)xinE(i=1nψ(eixinTβˆn)xin)=i=1nψ(yixinTβˆn)xini=1nφ(xinTβˆn)xin=i=1nφ(xinTβˆn)xin+OP(rn) 4.50

for |βˆn|δn. By Theorem 2.1 and (4.49), we have

Kn(βˆn)=i=1nψ(ei)xin+Op(τn(δn)logn+δni=1n|xin|4). 4.51

By (4.50) and (4.51), we have

i=1nφ(xinTβˆn)xin+Op(rn)=i=1nψ(ei)xin+Op(τn(δn)logn+δni=1n|xin|4). 4.52

By (4.52), φ(t)=tφ(0)+O(t2) as t0, and i=1nxinxinT=Ip, we have

i=1n{(xinTβˆn)φ(0)+O((xinTβˆn)2)}xini=1nψ(ei)xin=Op(τn(δn)logn+δni=1n|xin|4)Op(rn)

and

i=1nxinxinTφ(0)βˆni=1nψ(ei)xin=i=1nO((xinTβˆn)2)xin+Op(τn(δn)logn+δni=1n|xin|4)Op(rn).

Namely

φ(0)βˆni=1nψ(ei)xin=Op(τn(δn)logn)+Op(δni=1n|xin|4+i=1n(xinTβˆn)2xin+rn)=Op(τn(δn)logn)+Op(δni=1n|xin|4+|βˆn|2i=1n|xin|3+rn)=Op(τn(δn)logn)+Op(δnrn+δn2rn+rn)=Op(τn(δn)logn+δn2rn). 4.53

By m(t)=O(|t|λ)(t0) for some λ>0, we have

τn(δn)=2i=1n|xin|2(|xin|δn)2λ=2δn2λi=1n|xin|2+2λ. 4.54

Then it follows from (4.53) and (4.54) that

φ(0)βˆni=1nψ(ei)xin=Op(τn(δn)logn+δn2rn)=Op(i=1n|xin|2+2λδnλlogn+δn2rn) 4.55

for any δn, which implies

φ(0)βˆni=1nψ(ei)xin=Op(i=1n|xin|2+2λlogn+rn). 4.56

 □

Proof of Theorem 2.2

By Lemma 4.7, we have Theorem 2.2. □

Proof of Corollary 2.2

(1) By Lemma 4.7, we have

sup|βn|bn|K˜n(βn)|sup|βn|bn|K˜n(βn)K˜n(0)|+K˜n(0)=Oa.s.(Ln˜+Bn˜+hn), 4.57

where bn=n1/2(logn)3/2(loglogn)1/2+υ. Let

Θn(β)=i=1n[ρ(eixiTβ)ρ(ei)] 4.58

and

An(β)=i=1n01φ(xiTβ)xiTβdt. 4.59

Note that

ρ(ei)ρ(eixiTβ)=01ψ(eixiTβ)xiTβdt. 4.60

By (4.57)–(4.60), we have

sup|βn|bn|Θn(β)An(β)|=sup|βn|bn|i=1n01[ψ(eixiTβ)φ(xiTβ)]xiTβdt|=sup|βn|bn|01K˜n(βt)βdt|=Oa.s.((Ln˜+Bn˜+hn)bn). 4.61

It is easy to show that bn3i=1n|xi|3=O(nr˜n)bn3=o(nbn2). By φ(t)=tφ(0)+O(t2), we have

inf|βn|=bnAn(β)=inf|βn|=bn{i=1n01φ(xiTβ)xiTβdt}=inf|βn|=bn{i=1n01[φ(0)+φ(0)(xiTβ)+O((xiTβ)2)]xiTβdt}=inf|βn|=bn{i=1n[12φ(0)(xiTβ)2|0113O((xiTβ)3)|01]}=12φ(0)i=1nxiTxibn213inf|βn|=bn{bn2i=1nO(xiTxi)xiTβ}12φ(0)Snbn213bn2i=1n|xi|3bnO(1)16φ(0)nbn2lim infnλn/n. 4.62

By m(t)=O(n) as t0, we have (Ln˜+Bn˜+hn)bn=o(nbn2). Thus

inf|βn|=bnΘn(β)inf|βn|=bnAn(β)sup|βn|bn|Θn(β)An(β)|16φ(0)nbn2lim infnλn/n+Oa.s.((Ln˜+Bn˜+hn)bn)14φ(0)nbn2lim infnλn/n,a.s. 4.63

By the convexity of the function Θn(), we have

{inf|βn|bnΘn(β)14φ(0)nbn2lim infnλn/n}={inf|βn|=bnΘn(β)14φ(0)nbn2lim infnλn/n}. 4.64

Therefore the minimizer βˆn satisfies βˆn=Oa.s.(bn).

(2) Let |βˆn|bn. By a Taylor expansion, we have

i=1nφ(xiTβ)xi=i=1n[φ(0)xiTβ+O(|xiTβ|2)]xi=φ(0)Σnβ+O(bn2i=1n|xi|3). 4.65

Therefore (2) follows from Theorem 2.2 and (1). □

Acknowledgements

The author’s work was supported by the National Natural Science Foundation of China (No. 11471105, 11471223), and the Natural Science Foundation of Hubei Province (No. 2016CFB526).

Authors’ contributions

The author organized and wrote this paper. Further he examined all the steps of the proofs in this paper. The author read and approved the final manuscript.

Competing interests

The author declares to have no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Babu G.J. Strong representations for LAD estimators in linear models. Probab. Theory Relat. Fields. 1989;83:547–558. doi: 10.1007/BF01845702. [DOI] [Google Scholar]
  • 2.Bai Z.D., Rao C.R., Wu Y. M-estimation of multivariate linear regression parameters under a convex discrepancy function. Stat. Sin. 1992;2:237–254. [Google Scholar]
  • 3.Bardet J., Doukhan P., Lang G., Ragache N. Dependent Lindeberg central limit theorem and some applications. ESAIM Probab. Stat. 2008;12:154–172. doi: 10.1051/ps:2007053. [DOI] [Google Scholar]
  • 4.Berlinet A., Liese F., Vaida I. Necessary and sufficient conditions for consistency of M-estimates in regression models with general errors. J. Stat. Plan. Inference. 2000;89:243–267. doi: 10.1016/S0378-3758(99)00218-9. [DOI] [Google Scholar]
  • 5.Boente G., Fraiman R. Robust nonparametric regression estimation for dependent observations. Ann. Stat. 1989;17(3):1242–1256. doi: 10.1214/aos/1176347266. [DOI] [Google Scholar]
  • 6.Chen J., Li D.G., Zhang L.X. Bahadur representation of nonparametric M-estimators for spatial processes. Acta Math. Sin. Engl. Ser. 2008;24(11):1871–1882. doi: 10.1007/s10114-008-6589-2. [DOI] [Google Scholar]
  • 7.Chen X. Linear representation of parametric M-estimators in linear models. Sci. China Ser. A. 1993;23(12):1264–1275. [Google Scholar]
  • 8.Chen X., Zhao L. M-methods in Linear Model. Shanghai: Shanghai Scientific & Technical Publishers; 1996. [Google Scholar]
  • 9.Cheng C.L., Van Ness J.W. Generalized m-estimators for errors-in-variables regression. Ann. Stat. 1992;20(1):385–397. doi: 10.1214/aos/1176348528. [DOI] [Google Scholar]
  • 10.Dedecker J., Doukhan P. A new covariance inequality and applications. Stoch. Process. Appl. 2003;106:63–80. doi: 10.1016/S0304-4149(03)00040-1. [DOI] [Google Scholar]
  • 11.Dedecker J., Doukhan P., Lang G., Leon J.R., Louhichi S., Prieur C. Weak Dependence: With Examples and Applications. New York: Springer; 2007. [Google Scholar]
  • 12.Dedecker J., Prieur C. New dependence coefficients, examples and applications to statistics. Probab. Theory Relat. Fields. 2005;132:203–236. doi: 10.1007/s00440-004-0394-3. [DOI] [Google Scholar]
  • 13.Doukhan P., Klesov O., Lang G. Rates of convergence in some SLLN under weak dependence conditions. Acta Sci. Math. (Szeged) 2010;76:683–695. [Google Scholar]
  • 14.Doukhan P., Louhichi S. A new weak dependence condition and applications to moment inequalities. Stoch. Process. Appl. 1999;84:313–342. doi: 10.1016/S0304-4149(99)00055-1. [DOI] [Google Scholar]
  • 15.Doukhan P., Mayo N., Truquet L. Weak dependence, models and some applications. Metrika. 2009;69:199–225. doi: 10.1007/s00184-008-0216-1. [DOI] [Google Scholar]
  • 16.Doukhan P., Neumann M.H. Probability and moment inequalities for sums of weakly dependent random variables with applications. Stoch. Process. Appl. 2007;117:878–903. doi: 10.1016/j.spa.2006.10.011. [DOI] [Google Scholar]
  • 17. Doukhan, P., Wintenberger, O.: An invariance principle for weakly dependent stationary general models. Probab. Math. Stat. 27(1) (2007)
  • 18.Doukhan P., Wintenberger O. Weakly dependent chains with infinite memory. Stoch. Process. Appl. 2008;118:1997–2013. doi: 10.1016/j.spa.2007.12.004. [DOI] [Google Scholar]
  • 19.Fan J. Moderate deviations for M-estimators in linear models with ϕ-mixing errors. Acta Math. Sin. Engl. Ser. 2012;28(6):1275–1294. doi: 10.1007/s10114-011-9188-6. [DOI] [Google Scholar]
  • 20.Fan J., Yan A., Xiu N. Asymptotic properties for M-estimators in linear models with dependent random errors. J. Stat. Plan. Inference. 2014;148:49–66. doi: 10.1016/j.jspi.2013.12.005. [DOI] [Google Scholar]
  • 21.Freedman D.A. On tail probabilities for martingales. Ann. Probab. 1975;3(1):100–118. doi: 10.1214/aop/1176996452. [DOI] [Google Scholar]
  • 22.Gannaz I. Robust estimation and wavelet thresholding in partially linear models. Stat. Comput. 2007;17:293–310. doi: 10.1007/s11222-007-9019-x. [DOI] [Google Scholar]
  • 23.Gervini D., Yohai V.J. A class of robust and fully efficient regression estimators. Ann. Stat. 2002;30(2):583–616. doi: 10.1214/aos/1021379866. [DOI] [Google Scholar]
  • 24.He X., Shao Q. A general Bahadur representation of M-estimators and its application to linear regression with nonstochastic designs. Ann. Stat. 1996;24(8):2608–2630. doi: 10.1214/aos/1032181172. [DOI] [Google Scholar]
  • 25.Hu H.C. QML estimators in linear regression models with functional coefficient autoregressive processes. Math. Probl. Eng. 2010;2010:956907. doi: 10.1155/2010/956907. [DOI] [Google Scholar]
  • 26.Hu H.C. Asymptotic normality of Huber–Dutter estimators in a linear model with AR(1) processes. J. Stat. Plan. Inference. 2013;143(3):548–562. doi: 10.1016/j.jspi.2012.08.012. [DOI] [Google Scholar]
  • 27.Hu Y., Ming R., Yang W. Large deviations and moderate deviations for m-negatively associated random variables. Acta Math. Sci. 2007;27B(4):886–896. [Google Scholar]
  • 28.Huber P.J., Ronchetti E.M. Robust Statistics. 2. New Jersey: John Wiley & Sons; 2009. [Google Scholar]
  • 29.Hwang E., Shin D. Semiparametric estimation for partially linear regression models with ψ-weak dependent errors. J. Korean Stat. Soc. 2011;40:411–424. doi: 10.1016/j.jkss.2011.01.002. [DOI] [Google Scholar]
  • 30.Koul H.L. M-estimators in linear regression models with long range dependent errors. Stat. Probab. Lett. 1992;14:153–164. doi: 10.1016/0167-7152(92)90079-K. [DOI] [Google Scholar]
  • 31.Lai T.L. Asymptotic properties of nonlinear least squares estimates in stochastic regression models. Ann. Stat. 1994;22(4):1917–1930. doi: 10.1214/aos/1176325764. [DOI] [Google Scholar]
  • 32.Lehmann E.L. Elements of Large-Sample Theory. New York: Springer; 1998. [Google Scholar]
  • 33.Li I. On Koul’s minimum distance estimators in the regression models with long memory moving averages. Stoch. Process. Appl. 2003;105:257–269. doi: 10.1016/S0304-4149(02)00266-1. [DOI] [Google Scholar]
  • 34.Liang H., Jing B. Asymptotic normality in partial linear models based on dependent errors. J. Stat. Plan. Inference. 2009;139:1357–1371. doi: 10.1016/j.jspi.2008.08.005. [DOI] [Google Scholar]
  • 35.Lin Z., Bai Z. Probability Inequalities. Beijing: Science Press; 2010. [Google Scholar]
  • 36.Liptser R.S., Shiryayev A.N. Theory of Martingale. London: Kluwer Academic Publishers; 1989. [Google Scholar]
  • 37.Lô S.N., Ronchetti E. Robust and accurate inference for generalized linear models. J. Multivar. Anal. 2009;100:2126–2136. doi: 10.1016/j.jmva.2009.06.012. [DOI] [Google Scholar]
  • 38.Maller R.A. Asymptotics of regressions with stationary and nonstationary residuals. Stoch. Process. Appl. 2003;105:33–67. doi: 10.1016/S0304-4149(02)00263-6. [DOI] [Google Scholar]
  • 39.Nelson P.I. A note on strong consistency of least squares estimators in regression models with martingale difference errors. Ann. Stat. 1980;8(5):1057–1064. doi: 10.1214/aos/1176345142. [DOI] [Google Scholar]
  • 40.Nze P.A., Bühlmann P., Doukhan P. Weak dependence beyond mixing and asymptotics for nonparametric regression. Ann. Stat. 2002;30(2):397–430. doi: 10.1214/aos/1021379859. [DOI] [Google Scholar]
  • 41.Pere P. Adjusted estimates and Wald statistics for the AR(1) model with constant. J. Econom. 2000;98:335–363. doi: 10.1016/S0304-4076(00)00023-3. [DOI] [Google Scholar]
  • 42.Rao C.R., Zhao L.C. Linear representation of M-estimates in linear models. Can. J. Stat. 1992;20(4):359–368. doi: 10.2307/3315607. [DOI] [Google Scholar]
  • 43.Romano J.P., Wolf M. A more general central limit theorem for m-dependent random variables with unbounded m. Stat. Probab. Lett. 2000;47:115–124. doi: 10.1016/S0167-7152(99)00146-7. [DOI] [Google Scholar]
  • 44.Salibian-Barrera M., Aelst S.V., Yohai V.J. Robust tests for linear regression models based on τ-estimates. Comput. Stat. Data Anal. 2016;93:436–455. doi: 10.1016/j.csda.2014.09.012. [DOI] [Google Scholar]
  • 45.Valdora M., Yohai V.J. Robust estimators for generalized linear models. J. Stat. Plan. Inference. 2014;146:31–48. doi: 10.1016/j.jspi.2013.09.016. [DOI] [Google Scholar]
  • 46.Valk V.D. Hilbert space representations of m-dependent processes. Ann. Probab. 1993;21(3):1550–1570. doi: 10.1214/aop/1176989130. [DOI] [Google Scholar]
  • 47.Wu Q. Strong consistency of M estimator in linear model for negatively associated samples. J. Syst. Sci. Complex. 2006;19:592–600. doi: 10.1007/s11424-006-0592-4. [DOI] [Google Scholar]
  • 48.Wu W.B. Nonlinear system theory: another look at dependence. Proc. Natl. Acad. Sci. USA. 2005;102(40):14150–14154. doi: 10.1073/pnas.0506715102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Wu W.B. M-estimation of linear models with dependent errors. Ann. Stat. 2007;35(2):495–521. doi: 10.1214/009053606000001406. [DOI] [Google Scholar]
  • 50.Xiong S., Joseph V.R. Regression with outlier shrinkage. J. Stat. Plan. Inference. 2013;143:1988–2001. doi: 10.1016/j.jspi.2013.06.007. [DOI] [Google Scholar]
  • 51.Yang Y. Asymptotics of M-estimation in non-linear regression. Acta Math. Sin. Engl. Ser. 2004;20(4):749–760. doi: 10.1007/s10114-004-0378-3. [DOI] [Google Scholar]
  • 52.Zhou Z., Shao X. Inference for linear models with dependent errors. J. R. Stat. Soc. Ser. B. 2013;75(2):323–343. doi: 10.1111/j.1467-9868.2012.01044.x. [DOI] [Google Scholar]
  • 53.Zhou Z., Wu W.B. On linear models with long memory and heavy-tailed errors. J. Multivar. Anal. 2011;102:349–362. doi: 10.1016/j.jmva.2010.09.009. [DOI] [Google Scholar]

Articles from Journal of Inequalities and Applications are provided here courtesy of Springer

RESOURCES