Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2018 Apr 4.
Published in final edited form as: Electron J Stat. 2009 Dec 4;3:1305–1321. doi: 10.1214/09-EJS479

Semiparametric Minimax Rates

James Robins 1, Eric Tchetgen Tchetgen 2, Lingling Li 3, Aad van der Vaart 4
PMCID: PMC5884174  NIHMSID: NIHMS879384  PMID: 29629001

Abstract

We consider the minimax rate of testing (or estimation) of non-linear functionals defined on semiparametric models. Existing methods appear not capable of determining a lower bound on the minimax rate of testing (or estimation) for certain functionals of interest. In particular, if the semiparametric model is indexed by several infinite-dimensional parameters. To cover these examples we extend the approach of [1], which is based on comparing a “true distribution” to a convex mixture of perturbed distributions to a comparison of two convex mixtures. The first mixture is obtained by perturbing a first parameter of the model, and the second by perturbing in addition a second parameter. We apply the new result to two examples of semiparametric functionals:the estimation of a mean response when response data are missing at random, and the estimation of an expected conditional covariance functional.

AMS 2000 subject classifications: Primary 62G05, 62G20, 62G20, 62F25

Keywords and phrases: Nonlinear functional, nonparametric estimation, Hellinger distance

1. Introduction

Let X1, X2, …, Xn be a random sample from a density p relative to a measure ν on a sample space (𝒳, 𝒜). It is known that p belongs to a collection 𝒫 of densities, and we wish to estimate the value 𝒳(p) of a functional 𝒳:𝒫 → ℝ. In this setting the mimimax rate of estimation of 𝒳(p) relative to squared error loss can be defined as the root of

inf Tnsup p𝒫Ep|Tnχ(p)|2,

where the infimum is taken over all estimators Tn = Tn(X1, …, Xn). Determination of a minimax rate in a particular problem often consists of proving a “lower bound”, showing that the mean square error of no estimator tends to zero faster than some rate εn2, combined with the explicit construction of an estimator with mean square error εn2.

The lower bound is often proved by a testing argument, which tries to separate two subsets of the set {Pn: p ∈ 𝒫} of possible distributions of the observation (X1, …, Xn). Even though testing is a statistically easier problem than estimation under quadratic loss, the corresponding minimax rates are often of the same order. The testing argument can be formulated as follows. If Pn and Qn are in the convex hull of the sets {Pn: p ∈ 𝒫, 𝒳(p) ≤ 0} and {Pn: p ∈ 𝒫, 𝒳(p) ≥ εn} and there exist no sequence of tests of Pn versus Qn with both error probabilities tending to zero, then the minimax rate is not faster than a multiple of εn. Here existence of a sequence of tests with errors tending to zero (a perfect sequence of tests) is determined by the asymptotic separation of the sequences Pn and Qn and can be described, for instance, in terms of the Hellinger affinity

ρ(Pn,Qn)=dPndQn.

If ρ(Pn, Qn) is bounded away from zero as n → ∞, then no perfect sequence of tests exists (see e.g. Section 14.5 in [2]).

One difficulty in applying this simple argument is that the relevant (least favorable) two sequences of measures Pn and Qn need not be product measures, but can be arbitrary convex combinations of product measures. In particular, it appears that for nonlinear functionals at least one of the two sequences must be a true mixture. This complicates the computation of the affinity ρ(Pn, Qn) considerably. [1] derived an elegant nice lower bound on the affinity when Pn is a product measure and Qn a convex mixture of product measures, and used it to determine the testing rate for functionals of the type ∫ f(p) , for a given smooth function f:ℝ → ℝ, the function f(x) = x2 being the crucial example.

In this paper we are interested in structured models 𝒫 that are indexed by several subparameters and where the functional is defined in terms of the subparameters. It appears that testing a product versus a mixture is often not least favorable in this situation, but testing two mixtures is. Thus we extend the bound of [1] to the case that both Pn and Qn are mixtures. In our examples Pn is equal to a convex mixture obtained by perturbing a first parameter of the model, and Qn is obtained by perturbing in addition a second parameter. We also refine the bound in other, less essential directions.

The main general results of the paper are given in Section 2. In Section 3 we apply these results to two examples of interest.

2. Main result

For k ∈ ℕ let 𝒳=j=1k𝒳j be a measurable partition of the sample space. Given a vector λ = (λ1, …, λk) in some product measurable space Λ = Λ1 × ⋯ × Λk let Pλ and Qλ be probability measures on 𝒳 such that

  1. Pλ(𝒳j) = Qλ(𝒳j) = pj for every λ ∈ Λ, for some probability vector (p1, …, pk).

  2. The restrictions of Pλ and Qλ to 𝒳j depend on the jth coordinate λj of λ = (λ1, …, λk) only.

For pλ and qλ densities of the measures Pλ and Qλ that are jointly measurable in the parameter λ and the observation, and π a probability measure on Λ, define p = ∫ pλ(λ) and q = ∫ qλ (λ), and set

a=max jsupλχj(pλp)2pλdνpj,
b=max jsupλχj(qλpλ)2pλdνpj,
d=max jsupλχj(qp)2pλdνpj.

Theorem 2.1

If npj(1 ∨ ab) ≤ A for all j and pλ for positive constants A, B̲, B̅, then there exists a constant C that depends only on A, B̲, B̅ such that, for any product probability measure π = π1 ⊗ ⋯ ⊗ πk,

ρ(Pλndπ(λ),Qλndπ(λ))1Cn2(max jpj)(b2+ab)Cnd.

Proof

The numbers a, b and d are the maxima over j of the numbers a, b and d defined in Lemma 2.2, but with the measures Pλ and Qλ replaced there by the measures Pj,λj and Qj,λj given in (2.1). Define a number c similarly as

max jsupλ𝒳jp2pλdνpj.

Under the assumptions of the theorem c is bounded above by 2 /.

By applying Lemma 2.1 and next Lemma 2.2 we see that the left side is at least

Ej=1k(114r=2Nj(Njr)br12Nj2r=1Nj1(Nj1r)arb12Nj2cNj1d)1Ej=1k(114r=2Nj(Njr)br12Nj2r=1Nj1(Nj1r)arb12Nj2cNj1d),

because j=1k(1aj)1j=1kaj for any nonnegative numbers a1, …, ak. The expected values on the binomial variables Nj can be evaluated explicitly, using the identities, for N a binomial variable with parameters n and p,

Er=2N(Nr)br=E((1+b)N1Nb)=(1+bp)n1npb,
EN2cN1=np(cp+1p)n2(cnp+1p),
EN2r=1N1(N1r)ar=EN2((1+a)N11)=np(1+ap)n2(1+nap+npp)np(1p)n2p2.

Under the assumption that np(1∨abc) ≲ 1, the right sides of these expressions can be seen to be bounded by multiples of (npb)2, np and (np)2a, respectively. We substitute these bounds in the first display of the proof, and use the equality Σj pj = 1 to complete the proof.

Remark 2.1

If min pj ~ maxj pj ~ 1/n1+ε for some ε > 0, which arises for equiprobable partitions in k ~ n1+ε sets, then there exists a number n0 such that P(maxj Nj > n0) → 0. (Indeed, the probability is bounded by k(n maxj pj)n0+1.) Under this slightly stronger assumption the computations need only address Njn0 and hence can be simplified.

The proof of Theorem 2.1 is based on two lemmas. The first lemma factorizes the affinity into the affinities of the restrictions to the partitioning sets, which are next lower bounded using the second lemma. The reduction to the partioning sets is useful, because it reduces the n-fold products to lower order products for which the second lemma is accurate.

Define probability measures Pλj and Qλj on 𝒳j by

dPj,λj=1χj dPλpj,dQj,λj=1χj dQλpj. (2.1)

Lemma 2.1

For any product probability measure π = π1 ⊗ ⋯ ⊗ πk on Λ and every n ∈ ℕ,

ρ(Pλndπ(λ),Qλndπ(λ))=Ej=1kρj(Nj),

where (N1, …, Nk) is multinomially distributed on n trials with success probability vector (p1, …, pk) and ρj : {0, …, n} → [0, 1] is defined by ρj(0) = 1 and

ρj(m)=ρ(Pj,λjmdπj(λ),Qj,λjmdπj(λ)),m1.

Proof

Set P¯nPλndπ(λ) and consider this as the distribution of the vector (X1, …, Xn). Then, for pλ and qλ densities of Pλ and Qλ relative to some dominating measure, the left side of the lemma can be written as

ρ(Pλndπ(λ),Qλndπ(λ))=EP¯nj=1ki:Xiχjqλ(Xi)dπ(λ)j=1ki:Xiχjpλ(Xi)dπ(λ).

Because by assumption on each partitioning set 𝒳j the measures Qλ and Pλ depend on λj only, the expressions ∏i:Xi∈𝒳j qλ(Xi) and ∏i:Xi∈𝒳j pλ(Xi) depend on λ only through λj. In fact, within the quotient on the right side of the preceding display, they can be replaced by ∏i:Xi∈𝒳j qj, λj(Xi) and ∏i:Xi∈𝒳j pj, λj(Xi) for qj,λj and pj,λj densities of the measures Qj,λj and Pj,λj. Because π is a product measure, we can next use Fubini’s theorem and rewrite the resulting expression as

EP¯nj=1ki:Xiχjqj,λj(Xi)dπj(λj)j=1ki:Xiχjpj,λj(Xi)dπj(λj).

Here the two products over j can be pulled out of the square root and replaced by a single product preceding it. A product over an empty set (if there is no Xi ∈ 𝒳j) is interpreted as 1.

Define variables I1, …, In, In that indicate the partitioning sets that contain the observations: Ii = j if Xi ∈ 𝒳j for every i and j, and let Nj = (#1 ≤ in: Ii = j) be the number of Xi falling in 𝒳j.

The measure n arises as the distribution of (X1, …, Xn) if this vector is generated in two steps. First λ is chosen from π and next given this λ the variables X1, …, Xn are generated independently from Pλ. Then given λ the vector (N1, …, Nk) is multinomially distributed on n trials and probability vector (Pλ(𝒳1), …, Pλ(𝒳k)). Because the latter vector is independent of λ and equal to (p1, …, pk) by assumption, the vector (N1, …, Nk) is stochastically independent of λ and hence also unconditionally, under n, multinomially distributed with parameters n and (p1, …, pk). Similarly, given λ the variables I1, …, In are independent and the event Ii = j has probability Pλ(𝒳j), which is independent of λ by assumption. It follows that the random elements (I1, …, In) and λ are stochastically independent under n.

The conditional distribution of X1, …, Xn given λ and I1, …, In can be described as: for each partitioning set 𝒳j generate Nj variables independently from Pλ restricted and renormalized to 𝒳j, i.e. from the measure Pj,λj; do so independently across the partitioning sets; and attach correct labels {1, …, n} consistent with I1, …, In to the n realizations obtained. The conditional distribution under n of X1, …, Xn given In is the mixture of this distribution relative to the conditional distribution of λ given (I1, …, In), which was seen to be the unconditional distribution, π. Thus we obtain a sample from the conditional distribution under n of (X1, …, Xn) given (I1, …, In) by generating for each partitioning set 𝒳j a set of Nj variables from the measure Pj,λjNjdπj(λj), independently across the partitioning sets, and next attaching labels consistent with I1, …, In.

Now rewrite the right side of the last display by conditioning on I1, …, In as

EP¯nEP¯n[j=1ki:Ii=jqj,λj(Xi)dπj(λj)i:Ii=jpj,λj(Xi)dπj(λj)|I1,,In].

The product over j can be pulled out of the conditional expectation by the conditional independence across the partitioning sets. The resulting expression can be seen to be of the form as claimed in the lemma.

The second lemma does not use the partitioning structure, but is valid for mixtures of products of arbitrary measures on a measurable space. For λ in a measurable space Λ let Pλ and Qλ be probability measures on a given sample space (𝒳, 𝒜), with densities pλ and qλ relative to a given dominating measure ν, which are jointly measurable. For a given (arbitrary) density p define functions ℓλ = qλpλ and κλ = pλp, and set

a=supλΛκλ2pλdν,
b=supλΛλ2pλdν,
c=supλλp2pλdν,
d=supλΛ(μdπ(μ))2pλdν.

Lemma 2.2

For any probability measure π on Λ and every n ∈ ℕ,

ρ(Pλndπ(λ),Qλndπ(λ))114r=2n(nr)br12n2r=1n1(n1r)arb12n2cn1d.

Proof

Consider the measure P¯n=Pλndπ(λ), which has density p¯n(xn)=i=1npλ(xi)dπ(λ) relative to νn, as the distribution of (X1, …, Xn). Using the inequality E1+Y1EY2/8, valid for any random variable Y with 1 + Y ≥ 0 and EY = 0 (see for example [1], we see that

ρ(Pλndπ(λ),Qλndπ(λ))=EP¯n1+[i=1nqλ(Xi)i=1npλ(Xi)]dπ(λ)p¯n(X1,,Xn)118EP¯n[i=1nqλ(Xi)i=1npλ(Xi)]dπ(λ)2p¯n(X1,,Xn)2. (2.2)

It suffices to upper bound the expected value on the right side. To this end we expand the difference i=1nqλ(Xi)i=1npλ(Xi) as Σ|I|≥1iIc pλ(Xi) ∏iIλ(Xi), where the sum ranges over all nonempty subsets I ⊂ {1, …, n}. We split this sum in two parts, consisting of the terms indexed by subsets of size 1 and the subsets that contain at least 2 elements, and separate the square of the sum of these two parts by the inequality (A + B)2 ≤ 2A2 + 2B2.

If n = 1, then there are no subsets with at least two elements and the second part is empty. Otherwise the sum over subsets with at least two elements contributes two times

|I|2iIcpλ(xi)iIλ(xi)dπ(λ)2ipλ(xi)dπ(λ)dνn(xn)(|I|2iIcpλ(xi)iIλpλ(xi))2dπ(λ)dνn(xn)=|I|2iIcpλ(xi)iIλ2pλ(xi)dπ(λ)dνn(xn).

To derive the first inequality we use the inequality (EU)2/EV ≤ E(U2/V), valid for any random variables U and V ≥ 0, which can be derived from Cauchy-Schwarz' or Jensen's inequality. The last step follows by writing the square of the sum as a double sum and noting that all off-diagonal terms vanish, as they contain at least one term ℓλ(xi) and ∫ ℓλ = 0. The order of integration in the right side can be exchanged, and next the integral relative to νn can be factorized, where the integrals ∫ pλ are equal to 1. This yields the contribution 2 Σ|I|≥2 b|I| to the bound on the expectation in (2.2).

The sum over sets with exactly one element contributes two times

j=1nijpλ(xi)λ(xj)dπ(λ)2ipλ(xi)dπ(λ)dνn(xn). (2.3)

Here we expand

ijpλ(xi)ijp(xi)=ijpλ(xi)ij(pλκλ)(xi)=|I|1,jIiIcpλ(xi)iI(κλ)(xi),

where the sum is over all nonempty subsets I ⊂ {1, …, n} that do not contain j. Replacement of ∏ij pλ(xi) by ∏ij p(xi) changes (2.3) into

j=1nijp(xi)λ(xj)dπ(λ)2ipλ(xi)dπ(λ)dνn(xn)nj=1nijp2(xi)λ(xj)dπ(λ)2ipλ(xi)dπ(λ)dνn(xn)nj=1nijp2pμ(xi)λdπ(λ)2pμ(xj)dπ(μ)dνn(xn).

In the last step we use that 1/EV ≤ E(1/V) for any positive random variable V. The integral with respect to νn in the right side can be factorized, and the expression bounded by n2cn−1d. Four this this must be added to the bound on the expectation in (2.2).

Finally the remainder after substituting ∏ij p(xi) for ∏ij pλ(xi) in (2.3) contributes

j=1n|I|1,jIiIcpλ(xi)iI(κλ)(xi)λ(xj)dπ(λ)2ipλ(xi)dπ(λ)dνn(xn)(j=1n|I|1,jIiIcpλ(xi)iIκλpλ(xi)λPλ(xj))2dπ(λ)dνn(xn)nj=1n(|I|1,jIiIcpλ(xi)iIκλpλ(xi))2λ2pλ(xj)dπ(λ)dνn(xn)=nj=1n|I|1,jIiIcpλ(xi)iIκλ2pλ(xi)λ2pλ(xj)dπ(λ)dνn(xn).

We exchange the order of integration and factorize the integral with respect to νn to bound this by n2Σ|I|≥1,jI a|I|b.

3. Applications

3.1. Estimating the mean response in missing data models

Suppose that a typical observation is distributed as X = (Y A, A, Z) for Y and A taking values in the two-point set {0, 1} and conditionally independent given Z. We think of Y as a response variable, which is observed only if the indicator A takes the value 1, and are interested in estimating the mean response EY. The covariate Z is chosen such that it contains all information on the dependence between response and missingness indicator (“missing at random”). We assume that Z takes its values in 𝒵 = [0, 1]d.

The model can be parameterized by the marginal density f of Z relative to Lebesgue measure measure ν on 𝒵, and the probabilities b(z) = P(Y = 1|Z = z) and a(z)−1 = P(A = 1|Z = z). Alternatively, the model can be parameterized by the function g = f/a, which is the conditional density of Z given A = 1 up to the norming factor P(A = 1). Under this latter parametrization which we adopt henceforth, the density p of an observation X is described by the triple (a, b, g) and the functional of interest is expressed as 𝒳(p) = ∫ abg dν.

Define CMα[0,1]d as M times the unit ball of the Hölder space of α-smooth functions on [0, 1]d. For given positive constants α, β, γ, ϕ and M̲, M, we consider the models

  • 𝒫1={(a,b,g):aCMα[0,1]d,bCMβ[0,1]d,g=1/2,a,bM¯}.
  • 𝒫2={(a,b,g):aCMα[0,1]d,bCMβ[0,1]d,gCγ[0,1]d,a,bM¯}.

If (α + β)/2 ≥ d/4, then a n-rate is attainable over 𝒫2 (see [3]), and a standard “two-point” proof can show that this rate cannot be improved. Here we are interested in the case (α + β)/2 < d/4, when the rate becomes slower than 1/n. The paper [3] constructs an estimator that attains the rate n−(2α+2β)/(2α+2β+d) uniformly over 𝒫2 if

γ2γ+d>(αβd)(d2α2βd+2α+2β)γ(α,β). (3.1)

We shall show that this result is optimal by showing that the minimax rate over the smaller model 𝒫1 is not faster than n−(2α+2β)/(2α+2β+d).

In the case that α = β these results can be proved using the method of [1], but in general we need a construction as in Section 2 with Pλ based on a perturbation of the smoothest parameter of the pair (a, b) and Qλ constructed by perturbing in addition the coarsest of the two parameters.

Theorem 3.1

If (α + β)/2 < d/4 the minimax rate over 𝒫1 for estimating ∫ abg dν is at least n−2α−2β/(2α+2β+d).

Proof

Let H: ℝd → ℝ be a C function supported on the cube [0, 1/2]d with ∫ H dν = 0 and ∫ H2 = 1. Let k be the integer closest to n2d/(2α+2β+d) and let 𝒵1, …, 𝒵k be translates of the cube k−1/d[0, 1/2]d that are disjoint and contained in [0, 1]d. For z1, …, zk the bottom left corner of these cubes and λ = (λ1, …, λk) ∈ Λ = {−1, 1}k, let

aλ(z)=2+kα/di=1kλiH((zzi)k1/d),
bλ(z)=1/2+kβ/di=1kλiH((zzi)k1/d).

These functions can be seen to be contained in Cα[0, 1]d and Cβ[0, 1]d with norms that are uniformly bounded in k. We choose a uniform prior π on λ, so that λ1, …, λk are i.i.d. Rademacher variables.

We partition the sample space {0, 1}×{0, 1}×𝒵 into the sets {0, 1}×{0, 1}×𝒵j and the remaining set.

We parameterize the model by the triple (a, b, g). The likelihood can then be written as

(a1)1A(Z)(bY(Z)(1b)1Y(Z))A.

Because ∫ H dν = 0 the values of the functional ∫ abg dν at the parameter values (aλ, 1/1, 1/2) and (2, bλ, 1/2) are both equal to 1/2, whereas the value at (aλ, bλ, 1/2) is equal to

aλbλdν2=12+(1k)α/d+β/d(i=1kH((zzi)k1/d))2dν2=12+12(1k)α/d+β/d.

Thus the minimax rate is not faster than (1/k)α/d+β/d for k = kn such that the convex mixtures of the products of the perturbations do not separate completely as n → ∞. We choose the mixtures differently in the cases αβ and αβ.

αβ. We define pλ by the parameter (aλ, 1/2, 1/2) and qλ by the parameter (aλ, bλ, 1/2). Because ∫ aλ(λ) = 2 and ∫ bλ(λ) = 1/2, we have

p(X)pλ(X)dπ(λ)=(bY(Z)(1b)1Y(Z))A,
(pλp)(X)=(1A)(aλ2)(Z),
(qλpλ)(X)=A(bλ1/2)Y(1/2bλ)1Y,
(qp)(X)(qλpλ)(X)dπ(λ)=0.

Therefore, it follows that the number d in Theorem 2.1 vanishes, while the numbers a and b are of the orders k−2α/d and k−2β/d times

maxj𝒵j(i=1kλiH((zzi)k1/d))2dν1/k1,

respectively. Theorem 2.1 shows that

ρ(Pλndπ(λ),Qλndπ(λ))1Cn21k(k4β/d+k2α/dk2β/d).

For k ~ n2d/(2α+2β+d) the right side is bounded away from 0. Substitution of this number in the magnitude of separation (1/k)α/d+β/d leads to the rate as claimed in the theorem.

αβ. We define pλ by the parameter (2, bλ, 1/2) and qλ by the parameter (aλ, bλ, 1/2). The computations are very similar to the ones in the case αβ.

3.2. Estimating an expected conditional covariance

Suppose that we observe n independent and identically distributed copies of X = (Y, A, Z), where as in the previous section, Y and A are dichotomous, and Z takes its values in 𝒵 = [0, 1]d with joint density given by f. Let b(z) = P(Y = 1|Z = z) and a(z) = P(A = 1|Z = z). We note that

b(Z)=P(Y=1|A=1,Z)a(Z)+{1a(Z)}P(Y=1|A=0,Z)={P(Y=1|A=1,Z)P(Y=1|A=0,Z)}a(Z)+P(Y=1|A=0,Z)={P(Y=1|A=0,Z)P(Y=1|A=1,Z)}{1a(Z)}+P(Y=1|A=1,Z)

so that by combining the last two equations above, we can write

P(Y=1|A,Z)=Δ(Z){Aa(Z)}+b(Z)

where Δ (Z) = P (Y = 1|A = 1, Z) − P (Y = 1|A = 0, Z). This allows us to parametrize the density p of an observation by (Δ, a, b, f). The functional χ (p) is given by expected conditional covariance

Ef{covΔ,p,b(Y,A|Z)}=EΔ,p,b,f(Y A)abfdν (3.2)

We consider the models

  • 1={(Δ,a,b,f):Δ is unrestricted,aCMα[0,1]d,bCMβ[0,1]d,f=1,a,bM}
  • 2={(Δ,a,b,f):Δ is unrestricted,aCMα[0,1]d,bCMβ[0,1]d,fCMγ[0,1]d,a,bM}

We are mainly interested in the case (α + β) /2 < d/4 when the rate of estimation of χ (p) becomes slower than 1/n. The paper [3] constructs and estimator that attains the rate n−(2α+2β)/(2α+2β+d) uniformely over ℬ2 if equation 3.1 of the previous section holds. We will show that this rate is optimal by showing that the minimax rate over the smaller model ℬ1 is not faster than n−(2α+2β)/(2α+2β+d).

The first term of the difference on the right side of equation (3.2) can be estimated by the sample average n1i=1nYiAi at rate n−1/2. It follows that χ (p) can be estimated at the maximum of n−1/2 and the rate of estimation of ∫ ab f dν. In other words, to establish that the minimax rate for estimating χ (p) over ℬ1 is n−(2α+2β)/(2α+2β+d), we shall show that the minimax rate for estimating ∫ ab f dν over ℬ1 is n−(2α+2β)/(2α+2β+d).

Theorem 3.2

If (α + β) /2 < d/4 the minimax rate over ℬ1 for estimating ∫ ab f dν is at least n−2(α+β)/(2α+2β+d).

Proof

Under the parametrization (Δ, a, b, f), the density of an observation X is given by

([Δ(Z){Aa(Z)}+b(Z)]a(Z))Y A×([1Δ(Z){Aa(Z)}b(Z)]a(Z))(1Y)A×([Δ(Z){Aa(Z)}+b(Z)]{1a(Z)})Y(1A)×[{1Δ(Z){Aa(Z)}b(Z)}{1a(Z)}](1Y)(1A)×f(Z)

Suppose α < β and set

aλ(z)=1/2+δaλ(z)=1/2+kα/di=1kλiH((zzi)k1/d)
bλ(z)=1/2+δbλ(z)=1/2+kβ/di=1kλiH((zzi)k1/d)
Δλ(Z)=δbλ(Z)1/2δaλ(Z)

then at the parameters values (0, aλ, 1/2, 1), ∫ ab f dυ = 1/4 with a corresponding likelihood pλ = {aλ (Z)}A×[{1 − aλ (Z)}](1−A), whereas at parameter values (Δλ, aλ, bλ, 1), ∫ ab f dυ = 1/4 + n−2(α+β)/(d+2(α+β)) and the likelihood is given by

qλ(X)={aλ(Z)/2}Y A×{aλ(Z)/2}(1Y)A×([1/2+δbλ(Z)])Y(1A)([1/2aλ(Z)δbλ(Z)])(1Y)(1A)

so that

(qλpλ)(X)=(1A)×δbλ(Z)Y×{δbλ(Z)}(1Y)

And we conclude that (qp) (X) = ∫ (qλpλ) (X) (λ) = 0. Furthermore

(pλp)(X)={δaλ(Z)}A×[δaλ(Z)](1A)

so that

maxjsupλ𝒳j(qλpλ)2pλ1P(𝒳j)dυ=k2β/dmaxjsupλ𝒳j(i=1kλiH((zzi)k1/d))2pλ1P(𝒳j)dυk2β/d,
supλ𝒳j((qλpλ)dπ(λ))2pλ1P(𝒳j)dυ=0

and

maxjsupλ𝒳j(qλp)2pλ1P(𝒳j)dυ=k2α/dmaxjsupλ𝒳j(i=1kλiH((zzi)k1/d))2pλ1P(𝒳j)dυk2α/d

Therefore, it follows that the number d of Theorem 2.1 vanishes, while the numbers a and b are of order k−2α/d and k−2β/d respectively. Theorem 2.1. shows that

ρ(Pλndπ(λ),Qλndπ(λ))1Cn21k(k4β/d+k2β/dk2α/d)

which gives the desired result for the choice of k ~ n2d/(2α+2β+d).

Next, suppose α > β, set aλ (Z) and bλ (Z) as above, and let

Δλ(Z)=δaλ(Z)bλ(Z)(1/2δaλ(Z))a(Z)

then at the parameters values (0, 1/2, bλ, 1), ∫ ab f dυ = 1/4 with corresponding likelihood

pλ(X)=[(bλ(Z))]Y×[(1bλ(Z))](1Y)

whereas at parameter values (0, pλ, bλ, 1), ∫ ab f dυ = 1/4 + n−2(α+β)/(d+2(α+β)) with corresponding likelihood given by

qλ(X)=[bλ(Z)/2]Y×([aλ(Z)bλ(Z)/2])(1Y)A×[1/2δaλ(Z)bλ(Z)/2](1Y)(1A)

so that

(qλpλ)(X)=(1Y)×δa(Z)A×[δaλ(Z)](1A)

and we conclude that (qp) (X) = ∫ (qλpλ) (X) (λ) = 0. Furthermore

(pλp)(X)=δbλ(Z)Y×[δbλ(Z)](1Y)

so that

maxjsupλ𝒳j(qλpλ)2pλ1P(𝒳j)dυk2α/d,
supλ𝒳j((qλpλ)dπ(λ))2pλ1P(𝒳j)dυ=0

and

maxjsupλ𝒳j(pλp)2pλ1P(𝒳j)dυk2β/d

which yields the desired result by arguments similar to the previous case.

Contributor Information

James Robins, Department of Biostatistics and Epidemiology, School of Public Health, Harvard University.

Eric Tchetgen Tchetgen, Department of Biostatistics and Epidemiology, School of Public Health, Harvard University.

Lingling Li, Department of Population Medicine, Harvard Medical School and Harvard Pilgrim Health Care, Boston, MA, 02215.

Aad van der Vaart, Department of Mathematics, Vrije Universiteit, De Boelelaan 1081a, 1081 HV Amsterdam, The Netherlands.

References

  • 1.Birge L, Massart P. Estimation of integral functionals of a density. Annals of statistics. 1995;23:11–29. [Google Scholar]
  • 2.van der Vaart A. Asymptotic statistics. Cambridge University Press; 1998. [Google Scholar]
  • 3.Robins J, Li L, Tchetgen E, van der Vaart A. Higher order inuence functions and minimax estimation of nonlinear functionals. IMS Lecture Notes-Monograph Series Probability and Statistics Models:Essays in Honor of David A. Freedman. 2008;2:335–421. [Google Scholar]

RESOURCES