Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2011 May 13.
Published in final edited form as: Proc Am Math Soc. 2010;138(12):4331–4344. doi: 10.1090/S0002-9939-2010-10448-3

How many Laplace transforms of probability measures are there?

Fuchang Gao *, Wenbo V Li , Jon A Wellner
PMCID: PMC3093923  NIHMSID: NIHMS270996  PMID: 21572594

Abstract

A bracketing metric entropy bound for the class of Laplace transforms of probability measures on [0, ∞) is obtained through its connection with the small deviation probability of a smooth Gaussian process. Our results for the particular smooth Gaussian process seem to be of independent interest.

Keywords: Laplace Transform, bracketing metric entropy, completely monotone functions, smooth Gaussian process, small deviation probability

1 Introduction

Let μ be a finite measure on [0, ∞). The Laplace transform of μ is a function on (0, ∞) defined by

f(t)=0etyμ(dy). (1)

It is easy to check that such a function has the property that (−1)n f(n) (t) ≥ 0 for all non-negative integers n and all t > 0. A function on (0, ∞) with this property is called a completely monotone function on (0, ∞). A characterization due to Bernstein (c.f. Williamson (1956)) says that f is completely monotone on (0, ∞) if and only if there is a non-negative measure μ (not necessary finite) on [0, ∞) such that (1) holds. Therefore, due to monotonicity, the class of Laplace transforms of finite measures on [0, ∞) is the same as the class of bounded completely monotone functions on (0, ∞). These functions can be extended to continuous functions on [0, ∞), and we will call them completely monotone on [0, ∞).

Completely monotonic functions have remarkable applications in various fields, such as probability and statistics, physics and potential theory. The main properties of these functions are given in Widder (1941), Chapter IV. For example, the class of completely monotonic functions is closed under sums, products and pointwise convergence. We refer to Alzer and Berg (2002) for a detailed list of references on completely monotonic functions. Closely related to the class of completely monotonic functions are the so-called k-monotone functions, where the non-negativity of (−1)n f(n) is required for all integers nk. In fact, completely monotonic functions can be viewed as the limiting case of the k-monotone functions as k → ∞. In this sense, the present work is a partial extension of Gao (2008) and Gao and Wellner (2009).

Let be the class of completely monotone functions on [0, ∞) that are bounded by 1. Then

={f:[0,)[0,)|f(t)=0etxμ(dx),μ1}.

It is well known (see e.g. Feller (1971), Theorem 1, page 439) that the sub-class of with f(0) = 1 corresponds exactly to the Laplace transforms of the class of probability measures μ on [0, ∞). For a random variable with distribution function F(t) = P(Xt), define the survival function S(t) = 1 − F(t) = P(X > t). Thus the class

S={S:[0,)[0,)|S(t)=0etxμ(dx),μ=1}

is exactly the class of survival functions of all scale mixtures of the standard exponential distribution (with survival function et), with corresponding densities

p(t)=S(t)=0xextμ(dx),t0.

It is easily seen that the class P of such densities with p(0) < ∞ is also a class of completely monotone functions corresponding to probability measures μ on [0, ∞) with finite first moment. These classes have many applications in statistics; see e.g. Jewell (1982) for a brief survey. Jewell (1982) considered nonparametric estimation of a completely monotone density and showed that the nonparametric maximum likelihood estimator (or MLE) for this class is almost surely consistent. The bracketing entropy bounds derived below can be considered as a first step toward global rates of convergence of the MLE.

In probability and statistical applications, one way to understand the complexity of a function class is by way of the metric entropy for the class under certain common distances. Recall that the metric entropy of a function class F under distance ρ is defined to be log N(ε, F, ρ) where N(ε, F, ρ) is the minimum number of open balls of radius ε needed to cover F. In statistical applications, sometimes bracketing metric entropy is needed. Recall that bracket entropy is defined as log N[ ](ε, F, ρ) where

N[](ε,F,ρ):=min{n:f_1,f¯1,,f_n,f¯ns.t.ρ(f¯k,f_k)ε,Fk=1n[f_k,f¯k]},

and

[f_k,f¯k]={gF:f_kgf¯k}.

Clearly N(ε, F, ρ) ≤ N[ ](ε, F, ρ) and they are closely related in our setting below.

In this paper, we study the metric entropy of under the Lp(ν) norm given by

fLp(ν)p=0|f(x)|pν(dx),1p,

where ν is a probability measure on [0, ∞). Our main result is the following

Theorem 1.1

  1. Let ν be a probability measure on [0, ∞). There exists a constant C depending only on p ≥ 1 such that for any 0 < ε < 1/4,
    logN[](ε,,Lp(ν))Clog(Γ/γ)|logε|2,
    for any 0 < γ < Γ < ∞ such that ν([γ, Γ]) ≥ 1 − 4pεp. In particular, if there exists a constant K > 1, such that ν([εK, εK]) ≥ 1 − 4pεp, then
    logN[](ε,,Lp(ν))CK|logε|3.
  2. If ν is Lebesgue measure on [0, 1], then
    logN[](ε,,L2(ν))logN(ε,,L2(ν))|logε|3,
    where AB means there exist universal constants C1, C2 > 0 such that C1ABC2B.

As an equivalent result for part (ii) of the above theorem, we have the following important small deviation probability estimates for an associated smooth Gaussian process. In particular, it may be of interest to find a probabilistic proof for the lower bound directly.

Theorem 1.2

Let Y (t), t > 0, be a Gaussian process with covariance EY(t)Y (s) = (1 − ets)/(t + s), then for 0 < ε < 1

log(supt>0|Y(t)|<ε)|logε|3.

The rest of the paper is organized as follows. In Section 2, we provide the upper bound estimate in the main result by explicit construction. In Section 3, we summarize various connections between entropy numbers of a set (and its convex hull) and small ball probabilities for the associated Gaussian process. Some of our observations in a general setting are stated explicitly for the first time. Finally we identify the particular Gaussian process suitable for our entropy estimates. Then in Section 4, we obtain the required upper bound small ball probability estimate (which implies the lower bound entropy estimate as discussed in section 3) by a simple determinant estimate. This method of small ball estimates is made explicit here for the first time and can be used in many more problems. The technical determinant estimates are also of independent interests.

2 Upper Bound Estimate

In this section, we provide an upper bound for N[ ](ε, , ∥ · ∥Lp(ν)), where ν is a probability measure on [0, ∞) and 1 ≤ p ≤ ∞. Before we start, we note that is the convex hull of K := {K(t, ·) : t ∈ [0, ∞)} where for each t ∈ [0, ∞), K(t, ·) is a function on [0, ∞) defined by K(t, x) = etx. There are some general results on metric entropy of convex hulls conv(T) using the metric entropy of T. (cf. Dudley (1987), Ball and Pajor (1990), van der Vaart and Wellner (1996), Carl (1997), Carl et al. (1999), Li and Linde (2000), Gao (2004), etc.) For example, Carl et al. (1999) proved that if N(ε, T, ∥ · ∥) = O(εα), α > 0, then

logN(ε,conv(T),)=O(ε2α/(2+α)),

where ∥ · ∥ is any Banach space norm. Although these results are best possible for the general case, when applied to specific problems they could be far from being sharp. This is especially the case when the metric entropy of T grows at a polynomial rate. For example, in our case, because the functions ekx, k = 1, 2, … , n, have mutual L2[0, 1]-distance at least 12n3/2, we immediately have N(ε, K, ∥ · ∥L2[0,1]) ≥ −2/3. Thus, in the case p = 2 and with ν taken to be Lebesgue measure on [0, 1], the best upper bound we can hope to obtain using the general convex hull result quoted above is

logN(ε,,L2(ν))C2ε1/2,

which is a much larger (at least in the dependence on ) than upper bound

logN(ε,,L2(ν))C|logε|3

which we will obtain later in the section.

We will obtain our upper bound estimate by an explicit construction of ε-brackets under Lp(ν) distance.

For each 0 < ε < 1/4, we choose γ > 0 and Γ = 2mγ where m is a positive integer such that ν([γ, Γ]) ≥ 1 − 4p εp. We use the notation I(at < b) to denote the indicator function of the interval [a, b). Now for each f, we first write in block form

f(t)=I(0t<γ)f(t)+I(tΓ)f(t)+i=1mI(2i1γt<2iγ)f(t).

Then for each block 2i−1γt < 2iγ, we separate the integration limits at the level 22−i | log ε|/γ and use the first N terms of Taylor’s series expansion of eu with error terms associated with ξ = ξu,N, 0 ≤ ξ ≤ 1, to rewrite

f(t)=I(0t<γ)f(t)+I(tΓ)f(t)+i=1m(pi(t)+qi(t)+ri(t))

where

pi(t)=:I(2i1γt<2iγ)n=0N(1)ntnn!022i|logε|/γxnμ(dx)qi(t)=:I(2i1γt<2iγ)022i|logε|/γ(ξtx)N+1(N+1)!μ(dx)ri(t)=:I(2i1γt<2iγ)22i|logε|/γetxμ(dx).

We choose the integer N so that

4e2|logε|1N<4e2|logε|. (2)

Then, by using the inequality k! ≥ (k/e)k and the fact that 0 < ξ < 1, we have within the block 2i−1γt < 2iγ,

|qi(t)|022i|logε|/γ(tx)N+1(N+1)!μ(dx)|4logε|N+1(N+1)!(4e|logε|N+1)N+1e(N+1)ε4e2,

where we used tx ≤ 2iγ · 22−i| log ε|/γ = 4| log ε| in the second inequality above. This implies, due to disjoint supports of qi(t),

|i=1mqi(t)|ε4e2. (3)

Next, we notice that for t ≥ 2i−1γ and x ≥ 22−iγ−1| log ε|, etxε2. Thus

|i=1mri(t)|i=1mI(2i1γt<2iγ)22iγ1|logε|ε2μ(dx)ε2. (4)

Finally, because |f| ≤ 1 and ν([0, γ)) + ν([Γ, ∞)) ≤ 4pεp, we have

I0t<γf(t)+ItΓf(t)Lp(ν)ε/4.

Together with (3) and (4), we see that the set

R=:{i=1mqi(t)+i=1mri(t)+I(t<γ)f(t)+I(tΓ)f(t):f}

has diameter in Lp(ν)-distance at most ε2 + ε4e2 + ε/4 < ε/2.

Therefore, if we denote Pi = {pi(t) : f}, then the expansion of f above implies that i=1mPi+, and consequently, we have

N[](ε,,Lp(ν))N[](ε/2,i=1mPi,Lp(ν)).

For any 1 ≤ im and any piPi, we can write

pi(t)=I(2i1γt<2iγ)n=0Ni(1)nani(2iγ1t)n, (5)

where 0 ≤ ani ≤ |4 log ε|n/n!. Now we can construct

p¯i=I(2i1γt<2iγ)n=0N(1)nbni(2iγ1t)n,p_i=I(2i1γt<2iγ)n=0N(1)ncni(2iγ1t)n,

where

bni={ε2n+22n+2aniεifnis evenε2n+22n+2aniεifnis odd,cni={ε2n+22n+2aniεifnis evenε2n+22n+2aniεifnis odd.

Clearly, i(t) ≤ pi(t) ≤ i(t), and

|p¯ip_i|I(2i1γt<2iγ)n=0N|cnibni|(2iγ1t)nI(2i1γt<2iγ)n=0Nε2n+2(2iγ1t)nε2I(2i1γt<2iγ).

Hence

i=1mp_ii=1mpii=1mp¯ii=1mp_i+ε/2.

That is, the sets

P_=:{i=1mp_i:piPi,1im}andP¯=:{i=1mp¯i:piPi,1im}

form ε/2 brackets of i=1mPi in L-norm, and thus in Lp(ν)-norm for all 1 ≤ p < ∞.

Now we count the number of different realizations of and . Note that, due to the uniform bound on ani in (5) there are no more than

2n+1ε|4logε|nn!+1

realizations for bni. So, the number of realizations of i is bounded by

n=0N(2n+1ε|4logε|nn!+1).

Because n! > (n/e)n, for all 1 ≤ nN, we have

2n+1ε|4logε|nn!+13ε(8e|logε|n)n.

Thus, the number of realizations of i is bounded by

(3ε)N+1exp(n=1N(nlog|8elogε|nlogn))(3ε)N+1exp(N(N+1)2log|8elogε|1Nxlogxdx)(3ε)N+1exp(N(N+1)2log|8elogε|N22logN+N24)exp(C|logε|2)

for some absolute constant C, where in the last inequality we used the bounds on N given in (2).

Hence the total number of realizations of is bounded by exp (Cm| log ε|2). A similar estimate holds for the total number of realizations of , and we finally obtain

logN[](ε,,Lp(ν))Cm|logε|2

for some different constant C′. This finishes the proof since m = log2(Γ/γ).

3 Entropy of Convex Hulls

A lower bound estimate of metric entropy is typically difficult, because it often involves a construction of a well-separated set of maximal cardinality. Thus we introduce some soft analytic arguments to avoid this difficulty and change the problem into a familiar one in this section. The hard estimates are given in the next section.

First note that is just the convex hull of the functions ks(·), 0 < s < ∞, where ks(t) = ets. We recall a general method to bound the entropy of convex hulls that was introduced in Gao (2004). Let T be a set in ℝn or in a Hilbert space. The convex hull of T can be expressed as

conv(T)={n=1antn:tnT,an0,n,n=1an=1};

while the absolute convex hull of T is defined by

abconv(T)={n=1antn:tnT,n,n=1|an|1}.

Clearly, by using probability measures and signed measures, we can express

conv(T)={Ttμ(dt):μis a probability measure onT};abconv(T)={Ttμ(dt):μis a signed measure onT,μTV1}.

The following is clear:

conv(T)abconv(T)conv(T)conv(T).

Therefore, for any norm ∥ · ∥,

N(ε,conv(T),)N(ε,abconv(T),)[N(ε/2,conv(T),)]2.

In particular, at the logarithmic level, the two entropy numbers are comparable, modulo constant factors on ε. The benefit of using the absolute convex hull is that it is symmetric and can be viewed as the unit ball of a Banach space, which allows us to use the following duality lemma of metric entropy: there exist constants c1, c2, K1, K2 > 0 such that for all ε > 0,

K1logN(c1ε,abconv(T),2)logN(ε,B,T)K2logN(c2ε,abconv(T),2),

where B is the unit ball of the dual norm of ∥ · ∥, and ∥ · ∥T is the norm induced by T, that is,

xT:=suptT|t,x|=suptabconv(T)|t,x|.

Strictly speaking, the duality lemma remains a conjecture in the general case. However, when the norm ∥ · ∥ is a Hilbert space norm, this has been proved, see Tomczak-Jaegermann (1987), Bourgain et al. (1989), and Artstein et al. (2004).

A striking relation discovered by Kuelbs and Li (1993) says that the entropy number log N(ε, B, ∥ · ∥T) is determined by the Gaussian measure of the set

Dε=:{xH:xTε}

under some very weak regularity assumptions. For details, see Kuelbs and Li (1993), Li and Linde (1999), and also Corollary 2.2 of Aurzada et al. (2009). Using this relation, we can now summarize the connection between the metric entropy of convex hulls and the Gaussian measure of Dε as follows:

Proposition 3.1

Let T be a precompact set in a Hilbert space. For α > 0 and β ∈ ℝ, there exists a constant C1 > 0 such that for all 0 < ε < 1,

log(Dε)C1εα|logε|β

if and only if there exists a constant C2 > 0 such that for all 0 < ε < 1,

logN(ε,conv(T),)C2ε2α2+α|logε|2β2+α;

and for β > 0 and γ ∈ ℝ, there exists a constant C3 > 0 such that for all 0 < ε < 1,

log(Dε)C3|logε|β(log|logε|)γ

if and only if there exists a constant C4 > 0 such that for all 0 < ε < 1,

logN(ε,conv(T),2)C4|logε|β(log|logε|)γ.

Furthermore, the results also hold if the directions of the inequalities are switched.

The result of this proposition can be implicitly seen in Gao (2004), where an explanation of the relation between N(ε, B, ∥ · ∥T) and the Gaussian measure of Dε is also given.

Perhaps, the most useful case of Proposition 3.1 is when T is a set of functions: K(t, ·), tT, where for each fixed tT, K(t, ·) is a function in L2(Ω), and where Ω is a bounded set in ℝd, d ≥ 1. For this special case, we have

Corollary 3.2

Let X(t) = ∫Ω K(t, x)dB(x), tT, where K(t, ·) are square-integrable functions on a bounded set Ω in ℝd, d ≥ 1, and B(x) is the d-dimensional Brownian sheet on Ω. If is the convex hull of the functions K(t, ·), tT, then

log(suptT|X(t)|<ε)εα|logε|β

for α > 0 and β ∈ ℝ if and only if

logN(ε,,)ε2α2+α|logε|2β2+α;

and for β > 0 and γ ∈ ℝ,

log(suptT|X(t)|<ε)|logε|β(log|logε|)γ

if and only if

logN(ε,,2)|logε|β(log|logε|)γ.

The authors found this corollary especially useful. For example, it was used in Blei et al. (2007) and Gao (2008) to change a problem of metric entropy into a problem of small deviation probability of a Gaussian process which is relatively easier. The proof is given in Gao (2008) for the case Ω = [0, 1], and in Blei et al. (2007) for the case [0, 1]d. For the general case, it can be proved as easily. Indeed, the only thing we need to prove is that ℙ(Dε) can be expressed as the probability of the set suptT |X(t)| < ε. We outline a proof below. Let ϕn be an orthonormal basis of L2(Ω), then

X(t)=ΩK(t,s)dB(s)=n=1ξnΩK(t,s)ϕn(s)ds

where ξn are i.i.d standard normal random variables. Thus,

(Dε)={gL2(Ω):|Ωf(s)g(s)ds|<ε,f}={gL2(Ω):|TΩK(t,s)g(s)dsμ(dt)|<ε,μTV1}={n=1anϕn(s):n=1an2<,|n=1anTΩK(t,s)ϕn(s)dsμ(dt)|<ε,μTV1}={n=1anϕn(s):n=1an2<,suptT|n=1anΩK(t,s)ϕn(s)ds|<ε}=(suptT|X(t)|<ε).

Now back to our problem to estimate log N(ε, , ∥ · ∥2) in the statement (ii) of Theorem 1.1, where ∥ · ∥2 is the L2 norm with respect to Lebesgue measure on [0, 1], we notice that is the convex hull of the functions K(t, ·), t ∈ [0, ∞), on [0, 1], with K(t, s) = ets. Clearly, for each fixed t, K(t, ·) is a square-integrable function on the bounded set [0, 1]. Now, for this K, the corresponding X(t) is a Gaussian process on [0, ∞) with covariance

EX(t)X(s)=1etst+s,s,t0. (6)

Thus, the problem becomes how to sharply bound the probability

(supt(0,1]|X(t)|<ε).

This will be done in the next section.

4 Lower Bound Estimate

Let X(t), t ≥ 0 be the centered Gaussian process defined in (6). Our goal in this section is to prove that

log(supt0|X(t)|<ε)C|logε|3,

for some constant C > 0.

Note that for any sequence of positive numbers {δi}i=1n,

(supt0|X(t)|<ε)(max1in|X(δi)|<ε)=(2π)n/2(det)1/2max1in|yi|εexp(y,1y)dy1dyn(2π)n/2(det)1/2(2ε)nεn(det)1/2. (7)

where we use the covariance matrix

=(EX(δi)X(δj))1i,jn=(1eδiδjδi+δj)1i,jn.

To find a lower bound for det(Σ), we need the following lemma:

Lemma 4.1

If 0 < bij < aij for all 1 ≤ i, jn then

det(aijbij)det(aij)k=1nmax1lnbklaklper(aij).

where per(aij) is the permanent of the matrix (aij).

Proof

For notational simplicity, we denote cij = aijbij, then

det(aijbij)det(aij)=σ(1)σc1,σ(1)c2,σ(2)cn,σ(n)σ(1)σa1,σ(1)a2,σ(2)an,σ(n)=σ(1)σk=1n[c1,σ(1)ck1,σ(k1)](ck,σ(k)ak,σ(k))[ak+1,σ(k+1)an,σ(n)]σk=1n[a1,σ(1)ak1,σ(k1)](bk,σ(k))[ak+1,σ(k+1)an,σ(n)]k=1nmax1lnbklaklσ[a1,σ(1)ak1,σ(k1)](ak,σ(k))[ak+1,σ(k+1)an,σ(n)]=k=1nmax1lnbklaklper(aij).

In order to use Lemma 4.1 to estimate det(Σ), we set

aij=1δi+δj,andbij=eδiδjaij

for a specific sequence {δi}i=1n defined by

δmp+q=4p+m(m+q),0p<m,1qm

for n = m2.

Clearly, we have

0<bkl/akle2m4m,1k,ln=m2. (8)

It remains to estimate det(aij) and per(aij), which is given in the following lemma.

Lemma 4.2

For the matrix (aij) defined above, we have per(aij) ≤ 1, and det(aij) ≥ (240e)−2m3.

Proof

It is easy to see

per(aij)n!(maxi,jaij)n(m2)!(2m4m)m21

since aij ≤ (2m4m)−1 for 1 ≤ i, jn = m2.

To estimate det(aij), we use the Cauchy’s determinant identity, see Krattenthaler (1999),

det(aij)=det(1δi+δj)=1i<jn(δjδi)21i,jn(δj+δi)=12ni=1nδi1i<jn(δjδiδj+δi)2.

To estimate the last product, we partition the set {(i; j) : 1 < i < j < n = m2} into the three sets, and estimate each part separately. For 1 ≤ i < jn = m2, write i = mp + q and j = mr + s with 1 ≤ q, sm. Denote

A={(i,j):i=mp+q,j=mp+s,0pm1,1q<sm},B={(i,j):i=mp+q,j=m(p+1)+s,0pm2,1q,sm},C={(i,j):i=mp+q,j=mr+s,0pm3,p+2rm1,1q,sm}.

Thus, A, B and C form a partition of {(i, j) : 1 ≤ i < jn = m2}.

First, for (i, j) ∈ A,

δjδiδj+δi=sq2m+s+q>sq4m.

Thus

(i,j)A(δjδiδj+δi)2p=0m11q<sm(sq4m)2=k=1m1q=1mk(k4m)2mk=1m1(k4m)2m2=((m1)!(4m)m1)2m2(8e)2m3.

Second, for (i, j) ∈ B,

δjδiδj+δi=(4m+4s)(m+q)(4m+4s)+(m+q)15.

Thus we have

(i,j)B(δjδiδj+δi)2p=0m21q,sm5252m3.

Third, for (i, j) ∈ C, we have rp ≥ 2, and

(δjδiδj+δi)=4r(m+s)4p(m+q)4r(m+s)+4p(m+q)=124p(m+q)4r(m+s)+4p(m+q)>114rp1.

Thus, since Π(1 − xk) ≥ 1 − Σxk for 0 < xk < 1,

(i,j)C(δjδiδj+δi)2p=0m3r=p+2m11q,sm(114rp1)2k=1m2(14k)2m3(1k=1m24k)2m3(2/3)2m3.

Therefore, we have

1i<jn(δjδiδj+δi)2=(i,j)A(i,j)B(i,j)C(δjδiδj+δi)2(60e)2m3.

On the other hand, it is not difficult to see that

2ni=1nδi=2m2q=1mp=0m14p+m(m+q)<2m24m2(m1)/2+m3(2m)m2=43m3/2+m2/2+m2log4m<42m3

for m > 1. Hence,

det(aij)=(2ni=1nδi)1ii<jn(δjδiδj+δi)2(240e)2m3.

Now combining the two lemmas above, and using the estimate in (8), we obtain

det()(240e)2m3m2e2m4me16m3

provided that m is large enough. Plugging into (7), we have

(supt0|X(t)|<ε)e8m3εm2.

Minimizing the right-hand side by choosing m ≈ | log ε|/12, we obtain

(supt0|X(t)|<ε)exp((432)1|logε|3).

Statement (ii) of Theorem 1.1 follows by applying Corollary 3.2. At the same time, we also finished the proof of Theorem 1.2.

Acknowledgments

We owe thanks to a referee for a number of helpful suggestions.

References

  1. Alzer H, Berg C. Some classes of completely monotonic functions. Ann Acad Sci Fenn Math. 2002;27:445–460. [Google Scholar]
  2. Artstein S, Milman V, Szarek S, Tomczak-Jaegermann N. On convexified packing and entropy duality. Geom Funct Anal. 2004;14:1134–1141. [Google Scholar]
  3. Aurzada F, Ibragimov I, Lifshits M, van Zanten JH. Small deviations of smooth stationary Gaussian processes. Theory of Probability and Its Applications. 2009;53:697–707. [Google Scholar]
  4. Ball K, Pajor A. Geometry of Banach spaces (Strobl, 1989), vol 158 of London Math Soc Lecture Note Ser. Cambridge Univ. Press; Cambridge: 1990. The entropy of convex bodies with “few” extreme points; pp. 25–32. [Google Scholar]
  5. Blei R, Gao F, Li WV. Metric entropy of high dimensional distributions. Proc Amer Math Soc. 2007;135:4009–4018. [Google Scholar]
  6. Bourgain J, Pajor A, Szarek SJ, Tomczak-Jaegermann N. Geometric aspects of functional analysis (1987–88), vol 1376 of Lecture Notes in Math. Springer; Berlin: 1989. On the duality problem for entropy numbers of operators; pp. 50–63. [Google Scholar]
  7. Carl B. Metric entropy of convex hulls in Hilbert spaces. Bull London Math Soc. 1997;29:452–458. [Google Scholar]
  8. Carl B, Kyrezi I, Pajor A. Metric entropy of convex hulls in Banach spaces. J London Math Soc. 1999;60(2):871–896. [Google Scholar]
  9. Dudley RM. Universal Donsker classes and metric entropy. Ann Probab. 1987;15:1306–1326. [Google Scholar]
  10. Feller W. An introduction to probability theory and its applications. Second edition. II. John Wiley & Sons Inc; New York: 1971. [Google Scholar]
  11. Gao F. Entropy of absolute convex hulls in Hilbert spaces. Bull London Math Soc. 2004;36:460–468. [Google Scholar]
  12. Gao F. Entropy estimate for k-monotone functions via small ball probability of integrated Brownian motion. Electron Commun Probab. 2008;13:121–130. [Google Scholar]
  13. Gao F, Wellner JA. On the rate of convergence of the maximum likelihood estimator of a k-monotone density. Science in China, Series A Mathematics. 2009;52:1525–1538. doi: 10.1007/s11425-009-0102-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Jewell NP. Mixtures of exponential distributions. Ann Statist. 1982;10:479–484. [Google Scholar]
  15. Krattenthaler C. Sém Lothar Combin. Vol. 42. The Andrews Festschrift; Maratea, 1998: 1999. Advanced determinant calculus. Art. B42q, 67 pp. (electronic) [Google Scholar]
  16. Kuelbs J, Li WV. Metric entropy and the small ball problem for Gaussian measures. J Funct Anal. 1993;116:133–157. [Google Scholar]
  17. Li WV, Linde W. Approximation, metric entropy and small ball estimates for Gaussian measures. Ann Probab. 1999;27:1556–1578. [Google Scholar]
  18. Li WV, Linde W. Metric entropy of convex hulls in Hilbert spaces. Studia Math. 2000;139:29–45. [Google Scholar]
  19. Tomczak-Jaegermann N. Dualité des nombres d’entropie pour des opérateurs à valeurs dans un espace de Hilbert. C R Acad Sci Paris Sér I Math. 1987;305:299–301. [Google Scholar]
  20. van der Vaart AW, Wellner JA. Weak convergence and empirical processes. Springer Series in Statistics. Springer-Verlag; New York: 1996. With applications to statistics. [Google Scholar]
  21. Widder DV. The Laplace Transform. Princeton Mathematical Series, v. 6. Princeton University Press; Princeton, N. J: 1941. [Google Scholar]
  22. Williamson RE. Multiply monotone functions and their Laplace transforms. Duke Math J. 1956;23:189–207. [Google Scholar]

RESOURCES