Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2013 Dec 16.
Published in final edited form as: Commun Stat Theory Methods. 2007 Feb 15;30(2):10.1081/STA-100002037. doi: 10.1081/STA-100002037

On the Expected Values of Sequences of Functions

Deborah H Glueck 1, Keith E Muller 2
PMCID: PMC3864817  NIHMSID: NIHMS446107  PMID: 24353369

Abstract

We prove new extensions to lemmas about combinations of convergent sequences of distribution functions and absolutely continuous bounded functions. New lemma one, a generalized Helly theorem, allows computing the limit of the expected value of a sequence of functions with respect to a sequence of measures. Previously published results allow either the function or the measure to be a sequence, but not both. Lemma two allows computing the expected value of an absolutely continuous monotone function by integrating the probabilities of the inverse function values. Previous results were restricted to the identity function. Lemma three gives a computationally and analytically convenient form for the limit of the expected value of a sequence of functions of a sequence of random variables. This is a new result that follows directly from the first two lemmas. Although the lemmas resemble standard results and seem obviously true, we have found only similar looking and related but quite distinct results in the literature. We provide examples which highlight the value of the new results.

Keywords: Absolutely continuous, Inversion, Integrals

1. Introduction

Computing expectations, both analytically and numerically, remains one of the central problems in statistics. We present three new lemmas which aid both analytic and numerical calculation for some applications in small and large samples.

In particular, the first lemma allows calculation of the limits of expectations, when both the function and the random variable are converging to limits. The second lemma suggests natural transformations for the computation of expectations, a common statistical task. We suggest transformations based on probability distribution functions and their inverses. The required numerical functions are widely available in common statistical programming languages. The choice automatically simplifies numerical computation of expectations by leading to evaluating bounded functions on bounded regions. Although a transformation is a standard numerical technique, the use of probability functions eliminates the guess work involved in choosing the transformation, and provides an ideal match between statistical thinking and ease of computation. The third lemma combines the results and allows one to calculate an expectation when both the function and the random variable are converging to a limit.

The need for the results presented here arose from a desire to derive computable expressions for the power of certain hypothesis tests in multivariate regression with both fixed and random predictors. We provide examples of how we used the lemmas to numerically calculate small sample expectations, and to correctly find limiting results.

Calculating the expected value of variables and functions can be difficult, especially in the limit. Often either the function of integration, or the limits, or both, are unbounded. In attempting to derive large sample properties, a sequence of cumulative distribution functions (CDF's) may converge to a point mass, while their support converges to a set of measure zero. The Riemann integral fails under these conditions. Thus the lemmas that follow must be stated in terms of Lebesgue integrals computed with respect to probability measures.

Although the lemmas resemble standard results and seem obviously true, we have found only similar looking and related but quite distinct results in the literature. All depend on various restrictive regularity conditions that often arise in practice. For example, for a sequence of cumulative distribution functions, {Fn}, Serfling (1, p16) proved that if FnF then for any bounded continuous function g,

limng dFn=g dF. (1)

Pratt (2, p74) and Loeve (3, pl26) gave results similar to Equation 1. We give a more general result in which both the integrand and the measure are converging. Gibbons and Chakraborti (4, p37-38) mentioned a quantile transform result, without proof. They define the quantile function using the infimum. Our Corollary 3 is a special case of their result. We illustrate the value of the new results with three examples concerning power of certain multivariate tests.

2. Three Lemmas

Lemma 1

Consider a random variable, X, and a sequence of random variables, {X1, X2, …}, with corresponding CDFs F, and {F1, F2, …}. Suppose Fn converges to F, and thus Xn converges in distribution to X. Let {g1(x), g2(x), …} be a set of continuous bounded functions such that gn(x) converge uniformly to g(x), a continuous bounded function. Assume that ∀n ∫ gn dFn < ∞ and ∫ g dF < ∞. Taking the integrals with respect to the Lebesgue-Stieltjes probability measures induced by F and {F1, F2, …} (5, p69),

limngndFn=g dF. (2)

The special case of ∫ gd Fn corresponds to a Helly theorem, (5, p192–194). Also, the results of exercise 9-2 in Burrill (5, p195) indicate that less stringent regularity conditions would be hard to find for the special case of ∫ gnd F.

Proof. It suffices to show that ∀∊ > 0, ∃ M(∊) > 0 such that for n > M(∊),

|gnd Fng dF|<. (3)

By assumption, gn(x) converges uniformly to g(x). Thus, ∀∊ > 0, there exists a corresponding number M1(∊) > 0 such that for all n > M1 and for all x in the domain of g, |gn(x) − g(x)| < ∊/2 (6, p530). Thus, for n > M1(∊),

|gnd Fng dFn|=|(gng)d Fn|||gng|d Fn|<||/2|d Fn||/2||d Fn|/2. (4)

The last step follows because Fn is a cumulative probability distribution function, and hence ∀n, ∫ d Fn = 1. By part 3 of a theorem in Serfling (1, p16), ∀ ∊ > 0, we may conclude that ∃ M2(∊) > 0 such that for n > M2(∊),

|g dFng dF|</2. (5)

Now, ∀∊ > 0 and ∀x in the domain of g, choose n > max[M1(∊), M2(∊)]. Then

|gndFng dF|</2+/2=, (6)

with the inequality following from the triangle inequality.

Lemma 2

(Corollary to Lemma 2.1, 7, p243) Let X be a continuous random variable with density fx(x) and distribution function FX(x). Let g(x) be a real valued absolutely continuous function that is strictly monotone decreasing in x, so that g(x) > y iff x < g−1(y). Let

A={y:y0}{y:P{g(X)>y}>0}and={y:(y>0)}{y:P{g(X)<y}>0}.Thenɛ[g(X)]=AFX[g1(y)]dy{1FX[g1(y)]}dy. (7)

Proof. Note that

ɛ[g(X)]=AP{g(X)>y}dyP{g(X)<y}dy=AP{X<g1(y)}dyP{X>g1(y)}dy=AP{X<g1(y)}dy(1P{X<g1(y)})dy. (8)

The result follows.

Corollary 2.1

With the same conditions as in Lemma 1, and for b > a > 0, suppose ∀x ∈ ℛ, g(x) ∈ [a, b]. Then

ɛ[g(X)]=abFX[g1(y)]dy. (9)

Corollary 2.2

With the same conditions as in Lemma 1, consider instead h(x), a real valued absolutely continuous function that is strictly monotone increasing in x, so that h(x) > y iff x > h−1(y). Then

ɛ[h(X)]=A{1FX[h1(y)]}dyFX[h1(y)]dy. (10)

Lemma 3

Consider the continuous random variable, X, and the sequence of continuous random variables, {X1, X2, …}, with the same assumptions as Lemma 1.Let g(x)and the set {g1(x), g2(x), …} be real valued absolutely continuous bounded functions that are strictly monotone decreasing in x, so that gn(x) > y iff x<gn1(y). Suppose the sequence {g1(x), g2(x), …} converges uniformly to g(x). Assume that ∫ g dF < ∞ and ∀n, ∫ gn dFn < ∞. For b > a > 0, suppose ∀x ∈ ℛ, g(x) ∈ [a, b]. Then

limnɛ[gn(Xn)]=abFX[g1(y)]dy. (11)

Proof. Follows directly from Lemma 1 and Lemma 2.

Corollary 3

(Quantile transformation: see 4, §2.5, p37–38) Consider a real valued random variable X with strictly monotone distribution function Fx(x) and density function fx(x), defined on the interval (a, b), with −∞ < a < b < ∞. Let y = Fx(x). Then

ɛ(X)=01FX1(y)dy. (12)

When resorting to numerical techniques for calculating expectations, either the density or the region of integration, or both, may be infinite. This transformation reduces the problem to an integral of a bounded function over a bounded interval.

3. Examples

Example for Lemma 1

Glueck (8) considered taking the limit, under a sequence of Pitman local alternatives, of an approximation for the power of the Hotelling-Lawley trace statistic, with Gaussian predictors. The asymptotic power can be written as the expected value of a non-central F, with respect to the distribution of a random noncentrality value. In this setting, the Riemann integral is undefined because the support for the random noncentrality parameter converges to set of measure zero as the parameter converges to a point. Let F(ν1, ν2N, ωN) indicate a noncentral F random variable, with denominator degrees of freedom and noncentrality depending on N. Suppose

gN=Pr{F(ν1,ν2N,ωN)fN}. (13)

Consider integrating gN with respect to Fn, the distribution function of a sum of independent scaled χ2 random variables for which the scaling constants depend on ωN, and the degrees of freedom depend on N. With α the type 1 error rate, ccrit chosen so that Fχ2(ccrit; ab) = 1 − α, and limN→∞, ω = ωL,

limNgNdFN=1Fχ2(ccrit;ν1,ωL). (14)

Example for Corollary 2.1

Glueck (8) sought a computational form for the small sample power of the Hotelling-Lawley trace statistic, with Gaussian predictors. Suppose Fω(w) is the distribution function of ω, a sum of independent scaled χ2 random variables. Define

g(ω)=Pr{F(ν1,ν2,ω)<f}, (15)

with ω ∈ [0, ∞] and f chosen so that g(ω) ∈ [0, 1 − α]. Then

ɛ[g(ω)]=0g(ω)dFω=01αFω[F1(f;ν1,ν2,y)]dy. (16)

Example for Corollary 3

Muller and Pasour (9) defined

g(t)=Fχ2(ν1sν2sfRt;ν1s,ωs)Fχ2(ν1sν2sfLt;ν1s,ωs), (17)

and considered integrating

FV(z)=πs10zg(t)fχ2(t;ν2s)dt. (18)

They used a particular quantile transformation to produce a much better behaved numerical integral. If p = Fχ2(t; ν2s), then t=Fχ21 (p; ν2s) and dp = fχ2(t; ν2s)dt. If p0 = Fχ2(2s/σ2; ν2s) then

FV(z)=πs10p0gR[Fχ21(p;ν2s)]dpπs10p0gL[Fχ21(p;ν2s)]dp. (19)

Acknowledgments

The authors gratefully acknowledge the help of Drs. G. G. Koch, R. M. Hamer, L. M. LaVange, D. F. Ransohoff and P. W. Stewart. In addition, an anonymous Associate Editor stimulated us to clarify the motivation of the paper, provide more compelling examples, and improve the readability of the paper. Also, an anonymous Associate Editor brought the work of Gibbons and Chakraborti to our attention. Glueck was supported in part by grant number T32HS000589-04 from the Agency for Health Care Policy and Research to the University of Medicine and Dentistry of New Jersey, and by National Cancer Institute Grant IP30CA46934-01 to the University of Colorado. Muller's work supported in part by NCI program project grant P01 CA47 982-04.

References

  • 1.Serfling RJ. Approximation Theorems of Mathematical Statistics. John Wiley and Sons; New York: 1980. [Google Scholar]
  • 2.Pratt JW. On interchanging limits and integrals. Annals of Mathematical Statistics. 1960;31:74–77. [Google Scholar]
  • 3.Loève M. Probability Theory II. 4th. Springer-Verlag; New York: 1977. [Google Scholar]
  • 4.Gibbons JD, Chakraborti S. Nonparametric Statistical Inference. 3rd. Marcel Dekker; New York: 1992. [Google Scholar]
  • 5.Burrill CW. Measure, Integration and Probability. McGraw-Hill; New York: 1972. [Google Scholar]
  • 6.Courant R, John F. Introduction to Calculus and Analysis, Volume One. Interscience Publishers; New York: 1965. [Google Scholar]
  • 7.Ross S. A First Course in Probability. 2nd. Macmillan; New York: 1984. [Google Scholar]
  • 8.Glueck DH. Power for a Generalization of the General Linear Multivariate Model with Fixed and Random Predictors, Mimeo Series No 2158T. Institute of Statistics; Chapel Hill, North Carolina: 1996. [Google Scholar]
  • 9.Muller KE, Pasour VB. Bias in linear model power and sample size due to estimating variance. Communications in Statistics: Theory and Methods. 1997;26:839–851. doi: 10.1080/03610929708831953. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES