Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2017 Dec 1.
Published in final edited form as: Stat Probab Lett. 2016 Sep 10;119:317–325. doi: 10.1016/j.spl.2016.09.004

Double asymptotics for the chi-square statistic

Grzegorz A Rempała a, Jacek Wesołowski b
PMCID: PMC5383219  NIHMSID: NIHMS818460  PMID: 28392612

Abstract

Consider distributional limit of the Pearson chi-square statistic when the number of classes mn increases with the sample size n and n/mnλ. Under mild moment conditions, the limit is Gaussian for λ = ∞, Poisson for finite λ > 0, and degenerate for λ = 0.

Keywords: Pearson chi-square statistic, central limit theorem, Poisson limit theorem, weak convergence

1. Preliminaries

The Pearson chi-square statistic is probably one of the best-known and most important objects of statistical science and has played a major role in statistical applications ever since its first appearance in Karl Pearson’s work on “randomness testing” (Pearson, 1900). The standard test for goodness-of-fit with the Pearson chi-square statistic tacitly assumes that the support of the discrete distribution of interest is fixed (whether finite or not) and unaffected by the sampling process. However, this assumption may be unrealistic for modern ’big-data’ problems which involve complex, adaptive data acquisition processes (see, e.g., Grotzinger et al. 2014 for an example in astrobiology). In many such cases the associated statistical testing problems may be more accurately described in terms of triangular arrays of discrete distributions whose finite supports are dependent upon the collected samples and increase with the samples’ size (Pietrzak et al., 2016). Motivated by ’big-data’ applications, in this note we establish some asymptotic results for the Pearson chi-square statistic for triangular arrays of discrete random variables for which their number of classes mn grows with the sample size n. Specifically, let Xn,k, k = 1, . . . , n, be iid random variables having the same distribution as Xn, where

(Xn=i)=pn(i)>0,i=1,2,,mn<,n=1,2,

Recall that the standard Pearson chi-square statistic is defined as

χn2=ni=1mn(p^n(i)-pn(i))2pn(i), (1)

where the empirical frequencies n(i) are

p^n(i)=n-1k=1nI(Xn,k=i),i=1,,mn.

As stated above, in what follows we will be interested in the double asymptotic analysis of the weak limit of χn2, that is, the case when mn → ∞ as n → ∞.

Observe that χn2 given in (1) can be decomposed into a sum of two uncorrelated components as follows

χn2=n-1(Un+Sn)-n, (2)

where

Un=1klnI(Xn,k=Xn,l)pn(Xn,k) (3)

and

Sn=k=1n1pn(Xn,k)=k=1npn-1(Xn,k). (4)

The second equality above introduces notational convention we use throughout. Note that for fixed n the statistic S n is simply a sum of iid random variables and Un is an unnormalized U-statistic (see, e.g., Korolyuk and Borovskich, 2013). It is routine to check that

EUn=n(n-1)andESn=nmn

and consequently

Eχn2=mn-1.

Moreover, since we also have ℂov(Un, S n) = 0, it follows that

Varχn2=n-2(VarSn+VarUn)=n-1[Varpn-1(Xn)+2(n-1)(mn-1)].

When mn = m is a constant then the classical result (see, e.g., Shao, 2003, chapter 6) implies that the statistic χn2 asymptotically follows the χ2-distribution with (m−1) degrees of freedom. Consequently, when m is large the standardized statistic (χn2-(m-1))/2(m-1) may be approximated by the standard normal distribution. However, in the case when mn → ∞ as n → ∞ the matters appear to be more subtle and the above normal approximation may or may not be valid depending upon the asymptotic relation of mn and n, as described below. Since S n is a sum of iid random variables, the case when S n contributes to the limit of normalized χn2 may be largely handled with the standard theory for arrays of iid variables. Consequently, we focus here on a seemingly more interesting case when the asymptotic influence of Un dominates over that of S n. Specifically, throughout the paper we assume that as n, mn → ∞

(mnn)-1Varpn-1(Xn)0. (C)

Note that (C) implies n-1(Sn-nmn)/2mn0 in probability and, in particular, is trivially satisfied when Xn is a uniform random variable on the integer lattice 1, . . . , mn, that is, when pn(i)=mn-1 for i = 1 . . . , mn. Under condition (C) we get a rather complete picture of the limiting behavior of χn2. Our main results are presented in Section 2 where we discuss the Poissonian and Gaussian asymptotics. Some examples, relations to asymptotics known in the literature and further discussions are provided in Section 3. The basic tools used in our derivations are listed in the appendix. In what follows limits are taken as n → ∞ with mn → ∞ and d stands for convergence in distribution.

2. Poissonian and Gaussian asymptotics

We start with the case when a naive normal approximation for the standardized χn2 statistic fails. Indeed, as it turns out, when mn is asymptotically of order n2, we have the following Poisson limit theorem for χn2.

Theorem 2.1

Assume that the condition (C) holds, as well as

nmnλ(0,). (5)

Then

χn2-mn2mnd2λZ-λ2,Z~Pois(λ22) (6)

Proof

Due to (C) it suffices to consider the asymptotics of Un alone. We write

Un-n(n-1)n2mn=2mnnk=1nAn,k-n-12mn, (7)

where An,1 = 0 and for k = 2, . . . , n

An,k=mn-1j=1k-1I(Xn,j=Xn,k)pn(Xn,j)=mn-1pn-1(Xn,k)j=1k-1I(Xn,j=Xn,k). (8)

The above representation implies that to prove (6) we need only to show that k=1nAn,kdPois(λ22). To this end we will verify the conditions of Theorem A.1 in the appendix, due to Beśka, Kłopotowski and Słomiński (Beśka et al., 1982). Denote ℱn,0 = {∅, Ω} and ℱn,k = σ(Xn,1, . . . , Xn,k), k = 1, . . . , n. Then using the first form of An,k from (8) we see that

max1knE(An,kFn,k-1)=mn-1max1knj=1k-1E(I(Xn,j=Xn,k)pn(Xn,j)|Fn,k-1)=max1knk-1mn=n-1mn0

due to (5) and thus (A.1) holds. Similarly,

k=1nE(An,kFn,k-1)=k=1nk-1mn=n(n-1)2mnλ22 (9)

and thus (A.2) also follows with η=λ22. Since An,k ≥ 0 the required convergence in (A.3) (for any ε > 0) will follow from convergence of the unconditional moments

k=1nEAn,kI(An,k-1>ε)ε-2k=1n(EAn,k3-2EAn,k2+EAn,k). (10)

Using the second form of An,k from (8) we see that the conditional distribution of mn pn(Xn,k) An,k given Xn,k follows a binomial distribution Binom(k − 1, pn(Xn,k)). Since for M ~ Binom(r, p) we have 𝔼 M = rp, 𝔼 M2 = rp + r(r − 1) p2 and 𝔼 M3 = rp +3r(r − 1)p2 +r(r − 1)(r − 2)p3, we thus obtain

k=1nEAn,k=1mnk=1n(k-1)n22mnλ22,k=1nEAn,k2=1mn2k=1n((k-1)mn+(k-1)(k-2))n22mn+n33mn2λ22.

Similarly,

k=1nEAn,k3=1mn3k=1n((k-1)Epn-2(Xn)+3(k-1)(k-2)mn+(k-1)(k-2)(k-3))n22mn3Epn-2(Xn)+n3mn3+n44mn3.

Note that (C) and (5) imply mn-2Epn-2(Xn)1 and therefore

k=1nEAn,k3n22mn3E1pn2(Xn)λ22.

Combining the limits of the last three expressions we conclude that the right-hand side of (10) tends to zero and hence (A.3) of Theorem A.1 is also satisfied. The result follows.

Let us now consider the case nmn. As it turns out, under this condition the statistic χn2 is asymptotically Gaussian.

Theorem 2.2

Assume that condition (C) is satisfied and that there exists δ > 0 such that

supnmn-(1+δ)Epn-(1+δ)(Xn)< (11)

as well as

nmn. (12)

Then

χn2-mn2mndN,N~Norm(0,1). (13)

Remark 2.3

Note that under (C) the conditions (11) (with δ = 1) and (12) are implied by the condition n/mnλ ∈ (0, ∞).

Proof

As in Theorem 2.1, under our assumption (C) it suffices to show convergence in distribution to N ~ Norm(0, 1) of the normalized Un variable

Un-n(n-1)n(n-1)2(mn-1)=k=1nYn,k,

where

Yn,k=2n(n-1)(mn-1)j=1k-1(I(Xn,j=Xn,k)pn(Xn,j)-1)=2Bn,kn(n-1)(mn-1) (14)

and the last equality defines Bn,k. Since 𝔼(I(Xn,k = Xn, j)|ℱn,k−1) = pn(Xn, j) for any j = 1, . . . , k − 1, it follows that 𝔼(Yn,k|ℱn,k−1) = 0. Consequently, (Yn,k, ℱn,k)k=1,...,n are martingale differences. Therefore, to prove (13) we may use the Lyapounov version of the CLT for martingale differences (see Theorem A.2 in the appendix).

Due to (14) we have

E(Bn,k2Fn,k-1)=j=1k-1Var(I(Xn=Xn,j)Fn,k-1)pn2(Xn,j)+1ijk-1ov(I(Xn=Xn,i),I(Xn=Xn,j)Fn,k-1)pn(Xn,i)pn(Xn,j).

Since 𝕍ar(I(Xn = Xn, j)|ℱn,k−1) = pn(Xn, j)(1 − pn(Xn, j)) and

ov(I(Xn=Xn,i),I(Xn=Xn,j)Fn,k-1)=I(Xn,i=Xn,j)pn(Xn,i)-pn(Xn,i)pn(Xn,j)

we obtain

E(Bn,k2Fn,k-1)=j=1k-1(pn-1(Xn,j)-1)+1ijk-1(I(Xn,i=Xn,j)pn(Xn,i)-1).

Consequently, (A.4) is equivalent to

k=1nj=1k-1(pn-1(Xn,j)-mn)n(n-1)2(mn-1)+k=1n1ijk-1(I(Xn,i=Xn,j)pn(Xn,i)-1)n(n-1)2(mn-1)0. (15)

To show the above, we separately consider moments of the summands on the left-hand side of (15). For the first one, note that

k=1nj=1k-1(pn-1(Xn,j)-mn)=j=1n-1(n-j)(pn-1(Xn,j)-mn)=dj=1n-1j(pn-1(Xn,j)-mn)

where the last equality denotes the distributional equality of random variables. Therefore, using inequality (B.2) given in the appendix, we get (possibly with different universal constants C from line to line)

E|k=1nj=1k-1(pn-1(Xn,j)-mn)n(n-1)2(mn-1)|1+δCE|j=1n-1j(pn-1(Xn,j)-mn)|1+δn2+2δmn1+δCE|pn-1(Xn,j)-mn|1+δnδ-120j=1n-1j1+δn2+2δmn1+δCE|pn-1(Xn,j)-mn|1+δn3(1+δ)2(2+δ)n2+2δmn1+δCE|pn-1(Xn,j)-mn|1+δn1+δ2δmn1+δ.

In view of this and the elementary inequality |a + b|pC(|a|p + |b|p) valid for any p > 0 and any real a, b we have for some constants C1, C2

E|k=1nj=1k-1(pn-1(Xn,j)-mn)n(n-1)2(mn-1)|1+δC1n1+δ2δEpn-(1+δ)(Xn)mn1+δ+C2n1+δ2δ0.

For the numerator of the second part on the left hand side of (15) we may write

k=1n1ijk-1(I(Xn,i=Xn,j)pn(Xn,i)-1)=21i<jn-1(n-j)(I(Xn,i=Xn,j)pn(Xn,i)-1).

Moreover,

E(1i<jn-1(n-j)(I(Xn,i=Xn,j)pn(Xn,i)-1))2=1i<jn-1(n-j)2E(I(Xn,i=Xn,j)pn(Xn,i)-1)2,

since the expectations of the other terms resulting from squaring the large-bracketed first expression above are equal to zero. Consequently

E(1i<jn-1(n-j)(I(Xn,i=Xn,j)pn(Xn,i)-1))2=(mn-1)1i<jn-1(n-j)2Cmnn4

and thus for the squared expectation of the second term in (15) we get

E(k=1n1ijk-1(I(Xn,i=Xn,j)pn(Xn,i)-1)n(n-1)2(mn-1))2Cmn-10.

Note that here we used the fact that mn → ∞. To finish the proof we only need to show (A.5). Again we will rely on the representation of Yn,k given in (14). Note that

E|Yn,k|2+δCn-(2+δ)mn-(1+δ2)E(pn-(2+δ)(Xn,k)|j=1k-1(I(Xn,j=Xn,k)-pn(Xn,k))|2+δ).

Since I(Xn, j = Xn,k) − pn(Xn,k), j = 1, . . . , k − 1, are conditionally iid given Xn,k and

E((I(Xn,j=Xn,k)-pn(Xn,k))Xn,k)=0

then by conditioning with respect to Xn,k and applying Rosenthal’s inequality (see (B.1) in the appendix) to the conditional moment of the sum we obtain

k=1nE|Yn,k|2+δCn2+δmn1+δ2k=1nE(pn-(2+δ)(Xn)((k-1)pn(Xn)+[(k-1)pn(Xn)]1+δ2))C(n-δmn-(1+δ2)Epn-(1+δ)(Xn)+n-δ2mn-(1+δ2)Epn-(1+δ2)(Xn)). (16)

By virtue of the Schwartz inequality we obtain that

n-δ2mn-(1+δ2)Epn-(1+δ2)(Xn)=n-δ2mn-(1+δ2)Epn-12(Xn)pn-1+δ2(Xn)n-δ2mn-(1+δ)Epn-(1+δ)(Xn)0

in view of (11). Therefore, it only suffices to show that the first term in the last expression in (16) converges to zero. But this follows due to (11) and (12), since

Epn-(1+δ)(Xn)nδmn1+δ2=(mnn)δEpn-(1+δ)(Xn)mn1+δ0.

3. Discussion

We will now illustrate the results of the previous section with some examples as well as put them in a broader context of earlier work by others. For the sake of completeness, we first note

Remark 3.1. The case λ = 0

Consider nmn0. Then the last part of the right hand side of (7) converges to zero and we are left with the sum of non-negative random variables which satisfies

2mnnk=1nAn,k0.

To see the above, it suffices to consider the convergence of the first moments. To this end note that

2mnnk=1nEAn,k=2mnnk=1nk-1mn=n-1mn0.

The simple illustration of Theorem 2.2 is as follows.

Example 3.1

Let α ∈ [0, 1) and set pn(i) = (Cαiα)−1 for i = 1, . . . , mn. Here Cα=i=1mni-αmn1-α/(1-α) in view of the general formula

i=1mniβmnβ+1/(β+1)forβ>-1. (17)

Note that for 0 < α < 1 the condition (C) is equivalent to

n/mn (18)

and implies (12). Applying (17) again we see that for any δ > 0

Epn-(1+δ)(Xn)mn1+δ=Cαδi=1mniαδmn1+δmn(1-α)δmn1+αδ(1-α)δ(1+αδ)mn1+δ=(1-α)-δ(1+αδ)-1<

and therefore (11) is also satisfied. Hence, the conclusion of Theorem 2.2 holds true under (18) for 0 < α < 1.

Note that in the above example the assumption (5) of Theorem 2.1 cannot be satisfied for 0 < α < 1 (see (18)) but can hold for α = 0, that is, when the distribution is uniform. We remark that in our present setting such distribution is of interest, for instance, when testing for signal-noise threshold in data with large number of support points (Pietrzak et al., 2016). Combining the results of Theorems 2.1 and 2.2 and Remark 3.1 one obtains the following.

Corollary 3.2 (Asymptotics of χn2 for uniform distribution)

Assume that pn(i)=mn-1 for i = 1, 2, . . . , mn and n = 1, 2, . . . as well as

n/mnλ.

Then

χn2-mn2mnd{0whenλ=0,2λZ-λ2,Z~Pois(λ22)whenλ(0,),N~Norm(0,1)whenλ=.

We note that the asymptotic distribution of χn2 when both n and mn tend to infinity has been considered by several authors, typically in the context of asymptotics of families of goodness-of-fit statistics related to different divergence distances. Some of these results considered also the asymptotic behavior of such statistics not only under the null hypothesis (as we did here) but also under simple alternatives and hence are, in that sense, more general. However, when applied to the chi-square statistic under the null hypothesis they appear to be special cases of our theorems in Section 2. We briefly review below some of the most relevant results.

Tumanyan (1954, 1956) proved asymptotic normality of χn2 under the assumption min1≤imn npn(i) → ∞ which in the case of the uniform distribution is equivalent to n/mn → ∞, a condition obviously stronger than n/mn we use (see Corollary 3.2).

Steck (1957) generalized these results on normal asymptotics assuming among other conditions that infn n/mn > 0 which again is stronger than n/mn. He also obtained the Poissonian and degenerate limit in the case of uniform distribution, in agreement with the first two cases in our Corollary 3.2. The main result of Holst (1972) for the chi-square statistic gives normal asymptotics under the regime n/mnλ ∈ (0, ∞) and max1≤jn pn( j) < β/n which also is stronger than our assumptions. In the uniform case under this regime the result was proved earlier in Harris and Park (1971). The main result of Morris (1975) for the chi-square statistics gives asymptotic normality under n min1≤jn pn( j) > ε > 0 for all n ≥ 1, max1≤jn pn( j) → 0 and the ”uniform asymptotically negligible” condition of the form max1imnσn2(i)/sn20, where σn2(i)=2+(1-mnpn(i))2npn(i), i = 1, . . . , mn, and sn2=i=1mnσn2(i). In the case of the uniform distribution it gives asymptotic normality of χn2 under the condition n/mn > ε > 0, the result apparently weaker than the third part of Corollary 3.2.

Following the paper of Cressie and Read (1984) introducing the family of power divergence statistics (of which the chi-square statistic is a member), much effort was directed at proving asymptotic normality for wider families of divergence distances as well as for more than one multinomial independent sample, see e.g. Menéndez et al. (1998); Pérez and Pardo (2002) (in both papers the authors considered the regime n/mnλ ∈ (0, ∞)) and Inglot et al. (1991), Morales et al. (2003) (in both papers the authors considered the regime mn1+βlog2(n)/n0 and mnβmin1jnpn(j)>c>0 for some β ≥ 1) or Pietrzak et al. (2016) (with the regime n/mn → ∞). Note that for the asymptotic normality results all these regimes are again more stringent than what we consider here.

Finally, for completeness, we briefly address one of the scenarios when condition (C) does not hold.

Remark 3.3

Note that if mnnVarpn-1(Xn)0then the asymptotic behavior of standardized χn2 is the same as that of Zn=k=1nYn,k, where

Yn,k=pn-1(Xn,k)-mnnVarpn-1(Xn),k=1,,n.

Since for any fixed n ≥ 1 random variables Yn,k, k = 1, . . . , n, are iid (zero mean) and 𝕍ar Yn,k = n−1 it follows that {Yn,k, k = 1, . . . , n}n≥1 is an infinitesimal array. Therefore classical CLT for row-wise iid triangular arrays (cf., e.g., Shao, 2003, chapter 1) applies. Note also that the remaining case when mnnVarpn-1(Xn)λ(0,) appears more complicated and requires a different approach.

Acknowledgments

The research was conducted when the second author was visiting The Mathematical Biosciences Institute at OSU. Both authors thank the Institute for its logistical support and funding through US NSF grant DMS-1440386. The research was also partially funded by US NIH grant R01CA-152158 and US NSF grant DMS-1318886. The authors wish to gratefully acknowledge helpful comments made by the referee and the associate editor on the early version of the manuscript.

Appendix A. Limit Theorems

Below, for convenience of the readers, we recall some results which are used in the proofs. The first one is found in Beśka et al. (1982) and the second one is a version of the martingale CLT (see, e.g., Hall and Heyde, 1980).

Theorem A.1 (Poissonian conditional limit theorem)

Let {Zn,k, k = 1, . . . , n; n ≥ 1} be a double sequence of non-negative random variables adapted to a row-wise increasing double sequence of σ-fields {𝒢n,k−1, k = 1, . . . , n; n ≥ 1}. If for n → ∞

max1knE(Zn,kGn,k-1)0, (A.1)
k=1nE(Zn,kGn,k-1)η>0, (A.2)

and for any ε > 0

k=1nE(Zn,kI(Zn,k-1>ε)Gn,k-1)0, (A.3)

then k=1nZn,kdZ, where Z ~ Pois(η) is a Poisson random variable.

Theorem A.2 (Lyapunov-type martingale CLT)

Let {(Zn,k, ℱn,k) k = 1, . . . , n; n ≥ 1} be a double sequence of martingale differences. If

k=1nE(Yn,k2Fn,k-1)1 (A.4)

and

k=1nEYn,k2+δ0. (A.5)

then k=1nZn,kdN, where N ~ Norm(0, 1) is a standard normal random variable.

Appendix B. Moment Inequalities

The following moment inequalities are used in Section 2.

Rosenthal inequality

Rosenthal (1970). If X1, . . . , Xn are independent and centered random variables such that 𝔼|Xi|r < ∞, i = 1, . . . , n and r > 2 then

E|i=1nXi|rCrmax{i=1nEXir,(i=1nEXi2)r2}Cr(i=1nEXir+(i=1nEXi2)r2). (B.1)

MZ-BE inequality

Marcinkiewicz and Zygmund (1937) for r ≥ 2, von Bahr and Esseen (1965) for 1 ≤ r ≤ 2. If X1, . . . , Xn are independent and centered random variables such that 𝔼|Xi|r < ∞, i = 1, . . . , n then for r > 1

E|i=1nXi|rCrnri=1nEXir, (B.2)

where r=0(r2-1).

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

References

  1. Beśka M, Kłopotowski A, Słomiński L. Limit theorems for random sums of dependent d-dimensional random vectors. Probability Theory and Related Fields. 1982;61(1):43–57. [Google Scholar]
  2. Cressie N, Read TR. Multinomial goodness-of-fit tests. Journal of the Royal Statistical Society. Series B (Methodological) 1984:440–464. [Google Scholar]
  3. Grotzinger JP, Sumner D, Kah L, Stack K, Gupta S, Edgar L, Rubin D, Lewis K, Schieber J, Mangold N, et al. A habitable fluvio-lacustrine environment at Yellowknife Bay, Gale Crater, Mars. Science. 2014;343(6169):1242777. doi: 10.1126/science.1242777. [DOI] [PubMed] [Google Scholar]
  4. Hall P, Heyde CC. Martingale limit theory and its application. New York: Academic Press; 1980. includes indexes. [Google Scholar]
  5. Harris B, Park C. Indagationes Mathematicae (Proceedings) Vol. 74. Elsevier; 1971. The distribution of linear combinations of the sample occupancy numbers; pp. 121–134. [Google Scholar]
  6. Holst L. Asymptotic normality and efficiency for certain goodness-of-fit tests. Biometrika. 1972;59(1):137–145. [Google Scholar]
  7. Inglot T, Jurlewicz T, Ledwina T. Asymptotics for multinomial goodness of fit tests for a simple hypothesis. Theory of Probability & Its Applications. 1991;35(4):771–777. [Google Scholar]
  8. Korolyuk VS, Borovskich YV. Theory of U-statistics. Vol. 273. Springer Science & Business Media; 2013. [Google Scholar]
  9. Marcinkiewicz J, Zygmund A. Quelques théoremes sur les fonctions indépendantes. Fund Math. 1937;29:60–90. [Google Scholar]
  10. Menéndez M, Morales D, Pardo L, Vajda I. Asymptotic distributions of φ-divergences of hypothetical and observed frequencies on refined partitions. Statistica Neerlandica. 1998;52(1):71–89. [Google Scholar]
  11. Morales D, Pardo L, Vajda I. Asymptotic laws for disparity statistics in product multinomial models. Journal of Multivariate Analysis. 2003;85(2):335–360. [Google Scholar]
  12. Morris C. Central limit theorems for multinomial sums. The Annals of Statistics. 1975:165–188. [Google Scholar]
  13. Pearson K. On the criterion that a given system of deviations from the probable in the case of a correlated system of variables is such that it can be reasonably supposed to have arisen from random sampling. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science. 1900;50(302):157–175. [Google Scholar]
  14. Pérez T, Pardo J. Asymptotic normality for the Kϕ-divergence goodness-of-fit tests. J Comput Appl Math. 2002;145:301–317. [Google Scholar]
  15. Pietrzak M, Rempała GA, Seweryn M, Wesołowski J. Limit theorems for empirical Rényi entropy and divergence with applications to molecular diversity analysis. TEST. 2016:1–20. [Google Scholar]
  16. Rosenthal HP. On the subspaces ofl p (p¿ 2) spanned by sequences of independent random variables. Israel Journal of Mathematics. 1970;8(3):273–303. [Google Scholar]
  17. Shao J. Springer Texts in Statistics. Springer; 2003. Mathematical Statistics. [Google Scholar]
  18. Steck GP. Limit theorems for conditional distributions. Univ California Publ Statist. 1957;2(12):237–284. [Google Scholar]
  19. Tumanyan SK. On the asymptotic distribution of the chi-square criterion. Dokl Akad Nauk SSSR. 1954;94:1011–1012. [Google Scholar]
  20. Tumanyan SK. Asymptotic distribution of the chi-square criterion when the number of observations and number of groups increase simultaneously. Teor Veroyat Yeyo Primen. 1956;1(1):131–145. [Google Scholar]
  21. von Bahr B, Esseen CG. Inequalities for the r-th absolute moment of a sum of random variables, 1 ≦ r ≦ 2. The Annals of Mathematical Statistics. 1965;36(1):299–303. [Google Scholar]

RESOURCES