Skip to main content
Springer logoLink to Springer
. 2018 Aug 23;2018(1):221. doi: 10.1186/s13660-018-1811-y

Some mean convergence theorems for arrays of rowwise pairwise negative quadrant dependent random variables

Tapas K Chandra 1, Deli Li 2, Andrew Rosalsky 3,
PMCID: PMC6132388  PMID: 30839650

Abstract

For arrays of rowwise pairwise negative quadrant dependent random variables, conditions are provided under which weighted averages converge in mean to 0 thereby extending a result of Chandra, and conditions are also provided under which normed and centered row sums converge in mean to 0. These results are new even if the random variables in each row of the array are independent. Examples are provided showing (i) that the results can fail if the rowwise pairwise negative quadrant dependent hypotheses are dispensed with, and (ii) that almost sure convergence does not necessarily hold.

Keywords: Array of rowwise pairwise negative quadrant dependent random variables, Pairwise independent random variables, Weighted averages, Degenerate mean convergence, Stochastic domination, Almost sure convergence

Introduction

For a sequence of independent and identically distributed (i.i.d.) random variables {Xn,n1} with EX1=0, Pyke and Root [12] established the degenerate mean convergence law

j=1nXjnL10. 1.1

A considerably simpler proof of the limit law (1.1) was obtained by Dharmadhikari [4] who did not refer to the Pyke and Root [12] article. Chandra [3] established the following more general result for mean convergence of weighted averages. Its proof is more natural, straightforward, and powerful than that of Dharmadhikari [4]. Chandra’s [3] method is novel in the sense that the level of truncation does not depend on n (the sample size), whereas Dharmadhikari [4] used the truncation level n. The limit law (1.1) is obtained immediately from the Chandra [3] result by taking an,j=n1, 1jn, n1.

Theorem 1.1

(Chandra [3], Theorem 1)

Let {Xn,n1} be a sequence of pairwise i.i.d. random variables with EX1=0, and let {an,j,1jn,n1} be a triangular array of constants such that

supn1j=1n|an,j|<andlimnj=1nan,j2=0.

Then

j=1nan,jXjL10.

In the current work, we extend in Theorems 3.1 and 3.2 this degenerate mean convergence theorem of Chandra [3] in two directions:

  • (i)

    Our results pertain to weighted averages either from an array of random variables whose nth row is comprised of kn pairwise negative quadrant dependent random variables, n1 (Theorem 3.1) or from an array of random variables whose nth row is comprised of kn pairwise independent random variables, n1 (Theorem 3.2). No independence or dependence conditions are imposed between the random variables from different rows of the arrays. The Chandra [3] result considered weighted averages from a sequence of pairwise i.i.d. random variables.

  • (ii)

    The random variables that we consider are assumed to be stochastically dominated by a random variable which is a weaker assumption than the assumption of Chandra [3] that the random variables are identically distributed.

The third main result (Theorem 3.3) establishes for an array of random variables whose nth row is comprised of kn pairwise negative quandrant dependent random variables, n1 a degenerate mean convergence result for normed and centered row sums. In contradistinction to Theorems 3.1 and 3.2, weighted averages and stochastic domination play no role in Theorem 3.3. As in Theorems 3.1 and 3.2, no independence or dependence conditions are imposed between the random variables from different rows of the array in Theorem 3.3.

Definition 1.1

A finite set of random variables {X1,,XN} is said to be pairwise negative quadrant dependent (PNQD) if for all i,j{1,,N} (ij) and all x,yR,

P(Xix,Xjy)P(Xix)P(Xjy). 1.2

It is of course immediate that if X1,,XN are pairwise independent (a fortiori, independent) random variables, then {X1,,XN} is PNQD.

In many stochastic models, the classical assumption of independence among the random variables in the model is not a reasonable one; the random variable may be “repelling” in the sense that small values of any of the random variables increase the probability that the other random variables are large. Thus an assumption of some type of negative dependence is often more suitable. Pemantle [11] prepared an excellent survey on a general “theory of negative dependence”.

The choice of the adjective “negative” in the definition of PNQD random variables is due to the fact that (1.2) is equivalent to

P(Xj>yXix)P(Xj>y)

provided P(Xix)>0.

A collection of N PNQD random variables arises by sampling without replacement from a set of N2 real numbers (see, e.g., Bozorgnia et al. [2]). Li et al. [7] showed that for every set of N2 continuous distribution functions {F1,,FN}, there exists a set of PNQD random variables {X1,,XN} such that the distribution function of Xj is Fj, 1jN and such that for all j{1,,N1}, Xj and Xj+1 are not independent.

An array of random variables {Xn,j,1jkn,n1} is said to be rowwise PNQD if for each n1, the set of random variables {Xn,j,1jkn} is PNQD. There is interesting literature of investigation on the strong law of large numbers problem for row sums of rowwise PNQD arrays; see the discussion in Li et al. [7].

Definition 1.2

An array of random variables {Xn,j,1jkn,n1} is said to be stochastically dominated by a random variable X if there exists a constant D such that

P(|Xn,j|>x)DP(|DX|>x),x0,1jkn,n1. 1.3

Remark 1.1

Condition (1.3) is, of course, automatic with X=X1,1 and D=1 if the array {Xn,j,1jkn,n1} consists of identically distributed random variables.

Preliminary lemmas

Three lemmas will now be stated. Lemmas 2.1, 2.2, and 2.3 are used in the proof of Theorem 3.1, Lemma 2.3 is used in the proof of Theorem 3.2, and Lemmas 2.1 and 2.2 are used in the proof of Theorem 3.3.

Lemma 2.1 follows from Lemma 1 of Lehmann [6]; see Matuła [8] for a more direct proof.

Lemma 2.1

(Lehmann [6], Matuła [8])

Let the set of random variables {X1,,XN} be PNQD, and for each j{1,,N}, let fj:RR. If the functions f1,,fN are all nondecreasing or all nonincreasing, then the set of random variables {f1(X1),,fN(XN)} is PNQD.

The next lemma is well known (see, e.g., Patterson and Taylor [10]).

Lemma 2.2

Let the set of random variables {X1,,XN} be PNQD. Then

Var(j=1NXj)j=1NVar(Xj).

The following lemma is essentially due to Adler et al. [1].

Lemma 2.3

(Adler et al. [1])

Let {Xn,j,1jkn,n1} be an array of random variables which is stochastically dominated by a random variable X, and let D be as in (1.3). Then

E(|Xn,j|I(|Xn,j|>x))D2E(|X|I(|DX|>x)),x0,1jkn,n1.

Mainstream

The main results, Theorems 3.13.3, may now be established. These are new results even under the stronger hypothesis that the random variables in each row of the array are independent.

Theorem 3.1

Let {Xn,j,1jkn,n1} be an array of rowwise PNQD mean 0 random variables which is stochastically dominated by a random variable X with E|X|<. Let {an,j,1jkn,n1} be an array of constants such that

for each n1, either min1jknan,j0ormax1jknan,j0 3.1

and

supn1j=1kn|an,j|<andlimnj=1knan,j2=0. 3.2

Then

j=1knan,jXn,jL10 3.3

and, a fortiori,

j=1knan,jXn,jP0.

Proof

Let ϵ>0 be arbitrary. Set C=supn1j=1kn|an,j|. Let D< be as in (1.3). Since E|X|<, we can choose Aϵ(0,) such that

2CD2E(|X|I(D|X|>Aϵ))ϵ2and2CDAϵP(|DX|>Aϵ)ϵ2.

Let

Yn,j=Xn,jI(|Xn,j|Aϵ)+AϵI(|Xn,j|>Aϵ)AϵI(|Xn,j|<Aϵ),1jkn,n1

and

Zn,j=Xn,jI(|Xn,j|>Aϵ)AϵI(|Xn,j|>Aϵ)+AϵI(|Xn,j|<Aϵ),1jkn,n1.

Then

Xn,j=Yn,j+Zn,jandEYn,j+EZn,j=EXn,j=0,1jkn,n1,

and so

Xn,j=Yn,jEYn,j+Zn,jEZn,j,1jkn,n1.

It follows from Lemma 2.1 that {Yn,j,1jkn,n1} is an array of rowwise PNQD random variables. Again by Lemma 2.1, (3.1) ensures that {an,jYn,j,1jkn,n1} is an array of rowwise PNQD random variables. Note that |Yn,j|Aϵ, 1jkn, n1. Thus, for n1, by Lemma 2.2

E(j=1knan,j(Yn,jEYn,j))2=Var(j=1knan,jYn,j)j=1knan,j2Var(Yn,j)j=1knan,j2EYn,j2Aϵ2j=1knan,j20

by the second half of (3.2). Thus

j=1knan,j(Yn,jEYn,j)L20

and, a fortiori,

j=1knan,j(Yn,jEYn,j)L10. 3.4

Next, for n1, by Lemma 2.3 and (1.3)

E|j=1knan,j(Zn,jEZn,j)|2Ej=1kn|an,j|E|Zn,j|2j=1kn|an,j|(E(|Xn,j|I(|Xn,j|>Aϵ))+AϵP(|Xn,j|>Aϵ))2j=1kn|an,j|(D2E(|X|I(|DX|>Aϵ))+DAϵP(|DX|>Aϵ))2CD2E(|X|I(|DX|>Aϵ))+2CDAϵP(|DX|>Aϵ)ϵ2+ϵ2=ϵ 3.5

by the choice of Aϵ.

Combining (3.4) and (3.5) yields

lim supnE|j=1knan,jXn,j|=lim supnE|j=1knan,j(Yn,jEYn,j+Zn,jEZn,j)|lim supn(E|j=1knan,j(Yn,jEYn,j)|+E|j=1knan,j(Zn,jEZn,j)|)ϵ.

Since ϵ>0 is arbitrary,

limnE|j=1knan,jXn,j|=0;

that is, (3.3) holds. □

Remark 3.1

One of the reviewers so kindly called to our attention the article by Ordóñez Cabrera and Volodin [9] and suggested that we should provide a comparison between Theorem 3.1 above and Theorem 1 of that article. Both theorems are in the same spirit in that they both establish mean convergence for weighted averages from an array of rowwise PNQD mean 0 random variables. Ordóñez Cabrera and Volodin [9] introduced the following new integrability concept for an array of random variables {Xn,j,unjkn,n1} which is weaker than several well-known integrability notions. The array of random variables is said to be h-integrable with respect to an array of constants {an,j,unjkn,n1} if

supn1j=unkn|an,j|E|Xn,j|<andlimnj=unkn|an,j|E(|Xn,j|I(|Xn,j|>h(n)))=0,

where {h(n),n1} is a sequence of constants with 0<h(n). Ordóñez Cabrera and Volodin [9] established their Theorem 1 under an h-integrability assumption for the array. Suppose that un=1, n1. It is clear that the stochastic domination condition in Theorem 3.1 is indeed a stronger condition than the array being h-integrable. However, Theorem 1 of Ordóñez Cabrera and Volodin [9] has the condition limnh2(n)j=1knan,j2=0 which is stronger than the condition limnj=1knan,j2=0 in (3.2) of Theorem 3.1. Consequently, the two theorems being compared overlap with each other but neither theorem is contained in the other.

The next theorem is a version of Theorem 3.1 without assumption (3.1) for an array of random variables where the random variables in each row of the array are pairwise independent (which is a stronger assumption than the array being rowwise PNQD).

Theorem 3.2

Let {Xn,j,1jkn,n1} be an array of mean 0 random variables such that, for each n1, the random variables Xn,j,1jkn are pairwise independent. Suppose that the array {Xn,j,1jkn,n1} is stochastically dominated by a random variable X with E|X|<. Let {an,j,1jkn,n1} be an array of constants such that

supn1j=1kn|an,j|<andlimnj=1knan,j2=0.

Then

j=1knan,jXn,jL10

and, a fortiori,

j=1knan,jXn,jP0.

Proof

Let ϵ>0 be arbitrary, and let C, D, Aϵ, Yn,j, and Zn,j, 1jkn, n1 be as in the proof of Theorem 3.1. The pairwise independence assumption ensures that

Var(j=1knan,jYn,j)=j=1knan,j2Var(Yn,j),n1,

and (3.4) follows arguing as in the proof of Theorem 3.1. Moreover, (3.5) holds by the same argument as in the proof of Theorem 3.1. The rest of the proof is identical to that in Theorem 3.1. □

Remark 3.2

The cited result of Chandra [3] follows immediately from Theorem 3.2 by taking kn=n, n1 and Xn,j=Xj, 1jn, n1.

Remark 3.3

If the rowwise PNQD hypothesis in Theorem 3.1 is dispensed with, then the theorem can fail. To see this, let X be a nondegenerate mean 0 random variable, let kn=n, n1, and let

Xn,j=Xandan,j=n1,1jn,n1.

Then {Xn,j,1jn,n1} is not an array of PNQD random variables, (3.1) and (3.2) hold, but

graphic file with name M241.gif

This same example shows that Theorem 3.2 can fail without the pairwise independent hypothesis.

We now show via an example that the hypotheses to Theorems 3.1 and 3.2 do not necessarily ensure that j=1knan,jXn,j0 almost surely (a.s.).

Example 3.1

Let {Xn,n1} be a sequence of i.i.d. random variables with EX1=0 and E|X1|p= for some p>1. Set kn=n, n1, Xn,j=Xj, 1jkn, n1, and

a1,1=1,an,j={0,1jn,n1/p,j=n,n2.

Then (3.2) holds since

supn1j=1kn|an,j|=supn1n1/p=1<

and

limnj=1knan,j2=limnn2/p=0.

All of the hypotheses of Theorems 3.1 and 3.2 are satisfied and hence (3.3) holds.

Note that {j=1knan,jXn,j=n1/pXn,n1} is a sequence of independent random variables. Now, for arbitrary M1,

n=1P(|j=1knan,jXn,j|>M)=n=1P(|n1/pXn|>M)=n=1P(|X1|M>n1/p)=

since E|X1|p=. Then by the second Borel–Cantelli lemma,

P(|j=1knan,jXn,j|>M i.o. (n))=1,

and so

P(lim supn|j=1knan,jXn,j|=)=P(M=1{lim supn|j=1knan,jXn,j|M})P(M=1{|j=1knan,jXn,j|>M i.o. (n)})=1.

Thus,

lim supn|j=1knan,jXn,j|=a.s.

and so j=1knan,jXn,j0 a.s. fails.

We now establish Theorem 3.3. Throughout the rest of this section, for an array of random variables {Xn,j,1jkn,n1}, let Sn=j=1knXn,j, n1.

Theorem 3.3

Let {Xn,j,1jkn,n1} be an array of rowwise PNQD L1 random variables. Let g:[0,)[0,) be a continuous function with

g(0)=0andg2(v)vas 0<v.

Let {bn,n1} be a sequence of positive constants with bn and suppose that there exists a sequence of positive constants {αn,n1} such that

g(v)g(bn)αn(vbn)for all v>bn and n1. 3.6

Set

Vn,j=g1(|Xn,j|),1jkn,n1

and assume that EVn,j<, 1jkn, n1. Let {dn,n1} be a sequence of positive constants and suppose for some sequence of positive constants {cn,n1} with cn<bn, n1 that

αndnj=1knE((Vn,jbn)I(Vn,j>bn))0, 3.7
g2(bn)dn2bnj=1knE(Vn,jI(Vn,j>cn))0, 3.8
g2(bn)dn2bnj=1knEVn,j=O(1), 3.9
g2(bn)dn2j=1knP(Vn,j>bn)0, 3.10

and

g2(cn)cn=o(g2(bn)bn). 3.11

Then

SnESndnL10 3.12

and, a fortiori,

SnESndnP0.

Proof

For 1jkn and n1, set

Yn,j=Xn,jI(|Xn,j|g(bn))+g(bn)I(Xn,j>g(bn))g(bn)I(Xn,j<g(bn))

and

Zn,j=(Xn,jg(bn))I(Xn,j>g(bn))+(Xn,j+g(bn))I(Xn,j<g(bn)).

Then Yn,j+Zn,j=Xn,j, 1jkn, n1. Set Tn=j=1knYn,j, n1. We will show that

j=1knE|Zn,j|dn0 3.13

and

TnETndnL10. 3.14

To prove (3.13), note that for 1jkn and n1,

|Zn,j|=(Xn,jg(bn))I(Xn,j>g(bn))+(Xn,jg(bn))I(Xn,j<g(bn))=(|Xn,j|g(bn))I(Xn,j>g(bn))+(|Xn,j|g(bn))I(Xn,j<g(bn))=(|Xn,j|g(bn))I(|Xn,j|>g(bn))=(g(Vn,j)g(bn))I(g(Vn,j)>g(bn))=(g(Vn,j)g(bn))I(Vn,j>bn)αn(Vn,jbn)I(Vn,j>bn)(by (3.6)),

and hence

1dnj=1knE|Zn,j|αndnj=1knE((Vn,jbn)I(Vn,j>bn))0

by (3.7) proving (3.13).

To prove (3.14), note that for 1jkn and n1,

Yn,j2=Xn,j2I(|Xn,j|g(bn))+g2(bn)I(|Xn,j|>g(bn))=g2(Vn,j)I(g(Vn,j)g(bn))+g2(bn)I(g(Vn,j)>g(bn))=g2(Vn,j)I(Vn,jbn)+g2(bn)I(Vn,j>bn)=g2(Vn,j)Vn,jVn,jI(0<Vn,jcn)+g2(Vn,j)Vn,jVn,jI(cn<Vn,jbn)+g2(bn)I(Vn,j>bn).

Then for n1, since the set of random variables {Yn,j,1jkn} is PNQD by Lemma 2.1,

E(TnETndn)21dn2j=1knVar(Yn,j)(by Lemma2.2)1dn2j=1knEYn,j2=1dn2(j=1knE(g2(Vn,j)Vn,jVn,jI(0<Vn,jcn))+j=1knE(g2(Vn,j)Vn,jVn,jI(cn<Vn,jbn))+j=1knE(g2(bn)I(Vn,j>bn)))1dn2(g2(cn)cnj=1knEVn,j+g2(bn)bnj=1knE(Vn,jI(Vn,j>cn))+g2(bn)j=1knP(Vn,j>bn))=1dn2(g2(bn)bno(1)j=1knEVn,j+g2(bn)bnj=1knE(Vn,jI(Vn,j>cn))+g2(bn)j=1knP(Vn,j>bn))(by (3.11))=g2(bn)dn2bn(j=1knEVn,j)o(1)+g2(bn)dn2bnj=1knE(Vn,jI(Vn,j>cn))+g2(bn)dn2j=1knP(Vn,j>bn)=o(1)(by (3.9),(3.8), and (3.10)).

Thus

TnETndnL20

and hence (3.14) holds.

Finally, note that for n1,

SnESndn=j=1knYn,j+j=1knZn,jj=1knEYn,jj=1knEZn,jdn=j=1knZn,jj=1knEZn,jdn+TnETndn. 3.15

Now it follows from (3.13) that

E|j=1knZn,jj=1knEZn,jdn|2j=1knE|Zn,j|dn0. 3.16

The conclusion (3.12) follows from (3.15), (3.16), and (3.14). □

Corollary 3.1

Let {Xn,j,1jkn,n1} be a uniformly bounded array of rowwise PNQD random variables. Let {bn,n1} be a sequence of constants with 1<bn. Then

SnESnknbnL10. 3.17

Proof

Let dn=knbn, n1 and cn=bn, n1. Let g(v)=v, v0 and αn=1, n1. Set

Vn,j=g1(|Xn,j|)=|Xn,j|,1jkn,n1.

Since the array is comprised of uniformly bounded random variables, conditions (3.7), (3.8), (3.9), and (3.10) hold. Moreover, (3.11) also holds since cn=o(bn). The conclusion (3.17) follows from Theorem 3.3. □

Remark 3.4

If the rowwise PNQD hypothesis in Theorem 3.3 or Corollary 3.1 is dispensed with, then those results can fail. To see this, let {kn,n1} be a sequence of integers with 1<kn, let X be a bounded nondegenerate random variable, and set Xn,j=X, 1jkn, n1. Let bn=kn, n1. All of the conditions of Corollary 3.1 (hence of Theorem 3.3) are satisfied except for the rowwise PNQD hypothesis. The conclusions of Corollary 3.1 (hence of Theorem 3.3) fail since

graphic file with name M427.gif

We now show via an example that the hypotheses of Corollary 3.1 (hence of Theorem 3.3) do not necessarily ensure that

SnESnknbn0a.s. 3.18

Example 3.2

Let {Xn,n1} be a sequence of nondegenerate i.i.d. uniformly bounded random variables, and let

kn=n,bn=loglog(max{16,n}),dn=knbn,n1.

Let Xn,j=Xj, 1jn, n1. The hypotheses of Corollary 3.1 (hence of Theorem 3.3) are satisfied and so (3.12) and (3.17) hold. But by the Hartman and Wintner [5] law of the iterated logarithm,

lim supnSnESnknbn=2Var(X1)a.s.

and so (3.18) does not hold.

Corollary 3.2

Let {Xn,j,1jn,n1} be an array of identically distributed rowwise PNQD L1 random variables, and let {bn,n1} be a sequence of constants with 1<bn. If

nbnE((|X1,1|bn)I(|X1,1|>bn))0, 3.19

then

SnESnnbnL10. 3.20

Proof

We will apply Theorem 3.3 with g(v)=v, 0v< and

kn=n,αn=1,cn=bn,anddn=nbn,n1.

Then cn<bn, n1 and (3.6) and (3.11) are immediate. Condition (3.7) is the same as (3.19). Since E|X1,1|<, conditions (3.8) and (3.9) are immediate. Condition (3.10) reduces to

bnP(|X1,1|>bn)0

which holds since E|X1,1|<. The conclusion (3.20) follows from Theorem 3.3. □

Conclusions

For an array of rowwise PNQD random variables {Xn,j,1jkn,n1}, conditions are provided under which the following degenerate mean convergence laws hold:

  • (i)
    j=1knan,jXn,jL10,
    where EXn,j=0, 1jkn, n1, and {an,j,1jkn,n1} is an array of constants;
  • (ii)
    j=1kn(Xn,jEXn,j)dnL10,
    where {dn,n1} is a sequence of positive constants.

A version of the result in (i) is also obtained for an array of rowwise pairwise independent random variables and this result extends the result of Chandra [3]. Examples are provided showing that the above results can fail if the hypotheses are weakened and that a.s. convergence does not necessarily hold together with the L1 convergence.

Acknowledgments

Acknowledgements

The authors are grateful to the reviewers for carefully reading the manuscript and for offering helpful comments and suggestions.

Authors’ information

Tapas K. Chandra deceased before publication of this work was completed.

Authors’ contributions

All authors contributed equally and significantly in writing this article. All the authors read and approved the final manuscript.

Funding

The research of Deli Li was partially supported by a grant from the Natural Sciences and Engineering Research Council of Canada (Grant #: RGPIN-2014-05428).

Competing interests

The authors declare that they have no competing interests.

Footnotes

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Deli Li, Email: dli@lakeheadu.ca.

Andrew Rosalsky, Email: rosalsky@stat.ufl.edu.

References

  • 1.Adler A., Rosalsky A., Taylor R.L. Strong laws of large numbers for weighted sums of random elements in normed linear spaces. Int. J. Math. Math. Sci. 1989;12:507–529. doi: 10.1155/S0161171289000657. [DOI] [Google Scholar]
  • 2.Bozorgnia A., Patterson R.F., Taylor R.L. Limit theorems for dependent random variables; Proceedings of the First World Congress of Nonlinear Analysts (WCNA’92); Berlin: de Gruyter; 1996. pp. 1639–1650. [Google Scholar]
  • 3.Chandra T.K. On extensions of a result of S. W. Dharmadhikari. Bull. Calcutta Math. Soc. 1990;82:431–434. [Google Scholar]
  • 4.Dharmadhikari S.W. A simple proof of mean convergence in the law of large numbers. Am. Math. Mon. 1976;83:474–475. doi: 10.1080/00029890.1976.11994148. [DOI] [Google Scholar]
  • 5.Hartman P., Wintner A. On the law of the iterated logarithm. Am. J. Math. 1941;63:169–176. [Google Scholar]
  • 6.Lehmann E.L. Some concepts of dependence. Ann. Math. Stat. 1966;37:1137–1153. doi: 10.1214/aoms/1177699260. [DOI] [Google Scholar]
  • 7.Li D., Rosalsky A., Volodin A.I. On the strong law of large numbers for sequences of pairwise negative quadrant dependent random variables. Bull. Inst. Math. Acad. Sin. (N.S.) 2006;1:281–305. [Google Scholar]
  • 8.Matuła P. A note on almost sure convergence of sums of negatively dependent random variables. Stat. Probab. Lett. 1992;15:209–213. doi: 10.1016/0167-7152(92)90191-7. [DOI] [Google Scholar]
  • 9.Ordóñez Cabrera M., Volodin A.I. Mean convergence theorems and weak laws of large numbers for weighted sums of random variables under a condition of weighted integrability. J. Math. Anal. Appl. 2005;305:644–658. doi: 10.1016/j.jmaa.2004.12.025. [DOI] [Google Scholar]
  • 10.Patterson R.F., Taylor R.L. Strong laws of large numbers for negatively dependent random elements. Nonlinear Anal., Theory Methods Appl. 1997;30(7):4229–4235. doi: 10.1016/S0362-546X(97)00279-4. [DOI] [Google Scholar]
  • 11.Pemantle R. Towards a theory of negative dependence. J. Math. Phys. 2000;41:1371–1390. doi: 10.1063/1.533200. [DOI] [Google Scholar]
  • 12.Pyke R., Root D. On convergence in r-mean of normalized partial sums. Ann. Math. Stat. 1968;39:379–381. doi: 10.1214/aoms/1177698400. [DOI] [Google Scholar]

Articles from Journal of Inequalities and Applications are provided here courtesy of Springer

RESOURCES