Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2013 Mar 26.
Published in final edited form as: J Stat Plan Inference. 2012 Nov;142(11):2965–2975. doi: 10.1016/j.jspi.2012.04.017

Two-way model with random cell sizes

Steven F Arnold a, Panagis G Moschopoulos b,*
PMCID: PMC3608410  NIHMSID: NIHMS442817  PMID: 23538487

Abstract

We consider inference for row effects in the presence of possible interactions in a two-way fixed effects model when the numbers of observations are themselves random variables. Let Nij be the number of observations in the (i, j) cell, πij be the probability that a particular observation is in that cell and μij be the expected value of an observation in that cell. We assume that the {Nij} have a joint multinomial distribution with parameters n and {πij}. Then μ̄i. = Σjπijμijjπij is the expected value of a randomly chosen observation in the ith row. Hence, we consider testing that the μ̄i. are equal. With the {πij} unknown, there is no obvious sum of squares and F-ratio computed by the widely available statistical packages for testing this hypothesis. Let Ȳi.. be the sample mean of the observations in the ith row. We show that Ȳi.. is an MLE of μ̄i., is consistent and is conditionally unbiased. We then find the asymptotic joint distribution of the Ȳi.. and use it to construct a sensible asymptotic size α test of the equality of the μ̄i. and asymptotic simultaneous (1 − α) confidence intervals for contrasts in the μ̄i..

Keywords: Two-way model, Main effects, Analysis of variance, Unbalanced data, Multinomial cell sizes

1. Introduction

The two-way model with fixed effects for unbalanced data has received considerable attention in the past 30 years. There has been much discussion of the definition of the effects, calculation of sums of squares and definitions of hypotheses. The model is the subject of many texts, prominent among those Searle (1971, 1982), Graybill (1976), Scheffè (1959), and Arnold (1981). Several decompositions of sums of squares are presented in ANOVA tables, and testing the parameters of the model is routinely accomplished by subtracting sums of squares to isolate the contribution of a certain parameter. Searle’s (1971) R-notation, for example R(α/μ, β) is usually involved in testing the contribution of including α after μ and β in the model. The R-notation implies orthogonalizations of the design matrix that lead to sums of squares that test a host of hypotheses that depend on the cell sizes. Many hypotheses have interpretation problems and several articles have been written on the hypotheses implied by using Searle’s R(·/·) notation, e.g. see, Speed et al. (1978), Kutner (1974); for an extensive list see Macnaugton (1998).

This paper considers the same two-way model under the assumption that the overall sample size n is fixed but the cell sizes are themselves random variables following the multinomial distribution with unknown parameters πij. The setting has been introduced in Moschopoulos and Davidson (1985). Applications of the new setting arise mainly in survey studies. In these studies it is often not possible to sample from the individual cells. Instead, a sample of size n is taken from the overall population and is divided after the fact into the two-way cells according to sample characteristics. The following two examples will help in understanding this setting:

Example 1

A random sample size n of women is drawn from a metropolitan area for the purpose of studying fertility, measured by the number of children born to women (the response y-variable). It is assumed that fertility is dependent on religion, the row factor (Catholic (1) or Protestant (2)) and educational level, the column factor (Elementary (1), High School (2), College (3)). Sampling 300 women from the overall population results in 2 × 3 = 6 cell sizes nij, i = 1, 2; j = 1, 2, 3 that are random variables following the multinomial distribution (for large n). Note that sampling from each of the six classifications is not possible since religion and education are only known after the sample is drawn. In this example, let i represent the woman’s religion and j represent the woman’s educational level. In this situation, π21 represents the proportion of women who are Protestant with Elementary education; μ12 represents the expected number of children in Catholic women with High School education; μ¯2.π represents the expected number of children by Protestant women averaged across educational levels. In testing that the μ¯i.π, i = 1, 2 are all equal in this setting, we are testing the equality of the mean number of children for the two religions averaged over educational levels. Testing for no column effects with these weights is testing the equality between educational levels averaged over religion. For a real situation related to this, see Groat and Neal (1967).

Example 2

Consider an academic class size n viewed as a random sample from a population of similar students with the response y being the student performance during an examination. The purpose of the investigation here is to relate class performance to student anxiety, the row factor (High (1), Moderate (2), Low (3)) and student attitude towards the subject the column factor (Positive (1), Negative (2)). Note that sampling from each of the six classifications here is not possible as the student classifications are known only after the sample is drawn. Thus, the cell sizes are random following the multinomial distribution (for large n). As in the first example, here testing the equality of μ¯i.π, i = 1, 2, 3 is testing the equality of performances of High, Moderate and Low anxiety students averaged over student attitudes. For a real situation on this, see Galassi et al. (1981).

Random cell sizes but from a different perspective are considered in Psychology, consideration there given to random cell sizes arising as a result of the underlying treatment, see Weiss (1999, 2006).

As stated above, the two-way model with fixed effects and unbalanced data is not free of controversies. The main problems are the following: (1) the definition of main effects, i.e row, column and interaction effects, and (2) hypotheses tested by typical ANOVA decompositions and interpretations of such hypotheses. Concerning (1) above, as it turns out, in the case of the two-way model with multinomially distributed cell sizes we can provide a natural definition of main effects. This definition is not arbitrary but is a natural consequence of the sampling scheme and it is expressed in terms of parameters (i.e cell population proportions and the cell population means), see the next section. As for (2) above, we are concerned here with one particular hypothesis, namely the equality of row (column) means. Let μij, i = 1, …, r; j = 1, …, c be the mean of cell (i,j), and let πij be the probability of obtaining an observation in cell (i,j). We show that the i-th row mean is

μi=j=1cπijπi.μij, (1.1)

where πi.=j=1cπij. The hypothesis of equality of row means is

Ho:μiequal,i=1,,r. (1.2)

The hypothesis (1.2) above is equivalent to the hypothesis that the α-main effects (defined in Section 2) are zero. This hypothesis is well known in the literature but the means are defined with cell size weights, i.e. (see Searle, 1971)

μi=j=1cNijNi.μij. (1.3)

Obviously, when the cell sizes are random variables, testing the equality of means in (1.3) is not proper. Testing of the right hypothesis (1.2) is the second goal of this paper. However, there is no test derived from standard ANOVA sums of squares decompositions that is proper for testing (1.2). We develop an asymptotic (n → ∞) test for (1.2) and asymptotic 100 × (1 − α)% confidence intervals for contrasts of row means. In addition, we provide some numerical evaluations of the performance of the proposed test under the null hypothesis. Under the multinomial assumption for the cell sizes, it is possible that a cell size may be zero. The paper is rigorously considering the case of zero cells.

2. The two-way model with random cell sizes

The two-way model is often modeled in the following over-parameterized form that includes terms for main effects and interactions. We observe Yijk independent,

Yijk~N(θ+αi+βj+γij,σ2),i=1,,r;j=1,,c;k=1,,Nij, (2.1)

where θ, {αi}, {βj}, {γij} and σ2 are unknown parameters. In order to make these parameters identifiable, weights wij ≥ 0 are often chosen and it is assumed in addition that

ijwijαi=0,ijwijβj=0,iwijγij=0,jwijγij=0. (2.2)

It is well known that the problem of testing that the interactions γij = 0 does not depend on the weights, nor does the problem of testing that the main effects αi = 0 when it is assumed that the γij = 0 (e.g. see Arnold, 1981, pp. 93–96). Unfortunately, however, the problem of testing that the αi = 0 when the γij may be non-zero depends on the weights chosen. Different weights lead to different hypotheses. Let

μij=θ+αi+βj+γij,μ¯iW=jwijμij/jwij. (2.3)

Then testing that αi = 0 with the weights wij is the same as testing the equality of the μ¯iW. Therefore, the weights wij represent the relative importance of the observations in the ith row, jth cell out of all the observations in the ith row.

Now, suppose that the individuals observed are themselves a sample of size n( = Σi ΣjNij) from a very large population. In that case, we may assume that the cell sizes {Nij} have a joint multinomial distribution with parameters n, and {πij} written as

{Nij}~Mrc(n,{πij}),

where πij is the probability that a randomly chosen individual is in the ith row and the jth column. We assume that the πij are unknown parameters for this model. In this case, naturally, we use the weights

wij=πij. (2.4)

Using these weights, we see that the conditional expectation of an observation in row i is

E(Yijkrowi)=j=1cPr(columnjrowi)E(Yijkrowi,columnj)=j=1cπijπi.μij=μ¯i (2.5)

since

P(columnjrowi)=ηij=πij/jπij=πijπi. (2.6)

Therefore, in testing the equality of the μ̄i. we are testing that the expected value of the response variable is the same in all the rows.

Note

The expectation in (2.5) is approximate if some cells are empty in row i; the case of empty cells is treated in this paper, see Basic results. It is now interesting to note that if the artificial parameters θ, αi, βj and γij above are defined using the weights wij = πij, then

θ=ijπijμij,αi=μ¯i.-θ,βj=μ¯.j-θ,γij=μij-μ¯i.-μ¯.j+θ. (2.7)

3. Basic results

For simplicity we consider here the cell means μij instead of the over-parameterized model. Let Yijk be independent,

Yijk~N(μij,σ2),i=1,,r;j=1,,c;k=1,,Nij,

where

{Nij}~Mrc(n,{πij},)

i.e., the {Nij} are (jointly) multinomially distributed with parameters n and {πij}. We assume that

πij>0foralliandj

and that the πij are unknown and must be estimated from the data, but that n is known and fixed in advance. It is important to remember that the Nij are random variables, and may in fact be 0.

Let q be the number of cells (i,j) in which Nij > 0. (Note that q is a random variable dependent on the Nij). Let

Y¯ij.=kYijk/NijifNij>0,Y¯ij.=0ifNij=0, (3.1)
S2=ijk(Yijk-Y¯ij.)2/(n-q). (3.2)

Then it is easily shown that the sets {Ȳij.}, {Nij} together with S2 form a sufficient statistic for this model and that Ȳij., Nij/n and (nq)S2/n are the MLE’s of μij, πij and σ2 respectively. (Note that the MLE for μij is only unique if Nij > 0.) Furthermore, Nij/n is an unbiased estimator of πij. Unbiased estimators for the μij do not appear possible, due to the possibility that Nij = 0. Finally, note that conditionally on q,

(n-q)S2/σ2q~χn-q2 (3.3)

so that S2 is consistent and unbiased.

Now, with μ̄i. as in (2.5) and πi. and ηij as in (2.6), let Ni = ΣjNij and

π^i.=Ni/n,η^ij=Nij/NiifNi>0andη^ij=0ifNi=0

be MLE’s of πi. and ηij respectively. Further, let

Y¯i..=jη^ijY¯ij..

Note that Ȳi.. is the MLE for μ̄i. by the invariance principle, and is unique as long as Ni > 0. Note also that as long as Ni > 0,

Y¯i..=jkYijk/Ni. (3.4)

Finally, note that Ȳi.. has a point mass at 0, so is not a continuous random variable.

P(Y¯i..=0)=P(Ni=0)=(1-πi.)n. (3.5)

The Distribution of Ȳi

We are primarily interested in inference about the μ̄i. In particular, we are interested in testing that the μ̄i. are all equal. In order to draw inference about the μ̄i., we need to learn about the distribution of Ȳi... For notational convenience we let

ηi=(ηi1,,ηic),η^i=(η^i1,,η^ic),mi=(μi1,,μic).

The following random variables are useful in our development here and represent row-means weighted by cell sizes. For i = 1, …, r, let

μi=jNijNiμij=jη^ijμij=miη^i. (3.6)

We note here that testing for main effects in the presence of interaction and using the usual conditional F-test entails testing the hypothesis that the μ̃i’s are equal, see Searle, 1971, pp. 292–231, Arnold, 1981, pp. 93–96. Obviously this hypothesis makes no sense especially if the cell sizes are random variables. Define

eijk=Yijk-μij,e¯i..=jkeijk/NiifNi>0,e¯i..=0ifNi=0. (3.7)

If Ni > 0, then

Y¯i..-μi=jk(Yijk-μij)/Ni=e¯i... (3.8)

When Ni = 0,

Y¯i..=μi=e¯i..=0.

Therefore, for all Ni,

Y¯i..=μi+e¯i..=e¯i..+miη^i. (3.9)

Lemma 1

Conditionally on Ni, ēi.. and η̂i are independent.

Proof

If Ni = 0, then both ēi.. and η̂i are degenerate and hence independent. If Ni > 0, then ēi.. is the average of all Ni of the errors in the ith row, irrespective of which cell they are in. Since the errors are independently identically distributed, ēi.. does not depend on the relative proportions in the columns, and hence on η̂i.

This result implies that conditionally on N⃗(N1, N2, …, Nr), the Ȳi.., i = 1, …, r are not normally distributed; by (3.9), Ȳi.. is the sum of two independent random variables and if it were normal then both components should be normal, which obviously is not the case. This result is of importance in our development for the following reasons: The definition of μ̄i. depends only on μij and ηij and the distribution of N⃗ depends only on the πi.. Therefore, N⃗ is an ancillary statistic for this problem and it makes sense to work conditionally on N⃗. Note also that the Nij are not ancillary for the μ̄i., so that we should not work conditionally on the Nij.

Comment

Working conditionally on N⃗ = (N1, N2, …, Nr), is of course equivalent to making inferences about r means of non-normal sub-populations with the overall population being a mixture of these r sub-populations, each with mixing probability πi. This will lead to an approximate test statistic for the problem for large n and essentially reduces the two-way problem to a ‘one-way problem’. However, preliminary numerical evaluations showed that the ‘heuristic’ one-way ANOVA test of equality of row means in this case is totally inappropriate, because the distribution of the row-sample means is NOT normal. Obviously, any test statistic for testing the hypothesis that the μi.’s in (2.5) are equal would have to rely on the distribution of the row-sample mean. We first study the conditional distribution of the Ȳi.., given N⃗.

Theorem 1

  1. In the conditional distribution given N⃗, the Ȳi.., i = 1, …, r are independent.

  2. If Ni = 0, then E(Ȳi.. | N⃗) = Var(Ȳi.. | N⃗) = 0.

  3. If Ni > 0, then
    E(Y¯i..N)=μ¯i.,Var(Y¯i..N)=δiσ2/Ni,
    where
    δi=1+jηij(μij-μ¯i.)2/σ2. (3.10)

Proof

  1. Conditionally on the N⃗, the ēi.. are independent as are the η̂i, i = 1, …, r. Hence, using the lemma, the Y¯i..=e¯i..+miη^i are also independent.

  2. When Ni = 0, then Ȳi.. is degenerate at 0, so these results follow.

  3. If Ni > 0, then Niη^i conditional on N⃗ has the following multinomial distribution:
    Niη^N~Mc(Ni,ηi). (3.11)

Now,

E(e¯i..N)=0,E(μiN)=miE(η^iN)=miηi=μ¯i. (3.12)

and therefore,

E(Y¯i..N)=E(e¯i..N)+E(μiN)=0+μ¯i. (3.13)

To see the formula for the variance, note that

Var(e¯i..N)=σ2/Ni

and using the variance–covariance matrix of the multinomial proportions

η^i=(η^i1,,η^ic)

we obtain

Var(μiN)=mi(Cov(η^iN))mi=miVimi/Ni,

where Vi has (j, g) element Vijk given by

Vijj=ηij(1-ηij),Vijg=-ηijηig,jg. (3.14)

Since ēi.. and μ̃i are independent (conditionally on N⃗)

Var(Y¯i..N)=Var(e¯i..N)+Var(μiN)=(σ2/Ni)(1+miVimi/σ2)=σ2δi/Ni

and the theorem is proved.

Note from part (c) that Ȳi.. is a conditionally unbiased estimator of μ̄i > 0 but is not unbiased. In fact,

EY¯i..=μ¯i.P(Ni>0)=μ¯i.(1-(1-πi.)n).

The asymptotic distribution of Ȳi... is given in the following.

Theorem 2

  1. Ȳi.. is a consistent estimator of μ̄i

  2. Ni1/2(Y¯i..-μ¯i.)NdUi~N(0,δiσ2) as Ni → ∞.

Proof

(a) Note first that πi. > 0 so that

P(Ni>0)=1-(1-πi.)n1asn.

Therefore,

EY¯i..=E(EY¯i..N)=μ¯i.P(Ni>0)μ¯i.asn.

To show consistency we need to show that Var(Ȳi..) → 0. Recall that

Var(Y¯i..)=Var(E(Y¯i..N))+E(Var(Y¯i..N)). (3.15)

Now,

Var(E(Y¯i..N))=μ¯i.2(1-(P(Ni>0))2)μ¯i.2(1-12)=0.

Finally,

E(Var(Y¯i..N))=E(δiσ21NiP(Ni>0))=P(Ni>0)δiσ2m=1nm-1P(Ni=m)=P(Ni>0)δiσ2m=1nm-1(nm)πi.m(1-πi.)n-m=P(Ni>0)δiσ21(n+1)πi.m=1n(1+1m)(n+1m+1)πi.m+1(1-πi.)n-mP(Ni>0)δiσ22(n+1)πi.0asn.

Therefore, the Var(Ȳi..) and the bias of Ȳi.. converge to 0 and Ȳi.. is consistent. (b). By the asymptotic approximation of the multinomial proportions,

Ni1/2(η^i-ηi)NdWi~Nc(0,Vi),

where Vi is given in (3.14). Therefore,

Ni1/2(μi-μ¯i.)N=Ni1/2mi(η^i-ηi)NdN(0,miVimi).

Now, ēi.. is the average of all the errors in the ith row, so that

Ni1/2e¯i..N~N(0,σ2).

By (3.9) and the lemma above, conditionally on N, Ni1/2(μi-μ¯i) and N1/2ēi.. are independent. Therefore,

Ni1/2(Y¯i..-μ¯i.)N=(Ni1/2e¯i..+Ni1/2(μi-μ¯i.))NdN(0,σ2+miVimi).

However,

σ2+miVimi=σ2δi

and this proves the theorem.

Corollary

Qn=n1/2(Y¯1..-μ¯1.,,Y¯r..-μ¯r.)dQ~Nr(0,σ2T), as n → ∞, where T is a diagonal matrix whose ith element is δi/πi..

Proof

Note first that π̂i. = Ni/nπi. a.s. (almost surely) Therefore, Nia.s, so that we can use the previous theorem. From that theorem and the conditional independence of the Ȳi..,

Pn=(N11/2(Y¯1..-μ¯1.),,Nr1/2(Y¯r..-μ¯r.))NdP~Nr(0,σ2M),

where M is a diagonal matrix whose i-th diagonal element is δi. Since this limiting distribution does not depend on N⃗, unconditionally we have

PndP.

Now, π̂i is a consistent estimator of πi.. Therefore, by Slutsky’s theorem,

Qn=(π^1.-1/2Pn1,,π^r.-1/2Pnr)d(π1.-1/2P1,,πr.-1/2Pr)~N(0,T).

Now consider testing the hypothesis

Ho:μ¯i.equalforalli=1,,r. (3.16)

If the normal distribution in the corollary were exact and the δi/πi. were known, then the model would be a generalized linear model, and this hypothesis would be tested using the following (see (3.3) and recall that q is the number of non-empty cells):

F=niπi.(Y¯i..-Y)2/δi(r-1)S2, (3.17)

where

Y=i(πi.Y¯i../δi)/i(πi./δi)

and under the null hypothesis

F~Fr-1,n-q. (3.18)

In addition, Scheffe’ simultaneous confidence intervals for the set of contrasts between the μ̄i. (functions Σciμ̄i., Σci = 0) are given by

iciμ¯i.iciY¯i..±S[(r-1)Fα]1/2[ci2δi/nπi]1/2, (3.19)

where Fα is the upper 100(1 − α)% point of the F-distribution defined in (3.18). Of course, the δi and πi. would not be known in practice, so we estimate them in the obvious way. Let

π^i.=Ni/n,δ^i=1+jη^ij(Y¯ij.-Y¯i..)2/S2.

(By the invariance principle, π̂i. and δ̂i are the MLE’s of πi. and δi). We suggest using the test statistic:

F^n=nπ^i.(Y¯i..-Y^)2/δ^i(r-1)S2=iNi(Y¯i..-Y^)2/δ^i(r-1)S2, (3.20)

where

Y^=i(π^iY¯i../δ^i)i(π^i./δ^i). (3.21)

Since π̂i. and δ̂i are consistent (see the proof below), n should have an approximate F-distribution Fr − 1,n q under the null hypothesis. Similarly, we suggest using the following approximate confidence intervals for contrasts in the μ̄i.:

ciμ¯i.ciY¯i..±S[(r-1)Fα]1/2[ci2δ^i/Ni]1/2. (3.22)

We finish this section by showing that the test and confidence intervals given above are at least asymptotically correct.

Theorem 3

  1. Let n be defined in Eq. (3.20). Then under the null hypothesis that μi.., i = 1, …, r are equal,
    (r-1)F^ndH~χr-12,
    and hence the test which rejects the null hypothesis when
    FnFr-1,n-qα

    is an asymptotic size α test.

  2. The confidence intervals given in Eq. (3.19) are asymptotic 100(1 − α)% simultaneous confidence intervals for the set of contrasts given above.

Proof

  1. Note first that π̂ij = Nij/n are known to be consistent estimators of the πij so that the π̂i. and η̂ij are consistent estimators of the π̂i. and ηij. We have shown that the Ȳi.. and Sn2 are consistent estimators of μi. and σ2. Therefore, the δ̂i is consistent estimator of δi. Let Q⃗n and Q⃗ be defined as in the corollary above. Let
    π=(π1.,πr.),π^n=(π^1.,,π^r.),δ=(δ1,,δr),δ^n=(δ^1,,δ^r),h(Q,π,δ,σ2)=iπi.(Qi-Q)2/δiσ2,Q=i(πi.Qi/δi)/i(πi./δi).
    By the usual results on the generalized linear model,
    H=h(Q,π,δ,σ2)~χr-12.
    Therefore, by Slutsky’s theorem,
    h(Qn,π^n,δ^n,Sn2)dH~χr-1.2.
    If the μ̄i. are all equal, then
    (r-1)F^n=h(Qn,π^n,δ^n,Sn2)dH~χr-12.
    Now, qrc, so that nq→ ∞ and hence (r-1)Fr-1,n-qαχr-1.2. Therefore, the probability of rejecting under the null hypothesis is
    p(h(Qn,π^n,δ^n,Sn2)-(r-1)Fr-1,n-qα>0)P(H-χr-1,α2>0)=α.
  2. The simultaneous confidence intervals given in Eq. (3.22) are all satisfied if and only if
    h(Qn,π^n,δ^nSn2)(r-1)Fr-1,n-q.α.

    The remainder of the argument is similar to part (a).

Some additional comments

  1. For ease in defining the model, we have assumed that the πij > 0. A careful reading of the paper, however, will show that we have only used πi. > 0. If πi. = 0, we can just drop the ith row from the design, so that this is really no assumption.

  2. Again, for simplicity, we have assumed that the Yijk are normally distributed. However, the results are true under greater generality. In particular, the proof of Theorem 1 has only used that
    EYijk=μij,var(Yijk)=σ2.
    For Theorems 2 and 3, all we need is that
    Yijk=μij+eijk,
    where eijk are independently identically distributed with mean 0 and with finite variance σ2. We have only used the normality to establish that
    Ni1/2e¯i..dR~N(0,σ2),

    which can be established under these more general conditions using the central limit theorem.

  3. For simplicity, we have assumed that we were sampling from a very large population so that the distribution of the Nij could be assumed multinomial. If we sample from a smaller population we should replace the multinomial distribution with a hypergeometric distribution. Theorem 1 follows equally well for a multivariate hypergeometric (with population size m) if we replace the covariance matrix V for the multinomial with the covariance for the hypergeometric, getting
    δi=1+(m-n)ηij(μij-μ¯i.)2/(m-1)σ2.

    Theorem 2 can be similarly modified.

  4. As a final generalization, we mention a different sampling scheme. Suppose instead of getting a sample of size n from the whole population, we get independent samples of size Ni from each of the row classifications. As before, let Nij be the number of observations in the (i,j) cell. We assume that
    Ni=(Ni1,,Nic)~Mc(Ni,ηi),ηi=(ηi1,,ηic),
    where ηij is the proportion of the ith row class in column class j. Let Ȳi., ηij^, μ̄i., μ̃i, eijk and ēi.. be defined as in Section 2. The lemma is still true for this model, since it is conditional on the Ni. Similarly, Theorem 1 is conditional on the Ni, so that we see that the Ȳi.. are independent (since they are from independent samples) and
    E(Y¯i..)=μi.,var(Y¯i..)=δiσ2/Ni.
    (Note that for this model, we do not have to worry about the case Ni = 0.) Hence Ȳi.. is unbiased and consistent for this model. Similarly Theorem 2 is derived conditionally on the Ni, so that we have
    Ni1/2(Y¯i..-μ¯i.)dUi~N(0,δiσ2)asNi.
    Now suppose that
    n=iNi,Ni/nπi.asn

    for some constants πi.. Then Ni/nPπi. as Ni→ ∞, so that the corollary and Theorem 3 follow also. Therefore, the procedure derived in Section 2 with πi. replaced by limn→ ∞(ni/n) applies equally well to the model considered in this paragraph.

4. Numerical evaluations of the test in (3.20)

The proposed test Fn in (3.20) was evaluated numerically under the null hypothesis of equality of row means in (3.16) for several models of 2 × 2, 2 × 3 and 3 × 4 designs with simulation size 10,000. Tables 13, give the models, the cell probabilities, the cell means and the equal row means. The assumed interactions were non-zero but non-significant and the means are plotted in the associated Figs. 13. Entries in the tables are the percent of rejections out of the 10,000 samples, for various sample sizes. The size of the entries should be compared to 5%. As seen from all three tables, although asymptotic, the proposed test performs well in maintaining the correct Type-I error.

Table 1.

Type_I error rates of the new test for row effect of 2 × 2 factorial designs, with non-significant interaction.

Model Cell probabilities Cell means, SD Row means n = 30 n = 50 n = 70 n = 100
Model I (.3,.4,.1,.2) (50,15,35,27.5)
SD = 20
(30,30) 4.35 4.93 4.35 4.83
Model II (.3,.4,.1,.2) (50,30,5.71,55)
SD = 50
(38.571,38.571) 4.90 5.62 5.59 6.34
Model III (.3,.4,.2,.1) (50,15,40,10)
SD = 4
(30,30) 5.23 5.92 6.35 6.52
Model IV (.2,.4,.3,.1) (32,20,30,6)
SD = 10
(24.24) 4.95 5.84 6.24 6.71

Table 3.

Type I error rates of the new test, for row effect of 3 × 4 factorial designs, with non-significant interaction.

Model Cell probabilities Cell means, SD Row means n = 70 n = 100 n = 200
Model I (.8,.2,.15,.05,.1,.02,.04,.03,.05,.1,.1,.08) (50,15,20,10,25,27.5,10,23.54,15,15,40,12.11)
SD = 40
(21.88,21.88,21.88) 2.77 3.53 4.13
Model II (.08,.2,.15,.05,.1,.02,.04,.03,.05,.1,.1,.08) (40,30,20,50,35,27.5,10,45.63,30,15,20,63.83)
SD = 40
(30.63,30.63,30.63) 2.81 3.46 4.07
Model III (.08,.05,.15,.05,.1,.16,.04,.03,.05,.11,.1,.08) (40,30,20,50,35,30,45,3.33,20,15,30,60.74)
SD = 50
(30.91,30.91,30.91) 3.17 3.65 4.11
Model IV (.08,.02,.1,.05,.1,.33,.04,.03,.05,.02,.1,.08) (60,15,20,10,25,35,15,18.33,30,5,20,50)
SD = 50
(30.4,30.4,30.4) 3.38 3.72 4.94

Fig. 1.

Fig. 1

Graphs of means for 2 × 2 factorial designs.

Fig. 3.

Fig. 3

Graphs of means for 3 × 4 factorial designs.

Fig. 2.

Fig. 2

Graphs of means for 2 × 3 factorial designs.

Table 2.

Type I error rates of the new test, for row effect of 2 × 3 factorial designs, with non-significant interaction.

Model Cell probabilities Cell means, SD Row means n = 50 n = 70 n = 100
Model I (.2,.3,.1,.1,.1,.2) (50,15,20,35,27.5,23.75)
SD = 15
(27.5,27.5) 4.22 4.37 5.03
Model II (.2,.3,.1,.1,.1,.2) (40,35,20,45,45,23.33)
SD = 10
(34.17,34.17) 4.83 4.97 5.49
Model III (.2,.3,.1,.1,.1,.2) (65,5,60,30,45,30.83)
SD = 50
(34.17,34.17) 4.08 4.30 4.95
Model IV (.3,.1,.05,.05,.1,.4) (30,40,60,30,45,33.89)
SD = 30
(35.56,35.56) 3.98 4.23 4.38

Acknowledgments

Dr. Moschopoulos’s research was partially supported by Grants from the National Center for Research Resources 5G12RR008124 and the National Institute for Minority Health and Health Disparities Grant G12MD007592 from the National Institutes of Health. We express our thanks to Dr. Julia Bader of the Statistical Consulting Laboratory of UTEP for help with the simulations. Finally, we thank the editors and two anonymous referees for comments and revisions that improved the final form of the paper.

References

  1. Arnold SF. Theory of Linear Models and Multivariate Analysis. Wiley; New York: 1981. [Google Scholar]
  2. Galassi JP, Frierson HT, Jr, Sharer R. Behavior of high, moderate and anxious students during an actual test situation. Journal of Consulting and Clinical Psychology. 1981;49 (1):51–62. doi: 10.1037//0022-006x.49.1.51. [DOI] [PubMed] [Google Scholar]
  3. Groat HT, Neal AG. American Sociological Review. 1967. Social psychological correlates of urban fertility; pp. 945–959. [PubMed] [Google Scholar]
  4. Graybill FA. Theory and Application of linear Model. Duxburry; North Scituate, MA: 1976. [Google Scholar]
  5. Kutner MH. Hypothesis testing in linear models. The American Statistician. 1974;28:98–100. [Google Scholar]
  6. Macnaugton D. Which sums of squares are best in unbalanced analysis of variance?. Paper presented at the Joint Statistical Meetings; Boston. 1992; 1998. 〈 http://www.matstat.com/ss/easleaao.pdf〉. [Google Scholar]
  7. Moschopoulos PG, Davidson MI. Hypothesis testing in ANOVA under multinomial sampling. Sankhyâ Series B. 1985;47(Pt. 3):301–309. [Google Scholar]
  8. Scheffè H. The Analysis of Variance. Wiley; NY: 1959. [Google Scholar]
  9. Searle SR. Linear Models. Wiley; NY: 1971. [Google Scholar]
  10. Searle SR. Linear Models for Unbalanced Data. 1982. [Google Scholar]
  11. Speed FM, Hoching RR, Hackney OP. Methods of analysis of linear models with unbalanced data. Journal of the American Statistical Association. 1978;73:105–113. [Google Scholar]
  12. Weiss DJ. An analysis of variance test for random attrition. Journal of Social Behavior and Personality. 1999;14:433–438. [Google Scholar]
  13. Weiss DJ. Analysis of Variance and Functional Measurement: A Practical Guide. Oxford University Press; New York: 2006. [Google Scholar]

RESOURCES