Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2019 Mar 27;73(3):351–372. doi: 10.1111/stan.12173

Bayesian estimation of explained variance in ANOVA designs

Maarten Marsman 1,, Lourens Waldorp 1, Fabian Dablander 2, Eric‐Jan Wagenmakers 1
PMCID: PMC6618269  PMID: 31341338

Abstract

We propose to use the squared multiple correlation coefficient as an effect size measure for experimental analysis‐of‐variance designs and to use Bayesian methods to estimate its posterior distribution. We provide the expressions for the squared multiple, semipartial, and partial correlation coefficients corresponding to four commonly used analysis‐of‐variance designs and illustrate our contribution with two worked examples.

Keywords: analysis of variance, credible interval, effect size

1. INTRODUCTION

In the empirical sciences, researchers are commonly advised to report effect size (ES) interval estimates to complement the use of statistical hypothesis testing (e.g., Loftus, 1996; Cohen, 1992, 1994; Cumming, 2014). This is particularly true for psychology, where the need for reporting ES estimates is stressed in publication guidelines of several organizations and scientific journals (see Kelley & Preacher, 2012; Peng, Chen, Chiang, & Chiang, 2013 for recent surveys), but it also applies to other fields of science, such as biology (Nakagawa & Cuthill, 2007), genetics (Park et al., 2010), and marketing research (Fern & Monroe, 1996). Clearly, the use and interpretation of ES interval estimates is of practical importance and continues to attract much research interest (e.g., Huberty, 2002; Robinson, Whittaker, Williams, & Beretvas, 2003; Alhija & Levy, 2008; Fritz, Scherndl, & Kühberger, 2012; Grissom & Kim, 2012; Lakens, 2013; Fritz, Morris, & Richler, 2012).

Despite the continued interest in ES interval estimation, we have two concerns as to their development for the analysis of variance (ANOVA) model, which is arguably the most often applied statistical model across the empirical sciences. Our first concern is that ES measures for ANOVA models are often defined in a somewhat ad hoc manner and follow what Kelley and Preacher (2012, p. 139) call a definition‐by‐example approach, making it difficult to generalize the ES measures beyond the specific examples given. One case in point is the ω 2 measure for the two‐way ANOVA (Hays, 1963, p. 406) and the subsequent difficulty of generalizing it to a partial ω 2 measure (Keren & Lewis, 1979). Similar issues have been reported for the commonly used η 2 measure (Cohen, 1973; Kennedy, 1970; Levine & Hullet, 2002; Pierce, Block, & Aguinis, 2004). Thus, what is lacking is a general framework within which ES measures such as ω 2 and η 2 can be derived.

Our second concern is that ES interval estimation for ANOVA models has been treated almost exclusively from a classical perspective. Despite their popularity, classical confidence intervals (CIs) are often misunderstood and misinterpreted (Hoekstra, Morey, Rouder, & Wagenmakers, 2014; Morey, Hoekstra, Rouder, Lee, & Wagenmakers, 2016). Furthermore, Morey et al. (2016) demonstrated that current CI procedures for ω 2 are particularly problematic, as they can produce empty intervals and involve post hoc manual truncation at the boundary of the parameter space. In contrast to the classical CIs, Bayesian credible intervals are more straightforward to interpret (Wagenmakers, Morey, & Lee, 2016). In addition, credible intervals do not have the anomalies that beset some of the classical ES intervals for ANOVA designs. Despite their added value, however, Bayesian methods of ES estimation for ANOVA models have so far received little attention.

In light of these concerns, our goals are twofold. Our primary goal is to show that the theory on squared multiple correlations ρ 2—well established for linear regression models—also provides a unifying framework for ES measures in ANOVA designs. Several ES measures that are commonly reported in the ANOVA literature (e.g., ω 2, η 2, partial ω 2, and partial η 2) will be shown to be special cases of ρ 2. Although the connection between ES measures such as ω 2 and η 2 for the ANOVA model and the squared multiple correlations ρ 2 for the linear regression model has been noted before (e.g., Keren & Lewis, 1979; Maxwell, Camp, & Arvey, 1981; Sechrest & Yeaton, 1982; Cohen, Cohen, West, & Aiken, 2003), little is known about the assumptions that underpin this relation. We will detail these assumptions and indicate the relation for several commonly used ANOVA designs, thereby providing a formal basis for the development of ES measures for the ANOVA model.

Our secondary goal is to estimate the squared multiple correlations using Bayesian methods. Although the issue of Bayesian estimation of multiple correlations has received considerable attention in the context of linear regression (see, for instance, Geisser, 1965; Press & Zellner, 1978; Tiwari, Chib, & Jammalamadaka, 1989; Gross, 2000), it has received almost no attention in the context of ANOVA models. One notable exception is the work of Gelman and Pardoe (2006), which focused on hierarchical extensions of the classical ANOVA model (see also Gelman, 2005). While the approach of Gelman and Pardoe (2006) was based on a sample definition of ρ 2 (i.e., R 2), we instead derive the population expressions for ρ 2. The resulting expressions are found to be functions of the ANOVA model parameters, such that ρ 2 can be estimated from the posterior distribution of the ANOVA model parameters.

Below, we first discuss the general background, followed by a more targeted treatment of four popular ANOVA designs. We end by presenting two worked examples.

2. GENERAL BACKGROUND

Let (y,xT)T denote a (column) vector of variables with covariance matrix

Σ=Var(y)Cov(y,xT)Cov(x,y)Var(x)=ΨΔTΔΓ. (1)

We will refer to variable y as a dependent variable and the (vector of) variable(s) x as the independent variable(s). Given the covariance matrix Σ, we can express the relation between the dependent variable y and the independent variables x using the squared multiple correlation coefficient (Tate, 1966; Cohen, 1982; Mudholkar, 2014), as follows:

ρ2=ρyx2=1Σ|Ψ||Γ|=1|Γ|ΨΔTΓ1ΔΨ|Γ|=ΔTΓ1ΔΨ, (2)

provided that the covariance matrix Γ=Var(x) is nonsingular and that Ψ = Var( y) > 0. The squared multiple correlation coefficient ρ 2 is the maximum squared correlation that can be obtained between a single dependent variable y and any linear combination of the set of independent variables x (e.g., Anderson, 1958, p. 88).

Similar to most of the ES measures that have been proposed for the ANOVA model, the squared multiple correlation coefficient ρ 2, as defined in Equation (2), is a so‐called proportional reduction in error (PRE) measure (Reynolds, 1977). In general, a PRE measure expresses the proportion of the variance in an outcome y that is attributed to the independent variables x. Here, we treat the assignment in a design D as a random variable, and we decompose the variance of the dependent variable as follows:

Var(y)=Var(E(yx))+E(Var(yx)). (3)

Because the ANOVA model assumes homoscedasticity, the design affects only the conditional expectations (means) of the dependent variable y. In other words, Var(E( yx)) is the variance that can be attributed to random assignment in the design, and E(Var( yx)) is the residual variation that is constant across the cells in the design D. As a result, we can reexpress Equation (3) as

σY2=σD2+σE2,

where σD2=Var(E(yx)) denotes the design (or assignment) variance and σE2=E(Var(yx)) denotes the error variance. In this form, Equation (3) is used in the ANOVA literature to define PRE measures as

1σE2σY2=σD2σY2=σD2σD2+σE2. (4)

ANOVA ES measures defined in this manner include ω 2 (Hays, 1963, p. 382), η 2 (Hays, 1963, p. 547), and the intraclass correlation coefficient ρI2 (Hays, 1963, p. 424). In case of an ANOVA design with a single fixed factor, both ω 2 and η 2 specify or estimate σD2 as the sum of squares that are attributed uniquely to the fixed factor and σE2 as the residual variance (see Analysis I). When the factor is random, σD2 can be taken to be the variance of the random factor to define ρI2 (see Analysis IV).

The definition of ρ 2 in Equation (2) is of the same form as Equation (4)—note that σY2=Ψ and σD2=ΔTΓ1Δ. In order to apply Equation (2) to ANOVA, we have to define the joint model p( y,x), with x encoding the ANOVA design, from which we can derive Σ. To this aim, we first specify the ANOVA model as an ordinary linear regression model, utilizing dummy variables to encode the fixed effects (Jennings, 1967; Cohen, 1968). That is, for the fixed effects, we introduce indicator variables x i, with

xi=1,if subject is in groupi0,if subject is not in groupi.

We then have that the ANOVA model is a conditional normal distribution p( yx) that is of the same form as that used by Olkin and Tate (1961) to express correlations between discrete and continuous variables (see also Tate, 1954, 1966). To complete their model, Olkin and Tate (1961) specified a categorical (or multinomial) distribution for the indicator variables, that is,

p(x)=iπixi, (5)

where π i expresses the probability that a subject is assigned to a group i. With this multinomial distribution, we are able to cover many designs. In the derivation of ω 2, Hays (1963) assumed that the probabilities π i are uniform for every cell in the design, and expressions for η 2 typically generalize this assumption to incorporate unbalanced designs. To foreshadow our later results, note that posterior inference on the parameters of the ANOVA model will be unaffected by a particular choice of model p(x) for the design/indicator variables whenever the design model p(x) does not depend on the parameters of the ANOVA model.

Before we analyze four common ANOVA designs, we first define two correlation measures that are closely related to the squared multiple correlation: the semipartial correlation coefficient (also known as the part correlation coefficient) and the partial correlation coefficient. Suppose that we wish to express the contribution of some subset x A of the design variables x = {x A,x B} (i.e., xADA, xBDB, and D=DA×DB). The squared semipartial correlation ρyxA|xB2 is then defined as (Cohen, 1982, p. 308)

ρyxA|xB2=ρyx2ρyxB2=ΔTΓ1ΔΔBTΓB1ΔBΨ, (6)

where ρyxB2 is the squared multiple correlation coefficient between y and the (sub)set x B (i.e., the correlation obtained by removing any columns and rows in Σ that correspond to the set x A). The squared semipartial correlation ρyxA|xB2 expresses the proportion of the variance in the outcome y that can be uniquely attributed to the subset x A. It is the population value of the R 2‐change statistic that is commonly used in multiple linear regression. Consequently, the squared partial correlation coefficient is defined as (Cohen, 1982, p. 308)

ρyxA·xB2=ρyxA|xB21ρyxB2=ΔTΓ1ΔΔBTΓB1ΔBΨΔBTΓB1ΔB, (7)

where 1ρyxB2 denotes the proportion of variance in y that is not explained by x B. From this expression, we see that the squared semipartial correlation ρyxA|xB2 expresses the proportion of the variance that can be attributed to the subset x A after the unique contribution of x B has been completely removed.

3. ANALYSIS I: AN ANOVA DESIGN WITH ONE FIXED FACTOR

Consider an ANOVA design with a single fixed factor consisting of n levels, that is,

y=β0+i=1nβixi+ϵ,ϵN(0,σ2),

where x i is an indicator variable, with

xi=1,if subject is in groupi0,if subject is not in groupi

such that β 0 + β i is the expectation of y in group i, and it is assumed that the (residual) variance σ 2 is constant within each group. For completeness, we note here that the parameters in the conditional model p( yx) are not identified due to the use of dummy variables for the conditions. To identify the model parameters, we impose the sum‐to‐zero constraint β 1 + ⋯ + β n = 0. Upon assuming the multinomial distribution in Equation (5) for the indicator variables, it is then easy to confirm that we obtain the matrix Σ with the following elements:

Ψ=Var(y)=σ2+i=1nπi(βiβ)2Γij=Cov(xi,xj)=πi(1πi),ifi=jπiπj,ifijΔi=Cov(y,xi)=πi(βiβ),

where we have used β to denote the weighted average iπiβi.

For the definition of the squared multiple correlation coefficient in Equation (2), we require the inverse of

Γ=Var(x)=diag(𝛑)𝛑𝛑T, (8)

where 𝛑=(π1,π2,,πn)T, which is noninvertible due to the constraint ixi=1. Without loss of generality, however, we can omit the contribution of one of the x i's (c.f. Olkin & Tate, 1960). Let Γ˜ denote the matrix Γ with the ith row and column removed, and let Δ˜ denote the vector Δ with the i entry removed. In the Appendix, we omit the contribution of x n and show that

Δ˜TΓ˜1Δ˜=i=1nπi(βiβ)2.

Consequently, the expression for the squared multiple correlation coefficient is

ρyx2=i=1nπi(βiβ)2σ2+i=1nπi(βiβ)2, (9)

which is equation (3.12) from Olkin and Tate (1961).

In Equation (9), we see that the multinomial probabilities weigh the individual contributions from each group. In a balanced design, we assume that the weights (probabilities) are the same for each group (i.e., π i = n −1), and in that case, Equation (9) simplifies to

ρyx2=1ni=1nβi2σ2+1ni=1nβi2,

due to the sum‐to‐zero constraint; iπiβi=n1iβi=0. This is the expression for ω 2 given by Hays (1963, p. 382). Furthermore, replacing the cell probabilities with observed cell proportions gives η 2. Gelman and Pardoe (2006) replace iπi(βiβ)2 with the posterior expectation of 1n1i(βiβ¯)2=1n1iβi2, that is, the posterior expectation of the sample variance of the category parameters β i, which does not conform to an expression based on a multinomial distribution as the cell weights do not add to one. As a result, the factor contribution can be overestimated for designs with factors that have a small number of levels.

4. ANALYSIS II: AN ANOVA DESIGN WITH TWO FIXED FACTORS

In this section, we analyze an ANOVA design with two fixed factors. First, we consider a model with the two main effects but without the interaction, as this model is a direct extension of the single fixed factor discussed in the previous section. Second, we investigate the complete, full‐factorial model that includes both the two main effects as well as the interaction.

4.1. Two main effects model

Consider an ANOVA model with two fixed factors consisting of n 1 and n 2 levels, respectively, and no interaction effects, that is,

y=β0+i=1n1β1ix1i+j=1n2β2jx2j+ϵ,ϵN(0,σ2),

where β 0 + β 1i + β 2j is the expectation of y in cell (i,j) of the n 1 × n 2 design matrix, and x fi are indicator variables, with

xfi=1,if subject is in groupiof factorf0,if subject is not in groupiof factorf,

and it is assumed that the (residual) variance σ 2 is constant across all cells. Upon assuming independent multinomial distributions for the indicator variables of each factor (i.e., assuming that the manipulated factors are orthogonal), we obtain the matrix Σ with the following elements:

Ψ=Var(y)=σ2+f=12i=1nfπfi(βfiβf)2Γij=Cov(xfi,xfj)=πfi(1πfi),ifi=jandf=fπfiπfj,ifijandf=f0,ifffΔfi=Cov(y,xfi)=πfi(βfiβf),

where we have used β f to denote the weighted average iπfiβfi.

To compute the squared multiple correlation coefficient using (2), we require the inverse of

Γ=Cov(x1,x1)Cov(x1,x2)Cov(x2,x1)Cov(x2,x2)=Γ100Γ2,

where Γ f is the covariance matrix of the indicator variables for factor f as defined in Equation (8). Since Γ is a block‐diagonal matrix, we only require Γ11 and Γ21 to express Γ −1. Using the results from the Appendix, where Γ˜f denotes the matrix Γ f with the final column and row removed, for f = 1,2, we obtain

ρyx2=f=12i=1nfπfi(βfiβf)2σ2+f=12i=1nfπfi(βfiβf)2,

which is similar to the single‐fixed‐factor case that we have analyzed earlier.

We will now consider the semipartial and partial correlation for the unique contribution of the first factor after the contribution of the second factor has been partialed out. From the definition in Equation (6), we readily find the semipartial correlation

ρyx1|x22=i=1n1π1i(β1iβ1)2σ2+f=12i=1nfπfi(βfiβf)2,

and we can confirm that ρyx2=ρyx1|x22+ρyx2|x12. Similarly, we can find the partial correlation

ρyx1·x22=i=1n1π1i(β1iβ1)2σ2+i=1n1π1i(β1iβ1)2,

which is identical to the expression for the multiple squared correlation in a single‐fixed‐factor design.

4.2. Full‐factorial model

Next, consider the ANOVA model with both main effects and the interaction, that is,

y=β0+i=1n1β1ix1i+j=1n2β2jx2j+i=1n1j=1n2β3ijx3ij+ϵ,ϵN(0,σ2),

where β 0 + β 1i + β 2 j + β 3ij is the expectation of y in a cell (i,j) of the n 1 × n 2 design matrix, and x 3ij are indicator variables, with

x3ij=x1i×x2j=1,if subject is in cell(i,j)0,if subject is not in cell(i,j).

To identify the model, we again adopt the sum‐to‐zero constraint as follows:

i=1n1β1i=j=1n2β2j=i=1n1β3ij=j=1n2β3ij=0.

By assuming that the indicator variables encoding the main effects have independent multinomial distributions for each factor, we find that the indicator variables encoding the interaction effects also have a multinomial distribution, that is,

p(x3)=i=1n1j=1n2(π1i×π2j)x3ij=k=1n1×n2π3kx3k, (10)

where the latter provides shorthand notation that is useful later on.

With the two‐main‐effects‐only model, we could use the results from the Appendix and omit a row and column in Γ and Δ for each of the main effects, that is, rows and columns that correspond to x1n1 and x2n2, to express Δ˜TΓ˜1Δ˜. This is not the case for the full‐factorial design, however, since even after omitting the rows corresponding to x1n1 and x2n2 (and x3n3), the inverse Γ˜1 does not exist. We can think of two possible solutions to this problem. The first is to remove the contributions of x1n1 and x2n2 altogether, that is, remove a single row and column from both Γ 1 and Γ 2, and remove n 1 + n 2 − 1 rows and columns from Γ 3. However, as a result of leaving out these elements from Γ, and thus Δ, we lose much information about the design. We therefore take a different approach and use the main effects {x 1,x 2} in isolation from x 3 to express Δ˜TΓ˜1Δ˜, and vice versa. We will see that both the main effects and the interaction effects can be obtained with this approach.

First, we will study the model implied by only using the indicator variables encoding the interaction x 3 and ignore the {x 1,x 2} variables. That is, we will first use Γ=Var(x 3) and Δ=Cov(x 3,y) for Σ. Second, we will ignore x 3 and study the contribution of the main effects using Γ=Var({x 1,x 2}) and Δ=Cov({x 1,x 2},y) for Σ. Here, we can isolate the unique contribution of the main effects. By combining the two approaches, we can subsequently isolate the unique contributions of main effects and interaction and express these contributions in the form of semipartial and partial correlations. Thus, the solution comprises several steps that will be detailed below.

In the first step, we consider the use of x 3 encoding the interaction and ignore the {x 1,x 2} variables. It is easy to confirm that for this situation, Σ has the elements

Ψ=Var(y)=σ2+i=1n1j=1n2π1iπ2j(βijβ)2Γkl=Cov(x3k,x3l)=π3k(1π3k),ifk=lπ3kπ3l,ifklΔk=Cov(y,x3k)=Cov(y,x3ij)=π1iπ2j(βijβ),

where we have used the shorthand notation from Equation (10), that is, β ij to denote β 1i + β 2 j + β 3ij, β to denote β 1 + β 2 + β 3, and β 3 to denote the weighted average ijπijβ3ij. Since Γ is the covariance matrix of a single multinomial random variable, we have that (e.g., omitting x n as in the Appendix)

Δ˜TΓ˜1Δ˜=i=1n1j=1n2π1iπ2j(β1i+β2j+β3ijβ)2.

Thus, the squared multiple correlation coefficient can be expressed as

ρyx2=i=1n1j=1n2π1iπ2j(β1i+β2j+β3ijβ)2σ2+i=1n1j=1n2π1iπ2j(β1i+β2j+β3ijβ)2,

which, however, does not partition into separate terms for the main effects and the interaction when the cell probabilities π ij are not uniform. When the cell probabilities are uniform, that is, πij=1n1×n2, the squared multiple correlation coefficient equals

ρyx2=i=1n1β1i2n1+j=1n2β2j2n2+i=1n1j=1n2β3ij2n1×n2σ2+i=1n1β1i2n1+j=1n2β2j2n2+i=1n1j=1n2β3ij2n1×n2, (11)

which does partition into separate terms for the main effects and the interaction. However, that the total sum of squares partitions in this manner is insufficient to partial out an effect. We therefore now turn to the model implied by {x 1,x 2}.

In the second step, we consider the covariance matrix that is implied by ignoring the indicator variables encoding the interaction and only consider the indicator variables for the main effects, that is, {x 1,x 2}. It is easy to confirm that, for this situation, the covariance matrix Σ has the elements

Ψ=Var(y)=σ2+i=1n1j=1n2π3ij(βijβ)2Γij=Cov(xfi,xfj)=πfi(1πfi),ifi=jandf=fπfiπfj,ifijandf=f0,ifffΔ1i=Cov(y,x1i)=π1iβ1iβ1+jπ2jβ3ijβ3Δ2i=Cov(y,x2i)=π2iβ2iβ2+iπ1iβ3ijβ3,

where the covariance matrix Γ is identical to the covariance matrix for the two main effects analysis that we considered before. As a result, we find the squared multiple correlation coefficient, as follows:

ρy{x1,x2}2=1Ψi=1n1π1iβ1iβ1+jπ2jβ3ijβ32+j=1n2π2jβ2jβ2+iπ1iβ3ijβ32,

which does not simplify further in the case of nonuniform cell probabilities π. Assuming uniform cell probabilities, that is, π1i=n11 and π2j=n21, we observe that this expression of the squared multiple correlation simplifies to

ρy{x1,x2}2=i=1n1β1i2n1+j=1n2β2j2n2σ2+i=1n1β1i2n1+j=1n2β2j2n2+i=1n1j=1n2β3ij2n1×n2,

from which one can also recognize a semipartial correlation, as the contribution of the interaction effect has been removed from the numerator but not from the denominator.

We can now consider the semipartial and partial correlations that express the contribution of the first main effect. Since we have seen that ρy{x1,x2}2=ρy{x1,x2}|x32, we have

ρyx1|x22=ρy{x1,x2}2ρyx22=ρy{x1,x2}|x32ρyx2|x32=ρyx1|{x2,x3}2.

As a result, we find the following expression for the squared semipartial correlation for the first main effect:

ρyx1|{x2,x3}2=1Ψi=1n1π1iβ1iβ1+jπ2jβ3ijβ32.

Note that when the cell probabilities are assumed to be uniform, this expression simplifies to

ρyx1|{x2,x3}2=i=1n1β1i2n1σ2+i=1n1β1i2n1+j=1n2β2j2n2+i=1n1j=1n2βij2n1×n2.

From here, we can find the squared partial correlation as

ρyx1·{x2,x3}2=i=1n1β1i2n1σ2+i=1n1β1i2n1,

which is again simply the expression for the multiple squared correlation in a single‐fixed‐factor design.

We have shown how to express the semipartial and partial correlation (ES) measures that correspond to an ANOVA design with two fixed factors. In particular, we have seen simple expressions for the case with uniform cell probabilities, since, in that case, the total sum of squares σD2 decomposes into distinct terms for the main effects and the interaction (c.f. Equation (11)), that is,

σD2=σA2+σB2+σA×B2,

with the terms denoting the contribution of the main effect from a factor A, factor B, and the interaction between A and B, respectively. As a result, we see that the semipartial correlation ρyx1|{x2,x3}2 expresses the unique contribution of the first factor, that is,

σA2σE2+σA2+σB2+σA×B2.

For a long time, this expression was proposed as a “partial‐ω 2”/“partial‐η 2” in the extension of ω 2 and η 2 measures for single‐factor designs (Kerlinger, 1964; Kennedy, 1970) and consequently used by statistical software packages such as SPSS (Levine & Hullet, 2002; Pierce et al., 2004). Others suggested a different expression (e.g., Cohen, 1968, 1973; Keren & Lewis, 1979), that is,

σA2σE2+σA2, (12)

since it “can be shown that in the same way ω2 is a squared multiple correlation (12) is a squared partial correlation, with the artificial predictor variables designating B and the AB interaction partialled out” (Keren & Lewis, 1979, p. 123). However, this is true only for the particular case of uniform cell probabilities (e.g., ω 2), since, for the nonuniform case (e.g., η 2), the partial correlation is found to be

iπ1iβ1iβ1+jπ2jβ3ijβ32σ2+iπ1iβ1iβ1+jπ2jβ3ijβ32,

and it is clear that we cannot completely partial out the interaction. That is, the (semi)partial measures do not express unique contributions in the nonuniform case.

5. ANALYSIS III: AN ANOVA DESIGN WITH ONE FIXED FACTOR AND ONE COVARIATE

Consider an ANOVA design with a single fixed factor consisting of n levels and a (set of m) covariate(s), that is,

y=β0+i=1nβ1ix1i+j=1mβ2jx2j+ϵ,ϵN(0,σ2),

where β0+β1i+j=1mβ2jx2j is the expectation of y for a subject in group i with covariate (vector) x 2. As before, we assume that the indicator variables are multinomial random variables and that the distribution of the indicator variables is independent of that of the covariates, that is, p(x 1,x 2) = p(x 1)p(x 2). For the specification of the covariance matrix, we have to specify the covariance matrix Γ 2 = Var(x 2), which depends on the choice of p(x 2). The work of Olkin and Tate (1961) focused on the case where p(x 2) is a multivariate normal that depends on x 1.

We consider here the case of a single covariate, which allows us to provide an analytic expression for the squared correlation coefficient in the general case requiring only that we are able to specify Var(x 2), since it is easy to confirm that the covariance matrix Σ has the elements

Ψ=Var(y)=σ2+i=1n1π1i(β1iβ1)2+β22Var(x2)Γ1ij=Cov(x1i,x1j)=π1i(1π1i),ifi=jπ1iπ1j,ifijΓ12i·=Cov(x1i,x2)=0Γ2=Var(x2)Δ1i=Cov(y,x1i)=π1i(β1iβ1)Δ2=Cov(y,x2)=β2Var(x2).

If one is willing to assume that the distribution of the covariate is normal with mean μ and variance τ 2, for instance, then we may replace Var(x 2) with τ 2. Similarly, one could use the observed variance.

By assuming that the indicator variables and the covariate(s) are independent, we observe that the covariance matrix Γ is (block‐)diagonal, such that we can apply the results from the Appendix to obtain

ρy{x1,x2}=i=1n1π1i(β1iβ1)2+β22τ2σ2+i=1n1π1i(β1iβ1)2+β22τ2.

From here, it is simple to find the semipartial correlation for the covariate as

ρyx2|x1=β22τ2σ2+i=1n1π1i(β1iβ1)2+β22τ2

and, similarly, the partial correlation as

ρyx2·x1=β22τ2σ2+β22τ2.

When more than one covariate is used, it is clear that the expressions above depend on the covariance structure Γ 2. This is illustrated in the next analysis, where we consider a bivariate normal distribution for the random effects, which is one specific generalization of the results that we obtained here.

6. ANALYSIS IV: AN ANOVA DESIGN WITH ONE FIXED AND TWO CORRELATED RANDOM FACTORS

Consider an ANOVA design with a single fixed factor of n levels and a (set of m) random effect(s), that is,

y=β0+i=1nβ1ix1i+j=1mx2j+ϵ,ϵN(0,σ2),

where β0+β1i+j=1mx2j is the mean (expected) value of y for a subject in group i with random effect (values) x 2. In typical applications, it is assumed that the random effects have a (multivariate) normal distribution with a zero‐mean vector (for identification) and covariance matrix Λ. We consider here the situation for two random effects.

We assume that the distribution of the indicator variables for the manipulated factor (x 1) is a multinomial distribution (independent from the random effects) and that the distribution of the random effects is (bivariate) normal, with covariance matrix

Γ2=τ12ϱτ1τ2ϱτ1τ2τ22,

where ϱ denotes the correlation between the two random effects. It is easy to confirm that the covariance matrix Σ has the elements

Ψ=σ2+i=1n1π1i(β1iβ1)2+τ12+τ22+2ϱτ1τ2Γ1ij=Cov(x1i,x1j)=π1i(1π1i),ifi=jπ1iπ1j,ifijΓ12ij=Cov(x1i,x2j)=0Δ1i=Cov(y,x1i)=π1i(β1iβ1)Δ2j=Cov(y,x2j)=τj2+ϱτjτ3j.

As before, we have assumed that the manipulated factors are independent of other variables, so that the covariance matrix Γ is block‐diagonal. Applying the results from the Appendix to Γ 1 and from observing that

Γ21=1τ12τ22(1ϱ2)τ22ϱτ1τ2ϱτ1τ2τ12,

we find the squared correlation coefficient as follows:

ρyx2=i=1n1πi(β1iβ1)2+τ12+τ22+2ϱτ1τ2σ2+i=1n1πi(β1iβ1)2+τ12+τ22+2ϱτ1τ2,

which explicitly uses the covariance between the random effects.

We can now consider semipartial and partial correlations for the first random effect. The semipartial correlation (i.e., removing the contribution of the fixed factor and the second random effect) describing the contribution of the first random effect is equal to

ρyx21|{x1,x22}2=(1ϱ2)τ12σ2+i=1n1πi(β1iβ1)2+τ12+τ22+2ϱτ1τ2,

which was to be expected, as (1ϱ2)τ12 is the variance of x 21 after conditioning on x 22. That is, the partial correlation coefficient

ρyx21·{x1,x22}2=(1ϱ2)τ12σ2+(1ϱ2)τ12

is the intraclass correlation coefficient (Hays, 1963, p. 424).

7. BAYESIAN INFERENCE ON ρ 2

The correlation coefficients discussed in this paper are functions of the parameters from the correlation model

p(y,xβ,σ,𝛌)=p(y|x,β,σ)p(x|𝛌),

where p( y | x,β,σ) denotes the ANOVA model that depends on cell means β and the residual variance σ 2, and p(x | λ) is the distribution that we assume for the design variables and depends on the parameter vector λ (e.g., cell probabilities π and the population parameters of any random effects or covariates). By assigning prior probabilities p(β,σ) and p(λ) independently to the parameters {β,σ} and λ, we observe that conditional upon the observed data X and Y, the parameters {β,σ} and λ are conditionally independent, that is,

p(β,σ,𝛌|X,Y)=p(β,σ|X,Y)p(𝛌|X).

This factorization confirms that posterior inference on {β,σ} is unaffected by a specific choice of model p(x|λ).

Posterior inference on the correlation coefficients proceeds by simulation: Generate β and σ as a sample from p(β,σ | X,Y) and λ as a sample from p(λ | X). Given a simulated draw for each parameter of the joint model, that is, β , σ , and λ , we can compute the correlation measure of interest. Repeating this procedure many times produces samples from the posterior of the correlation coefficient that we may use to produce Monte Carlo estimates of quantities of interest, such as posterior medians, means, credible intervals, etc. When the prior distributions p(β,σ) and p(λ) are proper probability distributions, one can evaluate the prior distribution on the correlation measure in a similar way: Generate β , σ , and λ from the prior distributions and use these values to compute the correlation measure as a sample from the prior.

Further elaboration of the design model and its parameter λ offers an interesting avenue for further study, for instance, in the case of targeted, unbalanced, or multistage experiments. However, in regular experimental analyses, we are usually not interested in λ (except perhaps in the case of random effects), and we can simply use the observed values (i.e., cell proportions, means, and covariances of the covariates). We therefore only require draws from the posterior distributions of the ANOVA model parameters {β,σ}. Procedures to sample from these posterior distributions are described in many publications (see, for instance, Gelman et al., 2014; Gelman, 2005) and oftentimes use Markov chain Monte Carlo methods. A convenient tool is the BayesFactor package that is available for GNU‐R and works for many of the commonly used ANOVA designs (Morey & Rouder, 2015).

7.1. Example I: Incorporating prior information in a single‐fixed‐factor design

To illustrate the use of our approach to ρyx2, we consider data from an experiment performed by Gibson, Losee, and Vitiello (2014), which was part of a special issue “Replications of important results in social psychology” (Nosek & Lakens, 2013, 2014). In this experiment, Gibson et al. (2014) attempted to replicate the results of Shih, Pittinsky, and Ambady (1999) on stereotype susceptibility. Shih et al. (1999) had investigated the performance of Asian American women on a math test after the women had been surreptitiously attented to either their Asian identity (i.e., high maths ability stereotype), or their female identity (i.e., low maths ability stereotype), or have not been primed as such (control condition). The original results of Shih et al. (1999) suggested a relation between stereotype priming and maths performance: Subjects in the Asian identity condition outperformed subjects in the female identity and control conditions, and subjects in the control group outperformed subjects in the female identity condition.

In their replication attempt, Gibson et al. (2014) collected N = 158 participants: 52 in the Asian identity condition, 54 in the female identity condition, and 52 in the control condition. (The data for this example are available at https://osf.io/vnaqq/.) We analyzed the Gibson data as a single‐fixed‐factor design using a setup similar to that of Rouder, Morey, Speckman, and Province (2012), using noninformative Jeffreys priors on β 0 and σ 2 and g‐priors for β 1, β 2, and β 3 with the restriction that iβi=0, where we use β 1, β 2, and β 3 to refer to the departure of the overall mean for the Asian identity, female identity, and control conditions, respectively. For the squared correlation coefficient, we have used expression (9) and simply replaced the cell probabilities with the observed proportions.

In Table 1, we summarize the posterior distributions of the model parameters and the squared multiple correlation coefficient in terms of the posterior means, standard deviations, and the 2.5%, 50%, and 97.5% quantiles. Note that there is little variability in the mean departures between the three groups, indicating that the effect of the manipulation is small. This can also be observed from the posterior median of ρ 2 indicating that less than 2% of the total variance is attributed to the variability between conditions, with a small margin of uncertainty as indicated by the narrow 95% central credible interval that ranges from .001 to .062.

Table 1.

Posterior summaries for model parameters and the squared multiple correlation for Example I

Mean SD 2.5% 50% 95%
β 0 0.558 0.019 0.520 0.558 0.595
β 1 0.029 0.025 −0.020 0.029 0.079
β 2 −0.023 0.025 −0.072 −0.023 0.026
β 3 −0.006 0.025 −0.055 −0.006 0.043
σ 2 0.057 0.007 0.046 0.057 0.071
ρyx2
.019 .017 .001 .014 .062

Note. SD = standard deviation.

Note that Gibson et al. (2014) had available prior information from the original experiment of Shih et al. (1999). In particular, the original results suggested that the Asian identity group outperformed both the female identity and control groups and that the control group outperformed the female identity group. This information is easily incorporated into the prior distribution of the model parameters using the order constraint β 2 < β 3 < β 1. It should be clear that when the prior constraint aligns with the observed data, we obtain an estimate of ρ 2 that is at least as high as the estimate of ρ 2 obtained without including prior constraints, but we would obtain a lower estimate when the prior constraint does not align with the observed data.

In Table 2, we show the posterior means, standard deviations, and the 2.5%, 50%, and 97.5% quantiles for the parameters and the multiple correlation using the prior constraint β 2 < β 3 < β 1. From Table 2, we see that including the prior constraint increased the variability of the group mean discrepancies, whereas the error variability σ 2 remained the same. That is, ρ 2 increased. This was to be expected since the posterior results in Table 1 aligned with the hypothesis.

Table 2.

Posterior summaries for model parameters and the squared multiple correlation for Example I using the prior restriction β 2 < β 3 < β 1

Mean SD 2.5% 50% 95%
β 0 0.558 0.019 0.521 0.558 0.595
β 1 0.045 0.022 0.010 0.043 0.093
β 2 −0.040 0.018 −0.080 −0.038 −0.010
β 3 −0.006 0.015 −0.037 −0.005 0.024
σ 2 0.057 0.007 0.046 0.057 0.071
ρyx2
.026 .021 .002 .022 .079

Note. SD = standard deviation.

An advantage of a Bayesian approach to interval estimation, as opposed to a classical approach, is that the intervals and related statistics are derived from the posterior distribution. In Figure 1, we show histograms of 100,000 posterior samples for ρ 2, with the top panel showing the posterior distribution of ρ 2 without the prior constraint and the bottom panel showing the posterior distribution with the prior constraint. Since the posterior is a probability distribution, we can ask questions such as: “What is the (posterior) probability that ρ 2 > 0.05?” From Figure 1, it should be obvious that this probability is smaller for the estimated posterior distribution shown in the top panel than in the estimated posterior distribution in the bottom panel. To wit, the probability that ρ 2 > .05 equaled approximately 5% for the estimated posterior distribution in the top panel and approximately 12% for the one shown in the bottom panel. Observe that with the improper Jeffreys prior on σ 2, the prior distribution on ρyx2 is also improper. In fact, it will have all its mass placed at zero. This would mean that, in this case, the prior probability that ρyx2>0.05 is zero.

Figure 1.

stan12173-fig-0001

Histogram of 100,000 draws from the posterior distribution of ρ 2 given the unrestricted model in the top panel and the order‐restricted model in the bottom panel

7.2. Example II: An illustration of ρ 2‐change in a repeated‐measures design

To illustrate the use of the semipartial correlation coefficient as ρ 2‐change (Cohen, 1982), we consider data from Žeželj and Jokić (2014), which was also part of the Social Psychology special issue. (The data for this example are available at https://osf.io/z5te6/.) Žeželj and Jokić (2014) aimed at an exact replication of the results from Eyal, Liberman, and Trope (2008), who found that “people would judge immoral acts more harshly if presented to them as temporally distant rather than presented as temporally close” (Žeželj & Jokić, 2014, p. 225). The following experiment was conducted: Subjects were presented with three stories on morally offensive behaviors (eating one's dead pet, sexual intercourse with sibling, and dusting with the national flag) and were asked to imagine that the story would either happen the next day (near future) or the next year (distant future), after which participants were asked to judge the wrongness of these actions. Žeželj and Jokić (2014) collected data from N = 116 students, which were analyzed using a mixed‐ANOVA design with temporal distance and moral vignettes as between‐subject factors, that is, a 2 × 3 repeated‐measures design. We analyzed the data using the BayesFactor package (Morey & Rouder, 2015), using the main effects of temporal distance (fixed factor 1) and the moral vignettes (fixed factor 2), and a random participant factor.

Table 3 shows the posterior means, standard deviations, and the 2.5%, 50%, and 97.5% quantiles for the model parameters for this example. From Table 3, we see that there is little variability for the group mean deviations for the first factor—β 11 and β 12—but that there is nonnegligible variation in the mean deviations for the second factor—β 21, β 22, and β 23—and the contribution of the effect—τ 2. We will study their contributions, in turn, using semipartial correlations.

Table 3.

Posterior summaries for model parameters for Example II

Mean SD 2.5% 50% 95%
β 0 −2.383 0.191 −2.759 −2.383 −2.007
β 11 −0.056 0.181 −0.413 −0.055 0.299
β 12 0.056 0.181 −0.299 0.055 0.413
β 21 −1.535 0.164 −1.855 −1.535 −1.214
β 22 −1.416 0.164 −1.736 −1.416 −1.092
β 23 2.951 0.164 2.629 2.951 3.273
σ 2 4.688 0.437 3.907 4.663 5.613
τ 2 2.630 0.569 1.643 2.584 3.874

Note. SD = standard deviation.

For this example, the squared semipartial correlations are defined as

ρyx1|{x2,x3}2=i=1n1π1i(β1iβ1)2σ2+f=12i=1nfπfi(βfiβf)2+τ2ρyx2|{x1,x3}2=i=1n2π2i(β2iβ2)2σ2+f=12i=1nfπfi(βfiβf)2+τ2ρyx3|{x1,x2}2=τ2σ2+f=12i=1nfπfi(βfiβf)2+τ2,

where τ 2 is the variance of the random effect (i.e., subjects). The squared multiple correlation is defined as the sum of these three semipartial correlations, that is,

ρyx2=ρyx1|{x2,x3}2+ρyx2|{x1,x3}2+ρyx3|{x1,x2}2. (13)

From the posterior samples of the model parameters that we described in Table 3 and using the observed proportions π 1 = {.491,.509} and 𝛑2=13,13,13, we computed the squared correlation measures.

In Table 4, we summarize the posterior distributions of the squared semipartial correlations in terms of the posterior means, standard deviations, and the 2.5%, 50%, and 97.5% quantiles. As can be seen from Table 4 and was argued above, the amount of variation that can be attributed to the first fixed factor is very small, but the variation that can be uniquely attributed to the second fixed factor and the random factor is nonnegligible, with the posterior medians indicating that approximately 0.1%, 37.4%, and 22.2% of the total variance can be uniquely attributed to the first, second, and random factors, respectively, totaling approximately 60% of the variance being explained. While the 95% credible intervals for the semipartial correlations of the second factor (moral vignettes) and the random participant factor covered more than 10% of the total range, indicating that we are still uncertain about their exact values, the credible interval of the first factor (temporal distance) covered less than 2% of the total range, indicating here that we are fairly certain that temporal distance does not explain much variance.

Table 4.

Posterior summaries for the squared semipartial and multiple correlations for Example II

Mean SD 2.5% 50% 95%
ρyx1|{x2,x3}2
.003 .004 .000 .001 .015
ρyx2|{x1,x3}2
.373 .033 .308 .374 .437
ρyx3|{x1,x2}2
.223 .041 .147 .222 .307
ρyx2
.600 .037 .524 .601 .667

Note. SD = standard deviation.

The unique contributions of each of the three factors are visualized in Figure 2 as a ρ 2‐change, using 100,000 draws from the posterior distribution of each of the semipartial correlations. That is, in Figure 2, we show the amount of variance that is uniquely explained by the first fixed factor in the top panel, by the first and second fixed factors in the middle panel, and, finally, by the contribution for all three factors combined in the bottom panel.

Figure 2.

stan12173-fig-0002

Histograms of 100,000 draws from the posterior distribution of ρ 2‐change, with the top panel showing the unique contribution of the first fixed factor (x 1), the middle panel showing the joint contribution of both fixed factors (x 1 and x 2), and the bottom panel showing the joint contribution of all three factors (x 1, x 2, and x 3), that is, ρyx2

8. DISCUSSION

In this paper, we have shown that the theory on squared multiple correlation coefficients provides a useful framework for deriving ES measures for the ANOVA model, and we have analyzed several aspects that are commonly encountered in experimental designs. This allowed us to detail some expressions for the correlation coefficients and consequently clarify its underlying assumptions. As a result, we have shown that squared correlation measures encompass many of the existing ES measures that are now used for ANOVA, such as ω 2 and η 2, which are found to be special cases of the multiple correlation coefficient. Once the correlation expression is obtained, we note that it is easily evaluated using available software when the parameters of the design model are assumed fixed. In this manner, posterior interval estimates are simple to compute, which we have illustrated using two real‐data examples. Furthermore, we have shown in our examples how prior information can be included in the evaluation of the correlation coefficient and illustrated a Bayesian ρ 2‐change for the ANOVA model.

The analysis of the two‐way ANOVA design was complicated by the functional relation between the indicator variables for the main effects and the indicator variables encoding the interaction. We have solved this complication by combining two different approaches: one in which we only use the indicator variables encoding the interactions and one in which we only use the indicator variables that correspond to the main effects. In this manner, we were able to partial out the (unique) contribution of the main effects and the interaction in the particular case of uniform cell probabilities. Although not explicitly considered here, we believe that this approach will extend to situations involving more factors and interaction effects and to situations involving higher‐order interactions.

ACKNOWLEDGEMENT

This research was supported by the European Research Council (grant number 283876).

ANALYTIC EXPRESSION FOR Δ˜TΓ˜1Δ˜

Letting 𝛑˜=(π1,,πn1)T and D˜=diag(𝛑˜), we write the (n − 1) × (n − 1) matrix Γ˜ excluding the contribution of x n as

Γ˜=D˜𝛑˜𝛑˜T.

To apply Equation (2), we require the inverse (Sherman & Morrison, 1950), that is,

Γ˜1=D˜1+D˜1𝛑˜𝛑˜TD˜11𝛑˜TD˜1𝛑˜=D˜1+πn111T.

It is easy to verify that

Δ˜=D˜bT,

where Δ˜=Cov(x˜,y), x˜=(x1,,xn1)T, and b=(b1,,bn)T with bi=βii=1nπiβi=βiβ, such that

Δ˜TΓ˜1Δ˜=bD˜D˜1D˜bT+πn1bD˜11TD˜bT=bD˜bT+πn1b𝛑˜𝛑˜TbT=i=1n1πi(βiβ)2+πn1i=1n1πi(βiβ)2=i=1n1πi(βiβ)2+πn1i=1n1πiβii=1n1πiβ2=i=1n1πi(βiβ)2+πn1(βπnβn(1πn)β)2=i=1n1πi(βiβ)2+πn1(πnβπnβn)2=i=1nπi(βiβ)2.

Marsman M, Waldorp L, Dablander F, Wagenmakers E‐J. Bayesian estimation of explained variance in ANOVA designs. Statistica Neerlandica. 2019;73:351–372. 10.1111/stan.12173

REFERENCES

  1. Alhija, F. N.‐A. , & Levy, A. (2008). Effect size reporting practices in published articles. Educational and Psychological Measurement, 69, 245–265. [Google Scholar]
  2. Anderson, T. W. (1958). An introduction to multivariate statistical analysis. New York, NY: John Wiley & Sons. [Google Scholar]
  3. Cohen, J. (1968). Multiple regression as a general data‐analytic system. Psychological Bulletin, 70(6), 426–443. [Google Scholar]
  4. Cohen, J. (1973). Eta‐squared and partial eta‐squared in fixed factor ANOVA designs. Educational and Psychological Measurement, 33, 107–112. [Google Scholar]
  5. Cohen, J. (1982). Set correlation as a general multivariate data‐analytic method. Multivariate Behavioral Research, 17(3), 301–341. [DOI] [PubMed] [Google Scholar]
  6. Cohen, J. (1992). A power primer. Psychological Bulletin, 112(1), 155–159. [DOI] [PubMed] [Google Scholar]
  7. Cohen, J. (1994). The earth is round (p < .05). American Psychologist, 49(12), 997–1003. [Google Scholar]
  8. Cohen, J. , Cohen, P. , West, S. G. , & Aiken, L. S. (2003). Applied multiple regression/correlation analysis for the behavioral sciences (3rd ed.). Mahwah, NJ: Lawrence Erlbaum Associates Publishers. [Google Scholar]
  9. Cumming, G. (2014). The new statistics: Why and how. Psychological Science, 25, 7–29. [DOI] [PubMed] [Google Scholar]
  10. Eyal, T. , Liberman, N. , & Trope, Y. (2008). Judging near and distant virtue and vice. Journal of Experimental Social Psychology, 44, 1204–1209. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Fern, E. F. , & Monroe, K. B. (1996). Effect‐size estimates: Issues and problems in interpretation. Journal of Consumer Research, 23, 89–105. [Google Scholar]
  12. Fritz, A. , Scherndl, T. , & Kühberger, A. (2012). A comprehensive review of reporting practices in psychological journals: Are effect sizes really enough. Theory & Psychology, 23, 98–122. [Google Scholar]
  13. Fritz, C. O. , Morris, P. E. , & Richler, J. J. (2012). Effect size estimates: Current use, calculations and interpretation. Journal of Experimental Psychology: General, 141, 2–18. [DOI] [PubMed] [Google Scholar]
  14. Geisser, S. (1965). Bayesian estimation in multivariate analysis. Annals of Mathematical Statistics, 36, 150–159. [Google Scholar]
  15. Gelman, A. (2005). Analysis of variance—Why it is more important than ever. The Annals of Statistics, 33(1), 1–53. [Google Scholar]
  16. Gelman, A. , Carlin, J. B. , Stern, H. S. , Dunson, D. B. , Vehtari, A. , & Rubin, D. B. (2014). Bayesian data analysis(3rd ed.). Boca Raton, FL: Chapman & Hall/CRC press. [Google Scholar]
  17. Gelman, A. , & Pardoe, I. (2006). Bayesian measures of explained variance and pooling in multilevel (hierarchical) models. Technometrics, 48(2), 241–251. [Google Scholar]
  18. Gibson, C. E. , Losee, J. , & Vitiello, C. (2014). A replication attempt of stereotype susceptability (Shih, Pittinsky, & Ambady, 1999): Identity salience and shifts in quantitative performance. Social Psychology, 45(3), 194–198. [Google Scholar]
  19. Grissom, R. J. , & Kim, J. J. (2012). Effect sizes for research: Univariate and multivariate applications (2nd ed.). New York, NY: Routledge. [Google Scholar]
  20. Gross, A. L. (2000). Bayesian interval estimation of multiple correlations with missing data: A Gibbs sampling approach. Multivariate Behavioral Research, 35(2), 201–227. [DOI] [PubMed] [Google Scholar]
  21. Hays, W. L. (1963). Statistics. New York, NY: Holt, Rinehart & Wilson. [Google Scholar]
  22. Hoekstra, R. , Morey, R. D. , Rouder, J. N. , & Wagenmakers, E.‐J. (2014). Robust misinterpretation of confidence intervals. Psychonomic Bulletin & Review, 21(5), 1157–1164. [DOI] [PubMed] [Google Scholar]
  23. Huberty, C. J. (2002). A history of effect size indices. Educational and Psychological Measurement, 62(2), 227–240. [Google Scholar]
  24. Jennings, E. (1967). Fixed effects analysis of variance by regression analysis. Multivariate Behavioral Research, 2(1), 95–108. [DOI] [PubMed] [Google Scholar]
  25. Kelley, K. , & Preacher, K. J. (2012). On effect size. Psychological Methods, 17(2), 137–152. [DOI] [PubMed] [Google Scholar]
  26. Kennedy, J. J. (1970). The eta coefficient in complex ANOVA designs. Educational and Psychological Measurement, 30, 885–889. [Google Scholar]
  27. Keren, G. , & Lewis, C. (1979). Partial omega squared for ANOVA designs. Educational and Psychological Measurement, 39, 119–128. [Google Scholar]
  28. Kerlinger, F. N. (1964). Foundations of behavioral research. New York: Holt, Rinehart and Winston. [Google Scholar]
  29. Lakens, D. (2013). Calculating and reporting effect sizes to facilitate cumulative sciences: A practical primer for t‐tests and ANOVAs. Frontiers in Psychology, 4, 863. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Levine, T. R. , & Hullet, C. R. (2002). Eta squared, partial eta squared, and misreporting of effect size in communication research. Human Communication Research, 28(4), 612–625. [Google Scholar]
  31. Loftus, G. R. (1996). Psychology will be a much better science when we change the way we analyze data. Current Directions in Psychological Science, 5, 161–171. [Google Scholar]
  32. Maxwell, S. E. , Camp, C. J. , & Arvey, R. D. (1981). Measures of strength of association: A comparative examination. Journal of Applied Psychology, 66(5), 525–534. [Google Scholar]
  33. Morey, R. D. , Hoekstra, R. , Rouder, J. N. , Lee, M. D. , & Wagenmakers, E.‐J. (2016). The fallacy of placing confidence in confidence intervals. Psychonomic Bulletin and Review, 23(1), 103–123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Morey, R. D. , & Rouder, J. N. (2015). BayesFactor. R package version 0.9.11-1. https://CRAN.R-project.org/package=BayesFactor.
  35. Mudholkar, G. S. (2014). Multiple correlation coefficient In Balakrishnan N., Colton T., Everitt B., Piegorsch W., Ruggeri F. & Teugels J. L. (Eds.), Wiley StatsRef: Statistics reference online. 10.1002/9781118445112.stat00641 [DOI] [Google Scholar]
  36. Nakagawa, S. , & Cuthill, I. C. (2007). Effect size, confidence interval and statistical significance: A practical guide for biologists. Biological Reviews, 82, 591–605. [DOI] [PubMed] [Google Scholar]
  37. Nosek, B. A. , & Lakens, D. E. (2013). Call for proposals: Special issue of Social Psychology on "Replications of important results in social psychology". Social Psychology, 44, 59–60. [Google Scholar]
  38. Nosek, B. A. , & Lakens, D. (2014). Registered reports: A method to increase the credibility of published results. Social Psychology, 45(3), 137–141. [Google Scholar]
  39. Olkin, I. , & Tate, R. F. (1960). Multivariate correlation models with mixed discrete and continuous variables (Technical Report no. 58). Retrieved from Stanford University's Statistics Department website: https://statistics.sites.stanford.edu/sites/g/files/sbiybj6031/f/SOL%20ONR%2058.pdf
  40. Olkin, I. , & Tate, R. F. (1961). Multivariate correlation models with mixed discrete and continuous variables. The Annals of Mathematical Statistics, 32(2), 448–465. [Google Scholar]
  41. Park, J. H. , Wacholder, S. , Gail, M. H. , Peters, U. , Jacobs, K. B. , Chanock, S. J. , & Chatterjee, N. (2010). Estimation of effect size distribution from genome‐wide association studies and implications for future discoveries. Nature Genetics, 42, 570–575. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Peng, C. Y. J. , Chen, L. T. , Chiang, H. M. , & Chiang, Y. C. (2013). The impact of APA and AERA guidelines on effect size reporting. Educational Psychology Review, 25, 157–209. [Google Scholar]
  43. Pierce, C. A. , Block, R. A. , & Aguinis, H. (2004). Cautionary note on reporting eta‐squared values from multifactor ANOVA designs. Educational and Psychological Measurement, 64(6), 916–924. [Google Scholar]
  44. Press, S. J. , & Zellner, A. (1978). Posterior distribution for the multiple correlation coefficient with fixed regressors. Journal of Econometrics, 8, 307–321. [Google Scholar]
  45. Reynolds, H. T. (1977). The analysis of cross‐classifications. New York, NY: Free Press. [Google Scholar]
  46. Robinson, D. H. , Whittaker, T. A. , Williams, N. J. , & Beretvas, S. N. (2003). It's not effect sizes so much as comments about their magnitude that mislead readers. The Journal of Experimental Education, 72(1), 51–64. [Google Scholar]
  47. Rouder, J. N. , Morey, R. D. , Speckman, P. L. , & Province, J. M. (2012). Default Bayes factors for ANOVA designs. Journal of Mathematical Psychology, 56, 356–374. [Google Scholar]
  48. Sechrest, L. , & Yeaton, W. H. (1982). Magnitudes of experimental effects in social science research. Evaluation Review, 6(5), 579–600. [Google Scholar]
  49. Sherman, J. , & Morrison, W. J. (1950). Adjustment of an inverse matrix corresponding to a change in one element of a given matrix. Annals of Mathematical Statistics, 21(1), 124–127. [Google Scholar]
  50. Shih, M. , Pittinsky, T. L. , & Ambady, N. (1999). Stereotype susceptibility: Identity salience and shifts in quantitative performance. Psychological Science, 10, 80–83. [DOI] [PubMed] [Google Scholar]
  51. Tate, R. F. (1954). Correlation between a discrete and a continuous variable: Point‐biserial correlation. The Annals of Mathematical Statistics, 25(3), 603–607. [Google Scholar]
  52. Tate, R. F. (1966). Conditional‐normal regression models. Journal of the American Statistical Association, 61(314), 477–489. [Google Scholar]
  53. Tiwari, R. C. , Chib, S. , & Jammalamadaka, S. (1989). Bayes estimation of the multiple correlation coefficient. Communications in Statistics – Theory and Methods, 18(4), 1401–1413. [Google Scholar]
  54. Wagenmakers, E.‐J. , Morey, R. D. , & Lee, M. D. (2016). Bayesian benefits for the pragmatic researcher. Current Directions in Psychological Science, 25(3), 169–176. [Google Scholar]
  55. Žeželj, I. L. , & Jokić, B. R. (2014). Replication of experiments evaluating impact of psychological distance on moral judgment (Eyal, Liberman & Trope, 2008; Gong & Medin, 2012). Social Psychology, 45(3), 223–231. [Google Scholar]

Articles from Statistica Neerlandica are provided here courtesy of Wiley

RESOURCES