Skip to main content
eNeuro logoLink to eNeuro
. 2019 Aug 1;6(4):ENEURO.0205-19.2019. doi: 10.1523/ENEURO.0205-19.2019

Estimation for Better Inference in Neuroscience

Robert J Calin-Jageman 1,, Geoff Cumming 2
PMCID: PMC6709209  PMID: 31453316

Abstract

The estimation approach to inference emphasizes reporting effect sizes with expressions of uncertainty (interval estimates). In this perspective we explain the estimation approach and describe how it can help nudge neuroscientists toward a more productive research cycle by fostering better planning, more thoughtful interpretation, and more balanced evaluation of evidence.

Keywords: estimation, neuroscience methods, statistical inference

Significance Statement

The estimation approach to inference emphasizes reporting effect sizes with expressions of uncertainty (interval estimates). The estimation approach can serve as an adjuvant towards better inference: it pushes back against over-confident claims from inadequate samples, improves comparisons of results across contexts, normalizes the publication of negligible effects, and provides a straightforward approach for planning informative studies.

Estimation for better inference in neuroscience

Inference is at the heart of the scientific method: we collect finite datasets and then try to make reasonable generalizations about how the world works. Today eNeuro announces new author guidelines for statistical inference, enjoining the use of estimation along with or in place of hypothesis testing. Specifically, the guidelines ask authors to:

  • Pose quantitative research questions and report quantitative answers (effect sizes).

  • Countenance uncertainty in all statistical conclusions by reporting and interpreting the potential for error (interval estimates).

In some ways, this is a very subtle change in policy. The estimation approach to inference is based on the same mathematical foundations as hypothesis testing and still enables decision-making. Moreover, adopting estimation does not limit analysis options, as any hypothesis test (frequentist, Bayesian, bootstrap, etc.) can be re-expressed in terms of estimation. Why change, then? Because current norms for statistical inference are often misguided, leading to research that is wasteful, biased, and unreliable. The estimation approach can serve as an adjuvant toward better inference: it pushes back against over-confident claims from inadequate samples, improves comparisons of results across contexts, normalizes the publication of negligible effects, and provides a straightforward approach for planning informative studies. Estimation does not cure all ills, but this new policy can serve a vital role in eNeuro’s forward-looking efforts to promote strong rigor and reproducibility without sacrificing scientific vitality (Bernard, 2016).

In this commentary we give an overview of the estimation approach. We then give specific examples of how it can foster better inference. We conclude with some important caveats and clarifications and a list of resources that can help researchers make the transition to the estimation approach. Parts of this commentary are adapted from our previous work advocating the estimation approach (Calin-Jageman, 2017; Calin-Jageman and Cumming, 2019).

The estimation approach

Currently most neuroscientists approach inference through null-hypothesis significance testing. In this approach, we ask a qualitative question: Does this drug influence learning? The data collected is then reduced down to a test statistic and p value, and from these we make a qualitative conclusion: Yes, this drug influences learning. Results are often treated as definitive, so there can be little motivation to conduct replications: Why test the drug again now that it has been shown to work? Although this “one-and-done” approach is common, it is not the way null-hypothesis significance testing was meant to be used. For example, Fisher argued that “a scientific fact should be regarded as experimentally established only if a properly designed experiment rarely fails to give this level of significance” (Fisher, 1926 , p. 85).

Although the testing approach now dominates the neurosciences (Szucs and Ioannidis, 2017a), this was not always the case. Much of the most enduring and fruitful research in neuroscience was completed without recourse to p values (Hodgkin and Huxley, 1952; Olds and Milner, 1954; Scoville and Milner, 1957; Katz and Miledi, 1968; Bliss and Lomo, 1973; Sherrington et al., 2003). In fact, hypothesis testing is a special case of a broader and older statistical tradition: estimation. Estimation is sometimes called the “New Statistics” (Cumming, 2014), because adopting this approach would be new to the many scientists who have been trained only in the testing approach. To be clear, though, estimation is not a new approach to inference, just a different application of the same thinking underlying the testing approach.

The estimation approach involves a shift in how research questions are conceptualized and reported. Rather than a qualitative question, we pose a quantitative question: How much does this drug influence learning? This is answered with an effect size, which gives a quantitative answer to the research question, and an interval estimate, which helps express uncertainty: The drug improved memory by 10% with a 95% margin of error of 9%. The effect size expresses the magnitude of difference observed in the sample (10%). The interval estimate expresses some of the uncertainty in generalizing to the population. This can be expressed as an expected magnitude of error (95% margin of error of 9%) or as an interval around the effect size [95% confidence interval (1%, 19%)]. Because some effect sizes have asymmetric expected error, reporting an interval is usually preferred. Regardless of format, an interval estimate expresses some of the uncertainty in generalizing to the population. In this example, the interval estimate indicates a need to countenance a wide range of possible effect sizes (values ∼1% are compatible with the data and so are values ∼19%).

There are different approaches to constructing interval estimates. Every null-hypothesis test with a p value has a corresponding confidence interval that can be reported and interpreted (p values and confidence intervals both represent what are called frequentist approaches to statistical inference). In Bayesian statistics uncertainty can be expressed with a credible interval (Kruschke, 2013) or support interval (Wagenmakers et al., 2018b). There are also randomization-based approaches to quantifying uncertainty, such as bootstrapped intervals (Ho et al., 2018). Moreover, estimates can be constructed based on specific assumptions about population distributions (parametric) or with very minimal assumptions (nonparametric). Each approach has strengths and weaknesses; researchers should quantify uncertainty in the way that best suits their research purposes and then take care to provide interpretations properly grounded in the approach they have selected.

Although there is diversity within the estimation approach, the common theme is an emphasis on uncertainty, a key aspect of good statistical practice (Wasserstein et al., 2019). Focusing on uncertainty makes salient the tentative nature of any one study and highlights the need for direct replication. Replications can then be synthesized through meta-analysis, fostering cumulative science.

Estimation in action

To make the contrast between testing and estimation concrete, let us summarize the same data with both approaches (Fig. 1). Consider a recent report in Nature Neuroscience examining the effect of caffeine on memory (Borota et al., 2014). Participants studied images of objects and then received either 200 mg of caffeine (n = 35) or a placebo (n = 38). The next day, memory was evaluated. In the original report results were summarized using hypothesis testing, with a t test indicating a statistically significant enhancement in memory in the 200 mg caffeine group relative to the placebo group: t(71) = 2.0, p = 0.049. From this, a qualitative conclusion was drawn: “caffeine administration enhances memory consolidation in humans” (Borota et al., 2014, p. 21). Figure 1A depicts how such data are often presented: a bar graph with error bars representing SEs and an asterisk to denote a statistically significant difference (in the original paper, data from each participant was also shown).

Figure 1.

Figure 1.

Visualizations emphasizing testing versus estimation. Both plots are from the same experiment examining the effect of caffeine on memory (Borota et al., 2014). A, A traditional bar graph. The bars represent each group mean; the error bars represent the SEM. The * indicates a statistically significant difference, p = 0.05. B, An estimation plot of the same data. In this plot the small circles represent the individual participants. The large circles with error bars represent each group mean with their 95% confidence intervals. Critically, an estimation plot emphasizes the effect size of interest for this design: the difference between the group means. This is depicted on the “difference axis” on the right. The 0 point of this axis is based on the mean of a reference group (in this case the placebo group). The filled triangle shows the difference between groups in this sample. The shaded curve shows the entire distribution of expected sampling error for the difference between the means. The error bar on the triangle indicates the 95% confidence interval for the difference between means. The confidence interval represents the range of parameter values which remain compatible with the data; that is, the variety of effect sizes that are not rejected at α = 0.05. The range of compatible values is very long and includes values that would be impossible to reliably detect with feasible sample sizes. Given this, research conclusions should be tentative and expectations for replication should be tempered. This difference plot was generated using R. The data from A and B was extracted from Borota et al. (2014).

The estimation approach provides a different lens for interpreting the same data. In this approach we would ask a quantitative question: To what extent does caffeine improve memory? To answer this question we estimate the difference between the group means and then quantify uncertainty in this estimate due to expected sampling error: Caffeine is estimated to improve memory relative to the placebo group by 31% with a 95% confidence interval of (0.2%, 62%). Figure 1B graphically represents this information with an estimation plot that shows the effect size in the sample and the confidence interval for the effect. Critically, the confidence interval suggests considerable uncertainty about generalizing from the sample to the world at large. If the real effect were very large (62% increase), these data would not be especially surprising. Similarly, if the real effect were vanishingly small (0.2%), these data would not be especially surprising. Moreover, this wide range of possibilities is optimistic, as it is based on uncertainty due only to sampling error with the assumption that all other sources of error and bias are negligible.

Reflecting on this uncertainty makes it clear that although this study is statistically significant the sample collected is not adequate, yielding poor signal-to-noise (the margin of error is only fractionally smaller than the observed effect). In the testing approach this is classified as low power. A more illuminating label is uninformative; the study yields too much uncertainty to provide a clear answer to the research question. The most appropriate interpretation at this point would be very modest: caffeine probably does not impair memory consolidation.

This striking reappraisal of the caffeine and memory study does not come from changing our epistemic standards or the statistical model used to analyze the data. The 95% confidence interval for the caffeine and memory study is just an algebraic re-expression of the t test used in the original paper. Specifically, the 95% confidence interval contains all the parameter values that would not be rejected using a frequentist hypothesis test at the α = 0.05 level. All values outside of the confidence interval are parameter values that are rejected at α = 0.05. The comparison between caffeine and placebo is “statistically significant” at the 0.05 level because the null hypothesis of 0 is (just barely) outside of the 95% confidence interval. Reporting the statistical test tells us that we reject this one possible parameter value. It seems clear that we can do better science by thinking critically about the parameter values that remain compatible with the data. It is this range of values that should inform our assessment of practical significance, our theory, and our planning for subsequent experiments. Thus, the new policy at eNeuro is to report and interpret interval estimates either in place of or alongside hypothesis tests.

In making this change, eNeuro joins good company at the forefront of good statistical practice. In many fields of medicine there has long been an emphasis on estimation (International Committee of Medical Journal Editors, 1997). In the behavioral sciences the American Psychological Association (2010, p. 34) enjoins researchers to “base discussion and interpretation of results on point and interval estimates” (for review, see Fidler, 2010), and interval estimates are now reported in most papers published in top psychology journals (Giofrè et al., 2017).

Estimation thinking for an improved research cycle

Adopting estimation has several advantages (Cumming and Finch, 2001):

  • Focusing on effect sizes dovetails seamlessly with the development of quantitative theories and computational models.

  • Effect size estimates can be synthesized through meta-analysis, fostering cumulative science.

  • Estimation is easier to understand than testing, which is really a special case of estimation (Hogben, 1957, p. 320). Teaching estimation first can help trainees better understand the uses and limits of the testing approach.

  • Estimates are a natural choice for dissemination and communication with stakeholders. Modern journalistic standards specifically emphasize conveying magnitudes and uncertainty (https://www.healthnewsreview.org/).

Realizing these benefits requires more than just a rote change in how analyses are reported; it requires a different way of thinking about data that puts uncertainty at the forefront. This is critical because current norms for inference in neuroscience license problematic research practices. Neuroscience studies are often too small, yielding noisy results that are relatively uninformative ( aka, low-power research; Button et al., 2013; Dumas-Mallet et al., 2017; Szucs and Ioannidis, 2017b ). Somehow, though, this is not reflected in the public record: although noisy studies should rarely be able to detect effects, nearly all published papers report statistically significant results (Fanelli, 2012). This implausible “excess of significance” indicates that many unfavorable results have been discarded to the file drawer (Sterling, 1959; Sterling et al., 1995) or inappropriately coaxed under the threshold for statistical significance (Simmons et al., 2011; O’Boyle et al., 2017). Moreover, when statistical significance is obtained there is often insufficient attention to uncertainty, so interpretations are too confident and uninformative sample sizes are then copied forward, unwittingly perpetuating flawed research strategies. A system that is noisy and biased cannot be expected to yield reliable information, and indeed replicability seem to be low where these problems are prevalent (Boekel et al., 2015).

Here we discuss four key ways that estimation thinking can improve the neuroscience research cycle: (1) by better calibrating research conclusions to uncertainty, (2) by fostering planning and optimization toward generative lines of research, (3) by normalizing the use of inference to both rule in and rule out effects, and (4) by facilitating accurate comparison of results.

To make this discussion concrete we draw on a recent hot topic in neuroscience: the effects of intranasal oxytocin on trust and other social behaviors in humans. This topic of inquiry was initiated by a pair of prominent studies published in 2005 (Kosfeld et al., 2005; Zak et al., 2005). Since then, there has been an explosion of preclinical research and translational research on oxytocin and human social behavior. There are now grave concerns, however, about the way this research has been conducted and reported (Conlisk, 2011; Nave et al., 2015; Leng and Ludwig, 2016): studies have been too small, direct replication has been too rare, and negative results have been relegated to the file drawer. These concerns are so severe that a recent review concluded that the entire literature up to that point should be viewed with “healthy skepticism” because most published effects are probably spurious (Walum et al., 2016, p. 251). The tide now seems to be turning, with larger studies and direct replication helping to yield more clear and certain insight into how oxytocin might influence human social behavior (Liu et al., 2019). Looking back at the problems in this literature, though, brings into sharp relief fruitful avenues for improving the research cycle in neuroscience. The shortcomings we highlight are not atypical; there are several active lines of neuroscience research where selective reporting of noisy studies seems prevalent ( research on tDCS, for example; Horvath et al., 2015; Minarik et al., 2016 ; Medina and Cason, 2017).

Estimation helps calibrate research conclusions to uncertainty

The estimation approach helps guide interpretation of data: an interval estimate that is long relative to the scale of measurement requires cautious and tentative conclusions, whereas an interval estimate that is short can warrant stronger claims. This close calibration between uncertainty and interpretation is essential: it helps match research claims to the evidence, makes clear when additional direct replication is needed, and helps set realistic expectations for subsequent research. In contrast, researchers using the testing approach often base their conclusions solely on statistical significance, treating every statistically significant result as equally and completely compelling. The uncertainty blindness that can occur when p < 0.05 licenses unequivocal claims for research that should be treated as highly tentative and generates excessive confidence in the likelihood of replication (this has been called “the replication delusion”; Gigerenzer, 2018).

As an example, consider one of the studies that helped launch research on intranasal oxytocin and human trust (Kosfeld et al., 2005). In this experiment participants received an intranasal dose of oxytocin (n = 29) or placebo (n = 29) and then played an economic trust game. There was a statistically significant effect of oxytocin (t(56) = 1.8, p = 0.04, one-tailed; Fig. 2A, Trust context; see note at end on how these data were analyzed). In addition, there was a nonsignificant effect of oxytocin in a game involving only risk (p = 0.98; Fig. 2A, Risk context). From these results researchers made a categorical and unequivocal claim: “Oxytocin increases trust in humans” (Kosfeld et al., 2005, p. 673). Although we can all be partial to our own data, other scientists seem to have agreed with this sweeping interpretation. The study was published in Nature and quickly became a citation classic (>3000 citations as of January 2019, according to Google scholar). Even from the start, citations portrayed this study as unequivocal (Coan et al., 2006).

Figure 2.

Figure 2.

Visualizations emphasizing testing versus estimation. Both plots are from the same experiments examining the effect of oxytocin on social behavior (Kosfeld et al., 2005). A, A bar graph showing the effect of intranasal oxytocin on the amount invested in a trust (left) and risk (right) game. Each bar represents group means and error bars represent ± 1 SE. There is a statistically significant effect of oxytocin in the trust game (p = 0.04, on tailed), but not in the risk game. The difference in statistical significance status can give the impression that oxytocin specifically influences performance in the trust game, but a formal test for an interaction is not significant (p = 0.23). B, Estimation plots of the same data. The plots compare investment in the oxytocin and investment condition in the trust (left) and risk (right) games. Small circles represent individual participants. Large circles with error bars show group means with 90% confidence intervals (90% confidence was selected to match the stringency of the one-tailed test used in the original paper). The triangles represent the observed difference between groups, with 90% confidence intervals. The curves indicate the entire range of expected sampling error in estimating the mean difference. Note the considerable overlap in the estimated oxytocin effects in the trust and risk games. This correctly suggests that there are many compatible effect sizes in common and that these data do not support strong claims for an interaction. This figure was adapted with permission from Calin-Jageman and Cumming (2019).

This trajectory from a single significant result to widespread acceptance of a categorical claim is the norm in our field, and yet it often represents an egregious insensitivity to uncertainty. This can be appreciated by re-expressing the statistical test used in the oxytocin study as an estimate: mean trust increased by 17.4%, 90% CI (0.5%, 34.2%; Fig. 2B, left; a 90% CI is used to match the stringency of the one-tailed test used by Kosfeld et al., 2005). Summarized in this way, it is clear that this study is uninformative, with a sample size too small to support a clear answer to the research question. Although the data are compatible with oxytocin producing large changes in trust, they are also compatible with oxytocin producing infinitesimal changes of no practical significance. An appropriately cautious interpretation from these data alone would be that oxytocin likely does not impair trust in humans.

Focusing on uncertainty in current results is important because it helps set realistic expectations for future studies. Many researchers erroneously believe that statistical significance means “likely to replicate” (Gigerenzer, 2018). The estimation approach helps push back against this misconception, as an interval estimate can also serve as a prediction interval for what to expect with a direct replication. For example, a 95% confidence interval for a mean difference will “capture” the effect size of ∼83% of same-sized replications (Cumming and Maillardet, 2006; Cumming, 2008). That is, it is expected that 83% of the time the replication effect size will be within the original 95% confidence interval. Note that the capture rate is not 95% because both studies are subject to sampling error. Moreover, this expectation is optimistic, as it assumes the original and replication studies will differ only in terms of sampling error.

Examining the interval estimate from Kosfeld et al. (2005) gives a wide range of predicted outcomes for replication studies, including effect sizes that would be practically impossible to detect. Specifically, near the lower bound of the 95% confidence interval are effect sizes that would require many thousands of participants/group to regularly detect. Consistent with this prediction, replications of Kosfeld et al. (2005) have so far obtained primarily negligible effects. A recent meta-analysis suggests the effect is small, perhaps even exactly 0 [in SD units: 0.08, 95% CI (−0.12, 0.28); Nave et al., 2015]. Whereas a focus on statistical significance might suggest this disappointing outcome is surprising (Lai et al., 2012), a focus on estimation correctly shows that this is actually an expected possible outcome, one that is within the interval estimate from the original study [in SD units, Kosfeld et al. (2005) found an oxytocin effect in the trust context of 0.49, 90% CI (0.04, 0.93)]. This is perhaps the most useful aspect of the estimation approach: it can help properly calibrate our sense of surprise across a series of results.

Strong claims from uninformative samples are not unique to the oxytocin literature. In fact, it is remarkably common for neuroscientists to unwittingly conduct and cite research that is actually uninformative ( aka under-powered; Button et al., 2013; Dumas-Mallet et al., 2017; Szucs and Ioannidis, 2017b). This means that much of what currently passes for “established” should rightly be construed as highly tentative. Reporting interval estimates helps makes more clear when a study is too uncertain to be informative. This does not have to preclude publication; sometimes an inadequate sample is unavoidable and we must make do with highly uncertain results. The key, though, is that reporting and interpreting estimates will raise needed red flags, ensuring (1) that research conclusions will be appropriately tentative, (2) that the need for direct replication with larger samples will be clear, and (3) that expectations for replications will be appropriately broad.

Estimation fosters thoughtful research planning and optimization

Estimation is useful not only for interpreting completed studies but also for thoughtfully planning and optimizing the next study. First, estimation focuses on effect sizes and uncertainty, the inputs needed for planning samples. Second, estimation offers an intuitive approach to sample-size planning: planning for precision (Goodman, 1994; Kelley et al., 2003; Rothman and Greenland, 2018). In planning for precision (also known as the “Accuracy in Parameter Estimation” approach), researchers plan a sample size to obtain a desired level of precision (a desired margin of error). Unlike planning for power, planning for precision does not require a priori effect size expectations (though these are still helpful, if available). In addition, planning for precision plans to characterize the effect, not just to detect it, which means even if the effect is negligible the results are still informative and publishable. Planning for precision is amenable to sequential analysis (Kelley et al., 2018), so researchers can efficiently obtain a desired level of precision even in exploratory research where there is considerable uncertainty about variance in the dependent variable. The ease of planning for precision can help researchers meet their ethical obligation to avoid collecting both too little and too much data (ASA, 2016), ensuring research efforts are, to the extent possible, neither futile nor wasteful.

Forethought before initiating a study can often lead to sample-size sticker shock: obtaining an informative answer to a research question can require sample sizes that are not feasible. This is where optimization comes in: protocols can be tweaked to maximize effect sizes and minimize noise (Kraemer, 1991; MacKinnon, 2013; Meyvis and Van Osselaer, 2018). Optimization is a natural step when a laboratory’s focus is on these critical experimental outputs. In contrast, p values are generally too erratic (Cumming, 2008; Halsey et al., 2015) to guide optimization efforts.

Better planning is sorely needed in the neurosciences. Preclinical research seems to regularly proceed without an a priori sampling plan (Fritz et al., 2013; Tressoldi et al., 2013; Baker et al., 2014; Vankov et al., 2014; Tressoldi and Giofré, 2015). This neglect of best practices seems driven, in part, by a fundamental misconception that attaining statistical significance proves an adequate sample has been obtained ( Mole, 2017, 2018 ). From this flawed premise, researchers often feel comfortable setting sample sizes by following tradition or by chasing significance (Vankov et al., 2014; Goodhill, 2017). In reality, both of these approaches are problematic. Given the prevalence of uninformative research in neuroscience, relying on tradition risks copying forward the mistakes of the past ad infinitum. Even worse “backing into” a sample size by iteratively collecting data to obtain statistical significance provides repeated opportunities to capitalize on chance, decreasing the reliability of the results obtained (Anscombe, 1954; Simmons et al., 2011). Both of these problematic approaches to planning seem common in preclinical research, and this perpetuates the use of uninformative samples.

The oxytocin and human social behavior literature provides an acute illustration of poor planning. The study by Kosfeld et al. (2005) was one of the first to examine the effects of intranasal oxytocin on human social behavior. In this novel context it is not surprising that the sample size obtained was not well calibrated to the research question. What is surprising is that this issue was not widely recognized. Because of this, sample sizes from this exploratory work were copied forward for what should have been confirmatory work, plaguing the entire field with very poor signal-to-noise. Specifically, a recent meta-analysis found that the average published effect of oxytocin on human social behavior is fairly modest, ∼0.28 SD (Walum et al., 2016). Despite this, median sample size in this field is only 49 total participants, meaning expected sampling error is much larger (∼0.55 SD) than the typical reported effect. This is like trying to study ion channel structure with a magnifying glass: it does not mean that all the results are wrong, just that there is relatively little reason to believe them. It also means that most of these studies were launched without reasonable forethought, producing research conducted at considerable time and expense, but to little purpose.

Estimation can mitigate publication bias

With the estimation approach, an interval can provide evidence that an effect is meaningful (the whole interval estimate is in a range of practical significance). An estimate can also provide evidence that an effect is negligible (the whole interval estimate is in a range that is not practically significant). Of course, what counts as a meaningful effect size depends on the research context and requires judgment. What is critical is that with estimation thinking both outcomes are evaluated similarly: with thoughtful attention to uncertainty and careful consideration of factors that could bias the estimate (e.g., insufficient manipulation, experimenter bias, differential dropout, procedural error, etc.).

This even-handed weighing of evidence is essential to good science, where our analytic procedures must be capable of both ruling in and ruling out effects (and of reserving judgment due to an uninformative sample). The testing approach is also suitable for these needs. In practice, though, null hypothesis testing with p values is often used as though it can only demonstrate effects. This is like having a neural network that can only express LTP; noise will eventually saturate the system. This one-sided approach to research is due in part to incomplete training. Current training rarely includes exposure to Bayesian techniques or equivalence testing, the p value approach to testing for a negligible effect (Westlake, 1972; Shuirmann, 1987; Lakens et al., 2018). Moreover, half-truths are perpetuated, with trainees strongly cautioned that nonsignificant results may be too uncertain to interpret or merely an indicator of researcher incompetency. These are half-truths because they present a false specificity: these cautions apply to all research results. That is, significant findings can also be too uncertain to interpret (uninformative), and incompetency can produce spurious effects just as easily as it can obscure real ones. These widespread misunderstandings of the testing framework help fuel publication bias, yielding a distorted published literature and a “vast graveyard of undead theories” (Ferguson and Heene, 2012).

Again, the literature on intranasal oxytocin and human social behavior provides a cautionary example. Meta-analysis shows excess significance in the published literature on this topic (Walum et al., 2016), a sure tell that many nonsignificant findings have not been published or have been massaged toward statistical significance. Indeed, one laboratory has bravely opened its file drawer for inspection (Lane et al., 2016), reporting that the five statistically significant results it has published represent just 39% of the 13 different tests the laboratory had conducted. The laboratory had tried to publish the eight nonsignificant results, but these were “rejected time and time again” (Lane et al., 2016, p. 38). Thus, the published output of this laboratory would indicate a strong effect of oxytocin, but weighing all the data the laboratory has collected indicates only a negligible effect (ibid). Again, this is not atypical; excess significance has been detected across many domains of neuroscience research (Button et al., 2013), and there is an enormous body of literature showing that nonsignificant results are far less likely to be written up, submitted, and/or published than significant results (Song et al., 2010). A research cycle that suppresses unfavorable data cannot rightly be described as scientific.

Estimation facilitates accurate comparisons across results

Another important role of inference is in making comparisons to other contexts, conditions, or studies; this is the analysis of interactions. Within the testing approach, neuroscientists often fail to conduct a formal test for an interaction but instead rely on comparing statistical significance levels (Nieuwenhuis et al., 2011). This is invalid and frequently leads to spurious conclusions (Gelman and Stern, 2006). Two results can have the same effect size but differ in significance (e.g., due to different sample sizes). In addition, two results can be statistically significant and yet differ radically in effect size. The estimation approach helps researchers avoid this inferential trap and fosters accurate comparisons across sets of results.

For example, in the seminal experiment on intranasal oxytocin and trust (Kosfeld et al., 2005), researchers examined the effect of oxytocin on a game involving trust and on a control game involving only risk (Fig. 2A). Whereas oxytocin had a statistically significant effect on money transfer in the trust game (p = 0.04, one-tailed) it did not have a statistically significant effect in the no-trust game (p = 0.98). This suggests a possible interaction between oxytocin and trust, but this was not formally tested. Instead, the researchers relied on the difference in significance status to conclude that “oxytocin specifically affects trust in interpersonal interactions” (Kosfeld et al., 2005, p. 674). This is an invalid conclusion; formally testing for the interaction gives a nonsignificant result: p = 0.23. Within the testing framework, these data do not provide clear support of a claim for specificity.

When expressed only in terms of statistical significance, the analytic error in Kosfeld et al. (2005) is difficult to detect; it seems to have gone unnoticed by the researchers, reviewers, and numerous readers. In contrast, summarizing results with estimation makes it easier to accurately compare sets of results. For the trust experiment, oxytocin increased mean investment by $1.41, 90% CI [($0.04, $2.78); Fig. 2B, left]. In the non-trust experiment, oxytocin produced effectively no increase in mean investment: $0.01, 90%CI [(−$1.32, $1.35); Fig. 2B, right]. Although in the sample these are markedly different outcomes, there is substantial overlap in the interval estimates. This makes it clear “by eye” that there are many compatible effect sizes in common (Cumming and Finch, 2005), giving an intuitive sense that evidence for specificity is weak. To formally test for the interaction, we estimate the “difference in the difference”; the difference between each simple effect. Specifically, the oxytocin effect during the trust game ($1.41 increase) is compared with the oxytocin effect during the non-trust game ($0.01 increase), providing an estimated interaction of $1.40 90% CI (−$.52, $3.31). That is, in the sample a trust context strongly enhanced the oxytocin effect, but the data are also compatible with no interaction and even with a moderate enhancement in the non-trust context.

This analysis of the “difference in the difference” is just a quantitative way of expressing the interaction term in a 2 × 2 ANOVA. Arguably, though, the estimation approach is more transparent and easier to interpret. This clarity would be especially useful for neuroscience. Estimates of differences in results (interactions) have higher expected sampling error than estimates of simple effects, meaning that they are more likely to be uninformative (underpowered). Thus, estimation can help encourage formally correct comparisons across results in a way that is intuitive and sensitive to uncertainty.

Some important clarifications and caveats

Estimation can help improve every aspect of the neuroscience research cycle: helping us more carefully plan, more thoughtfully interpret, more accurately compare, and more completely report neuroscience research.

  • This new policy does not ban hypothesis testing; these may be reported alongside interval estimates. The use of test procedures that allow results to be judged negligible as well as meaningful are preferred (e.g., Bayesian approaches and equivalence tests; Westlake, 1972; Shuirmann, 1987; Lakens et al., 2018 ). Note that in most cases reporting an interval estimate already provides the information that would be conveyed in a hypothesis test.

  • Estimates should not be used as a surrogate for hypothesis testing (e.g., mindlessly checking whether the null value is contained inside or outside the interval estimate; Fidler et al., 2004). Authors should thoughtfully evaluate interval estimates and carefully calibrate research conclusions with respect to uncertainty. Where uncertainty is high, make clear the need for replication studies with greater precision.

  • Estimates must be aligned to the research question. One place to beware is with complex designs. These are often analyzed with an ANOVA and reported with a focus on a single omnibus F test with a form of η2 as the effect size. Omnibus tests rarely correspond to the research questions of interest, which are typically tested with a series of planned contrasts. The magnitude and uncertainty of these planned contrasts will usually be of critical interest.

  • Estimation is not a panacea. It helps highlight uncertainty but it does not overcome the problems that arise with selective reporting, flexible analysis, poor model specification, etc. Moreover, statistical outputs are not the only factor in generating scientific conclusions; this also requires careful attention to the design of the study, quality of the measurement, prior knowledge, and more (McShane et al., 2019).

  • Using estimation does not alter the need to clearly demarcate planned analyses from exploratory analyses. Preregistration can help make this distinction publicly verifiable.

  • Selection of confidence levels (95%, 99%, etc.) should not be rote, but should be based on an evaluation of the costs/benefits of making erroneous estimates.

  • There is nothing magic about the “ends” of an interval estimate; these are arbitrary relative to the selected level of confidence. Be careful not to draw sharp distinctions between values just inside an interval estimate versus those just outside: the boundaries are arbitrary and differences are matters of degree.

  • Interval estimates are optimistic in that they depend crucially on statistical assumptions, which may not be perfectly realized. This provides an additional reason for not treating the boundaries of interval estimates as definitive.

  • Estimation is for everyone and can be conducted within both the frequentist and Bayesian approaches to inference (Kruschke and Liddell, 2018). In this commentary, we focused on frequentist confidence intervals only because these directly re-express frequentist hypothesis tests, which are so pervasively used in the neuroscience literature.

  • Just as hypothesis testing can be misunderstood and misapplied, estimation can be misunderstood and misapplied. One common misconception applies to frequentist confidence intervals. Specifically, researchers often mistakenly apply the confidence level to their specific result, claiming (erroneously) that there is a 95% probability the interval contains the true value. In fact, for frequentist confidence intervals each specific result either contains the true value or does not. The probability statement applies not to the specific result but to the procedure of interval construction. This mirrors the ambivalence of science, where we can have confidence in the process but remain anxious about each individual study. In general, researchers should strive to make sure they understand the assumptions underlying the statistical approach they have selected and to interpret their results accordingly.

Resources for estimation

Learning about estimation

Software for estimation

Planning for precision

Notes on analysis of data from Kosfeld et al. (2005)

Kosfeld et al. (2005) compared group medians with nonparametric tests. For ease of interpretation we re-analyzed their data using parametric comparisons of means. This change in analysis strategy does not change the conclusions drawn here. Full details on how we extracted the data from Kosfeld et al. (2005) and analyzed it are contained in Calin-Jageman and Cumming (2019). In addition, the extracted data and analysis scripts are posted to https://osf.io/54n9q/.

Synthesis

Reviewing Editor: Christophe Bernard, INSERM & Institut de Neurosciences des Systèmes

Decisions are customarily a result of the Reviewing Editor and the peer reviewers coming together and discussing their recommendations until a consensus is reached. When revisions are invited, a fact-based synthesis statement explaining their decision and outlining what is needed to prepare a revision will be listed below. The following reviewer(s) agreed to reveal their identity: NONE.

Just a minor thing.

We cannot publish footnotes.

Can you include them in the main text?

thanks

References

  1. American Psychological Association (2010) Publication manual of the American Psychological Association. Washington, DC: American Psychological Association. [Google Scholar]
  2. Anscombe FJ (1954) Fixed-sample-size analysis of sequential observations. Biometrics 10:89 10.2307/3001665 [DOI] [Google Scholar]
  3. ASA (2016) Ethical guidelines for statistical practice: American Statistical Association. Available at http://www.amstat.org/asa/files/pdfs/EthicalGuidelines.pdf. [Google Scholar]
  4. Baker D, Lidster K, Sottomayor A, Amor S (2014) Two years later: journals are not yet enforcing the ARRIVE guidelines on reporting standards for pre-clinical animal studies. PLoS Biol 12:e1001756. 10.1371/journal.pbio.1001756 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bernard C (2016) Editorial: scientific rigor or rigor mortis? eNeuro 3:ENEURO.0176-16.2016. 10.1523/ENEURO.0176-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Bliss TV, Lomo T (1973) Long-lasting potentiation of synaptic transmission in the dentate area of the anaesthetized rabbit following stimulation of the perforant path. J Physiol 232:331–356. 10.1113/jphysiol.1973.sp010273 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Boekel W, Wagenmakers EJ, Belay L, Verhagen J, Brown S, Forstmann BU (2015) A purely confirmatory replication study of structural brain-behavior correlations. Cortex 66:115–133. 10.1016/j.cortex.2014.11.019 [DOI] [PubMed] [Google Scholar]
  8. Borota D, Murray E, Keceli G, Chang A, Watabe JM, Ly M, Toscano JP, Yassa MA ( 2014) Post-study caffeine administration enhances memory consolidation in humans. Nat Neurosci 17:201–203. 10.1038/nn.3623 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Button KS, Ioannidis JPA, Mokrysz C, Nosek BA, Flint J, Robinson ESJ, Munafò MR (2013) Power failure: why small sample size undermines the reliability of neuroscience. Nat Rev Neurosci 14:365–376. 10.1038/nrn3475 [DOI] [PubMed] [Google Scholar]
  10. Calin-Jageman RJ (2017) After p values: the new statistics for undergraduate neuroscience education. J Undergrad Neurosci Educ 16:E1–E4. [PMC free article] [PubMed] [Google Scholar]
  11. Calin-Jageman RJ, Cumming G (2019) The new statistics for better science: ask how much, how uncertain, and what else is known. Am Stat 73:271–280. 10.1080/00031305.2018.1518266 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Coan JA, Schaefer HS, Davidson RJ (2006) Lending a hand of the neural response to threat. Psychol Sci 17:1032–1039. 10.1111/j.1467-9280.2006.01832.x [DOI] [PubMed] [Google Scholar]
  13. Conlisk J (2011) Professor Zak’s empirical studies on trust and oxytocin. J Econ Behav Organ 78:160–166. 10.1016/j.jebo.2011.01.002 [DOI] [Google Scholar]
  14. Cumming G (2008) Replication and p intervals. Perspect Psychol Sci 3:286–300. 10.1111/j.1745-6924.2008.00079.x [DOI] [PubMed] [Google Scholar]
  15. Cumming G (2012) Understanding the new statistics: effect sizes, confidence intervals, and meta-analysis. New York: Routledge. [Google Scholar]
  16. Cumming G (2014) The new statistics: why and how. Psychol Sci 25:7–29. 10.1177/0956797613504966 [DOI] [PubMed] [Google Scholar]
  17. Cumming G, Calin-Jageman RJ (2017) Introduction to the new statistics: estimation, open science, and beyond. New York: Routledge. [Google Scholar]
  18. Cumming G, Finch SU (2001) Four reasons to use CIs. Educ Psychol Meas 61:532–574. [Google Scholar]
  19. Cumming G, Finch S (2005) Inference by eye: confidence intervals and how to read pictures of data. Am Psychol 60:170–180. 10.1037/0003-066X.60.2.170 [DOI] [PubMed] [Google Scholar]
  20. Cumming G, Maillardet R (2006) Confidence intervals and replication: where will the next mean fall? Psychol Methods 11:217–227. 10.1037/1082-989X.11.3.217 [DOI] [PubMed] [Google Scholar]
  21. Dumas-Mallet E, Button KS, Boraud T, Gonon F, Munafò MR (2017) Low statistical power in biomedical science: a review of three human research domains. R Soc Open Sci 4:160254. 10.1098/rsos.160254 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Ellis PD (2010) The essential guide to effect sizes. New York: Cambridge UP. [Google Scholar]
  23. Fanelli D (2012) Negative results are disappearing from most disciplines and countries. Scientometrics 90:891–904. 10.1007/s11192-011-0494-7 [DOI] [Google Scholar]
  24. Ferguson CJ, Heene M (2012) A vast graveyard of undead theories: publication bias and psychological science’s aversion to the null. Perspect Psychol Sci 7:555–561. 10.1177/1745691612459059 [DOI] [PubMed] [Google Scholar]
  25. Fidler F (2010) The American Psychological Association publication manual sixth edition: implications for statistics education. Proceedings of the Eighth International Conference on Teaching Statistics (ICOTS8, July, 2010), Ljubljana, Slovenia. Voorburg, The Netherlands: International Statistical Institute.
  26. Fidler F, Thomason N, Cumming G, Finch S, Leeman J (2004) Editors can lead researchers to confidence intervals, but can’t make them think: statistical reform lessons from medicine. Psychol Sci 15:119–126. 10.1111/j.0963-7214.2004.01502008.x [DOI] [PubMed] [Google Scholar]
  27. Fisher RA (1926) The arrangement of field experiments. J Minist Agric Gt Britain 33:503–513. [Google Scholar]
  28. Fritz A, Scherndl T, Kühberger A (2013) A comprehensive review of reporting practices in psychological journals: are effect sizes really enough? Theory Psychol 23:98–122. 10.1177/0959354312436870 [DOI] [Google Scholar]
  29. Gelman A, Stern H (2006) The difference between “significant” and “not significant” is not itself statistically significant. Am Stat 60:328–331. 10.1198/000313006X152649 [DOI] [Google Scholar]
  30. Gigerenzer G (2018) Statistical rituals: the replication delusion and how we got there. Adv Methods Pract Psychol Sci 1:198–218. 10.1177/2515245918771329 [DOI] [Google Scholar]
  31. Giofrè D, Cumming G, Fresc L, Boedker I, Tressoldi P (2017) The influence of journal submission guidelines on authors’ reporting of statistics and use of open research practices. PLoS One 12:e0175583. 10.1371/journal.pone.0175583 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Goodhill GJ (2017) Is neuroscience facing up to statistical power? arXiv:1701.01219. Available at http://arxiv.org/abs/1701.01219. [Google Scholar]
  33. Goodman SN (1994) The use of predicted confidence intervals when planning experiments and the misuse of power when interpreting results. Ann Intern Med 121:200. 10.7326/0003-4819-121-3-199408010-00008 [DOI] [PubMed] [Google Scholar]
  34. Halsey LG, Curran-Everett D, Vowler SL, Drummond GB (2015) The fickle p value generates irreproducible results. Nat Methods 12:179–185. 10.1038/nmeth.3288 [DOI] [PubMed] [Google Scholar]
  35. Ho J, Tumkaya T, Aryal S, Choi H, Claridge-Chang A (2018) Moving beyond p values: everyday data analysis with estimation plots. bioRxiv. Advance online publication. Retrieved July 26, 2018. doi: https://doi.org/10.1101/377978. [Google Scholar]
  36. Hodgkin AL, Huxley AF (1952) A quantitative description of membrane current and its application to conduction and excitation in nerve. J Physiol 117:500–544. 10.1113/jphysiol.1952.sp004764 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Hogben LT (1957) Statistical theory: the relationship of probability, credibility, and error. New York: Norton. [Google Scholar]
  38. Horvath JC, Forte JD, Carter O (2015) Quantitative review finds no evidence of cognitive effects in healthy populations from single-session transcranial direct current stimulation (tDCS). Brain Stimul 8:535–550. 10.1016/j.brs.2015.01.400 [DOI] [PubMed] [Google Scholar]
  39. International Committee of Medical Journal Editors (1997) Uniform requirements for manuscripts submitted to biomedical journals. N Engl J Med 336:309–316. [DOI] [PubMed] [Google Scholar]
  40. Katz B, Miledi R (1968) The role of calcium in neuromuscular facilitation. J Physiol 195:481–492. 10.1113/jphysiol.1968.sp008469 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Kelley K (2007) Methods for the behavioral, educational, and social sciences: an R package. Behav Res Methods 39:979–984. [DOI] [PubMed] [Google Scholar]
  42. Kelley K, Darku FB, Chattopadhyay B (2018) Accuracy in parameter estimation for a general class of effect sizes: a sequential approach. Psychol Methods 23:226–243. 10.1037/met0000127 [DOI] [PubMed] [Google Scholar]
  43. Kelley K, Maxwell SE, Rausch JR (2003) Obtaining power or obtaining precision. Eval Health Prof 26:258–287. 10.1177/0163278703255242 [DOI] [PubMed] [Google Scholar]
  44. Kline RB (2004) Beyond significance testing. Washington, DC: American Psychological Association. [Google Scholar]
  45. Kosfeld M, Heinrichs M, Zak PJ, Fischbacher U, Fehr E (2005) Oxytocin increases trust in humans. Nature 435:673–676. 10.1038/nature03701 [DOI] [PubMed] [Google Scholar]
  46. Kraemer HC (1991) To increase power in randomized clinical trials without increasing sample size. Psychopharmacol Bull 27:217–224. [PubMed] [Google Scholar]
  47. Kruschke JK (2014) Doing Bayesian data analysis: a tutorial with R, JAGS, and Stan, Ed 2. Amsterdam; London: Elsevier; . [Google Scholar]
  48. Kruschke JK (2013) Bayesian estimation supersedes the t test. J Exp Psychol Gen 142:573–603. 10.1037/a0029146 [DOI] [PubMed] [Google Scholar]
  49. Kruschke JK, Liddell TM (2018) The Bayesian new statistics: hypothesis testing, estimation, meta-analysis, and power analysis from a Bayesian perspective. Psychon Bull Rev 25:178–206. 10.3758/s13423-016-1221-4 [DOI] [PubMed] [Google Scholar]
  50. Lai J, Fidler F, Cumming G (2012) Subjective p intervals researchers underestimate the variability of p values over replication. Methodology 8:51–62. 10.1027/1614-2241/a000037 [DOI] [Google Scholar]
  51. Lakens D (2013) Calculating and reporting effect sizes to facilitate cumulative science: a practical primer for t-tests and ANOVAs. Front Psychol 4:863. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Lakens D, Evers ERK (2014) Sailing from the seas of chaos into the corridor of stability: practical recommendations to increase the informational value of studies. Perspect Psychol Sci 9:278–292. 10.1177/1745691614528520 [DOI] [PubMed] [Google Scholar]
  53. Lakens D, Scheel AM, Isager PM (2018) Equivalence testing for psychological research: a tutorial. Adv Methods Pract Psychol Sci 1:259–269. 10.1177/2515245918770963 [DOI] [Google Scholar]
  54. Lane A, Luminet O, Nave G, Mikolajczak M (2016) Is there a publication bias in behavioural intranasal oxytocin research on humans? Opening the file drawer of one laboratory. J Neuroendocrinol 28. 10.1111/jne.12384 [DOI] [PubMed] [Google Scholar]
  55. Leng G, Ludwig M (2016) Intranasal oxytocin: myths and delusions. Biol Psychiatry 79:243–250. 10.1016/j.biopsych.2015.05.003 [DOI] [PubMed] [Google Scholar]
  56. Liu Y, Li S, Lin W, Li W, Yan X, Wang X, Pan X, Rutledge RB, Ma Y (2019) Oxytocin modulates social value representations in the amygdala. Nat Neurosci 22:633–641. 10.1038/s41593-019-0351-1 [DOI] [PubMed] [Google Scholar]
  57. MacKinnon S (2013) Increasing statistical power in psychological research without increasing sample size. Open Science Collaboration. Available at: http://osc.centerforopenscience.org/2013/11/03/Increasing-statistical-power/.
  58. Marsman M, Wagenmakers EJ (2017) Bayesian benefits with JASP. Eur J Dev Psychol 14:545–555. 10.1080/17405629.2016.1259614 [DOI] [Google Scholar]
  59. Maxwell SE, Kelley K, Rausch JR (2008) Sample size planning for statistical power and accuracy in parameter estimation. Annu Rev Psychol 59:537–563. 10.1146/annurev.psych.59.103006.093735 [DOI] [PubMed] [Google Scholar]
  60. McShane BB, Gal D, Gelman A, Robert C, Tackett JL (2019) Abandon statistical significance. Am Stat 73:235–245. 10.1080/00031305.2018.1527253 [DOI] [Google Scholar]
  61. Medina J, Cason S (2017) No evidential value in samples of transcranial direct current stimulation (tDCS) studies of cognition and working memory in healthy populations. Cortex 94:131–141. 10.1016/j.cortex.2017.06.021 [DOI] [PubMed] [Google Scholar]
  62. Meyvis T, Van Osselaer SMJ (2018) Increasing the power of your study by increasing the effect size. J Consum Res. 1173 44:1157. [Google Scholar]
  63. Minarik T, Berger B, Althaus L, Bader V, Biebl B, Brotzeller F, Fusban T, Hegemann J, Jesteadt L, Kalweit L, Leitner M, Linke F, Nabielska N, Reiter T, Schmitt D, Spraetz A, Sauseng P (2016) The importance of sample size for reproducibility of tDCS effects. Front Hum Neurosci 10:453. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Mole (2018) Mea culpa, mea powerful culpa. J Cell Sci 131:jcs215707. [Google Scholar]
  65. Mole (2017) Peerless I. J Cell Sci 130:2079–2080. [Google Scholar]
  66. Nave G, Camerer C, McCullough M (2015) Does oxytocin increase trust in humans? A critical review of research. Perspect Psychol Sci 10:772–789. 10.1177/1745691615600138 [DOI] [PubMed] [Google Scholar]
  67. Nieuwenhuis S, Forstmann BU, Wagenmakers EJ (2011) Erroneous analyses of interactions in neuroscience: a problem of significance. Nat Neurosci 14:1105–1107. 10.1038/nn.2886 [DOI] [PubMed] [Google Scholar]
  68. O’Boyle EH, Banks GC, Gonzalez-Mulé E (2017) The chrysalis effect. J Manage 43:376–399. 10.1177/0149206314527133 [DOI] [Google Scholar]
  69. Olds J, Milner P (1954) Positive reinforcement produced by electrical stimulation of septal area and other regions of rat brain. J Comp Physiol Psychol 47:419–427. 10.1037/h0058775 [DOI] [PubMed] [Google Scholar]
  70. Pek J, Flora DB (2018) Reporting effect sizes in original psychological research: a discussion and tutorial. Psychol Methods 23:208–225. [DOI] [PubMed] [Google Scholar]
  71. Peters GJ, Crutzen R (2017) Knowing exactly how effective an intervention, treatment, or manipulation is and ensuring that a study replicates: accuracy in parameter estimation as a partial solution to the replication crisis. PsyArxiv. Advance online publication. Retrieved September 7, 2018. doi: 10.31234/osf.io/cjsk2. [Google Scholar]
  72. Rothman KJ, Greenland S (2018) Planning study size based on precision rather than power. Epidemiology 29:599–603. 10.1097/EDE.0000000000000876 [DOI] [PubMed] [Google Scholar]
  73. Scoville WB, Milner B (1957) Loss of recent memory after bilateral hippocampal lesions. J Neurol Neurosurg Psychiatry 20:11–21. 10.1136/jnnp.20.1.11 [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Sherrington R, Rogaev EI, Liang Y, Rogaeva EA, Levesque G, Ikeda M, Chi H, Lin C, Li G, Holman K, Tsuda T, Mar L, Foncin JF, Bruni AC, Montesi MP, Sorbi S, Rainero I, Pinessi L, Nee L, Chumakov I, Pollen D, et al. (2003) Cloning of a gene bearing missense mutations in early-onset familial Alzheimer's disease. Nature 375:754–760. [DOI] [PubMed] [Google Scholar]
  75. Shuirmann DJ (1987) A comparison of the two one-sided tests procedure and the power. J Pharmacokinet Biopharm 15:657–680. [DOI] [PubMed] [Google Scholar]
  76. Simmons JP, Nelson LD, Simonsohn U (2011) False-positive psychology: undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychol Sci 22:1359–66. 10.1177/0956797611417632 [DOI] [PubMed] [Google Scholar]
  77. Smithson M (2002) Confidence intervals. Thousand Oaks, CA: SAGE. [Google Scholar]
  78. Song F, Parekh S, Hooper L, Loke YK, Ryder J, Sutton AJ, Hing C, Kwok CS, Pang C, Harvey I (2010) Dissemination and publication of research findings: an updated review of related biases. Health Technol Assess 14:1–220. 10.3310/hta14080 [DOI] [PubMed] [Google Scholar]
  79. Sterling TD (1959) Publication decisions and their possible effects on inferences drawn from tests of significance: or vice versa. J Am Stat Assoc 54:30–34. 10.2307/2282137 [DOI] [Google Scholar]
  80. Sterling TD, Rosenbaum WL, Weinkam JJ (1995) Publication decisions revisited: the effect of the outcome of statistical tests on the decision to publish and vice versa. Am Stat 49:108 10.2307/2684823 [DOI] [Google Scholar]
  81. Szucs D, Ioannidis JPA (2017a) When null hypothesis significance testing is unsuitable for research: a reassessment. Front Hum Neurosci 11:390. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Szucs D, Ioannidis JPA (2017b) Empirical assessment of published effect sizes and power in the recent cognitive neuroscience and psychology literature. PLOS Biol 15:e2000797 10.1371/journal.pbio.2000797 [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Tressoldi PE, Giofré D (2015) The pervasive avoidance of prospective statistical power: major consequences and practical solutions. Front Psychol 6:1–4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Tressoldi PE, Giofré D, Sella F, Cumming G (2013) High impact = high statistical standards? Not necessarily so. PLoS One 8:e56180. 10.1371/journal.pone.0056180 [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Vankov I, Bowers J, Munafò MR (2014) On the persistence of low power in psychological science. Q J Exp Psychol 67:1037–1040. 10.1080/17470218.2014.885986 [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Wagenmakers EJ, Love J, Marsman M, Jamil T, Ly A, Verhagen J, Selker R, Gronau QF, Dropmann D, Boutin B, Meerhoff F, Knight P, Raj A, van Kesteren EJ, van Doorn J, Šmíra M, Epskamp S, Etz A, Matzke D, de Jong T, et al. (2018a) Bayesian inference for psychology: Part II. Example applications with JASP. Psychon Bull Rev 25:58–76. 10.3758/s13423-017-1323-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Wagenmakers EJ, Gronau QF, Dablender F, Etz A (2018b) The support interval. PsyArXiv. Advance online publication. Retrieved December 22, 2018. doi: 10.31234/osf.io/zwnxb. [Google Scholar]
  88. Walum H, Waldman ID, Young LJ (2016) Statistical and methodological considerations for the interpretation of intranasal oxytocin studies. Biol Psychiatry 79:251–257. 10.1016/j.biopsych.2015.06.016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Wasserstein RL, Schirm AL, Lazar NA (2019) Moving to a world beyond “p < 0.05”. Am Stat 73:1–19. 10.1080/00031305.2019.1583913 [DOI] [Google Scholar]
  90. Westlake WJ (1972) Use of confidence intervals in analysis of comparative bioavailability trials. J Pharm Sci 61:1340–1341. [DOI] [PubMed] [Google Scholar]
  91. Zak P, Kurzban R, Matzner W (2005) Oxytocin is associated with human trustworthiness. Horm Behav 48:522–527. 10.1016/j.yhbeh.2005.07.009 [DOI] [PubMed] [Google Scholar]

Articles from eNeuro are provided here courtesy of Society for Neuroscience

RESOURCES