In our target article (1), we note that signal detection theory evolved (2) from Fisher’s (3) null hypothesis significance testing (NHST) approach (H0: δ = 0, H1: δ ≠ 0) as elaborated by Neyman and Pearson (4), who additionally proposed considering the specific magnitude of the alternative hypothesis (H0: δ = 0, H1: δ = μ). We then argue that this perspective is a useful fiction, not an accurate depiction of underlying reality. The true, underlying effect size (δ) associated with an experimental protocol is more reasonably conceptualized as having been drawn from a continuous distribution, and it is never equal to zero.
Pek et al. (5) argue that the Fisherian perspective can be revised by dropping the null hypothesis fiction that δ = 0. We certainly agree. Following Jones and Tukey (6), they further argue that a modified goal of NHST might be to detect directionally correct effects. This perspective is based on an accurate binary depiction of underlying reality, namely, that δ is in one direction (δ > 0) or the other (δ < 0). Given that nonfictional premise, science can be properly construed as a signal detection problem.
This argument seems perfectly sound to us, and it is entirely consistent with our own view of original science. As we put it: “A P < 0.05 finding should not be regarded as a scientifically established discovery; instead, it should be regarded as a provisional finding, one that is likely in the right direction but with an observed effect size that is inflated to an unknown degree” (ref. 1, p. 5561). It is provisional because the underlying effect size might be too close to 0 to matter (hence the need for large-N direct replications of important findings). Unfortunately, Pek et al. (5) do not address our vision of science at all. Instead, they simply argue that it is possible to imagine an unconventional version of Fisherian NHST that views science as a signal detection problem. It is a fair point, but no one conducts science from that perspective. Our point is that science is not the signal detection problem it is ordinarily thought to be (Fig. 1). Even so, this account is a useful fiction for original science.
Fig. 1.
The standard Fisherian view of NHST science, as elaborated by Neyman and Pearson (H0: δ = 0, H1: δ = μ) (4). Many scientists use this signal detection model to compute statistical power and to assess statistical significance. For such purposes, we argue that the model serves as a useful fiction. However, others assume (mistakenly in our view) that this model provides an accurate depiction of underlying reality. They then use it to justify proposed reforms to science, such as encouraging large-N experiments for original research (e.g., ref. 9). Such reforms, we argue, run the risk of making science worse, not better.
Pek et al. (5) also argue that our complex simulations of science were pointless because regression to the mean associated with using a P < 0.05 filter can be demonstrated in a much more parsimonious way. However, our simulations were not conducted to demonstrate inevitable regression to the mean, a statistical phenomenon that we discussed well before (and without reference to) our simulations. Indeed, others called attention to this issue long before we did (e.g., ref. 7). Our simulations were conducted 1) to estimate regression to the mean in the replication results reported by OSC2015 (8) and 2) to illustrate the problem associated with large-N studies in original science. Pek et al. (5) do not address the actual purpose of our simulations. More to the point, they do not engage our vision of science at all.
Footnotes
The authors declare no competing interest.
References
- 1.Wilson B. M., Harris C. R., Wixted J. T., Science is not a signal detection problem. Proc. Natl. Acad. Sci. U.S.A. 117, 5559–5567 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Wixted J. T., The forgotten history of signal detection theory. J. Exp. Psychol. Learn. Mem. Cogn. 46, 201–233 (2020). [DOI] [PubMed] [Google Scholar]
- 3.R. A. Fisher , Statistical Methods for Research Workers (Oliver and Boyd, Edinburgh, 1925). [Google Scholar]
- 4.Neyman J., Pearson E. S., On the problem of the most efficient tests of statistical hypotheses. Philos. Trans. R. Soc. Lond., A Contain. Pap. Math. Phys. Character 231, 289–337 (1933). [Google Scholar]
- 5.Pek J., Wegener D. T., McClelland G. H., Signal detection continues to be part of science. Proc. Natl. Acad. Sci. U.S.A. 117, 13199–13200 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Jones L. V., Tukey J. W., A sensible formulation of the significance test. Psychol. Methods 5, 411–414 (2000). [DOI] [PubMed] [Google Scholar]
- 7.Schmidt F. L., What do data really mean? Research findings, meta-analysis, and cumulative knowledge in psychology. Am. Psychol. 47, 1173–1181 (1992). [Google Scholar]
- 8.Open Science Collaboration , PSYCHOLOGY. Estimating the reproducibility of psychological science. Science 349, aac4716 (2015). [DOI] [PubMed] [Google Scholar]
- 9.Button K. S., et al. , Power failure: Why small sample size undermines the reliability of neuroscience. Nat. Rev. Neurosci. 14, 365–376 (2013). [DOI] [PubMed] [Google Scholar]

