Significance
The widespread use of Bonferroni correction encumbers the scientific process and wastes opportunities for discovery presented by big data, because it discourages exploratory analyses by overpenalizing the total number of statistical tests performed. In this paper, I introduce the harmonic mean p-value (HMP), a simple to use and widely applicable alternative to Bonferroni correction motivated by Bayesian model averaging that greatly improves statistical power while maintaining control of the gold standard false positive rate. The HMP has a range of desirable properties and offers a different way to think about large-scale exploratory data analysis in classical statistics.
Keywords: big data, false positives, p-values, multiple testing, model averaging
Abstract
Analysis of “big data” frequently involves statistical comparison of millions of competing hypotheses to discover hidden processes underlying observed patterns of data, for example, in the search for genetic determinants of disease in genome-wide association studies (GWAS). Controlling the familywise error rate (FWER) is considered the strongest protection against false positives but makes it difficult to reach the multiple testing-corrected significance threshold. Here, I introduce the harmonic mean p-value (HMP), which controls the FWER while greatly improving statistical power by combining dependent tests using generalized central limit theorem. I show that the HMP effortlessly combines information to detect statistically significant signals among groups of individually nonsignificant hypotheses in examples of a human GWAS for neuroticism and a joint human–pathogen GWAS for hepatitis C viral load. The HMP simultaneously tests all ways to group hypotheses, allowing the smallest groups of hypotheses that retain significance to be sought. The power of the HMP to detect significant hypothesis groups is greater than the power of the Benjamini–Hochberg procedure to detect significant hypotheses, although the latter only controls the weaker false discovery rate (FDR). The HMP has broad implications for the analysis of large datasets, because it enhances the potential for scientific discovery.
Analysis of big data has great potential, for instance by transforming our understanding of how genetics influences human disease (1), but it presents unique challenges. One such challenge faces geneticists designing genome-wide association studies (GWAS). Individuals have typically been typed at around 600,000 variants spread across the 3.2 billion base-pair genome. With the rapidly decreasing costs of DNA sequencing, whole-genome sequencing is becoming routine, raising the possibility of detecting associations at ever more variants (2, 3). However, increasing the number of tests of association conventionally requires more stringent p-value correction for multiple testing, reducing the probability of detecting any individual association. The idea that analyzing more data may lead to fewer discoveries is counterintuitive and suggests a flaw of logic.
The problem of testing many hypotheses while controlling the appropriate false positive rate is a long-standing issue. The familywise error rate (FWER) is the probability of falsely rejecting a null in favor of an alternative hypothesis in one or more of all tests performed. Controlling the FWER in the presence of some true positives is challenging and considered the strongest form of protection against false positives (4). Unfortunately, the simple and widely used Bonferroni method for controlling the FWER is conservative, especially when the individual tests are positively correlated (5).
Model selection is an important setting affected by correlated tests, in which the same data are used to evaluate many competing alternative hypotheses. Reanalysis of the same outcomes across tests in GWAS causes dependence because of correlations between regressors in different models (6). Other phenomena, such as unmeasured confounders, can induce dependence, even when alternative hypotheses are not mutually exclusive, such as in gene expression analyses (7). The conservative nature of Bonferroni correction, particularly when tests are correlated, exacerbates the stringent criterion of controlling the FWER, jeopardizing sensitivity to detect true signals.
Simulations may be used to identify thresholds that are less stringent yet control the FWER. However, simulating can be time consuming; model-based simulations require knowledge of the dependency structure, which may be limited; and permutation-based procedures are not always appropriate (8).
The false discovery rate (FDR) offers an alternative to the FWER. Controlling the FDR guarantees that, among the significant tests, the proportion in which the null hypothesis is incorrectly rejected in favor of the alternative is limited (9). The widely used Benjamini–Hochberg (BH) procedure (9) for controlling the FDR shares with the Bonferroni method a robustness to positive correlation between tests (10) but is less conservative. These advantages have made FDR a popular alternative to FWER, in practice trading off larger numbers of false positives for more statistical power.
Combined tests offer a different way to improve power. By aggregating multiple hypothesis tests, combined tests are sensitive to signals that may be individually too subtle to detect, especially after multiple testing correction. Their conclusions, therefore, apply collectively rather than to individual tests. Fisher’s method (11) is perhaps the best known and has been widely used in gene set enrichment analysis, but it makes the strong assumption that tests are independent.
Bayesian model averaging offers a way to combine alternative hypotheses in the model selection setting. By comparing groups of alternative hypotheses against a common null, the null hypothesis may be ruled out collectively. In the case of GWAS, even if no individual variant shows sufficient evidence of association in a region, the model-averaged signal across that region may yet achieve sufficiently strong posterior odds (12, 13). Combining tests in this way makes an asset of more data by creating the potential for more fine-grained discovery when the signal is strong enough without the liability of requiring that all hypotheses are evaluated individually at the higher level of statistical stringency.
In this paper, I use Bayesian model averaging to develop a method, the harmonic mean p-value (HMP), for combining dependent p-values while controlling the strong-sense FWER. The method is derived in the model selection setting and is best interpreted as offering a complementary method to Fisher’s that combines tests by model averaging when they are mutually exclusive, not independent. However, the HMP is applicable beyond model selection problems, because it assumes only that the p-values are valid. It enjoys several remarkable properties that offer benefits across a wide range of big data problems.
Methods
Model-Averaged Mean Maximum Likelihood.
The original idea motivating this paper was to develop a classical analogue to the model-averaged Bayes factor by deriving the null distribution for the mean maximized likelihood ratio,
[1] |
with maximized likelihood ratios and weights , where .
The maximized likelihood ratio is a classical analogue of the Bayes factor and measures the evidence for the alternative hypothesis against the null given the data :
In a likelihood ratio test, the p-value is calculated as the probability of obtaining an as or more extreme if the null hypothesis were true:
For nested hypotheses (), Wilks’ theorem (14) approximates the null distribution of as when there are degrees of freedom.
The distribution of cannot be approximated by central limit theorem, because the LogGamma distribution is heavy tailed, with undefined variance when . Instead, generalized central limit theorem can be used (15), which states that, for equal weights () and independent and identically distributed s,
[2] |
where and are constants and is a Stable distribution with tail index . The specific Stable distribution is a type of Landau distribution (16) with parameters that depend on and (SI Appendix, section 1). Theory, supported by detailed simulations in SI Appendix, section 2, shows that (i) the assumptions of equal weights, independence, and identical degrees of freedom can be relaxed and that (ii) the Landau distribution approximation performs best when .
The Harmonic Mean p-Value.
Notably, when and the assumptions of Wilks’ theorem are met, the p-value equals the inverse maximized likelihood ratio:
and therefore, the mean maximized likelihood ratio equals the inverse HMP:
[3] |
Under these conditions, interpreting and the HMP is exactly equivalent. This equivalence motivates use of the HMP more generally because of the following.
-
i)
The Landau distribution gives an excellent approximation for with , and hence for .
-
ii)
Wilks’ theorem can be replaced with the simpler assumption that the p-values are well calibrated.
-
iii)
The HMP will capture similar information to for any degrees of freedom.
-
iv)
Combining s rather than s automatically accounts for differences in degrees of freedom.
A combined p-value, which becomes exact as the number of p-values increases, can be calculated as
[4] |
with the Landau distribution probability density function
Remarkably, however, the HMP can be directly interpreted, because it is approximately well calibrated when small. Using the theory of regularly varying functions (see ref. 17),
[5] |
This property suggests the following test, which controls the strong-sense FWER at level approximately for an HMP calculated on a subset of p-values :
[6] |
where . Directly interpreting the HMP using Eq. 6 constitutes a multilevel test in the sense that any significant subset of hypotheses implies that the HMP of the whole set is also significant, because
[7] |
Conversely, if the “headline” HMP is not significant, nor is the HMP for any subset . The significance thresholds apply no matter how many subsets are combined and tested.
The above properties show that directly interpreting the HMP (i) is a closed testing procedure (4) that controls the strong-sense FWER (SI Appendix, section 3); (ii) is more powerful than Bonferroni and Simes correction, because the HMP is always smaller than the p-values for those tests (SI Appendix, section 4); and therefore, (iii) produces significant results whenever the Simes-based BH procedure does, although BH only controls the less stringent FDR.
While direct interpretation of the HMP controls the strong-sense FWER, the level at which it does so is only approximately , and is in fact anticonservative, but only very slightly for small and small . Assessing the adjusted HMP, , against level calculated by inverting Eq. 4 permits a test that is exact up to the order of the Landau distribution approximation (Table 1). (Equivalently, one can compare the exact p-value from Eq. 4 with .) Simulations suggest that this exact test remains more powerful than Bonferroni, Simes, and therefore, BH (SI Appendix, section 4).
Table 1.
10 | 0.040 | 0.0094 | 0.00099 |
100 | 0.036 | 0.0092 | 0.00099 |
1,000 | 0.034 | 0.0090 | 0.00099 |
10,000 | 0.031 | 0.0088 | 0.00098 |
100,000 | 0.029 | 0.0086 | 0.00098 |
1,000,000 | 0.027 | 0.0084 | 0.00098 |
10,000,000 | 0.026 | 0.0083 | 0.00098 |
100,000,000 | 0.024 | 0.0081 | 0.00098 |
1,000,000,000 | 0.023 | 0.0080 | 0.00097 |
I recommend the use of this asymptotically exact test, available in the R package “harmonicmeanp” (https://CRAN.R-project.org/package=harmonicmeanp), on which all subsequent analyses in Results are based. Analyses based on direct interpretation of the HMP are also presented and reveal the practical differences between the approaches to be small for .
Choice of Weights.
I anticipate that the HMP will usually be used with equal weights, as are procedures such as Bonferroni correction and Simes’ test. SI Appendix, section 5 considers optimal weights. Based on Bayesian (18) and classical arguments and assuming that all tests have good power, the optimal weight is found to be proportional to the product of the prior probability of alternative hypothesis and the expectation of under . This optimal weighting would favor alternatives that are more probable a priori while penalizing those associated with more powerful tests.
Consequently, the use of equal weights can be interpreted as assuming that all alternative hypotheses are equally likely a priori and that all tests are equally powerful. If tests are not equally powerful for a given “effect size,” the equal power assumption implies that alternatives associated with inherently less powerful tests are expected to have larger effect sizes a priori, a testable assumption that has been used often in GWAS (19).
Results
The main result of this paper is that the weighted harmonic mean p-value of any subset of the p-values ,
[8] |
(i) combines the evidence in favor of the group of alternative hypotheses , (ii) is an approximately well-calibrated p-value for small values, and (iii) controls the strong-sense FWER at level approximately when compared against the threshold , no matter how many other subsets of the same p-values are tested ( and ). An asymptotically exact test is also available (Eq. 4). The HMP has several helpful properties that arise from generalized central limit theorem. It is:
-
i)
Robust to positive dependency between p-values.
-
ii)
Insensitive to the exact number of tests.
-
iii)
Robust to the distribution of weights .
-
iv)
Most influenced by the smallest p-values.
The HMP outperforms Bonferroni and Simes (5) correction. This advantage over Simes’ test means that whenever the BH procedure (9), which controls only the FDR, finds significant hypotheses, the HMP will find significant hypotheses or groups of hypotheses. The HMP complements Fisher’s method for combining independent p-values (11), because the HMP is more appropriate when (i) rejecting the null implies that only one alternative hypothesis may be true and not all of them or (ii) the p-values might be positively correlated and cannot be assumed to be independent.
HMP Enables Adaptive Multiple Testing Correction by Combining p-Values.
That the Bonferroni method for controlling the FWER can be overly stringent, especially when the tests are nonindependent, has long been recognized. In Bonferroni correction, a p-value is deemed significant if , which becomes more stringent as the number of tests increases. Since human GWAS began routinely testing millions of variants by statistically imputing untyped variants, a new convention was adopted in which a p-value is deemed significant if , a rule that implies that the effective number of tests is no more than . Several lines of argument were used to justify this threshold (20–22), most applicable specifically to human GWAS.
In contrast, the HMP affords strong control of the FWER while avoiding both simulation studies and the undue stringency of Bonferroni correction, an advantage that increases when tests are nonindependent. To show how the HMP can recover significant associations among groups of tests that are individually nonsignificant, I reanalyzed a GWAS of neuroticism (23), defined as a tendency toward intense or frequent negative emotions and thoughts (24). Genotypes were imputed for variants across individuals. I used the HMP to perform model-averaged tests of association between neuroticism and variants within contiguous regions of 10 kb, 100 kb, 1,000 kb, 10 Mb, entire chromosomes, and the whole genome, assuming equal weights across variants (SI Appendix, section 6).
Fig. 1 shows the p-value from Eq. 4 for each region adjusted by a factor to enable direct comparison with the significance threshold . Similar results were obtained from direct interpretation of the HMP (SI Appendix, Fig. S1). Model averaging tends to make significant and near-significant adjusted p-values more significant. For example, for every variant significant after Bonferroni correction, the model-averaged p-value for the corresponding chromosome was found to be at least as significant.
Model averaging increases significance more when combining a group of comparably significant p-values, e.g., the top hits in chromosome 9. The least improvement is seen when one p-value is much more significant than the others, e.g., the top hit in chromosome 3. This behavior is predicted by the tendency of harmonic means to be dominated by the smallest values. In the extreme case that one p-value dominates the significance of all others, the HMP test becomes equivalent to Bonferroni correction. This implies that Bonferroni correction might not be improved on for “needle-in-a-haystack” problems. Conversely, dependency among tests actually improves the sensitivity of the HMP, because one significant test may be accompanied by other correlated tests that collectively reduce the harmonic mean p-value.
In some cases, the HMP found significant regions where none of the individual variants were significant. For example, no variants on chromosome 12 were significant by Bonferroni correction nor by the conventional genome-wide significance threshold of . However, the HMP found significant 10-Mb regions spanning several peaks of nonsignificant individual p-values. One of those, variant rs7973260, which showed an individual p-value for association with neuroticism of , had been reported as also associated with depressive symptoms (). Such cross-association or “quasireplication,” in which a variant is nearly significant for the trait of interest and significant for a related trait, can be regarded as providing additional support for the variant’s involvement in the trait of interest (23).
In chromosome 3, individual variants were found to be significant by the conventional threshold of , but neither Bonferroni correction nor the HMP agreed that those variants or regions were significant at an FWER of . Indeed, the HMP found chromosome 3 nonsignificant as a whole. Variant rs35688236, which had the smallest p-value on chromosome 3 of , had not validated when tested in a quasireplication exercise that involved testing variants associated with neuroticism for association with subjective wellbeing or depressive symptoms (23).
These observations illustrate that the HMP adaptively combines information among groups of similarly significant tests where possible, while leaving lone significant tests subject to Bonferroni-like stringency, providing a general approach to combining p-values that does not require specific knowledge of the dependency structure between tests.
HMP Allows Large-Scale Testing for Higher-Order Interactions Without Punitive Thresholds.
Scientific discovery is currently hindered by avoidance of large-scale exploratory hypothesis testing for fear of attracting multiple testing correction thresholds that render signals found by more limited testing no longer significant. A good example is the approach to testing for pairwise or higher-order interactions between variants in GWAS. The Bonferroni threshold for testing all pairwise interactions invites a threshold times more stringent than the threshold for testing variants individually, and strictly speaking this must be applied to every test, even though this is highly conservative because of the dependency between tests. The alternative of controlling the FDR risks a high probability of falsely detecting artifacts among any genuine associations discovered. Therefore, interactions are not usually tested for.
To show how model averaging using the HMP greatly alleviates this problem, I reanalyzed human and pathogen genetic variants from a GWAS of pretreatment viral load in hepatitis C virus (HCV)-infected patients (25) (SI Appendix, section 7). Jointly analyzing the influence of human and pathogen variation on infection is an area of great interest, but it requires a Bonferroni threshold of when there are and variants in the human and pathogen genomes, respectively, compared with if testing the human and pathogen variants separately. In this example, and .
In the original study, a known association with viral load was replicated at human chromosome 19 variant rs12979860 in IFNL4 (), below the Bonferroni threshold of for tests. The most significant pairwise interaction that I found, assuming equal weights, involved the adjacent variant rs8099917 with . However, this did not meet the more stringent Bonferroni threshold of for 330 million tests (Fig. 2A). If the original study’s authors had performed and reported 330 million tests, they could have been compelled to declare the marginal association in IFNL4 nonsignificant, despite what intuitively seems like a clear signal.
Model averaging using the HMP reduces this disincentive to perform additional related tests. Fig. 2B shows that, despite no significant pairwise tests involving rs8099917, model averaging recovered a combined p-value of , below the multiple testing threshold of for the model-averaged tests. Additionally, two viral variants produced statistically significant model-averaged p-values of and at polyprotein positions 10 and 2,061 in the capsid and NS5a zinc finger domain (GenBank accession no. AQW44528), below the multiple testing threshold of for the 827 model-averaged tests.
These results show how model averaging using the HMP can assist discovery making by (i) encouraging tests for higher-order interactions when they otherwise would not be attempted and (ii) recovering lost signals of marginal associations after performing an “excessive” number of tests.
Untangling the Signals Driving Significant Model-Averaged p-Values.
When more than one alternative hypothesis is found to be significant, either individually or as part of a group, it is desirable to quantify the relative strength of evidence in favor of the competing alternatives. This is particularly true when disentangling the contributions of a group of individually nonsignificant alternatives that are significant only in combination.
Sellke et al. (18) proposed a conversion from p-values to Bayes factors which, when combined with prior information and test power through the model weights, produces posterior model probabilities and credible sets of alternative hypotheses. SI Appendix, section 5 details how the Bayes factors are approximately proportional to the weighted inverse p-values. This linearity mirrors the HMP itself, the inverse of which is an arithmetic mean of the inverse p-values.
After conditioning on rejection of the null hypothesis by normalizing the approximate model probabilities to sum to 100%, the probability that the association involved human variant rs8099917 was 54.4%. This signal was driven primarily by the three viral variants with the highest probability of interacting with rs8099917 in their effect on pretreatment viral load: position 10 in the capsid (10.9%), position 669 in the E2 envelope (8.7%), and position 2,061 in the NS5a zinc finger domain (11.4%) (Fig. 3). Even though the model-averaged p-value for the envelope variant was not itself significant, this revealed a plausible interaction between it and the most significant human variant rs8099917.
Discussion
The HMP provides a way to calculate model-averaged p-values, providing a powerful and general method for combining tests while controlling the strong-sense FWER. It provides an alternative to both the overly conservative Bonferroni control of the FWER, and the lower stringency of FDR control. The HMP allows the incorporation of prior information through model weights and is robust to positive dependency between the p-values. The HMP is approximately well calibrated for small values, while a null distribution, derived from generalized central limit theorem, is easily computed. When the HMP is not significant, neither is any subset of the constituent tests.
The HMP is more appropriate for combining p-values than Fisher’s method when the alternative hypotheses are mutually exclusive, as in model comparison. When the alternative hypotheses all have the same nested null hypothesis, the HMP is interpreted in terms of a model-averaged likelihood ratio test. However, the HMP can be used more generally to combine tests that are not necessarily mutually exclusive but that may have positive dependency, with the caveat that more powerful approaches may be available depending on the context. The HMP can be used alone or in combination: for example, with Fisher’s method to combine model-averaged p-values between groups of independent data.
The theory underlying the HMP provides a fundamentally different way to think about controlling the FWER through multiple testing correction. The Bonferroni threshold increases linearly with the number of tests, whereas the HMP is the reciprocal of the mean of the inverse p-values. To maintain significance with Bonferroni correction, the minimum p-value must decrease linearly as the number of tests increases. This strongly penalizes exploratory and follow-up analyses. In contrast, when the false positive rate is small, maintenance of significance with the HMP requires only that the mean inverse p-value remains constant as the number of tests increases. This does not penalize exploratory and follow-up analyses so long as the “quality” of the additional hypotheses tested, measured by the inverse p-value, does not decline.
Through example applications to GWAS, I have shown that the HMP combines tests adaptively, producing Bonferroni-like adjusted p-values for needle-in-a-haystack problems when one test dominates, but able to capitalize on numerous strongly significant tests to produce smaller adjusted p-values when warranted. I have shown how model averaging using the HMP encourages exploratory analysis and can recover signals of significance among groups of individually nonsignificant tests, properties that have the potential to enhance the scientific discovery process.
Supplementary Material
Acknowledgments
I thank Azim Ansari, Vincent Pedergnana, and Chris Spencer for sharing expertise; Gil McVean, Simon Myers, and Sarah Walker for insightful comments; and the Social Science Genetic Association Consortium, the Stratified Medicine to Optimise the Treatment of Patients with Hepatitis C Virus Infection (STOP-HCV) Consortium, and the HCV Research UK Biobank for sharing data. D.J.W. is a Sir Henry Dale Fellow, jointly funded by the Wellcome Trust and the Royal Society (Grant 101237/Z/13/Z). D.J.W. is supported by a Big Data Institute Robertson Fellowship, and is a member of the STOP-HCV Consortium, which is funded by Medical Research Council Award MR/K01532X/1.
Footnotes
The author declares no conflict of interest.
This article is a PNAS Direct Submission.
This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1814092116/-/DCSupplemental.
References
- 1.The Royal Society . Machine Learning: The Power and Promise of Computers That Learn by Example. The Royal Society; London: 2017. [Google Scholar]
- 2.Fuchsberger C, et al. The genetic architecture of type 2 diabetes. Nature. 2016;536:41–47. doi: 10.1038/nature18642. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.van Rheenen W, et al. Genome-wide association analyses identify new risk variants and the genetic architecture of amyotrophic lateral sclerosis. Nat Genet. 2016;48:1043–1048. doi: 10.1038/ng.3622. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Marcus R, Eric P, Gabriel KR. On closed testing procedures with special reference to ordered analysis of variance. Biometrika. 1976;63:655–660. [Google Scholar]
- 5.Simes RJ. An improved Bonferroni procedure for multiple tests of significance. Biometrika. 1986;73:751–754. [Google Scholar]
- 6.The International HapMap Consortium A haplotype map of the human genome. Nature. 2005;437:1299–1320. doi: 10.1038/nature04226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Leek JT, Storey JD. A general framework for multiple testing dependence. Proc Natl Acad Sci USA. 2008;104:18718–18723. doi: 10.1073/pnas.0808709105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Huang Y, Xu Y, Calian V, Hsu JC. To permute or not to permute. Bioinformatics. 2006;22:2244–2248. doi: 10.1093/bioinformatics/btl383. [DOI] [PubMed] [Google Scholar]
- 9.Benjamini Y, Hochberg Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. J R Stat Soc Series B. 1995;57:289–300. [Google Scholar]
- 10.Benjamini Y, Yekutieli D. The control of the false discovery rate in multiple testing under dependency. Ann Stat. 2001;29:1165–1188. [Google Scholar]
- 11.Fisher RA. Statistical Methods for Research Workers. 5th Ed Oliver and Boyd; Edinburgh: 1934. [Google Scholar]
- 12.Ferreira T, Marchini J. Modeling interactions with known risk loci: A Bayesian model averaging approach. Ann Hum Genet. 2011;75:1–9. doi: 10.1111/j.1469-1809.2010.00618.x. [DOI] [PubMed] [Google Scholar]
- 13.Crawford L, Zeng P, Mukherjee S, Zhou X. Detecting epistasis with the marginal epistasis test in genetic mapping studies of quantitative traits. PLoS Genet. 2017;13:e1006869. doi: 10.1371/journal.pgen.1006869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Wilks SS. The large-sample distribution of the likelihood ratio for testing composite hypotheses. Ann Math Stat. 1938;9:60–62. [Google Scholar]
- 15.Uchaikin VV, Zolotarev VM. Chance and Stability: Stable Distributions and Their Applications. VSP, Utrecht; The Netherlands: 1999. [Google Scholar]
- 16.Landau LD. On the energy loss of fast particles by ionization. J Phys USSR. 1944;8:201–205. [Google Scholar]
- 17.Mikosch T. Regular Variation, Subexponentiality and Their Applications in Probability Theory. Vol 99 Eindhoven University of Technology; Eindhoven, The Netherlands: 1999. [Google Scholar]
- 18.Sellke T, Bayarri MJ, Berger JO. Calibration of values for testing precise null hypotheses. Am Stat. 2001;55:62–71. [Google Scholar]
- 19.Speed D, Cai N, Johnson MR, Nejentsev S, Balding DJ. UCLEB Consortium Reevaluation of SNP heritability in complex human traits. Nat Genet. 2017;49:986–992. doi: 10.1038/ng.3865. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Pe’er I, Yelensky R, Altshuler D, Daly MJ. Estimation of the multiple testing burden for genomewide association studies of nearly all common variants. Genet Epidemiol. 2008;32:381–385. doi: 10.1002/gepi.20303. [DOI] [PubMed] [Google Scholar]
- 21.Dudbridge F, Gusnanto A. Estimation of significance thresholds for genomewide association scans. Genet Epidemiol. 2008;32:227–234. doi: 10.1002/gepi.20297. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Fadista J, Manning AK, Florez JC, Groop L. The (in) famous GWAS P-value threshold revisited and updated for low-frequency variants. Eur J Hum Genet. 2016;24:1202–1205. doi: 10.1038/ejhg.2015.269. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Okbay A, et al. Genetic variants associated with subjective well-being, depressive symptoms, and neuroticism identified through genome-wide analyses. Nat Genet. 2016;48:624–633. doi: 10.1038/ng.3552. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.McCrae RR, Costa PT., Jr . The five-factor theory of personality. In: Pervin LA, John OP, Robins RW, editors. Handbook of Personality: Theory and Research. 3rd Ed. Guilford; New York: 2008. pp. 159–181. [Google Scholar]
- 25.Ansari MA, et al. Genome-to-genome analysis highlights the effect of the human innate and adaptive immune systems on the hepatitis c virus. Nat Genet. 2017;49:666–673. doi: 10.1038/ng.3835. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.