Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2019 Mar 13.
Published in final edited form as: Ann Hum Genet. 2017 Dec 18;82(2):93–101. doi: 10.1111/ahg.12229

Fast permutation tests and related methods, for association between rare variants and binary outcomes

ARJUN SONDHI 1, KENNETH MARTIN RICE 1
PMCID: PMC6415917  NIHMSID: NIHMS1010322  PMID: 29250767

Summary

In large scale genetic association studies, a primary aim is to test for association between genetic variants and a disease outcome. The variants of interest are often rare, and appear with low frequency among subjects. In this situation, statistical tests based on standard asymptotic results do not adequately control the Type I error rate, especially if the case:control ratio is unbalanced. In this paper, we propose the use of permutation and approximate unconditional tests for testing association with rare variants. We use novel analytical calculations to efficiently approximate the true Type I error rate under common study designs, and in numerical studies show that the proposed classes of tests significantly improve upon standard testing methods. We also illustrate our methods in data from a recent case-control study, for genetic causes of a severe side-effect of a common drug treatment.

Keywords: association tests, binary outcomes, permutation tests, rare variants

1. Introduction

Association studies are often performed for binary traits, providing new knowledge of the genetic causes of human diseases, using data from case-control and cohort studies (Verhaaren et al., 2015; Opherk et al., 2014; Danjou et al., 2015; Hoffmann et al., 2015). Recent advances in sequencing technology have made it practical to type essentially every variant on the genome; to avoid spurious findings, very low Type I error rates must therefore be maintained (Hoggart et al., 2008). However, for the rare variants now being studied, standard analytic approaches do not reliably achieve their nominal rates, (Li and Leal, 2008; Xing et al., 2012; Ma et al., 2013) and may permit too many Type I errors. The problem can be particularly severe when the ratio of cases to controls is extreme. Adjustments that maintain control at the nominal rate can be conservative, leading to loss of power relative to methods that control Type I errors more accurately.

In this paper, motivated by work in a case-control study of rhabdomyolosis, we develop methods with improved control of the Type I error rate, when testing single rare variants for association with binary traits. In Section 2, we explain a novel numerical method that approximates the actual Type I error rate of a test statistic given sample size, significance level, and a variant’s expected frequency; we also show how the same basic ideas can be used in permutation and approximate unconditional tests, and how the ideas can be used when adjusting for covariates. In Section 3, we give the results of the numerical studies performed, demonstrating improvements over standard asymptotic tests. Section 4 applies these methods to data from a case-control study of statin-related rhabdomyolysis, and we conclude with a short discussion, including details of an R package that implements our methods.

2. Materials and Methods

With rare variants, homozygotes are so rare as to be negligible for analysis, and it suffices to consider whether subjects have any copies of the variant present (G=1) or not (G=0). This step also means it is simple to enumerate all possible datasets; given fixed numbers of cases and controls (m1 and m0 respectively), we need only consider the number of cases (r1) and controls (r0) with the variant (adjustment for covariates is considered in Section 2.3.).

For a variant with a given minor allele frequency (MAF), following Ma et al. (2013), we define expected number of minor allele carriers as EMAC=(m0+m1)×(1(1MAF)2). Under the null hypothesis of no association, it follows that

r1Binom(m1,EMACm0+m1)
r0Binom(m0,EMACm0+m1),

independently. Therefore, given a value of EMAC — or equivalently MAF — and the fixed numbers of cases and controls in the study at hand, we can simply write down the probability of seeing all possible datasets under the null hypothesis.

In theory, this direct enumeration allows exact calculation of the Type I error rate for any test: the Type I error rate is the sum of the probabilities of the datasets for which a significant test result is returned. Formally, a dataset (m0,m1,r0,r1) returns a significant test result when its associated p-value p(r0,r1;m0,m1)α, where p(r0,r1;m0,m1)=[|T|Tobs;H0] is the probability of test statistic T equaling or exceeding the observed value Tobs when the null hypothesis holds. The Type I error rate of the test at nominal level α is then defined as

T1ER(α)=0r0m0,0r1m1f(r0,r1;m0,m1,EMAC)1p(r0,r1;m0,m1)<α, (1)

where f(r0,r1;m0,m1,EMAC) denotes the probability of observing data r0,r1 under the null hypothesis, given m0, m1, and a specific EMAC. Although not discussed further, the approach is easily adapted to p-values that use lower tail areas -- below Tobs instead of above, or two-sided tests that examine tail areas beyond ±Tobs.

In practice, the sum of (m0+1)×(m1+1) terms in (1) may be large, making computation too slow for some purposes. However, for work on rare variants, almost all of the summands contribute negligibly to the overall Type I error rate. A practical solution is therefore to truncate the summation in (1) by zeroing-out terms that, in total, represent no more than a small fraction of the Type I error rate.

Taking this approach, in our work we will zero-out terms in (1) representing datasets for which r0+r1 exceeds the upper 10−12 quantile of the distribution of r1+r0, i.e. of Binom(m0+m1,EMACm0+m1) (Figure 1 gives a graphical description of this process). By setting these terms to zero, in this example, we therefore understate the Type I error rate by no more than 10−12, which is acceptable given our focus on Type I error rates near α=5×108, and maintain practical computation times, even for large studies. For example, performing this calculation for (m0,m1,EMAC)=(500,500,15) at α=5×108 using the standard Score test takes 0.05 seconds on a standard laptop. Without the zeroing-out method, the calculation takes 15.3 seconds, i.e. more than 300 times faster. Our choice of α corresponds to testing a million independent variants (Pe’er et al., 2008), a level that has been widely-adopted as the standard for genome-wide work.

Figure 1:

Figure 1:

Possible datasets and their contribution to T1ER for the standard Score test, for m0 = m2 = 1000 and EMAC = 15. In a), the red/green zones indicate datasets where the standard score test is significant/not-significant at nominal α=5×108; the blue zone shows terms that are not zeroed-out using the truncation described in Section 2. In b) the same situation is shown, zoomed in and with box size proportional to the probability of each dataset; the actual T1ER(α) is given by the sum of the box areas in the two red zones. Zeroing-out contributions beyond the blue region, where r0+r1>50, gives an approximation error in the p-value of no more than 10−12.

We emphasize that this approximation of the Type I error rate is entirely general; any test statistic T can be used, including the familiar Wald, Score, likelihood ratio test statistics (see e.g. Ma et al. (2013)) or more sophisticated choices such as the Firth test (Firth, 1993; Heinze et al., 2013). Accurate knowledge of these approximate tests enables users to better compare their performance at the nominal α.

The formulation of Type I error rate in (1) and its approximation can also directly inform construction of permutation and approximate unconditional tests, as we discuss in Sections 2.1 and 2.2 below. We briefly discuss adjusting for covariates in Section 2.3, for both forms of test.

2.1. Permutation tests

The ability of permutation tests to provide accurate p-values for association testing under minimal assumptions is well-know (Pitman, 1937; Huo et al., 2014; Nichols and Holmes, 2002; Anderson, 2001); where they are applicable, permutation tests are regarded by many analysts as the ‘gold standard’ method. For quantitative traits, a major drawback is that permutations must, in practice, be performed using random number generation (Boyett and Shuster, 1977). For analysis of binary traits this is not needed; we can instead enumerate all possible permutations and obtain accurate p-values.

Using the same notation as above, a permutation test requires an observed test statistic, Tobs, calculated on the observed data (m0,m1,r0,r1). We shall consider test statistics from standard score, Wald, likelihood ratio, and Firth test approaches, thus providing a permutation version of them. The test statistic is also calculated for each possible datasets (m0,m1,r0,r1) obtained by permuting binary outcomes (e.g. case/control labels) among all study subjects, or equivalently permuting the variant/non-variant carrier status among all subjects. Under permutation, the total number of minor allele carriers is the same as in observed data, that is, r0+r1=r0+r1, and under the null hypothesis of no association the probability of observing each dataset follows the hypergeometric distribution (Good, 2005):

f˜(r0,r1;m0,m1,r0+r1)=(r0+r1r1)(m0+m1r0r1m1r1)(m0+m1m1).

The permutation p-value is then defined as

pperm(r0,r1;m0,m1)=r0,r1f˜(r0,r1;m0,m1,r0+r1)1|Tr0,r1||Tobs|,

i.e. the sum of probabilities of datasets with the same number of allele carriers that result in more extreme test statistics than Tobs. The datasets enumerated in this method are illustrated in Figure 2.

Figure 2:

Figure 2:

Datasets used when calculating the p-value for the permutation version of the Score test, for observed data (m0,m1,r0,r1)=(1000,1000,5,10). The size of each square corresponds to the probability of observing the corresponding dataset under the null hypothesis. The p-value is represented by the sum of the areas of the squares in the two shaded `tails’ of the distribution, containing all datasets with Score test statistic at least as extreme as the observed data.

Permutation tests are exact, in the sense that the observed Type I error rate will always be less than or equal to α. This result is well-known and dates back to Fisher (Janssen and Pauls, 2003). In particular, for the rare variant setting, permutation tests will be fairly conservative. A mathematical explanation is given in the Appendix.

2.2. Approximate Unconditional (AU) tests

Approximate Unconditional (AU) tests (Storer and Kim, 1990) provide Type I error rates closer to the nominal level than permutation approaches. Unlike permutation tests, AU tests are not guaranteed to always strictly control the Type I error rate, but this anti-conservatism (where it occurs at all) is usually very mild.

Using the same notation as above, AU tests calculate a test statistic Tobs from the observed data (m0,m1,r0,r1) and from all possible datasets (m0,m1,r0,r1) but the restriction that r0+r1=r0+r1. The probability of observing each dataset under the null hypothesis is calculated using fitted binomial distributions, i.e.

f^(r0,r1;m0,m1,r0,r1)=(m0r0)(m1r1)(r0+r1m0+m1)r0+r1(1r0+r1m0+m1)m0+m1r0r1.

The AU test’s p-value is then defined as

pAU(r0,r1;m0,m1)=r0,r1f^(r0,r1;m0,m1,r0,r1)1|Tr0,r1||Tobs|, (2)

i.e. the sum of probabilities of datasets that result in more extreme test statistics thanTobs. We can then apply Equation (1) and write the Type I error rate as:

T1ER(α)=0r0m0,0r1m1f^(r0,r1;m0,m1,EMAC)1pAU(r0,r1;m0,m1)<α, (3)

Compared to the permutation test, the AU test’s p-value sums over many more possible datasets, allowing less crude approximation of the Type I error rate. This comes at the cost of using the same data to fit the null binomial models, and hence losing guaranteed control of the Type I error rate. However, in our setting a bigger practical concern is that taking a naïve approach to calculation in Equation (2) would require (m0+1)×(m1+1) evaluations for each p-value, which may be a burden, as with Equation (1). A much quicker approach that is still adequate in practice uses the same zeroing-out idea as before -- we only sum elements (r0,r1) in (2) for values of r0+r1 between the upper and lower 10−12 quantiles of the Binom(m0+m1,r0+r1m0+m1) distribution.

The datasets enumerated in this method are illustrated in Figure 3. As with the calculation of Type I errors in Section 2, the zeroing out leads to a slight understatement of the p-value compared to complete enumeration. However, understating the p-value by at most 2×1012 is a very minor concern when α=108, several orders of magnitude greater, and comes in return for a substantial speed increase. For example, computing an AU p-value under the Score test with data (m0,m1,r0,r1)=(5000,5000,10,50) takes 0.05 seconds on a standard laptop with zeroing out and 28.4 seconds without, i.e. over 500 times faster.

Figure 3:

Figure 3:

Datasets used when calculating p-value for the AU version of the Score test, for observed data (m0,m1,r0,r1)=(1000,1000,5,10). The size of each square corresponds to the probability of observing the corresponding dataset under the null hypothesis. In a), the p-value is represented by the sum of the areas of the squares in the two shaded areas, containing all datasets with Score test statistics at least as extreme as the observed data. In b), we show how truncation at the upper 10−12 quantile of the fitted distribution of r0 and r1 would zero out many datasets, making calculation much quicker.

The AU approach, like the permutation approach, is completely general, and AU versions of any test can be implemented. We shall use standard Score, Wald, likelihood ratio and Firth tests.

2.3. Adjusting for covariates

Both permutation and AU tests permit adjustment for covariates through stratification, i.e. only using information about association from within groups of subjects for whom confounding factors (for example ancestry) are held constant (Clayton et al., 1993).

Extending the previous notation, for stratified tests we now refer to vectors m0,m1,r0,r1, each of length q, where q is the number of strata defined by the levels of one or more categorical covariates. Indexing strata by i, with 1iq, for each stratum i the stratified test enumerates all possible strata-specific datasets (m0i,m1i,r0i,r1i) such that r0i+r1i=r0i+r1i, computing a test statistic for each. The test statistics Ti(r0i,r1i) from each strata are combined (by default they are added) to produce a single test statistic for the whole dataset; formally we define

Tr0,r1=i=1qTi(r0i,r1i).

The p-value, which as before compares this single test statistic to what might have been observed under the null, uses the hypergeometric distribution for each set of stratum-specific counts. We write the probability of observing specific datasets as

fˇ(r0,r1;m0,m1,r0,r1)=i=1qfˇi(r0i,r1i;m0i,m1i,r0i,r1i)

Where

fˇi(r0i,r1i;m0i,m1i,r0i,r1i)=(r0i+r1ir1i)(m0i+m1ir0ir1im1ir1i)(m0i+m1im1i)

and formally define the p-value as

pstrat.perm(r0,r1;m0,m1,r0,r1)=r0,r1fˇ(r0,r1;m0,m1,r0,r1)1|Tr0,r1||Tr0,r1|,

i.e. the sum of probabilities of datasets that result in more extreme test statistics than Tobs, where Tobs is the test statistic corresponding to the data that was observed.

The stratified AU test is constructed from the same steps as the permutation except for three differences, described earlier in Section 2.2. First, the datasets considered for each strata include any values of 0r0im0i and 0r1im1i. Second, the probabilities fˇ of each dataset are constructed from fitting a null binomial model within each strata. Third, summands within each strata are zeroed-out for which the total contribution is no more than 2×1012.

Our approach removes confounding effects by using stratified analysis. Implemented carefully, there is little to choose between use of stratification versus model-based regression adjustment. In line with Clayton and Hills (1993, Statistical Methods in Epidemiology, pg 273) we find it appealing that the stratification approach forces careful consideration of a which confounders are a priori most important to adjust for, and for stratification approaches to be based closely on the scientific question of interest. Moreover, categorizing confounding into strata is the only approach under which our enumeration approach for exact inference is feasible; regression-based alternatives with continuously-valued covariates and standard computing resources would have to compute p-values by some form of Monte Carlo method, with consequent Monte Carlo error and long compute times. While a limitation of the approach is that a finite number of strata do have to be defined, in practice we have found that stratifying data into commonly-used ancestry groups (European, African-American, etc) is generally sufficient to account for confounding issues present in analyses of complex disease traits.

3. Analytical calculation results

To illustrate analytical calculations, we set the total sample size to be N=10,000 —close to that seen in Section 4’s example—and considered case:control matching ratios of 1:1, 1:3, and 1:19, fixing the quantities m0 and m1. For EMAC ranging from 1 to 100, we enumerated all plausible observed data values of r0 and r1 under the null hypothesis of independence between the minor allele and case:control status. A “plausible” dataset here was defined as one where the sum r0+r1 was in between the lower and upper 10−12 quantiles of the null Binomial distribution. Permutation and AU versions of Score, Wald, likelihood ratio and Firth tests were examined. For comparison we also computed the standard Score, Wald, likelihood ratio, Firth tests, and Fisher’s exact test, which is itself a permutation test. For permutation, AU, and standard tests we also considered a regularized Wald test, which avoids undefined test statistics by adding 0.5 to each cell count when any count is zero. We set the nominal significance level at α=5×108, and calculated the Type I error rate for each test as in Equation (1), across the range of EMAC. Although we do not consider power calculations under an alternative hypothesis in this section, these are given in Appendix B for the AU Firth test.

As seen in Figure 4, the tests based on standard asymptotics do not adequately control the Type I error rate. In the balanced design, the tests are overly conservative, with the exception of the likelihood ratio test, which is anti-conservative. The Score test has very a large Type I error rate under the 1:3 ratio, so is presented separately. This is also true under the 1:19 ratio for the Score and Wald tests, which are omitted. The other tests continue to be conservative, and the likelihood ratio test’s Type I error rate is too large over certain ranges. The Firth test consistently performs the best.

Figure 4:

Figure 4:

T1ER versus EMAC for various standard tests at α=5×108 (dotted line) and total N=10,000, grouped by matching ratio. All calculations use the zeroing-out technique of Section 2, and so understate the true T1ER by no more than 10−12. Score and Wald tests are omitted from the final plot due to gross violation of the nominal α.

In Figure 5, we see that the permutation tests improve upon most of the standard tests, though remain more conservative than the regular Firth test. While these tests have the advantage of being exact, as the case:control ratio becomes more unbalanced, the Type I error rate becomes more conservative. Under the 1:19 ratio, all tests perform nearly identically, with the exception of the unregularized Wald test.

Figure 5:

Figure 5:

T1ER versus EMAC for permutation versions of standard tests as described in Section 2.1 at α=5×108 (dotted line) and total N=10,000, grouped by matching ratio. All T1ER rate calculations use the zeroing-out technique of Section 2, and so understate the true T1ER by no more than. 10−12

In Figure 6, we see that the AU tests show a large improvement over standard and permutation tests, especially in the AU likelihood ratio and AU Firth tests. Though they are not exact, the excess Type I error rate is mild. Note that under the 1:19 ratio, the Firth and likelihood ratio tests perform identically.

Figure 6:

Figure 6:

Figure 6:

T1ER versus EMAC, for AU versions of standard tests as described in Section 2.2, at α=5×108 (dotted line) and total N=10,000, grouped by matching ratio. All T1ER rate calculations use the zeroing-out technique of Section 2, and so understate the true T1ER by no more than 10−12.

4. Application: Rhabdomyolysis case-control study

The data comes from an exome-sequencing study, in which 9,763 subjects who used statins were considered; 211 cases with rhabdomyolysis and 9,552 controls. The original datatset consists of 2,194,116 variants sequenced. The rationale for this design are described in detail by Marciante et al. (2011). Our interest was primarily in assessing if there existed rare genetic variants associated with developing rhabdomyolysis in statin users. We defined ‘rare’ variants as those where less than or equal to 100 study participants carried the minor allele. Variants with less than 5 minor allele carriers were also removed, as these provide no ability to produce significant values at the low α threshold used in this form of study. Finally, for quality control, we filtered out variants with a genotyping rate of less than 0.85. We did not filter variants using the Hardy-Weinberg test. Applying these filters left 161,428 variants, and there are no covariates for which to adjust in this analysis.

We applied the AU and permutation versions of the likelihood ratio test, and the permutation and standard Firth test to all variants. The entire analysis took approximately 6.5 hours on a shared server, using a single CPU. The AU version of the Firth test was not used due to its high computational burden. The resulting QQ plot and a plot of the inflation (45 degree rotated QQ plot) observed are given in Figure 7.

Figure 7:

Figure 7:

QQ plot and 45 degree rotated QQ plot of -log10 p-values for rhabdomyolysis dataset, as described in Section 4. After quality control filtering, 161,428 variants are analyzed, with between 5 and 100 minor allele carriers each. For each method, the QQ plot shows the ordered p-values versus the corresponding expected value from null, i.e. Uniform(0,1) p-values. The blue cone shape indicates pointwise 95% prediction bounds for each ordered p-value. The rotated plot shows the same results, but where the y-axis shows the -log10 observed p-value minus the -log10 expected p-value; the blue cone has the same interpretation as before.

Based on our numerical results, it is reasonable to expect that the AU likelihood ratio test provides the best control of the Type I error rate. Applied to this dataset, while some granularity in the larger p-values is present on the left hand of both plots, we observe that the AU likelihood ratio test results in significantly less inflation than the standard Firth test. Therefore, although no variant was found to have an AU p-value less than α=5×108, we believe that the variants which deviate from the expected null p-value distribution under the AU test are more likely to be of interest than those under the other methods.

5. Discussion

We have developed and implemented association tests for rare genetic variants, that control Type I error rates better than standard asymptotic tests. Of the tests proposed in this paper, the AU version of the likelihood ratio and Firth tests perform the best, particularly when the ratio of cases to controls is extreme. However, the AU version of the Firth test has a notably higher computational burden than competitors. Therefore, we recommend the the AU likelihood ratio test, for large genome-wide studies. If an exact test is necessary, then a permutation test is recommended; though conservative, it tends to show an improvement over standard tests. We note that if the expected number of minor allele carriers is less than 20, then no test will perform adequately, and conservative control of the Type I error rate is the best achievable property. More generally, we note that in high-throughput rare variant work, unless effect sizes are large, power willl be limited even when Type I error rates are not controlled conservatively. As pointed out by a reviewer, in such settings an appealing alternative form analysis tests for association across a group of variants; examples include SKAT (Wu et al, 2011) but also “burden” tests, that collapse genotype across a region to a univariate measure. These approaches can improve power over single-variant approaches, by both combining multiple associaiton “signals”, but also by reducing the multiple testing burden. However, the inference they provide is region-specific and not variant-specific, leaving (for example) no strong indication of which variants may be causal. While not explored here, our permutation tests could be used with CAST (Morgenthaler and Thilly, 2007), a burden test that collapses genotype to presence/absence of a particular class of variants.

The methods described here have been implemented in an R package, AUtests. This package contains the functions basic.tests, perm.tests, and au.tests, which implement all the respective standard, permutation, and AU tests for a given vector of counts (m0,m1,r0,r1), returning a vector of p-values. The AU Firth test is implemented in a separate function, au.firth, due to its increased computational time. For a typical dataset(m0,m1,r0,r1) =(10000,10000.50,50), on a standard laptop, the basic.tests function takes 0.03 seconds of CPU time, the perm.tests function takes 0.21 seconds, the au.tests function takes 0.39 seconds, and the au.firth function takes 51 seconds. To account for covariates, appropriately categorized, the package also contains the functions au.test.strat and perm.test.strat, which implement stratified AU and permutation likelihood ratio tests. The package is available on CRAN.

6. Acknowledgments

Research reported in this paper was supported by the National Institute on Aging of the National Institutes of Health under award numbers U01AG049505 and U01AG049507, and by the National Heart, Lung, and Blood Institute of the National Institutes of Health under award number R01HL078888. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

A. Appendix: exact control of permutation tests

In this appendix, we show that permutation tests give exact control of the Type I error rate. Rewriting equation (1) as a double summation over the observed minor allele count (t:=r0+r1) and the number of these in the controls, we obtain

T1ER(α)=0tm0+m1g(t;m0,m1)r0f(r0;m0,m1,t)1p(r0,r1;m0,m1)<α

where g() denotes the probability of the observed minor allele count, and f() gives the probability of the observed counts in cases and controls given the minor allele count r0+r1 --- so f() supports values of r0 between max(0,tm1) and min(m0,t).

By construction, the inner sum always gives a value less than or equal to α; the outer sum averages these, and so is similarly bounded. However, particularly for rare variants, the inner sum considers a small set of possible permutations, as illustrated in Figure 2. While this makes the test fast enough that zeroing-out is not required, it means that for small α, the actual Type I error rate, while below α, will be quite conservative for many values of m0 and m1.

B. Appendix: AU test power calculations

In this section, we show power calculations for the AU Firth test, under the same matching ratios considered in the main paper and over the same range of EMAC. Specifically, we perform analytical calculations similarly as in Section 3, but consider data probabilities from a Binomial distribution where having a minor allele is related to case:control status through an odds ratio. We consider odds ratios ranging from 1.5 to 5. In Figure 8, we observe that power decreases as case:control ratios become more skewed. In particular, the extremely unbalanced 1:19 ratio requires a very large association in order to have meaningful power, even at higher minor allele counts.

Figure 8:

Figure 8:

Power curves giving the probability of rejecting the null hypothesis of independence by expected minor allele count. Different curves correspond to different odds ratios. Each panel corresponds to a different case:control matching ratio with an overall sample size of 20,000. Left: 1:1, middle: 1:3, right: 1:19.

Footnotes

Method development: KMR, AS; Method implementation: AS; Data analysis: AS, Writing: AS, KMR

7 Conflict of interest statement

The authors have no conflicts of interest to declare.

References

  1. Anderson MJ. (2001). Permutation tests for univariate or multivariate analysis of variance and regression. Canadian Journal of Fisheries and Aquatic Sciences, 58(3):626–639. [Google Scholar]
  2. Boyett JM. and Shuster JJ. (1977). Nonparametric one-sided tests in multivariate analysis with medical applications. Journal of the American Statistical Association, 72(359):665–668. [Google Scholar]
  3. Clayton D, Hills M, and Pickles A (1993). Statistical models in epidemiology, volume 161 IEA. [Google Scholar]
  4. Danjou F, Zoledziewska M, Sidore C, Steri M, Busonero F, Maschio A, Mulas A, Perseu L, Barella S, Porcu E, Pistis G, Pitzalis M, Pala M, Menzel S, Metrustry S, Spector T, Leoni L, Angius A, Uda M, Moi P, Thein S, Galanello R, Abecasis G, Schlessinger D, Sanna S, and Cucca F (2015). Genome-wide association analyses based on whole-genome sequencing in sardinia provide insights into regulation of hemoglobin levels. Nature Genetics. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Firth D (1993). Bias reduction of maximum likelihood estimates. Biometrika, 80(1):27–38. [Google Scholar]
  6. Good P (2005). Permutation, Parametric and Bootstrap Tests of Hypotheses. Springer; New York. [Google Scholar]
  7. Heinze G, Ploner M, Dunkler D, and Southworth H (2013). logistf: Firth’s bias reduced logistic regression. R package version 1.21. [Google Scholar]
  8. Hoffmann TJ, Van Den Eeden SK, Sakoda LC, Jorgenson E, Habel LA, Graff RE, Passarelli MN, Cario CL, Emami NC, Chao CR, Ghai NR, Shan J, Ranatunga DK, Quesenberry CP, Aaronson D, Presti J, Wang Z, Berndt SI, Chanock SJ, McDonnell SK, French AJ, Schaid DJ, Thibodeau SN, Li Q, Freedman ML, Penney KL, Mucci LA, Haiman CA, Henderson BE, Seminara D, Kvale MN, Kwok P-Y, Schaefer C, Risch N, and Witte JS. (2015). A large multiethnic genome-wide association study of prostate cancer identifies novel risk variants and substantial ethnic differences. Cancer Discovery, 5(8):878–891. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Hoggart CJ, Clark TG, De Iorio M, Whittaker JC, and Balding DJ. (2008). Genome-wide significance for dense snp and resequencing data. Genetic Epidemiology, 32(2):179–185. [DOI] [PubMed] [Google Scholar]
  10. Huo M, Heyvaert M, den Noortgate WV, and Onghena P. (2014). Permutation tests in the educational and behavioral sciences. Methodology, 10(2):43–59. [Google Scholar]
  11. Janssen A and Pauls T (2003). How do bootstrap and permutation tests work? Annals of statistics, pages 768–806. [Google Scholar]
  12. Li B and Leal SM. (2008). Methods for detecting associations with rare variants for common diseases: Application to analysis of sequence data. The American Journal of Human Genetics, 83(3):311–321. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Ma C, Blackwell T, Boehnke M, Scott LJ, and the GoT2D investigators (2013). Recommended joint and meta-analysis strategies for case-control association testing of single low-count variants. Genetic Epidemiology, 37(6):539–550. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Marciante KD, Durda JP, Heckbert SR, Lumley T, Rice K, McKnight B, Totah RA, Tamraz B, Kroetz DL, Fukushima H, et al. (2011). Cerivastatin, genetic variants, and the risk of rhabdomyolysis. Pharmacogenetics and genomics, 21(5):280. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Morgenthaler S, & Thilly WG (2007). A strategy to discover genes that carry multi-allelic or mono-allelic risk for common diseases: a cohort allelic sums test (CAST). Mutation Research/Fundamental and Molecular Mechanisms of Mutagenesis, 615(1), 28–56. [DOI] [PubMed] [Google Scholar]
  16. Nichols TE. and Holmes AP. (2002). Nonparametric permutation tests for functional neuroimaging: A primer with examples. Human Brain Mapping, 15(1):1–25. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Opherk C, Gonik M, Duering M, Malik R, Jouvent E, Hervé D, Adib-Samii P, Bevan S, Pianese L, Silvestri S, Dotti MT, De Stefano N, Liem M, Boon EM, Pescini F, Pachai C, Bracoud L, Müller-Myhsok B, Meitinger T, Rost N, Pantoni L, Lesnik Oberstein S, Federico A, Ragno M, Markus HS, Tournier-Lasserve E, Rosand J, Chabriat H, and Dichgans M. (2014). Genome-wide genotyping demonstrates a polygenic risk score associated with white matter hyperintensity volume in cadasil. Stroke, 45(4):968–972. [DOI] [PubMed] [Google Scholar]
  18. Pe’er I, Yelensky R, Altshuler D, and Daly MJ. (2008). Estimation of the multiple testing burden for genomewide association studies of nearly all common variants. Genetic epidemiology, 32(4):381–385. [DOI] [PubMed] [Google Scholar]
  19. Pitman EJG. (1937). Significance tests which may be applied to samples from any populations. Supplement to the Journal of the Royal Statistical Society, 4(1):119–130. [Google Scholar]
  20. Storer BE. and Kim C. (1990). Exact properties of some exact test statistics for comparing two binomial proportions. Journal of the American Statistical Association, 85(409):pp. 146–155. [Google Scholar]
  21. Verhaaren BF, Debette S, Bis JC, Smith JA, Ikram MK, Adams HH, Beecham AH, Rajan KB, Lopez LM, Barral S, van Buchem MA, van der Grond J, Smith AV, Hegenscheid K, Aggarwal NT, de Andrade M, Atkinson EJ, Beekman M, Beiser AS, Blanton SH, Boerwinkle E, Brickman AM, Bryan RN, Chauhan G, Chen CP, Chouraki V, de Craen AJ, Crivello F, Deary IJ, Deelen J, De Jager PL, Dufouil C, Elkind MS, Evans DA, Freudenberger P, Gottesman RF, Guðnason V, Habes M, Heckbert SR, Heiss G, Hilal S, Hofer E, Hofman A, Ibrahim-Verbaas CA, Knopman DS, Lewis CE, Liao J, Liewald DC, Luciano M, van der Lugt A, Martinez OO, Mayeux R, Mazoyer B, Nalls M, Nauck M, Niessen WJ, Oostra BA, Psaty BM, Rice KM, Rotter JI, von Sarnowski B, Schmidt H, Schreiner PJ, Schuur M, Sidney SS, Sigurdsson S, Slagboom PE, Stott DJ, van Swieten JC, Teumer A, TÃ, glhofer AM, Traylor M, Trompet S, Turner ST, Tzourio C, Uh H-W, Uitterlinden AG, Vernooij MW, Wang JJ, Wong TY, Wardlaw JM, Windham BG, Wittfeld K, Wolf C, Wright CB, Yang Q, Zhao W, Zijdenbos A, Jukema JW, Sacco RL, Kardia SL, Amouyel P, Mosley TH, Longstreth WT, DeCarli CC, van Duijn CM, Schmidt R, Launer LJ, Grabe HJ, Seshadri SS, Ikram MA, and Fornage M. (2015). Multiethnic genome-wide association study of cerebral white matter hyperintensities on mri. Circulation: Cardiovascular Genetics, 8(2):398–409. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Wu MC, Lee S, Cai T, Li Y, Boehnke M, & Lin X (2011). Rare-Variant Association Testing for Sequencing Data with the Sequence Kernel Association Test. American Journal of Human Genetics, 89(1), 82–93. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Xing G, Lin C-Y, Wooding SP, and Xing C. (2012). Blindly using wald’s test can miss rare disease-causal variants in case-control association studies. Annals of Human Genetics, 76(2):168–177. [DOI] [PubMed] [Google Scholar]

RESOURCES