Abstract
The use of weights provides an effective strategy to incorporate prior domain knowledge in large-scale inference. This paper studies weighted multiple testing in a decision-theoretic framework. We develop oracle and data-driven procedures that aim to maximize the expected number of true positives subject to a constraint on the weighted false discovery rate. The asymptotic validity and optimality of the proposed methods are established. The results demonstrate that incorporating informative domain knowledge enhances the interpretability of results and precision of inference. Simulation studies show that the proposed method controls the error rate at the nominal level, and the gain in power over existing methods is substantial in many settings. An application to a genome-wide association study is discussed.
Keywords: Class weights, Decision weights, Multiple testing with groups, Prioritized subsets, Value to cost ratio, Weighted p-value.
1. Introduction
In large-scale studies, relevant domain knowledge, such as external covariates, scientific insights and prior data, is often available alongside the primary data set. Exploiting such information in an efficient manner promises to enhance both the interpretability of scientific results and precision of statistical inference. In multiple testing, the hypotheses being investigated often become “unequal” in light of external information, which may be reflected by differential attitudes towards the relative importance of testing units or the severity of decision errors. The use of weights provides an effective strategy to incorporate informative domain knowledge in large-scale testing problems.
In the literature, various weighting methods have been advocated for a range of multiple comparison problems. A popular scheme, referred to as the decision weights or loss weights approach, involves modifying the error criteria or power functions in the decision process (Benjamini and Hochberg, 1997). The idea is to employ two sets of positive constants a = {ai : i = 1,⋯ ,m} and b = {bi : i = 1,⋯ ,m} to take into account the costs and gains of multiple decisions. Typically, the choice of the weights a and b reflects the degree of confidence one has toward prior beliefs and external information. It may also be pertinent to the degree of preference that one has toward the consequence of one class of erroneous/correct decisions over another class based on various economical and ethical considerations. For example, in the spatial cluster analysis considered by Benjamini and Heller (2007), the weighted false discovery rate was used to reflect that a false positive cluster with larger size would account for a larger error. Another example arises from genome-wide association studies (GWAS), where prior data or genomic knowledge, such as prioritized subsets (Lin and Lee, 2012), allele frequencies (Lin et al., 2014) and expression quantitative trait loci information (Li et al., 2013), can often help to assess the scientific plausibility of significant associations. To incorporate such information in the analysis, a useful strategy is to up-weight the gains for the discoveries in preselected genomic regions by modifying the power functions in respective testing units (P˜ena et al., 2011; Sun et al., 2015; He et al., 2015). We assume in this paper that the weights have been pre-specified by the investigator. This is a reasonable assumption in many practical settings. For example, weights may be assigned according to economical considerations (Westfall and Young, 1993), external covariates (Benjamini and Heller, 2007; Sun et al., 2015) and biological insights from prior studies (Xing et al., 2010).
We mention two alternative formulations for weighted multiple testing. One popular method, referred to as the procedural weights approach, involves the adjustment of the pvalues from individual tests. In GWAS, Roeder et al. (2006) and Roeder and Wasserman (2009) proposed to utilize linkage signals to up-weight the p-values in preselected regions and down-weight the p-values in other regions. It was shown that the power to detect association can be greatly enhanced if the linkage signals are informative, yet the loss in power is small when the linkage signals are uninformative. Another useful weighting scheme, referred to as the class weights approach, involves allocating varied test levels to different classes of hypotheses. For example, in analysis of the growth curve data (Box, 1950), Westfall and Young (1993, page 186) proposed to allocate a higher family-wise error rate (FWER) to the class of hypotheses related to the primary variable “gain” and a lower FWER to the secondary variable “shape”.
We focus on the decision weights approach in the present paper. This weighting scheme is not only practically useful for a wide range of applications, but also provides a powerful framework that enables a unified investigation of various weighting methods. Specifically, the proposal in Benjamini and Hochberg (1997) involves the modification of both the error rate and power function. The formulation is closely connected to the classical ideas in compound decision theory that aim to optimize the tradeoffs between the gains and losses when many simultaneous decisions are combined as a whole. Our theory reveals that if the goal is to maximize the power subject to a given error rate, then the modifications via decision weights would lead to improved multiple testing methods with sensible procedural weights or class weights, or both. For example, in GWAS, the investigators can up-weight the power functions for discoveries in genomic regions that are considered to be more scientific plausible or biologically meaningful; this would naturally up-weight the p-values in these regions and thus yield weighting strategies similar to those suggested by Roeder and Wasserman (2009). In large clinical trials, modifying the power functions for respective rejections at the primary and secondary end points would correspond to the allocation of varied test levels across different classes of hypotheses, leading to weighting strategies previously suggested by Westfall and Young (1993).
The false discovery rate (FDR; Benjamini and Hochberg, 1995) has been widely used in large-scale multiple testing as a practical and powerful error criterion. Following Benjamini and Hochberg (1997), we generalize the FDR to weighted false discovery rate (wFDR), and develop optimal procedures for wFDR control under the decision weights framework. We first construct an oracle procedure that maximizes the weighted power function subject to a constraint on the wFDR, and then develop a data-driven procedure to mimic the oracle and establish its asymptotic optimality. The numerical results show that the proposed method controls the wFDR at the nominal level, and the gain in power over existing methods is substantial in many settings. Our optimality result marks a clear departure from existing works in the literature on covariate-assisted inference, which aims to incorporate the external information by deriving optimal procedural weights for p-values (Roeder and Wasserman, 2009; Roquain and van de Wiel, 2009) or constructing covariate-adjusted test statistics (Ferkingstad et al. 2008; Zablocki et al. 2014; Cai et al. 2016) subject to the unweighted FDR and power function.
Our research also makes a novel contribution to the theory of optimal ranking in multiple testing. Conventionally, a multiple testing procedure operates in two steps: ranking the hypotheses according to their significance levels and then choosing a cutoff along the rankings. It is commonly believed that the rankings remain the same universally at all FDR levels. For example, the ranking based on p-values or adjusted p-values in common practice is invariant to the choice of the FDR threshold. The implication of our theory is interesting, for it claims that there does not exist a ranking that is universally optimal at all test levels. Instead, the optimal ranking of hypotheses depends on the pre-specified wFDR level. That is, the hypotheses may be ordered differently when different wFDR levels are chosen. This point is elaborated in Section A of the Supplementary Material. See also Roeder and Wasserman (2009) and Roquain and van de Wiel (2009), where the FDR level α is included in constructing the procedural weights.
The rest of the article is organized as follows. Section 2 discusses a general framework for weighted multiple testing. Sections 3 and 4 develop oracle and data-driven wFDR procedures and establish their optimality properties. Practice guidelines and numerical results are presented in Section 5. Section 6 concludes the article with a discussion of related and future works. Proof of Theorem 1 is given in Section 7. Proofs of other theoretical results and additional discussions are provided in the Supplementary Material.
2. Problem Formulation
This section discusses a decision weights framework for weighted multiple testing. We first introduce the model and notation and then discuss modified error criteria and power functions.
2.1. Model and notation
Suppose that m hypotheses H1,⋯ ,Hm are tested simultaneously based on observations X1,⋯ ,Xm. Let θ = (θ1,⋯ ,θm) ∈ {0,1}m denote the true state of nature, where 0 and 1 stand for null and non-null cases, respectively. Assume that observations Xi are independent and distributed according to the following model
(2.1) |
where F0i and F1i are the null and non-null distributions for Xi, respectively. Denote by f0i and f1i the corresponding density functions. Suppose that the unknown states θi are Bernoulli (pi) variables, where pi = P(θi = 1). The mixture density is denoted by f·i = (1 − pi)f0i + pif1i.
Consider the widely used random mixture model (Efron et al., 2001; Storey, 2002; Genovese and Wasserman, 2002)
(2.2) |
This model, which assumes that all observations are identically distributed according to a common distribution F, can sometimes be unrealistic in applications. In light of domain knowledge, the observations are likely to have different distributions. For example, in the context of a brain imaging study, Efron (2008) showed that the proportions of activated voxels are different for the front and back halves of a brain. In GWAS, certain genomic regions contain higher proportions of significant signals than other regions. In the adequate yearly progress study of California high schools (Rogosa, 2003), the densities of z-scores vary significantly from small to large schools. We shall develop theory and methodology for model (2.1), which allows possibly different non-null proportions and densities and is applicable to a wide range of settings.
The multi-group model considered in Efron (2008) and Cai and Sun (2009), which has been widely used in applications, is an important case of the general model (2.1). The multi-group model assumes that the observations can be divided into K groups. Let denote the index set of the observations in group k, k = 1,⋅,K. For each i ∈ , θi is distributed as Bernoulli(pk), and Xi follows a mixture distribution:
(2.3) |
where f0k and f1k are the null and non-null densities for observations in group k. This model will be revisited in later sections. See also Ferkingstad et al. (2008), Hu et al. (2010) and Liu et al. (2016) for related works on multiple testing with groups.
2.2. Weighted error criterion and power function
This section discusses a generalization of the FDR criterion in the context of weighted multiple testing. Denote the decisions for the m tests by δ = (δ1,⋯ ,δm) ∈ {0,1}m, where δi = 1 indicates that Hi is rejected and δi = 0 otherwise. Let ai be the weight indicating the severity of a false positive decision. For example, ai is taken as the cluster size in the spatial cluster analyses conducted in Benjamini and Heller (2007) and Sun et al. (2015). As a result, rejecting a larger cluster erroneously corresponds to a more severe decision error. To incorporate domain knowledge in multiple testing, Benjamini and Hochberg (1997) defined the weighted FDR wFDRBH = E{Q(a)}, where
(2.4) |
We consider a slightly different version of the wFDR:
(2.5) |
In Section D of the Supplementary Material, we show that the definitions in (2.4) and (2.5) are asymptotically equivalent. The main consideration of using (2.5) is to facilitate our theoretical derivations and obtain exact optimality results.
To assess the effectiveness of different wFDR procedures, we define the expected number of true positives
(2.6) |
where bi is the weight indicating the power gain when Hi is rejected correctly. The use of bi provides an effective scheme to incorporate domain knowledge. In GWAS, larger bi can be assigned to pre-selected genomic regions to reflect that the discoveries in these regions are more biologically meaningful. In spatial data analysis, correctly identifying a larger cluster that contains signal may correspond to a larger bi, indicating a greater gain.
By combining the concerns on both the error criterion and power function, the goal in weighted multiple testing is to
(2.7) |
The optimal solution to (2.7) is studied in the next section.
3. Oracle Procedure for wFDR Control
The basic framework of our theoretical and methodological developments is outlined as follows. In Section 3.1, we assume that pi, f0i, and f·i in the mixture model (2.1) are known by an oracle and derive an oracle procedure that maximizes the ETP subject to a constraint on the wFDR. Connections to the literature is given in Section 3.2. In Section 4, we develop a data-driven procedure to mimic the oracle and establish its asymptotic validity and optimality.
3.1. Oracle procedure
The derivation of the oracle procedure involves two key steps: the first is to derive the optimal ranking of hypotheses and the second is to determine the optimal threshold along the ranking that exhausts the pre-specified wFDR level. We discuss the two issues in turn. Consider model (2.1). Define the local false discovery rate (Lfdr, Efron et al. 2001) as
(3.1) |
The wFDR problem (2.7) is equivalent to the following constrained optimization problem
(3.2) |
Let S− = {i : Lfdri ≤ α} and S+ = {i : Lfdri > α}. Then the constraint in (3.2) can be equivalently expressed as
(3.3) |
Consider an optimization problem which involves packing a knapsack with a capacity given by the right hand side of equation (3.3). Every available object has a known value and a known cost (of space). Clearly rejecting a hypothesis in S− is always beneficial as it allows the capacity to expand, and thus promotes more discoveries. Hence the key issue boils down to how to efficiently utilize the capacity (after all hypotheses in S− are rejected) to make as many discoveries as possible in S+. Observe that each rejection in S+ would simultaneously increase the power and decrease the capacity. Intuitively, we should sort all hypotheses in S+ in an decreasing order of the value to cost ratio (VCR). Equations (3.2) and (3.3) suggest that
(3.4) |
To maximize the power, the ordered hypotheses are rejected sequentially until maximum capacity is reached.
The above considerations motivate us to consider the following class of decision rules , where
(3.5) |
We briefly explain some important operational characteristics of testing rule (3.5). First, if we let t > 0, then the equation implies that for all i ∈ S−; hence all hypotheses in S− are rejected as desired. (This explains why the VCR is not used directly in (3.5), given that the VCR is not meaningful in S−.) Second, a solution path can be generated as we vary t continuously from large to small. Along the path δ∗(t) sequentially rejects the hypotheses in S+ according to their ordered VCRs. Denote by H(1),⋯ ,H(m) the hypotheses sequentially rejected by δ∗. (The actual ordering of the hypotheses within S− does not matter in the decision process since all are always rejected.)
The next task is to choose a cutoff along the ranking to achieve exact wFDR control. The difficulty is that the maximum capacity may not be attained by a sequential rejection procedure. To exhaust the wFDR level, we permit a randomized decision rule. Denote the Lfdr values and the weights corresponding to H(i) by Lfdr(i), a(i), and b(i). Let
(3.6) |
denote the capacity up to jth rejection. According to the constraint in equation (3.2), we choose k = max{j : C(j) ≤ 0} so that the capacity is not yet reached when H(k) is rejected but would just be exceeded if H(k+1) is rejected. The idea is to split the decision point at H(k+1) by randomization.
Let U be a Uniform (0,1) variable that is independent of the truth, the observations, and the weights. Define
Let be an indicator, which takes value 1 if event A occurs and 0 otherwise. We propose the oracle decision rule , where
(3.7) |
Remark 1 The randomization step is only employed for theoretical considerations to enforce the wFDR to be exactly α. Thus the optimal power can be effectively characterized. Moreover, only a single decision point at H(k+1) is randomized, which has a negligible effect in large-scale testing problems. We do not pursue randomized rules for the data-driven procedures developed in later sections.
Let wFDR(δ) and ETP(δ) denote the wFDR and ETP of a decision rule δ, respectively. Theorem 1 shows that the oracle procedure (3.7) is valid and optimal for wFDR control.
Theorem 1 Consider model (2.1) and oracle procedure δOR defined in (3.7). Let be the collection of decision rules such that for any δ ∈ , wFDR(δ) ≤ α. Then we have
-
(i).
wFDR(δOR) = α.
-
(ii).
ETP(δOR) ≥ ETP(δ) for all δ ∈ .
3.2. Comparison with the optimality results in Spjøtvoll (1972), Benjamini and Hochberg (1997) and Storey (2007)
Spjøtvoll (1972) showed that the likelihood ratio (LR) statistic
(3.8) |
is optimal for the following multiple testing problem
(3.9) |
where ∩H0i and ∩H1i denote the intersections of the nulls and non-nulls, respectively. The error criterion E∩H0i is referred to as the intersection tests error rate (ITER).
A weighted version of problem (3.9) was considered by Benjamini and Hochberg (1997), where the goal is to
(3.10) |
The optimal solution to (3.10) is given by the next proposition.
Proposition 1 (Benjamini and Hochberg, 1997). Define the weighted likelihood ratio (WLR)
(3.11) |
Then the optimal solution to (3.10) is a thresholding rule of the form , where tα is the largest threshold that controls the weighted ITER at level α.
The ITER is very restrictive in the sense that the expectation is taken under the conjunction of the null hypotheses. The ITER is inappropriate for mixture model (2.1) where a mixture of null and non-null hypotheses are tested simultaneously. To extend intersection tests to multiple tests, define the per family error rate (PFER) as
(3.12) |
The power function should be modified correspondingly. Therefore the goal is to
(3.13) |
The key difference between the ITER and PFER is that the expectation in (3.12) is now taken over all possible combinations of the null and non-null hypotheses. The optimal PFER procedure is given by the next proposition.
Proposition 2 Consider model (2.1) and assume continuity of the LR statistic. Let be the collection of decision rules such that for every δ ∈ , PFER(δ) ≤ α. Define the weighted posterior odds (WPO)
(3.14) |
Denote by QPF (t) the PFER of . Then the oracle PFER procedure is , where and tPF = sup{t : QPF (t) ≤ α}.
This oracle rule satisfies:
-
(i).
ETP(δPF ) = α.
-
(ii).
ETP(δPF ) ≥ ETP(δ) for all δ ∈ .
Storey (2007) proposed the optimal discovery procedure (ODP) that aims to maximize the ETP subject to a constraint on the expected number of false positives (EFP). The ODP extends the optimality result in Spjøtvoll (1972) from the intersection tests to multiple tests. However, the formulation of ODP does not incorporate weights. As a result, the ODP procedure is a symmetric rule where all hypotheses are exchangeable. Symmetric rules are in general not suitable for weighted multiple testing where it is desirable to incorporate external information and treat the hypotheses differently.
Our formulation (2.7) modifies the conventional formulations in (3.10) and (3.13) to the multiple testing situation with an FDR type criterion. These modifications lead to methods that are more suitable for large-scale scientific studies. The oracle procedure (3.7) uses the VCR (3.4) to rank the hypotheses and is an asymmetric rule. The VCR, which optimally combines the decision weights, significance measure (Lfdr) and test level α, produces a more powerful ranking than the WPO (3.14) in the wFDR problem. Section A in the supplementary material provides a detailed discussion on the ranking issue.
4. Data-Driven Procedures and Asymptotics
The oracle procedure (3.7) cannot be implemented in practice since it relies on unknown quantities such as Lfdri and t∗. This section develops a data-driven procedure to mimic the oracle. We first propose a test statistic to rank the hypotheses and discuss related estimation issues. A step-wise procedure is then derived to determine the best cutoff along the ranking. Finally, asymptotic results on the validity and optimality of the proposed procedure are presented.
4.1. Proposed test statistic and its estimation
The oracle procedure utilizes the ranking based on the VCR (3.4). However, the VCR is only meaningful for the tests in S+ and becomes problematic when both S− and S+ are considered. Moreover, the VCR could be unbounded, which would lead to difficulties in both numerical implementations and technical derivations. We propose to rank the hypotheses using the following statistic (in increasing values)
(4.1) |
As shown in the next proposition, Ri always ranks hypotheses in S− higher than hypotheses in S+ (as desired), and yields the same ranking as that by the VCR (3.4) for hypotheses in S+. The other drawbacks of VCR can also be overcome by Ri: Ri is always bounded in the interval [−1,1] and is a continuous function of the Lfdri.
Proposition 3 (i) The rankings generated by the decreasing values of VCR (3.4) and increasing values of Ri (4.1) are the same in both S− and S+. (ii) The ranking based on increasing values of Ri always puts hypotheses in S− ahead of hypotheses in S+.
Next we discuss how to estimate Ri; this involves the estimation of the Lfdr statistic (3.1), which has been studied extensively in the multiple testing literature. We give a review of related methodologies. If all observations follow a common mixture distribution (2.2), then we can first estimate the non-null proportion p and the null density f0 using the methods in Jin and Cai (2007), and then estimate the mixture density f using a standard kernel density estimator (e.g. Silverman, 1986). If all observations follow a multi-group model (2.3), then we can apply the above estimation methods to separate groups to obtain corresponding estimates , and , k = 1,⋯ ,K. The theoretical properties of these estimators have been established in Sun and Cai (2007) and Cai and Sun (2009). In practice, estimation problems may arise from more complicated models. Related theories and methodologies have been studied in Storey (2007), Ferkingstad et al. (2008), and Efron (2008, 2010); theoretical supports for these estimators are yet to be developed.
The estimated Lfdr value for Hi is denoted by . By convention, we take if . This modification only facilitates the development of theory and has no practical effect on the testing results (since rejections are essentially only made for small ’s). The ranking statistic Ri can therefore be estimated as
(4.2) |
The performance of the data driven procedure relies on the accuracy of the estimate ; some technical conditions are discussed in the next subsection. The finite sample performance of different Lfdr estimates are investigated in Section E.4 of the Supplementary Material.
4.2. Proposed testing procedure and its asymptotic properties
Consider defined in (4.2). Denote by the estimate of excessive error rate when Hi is rejected. Let be the ordered test statistics (in increasing values). The hypothesis and estimated excessive error rate corresponding to are denoted by H(i) and . The idea is to choose the largest cutoff along the ranking based on so that the maximum capacity is reached. Motivated by the constraint in (3.2), we propose the following step-wise procedure.
Procedure 1 (wFDR control with general weights). Rank hypotheses according to in increasing values. Let . Reject H(i), for i = 1,...,k.
It is important to note that in Procedure 1, is used in the ranking step whereas (or a weighted transformation of ) is used in the thresholding step. The ranking by is in general different from that by . In some applications where the weights are proportional, i.e. a = c · b for some constant c > 0, then the rankings by and are identical. Specifically is then monotone in . Further, choosing the cutoff based on is equivalent to that of choosing by a weighted . This leads to an Lfdr based procedure (Sun et al., 2015), which can be viewed as a special case of Procedure 1.
Procedure 2 (wFDR control with proportional weights). Rank hypotheses according to in increasing values. Denote the hypotheses and weights corresponding to by H(i) and a(i). Let
Reject H(i), for i = 1,...,k.
Next we investigate the asymptotic performance of Procedure 1. We first give some regularity conditions for the weights. Our theoretical framework requires that the decision weights must be obtained from external sources such as prior data, biological insights, or economical considerations. In particular, the observed data {Xi : i = 1,⋯ ,m} cannot be used to derive the weights. The assumption is not only crucial in theoretical developments, but also desirable in practice (to avoid using data twice). Therefore given the domain knowledge, the decision weights do not depend on observed values. Moreover, a model with random (known) weights is employed for technical convenience, as done in Genovese et al. (2006) and Roquain and van de Wiel (2009). We assume that the weights are independent with each other across testing units. Formally, denote ei the external domain knowledge for hypothesis i, we require the following condition.
Condition 1 (i) for 1 ≤ i ≤ m. (ii) (ai,bi) and (aj,bj) are independent for i ≠ j.
In weighted multiple testing problems, the analysis is always carried out in light of the external information ei implicitly. The notation of conditional distribution on ei will be suppressed when there is no ambiguity. In practice, the weights ai and bi are usually bounded. We need a weaker condition in our theoretical analysis.
Condition 2 (Regularity conditions on the weights.) Let C and c be two positive constants. E(supi ai) = o(m),E(supi bi) = o(m), , and min{E(ai),E(bi)} ≥ c.
A consistent Lfdr estimate is needed to ensure the large-sample performance of the data-driven procedure. Formally, we need the following condition.
Condition 3 . Also, , where Lfdr is an independent copy of Lfdri.
Remark 2 Condition 3 is a reasonable assumption in many applications. We give a few important scenarios where Condition 3 holds. Suppose we observe z-values from random mixture model
where (μi,σi) = (μ0,σ0) if θi = 0, (μi,σi) ≠ (μ0,σ0) if θi = 1, and F(μ,σ) is a general bivariate distribution. Let p denote the proportion of non-null cases. Then , the estimator proposed in Jin and Cai (2007), satisfies under mild regularity conditions. Moreover, it is known that the kernel density estimator satisfies . It follows from Sun and Cai (2007) that Condition 3 holds when the above estimators are used. If the null distribution is unknown, then we can use the method in Jin and Cai (2007) to estimate the null parameters (μ0,σ0). Then under certain regularity conditions, we can show that , and Condition 3 holds. For the multi-group model (2.3), let , and be estimates of pk, fk0, and fk such that k = 1,⋯ ,K. Let if . It follows from Cai and Sun (2009) that Condition 3 holds when we apply Jin and Cai’s estimators and standard kernel estimates to the groups separately.
The oracle procedure (3.7) provides an optimal benchmark for all wFDR procedures. The next theorem establishes the asymptotic validity and optimality of the data-driven procedure by showing that the wFDR and ETP levels of the data-driven procedure converge to the oracle levels as m → ∞.
Theorem 2 Assume Conditions 1–3 hold. Denote by wFDRDD the wFDR level of the data-driven procedure (Procedure 1). Let ETPOR and ETPDD be the ETP levels of the oracle procedure (3.7) and data-driven procedure, respectively. Then we have
-
(i).
wFDRDD = α + o(1).
-
(ii).
ETPDD/ETPOR = 1 + o(1).
Corollary 1 Suppose we choose ai = 1 for all i. Then under the conditions of the theorem, our data-driven procedure controls the unweighted FDR at level α + o(1).
5. Practical Issues and Numerical Results
The wFDR framework provides a useful approach to integrate domain knowledge in multiple testing. However, the weights must be chosen with caution to avoid improper manipulation of results. Under our formulation, we assume that the weights are pre-specified, while stressing that the task of assigning “correct” weights is a critical issue in practice. This section first states a few practical guidelines in weighted FDR analysis, then describes general strategies in choosing weights for a range of application scenarios. Finally we illustrate the proposed methodology via an application to GWAS.
5.1. Practical guidelines and examples on choosing weights
We first state a few important practical guidelines to ensure that the weighted FDR analysis is conducted properly. The first guideline is in particular important for a valid FDR control.
The weights must be “external” in the sense that they do not depend on the primary data such as the z-values or p-values on which the multiple tests are performed. As a practical guideline, we recommend “choosing external weights before seeing the data” as a standard rule in all weighted FDR analyses.
The choice of weights requires scientific motivations or economic considerations. For a statistical analysis involving any weighting schemes, the procedure for obtaining weights must be specified beforehand in the analysis plan, and needs be disclosed and justified carefully.
To prevent researchers from setting weights to find desired significant results, extreme weights are in general not acceptable. It is recommended to use moderate weights to incorporate domain knowledge and carry out a sensitivity analysis afterwards to avoid manipulations of results.
Next we discuss a range of scenarios that naturally give rise to weighted FDR analyses. We use these examples to illustrate how to assign weights properly in light of various economic and scientific considerations. Moreover, we emphasize how the above practical guidelines can be implemented in respective contexts to avoid manipulations of results.
Example 1. Spatial cluster analysis.
In multiple testing, relevant domain knowledge or external data can be exploited to form more meaningful testing units such as sets or clusters. For example, in a spatial setting (Pacifico et al. 2004; Benjamini and Heller 2007; Sun et al. 2015), individual locations can be aggregated into clusters to increase the signal to noise ratio and scientific interpretability. It is natural to consider weighted error rates when clusters are heterogeneous. We state a few guidelines for the choice of weights.
-
(i)
The weights should be pre-specified by experts to reflect various practical and economicconsiderations based on relevant spatial covariates such as cluster sizes, population densities, poverty rates, and urban vs. rural regions.
-
(ii)
In cluster-wise analyses, weights may be used to modify both the error rate and powerfunction. For example, symmetric weights (i.e. ai = bi for all i) may be assigned to reflect that the gains and losses in the decision process are proportional. A careful investigation of the optimal ranking of clusters (Section A in the Supplementary Material) is helpful to improve the power of analysis.
Example 2. Multiple endpoints clinical trials.
In comparing treatments with multiple endpoints (e.g. Dmitrienko et al. 2003), the error rates may be modified by assigning different weights to primary and secondary end points; this promises to increase the power in discovering more important clinical findings (Young and Gries 1984; Westfall and Young 1993). Benjamini and Cohen (2016) discussed two weighting regimes in clinical trials: the first regime requires that any primary endpoint carries a larger weight than any secondary one; and the second regime in addition requires that the combined weight of the primary endpoints is greater than the combined weight of the secondary ones. To avoid manipulations of results, Benjamini and Cohen (2016) suggested that a weighted FDR analysis should obey the following rules:
-
(i)
The tradeoffs between primary and secondary end points should be evaluated prior toconducting the trial.
-
(ii)
The weights should be given before seeing the data.
-
(iii)
The details of obtaining weights should be revealed in the trial protocol.
Example 3. Prioritized subset analysis.
In GWAS data analysis, finding diseasecausing SNPs in the immense data is very challenging. Weighted FDR analysis provides a powerful approach to integrate genomic knowledge in inference with prioritized SNP subsets. For example, Roeder et al. (2006) and Zablocki et al. (2014) assigned higher weights to pre-selected genomic regions to increase the power in detecting disease-associated SNPs that are more biologically plausible. In large-scale multiple testing with prioritized subset, we have the following requirements regarding the choice of weights:
-
(i).
It is not allowed to use the primary data (e.g. significant levels of disease association) to derive weights; such practice would distort the null distribution of p-values and lead to the failure of the wFDR procedure.
-
(ii).
The weights must be obtained based on either prior knowledge regarding the biological importance of an association, or external knowledge such as the linkage disequilibrium (LD) correlations between SNPs. This requirement can effectively avoid selection bias because the weights are conditionally independent from the significance levels of SNP associations (cf. our Condition 1). See also Section 5.(a) in Benjamini et al. (2009) for related discussions.
-
(iii).
To avoid further aggravating the multiplicity issue in large-scale testing problems, the weights should only be used to modify the power function. The unweighted FDR criterion is recommended to remain unchanged. This ensures greater power to reject more biologically relevant hypotheses without increasing the FDR (cf. our Corollary 1). See also Theorem 1 in Genovese et al. (2006) and Section 5.(a) in Benjamini et al. (2009) for related theoretical results and discussions.
5.2. Decision weights vs. procedural weights
The formulation (2.7) is different from the “procedural weights” approach, for example, in Roeder et al. (2006) and Zablocki et al. (2014), where the goal is to exploit external information by constructing weights to increase the total number of rejections (covariateassisted inference). The decision weights framework integrates the external information in a very different way; the goal is to improve the interpretability of results instead of statistical power (although the power may be gained as a byproduct, see Section E.5 in the Supplementary Material). Under the wFDR framework, the hypotheses are of “unequal” importance in light of domain knowledge. For instance, in spatial cluster analysis, a false positive cluster with larger size would account for a larger error; hence it is not sensible to use the conventional FDR definition that treats all testing units equally. The wFDR framework is different from those in Genovese et al. (2006), Ferkingstad et al. (2008) and Cai et al. (2016), where the gains and losses in various decisions are exchangeable, and conventional FDR and power definitions are suitable.
5.3. Application to Framingham Heart Study (FHS)
In this section, we implement the proposed method for analyzing a data set from Framingham Heart Study (Fox et al., 2007; Jaquish, 2007). A brief description of the study, the implementation of our methodology, and the results are discussed in turn.
The goal of the study is to decipher the genetic architecture behind the cardiovascular disorders for the Caucasians. Started in 1948 with nearly 5,000 healthy subjects, the study is currently in its third generation of the participants. The biomarkers responsible for the cardiovascular diseases, for e.g., body mass index (BMI), weight, blood pressure, and cholesterol level, were measured longitudinally. Since the mutation SNPs (within the block of DNA) are commonly passed on to descendants, it is a standard practice to collect data over several generations. The specific goal of our study is to identify disease-associated SNPs in the second generation while paying more attention to those that remain stable over the first two generations. Identifying significant SNPs (or SNP blocks) that are preserved across generations can improve both biological plausibility and statistical replicability. In our study design, we use 399 subjects in the first generation to serve as the baseline to construct weights, and use the information from the 578 subjects in the second generation (with 310 males and 268 females) to conduct the primary analysis. It is important to note that the goal of our weighted FDR analysis is different from that in Roeder et al. (2006) and Zablocki et al. (2014), where prior likelihood of association with the phenotype is utilized for constructing “procedural weights” to improve the unweighted power. In contrast, we use decision weights to modify the power function so that discoveries in certain preselected regions receive a higher priority. The main consideration is not to increase the total number of discoveries, but to focus on findings are more relevant to our study goal (identifying stable mutations across generations). In our study it is not sensible to use the conventional power definition where the discoveries from all regions are treated equally.
We consider the BMI as the response variable and develop a dynamic model to detect the SNPs associated with the BMI. Let Yi(tij) denote the response (BMI) from the i-th subject at time tij, j = 1,...,Ti. Consider the following model for longitudinal traits:
(5.1) |
where f(·) is the general effect of time that is modeled by a polynomial function of suitable order, selected by AIC or BIC model selection criterion, βk is the effect of the k-th SNP on the response and Gik denotes the genotype of the i-th subject for the k-th SNP. We also consider the random intercepts and random slopes, denoted γ0i and γ1i, respectively, for explaining the subject-specific response trajectories. A bivariate normal distribution for γi = (γ0i,γ1i) is assumed. Moreover, we assume that the residual errors are normally distributed with zero mean, and covariance matrix Σi with an order-one auto-regressive structure. We fit model (5.1) for each SNP and obtain the estimate of the genetic effect . If we reject the null hypothesis H0 : βk = 0 vs. H1 : βk ≠ 0, then we conclude that the k-th SNP has a significant association with the BMI. Since we have nearly 5 million SNPs, the false discovery rate needs to be controlled for making scientifically meaningful inference. For each k, we take standardized as our z-scores and obtain the estimated ranking statistic as described in (4.2).
To incorporate external domain knowledge, we construct weights by utilizing the pvalues obtained from first generation participants. This information is secondary in the sense that the first generation data are not directly used in multiple testing but are only employed to provide supplementary support in inference. In the FHS study, there are 399 subjects in the first generation for which we have obtained the trait values (such as BMI) and genetic information on the same set of SNPs. Using the p-values obtained from the first generation data, we partition the SNPs into three groups: less than 0.001, between 0.001 and 0.01, and greater than 0.01. The groups are denoted , j = 1,2,3. We implement the proposed wFDR method using the p-values computed from second generation participants, with group-wise weights bi = cj if i ∈ . Note that bi are the weights in respective power functions with a larger bi indicating a higher priority or a stronger belief. Since assigning a value to the biological importance of an association can be quite subjective, we propose to use different combinations of weights to investigate the sensitivity of testing results:
Setting 1: ai = 1 for all i, bi = cj if i ∈ Gj, (c1,c2,c3) = (1,1,1)
Setting 2: ai = 1 for all i, bi = cj if i ∈ Gj, (c1,c2,c3) = (4,2,1)
Setting 3: ai = 1 for all i, bi = cj if i ∈ Gj, (c1,c2,c3) = (10,5,1).
Setting 1 corresponds to an unweighted FDR analysis. Settings 2 and 3 reflect our belief that the correct rejections from groups 1 and 2 are more “powerful” or more “relevant” when compared to those from group 3. This is sensible as one of the goals in the FHS study is to discover significant SNPs that remain stable across generations. In our analysis, the weights ai = 1 are chosen for all SNPs; this ensures that the FDR will not be increased by weighting. We only modify the weights bi in the power functions to reflect that discoveries in certain pre-selected regions are of greater importance. In practice bi may be derived from expert opinion to incorporate scientific or economic considerations. We stress that in GWAS, the details of the process in choosing weights, which requires a separate discussion or justification, should be revealed.
We implement the proposed wFDR method to select most significant SNPs. The results for the sensitivity analysis with different choices of weights are summarized in Table 1. The table shows the groups sizes, the group-wise threshold levels, and the number of SNPs selected from different “groups” for each weight combination at two different FDR levels α = 0.10 and α = 0.05.
Table 1:
Number of rejections and thresholds by varying choice of weights between groups.
Choice of Weights | |||||||
---|---|---|---|---|---|---|---|
Group | Total | (1, 1, 1) | Threshold | (4, 2, 1) | Threshold | (10, 5, 1) | Threshold |
α = 0.10 | |||||||
1043 | 3 | 6.69 × 10−4 | 5 | 1.70 × 10−3 | 53 | 3.17 × 10−2 | |
5277 | 23 | 9.01 × 10−4 | 33 | 1.52 × 10−3 | 57 | 4.53 × 10−3 | |
220392 | 2019 | 2.92 × 10−3 | 1997 | 2.88 × 10−3 | 1663 | 2.31 × 10−3 | |
Overall | 226712 | 2045 | – | 2035 | – | 1773 | – |
α = 0.05 | |||||||
1043 | 1 | 5.22 × 10−6 | 2 | 3.95 × 10−4 | 4 | 7.43 × 10−4 | |
5277 | 13 | 4.22 × 10−4 | 19 | 5.20 × 10−4 | 21 | 6.61 × 10−4 | |
220392 | 159 | 1.17 × 10−4 | 147 | 1.06 × 10−4 | 131 | 1.01 × 10−4 | |
Overall | 226712 | 173 | – | 168 | – | 156 | – |
From Table 1 we can see that the unweighted analysis [i.e. the weights combination (1, 1, 1)] consistently selects higher number of SNPs than the other two weights combinations. Meanwhile, it selects fewer number of SNPs from groups 1 and 2. Here we have aimed to prioritize certain pre-selected regions and discover significant SNPs that are more biologically relevant. In our analysis, groups 1 and 2 are believed to be more “informative.” At α = 0.10, it is interesting to note that the number of rejections from groups 1 and 2 are increased by several folds but the number of rejections from group 3 is only decreased by a small proportion.
In Table 1, we also report the group-wise thresholds for the p-values. The value of the threshold is computed as the maximum of the p-values of the hypotheses rejected in the specific group and weight combination. We can see that the threshold for group 3 decreases in weights, which indicates that the rejection criteria become more stringent with higher weights. In contrast, the thresholds for groups 1 and 2 increase with weights, which allows for more discoveries in pre-selected regions. All these above observations are sensible and in agreement with the intuitions of our proposed methodology.
5.4. Simulation
In all simulation studies, we consider a two-point normal mixture model
(5.2) |
The nominal wFDR is fixed at α = 0.10. We consider the comparison of different methods under the scenario where there are two groups of hypotheses and within each group the weights are proportional.
The proposed method (Procedure 1 in Section 4.2) is denoted by 4 DD. Other methods to be compared include:
The wFDR method proposed by Benjamini and Hochberg (1997); denoted by 2 BH97. In simulations where ai = 1 for all i, BH97 reduces to the well-known step-up procedure in Benjamini and Hochberg (1995), denoted by BH95.
A stepwise wFDR procedure, which rejects hypotheses along the WPO (3.14) ranking sequentially and stops at , with defined in Section 4.2. The method is denoted by ◦ WPO. Following similar arguments in the proof of Theorem 2, we can show that the WPO method controls the wFDR at the nominal level asymptotically. This is also verified by our simulation results. Meanwhile, we expect that the WPO method will be outperformed by the proposed method (4 DD), which operates along the more efficient VCR ranking.
The adaptive z-value method in Sun and Cai (2007), denoted by + AZ. AZ is valid and optimal in the unweighted case but suboptimal in the weighted case.
To save space, this section only presents results on group-wise weights. Our setting is motivated by our application to GWAS, where the hypotheses can be divided into groups: those in preselected regions and those in other regions. It is desirable to assign varied weights to separate groups to reflect that the discoveries in preselected regions are more biologically meaningful. In Section E.2 of the Supplementary Material, we compare our methods with existing methods using general weights ai and bi that are generated from probability distributions. We also provide additional numerical results such as the comparison of various wFDR definitions, the finite sample performance of Lfdr, and the impacts of weights on the power of different wFDR procedures.
The first simulation study investigates the effect of weights. Consider two groups of hypotheses with group sizes m1 = 3000 and m2 = 1500. In both groups, the non-null proportion is p = 0.2. The null and non-null distributions are N(0,1) and N(1.9,1), respectively. We fix ai = 1 for all i. Hence BH97 reduces to the method proposed in Benjamini and Hochberg (1995), denoted by BH95. The wFDR reduces to the regular FDR, and all methods being considered are valid for FDR control. For hypotheses in group 1, we let c1 = ai/bi. For hypotheses in group 2, we let c2 = ai/bi. We choose c1 = 3 and vary c2. Hence the weights are proportional within respective groups and vary across groups. The second simulation study investigates the impacts of signal strength. The results are presented in Section E.1 of the Supplementary material.
In the simulation, we apply the four methods above to the simulated data set and obtain the wFDR and ETP levels by averaging the multiple testing results over 200 replications. In Figure 1, we plot the wFDR levels and ETP of different methods as functions of c2, which is varied over [0.1,0.8]. Panel (a) shows that all methods control the wFDR under the nominal level, and the BH97 method is conservative. Panel (b) shows that the proposed method dominates all existing methods. The proposed method is followed by the WPO method, which outperforms all unweighted methods (AZ and BH95) since bi, the weights in the power function, are incorporated in the testing procedure. The BH97 (or BH95) has the smallest ETP. As c2 approaches 1 or the weights ai and bi equalizes, the relative difference of the various methods (other than BH95) becomes less.
Figure 1:
Comparison under group-wise weights: □ BH97 (or BH95), o WPO, Δ DD (proposed), and + AZ. The efficiency gain of the proposed method increases as c1 and c2 become more distinct.
6. Discussion
The FDR provides a practical and powerful approach to large-scale multiple testing problems and has been widely used in a wide range of scientific studies. The wFDR framework extends the FDR paradigm to integrate useful domain knowledge in simultaneous inference. The weights must be chosen with caution to avoid improper manipulation of results.
In the multiple testing literature, procedural, decision, and class weights are often viewed as distinct weighting schemes and have been mostly investigated separately. Although this paper focuses on the decision weights approach, the decision-theoretic framework enables a unified investigation of other weighting schemes. For example, a comparison of the LR (3.8) and WLR (3.11) demonstrates how the LR statistic may be adjusted optimally to account for the decision gains or losses. This shows that procedural weights may be derived in the decision weights framework. Moreover, the difference between the WLR (3.11) and WPO (3.14) shows the important role that pi plays in multiple testing. In particular the WPO (3.14) provides important insights on how prior beliefs may be incorporated in a decision weights approach to derive appropriate class weights. To see this, consider the multi-class model (2.3). Following the arguments in Cai and Sun (2009), we can conclude that in order to maximize the power, different FDR levels should be assigned to different classes. Similar suggestions for varied class weights have been made in Westfall and Young (1993, pages 169 and 186). These examples demonstrate that the decision weights approach provides a powerful framework to derive both procedural weights and class weights.
Our formulation requires that the weights must be pre-specified based on external domain knowledge. It is of interest to extend the work to the setting where the weights are unknown. Due to the variability in the quality of external information, subjectivity of investigators, and complexity in modeling and analysis, a systematic study of the issue is beyond the scope of the current paper. The optimal choice of weights depends on statistical, economic and scientific concerns jointly. Notable progresses have been made, for example, in Roeder and Wasserman (2009) and Roquain and van de Wiel (2009). However, these methods are mainly focused on the weighted p-value approach under the unweighted FDR criterion, hence do not apply to the framework in Benjamini and Hochberg (1997). Moreover, the optimal decision rule in the wFDR problem in general is not a thresholding rule based on the adjusted p-values. Much work is still needed to derive decision weights that would optimally incorporate domain knowledge in large-scale studies.
7. Proofs
This section proves Theorem 1. The proofs of other results are given in the Supplementary Material.
Proof of Part (i) of Theorem 1.
To show that wFDR(δOR) = α, we only need to establish hat
where the notation EU,a,b,X denotes that the expectation is taken over U,a,b, and X. According to the definitions of the capacity function C(·) and threshold t∗, we have
It follows from the definition of p* that , where the notation EU|a,b,X indicates that the expectation is taken over U while holding (a,b,X) fixed. Therefore
(7.1) |
and the desired result follows.
Proof of Part (ii) of Theorem 1.
Let δ* be an arbitrary decision rule such that wFDR(δ∗) ≤ α. Then
(7.2) |
The notation means that the expectation is taken to average over potential randomization conditional on the observations and weights.
Let and . For i ∈ , we have and hence bi(1 − Lfdri) ≥ t∗ai(Lfdri − α). Similarly for i ∈ , we have and so bi(1 − Lfdri) ≤ t∗ai(Lfdri − α). Thus
Note that is perfectly determined by X except for (k + 1)th decision. Meanwhile,b(k=0)(1-Lfdr(k+1) – t*a(k+1) (Lfdr(k+1) –α)=0 by our choice of t∗. It follows that
(7.3) |
Recall that the power function is given by for any decision rule δ. Combining equations (7.1) – (7.3) and noting that t∗ > 0, we claim that ETP(δOR) ≥ ETP(δ∗) and the desired result follows.
Supplementary Material
References
- [1].Benjamini Y, and Cohen R (2016), “Weighted False Discovery Rate Controlling Procedures for Clinical Trials,” Biostatistics, Advance Access. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [2].Benjamini Y, and Heller R (2007), “False Discovery Rates for Spatial Signals,” Journal of the American Statistical Association, 102, 1272–1281. [Google Scholar]
- [3].Benjamini Y, Heller R, and Yekutieli D (2009), “Selective Inference in Complex Research,” Philosophical Transactions of the Royal Society A, 367, 4255–4271. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [4].Benjamini Y, and Hochberg Y (1995), “Controlling the False Discovery Rate: A Practical and Powerful Approach to Multiple Testing,” Journal of the Royal Statistical Society, Series B, 57, 289–300. [Google Scholar]
- [5].Benjamini Y, and Hochberg Y (1997), “Multiple Hypotheses Testing with Weights,” Scandinavian Journal of Statistics, 24, 407–418. [Google Scholar]
- [6].Box GEP (1950), “Problems in the Analysis of Growth and Wear Curves,” Biometrics, 6, 362–389. [PubMed] [Google Scholar]
- [7].Cai TT, and Sun W (2009), “Simultaneous Testing of Grouped Hypotheses: Finding Needles in Multiple Haystacks,” Journal of the American Statistical Association, 104, 1467–1481. [Google Scholar]
- [8].Cai TT, and Sun W (2016), “Optimal Screening and Discovery of Sparse Signals with Applications to Multistage High-Throughput Studies,” Journal of the Royal Statistical Society, Series B, Accepted. [Google Scholar]
- [9].Cai TT, Sun W, and Wang W (2016), “CARS: Covariate Assisted Ranking and Screening for Large-Scale Two-Sample Inference,” Technical Report. [Google Scholar]
- [10].Dmitrienko A, Offen WW, and Westfall PH (2003), “Gatekeeping Strategies for Clinical Trials that Do Not Require All Primary Effects to be Significant,” Statistics in Medicine, 22, 2387–2400. [DOI] [PubMed] [Google Scholar]
- [11].Efron B (2008), “Simultaneous Inference: When Should Hypothesis Testing Problems be Combined?” The Annals of Applied Statistics, 2, 197–223. [Google Scholar]
- [12].Efron B (2010), Large-Scale Inference: Empirical Bayes Methods for Estimation, Testing, and Prediction, New York: Cambridge University Press. [Google Scholar]
- [13].Efron B, Tibshirani R, Storey JD, and Tusher V (2001), “Empirical Bayes Analysis of a Microarray Experiment,” Journal of the American Statistical Association, 96, 1151–1160. [Google Scholar]
- [14].Ferkingstad E, Frigessi A, Rue H, Thorleifsson G, and Kong A (2008), “Unsupervised Empirical Bayesian Multiple Testing with External Covariates,” The Annals of Applied Statistics, 2, 714–735. [Google Scholar]
- [15].Fox CS, Heard-Costa N, Cupples LA, Dupuis J, Vasan RS, and Atwood LD (2007), “Genome-Wide Association to Body Mass Index and Waist Circumference: The Framingham Heart Study 100K Project,” BMC Medical Genetics, 8, S18. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Genovese CR, Roeder K, and Wasserman L (2006), “False Discovery Control with P-Value Weighting,” Biometrika, 93(3), 509–524. [Google Scholar]
- [17].Genovese C, and Wasserman L (2002), “Operating Characteristics and Extensions of the False Discovery Rate Procedure,” Journal of the Royal Statistical Society, Series B, 64, 499–517. [Google Scholar]
- [18].He L, Sarkar SK, and Zhao Z (2015), “Capturing the Severity of Type II Errors in High-Dimensional Multiple Testing,” Journal of Multivariate Analysis, 142, 106–116. [Google Scholar]
- [19].Hu JX, Zhao H, and Zhou HH (2010), “False Discovery Rate Control With Groups,” Journal of the American Statistical Association, 105, 1215–1227. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [20].Jaquish CE (2007), “The Framingham Heart Study, on its Way to Becoming the Gold Standard for Cardiovascular Genetic Epidemiology?” BMC Medical Genetics, 8, 63. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [21].Jin J, and Cai TT (2007), “Estimating the Null and the Proportion of Nonnull Effects in Large-Scale Multiple Comparisons,” Journal of the American Statistical Association, 102, 495–506. [Google Scholar]
- [22].Li L, Kabesch M, Bouzigon E, Demenais F, Farrall M, Moffatt MF, Lin X, and Liang L (2013), “Using eQTL Weights to Improve Power for Genome-Wide Association Studies: A Genetic Study of Childhood Asthma,” Frontiers in Genetics, 4, Article 103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [23].Lin W-Y, and Lee W-C (2012), “Improving Power of Genome-Wide Association Studies with Weighted False Discovery Rate Control and Prioritized Subset Analysis,” PLoS ONE, 7, e33716. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [24].Lin C-Y, Xing G, Ku H-C, Elston RC, and Xing C (2014), “Enhancing the Power to Detect Low-Frequency Variants in Genome-Wide Screens,” Genetics, 196, 1293–1302. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [25].Liu Y, Sarkar SK, and Zhao Z (2016), “A New Approach to Multiple Testing of Grouped Hypotheses.” Journal of Statistical Planning and Inference, 179, 1–14. [Google Scholar]
- [26].Pacifico MP, Genovese C, Verdinelli I, and Wasserman L (2004), “False Discovery Control for Random Fields,” Journal of the American Statistical Association, 99, 1002–1014. [Google Scholar]
- [27].P˜ena EA, Habiger JD, and Wu W (2011), “Power-Enhanced Multiple Decision Functions Controlling Family-Wise Error and False Discovery Rates,” The Annals of Statistics, 39, 556–583. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [28].Roeder K, Bacanu S-A, Wasserman L, and Devlin B (2006), “Using Linkage Genome Scans to Improve Power on Association in Genome Scans,” The American Journal of Human Genetics, 78, 243–252. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [29].Roeder K, and Wasserman L (2009), “Genome-Wide Significance Levels and Weighted Hypothesis Testing,” Statistical Science, 24, 398–413. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [30].Rogosa D (2003), “Student Progress in California Charter Schools,” 1999–2002. Available at http://statweb.stanford.edu/rag/draft/charter9902.pdf.
- [31].Roquain E, and van de Wiel MA (2009), “Optimal Weighting for False Discovery Rate Control,” Electronic Journal of Statistics, 3, 678–711. [Google Scholar]
- [32].Silverman BW (1986), Density Estimation for Statistics and Data Analysis, London: Chapman & Hall. [Google Scholar]
- [33].Spjøtvoll E (1972), “On the Optimality of Some Multiple Comparison Procedures,” The Annals of Mathematical Statistics, 43, 398–411. [Google Scholar]
- [34].Storey JD (2002), “A Direct Approach to False Discovery Rates,” Journal of the Royal Statistical Society, Series B, 64, 479–498. [Google Scholar]
- [35].Storey JD (2007), “The Optimal Discovery Procedure: A New Approach to Simultaneous Significance testing,” Journal of the Royal Statistical Society, Series B, 69, 347–368. [Google Scholar]
- [36].Sun W, and Cai TT (2007), “Oracle and Adaptive Compound Decision Rules for False Discovery Rate Control,” Journal of the American Statistical Association, 102, 901–912. [Google Scholar]
- [37].Sun W, Reich B, Cai TT, Guindani M, and Schwartzman A (2015), “False Discovery Control in Large-Scale Spatial Multiple Testing,” Journal of the Royal Statistical Society, Series B, 77, 59–83. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [38].Westfall PH, and Young SS (1993), Resampling-Based Multiple Testing, New York: John Wiley & Sons. [Google Scholar]
- [39].Xing C, Cohen JC, and Boerwinkle E (2010), “A Weighted False Discovery Rate Control Procedure Reveals Alleles at FOXA2 that Influence Fasting Glucose Levels,” The American Journal of Human Genetics, 86, 440–446. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [40].Young SS, and Gries CL (1984), “Exploration of the Negative Correlation Between Proliferative Hepatocellular Lesions and Lymphoma in Rats and Mice – Establishment and Implications,” Fundamental and Applied Toxicology, 4, 632–640. [DOI] [PubMed] [Google Scholar]
- [41].Zablocki RW, Schork AJ, Levine RA, Andreassen OA, Dale AM, and Thompson WK (2014), “Covariate-Modulated Local False Discovery Rate for Genome-Wide Association Studies,” Bioinformatics, 30(15), 2098–2104. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.