Abstract
The perturbation of a transcription factor should affect the expression levels of its direct targets. However, not all genes showing changes in expression are direct targets. To increase the chance of detecting direct targets, we propose a modified two-group model where the null group corresponds to genes which are not direct targets, but can have small non-zero effects. We model the behavior of genes from the null set by a Gaussian distribution with unknown variance τ2. To estimate τ2, we focus on a simple estimation approach, the iterated empirical Bayes estimation. We conduct a detailed analysis of the properties of the iterated EB estimate and provide theoretical guarantee of its good performance under mild conditions. We provide simulations comparing the new modeling approach with existing methods, and the new approach shows more stable and better performance under different situations. We also apply it to a real data set from gene knock-down experiments and obtained better results compared with the original two-group model testing for non-zero effects.
Keywords: empirical Bayes, ϵ-contamination, gene knock-down
1. Introduction
The transcriptional regulatory networks, formed by transcription factors(TFs) and their targets, are believed to play an important role regulating embryonic stem(ES) cell pluripotency(Niwa et al. (1998, 2000); Chambers and Smith (2004); Loh et al. (2006); Kim et al. (2008); Chen et al. (2008)). A multitude of inference methods exist in the literature for the identification of such networks using observational gene expression data(Friedman et al. (2000); Murphy et al. (1999); Kim et al. (2004); Lebre et al. (2010)). On the other hand, there is also intense interest in using perturbation experiments in the study of gene regulation. For example, the TF knock-down experiment is expected to very informative identifying potential targets of a TF because it depicts a less complex picture and can provide evidence for causal relationships(Geier et al. (2007); Werhli et al. (2006)).
Traditionally, potential targets of the TF are usually identified as the subset of differentially expressed genes between the control and experiment group. However, when an important TF has been knocked down, it is almost always the case that the proportion of significantly changed genes is much larger than expected(Ivanova et al. (2006); Zhou et al. (2007)). As a concrete example, consider the data set analyzed in this study, which is from the knockdown experiment for two TFs which play an important role regulating embryonic stem (ES) cell pluripotency(see section 5 for details). In this data set, the number of differentially expressed genes is very large, while the number of likely direct targets (from external CHIP-seq data assessing TF binding) is significantly smaller.
There are two popular explanations for this phenomenon:
The theoretic null distribution of the test statistics(often z-score and other analogous quantities) for zero effect is not accurate.
There are a large number of genes showing non-zero but small changes of gene expression level, as effects of the perturbation.
Proposed solutions include modifying the null distribution of z-score empirically(Efron (2007, 2008)) and applying a cut-off to fold-change as a second-layer filter; the latter has been extremely popular in practice(Nichols et al. (1998); Zhou et al. (2007); Vaes et al. (2014)). While both of these approaches can narrow down the selected, the former tackles the problem mainly based on the first explanation while the latter adopts the second implicitly, and results can be different in general(Witten and Tibshirani (2007)). For the knockdown experiment, the latter seems preferable because it considers both the change magnitude and the non-zero significant level, which is more related to what scientists care about; however, this approach lacks a natural quantitative justification.
Here, we propose a simple model to combine these two perspectives. By using a Gaussian distribution with unknown variance to describe the underlying behavior of genes in the null group, our model assumes that there can be relatively small non-zero effects even for the null genes. Assuming that the number of genes with large effect size is small, we test for the presence of such large effects relative to the background null variance. Although this model is motivated by the knockdown experiment, it can be applied to more general multiple-testing setting where both the significance and the effect size matter.
Our approach is related to the method of maximal agreement cut by Henderson and Newton (2015). They similarly pointed out that testing approaches which measure evidence against the null hypothesis tend to over-populate the candidate list with those associated with small variance, while approaches that consider only the magnitude will overlook the noise. While sharing the same spirit, our method does not aim to find the top α% subset of genes maximizing the expected overlap with the truth, with some assumed prior for all genes. Instead, we are interested in identifying the subset of genes which could not be described well by the prior describing the majority. In the setting of knock-down experiment, our model describes a scenario different from the one assumed in approaches testing for differentially expressed genes. In our model, it is assumed that, perhaps due to propagation through the gene regulatory network, when we collect the data, a lot and even all genes may have been influenced once a TF has been knocked down, and we take this possibility into consideration. We show that in this scenario, it is still possible to test for strong effect if (1) the direct target tends to have larger effect size, and (2) there are enough null hypotheses to estimate the variance under the null.
The paper is organized as follows: We describe our model in section 2.1 and the procedure to estimate the null variance in section 2.2. In section 3, we study the properties of the estimating procedure of τ2; in section 4, we extended the model to the non-centered case and the case of two sample testing with unequal variance. We provide simulations in section 5 and real data examples in section 6.
2. Statistical model and estimation procedure
2.1. Statistical model
We assume there is a control group with m0 replicates and an experiment group with m1 replicates after knocking down one TF of interest. The expression levels for N genes are measured for each replicate. Let xi, j be the measurement for gene i in replicate j from the experiment group, and zi, j be the measurement for gene i in replicate j from the control group. Without loss of generality, assume the mean level of zi, j is 0 and the mean level of xi, j is μi:
To do inference on μi, we can look at the two-sample test statistics,
Or the paired sample test statistics to remove the batch effect (m0 = m1),
As there is no fundamental difference between these two tests in our later analysis, we will omit the notation zi and use the following common notations for simplicity:
where , with n being the effective sample size and is the usual unbiased variance estimate of . In the two-sample case, , m = m1 + m2, k = 2 and in the paired sample case, n = m = m1, k = 1. Following the widely used two group model(Efron (2008)), let A0, A1 denote the sets of nulls and non-nulls respectively and denote the proportion of non-nulls. We assume that the μis in A0 and A1 are generated from different distributions:
For each gene i ∈ A1, gi is some unknown density function. The parameter t can be viewed as describing the range of normal behaviour.
In contrast to the original two-group model, which corresponds to t = 0, we allow t to take positive values. By relaxing this assumption on t, we are able to detect relatively abnormal behavior compared with the background signal. If we know t, the p-value for the new null hypothesis for gene i can be derived.
Let , under the null hypothesis, ui is a Welch statistics(Welch (1947)) in the limit case with the degree of freedom for the first “variance estimate” τ2 being ∞. Usual analysis controlling False discovery rate (FDR) or family wise error rate (FWER) carries through under our extended model in this case.
We emphasize that the parameter t itself is informative because it characterizes how influential a stimulus is – in our case, how dramatically the whole system changes after we have knocked down a TF. The value of t reflects the importance of the TF: it can be set according to either prior knowledge or estimated from the data. The second approach is usually more feasible, as we lack a quantitative characterization of this kind of importance, and it can vary under different environments even for the same TF.
2.2. Estimation procedure
For marginalizing out μi. If A0 is known, the empirical Bayes estimate of τ2 with estimated variance for is given by . Let δ be a pre-determined small value, the adjusted form , is often preferred to reduce the error for small τ2 and to ensure non-negativity. Similar form of estimate is analyzed by Johnstone (2001b,a) in the context of estimating non-centrality of χ2 - distribution with known variance.
Let Fi(.) denote the distribution of when i ∈ A0 which is the distribution of the square of a Welch’s statistic as mentioned before, and . The Iterated empirical Bayes estimation(ITEB) procedure is given below, which starts from the whole set (as A0 ∪ A1) and then iteratively remove potential outliers on the tail based on the current estimate of τ2. It stops when no point needs to be further removed.
Iterated empirical Bayes estimation(ITEB) of τ2
Input: , significance level α1, α2 and δ. By default, α1 = 0.1, α2 = 0.01 and .
Output: , the estimated τ2.
Initialization: S0 = {1,2,…,N} be the initial estimate of the null set, and .
For k = 1,2,…, do
Update the p value for each gene . The ordered p values from small to large are p(1), p(2),…, p(N). Let i* be the largest index, such that .
Let , and and remove . Update Sk = Sk−1 \ Jk and .
If Sk = Sk−1, return
We have also described two other estimation methods and compared them with ITEB in the Supplement: the truncated MLE method and the central matching(CM) method. The detailed descriptions of the truncated MLE and the CM estimator are given in Appendix C. These two methods have also been applied to estimating the empirical null distribution in the traditional two group test problem(Efron (2012); Efron et al. (2001)), and we have adapted them to our problem here. In Appendix D, we compare performances of the three estimators in different scenarios and discuss their strengths and weaknesses. ITEB is most computationally efficient and is found to have better performance overall when the non-null proportion γ is small. Thus, we will focus on ITEB and we provide detailed analysis of its properties.
3. Properties of ITEB
We study the estimation quality of ITEB as the number of hypotheses N → ∞ For simplicity, we analyze the algorithm under following mild conditions and notations with δ = 0. Let . The degree of freedom for the variance estimate is m for all i, and K := the number of iterations needed for the algorithm to stop. Since in this section we only use the mean level and its estimated variance , with slight abuse of notation, let . For the two levels α1 and α2 in the ITEB algorithm, we let be a fixed value, and let α2 → 0 at a slow rate to simplify the notations in the proof (we always let bounded away from 0).
Assumption 3.1. The degree of freedom for variance estimates m ≥ 5 is a constant and the non-null proportion γ < 1 − c for some positive constant c. The ratio of variances of different genes is bounded: there exists a positive constant C such that .
Without loss of generality, we rescale , then . We do not require τ2 to be positive or a constant. It can be 0 or decay to 0 as N → ∞.
Assumption 3.2. There exist constants L and ϵ, such that
Remark 3.3. If i ∈ A0, we have Assumption 3.2 states that, for i ∈ A1, has expectation non-negligibly bigger than τ2, and it is not purely driven by observations from its tail.
Assumption 3.4. The non-null proportion γ → 0 as N → ∞.
Theorem 3.5. (Lower bound of the variance estimate) Let RK :=|JK ∩ A1|, where JK is the rejected set at the last step k = K from ITEB. Let . Under Assumption 3.1 and 3.2, we have .
As for all l, Corollary 3.6 is a direct result of Theorem 3.5.
Corollary 3.6. Under Assumption 3.1 and 3.2, for any .
Theorem 3.7. (Upper bound of the variance estimate) Under Assumption 3.1 and 3.4, suppose α1 > 0 is fixed and α2 → 0 at a rate slow enough: γλ(α2) → 0. Then, for any δ > 0, we have .
We next show that these results can usually lead to good performance in the follow-up analysis in practice. Theorem 3.8 states that our estimate of τ2 can successfully control the FDR if we reject the hypotheses in the set JK.
Theorem 3.8. (FDR control) Under Assumption 3.1 and 3.2, if we reject all hypothesis in JK, we have .
Remark 3.9. Note that at the given level α1, α2 in the ITEB algorithm, will correspond to the set of rejections using the BFI(Benjamini-Hochberg procedure)(Benjamini and Hochberg (1995)) and JK will correspond to the set of rejections which are both rejected by the BH procedure and with p-values no greater than α2. The extra requirement that the p-value is no greater than a reasonable small value α2 is desirable in many large scale hypotheses testing settings, including the knock-down experiment.
Theorem 3.10. (Power analysis) Let and let be the oracle decision rule knowing τ2: for any level α, . Let where for i ∈ A1. Under Assumption 3.1 and 3.4, if we further assume that the density of is upper bounded by a constant, and the tail probability for decays sufficiently fast:
Then we have .
Remark 3.11. Recall that the follow-up p value pi for hypothesis i is . Thus Theorem 3.10 says that the test ϕi based on the estimated variance is asymptotically as powerful as the optimal test based on the (unknown) true variance τ2.
Proofs of Theorem 3.5, 3.7, 3.8 and 3.10 are given in the Appendix A.
4. Extensions
4.1. Non-centered null distribuion
We have been assuming that μi in the null set is generated according to N(0, τ2) . The ITEB estimation approach is easily extended to the setting where the null distribution might not be centered and μi ~ N(ϵ, τ2) with a small non-centrality ϵ for i ∈ A0. In ITEB, ϵ can be approximated by
where δ0 > 0 is a reasonable cut-off and we treat xi ∈ (−δ0, δ0) to be from A0. We can form an ITEB estimation for the non-centered case by replacing with in the ITEB algorithm.
4.2. Two sample test with unequal variance
For hypothesis i, the observations from the experiment and control groups, xi, j and zi, j, can have different variances. It is straightforward to generalize ITEB to this situation if we want to perform a two-sample test. We know that ITEB takes in {} and {}. In the unequal variance setting, we can estimate by:
whose degree of freedom is approximated by
We approximate Fi(.), the distribution of the test statistics , by F1,df(.), the F distribution with degree of freedoms (1, df), where df is approximated by (Satterthwaite (1946)).
5. Simulation: Detection of large signal
We consider the two-sample setting with equal-variance and generate data under various values of t and non-null proportion . Specifically, we fix N = 15000, m = 5 for both the control and experiment group, for any given t and γ, where γ = 1%,5% and τ = 0, 0.1,…,1, 1.5, 2, 2.5, 3 , we generate the true mean and variance as below.
1. Let μi = 0 in the control group, and in the experiment group, we generate them as following:
where U[1, max(3, 10τ)] is the uniform distribution between 1 and max(3, 10τ), and the signs of μis will be half positive and half negative.
We sample the variances from its empirical distribution from the real data set, and we scale them to have mean level 1.
We compare the following approaches:
ITEB estimate of τ2, followed the Welch’s t-test.
t-test with the null hypothesis testing for non-zero effect.
EBarray(Kendziorski et al. (2003); Yuan and Kendziorski (2006)), which is a two group empirical Bayes method providing a posterior probability for having non-zero effect. We choose the “LNNMV” method to fit the data as suggested by the authors.
Fold-change rankings with a threshold for t-test p-values being 10−5 (referred to as “fchange”). For hypotheses that do not pass this cut-off, we rank them based on p values from the t-test, after those who have passed the cut-off.
Rvalue ranking, which finds the α% of genes maximizing the overlap between this gene list and gene list with differential change above the upper α% quantile of a prior normal distribution(Henderson and Newton (2015)).
The proposed model, the rvalue ranking and the fchange all consider the magnitude and significance explicitly. If the p-value cut-off is correctly set for the fchange, we expect these three methods to share more in common in terms of their ROC curves (average sensitivity versus FDP), while only the proposed model allows you to set a cut-off based the significance level of whether a hypothesis is different from the background. LNNMV models observations from the control and the experiment directly instead of modeling their differences, also, although it smoothes the variance estimations with an inverse Chi-square prior, the prior is not related to the effect sizes across hypotheses.
Figures 1 and 2 provide ROC curves across 20 repetitions as the significance level in the testing step is varied. We see that the new approach performs the best or one of the best across different experiments. When t = 0, ITEB procedure and t-test behave similarly, and ITEB becomes almost identical as the fchange when t is large. ITEB and rvalue result in very similar results across different settings (ITEB is slightly better in the more sparse case with γ = 1%), however, besides that ITEB provides information about where to set a cut-off, the ITEB approach is much faster considering the run time. All methods except for the t-test are stable across a wide range of t.
Figure 1.

ROC curve for γ = 1 %.
Figure 2.

ROC curve for γ = 5%.
6. Real Data Examples
In this section, we apply our approach to data from two knock-down experiments described below. The quality of the results are evaluated by the enrichment of ChIP-seq peaks (for the perturbed TF) in active enhancers/promoters for the selected genes. Note that the ChIP-seq data for ES cells are external to the data used to select the genes, and it provides an orthogonal information to access how likely the selected genes are direct targets of the TFs.
We perform gene knock-down experiments on 2 TFs on the mouse ES cell line R1. For each TF, RNA interference (RNAi) delivered using nucleofection was used to knock down its expression. Puromycin selection was introduced 18 h later at 1 μ g/ml, and the medium was changed daily. 30 h, 48 h, and 72 h after puromycin selection, the cells were collected for RNA isolation. After the experiments, Microarray hybridizations were performed on the MouseRef-8 v2.0 expression beadchip arrays (lllumina, CA). More details of the experiments can be found in Appendix E. Quantile normalization is performed in the first step to reduce the batch effect, and for the same reason, for each sample in the experiment group, we consider the paired test statistics with each pair being a pair of independent experiment and control samples from the same batch and time point. We have 8 paired observations for both POU5F1 and NANOG, and we take the log difference between the gene expression levels in a knock-down sample and its corresponding control sample to further reduce the batch effect. Table 1 summarizes the data we have at different time points. Figure 3 shows results from nine random realizations of the t-SNE(Maaten and Hinton (2008)) plot using the top 1000 genes with largest variance across experiments. Each data point in the t-SNE plot represents one sample (paired) in the experiment. We use the colors black, red and green to represent data at time points 30 hr, 48 hr and 72 hr respectively. From the results, we see that differences between time points within the same knock-down experiment is comparable to the differences across batches, and they are very small compared with the differences between two knock-down experiments. To compare the targets of two TFs, we will regard different times points in the same knock-down experiment as replicates of each other.
Table 1.
Information of the knock-down data sets.
| 30 hr | 48 hr | 72 hr | |
|---|---|---|---|
| POU5F1 | 4 pairs | 4 pairs | – |
| NANOG | 2 pairs | 4 pairs | 2 pairs |
Figure 3.
Clustering with t-SNE algorithm. The nine plots here are 9 different random tsne-plot realizations. Black, red, and green colors represent experiments from 30 hr, 48 hr and 72 hr, respectively. Note that there are two replicates from the same batch in a single experiment (same day, same TF, same batch) that are almost identical to each other.
ChIP-seq data and enhancer-gene association data:
To evaluate the quality of the selected gene set, we utilize two external data sets: the ChIP-seq data is from Chen et al. (2008) and the enhancer-gene association data is from Mumbach et al. (2017). The ChIP-seq data contains results using chromatin immunoprecipitation coupled with ultra-high-throughput DNA sequencing (ChIP-seq) to map the binding locations of 13 sequence-specific TFs, including POU5F1 and NANOG. The enhancer-gene association data is generated by using the HiChIP method where the authors performed H3K27ac HiChIP in mouse ES cells. H3K27ac is a histone modification mark characteristic of active enhancers and promoters in the cell. HiChIP using H3K27ac mark as bait will provide enhancer-gene interaction information. We can evaluate the quality of the selected gene set by examining whether the binding sites of POU5F1 and NANOG are enriched near the active enhancers/promoters of the selected genes in the ES cell.
Let us call the approach based on p-value using a simple t-test S0, the approach based on EBarray S1, and the approach based on p-value using ITEB S2. We will focus on those genes with significant decrease in their expression levels after POU5F1/NANOG knock-down. For each method, we set the cut-off using the BH procedure with targeted FDR level at 0.01.
Accordingly, we set the the cut-off level α1 = 0.01 and we select 87 genes after knocking down POU5F1 and 43 genes after knocking down NANOG using S2. These numbers are 2274 and 1267 using S0, 2362 and 886 using S1. Neither S0 nor S1 provides informative candidate lists with this criterion. To have a meaningful comparison, we also consider the case where we control FWER at 0.01, which is quite a stringent criterion and under which, S0 selects 144 genes for POU5F1 and 49 genes for NANOG, S1 selects 1091 genes for POU5F1 and 254 genes for NANOG.
We say that there is supporting evidence of a gene being a direct target of a TF if this TF has at least one ChIP-seq peak within x kilobase(kb) away from the gene’s active enhancers/promoter. As we change x in a large range of value, Figure 4 gives the percentages of genes with this supporting evidence in the selected gene sets using S0, S1 with FWER control and S2 with FDR control. Figure 4 also shows the percentages of all genes with this supporting evidence (referred to as “all” in the figure), and the percentages of bottom 2000 genes with this supporting evidence (referred to as “bottom”). The bottom 2000 genes are those showing less significance after the knock-down based on the ITEB p-value (we consider this set to be the negative control). From Figure 4, we see that the ChIP-seq enrichment is quite significant for the selected genes comparing with both the negative controls and all genes. The selected gene set using S2 is significantly better than the gene set selected using S1 and S0. Table 2 shows the result with x = 20.
Figure 4.
Percent of genes with ChIP-seq nearby versus x for the selected gene sets (S0/S1 genes selected using t-test/EBarray with FWER control, S2 genes selected based on ITEB with FDR control, bottom: the bottom 2000 genes which show the smallest changes, all: all genes). The x-axis is the threshold we use to define whether a gene has a ChIP-seq peak near its enhancer/promoter and the y-axis is the percentage of selected genes with ChIP-seq nearby.
Table 2.
S0, S1, S2 results with FDR/FWER level set at 0.01
| TF | percent(negative control) | size(S0) | percent(S0) | size(S1) | percent(S1) | size(S2) | percent(S2) | |
|---|---|---|---|---|---|---|---|---|
| FDR | POU5F1 | 0.21 | 2274 | 0.48 | 2362 | 0.48 | 87 | 0.73 |
| NANOG | 0.32 | 1267 | 0.62 | 886 | 0.61 | 43 | 0.81 | |
| FWER | POU5F1 | 0.21 | 144 | 0.62 | 1091 | 0.53 | 31 | 0.74 |
| NANOG | 0.32 | 49 | 0.74 | 254 | 0.64 | 20 | 0.8 |
The third, fifth, seventh and nineth columns are the percent of genes with Chip-seq+Hi-C support in the negative control set, the gene set selected by S0 (t-test), the gene sets selected by S1 (EBarray) and S2 (ITEB t-test) respectively.
Figure 5 provides quality evaluation of the top K genes in different ranking lists respectively. The vertical grey line is where the top 50 genes is. Besides S0, S1 and S2, we also include the results using rvalues and fchange. Comparing S2 with S0 and S1, not only S2 provides a candidate gene set with much smaller size and higher quality, it provides a gene ranking list with higher quality. In terms of the gene ranking list, the rvalue provides a similar list as that from ITEB (~90% overlappings in the top 200 genes for both TFs).
Figure 5.
Percent of genes with ChIP-seq nearby versus selected gene size. The x-axis is the threshold of the ranking, and we only consider the top k genes from each ranking list.
7. Discussion
In this paper, motivated by the problem of identifying TF targets based on data from the knock-down experiment, we have proposed to test for large effect size instead of non-zero effect size in the two-group model where a Gaussian distribution with non-zero variance is used for the effect in the null group. We have considered three approaches(ITEB, truncated MLE and CM) to estimate this non-zero variance adaptively, and recommend ITBE for its computational efficiency, strong performance in simulation and attractive theoretical properties. Although we have focused on the Gaussian setting here, the idea of testing for strong signal and the approaches to estimate the null distribution can be applied to problems involving other data types. The model itself is related to the “g-modeling”(Carroll and Hall (1988); Efron (2014)), the “ϵ-contamination”(Huber (1964); Chen et al. (2016)) and the “robust Bayesian analysis”(Berger and Berliner (1986); Gaver and O’ Muircheartaigh (1987); Berger et al. (1994)). However, it should be noted that our approach has a different goal from the g-modellings. We only estimate the shape of the null distribution of the mean parameter while the g-modeling models the marginal distribution of the mean parameter considering both the nulls and the non-nulls. Our purpose for estimating the null effect distribution is to set a cut-off for the strong signals adaptively, which is not the case for the g-modelling. Although our model can be considered as a special case of the ϵ-contamination model in the parameter space and a special case of the robust empirical Bayesian analysis, the use of these models for large effects have not been studied and, to the best of our knowledge, methods with provable guarantees on power and FDR have not been demonstrated previously.
Supplementary Material
Acknowledgments
We would like to thank Zhou Fan for his helpful feedbacks. This work is supported by NIH Grants R01HG007834 and R01GM109836, NSF Grants DMS1721550 and DMS1811920.
Footnotes
SUPPLEMENTARY MATERIAL
Online Supplementary Materials.pdf: This is the Appendix contains (1) proofs to Theorems in the paper (Appendix A, B), (2) details of estimating τ2 using the truncated MLE and the CM methods (Appendix C), (3) simulations comparing ITEB, truncated MLE and the CM methods (Appendix D), and (4) details of how the knock-down experiment is performed (Appendix E) (.pdf file)
Data and code: Inside the directory “data”, we have data files containing gene expression levels from the knock-down experiments (*.csv files) and a file containing experiment conditions for each column of the previous files (*.xlsx). Inside the directory “code”, we provide R code used to generate results in section 5. (.zip file)
References
- Benjamini Y and Hochberg Y (1995). Controlling the false discovery rate: a practical and powerful approach to multiple testing. Journal of the royal statistical society. Series B Methodological), 289–300. [Google Scholar]
- Berger J and Berliner LM (1986). Robust bayes and empirical bayes analysis with ε-contaminated priors. The Annals of Statistics, 461–486. [Google Scholar]
- Berger JΟ, Moreno E, Pericchi LR, Bayarri MJ, Bernardo JM, Cano JA, De la Horra J, Martín J, Ríos-Insúa D, Betrò B, et al. (1994). An overview of robust bayesian analysis. Test 3 (1), 5–124. [Google Scholar]
- Carroll RJ and Hall P (1988). Optimal rates of convergence for deconvolving a density. Journal of the American Statistical Association 83 (404), 1184–1186. [Google Scholar]
- Chambers I and Smith A (2004). Self-renewal of teratocarcinoma and embryonic stem cells. Oncogene 23 (43), 7150. [DOI] [PubMed] [Google Scholar]
- Chen M, Gao C, Ren Z, et al. (2016). A general decision theory for huber’s ε-contamination model. Electronic Journal of Statistics 10 (2), 3752–3774. [Google Scholar]
- Chen X, Xu H, Yuan P, Fang F, Huss M, Vega VB, Wong E, Orlov YL, Zhang W, Jiang J, et al. (2008). Integration of external signaling pathways with the core transcriptional network in embryonic stem cells. Cell 133 (6), 1106–1117. [DOI] [PubMed] [Google Scholar]
- Efron B (2007). Correlation and large-scale simultaneous significance testing. Journal of the American Statistical Association 102 (477), 93–103. [Google Scholar]
- Efron B (2008). Microarrays, empirical bayes and the two-groups model. Statistical science, 1–22. [Google Scholar]
- Efron B (2012). Large-scale inference: empirical Bayes methods for estimation, testing, and prediction, Volume 1 Cambridge University Press. [Google Scholar]
- Efron B (2014). Two modeling strategies for empirical bayes estimation. Statistical science: a review journal of the Institute of Mathematical Statistics 29 (2), 285. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Efron B, Tibshirani R, Storey JD, and Tusher V (2001). Empirical bayes analysis of a microarray experiment. Journal of the American statistical association 96 (456), 1151–1160. [Google Scholar]
- Friedman N, Linial M, Nachman I, and Pe’er D (2000). Using bayesian networks to analyze expression data. Journal of computational biology 7 (3–4), 601–620. [DOI] [PubMed] [Google Scholar]
- Gaver DP and O’Muircheartaigh IG (1987). Robust empirical bayes analyses of event rates. Technometrics 29 (1), 1–15. [Google Scholar]
- Geier F, Timmer J, and Fleck C (2007). Reconstructing gene-regulatory networks from time series, knock-out data, and prior knowledge. BMC systems biology 7(1), 11. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Henderson NC and Newton MA (2015). Making the cut: improved ranking and selection for large-scale inference. Journal of the Royal Statistical Society: Series B (Statistical Methodology). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Huber PJ (1964). Robust estimation of a location parameter. The annals of mathematical statistics, 73–101. [Google Scholar]
- Ivanova N, Dobrin R, Lu R, Kotenko I, Levorse J, DeCoste C, Schafer X, Lun Y, and Lemischka IR (2006). Dissecting self-renewal in stem cells with rna interference. Nature 442 (7102), 533. [DOI] [PubMed] [Google Scholar]
- Johnstone I (2001a). Thresholding for weighted χ2. Statistica Sinica, 691–704. [Google Scholar]
- Johnstone IM (2001b). Chi-square oracle inequalities. Lecture Notes-Monograph Series, 399–418. [Google Scholar]
- Kendziorski C, Newton M, Lan H, and Gould M (2003). On parametric empirical bayes methods for comparing multiple groups using replicated gene expression profiles. Statistics in medicine 22 (24), 3899–3914. [DOI] [PubMed] [Google Scholar]
- Kim J, Chu J, Shen X, Wang J, and Orkin SH (2008). An extended transcriptional network for pluripotency of embryonic stem cells. Cell 132 (6), 1049–1061. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kim S, Imoto S, and Miyano S (2004). Dynamic bayesian network and nonparametric regression for nonlinear modeling of gene networks from time series gene expression data. Biosystems 75 (1), 57–65. [DOI] [PubMed] [Google Scholar]
- Lebre S, Becq J, Devaux F, Stumpf ΜP, and Lelandais G (2010). Statistical inference of the time-varying structure of gene-regulation networks. BMC systems biology 4 (1), 130. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Loh Y-FL, Wu Q, Joon-Lin C, Vega VB, Zhang W, Chen X, Bourque G, Joshy G, Leong B, Liu J, et al. (2006). The oct4 and nanog transcription network regulates pluripotency in mouse embryonic stem cells. Nature genetics 38 (4), 431. [DOI] [PubMed] [Google Scholar]
- Maaten L. v. d. and Hinton G (2008). Visualizing data using t-sne. Journal of machine learning research 9 (Nov), 2579–2605. [Google Scholar]
- Mumbach MR, Satpathy AT, Boyle EA, Dai C, Gowen BG, Cho SW, Nguyen ML, Rubin AJ, Granja JM, Kazane KR, et al. (2017). Enhancer connectome in primary human cells identifies target genes of disease-associated dna elements. Nature genetics 49 (11), 1602. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Murphy K, Mian S, et al. (1999). Modelling gene expression data using dynamic bayesian networks Technical report, Technical report, Computer Science Division, University of California, Berkeley, CA. [Google Scholar]
- Nichols J, Zevnik B, Anastassiadis K, Niwa H, Klewe-Nebenius D, Chambers I, Scholer H, and Smith A (1998). Formation of pluripotent stem cells in the mammalian embryo depends on the pou transcription factor oct4. Cell 95 (3), 379–391. [DOI] [PubMed] [Google Scholar]
- Niwa Η, Burdon T, Chambers I, and Smith A (1998). Self-renewal of pluripotent embryonic stem cells is mediated via activation of stat3. Genes & development 12 (13), 2048–2060. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Niwa H, Miyazaki J.-i., and Smith AG (2000). Quantitative expression of oct-3/4 defines differentiation, dedifferentiation or self-renewal of es cells. Nature genetics 24 (4), 372. [DOI] [PubMed] [Google Scholar]
- Satterthwaite FE (1946). An approximate distribution of estimates of variance components. Biometrics bulletin 2 (6), 110–114. [PubMed] [Google Scholar]
- Vaes E, Khan M, and Mombaerts P (2014). Statistical analysis of differential gene expression relative to a fold change threshold on nanostring data of mouse odorant receptor genes. BMC bioinformatics 75 (1), 39. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Welch BL (1947). The generalization ofstudent’s’ problem when several different population variances are involved. Biometrika 34 (1/2), 28–35. [DOI] [PubMed] [Google Scholar]
- Werhli AV, Grzegorczyk M, and Husmeier D (2006). Comparative evaluation of reverse engineering gene regulatory networks with relevance networks, graphical gaussian models and bayesian networks. Bioinformatics 22 (20), 2523–2531. [DOI] [PubMed] [Google Scholar]
- Witten D and Tibshirani R (2007). A comparison of fold-change and the t-statistic for microarray data analysis. Analysis 1776, nbsp58–85. [Google Scholar]
- Yuan M and Kendziorski C (2006). A unified approach for simultaneous gene clustering and differential expression identification. Biometrics 62 (4), 1089–1098. [DOI] [PubMed] [Google Scholar]
- Zhou Q, Chipperfield H, Melton DA, and Wong WH (2007). A gene regulatory network in mouse embryonic stem cells. Proceedings of the National Academy of Sciences 104 (42), 16438–16443. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.



