Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2011 Mar 1.
Published in final edited form as: J Biopharm Stat. 2010 Mar;20(2):267–280. doi: 10.1080/10543400903572746

A Semiparametric Bayesian Approach for Estimating the Gene Expression Distribution

Fei Zou 1, Hanwen Huang 1, Joseph G Ibrahim 1
PMCID: PMC3045782  NIHMSID: NIHMS274546  PMID: 20309758

Abstract

Gene expression microarrays are powerful tools for global comparison and estimation of gene expression. Many microarray studies have demonstrated biologically plausible results with only a few arrays, leading to a mis-perception that a handful of hybridized arrays can always find something meaningful. From a statistical point of view, it is important to prospectively estimate required sample sizes prior to undertaking a microarray experiment. However, all sample size calculations need to directly or indirectly estimate the unknown distribution of the effect sizes of gene expression intensities. A parametric mixture model has been developed for relating the sample size directly to the false discovery rate (FDR), the most popular multiple comparison control criteria. In this paper,we extend the parametric mixture model and propose a robust semiparametric Dirichlet process mixture model, where the parametric distribution of gene expressions is no longer specified. This analysis is performed in a Bayesian inference framework using Markov chain Monte Carlo. The usefulness of the method is illustrated by simulations and a real murine lung study.

1 Introduction

Gene expression microarrays are powerful tools for global comparison and estimation of gene expression. One of main applications of microarray studies is to detect genes that are differentially expressed under two (or more) conditions. Early approaches had used simple thresholds for ratios of expression estimates under the two conditions (Chen et al 1997), while ordinary t-tests and Wilcoxon tests (Dudoit et al 2002 and Troyanskaya et al 2002) have later been used for more rigorous statistical analyses. More sophisticated ideas are further employed in regularized t-tests, for example by adding a constant to the variance estimate in the t denominator (Tusher et al 2001 and Efron et al 2001).

Many microarray studies have demonstrated biologically plausible results with only a few arrays (Yoon et al 2002), leading to a mis-perception that a handful of hybridized arrays can always find something meaningful. It is plausible that in some cases surprisingly few arrays may be sufficient, and the current dominance of microarray research in cancer has produced an optimistic view of microarray studies that may be unwarranted in other settings, such as examining expression differences among closely related rodent strains. From a statistical point of view, it is important to prospectively estimate required sample sizes prior to undertaking an experiment. Using ANOVA analysis of gene expression, Black and Doerge (2002) propose a parametric approach assuming lognormal or gamma distributions for gene expression intensities. Lee and Whitmore (2002) also describe sample size calculations for the ANOVA model, under several different experimental designs. Pan et al (2002) discuss sample size calculations by applying a nonparametric statistical method, the normal mixture model approach. All these methods relate power to sample size while controlling for the Type I error. Hu et al (2005) relate the sample size directly to the false discovery rate (FDR), as this criterion is often used for controlling for multiple comparisons.

All sample size calculations need to directly or indirectly estimate the unknown distribution of the effect sizes of the gene expression intensities. In the statistical literature, there are many proposed methods for estimating a distribution parametrically or non-parametrically. Hu et al (2005) employ three-component parametric mixture models to estimate the effect size distribution. Alternatively, the non-parametric methods of Robbins (1955), Laird (1978; 1981), Lindsay (1983), Wu et al (2006), Ruppert et al (2007) and Guan et al (2008) can be used for estimating the effect size distribution. However, these frequentist nonparametric methods assume a (known) fixed number of mixture components. Pan et al (2002) fit a series of normal mixture models with a varying number of mixture components and then use AIC or BIC to pick up the number of mixture components corresponding to the first local minimum of AIC or BIC. In the Bayesian literature, the mixture of Dirichlet process (MDP) model (Ferguson 1973; Blackwell and MacQueen 1973; Berry and Christensen 1979; Escobar and West 1995; Liu 1996; MacEachem 1994; MacEachem and Müller 1998; Kleinman and Ibrahim 1998; Neal 2000; Rasmussen 2000 and Ishwaran and James 2002) have been widely used in nonparametric Bayesian models. The Dirichlet process (DP) provides a nonparametric prior specifications for the parameters of a mixture model and allows the number of mixture components to grow as the size of the data grows. While classical parametric modeling techniques assume that a fixed number of components, the DP approach makes no such assumptions. Tang et al (2007) apply Dirichlet mixtures of beta densities to the distribution of p-values in estimating positive FDR.

This paper extends the three-component parametric mixture models in Hu et al (2005) to the DP mixture model in estimating the effect size distribution of the gene expressions. Our method is described in the next section with details of the Markov chain Monte Carlo (MCMC) steps for fitting the model. Simulation studies and the analysis of real datasets are presented in Section 3. We conclude the paper with some remarks and discussion.

2 Methods

2.1 Notation and assumptions

Here we largely follow the notation in Hu et al (2005). Let x1j,i(j = 0, ⋯ ,n1) and x2j,i(j = 1, ⋯ ,n2) denote the expression levels for a single gene i (i = 1, ⋯ ,m) under two conditions. For simplicity, we assume n1 = n2 = n. For a single gene, we assume that the gene expression estimates are normally distributed within each condition. Let μ1,i and μ2,i denote the true mean expressions for gene i under conditions 1 and 2, and σ1,i2 and σ2,i2 denote the corresponding variances. For testing purposes, the hypotheses are H0 : μ1,i = μ2,i versus H1 : μ1,iμ2,i, and we use the statistic ti=n(x1,ix2,i)s1,i2+s2,i2, where x1,i=j=1nx1j,in,x2,i=j=1nx2j,ins1,i2=j=1n(x1j,ix1,i)2(n1) and s2,i2=j=1n(x2j,ix2,i)2(n1). Then ti follows the student t-distribution 2n – 2 degrees of freedom and non-centrality parameter nδi. The effect size

δi=μ1,iμ2,iσ1,i2+σ2,i2 (1)

governs the power to detect departures from H0, and has a biological interpretation as a scaled expression difference. We can now restate the hypotheses as H0 : δi = 0 versus H1 : δi ≠ 0.

With thousands of genes assayed in a typical microarray experiment, we can treat the δis as realizations from a common CDF, F , represented by a random variable Δ. The distributions of the tis and derived p-values depend entirely on F. Because the sign of δi has biological importance, F is typically decomposed as the following three component mixture:

F(δ)~π0F0(δ)+π1F1(δ)+π2F2(δ). (2)

Here π0, π1 and π2 are the probabilities that Δ is 0, positive or negative respectively, and π0 + π1 + π2 = 1. F0(δ) = I0(δ) is the cumulative distribution function (CDF) of the random variable who has a point mass at 0, whereas F1 and F2 are the conditional CDFs for a positive and negative Δ. Equation (2) encompasses a large number of practical situations, where F1 and F2 can be discrete or continuous.

Given δ, the conditional distribution of the p-value of a gene is

P(P<p0δ)=1G(tp02;nδ)+G(tp02;nδ), (3)

where G(t;nδ) is the CDF of the non-central student t-distribution with 2n – 2 degrees of freedom and noncentrality parameter nδ and tα is the 1 – α quantile of the central t distribution with 2n – 2 degrees of freedom. The marginal distribution of P can be written as

P(P<p0)=δP(P<p0δ)dF(δ)=π0p0+π1δ?0P(P<p0δ)dF1(δ)+π2δ<0P(P<p0δ)dF2(δ). (4)

Thus, the relationship between the FDR and p0 is

FDR(p0)=P(P<p0,δ=0)P(p<p0)=π0p0P(p<p0), (5)

which depends on n through (3). The relationship (5) depends on the underlying distribution of δ. For known π0, π1, π2, F1 and F2, through (5), we can estimate the required sample size n in order to achieve a desired FDR value for a given p-value threshold p0. Therefore, the sample size calculation reduces to the estimation of π0, π1, π2, F1 and F2. For such purposes, Hu et al (2005) propose a parametric method by considering three situations: (1) F1 (and F2) has a point mass at (positive) a1 (and (negative) a2); (2) the exponential mixture model where F1 and F2 follow positive and negative exponential distributions, and (3) the normal mixture model where F1 and F2 are two normal densities truncated at 0. However, for real datasets, the above parametric assumptions are likely to be violated. Below, we relax the parametric assumptions and estimate F nonparametrically via the MDP model.

2.2 Mixture Dirichlet Process Model

In this section, we propose the MDP model for δ, derive the full conditional distributions, and demonstrate how to implement the Gibbs sampling. Parametric Bayesian modeling typically assumes that δ follows a prior parametric distribution with known hyperparameters Ψ0. That is,

tiδi~t(δi,d),δiΨ0~h(δiΨ0), (6)

where t(δ, d) is the non-central student t-distribution with d degrees of freedom and non-centrality parameter δ and h(δ0) is a known distribution. The MDP model removes the distributional assumption on the δis and assumes that they follow an unspecified distribution G, which itself follows a Dirichlet process prior:

tiδi~t(δi,d),δi~G,GM~DP(MG0). (7)

The parameters of the Dirichlet process, DP(M · G0) are G0 and a positive scalar M. In practice, the base measure G0 is set to a parametric distribution that (hopefully) approximates the true nonparametric shape of G. The scalar M reflects our prior belief about how similar the nonparametric distribution G is to the base measure G0. As M → ∞, GG0, the MDP model leads to the fully parametric case and the base measure is the prior distribution for δ.

Note that the t-distribution can be obtained as a scale mixture of normal distributions as follows:

tiδi,τi~N(δi,τi1)τi~Gamma(d2,d2), (8)

where N(μ, σ2) denotes the normal distribution with mean μ and variance σ2 and Gamma(a, b) denotes the Gamma distribution with shape parameter a and scale parameter b. The marginal distribution of ti given δi is therefore t(δi, d). Thus the MDP model (7) can be re-expressed as the following,

tiδi,τi~N(δi,τi1),τi~Gamma(d2,d2),δi~G,G~DP(MG0). (9)

This model, as will be shown below, allows direct Gibbs sampling with closed-form full conditional distributions.

For the base measure G0, we set it to a three-component normal mixture

G0(δ)~p0I0(δ)+p1N+(μ01,τ011)+p2N(μ02,τ021), (10)

where the indicator function Ix(y) = 1 (or 0) if y = x (or yx); N+, N are positively and negatively truncated normal distributions, respectively, and p0, p1, p2, μ01, μ02, τ01, τ02 are unspecified hyperparameters with hyper priors

(p0,p1,p2)~Dirichlet(p0,p1,p2), (11)
μ01~N(μ01,τ011), (12)
μ02~N(μ02,τ021), (13)
τ01~Gamma(α102,β102), (14)
τ02~Gamma(α202,β202), (15)

respectively. Here Dirichlet(Dirichlet(p0,p1,p2)) denotes the Dirichlet distribution of order 3 with parameters p0,p1,p2?0. Last, we set the prior of M to Gamma(Ma, Mb).

The MDP model (9) allows us to get the full conditional for each of the unknown parameters via Markov chain Monte Carlo, which we describe in detail below. Let θ be the vector of all unknown parameters and let θv be the remaining vector of θ after removing element v from θ, and t = (t1, ⋯ ,tm)’.

Step 0: Assign starting values to τi, δi, i = 1, ⋯ ,m, and P = (p0, p1, p2), μ01, μ02, τ01, τ02.

Step 1: Smple τi, i = 1, ⋯ ,m according to the following conditional distribution:

τiθτi,t~Gamma(d+12,d+(tiδi)22). (16)

Step 2: Sample δi, i = 0, ⋯ ,m. As derived in Escobar (1994), the full conditional of δi is proportional to

kiqikIδk(δi)+Mqi0G0(δi)N(tiδi,τi), (17)

where δi = (δ1, ⋯ ,δi−1, δi+1, ⋯ ,δm), and N(δi,τi1) is the density function of N(δi,τi1) at ti. Let qik=N(tiδk,τi1),k=1,,i1,i+1,,m, and qi0=N(tiδ,τi)G0(δ)dδ, which equals

qi0=N(tiδ,τi1)G0(δ)dδ=p0N(ti0,τi1)+p10N+(tiδ,τi1)N(δμ01,τ011)dδ+p20N(tiδ,τi1)N(δμ02,τ021)dδ. (18)

The qik’s and Mqi0 need to be further normalized as below to find the selection probabilities. Let C−1 = ∑ki qik+Mqi0. With probability Cqik, δi is drawn from Iδk(δi) and with probability CMqi0, δi is drawn from the distribution

aI0(δi)+bN+(δiτiti+τ01μ01τi+τ01,1τi+τ01)+cN(δiτiti+τ02μ02τi+τ02,1τi+τ02),

where

a=sp0,b=sp1(τ01τi+τ01)12N(μ10,1)N(τ0112μ010,1)exp[12(τ01μ012μ12)], (19)
c=sp2(τ02τi+τ02)12N(1μ20,1)N(τ0212μ020,1)exp[12(τ02μ022μ22)], (20)

with s being the normalizing constant such that a + b + c = 1, and

μ1=τiti+τ01μ01(τi+τ01)12,μ2=τiti+τ02μ02(τi+τ02)12. (21)

We therefore sample δi(j+1),i=0,,m according to (17). After sampling the δis for all subjects, we group them into clusters such that subjects within each cluster have the same δ value. Let L(0 < Lm) be the number of unique δ values. Denote these unique values as zl, l = 1, ⋯ ,L. Note that knowing δi is equivalent to knowing the zl’s and the cluster that subject i belongs to.

One additional step in the sampling is recommended by Bush and MacEachern (1996) in order to improve the convergence for the Gibbs sampler. To speed the mixing over the entire parameter space, they suggest sampling the zls according to the conditional density of zl after assigning subjects to their clusters:

p(zlθ,t)iclusterlp(tizl,τi)G0(zl)

which implies that

p(z1θ,t)=a~I0(z1)+b~N+(zlμ~1,1ilτi+τ01)+c~N(zlμ~2,1ilτi+τ02), (22)

where

a~=s~p0,b~=s~p1(τ01ilτi+τ01)12N(μ~10,1)N(τ0112μ010,1)exp[12(τ01μ012μ~12)], (23)
c~=s~p2(τ02ilτi+τ02)12N(μ~20,1)N(τ0212μ020,1)exp[12(τ02μ022μ~22)], (24)

with s~ being the normalizing constant to ensure that a~+b~+c~=1, and

μ~1=ilτiti+τ01μ01(ilτi+τ01)12,μ~2=ilτiti+τ02μ02(ilτi+τ02)12. (25)

Step 3: Sample (p0, p1, p2): The full conditional of p = (p0, p1, p2) is

[(p0,p1,p2)θ(p0,p1,p2),t]~Dirichlet(p~0+n~0,p~1+n~1,p~2+n~2), (26)

where n0,n1,n2 are the number of zls belonging to the clusters whose δi are equal to, greater than, or less than 0 respectively, that is n0+n1+n2=L.

Step 4: Sample the hyperparameters μ01, μ02, τ01, τ02 from their full conditionals:

μ01θμ01,t~N(τ01l=1LI(δl?0)δl+τ~01μ~01n~1τ01+τ~01,1n~1τ01+τ~01), (27)
μ02θμ02,t~N(τ02l=1LI(δl<0)δl+τ~02μ~02n~2τ02+τ~02,1n~2τ02+τ~02), (28)
τ01θτ01,t~Gamma(n~1+α~102,l=1LI(δl?0)(δlμ01)2+β~102), (29)
τ02θτ02,t~Gamma(n~2+α~202,l=1LI(δl<0)(δlμ02)2+β~202), (30)

respectively.

Step 5. Sample M in two steps: we first sample a latent variable η from

η~Beta(M+1,N), (31)

then we sample M from

Mη~πηΓ(Ma+L,Mblog(η))+(1πη)Γ(Ma+L1,Mblog(η)), (32)

where

πη1πη=Ma+L1N(Mblog(η)). (33)

Steps 1-5 complete one cycle of the Gibbs sampler. From the Gibbs samples, one can easily obtain Bayesian estimates of the parameters of primary interest, such as the FDR shown in Equation (5).

3 Numerical Studies

In this section, we test the performance of our method using both simulated and real datasets. Similar to Hu et al (2005), we first simulated two sets of data from mixture exponential and mixture normal distributions for F . The range of π0 is from 0.3 to 0.9, which covers real biologic scenarios. In the exponential setup, we chose a1 = a2 = 1/λ = 0.5, 1, 2.5, 5, which represent an F with comparatively short, medium and heavy tails, respectively. In the normal setup, we chose μ1 = μ2 = 0, and σ1 = σ2 = 0.5, 1, 2.5, 5 which again represent an F with comparatively short, medium and heavy tails. We set the hyperparameters p0=p1=p2=1, μ01=μ02=0,τ01=τ02=1, α10=α20=β10=β20=1, and Ma = Mb = 1.

The length of the chain consisted of 23000 Gibbs samples. The first 3000 samples (the burn-in) were discarded. The remaining samples were saved for every 20 samples to reduce serial correlations. The final posterior samples contained 1000 samples and were used for the subsequent Bayesian analysis. Table 1 shows the posterior mean and the 90% credible interval for one of the primary parameters, π0. The upper part is for the exponential mixture settings and the lower part is for the normal mixture settings. Figure 1 shows one example of the posterior distribution for π0 = 0.7. For all the π0 values simulated, the farther the two non-zero components of F are from zero, the better our method performs. This is intuitive since for those distributions with two non-zero components of F close to 0, the resulting distributions of the test statistics tis are centered around 0 and thus it becomes harder to distinguish the two non-zero components from 0. This statement can be further illustrated by Figure 2, in which the Q-Q plots of the estimated F distributions against the simulated ones are provided. Comparing the exponential mixture results (upper panel) to the normal mixture ones (lower panel), we find that our approach fits the exponential mixture model better. This is consistent with the above conclusion since the two non-zero exponential distributions have longer tails than its two counterpart normal distributions when a = σ.

Table 1.

π0 and their estimates obtained from the simulated data

Exponential Mixture
π 0 a=5 a=2.5 a=1 a=0.5
0.3 0.31(0.1,0.43) 0.26(0.002,0.48) 0.49(0.06,0.72) 0.42(0,0.88)
0.5 0.52(0.29,0.64) 0.55(0.25,0.72) 0.49(0,0.81) 0.60(0.06,0.92)
0.7 0.69(0.54,0.75) 0.68(0.46,0.78) 0.74(0.11,0.87) 0.75(0.05,0.92)
0.9 0.92(0.82,0.96) 0.91(0.74,0.97) 0.89(0.12,0.99) 0.97(0.93,1)
Normal Mixture
π 0 σ = 5 σ = 2.5 σ = 1 σ = 0.5
0.3 0.28(0.001,0.45) 0.43(0.004,0.61) 0.50(0,0.81) 0.86(0.18,1)
0.5 0.42(0.02,0.57) 0.45(0.009,0.67) 0.32(0,0.72) 0.92(0.46,1)
0.7 0.66(0.38,0.75) 0.68(0.45,0.80) 0.64(0.01,0.87) 0.59(0,0.94)
0.9 0.89(0.66,0.97) 0.90(0.66,0.98) 0.97(0.81,1) 0.97(0.88,1)

The fifth and ninety-fifth percentiles of the posterior distribution are given in parentheses.

Figure 1.

Figure 1

Histograms for the posterior distributions of π0 for the four simulation studies for the mixture exponential model (upper) and mixture normal distribution (lower). π0 = 0.7.

Figure 2.

Figure 2

Q-Q plots for the underlying F distribution in a simulation study (n=50, m=10000). The lower panel is for the normal mixture model, the upper panel is for the exponential mixture model, and π0 = 0.7.

We also reconstruct the distributions of the test statistics based on the estimated underlying distributions of δ and compare them with the true distributions of the test statistics (Figure 3). It is clear that the reconstructed distributions of the test statistics match their corresponding true distributions well, regardless of how heavy the tail for δ is. Further, our method always performs better than the parametric estimations in Hu et al (2005) unless the parametric model assumptions hold exactly.

Figure 3.

Figure 3

Q-Q plots for the observed t-statistics in a simulation study (n=50, m=10000). The lower panel is for the normal mixture model, the upper panel is for the exponential mixture model, and π0 = 0.7.

Next, we demonstrate the effectiveness of our method on a real dataste. we considre a murine lung microarray dataset submitted to Gene Expression Omnibus (E. Hoffman, Children’s National Medical Center, GDS251). The study used the Affymetrix U74Av2 array (12488 probesets) to compare samples from the two mouse strains: C57BL6/J (n=12) and Balb/c (n=12) in their sensitivity with pulmonary fibrosis. This data has been analyzed by Hu et al (2005) and we compare our results to theirs.

The 95% incredible intervals of the πis are (0.27, 0.47), (0.33, 0.55) and (0.09, 0.29) for π0, π1 and π2, respectively. These results are somewhat different from the estimations in Hu et al (2005). The difference between π1 and π2 indicates that the distribution of δ is not symmetric in the positive and the negative components. The fitted distribution for δ is shown in the right panel of Figure 4. The Q-Q plot of 12488 observed t-statistics versus the estimated ones is given in the left panel of Figure 4. Compared to the Q-Q plot (Figure 3) in Hu et al (2005), our nonparametric model clearly provides a better fit.

Figure 4.

Figure 4

Real murine lung study: the left panel is the Q-Q plot of the observed t-statistics vs the estimated ones; the right panel displays the histogram of the posterior distribution of F .

4 Discussion

We have described a Bayesian semiparametric approach to estimate the underlying gene expression distribution F . Our method involves specifying a nonparametric prior for the distribution of the scaled gene expression difference via a mixture of Dirichlet processes. The resulting model is estimated via the Gibbs sampler. Compared to the parametric mixture model, our nonparametric Bayesian method is more flexible and robust in modeling the gene expression distribution. Once the distribution of the effect size is estimated, we can easily apply this method to the FDR-based sample size calculations in Hu et al (2005). The biases in estimating F with short tails is not unique to our method. All parametric and non-parametric methods share similar (identification) problems. For such situations, increasing the sample size may be necessary in improving estimation accuracy. In this paper, only a two-sample design is considered. One future extension is to extend the method to other experimental designs. More general multi-sample designs or general linear regression modeling require a more careful consideration regarding δ, but the essence of the approach remains the same.

Acknowledgments

This research is supported in part by NIH grant R01 GM074175.

References

  • [1].Berry DA, Christensen R. Empirical bayes estimation of a Binomial parameter via mixtures of Dirichlet processes. The Annals of Statistics. 1979;7:558–568. [Google Scholar]
  • [2].Black MA, Doerge RW. Calculation of the minimum number of replicate spots required for detection of significant gene expression fold change in microarray experiments. Bioinformatics. 2002;18:1609–1616. doi: 10.1093/bioinformatics/18.12.1609. [DOI] [PubMed] [Google Scholar]
  • [3].Blackwell D, MacQueen JB. Ferguson distributions via Polya urn schemes. Annals of Statistics. 1973;1:353–355. [Google Scholar]
  • [4].Bush CA, MacEachern SN. A semi-parametric Bayesian model for randomized block designs. Biometrika. 1996;83:275–285. [Google Scholar]
  • [5].Chen Y, Dougherty ER, Bittner ML. Ratio-based decisions and the quantitative analysis of cDNA microarray images. J. Biomed. Optics. 1997;2:364–367. doi: 10.1117/12.281504. [DOI] [PubMed] [Google Scholar]
  • [6].Dudoit S, Yang YH, Callow MJ, Speed TP. Statistical methods for identifying differentially expressed genes in replicated cdna microarray experiments. Statistica Sinica. 2002;12:111–139. [Google Scholar]
  • [7].Efron B, Tibshirani R, Storey JD, Tusher V. Empirical bayes analysis of a microarray experiment. Journal of the American Statistical Association. 2001;96:1151–1160. [Google Scholar]
  • [8].Escobar MD. Estimating normal means with a Dirichlet process prior. Journal of American Statistical Association. 1994;89:268–277. [Google Scholar]
  • [9].Escobar MD, West M. Bayesian density estimation and inference using mixtures. Journal of American Statistical Association. 1995;90:577–588. [Google Scholar]
  • [10].Ferguson TS. A Bayesian analysis of some nonparametric problems. The Annals of Statistics. 1973;1:209–230. [Google Scholar]
  • [11].Guan Z, Wu B, Zhao H. Application of Bernstein polynomial in the estimation of false-discovery-rate. Statistica Sinica. 2008;18:905–923. [Google Scholar]
  • [12].Hu J, Wright FA, Zou F. Practical FDR-based sample size calculations in microarray experiments. Bioinformatics. 2005;21:3264–3272. doi: 10.1093/bioinformatics/bti519. [DOI] [PubMed] [Google Scholar]
  • [13].Ishwaran H, James LF. Approximate Dirichlet process computing in finite normal mixtures: smoothing and prior information. Journal of Computational and Graphical Statistics. 2002;11:508–532. [Google Scholar]
  • [14].Kleinman KP, Ibrahim JG. A Semi-parametric bayesian approach to the random effects model. Biometrics. 1998;54:921–938. [PubMed] [Google Scholar]
  • [15].Lee MT, Whitmore GA. Power and sample size for dna microarray studies. Statistics in Medicine. 2002;21:3543–3570. doi: 10.1002/sim.1335. [DOI] [PubMed] [Google Scholar]
  • [16].Laird NM. Nonparametric maximum likelihood estimation of a mixing distribution. J. Amer. Statist. Assoc. 1978;73:805–811. [Google Scholar]
  • [17].Laird NM. Empirical Bayes estimates using nonparametric maximum likelihood estimate for the prior. J. Statist. Comput. Simulation. 1981;15:211–220. [Google Scholar]
  • [18].Lindsay BG. The geometry of mixture likelihoods: a general theory. Ann Stat. 1983;11:86–94. [Google Scholar]
  • [19].Liu JS. Nonparametric hierarchical Bayes via sequential imputations. The Annals of Statistics. 1996;24:911–930. [Google Scholar]
  • [20].MacEachern SN. Estimating normal means with conjugate style Dirichlet process prior. Communication in Statistics Simulation and Computation. 1994;23:727–741. [Google Scholar]
  • [21].MacEachern SN, Müller P. Estimating mixture of Dirichlet process models. Journal of Computational and Graphical Statistics. 1998;7:223–238. [Google Scholar]
  • [22].Neal RM. Markov chain sampling methods for Dirichlet process mixture models. Journal of Computational and Graphical Statistics. 2000;9:249–265. [Google Scholar]
  • [23].Pan W, Lin J, Le C. How many replicates of arrays are required to detect gene expression changes in microarray experiments? A mixture model approach. Genome Biolology. 2002;3 doi: 10.1186/gb-2002-3-5-research0022. research0022.1-0022.10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [24].Rasmussen CE. The Infinite Gaussian Mixture Model. In: Solla SA, Leen TK, Muller K-R, editors. Adv. Neur. Inf. Proc. Sys. Vol. 12. MIT Press; 2000. pp. 554–560. [Google Scholar]
  • [25].Robbins H. An Empirical Bayes approach to atatistics; Proceedings of the Third Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics; 1955.pp. 157–163. [Google Scholar]
  • [26].Ruppert D, Nettleton D, Hwang J. T. Gene. Exploring the information in P-values for the analysis and planning of multiple-test experiments. Biometrics. 2007;63:483–495. doi: 10.1111/j.1541-0420.2006.00704.x. [DOI] [PubMed] [Google Scholar]
  • [27].Tang Y, Ghosal S, Roy A. Nonparametric Bayesian estimation of positive false discovery rates. Biometrics. 2007;63:1126–1134. doi: 10.1111/j.1541-0420.2007.00819.x. [DOI] [PubMed] [Google Scholar]
  • [28].Troyanskaya OG, Garber ME, Brown P, Botstein D, Altman RB. Nonparametric methods for identifying differentially expressed genes in microarray data. Bioinformatics. 2002;18:1454–1461. doi: 10.1093/bioinformatics/18.11.1454. [DOI] [PubMed] [Google Scholar]
  • [29].Tusher VG, Tibshirani R, Chu G. Significance analysis of microarrays applied to the ionizing radiation response. Proceedings of the National Academy of Sciences USA. 2001;98:5116–5121. doi: 10.1073/pnas.091062498. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [30].Wu B, Guan Z, Zhao H. Parametric and nonparametric FDR estimation revisited. Biometrics. 2006;62:735–744. doi: 10.1111/j.1541-0420.2006.00531.x. [DOI] [PubMed] [Google Scholar]
  • [31].Yoon H, Liyanarachchi S, Wright FA, Davuluri R, Lockman JC, de la Chapelle A, Pellegata NS. Gene expression profiling of isogenic cells with different tp53 gene dosage reveals numerous genes that are affected by tp53 dosage and identifies cspg2 as a direct target of p53. Proceedings of the National Academy of Sciences USA. 2002;99:15632–15637. doi: 10.1073/pnas.242597299. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES