Abstract
Finite mixtures of Gaussian distributions are known to provide an accurate approximation to any unknown density. Motivated by DNA repair studies in which data are collected for samples of cells from different individuals, we propose a class of hierarchically weighted finite mixture models. The modeling framework incorporates a collection of k Gaussian basis distributions, with the individual-specific response densities expressed as mixtures of these bases. To allow heterogeneity among individuals and predictor effects, we model the mixture weights, while treating the basis distributions as unknown but common to all distributions. This results in a flexible hierarchical model for samples of distributions. We consider analysis of variance–type structures and a parsimonious latent factor representation, which leads to simplified inferences on non-Gaussian covariance structures. Methods for posterior computation are developed, and the model is used to select genetic predictors of baseline DNA damage, susceptibility to induced damage, and rate of repair.
Keywords: Comet assay, Finite mixture model, Genotoxicity, Hierarchical functional data, Latent factor, Samples of distributions, Stochastic search
1. INTRODUCTION
Molecular epidemiology studies increasingly make use of samples of cells from immortalized cell lines corresponding to different individuals. Typically, the number of cells collected can be large even when the number of individuals (cell lines) is small to moderate. Although analysts often simplify the data by focusing on distributional summaries, the natural response for a study subject is a random distribution. The focus of inference is then on assessing how the random distribution changes across cell lines and with predictors. As in functional data analysis (Ramsay and Silverman, 1997), in which the response for an individual is a random function, it is appealing to limit parametric assumptions.
Our motivation is drawn from a study using single-cell electrophoresis (comet assay) to study genetic factors predictive of DNA damage and repair. The cell lines employed in this study come from the STET, a resequencing project including 90 unrelated male and female individuals from different ethnic backgrounds. For these individuals, unique combinations of single nucleotide polymorphisms (SNPs) were used to define haplotype categories for 20 candidate DNA repair genes. Single-cell electrophoresis, also known as comet assay (Östling and Johanson, 1984, Singh and others, 1988), was then used to measure the frequency of DNA strand breaks in cell lines for the 90 STET individuals. Replicate samples (100 cells) from each cell line were allocated to 1 of 3 groups: (1) analyzed without treatment, (2) analyzed immediately after exposure to a known genotoxic agent (hydrogen peroxide, H2O2), or (3) analyzed after allowing 10 min for DNA repair following exposure to H2O2.
The comet assay is commonly used in genotoxicity experiments to quantify DNA damage arising from single-strand breaks and alkyl labile sites on the individual cell level. Figure 1 shows 2 cells analyzed with the comet assay. Cells with few strand breaks have a spherical shape, while those with a high frequency of strand breaks present a tail of DNA streaming out from the nucleoid, forming a comet-like appearance. Summary statistics of the amount of DNA in the tail are used as surrogates for the amount of damage. Motivated by the results of Dunson and others (2003), we focus in this article on one particular surrogate.
The Olive tail moment (Olive and others, 1990) is defined as the percentage of DNA in the tail of the comet multiplied by the length between the center of the comet's head and tail as is automatically calculated from the cell image by standard software like Komet™ or Vis Comet™.
To demonstrate the challenges involved in the analysis, we present in Figure 2 the distribution of the Olive tail moment for one of the cell lines in the sample at 3 time points of the Olive tail moment each based on 100 cells. Note that these distributions are non-Gaussian, violating the assumptions of typical analysis of variance (ANOVA) models, and that the shape of the distribution changes dramatically. Another issue is that our interest focuses not on estimating these distributions but in studying heterogeneity among individuals in susceptibility to damage and repair rates. Also, we do not expect samples at different times points to be uncorrelated. Indeed, cells from individuals with large initial damage might show increased susceptibility to damage, or individuals suffering a lot of damage might recover at a faster rate. Finally, we note that it is not possible to examine the same cells before and after damage because cells are destroyed in applying the comet assay. Therefore, longitudinal methods are not helpful in this setup. In summary, flexible models for random distributions are needed, which should allow for the inclusion of highly structured hierarchical models in order to account for the special characteristics of the data.
In modeling hierarchical functional data, one can potentially allow heterogeneity among individuals in curve data by using a spline and/or wavelet model with individual-specific basis coefficients (Bigelow and Dunson, 2005), (Thompson and Rosen, 2008), (Morris and Carroll, 2006). When the data consist of random distributions instead of random curves, this approach can be modified by specifying the random distribution as a mixture of a finite number of basis distributions. In particular, mixtures of Gaussians provide an appealing choice; finite mixtures of a moderate number of Gaussians can produce an accurate approximation of any smooth density.
Our proposed approach is based on generalizing finite mixtures of Gaussians to place a hierarchical regression model on the mixture weights, while keeping the basis distributions constant across groups. A conceptually related approach for functional data analysis has recently been presented by Behseta and others (2005). In this paper, we focus on multilevel probit regression models, incorporating latent factors to allow flexible modeling of dependence. The idea of placing a regression model on the mixture probabilities is borrowed from the latent class modeling literature. Although latent class regression methods have previously been applied to functional data (Muthen and Shedden, 1999), to our knowledge such approaches have not been used to obtain flexible models of samples of distributions.
There is a growing Bayesian literature on models for dependent random distributions, most based on generalizations of the Dirichlet process (Ferguson, 1973, Ferguson, 1974, Sethuraman, 1994). Some relevant examples include the dependent Dirichlet process (MacEachern, 1999, MacEachern, 2000), which has spawned the works of DeIorio and others (2004) and Gelfand and others (2005), the order-dependent Dirichlet process (Griffin and Steel, 2006), the hierarchical Dirichlet process (Teh and others, 2006), and the linear combination of draws from independent Dirichlet processes (Müller and others, 2004, Dunson and others, 2007). However, although most of these methods are rather general in nature and provide full support on the space of absolutely continuous distributions, practical application and interpretation in large and complex hierarchical settings like the DNA repair study can be difficult. Indeed, these approaches do not allow for a hierarchical model specification like the one we describe in this paper.
As mentioned by several authors, including Mengersen and Robert (1996), Richardson and Green (1997), and Green and Richardson (2001), finite mixture models provide an accurate approximation to fully nonparametric Bayes approaches in many cases, with some distinct advantages. In developing our finite mixture model, we were motivated in particular by 2 issues: (1) interpretability and (2) ease of computation, given that the data in the National Institutes of Environmental Health Sciences (NIEHS) DNA repair study consist of information for over 20 000 cells. Fully nonparametric Bayes methods that allow the number of mixture components to be unknown tend to be both computationally intensive and subject to difficulties in interpreting latent class category-specific results, as the class definitions change across Monte Carlo Markov chain (MCMC) iterations.
The paper is organized as follows. Section 2 develops the general formulation of the hierarchically weighted finite mixture (HWFM) models and discusses prior elicitation, while Section 3 describes specific choices of the hierarchical structure in the context of the DNA repair study. Section 4 describes efficient computational implementations for this class of models. In Section 5, we apply the HWFM models to answer relevant questions on the DNA repair study. Finally, Section 6 contains a brief discussion on this class of models and the application at hand, as well as possible extensions.
2. HIERARCHICALLY WEIGHTED GAUSSIAN MIXTURES
2.1. Finite mixtures
Consider initially the case in which data consist of i.i.d. draws from a single unknown distribution, so that yjf, for j = 1,…,m. For example, the data y = (y1,…,ym)′ may consist of measures of DNA damage for m untreated cells drawn from a single cell line. In this i.i.d. case, we focus on the following model:
(2.1) |
where k ∈ {1,…,K} indexes the component number, ωk is a probability weight on the kth component (0 < ωk < 1, =1), and {θk, σk2} are the normal mean and variance for the kth component density.
In order to generalize (2.1) to allow covariates and hierarchical dependency structures, one can potentially choose a hierarchical regression structure for the weights {ωk} and/or the component-specific parameters {θk,σk}. However, by fixing the set of component-specific parameters, the components correspond to common, data-defined, latent classes that can be easily interpreted in the context of each specific application. For example, in exposing samples of cells to genotoxic agents, it is not possible to treat all cells with the same dose due to the inability to spread the compound equally over the entire cell culture tray, and a subset of cells may remain effectively untreated. Hence, at a given dose, the response distribution can be realistically modeled as a mixture of the distributions at lower doses and an innovation distribution. Similar behavior occurs as repair times vary, with cells that repair more rapidly mimicking cells that have suffered little damage but have slower rates of repair. In this setting, we find it appealing to visualize a set of (unknown but fixed) basis distributions, (·| θk, σk2), for k=1,…,K, representing different levels of cellular damage, with weights that vary depending on experimental conditions (before or after exposure, repair time) and measured/unmeasured genetic factors.
Prior to exposure to the genotoxic agent, the distribution of DNA damage among cells in the different cell lines could be characterized by assigning relatively high weight to component densities that allocate high probability to values close to 0. In contrast, after exposure, the weights on stochastically larger components would be expected to increase. Also, in characterizing the distributions of DNA damage following repair, the weights on stochastically larger basis densities may be relatively small for cell lines established for individuals having high rates of repair. Thus, by allowing the specific weights to vary across the different cell lines, heterogeneity can be accommodated.
2.2. Hierarchical structure
Consider now observations that are indexed by categorical variables, such as cell line (i=1,…,n) and treatment group (t=1,2,3). In order to allow the mixture weights to vary with treatment groups and (initially unmeasured) cell line–specific factors, one can define a model for the latent class indicator itj ∈{1,…,K} for the jth cell in the tth treatment group for the ith cell line. Note that itj=k denotes that the cell belongs to the kth mixture component.
A computationally convenient and flexible structure is provided by the continuation ratio probit model:
where Φ(·) denotes the standard normal distribution function. Note that this model allows a different weight for each component in the mixture for every cell line × treatment group combination, effectively allowing the distribution of the cellular damage to change freely. The use of the probit transformation to define the weights allows us to restate the model using normally distributed latent variables (see Section 4), enabling us to incorporate most of the standard Bayesian machinery into the model for the distribution. Additionally, the probit transformation induces a natural scale in the transformed weights that simplifies prior elicitation, as discussed below. The use of a continuation ratio probit model is reminiscent of the stick-breaking construction for the weights of the Dirichlet process (Sethuraman, 1994). Indeed, if αituN(αitu|0,1), then Φ(αitu)∼ Uni[0,1]= Be(1,1) and we recover a special case of the finite stick-breaking construction in Ishwaran and James (2001), creating independent priors for each cell line at each experimental condition. By allowing for different dependence across the αs, we allow the shape of the distributions to be dependent across individuals and experimental conditions while allowing for flexible and rich weight structures. Similar approaches, using a multinomial logit model, have been used in machine learning to construct spatially weighted mixtures for image segmentation (Figueiredo and others, 2007) and in latent class analysis (Muthen and Shedden, 1999).
For the parameters of the normal component, we propose standard, conditionally conjugate priors
where N(·|θ,σ2)1A denotes the normal distribution with mean θ and variance σ2 restricted to the set A. Order constraints on the parameters of the mixture components have been regularly used in the literature to ensure identifiability (see, e.g. Mengersen and Robert, 1996, and Richardson and Green, 1997) by attempting to prevent label switching due to the inherent symmetry of the parameter space in mixture model. However, label switching is not a concern for us since we are not interested in component-specific inference (Stephens, 2000). Instead, introducing an order constraint on the means allows us to interpret the components of the mixtures as corresponding to increasingly higher levels of cellular damage, as described above.
The specific form for the transformed weights αitk has so far been left open, which endows the model with a great deal of flexibility and generality. In general, we assume that the vector of transformed weights α follows a parametric distribution, possibly dependent on a set of hyperparameters η, that is, α∼p(α|η). Choosing p to be a Gaussian distribution will typically yield conditionally conjugate distributions, simplifying computation through Gibbs sampling. In Section 3, we discuss 3 interesting choices for such models, in the context of the DNA repair studies.
2.3. Bayesian model selection and mixture priors
In this section, we present a brief review of Bayesian model selection; for a more detailed discussion see Kass and Raftery (1995). Given 2 models M1:p(y|φ1) and M2:p(y|φ1) with associated prior distributions p1(φ1) and p2(φ2), the Bayes factor of model 1 versus model 2 is defined as
Let πi denote the prior probability of model i. Given prior odds for model 2 versus model 1, π2/π1, these can be updated using the Bayes factor to yield posterior odds, O21=B21π2/π1. The posterior odds measures the relative strength of the evidence in favor of each model provided by the data. Indeed, posterior odds can be transformed to posterior probabilities for each of the models by noting that Pr(M1|y)=(1+O21π2/π1)−1. Posterior probabilities over 0.76 provide substantial evidence in favor of M1, while probabilities over 0.99 provide decisive evidence (Jeffreys, 1961, Kass and Raftery, 1995).
Computation of Bayes factors can be a complex task, as they require the calculation of multidimensional integrals. Mixture priors (in contrast to the mixture likelihoods we discussed in Section 2.1) can be used to simplify this operation by transforming the model comparison problem into an inference problem. In the context of nested models, it is common to use zero-inflated priors
where p*(φ) is the prior for the parameter of interest under the alternative (more complex) model and 1x(·) denotes the degenerate distribution with all its mass at x. The updated value π2|y corresponds to the posterior probability of the null model (see George and McCulloch, 1997, and Clyde and George, 2004, for excellent reviews).
3. HIERARCHICAL MODELING IN DNA REPAIR STUDIES
3.1. Modeling and testing heterogeneity
Our first goal is to formally assess heterogeneity across individuals in the frequency of DNA strand breaks. Although exploratory analysis of the data reveals some important differences in the shape of the distributions across subjects, developing a formal test is a necessary preliminary step before building a more complicated model structure. Besides, the problem of identifying heterogeneity in populations of distributions is a complex problem that appears in many different applications.
Given this goal, an ANOVA-like structure for the transformed weights and suitable zero-inflated priors seem a natural choice. Therefore, we consider the model
(3.1) |
where {γk} controls the baseline probability of latent class k, {δi} controls the baseline susceptibility of individual i, {βt} controls the population-wide baseline damage at time t, and the remaining terms represent dependence of the weights on possible interactions. To ensure identifiability of the parameters, we let δ1 = β1 = (δβ)11 = (δγ)11 = (βγ)11 = (δβγ)111 = 0, so γu, u = 1,…,k, determines the probability of the kth component for the baseline damage in the first individual in the sample.
In this ANOVA-like structure, hypotheses related to differences between groups, cell line–specific effects, and interactions can be assessed by considering nested models, with appropriate terms in the linear predictor excluded. In particular, we are interested in testing differences attributable to cell line effects. We therefore write γ = (γ1,…,γK)′ and let λ =(λ(−r)′, λ(r)′)′ denote the remaining parameters, with the subvector λ(r)′ =(δ′,δβ′, δγ′, δβγ)′ denoting coefficients associated with cell line (individual)–specific effects and λ(−r)′=(β′, βγ′) measuring differences attributable to repair time. Then, we let
This prior for γ and the identifiability constraints are intended to approximately enforce the same prior probability for each component. See Section 4.2 for details.
The prior on λ(−r) has a ridge regression–type shrinkage structure intended to stabilize estimation, while the mixture prior for the cell line–specific coefficients λ(r) allows us to assess evidence of heterogeneity among individuals. Zero values for these coefficients correspond to a null model with no heterogeneity. We choose a prior that assigns 0.5 probability to the model with no heterogeneity and choose independent standard normal priors for the coefficients included in the model. As we discussed in Section 2.3, related mixture priors have been widely used in model selection problems. Note that only 2 models are being considered in this particular setup, instead of all possible nested models. This greatly reduces computational burden and improves the mixing of our algorithm, without detracting from our goal. The unit prior variance was chosen to assign high probability to a range of plausible values for the regression coefficients included in the model, taking the probit scale into consideration.
3.2. Modeling heterogeneity in DNA repair
The main interest of the DNA repair study focuses on understanding heterogeneity in the rates of DNA repair, adjusting for baseline damage and susceptibility to induced damage. To address this goal, we need to place additional structure on the model: damage levels at different time points cannot be directly interpreted as the quantities of interest, as they might be correlated.
For this purpose, we replace the cell line–specific terms λ(r) in the previous linear predictor in (3.1) by a factor analytic model containing cell line–specific latent traits
with {γk} intercepts as before, {βt} group differences, {(βγ)kt} interactions,
where the upper diagonal terms of Λ have been set to 0 for identifiability. The factor analytic term, λt′ ηi, accounts for heterogeneity among the cell lines, with ηi1, ηi2, and ηi3 normal latent traits measuring the ith cell line's level of initial DNA damage, susceptibility to induced damage, and rate of repair relative to the other cell lines. Due to the conditional structure, ηi2 can be interpreted as a susceptibility adjusting for baseline damage, while ηi3 (measuring the change in the distribution along 10 min elapsed since the cells were damaged) can be interpreted as a repair rate adjusting for baseline and induced damage. Within-cell line correlation between groups is captured by the parameters λ21, λ31, and λ32. To fix the scale of the latent traits for identifiability purposes, we let var(ηit)=1 for t = 1,2,3. In addition, we initially let E(ηit)=0 for t = 1,2,3, though we will later consider methods for incorporating diplotypes.
Under this structure, the level of heterogeneity among the cell lines is controlled by the magnitude of the factor loadings parameters λtt, t=1,2,3. The null hypothesis of homogeneity in baseline damage corresponds to λ11 =0. The null hypothesis of homogeneity in susceptibility corresponds to λ22 =0 and homogeneity in repair rates corresponds to λ33 =0. Models excluding the main effects by letting λtt =0 should automatically exclude any correlation effects by also fixing λtt′ = 0, ∀ t′ < t, thereby reducing the number of models for the covariance structure to 30.
Again, by using mixture priors with point masses at 0 for the factor loadings, we can effectively move within this model space
in which the sign constraint on λtt is used to ensure identifiability. Because the total number of models involved is small and predictive densities can be calculated in closed form, we opt for sampling over the full model space instead of the one-component-at-a-time scheme typical of stochastic search variable selection (SSVS). This improves the efficiency of the Gibbs sampler, avoiding problems with slow mixing due to a tendency to remain in local regions of the model space for long intervals. The models are given equal probability a priori, Pr(Mi)=1/30 for all i.
3.3. Haplotype selection
The primary advantage of the model structure presented above is that it contains a single cell line-specific summary of DNA repair capability capturing the characteristics of the individual distributions. However, while differences in the repair rates are of interest, the main focus of the study is on relating those differences to genetic factors.
Suppose that G candidate genes have been preselected for study, with the gth gene having ng variants in the population. Here, `variants’ refer to haplotypes, which are unique combinations of SNPs. An individual's diplotype for a specific gene is defined by the pair of haplotypes for that gene, one inherited from the mother and one from the father. For the 20 genes that were preselected in the NIEHS study, the number of diplotypes ranges from 3 to 24, with a total of 224. This results in a huge number of unique combinations of diplotypes for the 20 genes.
Given the small number of subjects and the large number of diplotypes involved in the analysis, we consider only additive effects of the genes on the latent factor, so that ηit=uiμt, where ui = (ui1′,…,uiG′)′ is a vector of indicators of the variant category for each of the G genes and μt is a vector of regression coefficients. Extensions to include interactions are straightforward but would not be useful in this specific example since the data on interactions are very sparse. Note that uig contains ng−1 indicators when the gth gene has ng haplotypes, and the intercept term is excluded for identifiability (for the same reason that E(ηit) was set to 0 previously).
Taking advantage of the normal linear regression structure of the models for each of the latent traits, we can apply Bayesian variable selection methods for subset selection in regression, using once again mixture priors with point mass at 0 for the regression coefficients. This structure can be formalized through prior distributions of the form
and a SSVS procedure can then be used to identify genes that can influence the repair capability. If a gene has no effect on the repair rate, the coefficients for each of the subcategories should be 0. The previous approach could be easily extended to identify genes associated with higher susceptibilities to damage in the individuals.
4. INFERENCE
4.1. Posterior computation
By augmenting the observed data with the latent class indicators, itj}, and with latent normal random variables underlying each itj, we can obtain a simple Gibbs sampling algorithm for posterior computation. This algorithm relies on a strategy related to Albert and Chib (1993), but by using a continuation ratio probit instead of generalized probit structure, we avoid the problems in the updating of the threshold parameters mentioned in Johnson and Albert (1999). A related strategy was used by Dunson (2006) in multistate modeling of multiple event data.
Specifically, let zitjk ∼N(αitk,1) and define itj = k if zitju < 0 for all u < k and zitjk > 0, with itj = K if zitju < 0 for all u ≤K−1. The joint posterior distribution of the parameters and latent variables is proportional to
Derivation of the full conditional posterior distributions for each of the unknowns in the model follows by standard algebra. After choosing initial values for the parameters, and given a specific form for p(α|η), the algorithm iterates through the following steps:
Step 1: The latent variables zitjk and itj are updated in block by
- Sampling itj from a discrete distribution with probabilities
- Generating zitjk|itj from
with
Step 2: The parameters for the normal components in the mixture can be generated from:
- For the means, the first component is sampled from
while for k = 2, …, K,
where rk = and hk = . - The variances σk2 are a posteriori conditionally independent, yielding
where rk = .
Step 3: Sample α from
Data augmentation strategies of this kind allow for implementations that rely only on Gibbs samplers, rather than on more general MCMC schemes requiring simultaneous proposals of large numbers of parameters or rejection samplers that could generate even worse mixing issues by forcing us to sample one parameter at a time.
4.2. Prior elicitation
Consider first eliciting hyperparameters {}k=1K and {κθk2}k=1K corresponding to the location of the Gaussian components and aσ and bσ corresponding to their variances. These hyperparameters need to be chosen to ensure that the mixture spans the expected range of observed values with high probability. In practice, we have experimented with 2 types of choices, with essentially equivalent results: (1) Having all prior means {}k=1K equal to the global mean (or global median) of all observations in the sample, and setting all κθk2 equal to half the range of the observed data for all k (a rough estimate of dispersion), and (2) setting equal to the k/(K+1) quantile of the sample, with κθk2 equal to one-eighth of the range. Sensitivity was assessed by halving and doubling the values of κθk2 under each of these 2 scenarios. Under a similar argument, aσ and bσ were chosen so that (σk2)=bσ/(aσ−1) is equal to half the range of the observations. Note that in every scenario we have employed proper priors to avoid identifiability issues with mixture models. See Mengersen and Robert (1996), Natarajan and McCulloch (1998), and references therein for a discussion of this point.
Next, we consider the prior structure on the weights ωitk. As discussed above, the use of a continuation ratio probit model along with normal priors for the transformed weights is convenient, as it greatly simplifies implementation of the model. In particular, the transformed mixture weights {αitk} can be sampled in step 3 above from conditionally normal distributions. Hyperparameter choice is also simplified. A common assumption in mixture models is that all components have the same probability a priori. In the current context, this can be approximately enforced by setting (αjtk) = Φ−1(1/(K−k+1)). Additionally, (αjtk)≈1 because we expect the continuation ratio Φ(αjtk) to be between 0.002 and 0.998 with 0.99 probability. Smaller values for (αjtk) lead to strong restrictions on the set of weights, discouraging small ones (especially for the first few components in the mixture). On the other hand, larger variances do not impose restrictions on the set of weights but can adversely affect model selection.
We concentrate the rest of the discussion on the heterogeneity model described in Section 3.1, but similar arguments apply to the other models considered in the paper. It is well known that, unlike in estimation problems, the effect of the prior in model selection does not necessarily vanish as the sample size grows (Kass and Raftery, 1995). In particular, Bayes factors and posterior probabilities obtained using noninformative and flat priors tend to overly favor the null model even in large samples due to Lindley's paradox. In the heterogeneity model, this means that large variances for the normal component of p(λ(r)) can seriously bias the results. Therefore, we avoid variances larger than 1 for these parameters. On the other hand, the parameters γk or λ(−r) are common to both models under consideration, and improper or flat priors should produce reasonable results. Our sensitivity analysis confirms that model selection results are not dramatically affected by increasing their variances.
5. UNDERSTANDING HETEROGENEITY IN DNA REPAIR STUDIES
The data from the DNA repair study were analyzed using the models described in Section 3. Eight mixture components were judged sufficient to flexibly characterize changes in the density across cell lines and treatment groups, while limiting the risk of overfitting. Inferences were robust in our sensitivity analysis for K ranging between 8 and 15, but the quality of the fit, as assessed through the plots described in Section 5.5, was compromised for K < 8. We considered different options for the hyperparameters , including = 8 for all k (about the empirical mean of the data) and fixing to the k/(K+1)th quantile of the data, all leading to equivalent results. The prior variance was taken to be κθ2 = 25, so as to cover the range of the data, while the hyperparameters for σk2 were set to aσ = 1.0, bσ = 1/2, so that (σk2) = 2 a priori. Again, results were robust to reasonable changes in these parameters.
The Gibbs sampler was run for 60 000 iterations following a 10 000 iteration burn-in period. Code was implemented in FORTRAN, and the longest running time was around 26 h on a 2.80-GHz Intel Pentium 4 computer in the case of the haplotype selection model described in Section 3.3. Examination of diagnostic plots showed adequate mixing and no evidence of lack of convergence. In order to corroborate this observation, we used the Gelman–Rubin convergence test (Gelman and Rubin, 1992), which compares the variability within and between multiple runs of the sampler with overdispersed starting values; we monitored the mean, variance, and skewness of the fitted distributions as our parameters of interest. In every case, confidence intervals for the convergence statistic R contained the reference value 1, as expected for nondivergent chains. Agreement of fitted and empirical distributions, assessed through quantile–quantile plots (see Section 6.4), was adequate for most subjects and experimental conditions.
5.1. Simulation study
As a preliminary to the analysis of the DNA repair data set, a simulation study was run addressing some of the frequency properties of our density comparison approach. The study demonstrated that the models are indeed capable of detecting differences across populations as long a moderate number of components are incorporated in the mixture. This is true even for distribution that has similar mean and variances but differs in higher moments. Additional details can be found in Section 1 of the supplementary material, available at Biostatistics online (http://www.biostatistics.oxfordjournals.org).
5.2. Modeling and testing heterogeneity in Olive tail moment
The estimated posterior probability of the null hypotheses (homogeneity among the cell lines) was 0.0004, and the Bayes factor for the alternative hypothesis was ≈2500. Hence, there was clear evidence of heterogeneity among the cell lines, justifying a more detailed analysis of the data.
5.3. Modeling heterogeneity in DNA repair
The chain visited only 4 out of 30 models; since sampling was performed over the whole model space and not one variable at a time, we believe that lack of mixing is not an issue. The model with highest posterior probability (0.93) contains 4 factor loadings, corresponding to the main effects λ11, λ22, and λ33, along with the correlation between susceptibility and repair rate λ32. The other models include the same 4 variables plus λ21, λ31, or both, with respective posterior probabilities 0.03, 0.02, and 0.01. This reveals clear evidence of heterogeneity in the 3 latent traits, no evidence of dependence between initial damage and repair rate or susceptibility, and decisive evidence of a negative dependence between the repair rate and the susceptibility to damage (since Pr(λ32≤0)≅1). This can be explained by the speed of the repair mechanisms for oxidative damage, with damage being actively repaired even as cells are being exposed.
Figure 3 shows the estimated latent repair rates, ηi3, for each of the cell lines. It is important to emphasize that these values cannot be interpreted in absolute terms, but only relative to other members of the same population. For example, individuals 48, 50, 53, and 82 have the higher repair rates compared with other members of the population, while individuals 4, 15, 22, and 27 have a relatively very low repair rate. Therefore, only a careful choice of the study population, like in the STET project, allows generalizations to the general population.
5.4. Haplotype selection
We focus on 20 candidate DNA repair genes for illustration, resulting in a total of 224 parameters characterizing differences among individuals in repair rates. The genes considered, along with their associated number of diplotypes and posterior probability of influence, are shown in Table 1.
Table 1.
Gene | Number of diplotypes | Posterior probability |
XRCC3 | 8 | 0.89 |
POLG | 7 | 0.83 |
ADPRT | 20 | 0.80 |
ERCC1 | 9 | 0.77 |
ERCC6 | 24 | 0.76 |
POLD1 | 15 | 0.75 |
XPA | 20 | 0.66 |
POLI | 11 | 0.64 |
ERCC5 | 14 | 0.63 |
XRCC2 | 10 | 0.54 |
ERCC3 | 10 | 0.50 |
FEN1 | 3 | 0.50 |
XRCC2 | 10 | 0.50 |
PCNA | 9 | 0.49 |
OGG1 | 6 | 0.39 |
POLL | 6 | 0.33 |
POLD2 | 10 | 0.30 |
LIG1 | 16 | 0.19 |
POLB | 6 | 0.18 |
LIG3 | 10 | 0.15 |
Our prior probability on any gene being significant was set to π0=0.5. Results for the factor loading are similar to those obtained in Section 5.3 and therefore not discussed again. Following convention, we consider genes with a posterior probability greater than 0.75 as potentially involved in the repair mechanisms of oxidative damage, yielding 6 candidates: XRCC3, POLG, ADPRT, ERCC6, POLD1, and ERCC1. At the other extreme, there were 3 genes that had less than 0.25 posterior probability, including LIG3, POLB, and LIG, which therefore do not seem to play a relevant role in this specific repair process.
Some additional information can be obtained by looking at the joint distribution of the diplotypes rather than at its marginal distribution. The 2 most visited models include XRCC2, XRCC3, POLD1, POLG, POLI, ERCC1, ERCC6, and PCNA with a posterior probability of 0.0015 (note that ADPRT is not in the list) and the second most visited model contains ADPRT, XPA, XRCC2, XRCC3, POLD1, POLD2, POLG, POLI, ERCC1, ERCC5, ERCC6, FEN1, and PCNA, with a posterior probability of 0.00145, showing that no combination of genes is clearly preferred to explain the variations across individuals. On the other hand, XRCC3, POLD1, POLG, and ERCC1 appear in all 15 most visited models, while ERCC6 and ADPRT are also often present among the most visited models (13 and 12 times, respectively), with no other gene present with similar regularity.
5.5. Assessing model fit
As a way to assess model fit, we computed quantile–quantile plots of the predictive distribution for each individual against its corresponding empirical distribution. As an illustration, we show in Figure 4 the resulting plots for one individual at each time point for the model in Section 5.4. Dotted lines correspond to a 95% probability interval for the quantile of the predictive distribution.
In general, these plots demonstrate a reasonable fit of the model to the data. However, a slight lack of fit can be observed for some individuals in the high quantiles of the distribution (typically, over 90%), where the predictive distribution would indicate slightly more extreme values than actually observed.
6. DISCUSSION
Motivated by data from comet assay studies of DNA damage and repair, this article has proposed an approach for Bayesian hierarchical density regression, allowing an outcome distribution to change flexibly with multiple predictors and across subjects. The basic idea underlying our method is to use a finite mixture model, with the probability weights following a hierarchical model. Efficient posterior computation is facilitated by using a continuation ratio probit structure with data augmentation. This scheme enables us to take advantage of the well-developed literature on Gaussian models and priors developed for variable selection in linear regression, even though the outcome distributions are clearly nonnormal. In this particular setting, the low dimensionality allows us to sample over the whole model space at once, ensuring good mixing and avoiding common pitfalls.
By incorporating latent factors measuring baseline damage, susceptibility, and rate of repair, we were able to perform inferences on the factors varying across cell lines in the STET database. The finding of clear evidence of heterogeneity in rate of repair raised our interest in identifying factors explaining the differences between individuals. Using candidate genes provided by the investigators, we ran a stochastic search variable selection procedure to identify genes that may be of interest for further study.
Some comments on our SSVS scheme are necessary. Note that our choice of π0 = 0.5 implies that we expect, a priori, at least 6 genes to be significant with a probability 0.94, which is about the number of significant genes in our posterior analysis. This would seem to indicate that the relatively large posterior probabilities obtained might be driven in part by our prior selection. Therefore, our analysis at this stage is necessarily exploratory but allows us to provide some guidance as to where to focus future more detailed studies on repair mechanisms.
FUNDING
This research was partially supported by the Intramural Research Program of the NIH, National Institute of Environmental Health Sciences.
Supplementary Material
Acknowledgments
We would like to acknowledge Leping Li and David Umbach for their helpful comments on the manuscript and Mary Watson for her support with the data files involved. We also appreciate the suggestions from the editor, associate editor, and a second referee, which greatly improved the quality of the manuscript. Conflict of Interest: None declared.
References
- Albert JH, Chib S. Bayesian analysis of binary and polychotomous response data. Journal of the American Statistical Association. 1993;88:669–679. [Google Scholar]
- Behseta S, Kass RE, Wallstrom GL. Hierarchical models for assessing variability among functions. Biometrika. 2005;92:419–434. [Google Scholar]
- Bigelow J, Dunson DB. Semiparametric classification in hierarchical functional data analysis. Technical Report. Institute of Statistics and Decision Sciences. 2005 [Google Scholar]
- Clyde M, George E. Model uncertainty. Statistical Science. 2004;19:81–94. [Google Scholar]
- DeIorio M, Müller P, Rosner G, MacEachern SN. An ANOVA model for dependent random measures. Journal of the American Statistical Association. 2004;99:205–215. [Google Scholar]
- Dunson DB. Bayesian dynamic modeling of latent trait distributions. Biostatistics. 2006;7:551–568. doi: 10.1093/biostatistics/kxj025. [DOI] [PubMed] [Google Scholar]
- Dunson DB, Pillai N, Park J-H. Bayesian density regression. Journal of the Royal Statistical Society, Series B. 2007;69:163–183. [Google Scholar]
- Dunson DB, Watson M, Taylor JA. Bayesian latent variable models for median regression on multiple outcomes. Biometrics. 2003;59:296–304. doi: 10.1111/1541-0420.00036. [DOI] [PubMed] [Google Scholar]
- Ferguson TS. A Bayesian analysis of some nonparametric problems. Annals of Statistics. 1973;1:209–230. [Google Scholar]
- Ferguson TS. Prior distributions on spaces of probability measures. Annals of Statistics. 1974;2:615–629. [Google Scholar]
- Figueiredo MA, Cheng DS, Murino V. Clustering under prior knowledge with application to image segmentation. In: Schölkopf B, Platt J, Hoffman T, editors. Advances in Neural Information Processing Systems. Volume 19. Cambridge, MA: MIT Press; 2007. pp. 401–408. [Google Scholar]
- Gelfand AE, Kottas A, MacEachern SN. Bayesian nonparametric spatial modeling with Dirichlet process mixing. Journal of the American Statistical Association. 2005;100:1021–1035. [Google Scholar]
- Gelman A, Rubin DB. Inference from iterative simulation using multiple sequences. Statistical Science. 1992;7:457–472. [Google Scholar]
- George EI, McCulloch RE. Approaches for Bayesian variable selection. Statistica Sinica. 1997;7:339–374. [Google Scholar]
- Green P, Richardson S. Modelling heterogeneity with and without the dirichlet process. Scandinavian Journal of Statistics. 2001;28:355–375. [Google Scholar]
- Griffin JE, Steel MFJ. Order-based dependent dirichlet processes. Journal of the American Statistical Association. 2006;101:179–194. [Google Scholar]
- Ishwaran H, James LF. Gibbs sampling methods for stick-breaking priors. Journal of the American Statistical Association. 2001;96:161–173. [Google Scholar]
- Jeffreys H. Theory of Probability. 3rd edition. Oxford: Oxford University Press; 1961. [Google Scholar]
- Johnson VE, Albert JH. Ordinal Data Modeling. New York: Springer; 1999. [Google Scholar]
- Kass RE, Raftery AE. Bayes factors. Journal of the American Statistical Association. 1995;90:773–795. [Google Scholar]
- MacEachern SN. Proceedings Section on Bayesian Statistical Sciences. Alexandria, Virginia: Am. Statist. Assoc; 1999. Dependent nonparametric processes; pp. 50–55. [Google Scholar]
- MacEachern SN. Dependent Dirichlet processes. Technical Report. Department of Statistics: Ohio State University; 2000. [Google Scholar]
- Mengersen KL, Robert CP. Testing for mixtures: a Bayesian entropic approach (with discussion) In: Berger JO, Bernardo JM, Dawid AP, Lindley DV, Smith AFM, editors. Bayesian Statistics. Vol. 5. Oxford: Oxford University Press; 1996. pp. 255–276. [Google Scholar]
- Morris J, Carroll R. Wavelet-based functional mixed models. Journal of the Royal Statistical Society, Series B. 2006;68:179. doi: 10.1111/j.1467-9868.2006.00539.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Müller P, Quintana F, Rosner G. Hierarchical meta-analysis over related non-parametric Bayesian models. Journal of Royal Statistical Society, Series B. 2004;66:735–749. [Google Scholar]
- Muthen B, Shedden K. Finite mixture modeling with mixture outcomes using the EM algorithm. Biometrics. 1999;55:463–469. doi: 10.1111/j.0006-341x.1999.00463.x. [DOI] [PubMed] [Google Scholar]
- Natarajan R, McCulloch RE. Gibbs sampling with diffuse proper priors: a valid approach to data-driven inference? Journal of Computational and Graphical Statistics. 1998;7:267–277. [Google Scholar]
- Olive P, Banath J, Durand R. Heterogeneity in radiation-induced DNA damage and repair in tumour and normal cells measured using the `comet’ assay. Radiation Research. 1990;112:86–94. [PubMed] [Google Scholar]
- Östling O, Johanson K. Microelectrophoretic study of radiation-induced DNA damage in individual mammalian cells. Biochemical and Biophysical Research Communications. 1984;123:291–298. doi: 10.1016/0006-291x(84)90411-x. [DOI] [PubMed] [Google Scholar]
- Ramsay J, Silverman B. Functional Data Analysis. New York: Springer; 1997. [Google Scholar]
- Richardson S, Green P. On Bayesian analysis of mixtures with an unknown number of components. Journal of the Royal Statistical Society, Series B. 1997;59:731–792. [Google Scholar]
- Sethuraman J. A constructive definition of Dirichlet priors. Statistica Sinica. 1994;4:639–650. [Google Scholar]
- Singh N, McCoy M, Tice R, Schneider E. A simple technique for quantitation of low levels of DNA damage in individual cells. Experimental Cell Research. 1988;175:184–191. doi: 10.1016/0014-4827(88)90265-0. [DOI] [PubMed] [Google Scholar]
- Stephens M. Bayesian analysis of mixture models with an unknown number of components—an alternative to reversible jump methods. Annals of Statistics. 2000;1:40–74. [Google Scholar]
- Teh YW, Jordan MI, Beal MJ, Blei DM. Sharing clusters among related groups: hierarchical Dirichlet processes. Journal of the American Statistical Association. 2006;101:1566–1581. [Google Scholar]
- Thompson WK, Rosen O. A Bayesian model for sparse functional data. Biometrics. 2008;64:54–63. doi: 10.1111/j.1541-0420.2007.00829.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.