Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2016 Nov 1.
Published in final edited form as: Biometrics. 2016 Feb 12;72(3):955–964. doi: 10.1111/biom.12482

Bayesian Inference for Latent Biologic Structure with Determinantal Point Processes (DPP)

Yanxun Xu 1,2,*, Peter Müller 3,**, Donatello Telesca 4,***
PMCID: PMC5088621  NIHMSID: NIHMS757534  PMID: 26873271

Summary

We discuss the use of the determinantal point process (DPP) as a prior for latent structure in biomedical applications, where inference often centers on the interpretation of latent features as biologically or clinically meaningful structure. Typical examples include mixture models, when the terms of the mixture are meant to represent clinically meaningful subpopulations (of patients, genes, etc.). Another class of examples are feature allocation models. We propose the DPP prior as a repulsive prior on latent mixture components in the first example, and as prior on feature-specific parameters in the second case. We argue that the DPP is in general an attractive prior model for latent structure when biologically relevant interpretation of such structure is desired. We illustrate the advantages of DPP prior in three case studies, including inference in mixture models for magnetic resonance images (MRI) and for protein expression, and a feature allocation model for gene expression using data from The Cancer Genome Atlas. An important part of our argument are efficient and straightforward posterior simulation methods. We implement a variation of reversible jump Markov chain Monte Carlo simulation for inference under the DPP prior, using a density with respect to the unit rate Poisson process.

Keywords: Biomedical, Determinantal point process, Latent structure, Repulsive, Reversible jump Markov chain Monte Carlo

1. Introduction

Independent priors for latent structure are almost never appropriate in biomedical inference. Nevertheless, they are widely used, simply for technical convenience and the lack of good alternatives. In this article we argue for an attractive class of such alternative models in typical inference problems in biostatistics and bioinformatics.

We discuss the use of the determinantal point process (DPP) for modeling latent biologic structure. In particular, we focus on mixture models and feature allocation problems, when the latent components are to be interpreted as biologically meaningful structure. For example, in the case of a mixture model, we might want to interpret components of a mixture as clinically meaningful patient subpopulations. Similarly, when using feature allocation to model latent tumor cell subpopulations we might want to interpret the features as substantially distinct subclones (Xu et al., 2015). In both cases, an important aspect of the problem is the preference for the latent elements being diverse. Such inference is poorly formalized by traditionally used independent priors. We suggest the DPP prior as an attractive alternative to implement repulsive priors. The use of the DPP for mixture models is not novel. It was originally proposed in Affandi et al. (2013), but remains curiously under-used in biomedical literature. The contribution of the following discussion is the emphasis on problems with small to moderate size mixtures, the extension to inference for general latent structure, and the detailed posterior Markov chain Monte Carlo (MCMC) scheme, including easy to implement transdimensional posterior simulation across different size latent structures.

For the moment we restrict attention to parametric mixture models, to be specific and also because such models are perhaps the most common models for latent structure in biomedical applications. For example, popular Bayesian models for clustering and inference on patient subpopulations are variations of the following model. Let yi denote a response for the i-th patient. We assume

yi~h=1Hwhp(yi|μh), (1)

i = 1, …, n, including possibly H = ∞. The component-specific sampling model p(yi | μh) could be, for example, a survival model with parameters μh, possibly including a regression on patient covariates. The use of independent priors for component-specific parameters μh then gives rise to concerns about over-fitting that generates redundant mixture components with similar parameters, leading to unnecessarily complex models and poor interpretability. In particular, such over-fit compromises the interpretation of the mixture components as biologically meaningful structure. Rousseau and Mengersen (2011) argued that such concerns were asymptotically partially mitigated with carefully chosen priors. Alternatively, Petralia et al. (2012) proposed a class of repulsive priors for mixture components. The proposed repulsive prior was based on a distance metric in which small distances were penalized. They showed that using repulsive priors on location parameters resulted in better separated clusters, while keeping the density estimation accurate. However, posterior computations are complex and do not readily extend to high dimensional cases.

An alternative interpretation of (1) is as a mixture, yi ∼ ∫p(yi | μ)dG(μ), with respect to a discrete probability measure G = Σwhδμh. If the model is completed with a Dirichlet process (DP) prior on G the popular DP mixture model is obtained. See, for example, Ghoshal (2010) for a review of such nonparametric Bayesian models. Importantly, the DP prior includes independence across μh.

For later reference note that (1) can be equivalently written as a hierarchical model with latent indicators si,

yi|si=k~p(yi|μk)andp(si=k)=wk. (2)

Interpreting the latent indicators as cluster membership indicators, model (2) includes inference on a random partition s = (s1, …, sn) of {1, …, n}. Let Sk = {i : si = k} denote the k-th cluster. To avoid the notion of empty clusters, that is |Sk| = 0, we re-arrange the indexing of the μh to start with h = 1, …, K corresponding to non-empty clusters. Again, an independent prior on the cluster-specific parameters μh complicates a meaningful interpretation of posterior inference on the random partition s.

In this article we argue for an alternative model that replaces the independent prior on μh by the repulsive DPP (Macchi, 1975). Recent reviews of the DPP appear in Lavancier et al. (2015) and, specifically for finite state spaces, in Kulesza and Taskar (2012). The use of the DPP as a prior for statistical inference in mixture models, we believe, is first discussed in Affandi et al. (2013). The main contributions of this article are the recognition of the DPP as an attractive prior for latent features in general latent structure models, including mixture models and latent feature allocation as specific examples; the discussion of DPP mixtures specifically when one wants to interpret latent structure as biologically meaningful features; and an easily implemented posterior simulation scheme for a moderate number of latent structures, as is typical for biomedical inference problems. Posterior simulation is implemented as a variation of reversible jump (RJ) MCMC simulation (Green, 1995) for a density with respect to the unit rate Poisson process.

2. Motivating Example

Magnetic resonance imaging (MRI) is an effective technique for studying the human brain. For example, MRI volume estimates of white matter (WM), gray matter (GM), cerebrospinal fluid (CSF), and their spatial distribution help the diagnosis of degenerative brain illnesses, like Alzheimer's disease (DeCarli et al., 1992). Therefore, accurate clustering of MRI data according to tissue types is vital to diagnosis and clinical research. To illustrate, we download a sample of simulated imaging data from BrainWeb (Cocosco et al., 1997) for slice number 92. Figure 1a depicts the ground truth components for CSF, WM, and GM. We implement inference under model-based clustering with a DPP prior and a similar model based on the widely used Dirichlet process mixture (DPM) model. Model details will be discussed later. For the moment we only intend to highlight the nature of the inference under the two models to motivate the upcoming discussion. Figure 1c shows the posterior distribution p(K | data) on the number of clusters estimated under the DPP prior (left panel) and the DPM prior (right panel). As shown in Figure 1b, the DPP clustering model identifies four clusters, three of which match the simulation truth and the last one is simulated noise. In contrast, inference under the DPM model finds seven clusters, only three of them having a meaningful explanation.

Figure 1.

Figure 1

BrainWeb images. Panel (a) shows the three true tissue types: CSF, WM and GM. Panel (b) shows the estimated tissue types under the DPP prior. Panel (c) shows the estimated number of tissue types by DPP (left) and DPM (right). Panel (d) shows the estimated tissue types under the DPM prior.

3. Determinantal Point Process (DPP)

3.1. Definition

The DPP defines a point process on S ⊆ ℜD, that is, a random point configuration X = {x1, …, xK} with xkS. We first define it for a finite state space, S = {ω1, …, ωN}. Let C denote an (N × N) positive semidefinite matrix, constructed, for example, as Cij = C(ωi, ωj) with a covariance function C(ωi, ωj). Let CA denote the submatrix of rows and columns indicated by AS. In later applications we will identify xk as μk in mixture models like (2), latent feature allocations etc. For the moment we consider a generic random point configuration X, defined as

p(X=A)=det(CA)/det(C+I) (3)

as a probability distribution on the 2N possible point configurations XS. This defines a subclass of DPPs known as L-ensembles. It is easy to see why (3) defines a repulsive point process if one interprets the determinant as the volume of a parallelotope spanned by the column vectors of CA. Equal or similar column vectors span less volume than very diverse ones. Equation (3) can be shown to imply the marginal probabilities

p(AX)det(MA) (4)

for M = C(I + C)−1 (Kulesza and Taskar, 2012), where MA is a submatrix of M. Equation (4) defines a DPP on a finite state space S. Every L-ensemble is a DPP. But not every DPP is an L-ensemble. For singular (IM) we can define (4), but not (3). A good review of DPP models for finite state spaces, including the derivation of the normalizing constant in (3) appears in Kulesza and Taskar (2012).

For a continuous state space S ⊆ ℜD, we define an L-ensemble by a density f(X) with respect to the unit rate Poisson process as

f(X)=det(CX)/h=1(λh+1). (5)

for X = {x1, …, xK}. As before, CX is a (K × K) matrix with (i, j) entry defined by a continuous covariance function C(xi, xj). The λh's are the eigenvalues of the associated kernel operator ∫SC(x, y)h(y)dy. Similar to the case of a finite state space, it is possible to generalize (5) to the slightly larger class of DPP models (Lavancier et al., 2015). However, for the rest of this discussion we will consider L-ensembles and work with the kernel C(xi, xj) only.

For continuous DPP kernels, the eigenvalues λh are generally unknown except for a few kernels such as a squared exponential kernel. Several numerical methods are used to approximate eigenvalues and corresponding eigenfunctions (Lavancier et al., 2015). We build on Kulesza and Taskar (2010) and decompose the kernel function C as

C(x,x)=q(x)c(x,x)q(x)x,xχandc(x,x)=1, (6)

where q(x) is the quality function and c(x, y) is the similarity kernel. For a multivariate x = (x1, …, xD) ∈ ℜD we use

q(x)=d=1D12πσqexp{xd22σq2}andc(x,x)=exp{d=1D(xdxd)2θ2},

for which Zhu et al. (1998) gives analytic results for the eigenvalues and eigenfunctions. Eigenvalues λh are given by:

λh=d=1D2aa+b+c(ba+b+c)hd1, (7)

where h = (h1, …, hD) is a multivariate index, a=14σq2, b=1θ2, and c=a2+2ab. Here, θ and σq are hyperparameters that define the kernel function.

We write X ∼ DPP(C, θ, σq) for X = {x1, …, xK} generated by a DPP model with a kernel function C(·, ·) that is indexed with parameters θ, σq, and we write DPP(C) when C(·, ·) involves no unknown hyperparameters.

3.2. Posterior Simulation

Later we will use the DPP as prior probability model for latent structure, including latent clustering and feature allocation. In both cases, an important step in the posterior simulation will be a transition probability to change the number of atoms in the DPP. We discuss a reversible jump (RJ) scheme to implement such transition probabilities using the density (5) with respect to the unit rate Poisson process. Let ΩK denote the σ-algebra for size K point configurations, and Ω=K=0ΩK. We define an MCMC transition probability that allows a move from FK ∈ ΩK to FK+1 ∈ ΩK+1 or FK−1 ∈ ΩK−1. The algorithm combines the MCMC simulation for a point process from Geyer and Møller (1994) with the deterministic transformation that is included in the reversible jump (RJ) scheme of Green (1995). The construction parallels the construction of Green (1995), with only a minor variation that is needed to reduce the integral with respect to the unit rate Poisson process to an integral with respect to Lebesgue.

Assume the current state is x = {x1, …, xK} and we consider two transition probabilities, Pu(dy | x) which proposes a move to a size K + 1 point configuration (“up” move) and Pd(dx | y) which proposes a move to a size K − 1 point configuration (“down” move). For example, Pu could be proposing to split one of the atoms in x into two daughters, thereby incrementing K by one; and Pd could involve merging two points in x. Let q(x) denote the probability of choosing Pu, and let Au(x, y) and Ad(y, x) denote the acceptance probability for a proposal y. Finally, let f(x) denote the density (5) with respect to the unit rate Poisson process μ(·). The detailed balance condition becomes

FK+1(1q(y))[FkAd(y,x)Pd(dx|y)]f(y)dμ(y)=FKq(x)[Fk+1Au(x,y)Pu(dy|x)]f(x)dμ(x). (8)

Assume that there are nup(x) possible up moves, j = 1, …, nup(x). For example, if the up move involves splitting one of the atoms of the size K point configuration x, we could choose one of the nup(x) = K points to split. Let quj(x) denote the probability of selecting the j-th transition probability. That is Pu(dy | x) = Σjquj(x)Puj(dy | x). Similarly, Pd(dx | y) = Σjqdj(y)Pdj(dx | y). A sufficient condition for detailed balance is that equation (8) holds for pairs of moves, Puj, Pdf′ that are defined and linked in the following sense. We assume that Puj is constructively defined by (i) generating an auxiliary variable uqu(u | x); (ii) a deterministic, invertible transformation y = T(x, u); and (iii) the matching down move Pdj′ is defined by x=T11(y). Here T11(y) denotes the first element of T−1(y) = (x, u). The detailed balance condition becomes

(1q(y))qdj(y)[Ad(y,x)I{x=T11(y)FK;yFK+1}]×f(y)dμ(y)=q(x)quj(x)[Au(x,y)I{xFK;y=T(x,u)FK+1}qu(u|x)du]f(x)dμ(x).

We replaced the range of integration by an indicator for xFK and yFK+1. Next we use FKh(x)dμ(x)=e|S|K!h(x)I(xFK)dx1dxK. That is, a unit rate Poisson process restricted to size K point configurations looks exactly like K i.i.d. uniform random variables on S (Kingman, 1992). The extra factor e−|S||S|K/K! arises from the probability of a size K point configuration. Note that x = {x1, …, xK} remains the (unordered) point configuration. We get

(1q(y))qdj(y)[Ad(y,x)I{xFk;yFK+1}]×f(y)(K+1)!dy1dyK+1=q(x)quj(x)×[Au(x,y)I{xFk;yFK+1}qu(u|x)du]×f(x)K!dx1dxK, (9)

still using x=T11(y) on the left and y = T(x, u) on the right hand side. Finally, we use a change of variables, substituting dy1dyK+1 by dx1dxKdu|J| with the Jacobian J = ∂T/∂x1∂xK∂u. A sufficient condition for (9) is the equality of the two integrands, (1q(y))qdj(y)Ad(y,x)f(y)K+1|J|=q(x)quj(x)Au(x,y)qu(u|x)f(x), for xFK and y = T(x, u) ∈ FK+1. The condition is verified for Au(x, y) = min{1, ρ(x, y)} and Ad(y, x) = min{1, 1/ρ(x, y)} with

ρ(x,y)=f(y)(K+1)f(x)1q(y)q(x)qdj(y)quj(x)1qu(u|x)|J|. (10)

Acceptance probability (10) defines essentially the RJ algorithm of Green (1995). The only minor difference is the extra step of representing the probability of a point configuration with respect to the unit rate Poisson process by a probability of the ordered K-tuple (x1, …, xK). Geyer and Møller (1994) use the latter for a birth and death Markov chain Monte Carlo, and without the deterministic transformation. For posterior simulation conditional on data yp(y | x, θ) multiply with an additional likelihood ratio in (10). Here θ are additional parameters in the sampling model, beyond x.

In summary, we have shown that the density with respect to the unit rate Poisson process can be used to construct a RJ MCMC, essentially as if it were a density with respect to Lebesgue. A similar argument holds for Metropolis-Hastings transition probabilities, without a change in the size of the point configuration x.

4. DPP Clustering

4.1. Motivation and Model

Clustering is fundamental to exploratory analysis of bioinformatics data. For instance, elucidating patterns of gene expression and identifying sets of genes that behave similarly under certain biologic conditions is important in the study of functional genomics and proteomics. Clustering also can be applied to develop targeted therapies. We first cluster the patient samples into several subgroups based on protein activation (or some other patient baseline characteristics), then correlate patient clusters with overall survival and investigate subgroup-specific therapies. These and similar applications in biomedical inference motivate the following model.

We start with a mixture of normals sampling model, as it is widely used in clustering and density estimation. Here, we show simulation with the univariate sampling model (2). In Web Appendices A and B we show a straightforward extension to a multivariate mixture, including a brief simulation study. We assume that data yn={yi}i=1n are generated from yi~k=1KwkN(|μk,σk2), i = 1, …, n, with unknown K. This is a special case of (1) with random K and with a normal kernel p(yi|μk,σk2). The model implies a prior for a random partition s = (s1, …, sn), as in (2). Often the inference goal is to identify latent clusters Sk = {i : si = k} that correspond to meaningful biologic conditions or to identify subpopulations that are sufficiently diverse to be considered for different clinical decisions such as treatment allocation. The protein data analysis for kidney cancer patients, in Section 4.2, is a typical example. In such problems an independent prior on μk has the undesirable feature of allowing for very similar, or even identical (in the case of a discrete parameter space) μk. To interpret different terms in the mixture as meaningful structure in the population, we prefer a repulsive prior on the μk, that is, a probability model that favors a priori very distinct values μk. The repulsive property and the relative computational simplicity make the DPP an appealing choice. Kwok and Adams (2012) applied the DPP as repulsive prior in latent variable models. However, for lack of efficient computational algorithms their method was restricted to MAP (maximum a posterior) inference. Affandi et al. (2013) proposed a Gibbs sampling technique for inference with DPP priors under fixed K (K-DPP). The posterior simulation scheme from Web Appendix A allows us to implement inference under an unconstrained DPP prior, including a random size (K) point configuration.

The DPP mixture model

We complete the sampling model (2) with a DPP prior on the cluster-specific parameters μk:

yi|si=k~p(yi|μk)andp(si=k)=wk.μ={μ1,,μK}~DPP(C,θ,σq), (11)

using the kernel function in (6). Recall that θ, σq are hyperparameters in the definition of C. Finally we use hyperpriors w | K, δ ∼ Dir(δ, …, δ), 1/σk2~Ga(a0,b0), θ~N(a1,b12), and σq~N(a2,b22). Here Ga(a, b) refers to a Gamma distribution with mean a/b. The model can be easily extended to multivariate responses using multivariate normal and inverse-Wishart priors. Posterior inference is carried out using MCMC simulations. Details are shown in Web Appendix A.

Two simulation studies

We carry out two simulation studies to evaluate the performance of the repulsive DPP prior in clustering and density estimation, with both univariate and multivariate responses. Results are summarized in Figure 2. Details of the simulation setup and more results are shown in Web Appendix B. See there also for more discussion of Figure 2 and for a statement of the multivariate version of the DPP mixture model. The results show that the DPP prior leads to a sparser representation with interpretable clusters compared to DPM, while maintaining a good fit for the density estimate, making it a preferable prior model for applications where such parsimony is desired.

Figure 2.

Figure 2

Simulation: DPP mixture model. The upper panel shows the histograms of two simulated datasets with true density (solid), estimated density by DPP prior (dashed) and DPM prior (dotted). The lower panels present the histograms of the estimated number of clusters by DPP prior (2nd row) and DPM prior (3rd row).

4.2. KIRC Protein Data Analysis

We implement inference under the proposed DPP mixture model for protein expression data from Yuan et al. (2014) with n = 243 samples and D = 17 protein markers for kidney renal clear cell carcinoma (KIRC). See equation (4) in Web Appendix A.2 for a statement of the DPP mixture model (4.1) with a multivariate normal kernel p(yi | μk). Inference goals include correlating protein expression with patients' overall survival. The n = 243 KIRC samples are classified into three clusters by the proposed DPP mixture model. As shown in Figure 3a, patients stratified by these three DPP groups exhibit very distinct survival patterns (p-value under a log-rank test is p = 0.00027). Proteins that are correlated with better prognosis are relatively elevated in cluster 2 (the best survival group) while the proteins correlated with worst survival are relatively elevated in clusters 1 and 3 (especially cluster 3, the worst survival group) (Figure 3b). These results suggest that inference under the DPP prior can successfully classify patients into biologically meaningful groups based on molecular profiles.

Figure 3.

Figure 3

KIRC data. Panel (a) shows a Kaplan–Meier plot of overall survival in the KIRC core set stratified by three clusters identified under DPP prior. Panel (b) shows the top differentially expressed protein markers among three DPP clusters. Columns correspond to patients, rows correspond to proteins. Panel (c) is the histogram of the number of clusters identified under DPM prior. Panel (d) shows a heatmap of seven DPM clusters for top differentially expressed protein markers. The sizes of the seven clusters are 163, 19, 39, 14, 6, 1, and 1, respectively. This figure appears in color in the electronic version of this article.

In contrast, under the DPM mixture model, K = 7 clusters are identified, estimated as the mode of p(K | y) shown in Figure 3c. Most of the seven clusters have small size (< 20) while 2/3 of the samples are allocated to one cluster (the red bar in Figure 3d). The clusters are not easily interpreted (Figure 3d).

In summary, inference under the DPP prior provides fewer clusters and gives more interpretable results in molecular profile-based classifications than inference under a comparable DPM prior.

5. A DPP Feature Allocation Model

5.1. Motivation and Model

Breast cancer is a heterogeneous disease in terms of molecular alterations and clinical responses. Gene expression profiling can provide valuable information for understanding this complexity and consequently for predicting clinical outcomes. Here, we consider a study reported in Chen et al. (2015) who aim to characterize gene expression profiles by a small number of underlying distinct molecular drivers. These latent molecular drivers should be linked to different subsets of samples. This motivates us to propose the model below which formalizes this preference by using a DPP prior for the pattern of how molecular drivers (the columns of the matrix Z below) are linked to samples (rows of Z). Let Y denote the observed n × S data matrix with rows representing samples and columns representing genes. Let Z be an n × K binary matrix with zik = 1 if molecular driver k presents in sample i, and 0 otherwise. That is, the k-th column zk defines the subset Gk = {i : zik = 1} of samples that are linked with the k-th molecular driver. The entire matrix Z defines a multiset {Gk, k = 1, …, K}. Such multisets are known as feature allocation (Broderick et al., 2013) and are popular tools in machine learning to implement inference about overlapping subsets of experimental units (customers etc.). The special case of non-overlapping subsets that cover all samples, that is, GkG = ∅ and ∪Gk = {1, …, n), is a partition. See Broderick et al. (2013) for a recent review. We use the feature allocation matrix Z to construct a sampling model for the breast cancer gene expression data Y:

Y=Zβ+E, (12)

where β is a K × S loading matrix with each entry βkj weighing the contribution of gene j to the driver k and E = [eij] is an error matrix with eijN(0, σ2), independently. This defines a sampling model for the observed gene expressions Y in terms of assumed latent structure Z. That is, yij~N(k=1KZikβkj,σ2).

The key assumption in a feature allocation model is the prior model on Z. A technically convenient and traditional prior is the Indian buffet process (IBP) (Ghahramani and Griffiths, 2006). One of the key properties of the IBP, in the context of this application, is the implied independence across columns of the binary matrix (re-arranging columns in left ordered form or by other constraints introduces a trivial form of dependence). This independence is undesirable for the desired inference on molecular drivers. In particular, independence across columns implies a positive prior probability for identical columns, which is meaningless in the interpretation of columns as distinct molecular drivers.

DPP feature allocation

In contrast to the IBP, a DPP prior on the columns zk formalizes the desired parsimony in identifying latent molecular drivers. We assume

{zk,k=1,,K}~DPP(C)withC(zl,zl)=exp{i=1n(zilzil)2θ2}. (13)

With large n, there is no effective way to decompose the kernel matrix C (N × N matrix with N = 2n) and to compute the eigenvalues and the corresponding eigenvectors. We therefore fix θ in (13) and complete the model with a conditionally conjugate prior on the coefficients, βkj~i.i.d.N(0,τ2), and hyperpriors 1/σ2 ∼ Ga(a0, b0), 1/τ2 ∼ Ga(a1, b1).

DPP-K feature allocation

In the upcoming applications we find it convenient to work with a slight variation of model (13). Let DPPK(C) denote a DPP prior restricted to a fixed number of atoms, K, and define

{zk,k=1,,K}|K~DPPK(C)andK~p(K) (14)

for some prior p(K). We refer to the model as the DPP-K feature allocation model. The reason for introducing the DPP-K model is that it facilitates a computationally efficient posterior simulation scheme. Under (13) we require a RJ type implementation, following the general scheme in Section 3.2. However, in some applications it is difficult to construct proposal distributions that lead to reasonably mixing Markov chains. Instead, we propose in Web Appendix C an alternative MCMC scheme under (14), which we find to work well for applications with feature allocation problems. Posterior inference in model (13) as well as in (14) is implemented by MCMC posterior simulation. Details are shown in Web Appendix C.

Simulation study

We carry out a simulation study to compare inference under the DPP-K feature allocation prior versus the IBP prior. Results are summarized in Figure 4. More discussion of the results and details of the simulation setup are in Web Appendix D. With fewer latent features, inference under the DPP prior can better and more parsimoniously recover the simulation truth than under a standard IBP prior in this simulated example.

Figure 4.

Figure 4

Simulation: DPP feature allocation model. Panels (a-c) show the true feature allocation matrix Zo and the estimate under the DPP prior and the IBP prior, respectively. Panels (d-f) show the true feature mean βo and the estimated β̂ under the DPP prior and under the IBP prior, respectively. Panels (g-h) are histograms of the number of latent features identified under DPP prior and IBP prior, respectively.

5.2. Breast Cancer (BRCA) Data Analysis

We analyze the TCGA BRCA mRNA expression data (The Cancer Genome Atlas Network, 2012). We focus on n = 150 tumor samples classified as basal-like, HER2-enriched (HER2) and luminal A (LumA) subtypes by PAM50, a well-established 50-gene signature for distinguishing the gene expression-based “intrinsic” subtypes of breast cancer (Parker et al., 2009). Among those three subtypes, the HER2-enriched subtype is well studied. There are effective therapeutic drugs developed for targeting HER2 breast cancer. The basal-like subtype (also known as triple-negative breast cancer due to its lacking of expression of estrogen receptor (ER), progesterone receptor (PR) and HER2), and the LumA subtype, which is known to have lowest overall mutation rate, are poorly understood. As a result, there is currently no effective targeted therapy for these two subtypes, leaving chemotherapy as the main therapeutic treatment. A better characterization of basal-like and LumA subtypes at the molecular level is needed for clinical studies.

We implement inference under the proposed DPP-K latent feature model and identify K = 3 latent features. Figure 5a shows the posterior inferred latent feature matrix Z with different breast cancer subtypes samples marked on the left. The basal-like, HER2 and LumA samples show clear and distinct patterns: 35 of 50 basal-like samples exhibit the first and third features and 44 of 50 are depleted with respect to the second feature; 43 of 50 HER2 samples exhibit the first two features and 33 of 50 are depleted with respect to the third feature; 48 of 50 LumA samples exhibit the second and third features and 47 of 50 are depleted with the first feature. More biological findings are discussed in Web Appendix E.

Figure 5.

Figure 5

BRCA data. Estimated feature allocation matrix under the DPP prior (panel a) and under the IBP prior (b).

For comparison, we analyze the same BRCA dataset under a model with an IBP prior. It identifies 27 latent features, of which 17 are active in less than 4 samples. Figure 5b shows the estimated latent feature matrix Z. The first three features identified under the IBP prior can distinguish different breast cancer subtypes: 44 of 50 basal-like samples exhibit the first feature; 4 of 50 HER2 samples and none of LumA samples exhibit the first feature; 48 of 50 LumA samples exhibit the third feature. However, for the remaining 24 features, we can not observe any pattern for different breast cancer subtypes: these features were sparsely scattered across all samples. This is a good example of how the independent prior across features, as it is implied in the IBP model, leads to a lack of parsimony and difficult interpretability in the latent structure. In summary, the DPP prior model provides a less complicated representation and more interpretable features than inference under the IBP prior model.

6. Supplementary Materials

Web appendices and figures referenced in Sections 4.1, 5.1, and 5.2, as well as the code and data, are available with this article at the Biometrics website on Wiley Online Library.

Supplementary Material

Supp Info
Supp Material Code

Acknowledgments

Peter Müller and Yanxun Xu's research is partly supported by NIH grant R01 CA132897.

References

  1. Affandi RH, Fox E, Taskar B. Approximate inference in continuous determinantal processes. Advances in Neural Information Processing Systems. 2013:1430–1438. [Google Scholar]
  2. Broderick T, Jordan MI, Pitman J, et al. Cluster and feature modeling from combinatorial stochastic processes. Statistical Science. 2013;28:289–312. [Google Scholar]
  3. Broderick T, Pitman J, Jordan MI. Feature allocations, probability functions, and paintboxes. Bayesian Analysis. 2013;8:801–836. [Google Scholar]
  4. Chen M, Gao C, Zhao H. Posterior contraction rates of the phylogenetic indian buffet processes. Bayesian Analysis. 2015 doi: 10.1214/15-BA958. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Cocosco CA, Kollokian V, Kwan RKS, Pike GB, Evans AC. NeuroImage. Vol. 5. Princeton, New Jersey, USA: Citeseer; 1997. Brainweb: Online interface to a 3D MRI simulated brain database; p. 425. [Google Scholar]
  6. DeCarli C, Maisog J, Murphy DG, Teichberg D, Rapoport SI, Horwitz B. Method for quantification of brain, ventricular, and subarachnoid CSF volumes from MR images. Journal of Computer Assisted Tomography. 1992;16:274–284. doi: 10.1097/00004728-199203000-00018. [DOI] [PubMed] [Google Scholar]
  7. Geyer CJ, Møller J. Simulation procedures and likelihood inference for spatial Point processes. Scandinavian Journal of Statistics. 1994;21:359–373. [Google Scholar]
  8. Ghahramani Z, Griffiths TL. Advances in Neural Information Processing Systems. Boston, MA: IEEE Conference on Neural Information Processing Systems–Natural and Synthetic, Massachusetts Institute of Technology Press; 2006. Infinite latent feature models and the Indian buffet process; pp. 475–482. [Google Scholar]
  9. Ghoshal S. The Dirichlet process, related priors and posterior asymptotics. In: Hjort NL, Holmes C, Müller P, Walker SG, editors. Bayesian Nonparametrics. New York, NY, USA: Cambridge University Press; 2010. pp. 22–34. [Google Scholar]
  10. Green PJ. Reversible jump Markov chain Monte Carlo computation and Bayesian model determination. Biometrika. 1995;82:711–732. [Google Scholar]
  11. Kingman JFC. Poisson Processes. New York, NY, USA: Oxford University Press; 1992. [Google Scholar]
  12. Kulesza A, Taskar B. Structured determinantal point processes. Advances in Neural Information Processing Systems. 2010:1171–1179. [Google Scholar]
  13. Kulesza A, Taskar B. Determinantal point processes for machine learning. Machine Learning. 2012;5:123–286. [Google Scholar]
  14. Kwok JT, Adams RP. Priors for diversity in generative latent variable models. Advances in Neural Information Processing Systems. 2012:2996–3004. [Google Scholar]
  15. Lavancier F, Møller J, Rubak E. Determinantal point process models and statistical inference. Journal of the Royal Statistical Society: Series B, (Statistical Methodology) 2015;77:853–877. [Google Scholar]
  16. Macchi O. The coincidence approach to stochastic point processes. Advances in Applied Probability. 1975;7:83–122. [Google Scholar]
  17. Parker JS, Mullins M, Cheang MC, Leung S, Voduc D, Vickery T, Davies S, Fauron C, He X, Hu Z, Quackenbush JF. Supervised risk predictor of breast cancer based on intrinsic subtypes. Journal of Clinical Oncology. 2009;27:1160–1167. doi: 10.1200/JCO.2008.18.1370. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Petralia F, Rao V, Dunson DB. Repulsive mixtures. Advances in Neural Information Processing Systems. 2012:1889–1897. [Google Scholar]
  19. Rousseau J, Mengersen K. Asymptotic behaviour of the posterior distribution in overfitted mixture models. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 2011;73:689–710. [Google Scholar]
  20. The Cancer Genome Atlas Network. Comprehensive molecular portraits of human breast tumours. Nature. 2012;490:61–70. doi: 10.1038/nature11412. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Xu Y, Müller P, Yuan Y, Gulukota K, Ji Y. MAD Bayes for tumor heterogeneity – feature allocation with exponential family sampling. Journal of the American Statistical Association. 2015;110:503–514. doi: 10.1080/01621459.2014.995794. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Yuan Y, Van Allen EM, Omberg L, Wagle N, Amin-Mansour A, Sokolov A, Byers LA, Xu Y, Hess KR, Diao L, Han L. Assessing the clinical utility of cancer genomic and proteomic data across tumor types. Nature Biotechnology. 2014;32:644–652. doi: 10.1038/nbt.2940. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Zhu H, Williams C, Rohwer R, Morciniec M. Gaussian regression and optimal finite dimensional linear models. NATO ASI Series Series F: Computer and System Sciences. 1998:167–184. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supp Info
Supp Material Code

RESOURCES