Abstract
We develop a generalized method of moments (GMM) approach for fast parameter estimation in a new class of Dirichlet latent variable models with mixed data types. Parameter estimation via GMM has computational and statistical advantages over alternative methods, such as expectation maximization, variational inference, and Markov chain Monte Carlo. A key computational advantage of our method, Moment Estimation for latent Dirichlet models (MELD), is that parameter estimation does not require instantiation of the latent variables. Moreover, performance is agnostic to distributional assumptions of the observations. We derive population moment conditions after marginalizing out the sample-specific Dirichlet latent variables. The moment conditions only depend on component mean parameters. We illustrate the utility of our approach on simulated data, comparing results from MELD to alternative methods, and we show the promise of our approach through the application to several datasets. Supplementary materials for this article are available online.
Keywords: Generalized method of moments, Latent Dirichlet allocation, Latent variables, Mixed membership model, Mixed scale data, Tensor factorization
1. Introduction
Many modern statistical applications require the analysis of large-scale, heterogeneous data types including continuous, categorical, and count variables. For example, in social science, data often consist of collections of different data types (e.g., height, gender, and age); in population genetics, researchers are interested in analyzing genotype (integer-valued) and heterogeneous traits (e.g., blood pressure, BMI, alcoholic drinks per day). Often data take the form of an n × p matrix Y = (y1, …, yn)T, with yi (yi1, …, yip)T a p dimensional vector of measurements of varying types for subject i, for i =1, …, n.
This article focuses on (1) developing a new class of generalized latent variable models for mixed data types, and (2) performing fast and robust parameter estimation using the generalized method of moments (GMM). Many models have been proposed to address the first goal (reviewed below), but often these models lack robustness both statistically and computationally. In addition, parameter estimation in these models is often both inefficient and unstable, which necessitates our second goal. Hence, there is a clear need for new classes of models and corresponding robust and efficient approaches for routine inference under these models in general applications.
For modeling mixed scale data, there are two general strategies that have been most commonly used in the literature. The first is to assume an underlying Gaussian model, and then to characterize the feature dependence through a structured model of the Gaussian covariance (Muthén 1984). Such models are routinely used in the social science literature, focusing almost entirely on the case in which data are categorical or continuous, with the categorical variables arising via thresholding of Gaussian variables. These models are related to Gaussian copula models, which have been developed for broad classes of mixed scale data also including counts (Murray et al. 2013). The second approach is to define an exponential family distribution for each of the p variables, inducing dependence between variables through generalized linear models containing shared latent variables (Sammel, Ryan, and Legler 1997; Moustaki and Knott 2000; Dunson 2003).
There are a number of disadvantages to the above approaches. The underlying Gaussian framework is restrictive in forcing continuous variables to be Gaussian and non-continuous variables to be categorical. The exponential family latent variable models are potentially more flexible, but in practice suffer in several respects. Most notably, there is a fundamental lack of robustness due to the model structure, which arises because of the dual role of the latent variables in controlling sample dependence and the shape of the marginal distributions. For example, if there are two count variables, yi1 and yi2, that are highly correlated, then both counts must load strongly on a similar set of latent variables, which will induce over-dispersion in the marginal distributions with the variance much larger than the mean. There is no possibility of having highly correlated counts that do not have high over-dispersion. Copula models solve this problem by imposing a restrictive Gaussian copula covariance structure. Another fundamental issue is computation, which tends to rely on expectation-maximization (EM) or Markov chain Monte Carlo (MCMC) algorithms that alternate between updating latent variables and population parameters, intrinsically leading to slow convergence, inefficiency, and instability.
One promising approach to address the above issues is to rely on a generalized method of moments (GMM) estimator, which relates to the second goal in this article. The GMM estimator is robust to misspecification of the higher-order moments, while also leading to optimal efficiency among all estimators using only information on the initial moments (Hansen 1982). Applications of GMM to structural models with latent variables have a long history, with early examples including Bentler (1983) and Anderson and Gerbing (1988), among others. More recently, Gallant, Giacomini, and Ragusa (2013) applied GMM to a specific class of latent variable models by defining moment conditions based on the complete data, including the latent variables. In contrast, Bollen, Kolenikov, and Bauldry (2014) relied on a model-implied instrumental variable GMM estimator. Generally, current GMM methods focus on latent variable models that satisfy restrictive assumptions or require the instantiation of latent variables in a computationally intensive estimation algorithm.
The focus of our work is on a broad new class of latent variable models for mixed scale data. This eliminates some of the problems of current exponential family latent variable models, while also enabling derivation of an efficient GMM implementation that marginalizes out latent variables. With the first goal in mind, we focus on generalized mixed membership models, which incorporate Dirichlet latent variables. Mixed membership models have a rich literature with applications ranging from inference of ancestral populations in genomic data (Pritchard, Stephens, and Donnelly 2000a; Pritchard et al. 2000b) to topic modeling of documents (Blei, Ng, and Jordan 2003). Using Dirichlet latent variables to define cluster membership allows samples to partially belong to each of k latent components. Our model class includes latent Dirichlet allocation (Blei, Ng, and Jordan 2003) and simplex factor models (Bhattacharya and Dunson 2012) as special cases; however, we go beyond current approaches by allowing mixed scale data.
For the second goal, we develop an efficient moment tensor approach for parameter estimation with the Dirichlet latent variables marginalized out. This objective is related to recent moment tensor methods developed for latent variable models including mixtures of Gaussians, hidden Markov models, mixed membership models, and stochastic block models (Arora, Ge, and Moitra 2012; Anandkumar, Hsu, and Kakade 2012a; Anandkumar et al. 2012b; Hsu and Kakade 2013; Anandkumar et al. 2014a, 2014b). This recent work finds moment tensors of these latent variable models that have a symmetric PARAFAC (parallel factors) tensor decomposition (Kiers 2000). Parameter estimation using a moment tensor approach is performed as follows: (1) third-order empirical moment tensors are transformed to an orthogonal decomposable form; (2) orthogonal decompositions are performed using tensor power methods or singular value decompositions; (3) parameter estimates are recovered from the decomposition of the moment tensors. The moment tensor approach offers substantial computational advantages over other approximation methods, as illustrated in various applications (Tung and Smola 2014; Anandkumar et al. 2014a; Colombo and Vlassis 2015).
In this work, we develop a moment tensor approach for generalized mixed membership models. Our approach, moment estimation for latent Dirichlet models (MELD), differs from previous approaches in that the moment functions are defined for variables with different generative distributions. This goes beyond previous moment tensor approaches that construct moment tensors for homogeneous data distributions. Our approach constructs second- and third-order moments for variables with different data types. Parameter estimates are performed in the GMM framework by minimizing quadratic forms of the moment functions using a coordinate descent algorithm. Our algorithm scales as either O(p2) or O(p3), depending on whether third moments are used in the GMM procedure. In the setting where n is large and p is moderate, our algorithm shows clear computational advantages over alternative methods such as MCMC or EM. We also provide results regarding the asymptotic efficiency of our estimators.
In Section 2, we characterize the class of generalized Dirichlet latent variable models in MELD. In Section 3, we briefly review GMM and introduce the estimation procedure used in MELD. Asymptotic properties of the GMM estimator are also discussed. In Section 4, we evaluate the performance in a simulation study. We apply our method to three examples in Section 5 and conclude with a discussion in Section 6.
2. Generalized Dirichlet Latent Variable Models
In this section, we specify a generalized Dirichlet latent variable model. Let yi = (yi1, …, yip)T denote a vector of p measurements having heterogeneous data types (e.g., continuous, categorical, and counts) over subjects i = 1, …, n. We are interested in flexible models for joint distributions of the data. A simple way to achieve this is to assume the elements of yi are conditionally independent given a component index , with the densities for each variable in each component having a parametric form. This is inflexible due to the assumption that each subject i belongs to exactly one component. We instead allow subjects to partially belong to each of the k components. We assume that the elements of yi are conditionally independent given a latent Dirichlet vector , with ∆k−1 denoting the (k – 1) probability simplex. In particular, we let
| (1) |
where gj (ϕjh) is the density of the jth variable specific to component h. The elements of xi are interpretable as probability weights for subject i on each of k components; in this setting, pure subjects have weight vectors with all zeros except for a single one.
The corresponding density gj (ϕjh) is absolutely continuous with respect to a dominating measure . In this article, we focus on exponential family models for the density, but the procedures we propose will hold for general parametric forms. Our model in Equation (1) generalizes current mixed membership models to allow mixed data types. The parameters are mixture component parameters shared by all subjects. A full likelihood specification is completed by choosing a population distribution for the latent variable vector, xi ∼ P, with P a distribution on the simplex ∆k−1. We place a Dirichlet distribution on this latent variable vector xi ∼ Dir(α) with parameter α = (α1, …, αk)T.
Let mi = (mi1, … mip)T denote a membership vector for subject i, where indicates the component that feature j in subject i is generated from. We specify the following generative model:
| (2) |
When yi is a multivariate categorical vector and g j represents a multinomial distribution, Bhattacharya and Dunson (2012) proposed a simplex factor model. Their model reduces to latent Dirichlet allocation (Blei, Ng, and Jordan 2003) in the special case in which yij is the number of occurrences of word j in document i, and gj (ϕjh) is a Poisson distribution, with scalar ϕ jh indicating the rate of occurrence of word j in topic h. We also assume xi ~ Dir(α), but allow a general form for the component specific densities. This leads to the following likelihood after marginalizing out the latent variables:
Categorical data.
Let , for j = 1, …, p, so that we have multivariate categorical data that can be organized as a p-way contingency table. Let c = (c1, …, cp)T with and . Then is a probability tensor (Dunson and Xing 2009) satisfying and . In this case, is a multinomial distribution, with a probability vector. As shown by Bhattacharya and Dunson (2012), marginalizing out the mixture proportion xi leads to a Tucker decomposition of π
Mixed data types.
When yi includes variables of mixed data types, gj (ϕjh) belongs to the exponential family of distributions with mean parameter ϕjh. We are interested in estimating mean parameters with . For categorical variables yij, ϕjh is a probability vector indicating the distribution of the categories in component h; for noncategorical variables yij, ϕjh is a scalar term indicating the mean parameter of that feature for component h.
3. Generalized Methods of Moments Estimation
In this section, we provide a procedure for parameter estimation in generalized Dirichlet latent variable models based on a GMM framework. We first state the moment functions and then propose a two-stage procedure for parameter estimation.
3.1. Moment Functions Used in MELD
The idea behind method of moments (MM) algorithms is to derive a list of moment equations that have expected value zero at the true parameter values. Given observations , MM algorithms specify ℓ moment functions and define a moment vector such that at the true parameter . The moment vector is then used to specify an estimator by solving the following ℓ sample equations with p unknowns:
| (3) |
where is the sample estimate of . When f (yi, θ) is linear in θ (e.g., linear regression with instrumental variables) with independent moment functions and ℓ = p, then θ can be uniquely determined. When ℓ> p, the system in Equation (3) might be over-determined, in which case standard MM cannot be applied.
Generalized method of moments (GMM) minimizes the following quadratic form rather than solving a linear system:
| (4) |
where An is the asymptotically optimal positive semidefinite matrix:
In the over-determined setting, minimizing the quadratic form still well defined. Hansen’s two stage procedure (Hansen 1982) obtains an initial estimate of using a suboptimal weight matrix, such as the identity. This initial estimate is used to calculate An, which is then plugged into Equation (4) to obtain the final estimate (Hall 2005), which is consistent and asymptotically normal (Hansen 1982).
Applying GMM to all of the parameters in the generalized Dirichlet latent variable model is not feasible due to the dimensionality of the parameter space. Higher-order moment functions are complex and involve large numbers of unknowns. We will make use of a tensor decomposition and define moment functions that only depend on lower order moments. This extends recent moment tensor approaches to our proposed broad class of latent variable models (Arora, Ge, and Moitra 2012; Anandkumar, Hsu, and Kakade 2012a; Anandkumar et al. 2012b; Hsu and Kakade 2013; Anandkumar et al. 2014a, 2014b). One key innovation in our procedure is that we allow for lower order interaction moments across variables, where each variable can have a different distribution. The moment functions will have rank one decompositions of the component mean parameters over the heterogeneous variables. These heterogeneous low-order moment functions lead to fast and robust parameter estimation in MELD, as we show in the simulation and application sections.
We now define the moment functions. Recall that yij is the jth variable for subject i, which may include various data types, including continuous, categorical, or count data. Each categorical variable yij is encoded as a vector bij of length dj containing all zeros except for a one in position yij. If yij is a non-categorical variable, then bij = yij is a scalar value. In the following, we assume that there are k latent components and that the latent variable xi follows a Dirichlet distribution with parameters α = (α1, …, αk)T and . Given an observation bij and hyperparameter α, we define moment functions based on second and third moments:
| (5) |
| (6) |
We use ◦ to denote an outer product, and ×s indicates multiplication of a tensor with a matrix for mode s. The means μ j take the values for j = 1, …, p. The second moment function is a dj × dt matrix, and the third moment function is a dj × ds × dt tensor. We set dj = 1 when variable j is noncategorical. The matrix is a k × k matrix with along the diagonals and zero for all other entries. is a three-way tensor with the diagonal entry for h = 1, …, k and all the other entries are zero.
The following theorem states that, at the true parameter value, the expectations of the moment functions are zero. The proof can be found in Appendix A.
Theorem 1 (Moment conditions in MELD). The expectations of the second moment matrix and third moment tensor defined in Equations (5) and (6) are zero at true model parameter values
with the expectations taken with respect to yi.
3.2. Two Stage Optimal Estimation
To fix notation, we first restate Hansen’s two stage GMM procedure as:
estimate ;
given calculate Sn and set ;
compute as the final parameter estimate.
We specify two versions of MELD, which we denote as MELD-2 and MELD-3, depending on whether we use only the second moment matrix or both the second and third moment matrices for estimation.
We now define the second and third moment vectors. We denote as the second moment vector, which is constructed by stacking the second moment matrices in Equation (5) as follows:
| (7) |
As the moment matrix is symmetric in the sense that we only consider the case where j < t. Assuming d levels for each element of yi results in a moment vector of dimension p(p – 1)d2/2. The third moment vector is constructed by stacking the second moment matrices and third moment tensors in Equations (5) and (6)
| (8) |
The moment matrices follow the same order as in . The third moment tensors are ordered such that the index on the right runs faster than the index on the left in the subscript. The third moment tensor is symmetric with respect to any permutations of its indices, so we include only the moment tensors with j < s < t. The dimension of this moment vector is with the assumption of d levels for each element of yi.
Based on the two moment vector versions, we define the following two quadratic functions used for our GMM estimation:
| (9) |
| (10) |
where and are sample estimates of the expectations of the moment vectors
and and are two positive semidefinite matrices. We emphasize that when calculating , the μ j and μt in are replaced by their sample estimates and , instead of their parametric counterparts and . In doing this the second moment matrix becomes a function of Φ only through , which is a rank one summation of ϕjh and ϕth. This method facilitates development of a fast coordinate descent algorithm for parameter estimation. Similar methods are used for calculating .
In the first stage, we set the weight matrix to an identity matrix. Then, the quadratic functions in Equations (9) and (10) can be rewritten as follows:
where and are defined as
with μj, μs and μt replaced by , and respectively in and . indicates the Frobenius norm, the element-wise sum of squares.
We obtain a first-stage estimator of Φ by minimizing the quadratic forms using Newton-Raphson. Note that, after we substitute μ j by its sample estimate, only the last term of and involves unknown parameter Φ. For simplicity, we denote
The two quantities and can be computed directly from the samples. We optimize ϕjh with other mean parameters fixed. After calculating the gradient and Hessian of , the update rule simply becomes
| (11) |
where and is the hth diagonal entry of .
The update rule for ϕjh with can be calculated as
| (12) |
where and is the hth diagonal entry of . The derivations can be found in Appendix E. After updating ϕjh using the above equations, we retract ϕjh to its probability simplex when yij is a categorical variable.
We find the optimization procedure to be robust to parameter initialization. Our default initialization sets the parameters to the observed frequencies for categorical variables and to the empirical means for noncategorical variables. We also initialized the parameters of categorical variables to equal weights or a draw from the uniform distribution over the simplex with almost identical results. Although the optimization problem in Equations (9) and (10) is nonconvex, our algorithm tended to find a local minimum after a small number of iterations. Our convergence criteria corresponded to a difference less than 1 × 10−5 in the objective between two iterations.
MELD-2 minimizes the quadratic form Q(2)(Φ, A(2)), while MELD-3 minimizes Q(3)(Φ, A(3)). After an initial consistent estimator of Φ is found, we calculate the asymptotic covariance matrix of the moment vectors Sn and define a new weight matrix An = (Sn)−1 for second stage GMM estimation. The form of Sn can be derived analytically for both MELD-2 and MELD-3, and we provide the results in the supplement materials. The calculation of An requires the inversion of a full-rank dense matrix Sn, with computation scaling as O(p2d2 ) for MELD-2 and O(p3d3 ) for MELD-3. When including the off-diagonal terms in the weight matrix, the updating rules become intrinsically complicated. In practice, we only extract the diagonal elements of Sn and let An = 1/diag[(Sn)] in the second stage. This approximation has been used in previous GMM implementations (Jöreskog and Sörbom 1987). The gradient descent update equations can be found by slightly modifying Equations (11) and (12) to include weights.
Note that the moment functions do not solve the identifiability problems with respect to Φ. When α is a constant vector, any permutation τ of 1, …, k with for all j = 1, …, p satisfies the moment condition. This problem is inherited from the label switching problem in mixture models. A similar situation occurs when there are ties in α and the permutation is restricted to each tie. In real applications, we typically do not find ties, and our parameter estimates tend to be close to the true parameter values.
3.2.1. Algorithmic Guidance
In deciding which version of MELD to apply to a particular problem, the main considerations are how to balance robustness to misspecification, statistical efficiency, and computational time. MELD-3 will have better statistical efficiency (when the model is correctly specified), greater computational cost as a function of p as it scales as O(p3 ), and is less robust to model misspecification. In practice, this tradeoff is somewhat complex, as it is difficult to test to what extent the data comply with the modeling assumptions on the third moment. As a general rule of thumb, we recommend using MELD-2 when the additional computational cost of MELD-3 is prohibitive. If p is modest (as in the political-economic data below), and thus the additional computational burden is small, we would generally recommend MELD-3.
The performance of the one-stage versus two-stage estimator relies on whether the matrix An is estimated accurately. If the variance matrix is accurately estimated, the two-stage method will not perform worse than the one-stage method. However, we find that, when the sample size is small or the model is misspecified, the variance can be poorly estimated and the one-stage method can outperform the two stage method. In general, if the sample size is small in comparison to the number of variables, we suggest using the one-stage estimator.
3.3. Properties of Parameter Estimates
We use GMM asymptotic theory to show that parameter estimates in MELD are consistent. We assume the following regularity conditions on the moment vector and the parameter space Θ.
Assumption 1 (Regularity conditions (Assumptions 3.2, 3.9, and 3.10 (Hall 2005))).
(1) is continuous on Θ for all ; (2) and continuous for ; 3) Θ is compact and .
Remark 1. Conditions (1) and (2) are satisfied in MELD. Condition (3) is satisfied automatically for categorical variables, noting that is compact. For noncategorical variables, we can restrict support to a large compact domain such as closed intervals in without sacrificing practical performance.
With these conditions, we further assume that the weight matrix An converges to a positive definite matrix A in probability. We define the population analogs of the quadratic functions as
| (13) |
We have the following lemma showing the uniform convergence of .
Lemma 1 (Uniform convergence (Lemma 3.1 (Hall 2005))). Under the regularity conditions in Assumption 1
Theorem 2 (Consistency). Under the same conditions in Lemma 1, the estimator that minimizes Qn(Φ;An) in MELD-2 and MELD-3 converges to the true parameter Φ0 in probability.
The proof can be found in Appendix B.
The asymptotic normality of can also be established by assuming the following conditions on .
Assumption 2 (Conditions on (Assumptions 3.5, 3.12, and 3.13. (Hall 2005))).
(1) exists and is continuous on Θ for all ; (2) Φ0 is an interior point of Θ; (3) ; (4) G0(Φ) is continuous on some neighborhood of Φ0; (5) the sample estimate Gn(Φ) uniformly converges to G0(Φ).
Remark 2. We derive for both MELD-2 and MELD-3 in Appendix D. Conditions 1, 3, and 4 are satisfied in MELD. Condition 5 can be shown with continuousness of the derivative and the compactness of Θ.
Theorem 3 (Asymptotic normality). With Assumptions 1 and 2, we have
with
where and .
The proof can be found in Appendix C. The optimal estimator can be obtained so that the weight matrix An → A = (S)−1 (Hansen 1982).
3.4. Model Selection Using Goodness of Fit Tests
We use goodness-of-fit tests to choose the number of latent components k. With the optimal weight matrix An → (S)−1, Wald-, Lagrange multiplier- (score), and likelihood ratio-type tests can be constructed. We could construct a sequence of test statistics with increasing k to quantify the goodness-of-fit in MELD. However, this approach requires the calculation of the optimal weight matrix and large matrix inversions. Instead, we quantify goodness of fit using a fitness index (FI) (Bentler 1983). Practically, we show in an extensive simulation study that the FI has low error in estimating k in MELD.
The FI is defined based on the value of the objective function evaluated at the parameter estimate (Bentler 1983)
| (14) |
where en is the concatenation of with j < t for MELD-2 and the concatenation of both with j < t and with j < s < t for MELD-3. This FI may be constructed for any weight matrix An. It can be viewed as a normalized objective function; thus, the FI has values less than one. Larger values of the FI indicate a better fit.
4. Simulation Study
In this section, we evaluate the accuracy and run time of MELD-2 and MELD-3 in simulations with categorical and mixed data types. In the first stage, an identity weight matrix is used. In the second stage, we set An = 1/diag[(Sn)]. We denote MELD-2-j and MELD-3- j as the procedure using up to the jth stage of MELD-2 and MELD-3, respectively, for j = 1, 2.
4.1. Categorical Data
For a categorical data simulation, we considered a low-dimensional setting (p = 20) so that both second- and third-order moment functions may be efficiently calculated. A simulation with a moderate dimension (p = 100) is summarized in the supplementary materials.
We assumed each of the 20 categorical variables has d = 4 levels. We set the number of components to k = 3 and generated ϕjh from Dir(0.5, 0.5, 0.5, 0.5) with h = 1, …, k; α was set to (0.1, 0.1, 0.1)T. We drew n = {50, 100, 200, 500, 1000, 10000} samples from the model in Equation (1) with gj (ϕjh) a multinomial distribution. For each value of n, we generated ten independent datasets. We ran MELD for different values of k = {1, …, 5}. The FI criterion consistently chose the correct number of latent components k in first stage estimation (Table 1). For second-stage estimation, FI did not perform as well. The trajectories of the objective functions under different values of k are shown in Figure S1. Convergence of parameter estimates for the ten simulated datasets under n = 1000 and k = 3 is shown in Figure S2. MELD-2 converged in about 25 iterations and MELD-3 converged in about 10 iterations.
Table 1.
Goodness-of-fit tests using the fitness index (FI) in categorical simulation. Larger values of FI indicate better fit, with the maximum at one. Results shown are based on ten simulated datasets for each value of n. Standard deviations of FI are provided in parentheses. Bold type indicates the best FI across k.
| n | k | MELD-2–1 | MELD-2–2 | MELD-3–1 | MELD-3–2 |
|---|---|---|---|---|---|
| 50 | 1 | 0.824(0.020) | − 1.865(4.950) | 0.585(0.031) | − 27.987(40.128) |
| 2 | 0.908(0.011) | − 1.285(2.630) | 0.745(0.032) | − 27.240(54.415) | |
| 3 | 0.930(0.004) | 0.570(0.044) | 0.775(0.022) | − 27.713(42.607) | |
| 4 | 0.901(0.010) | 0.588(0.032) | 0.735(0.023) | 0.254(0.090) | |
| 5 | 0.795(0.033) | 0.686(0.021) | 0.653(0.024) | 0.305(0.091) | |
| 100 | 1 | 0.860(0.012) | − 0.521(2.142) | 0.651(0.021) | − 0.031(0.703) |
| 2 | 0.930(0.011) | 0.282(0.644) | 0.795(0.030) | − 4.457(8.924) | |
| 3 | 0.960(0.005) | 0.677(0.044) | 0.851(0.009) | − 4.868(11.889) | |
| 4 | 0.942(0.012) | 0.691(0.042) | 0.822(0.010) | 0.225(0.019) | |
| 5 | 0.863(0.046) | 0.782(0.022) | 0.768(0.044) | 0.232(0.080) | |
| 200 | 1 | 0.869(0.012) | 0.679(0.060) | 0.682(0.021) | 0.298(0.070) |
| 2 | 0.940(0.007) | 0.699(0.047) | 0.838(0.014) | 0.306(0.054) | |
| 3 | 0.980(0.001) | 0.761(0.061) | 0.919(0.004) | 0.278(0.049) | |
| 4 | 0.967(0.006) | 0.780(0.019) | 0.891(0.008) | 0.287(0.050) | |
| 5 | 0.911(0.017) | 0.824(0.008) | 0.864(0.015) | 0.286(0.042) | |
| 500 | 1 | 0.882(0.007) | 0.783(0.022) | 0.713(0.012) | 0.414(0.080) |
| 2 | 0.948(0.006) | 0.799(0.019) | 0.870(0.013) | 0.427(0.080) | |
| 3 | 0.992( <0.001) | 0.884(0.024) | 0.966(0.001) | 0.388(0.073) | |
| 4 | 0.983(0.004) | 0.874(0.026) | 0.938(0.005) | 0.365(0.065) | |
| 5 | 0.937(0.014) | 0.894(0.006) | 0.921(0.007) | 0.353(0.042) | |
| 1000 | 1 | 0.888(0.003) | 0.828(0.006) | 0.729(0.008) | 0.571(0.017) |
| 2 | 0.951(0.003) | 0.855(0.009) | 0.881(0.008) | 0.615(0.030) | |
| 3 | 0.996( <0.001) | 0.950(0.005) | 0.982(0.001) | 0.609(0.031) | |
| 4 | 0.989(0.002) | 0.938(0.004) | 0.961(0.003) | 0.579(0.030) | |
| 5 | 0.953(0.010) | 0.932(0.006) | 0.951(0.006) | 0.550(0.034) | |
| 10000 | 1 | 0.920(0.001) | 0.853(0.003) | 0.812(<0.001) | 0.582(0.009) |
| 2 | 0.980(<0.001) | 0.914(0.002) | 0.940(<0.001) | 0.725(0.005) | |
| 3 | 0.999( <0.001) | 0.992(0.001) | 0.998( <0.001) | 0.900(0.008) | |
| 4 | 0.990(0.003) | 0.977(0.004) | 0.981(<0.001) | 0.861(0.008) | |
| 5 | 0.885(0.005) | 10.929(0.002) | 0.981(<0.001) | 0.856(0.008) |
We compared MELD with the simplex factor model (SFM) (Bhattacharya and Dunson 2012) and latent Dirichlet allocation (LDA) (Blei, Ng, and Jordan 2003). For the LDA model, we used the topicmodels package in R (Hornik and Grün 2011) with both collapsed Gibbs sampling and variational EM. In our notation, yij is a word and yi is a document. Our generative model is more flexible than LDA in that each variable was allowed to have its own k component distributions with parameter ϕj1, …, ϕjk. For SFM and LDA, we ran 10,000 steps of MCMC with fixed k and a burn-in of 5000 iterations. We collected every 50th sample to thin the chain. For LDA variational EM, we used the package defaults. Dirichlet parameters were set to α = 0.1 and β = 0.5 for mixture proportions and topic distributions, respectively. Mean squared errors (MSEs) were calculated for each variable yij by recovering the associated membership variable mij from the estimated parameters, calculating the estimate of yij using the estimated Φj parameter for membership variable mij, and then computing the MSE of that estimated yij with the true value (see supplementary materials). MSEs for different values of k are shown in Figure 1, and the running times of different methods are reported in Table S1. The LDA implementation in topicmodels does not allow k = 1, so we did not consider this case. The results show that MELD-2–1 was the most accurate and the fastest in most cases. MELD-3–2 did not perform well when n was small (i.e., n = 50), but estimation accuracy improved for n larger. SFM had comparable MSEs when n was small. However, when n = 500, 1000, or 10,000, MELD outperformed SFM. The model we simulated from is more flexible than the standard LDA model, so it was not surprising that LDA did not perform as well as MELD and SFM.
Figure 1.

Comparison of mean squared error (MSE) of estimated parameters in categorical simulations. For SFM and LDA with Gibbs sampling, posterior means of parameters were calculated using 100 posterior draws on each of the ten simulated datasets. MSEs were calculated for each variable yij by recovering the associated membership variable mij from the estimated parameters, calculating the estimate of yij using the estimated Φj parameter for membership variable mij, and then computing the MSE of that estimated yij with the true value. Details of MSEs and their standard deviations are in Supplementary Table S2– S4.
To provide a comparison more favorable to LDA, we generated samples from the LDA model. This is equivalent to setting for h = 1, …, k. We set the same values for α and k, and we draw ϕjh from Dir(0.5, 0.5, 0.5, 0.5). The MSEs with different contamination levels (see below) are shown in Tables S5–S7. When there was no contamination and n was small (i.e., n = 50 or 100), LDA had the highest accuracy in parameter estimation. However, when the sample size was large or there was contamination, MELD outperformed LDA.
We also found that MELD using two-stage estimation had larger MSEs compared to MELD using one-stage estimation. This was especially true when the number of observations was small or the number of latent components was misspecified. FIs for two-stage estimation were less accurate. The loss in accuracy appeared to be a result of poorly estimating the diagonal weight matrix. We confirmed this hypothesis by calculating mean squared distances between true and estimated diagonal weight matrices for different n and k (Table S8; Figure S3).
We further evaluated performance in the presence of contamination. For each simulated dataset, we randomly replaced a proportion of observations (4% and 10%) with draws from a discrete uniform distribution. MSEs under different values of k are shown in Figure 1. With 4% contamination, MELD had the most accurate parameter estimation in almost all cases. MELD-2–1 performed the best, followed by MELD-3–1. When we increased contamination to 10%, MELD had the most accurate MSE in all cases. MELD-2–1 performed best, followed by MELD-2-MELD-2 consistently performed better than MELD-3, as expected given the robustness of using only low-order moments.
To evaluate the performance of MELD in terms of selecting the correct number of components, we varied the number of components k = {6, 7, 8, 9, 10} and generated 10,000 observations using the same simulation model. We found that MELD-3 was able to estimate the correct number of components for k ≤ 8, MELD-2–2 was accurate for k ≤ 7 and MELD-2–2 for k = 6 (see Tables S9– S11). For k ≥ 9, none of the variants of MELD were able to correctly estimate the number of components (see Tables S12 and S13).
4.2. Mixed Data Types
We considered a simulation setting mimicking applications in which DNA sequence variations influence a quantitative trait. An additional simulation combining categorical, Gaussian, and Poisson variables is summarized in the Supplementary Materials. We generated a sequence of nucleotides {A, C, G, T} at 50 genetic loci along with a continuous or integer-valued trait, resulting in p = 51 variables. We set k = 2 latent components and simulated n = 1000 individuals, with the first 500 from the first component and the last 500 from the second component. The continuous traits were drawn from and , while count traits were drawn from Poisson (5) and Poisson(10), respectively, for the two components. We chose eight loci J = {2, 4, 12, 14, 32, 34, 42, 44} to be associated with the trait. The multinomial parameters for each of the two components were randomly drawn from Dir(0.5, 0.5, 0.5, 0.5). The association between the trait and the eight loci is given by marginalizing the component variables. The distribution for the remaining 42 loci was fixed across individuals independently of the trait as multinomial with parameters generated from Dir(0.5, 0.5, 0.5, 0.5). Ten datasets were simulated from the model in Equation (1). To assess robustness, we added contamination (e.g., genotyping errors) by randomly replacing 4%, 10%, and 20% of the nucleotides with values uniformly generated from {A, C, G, T}.
We ran MELD-2–1 and chose the number of components k = {1, …, 5}. The fitness test FI chose the correct value of k on all ten datasets (Table S14). For each genomic locus, we calculated its marginal frequency according to the simulated data, and then computed the averaged Kullback–Leibler (KL) distance between the estimated component distributions and the marginal frequency as follows:
| (15) |
A smaller averaged KL distance suggests that the component distributions were closer to the marginal distribution, implying that the locus frequency was not differentiated across components. The set J corresponded exactly to the eight loci with largest averaged KL distance in the MELD results in most cases (Table 2).
Table 2.
Quantitative trait association simulation with 50 nucleotides and one response. Nucleotides not in J = {2, 4, 12, 14, 32, 34, 42, 44} are labeled by an underline and missing nucleotides are crossed out. Results shown are for one of the ten simulated datasets. The complete results can be found in Tables S15 and S16.
| Response | Contamination | Q(2)(Φ) 1st stage | Bayesian copula factor model |
|---|---|---|---|
| Gaussian | 0% | {2, 4, 12, 14, 32, 34, 42, 44} | {1, 2, , 12, 14, 32, 34, 42, 44} |
| 4% | {2, 4, 12, 14, 32, 34, 42, 44} | {2, , 5, 12, 14, 32, 35, , 42, 44} | |
| 10% | {2, 4, 12, 14, 32, 34, 42, 44} | {2, , 5, 12, 14, 25, 32, , 42, 44} | |
| 20% | {2, 4, 12, 14, 32, 34, 42, 44} | {1, 2, 4, , , 25, 32, 34, 42, 44} | |
| Poisson | 0% | {2, 4, 12, 14, 32, 34, 42, 44} | {, 3, 4, 12, , 32, 34, 37, 42, 44} |
| 4% | {2, 4, 12, 14, 32, 34, 42, 44} | {, 3, 4, 12, 14, 32, 34, 42, 44} | |
| 10% | {2, 4, 12, 14, 32, 34, 42, 44} | {, 3, 4, 12, 14, 32, 34, 42, 44} | |
| 20% | {2, 4, 7, 12, 14, 32, 34, , 44} | {, 3, 4, 12, 14, 30, 32, 34, , 44} |
We compared results from MELD with results from the Bayesian copula factor model (Murray et al. 2013), which estimates a correlation matrix C between variables. From this estimate, we computed partial correlations between the response variable (trait) and each genetic locus (Hoff 2007). We ran MCMC for the Bayesian copula factor model for 10,000 iterations with the correct value for k, a burn-in of 5000 iterations, and saving every 50th iteration. We selected genomic locations for which the 95% posterior intervals of the partial correlation did not include zero. Table 2 shows the eight loci having the largest absolute values of the posterior mean partial correlation. The Bayesian copula factor model had much higher error rates than MELD.
5. Applications of MELD
Promoter sequence analysis.
We applied MELD to gene transcription promoter data available in the UCI machine learning repository (Lichman 2013). The data include n = 106 nucleotide sequences {A, C, G, T} of length 57. Of those, 53 sequences are located in promoter regions, and 53 sequences are located in nonpromoter regions. The goal is binary classification: using nucleotide sequence alone, we would like to predict whether or not the sequence is in a promoter or a non-promoter region.
We included the promoter or nonpromoter status of the sequences as an additional binary variable, giving us p = 57 + 1 categorical variables. We applied MELD Q(2)(Φ) with first-stage estimation on the full data and also separately on the subsets of promoter and non-promoter sequences. We set k = {1, …, 8}. For k = 2, MELD converged in 2.13 seconds, compared with SFM, which took 41.6 seconds to perform 10,000 MCMC iterations. The FI test selected k = 2 for the full data, k = 2 for the promoter data, and k = 1 for the nonpromoter data (Table S17).
We set k = 2 in the following analysis. For each nucleotide position, we calculated the averaged KL distance between the estimated component distributions and its marginal distribution using Equation (15). An interpretation of the averaged KL distance is that it quantifies stratification of each nucleotide distribution across components: a larger value indicates greater stratification, suggesting that the nucleotide is important in defining and differentiating the components. For the full dataset, we observed approximately two peaks of the averaged KL distance, one around the 15th nucleotide and one around the 42nd (Figure S5). The first peak corresponds to the start of the biologically conserved region for promoter sequences (Harley and Reynolds 1987). For MELD applied only to promoter sequences, this peak was reduced, suggesting that, at approximately the 15th nucleotide, the components all included similarly well conserved distributions of this nucleotide. The estimated component membership variables also showed the importance of nucleotides around the 15th locus (Figures S6 and S7). For the peak around the 42nd nucleotide, this phenomenon was reversed: the increased averaged KL distance remained in promoter sequences but diminished in non-promoter sequences. One possible explanation is that this region was conserved uniformly in the non-promoter region, but differentially conserved in the promoter region. Applying MELD to the entire data and separately within each group of sequences (promoter, nonpromoter) provides insight into the sequence differences across the two groups.
Political-economic risk data.
In a second application, we applied MELD to political-economic risk data (Quinn 2004), which include five proxy variables of mixed types measured for 62 countries. The dataset has been analyzed using a mixed Gaussian and probit factor model (Quinn 2004) and a Bayesian copula factor model (Murray et al. 2013). The data are available in the MCMCpack package. There are three categorical variables and two real valued variables (Table 3).
Table 3.
Variables in the political-economic risk data
| Variable | Type | Explanation |
|---|---|---|
| ind.jud | binary | An indicator variable that measures the independence of the national judiciary. This variable is equal to one if the judiciary is judged to be independent and equal to zero otherwise. |
| blk.mkt | real | Black-market premium measurement. Original values are measured as the black-market exchange rate (local currency per dollar) divided by the official exchange rate minus one. Quinn (2004) transformed the original data to log scale. |
| lack.exp.risk | ordinal | Lack of appropriation risk measurement. Six levels with coding 0 < 1 < 2 < 3 < 4 < 5. |
| lack.corrup | ordinal | Lack of corruption measurement. Six levels with coding 0 < 1 < 2 < 3 < 4 < 5. |
| gdp.worker | real | Real gross domestic product (GDP) per worker in 1985 international prices. Recorded data are log transformed. |
We applied MELD with k = {1, …, 5} using both MELD-2–1 and MELD-3–1. The FI criterion selected k = 4 for MELD-2–1 and k = 3 for MELD-3–1 (Table S18). We chose results from MELD-3–1 with k = 3 for further analysis.
The estimated component parameters for the five variables showed distinct interpretations of the three components (Figure S8). We might interpret the three components as low-risk, intermediate-risk, and high-risk political–economic status, respectively. The first component had a high probability of independence of the national judiciary (ind.jud being one) and a low measurement of black-market premium. The first component also had a high probability of observing 4th and 5th levels in lack.exp.risk and 3rd, 4th, and 5th levels in lack.corrup. The mean of the GDP per worker was highest among the three components. The second component had a relatively high probability of being zero in ind.jud and a large mean value of blk.mkt. Both of lack of appropriation risk measurement and lack of corruption measurement put higher weights on lower category numbers (0, 1, and 2), indicating more risk and higher levels of corruption; the GDP per worker was still high. We might interpret this component as a society being relatively unstable while still having a good economic forecast, meaning that GDP per worker is high, possibly through the black market. The last component had the least judicial independence as quantified by the probability of ind.jud being zero. The black-market premium is also low, as is the lack of risk level and lack of corruption level. The GDP per worker is by far the lowest among the three components. We might interpret this component as societies being the most unstable with the greatest economic risk. We found that, although the three components had distinct stratification, each country was a mixture of the three components (Figure S9).
Human single nucleotide polymorphism data.
As a third application, we applied MELD to Human HapMap phase 3 (HM3) genotype data (Stranger et al. 2012). The HM3 data consist of 608 individuals from diverse ancestral populations. We focused on single nucleotide polymorphisms (SNPs) on chromosome 21 for which the frequency of the less frequent variant is greater than 5%. SNPs were represented by the number of copies of the less frequent variant at each genetic locus {1,2,3}. After selecting every 10th locus to avoid including well-correlated SNPs, we were left with 1672 SNPs for this application.
We applied MELD-2–1 to estimate genotype distributions in different populations. We considered different numbers of ancestral populations, including k = {1, 2, 3, 4, 5}. On this analysis, FI chose k = 2 (Table S19). The bands in the estimated membership variables shown in Figure S10 suggest a clear population structure among the 608 individuals. We calculated average KL distance for the 1,672 SNPs. The five SNPs with highest average KL distance are shown in Table S20, along with their genotypes in the samples. As with the previous applications, these SNPs define and differentiate the two inferred ancestral populations.
6. Discussion
In this article, we developed a new class of latent variable models with Dirichlet-distributed latent variables for mixed data types. These generalized latent Dirichlet variable models extend previous mixed membership models, such as LDA (Blei, Ng, and Jordan 2003) and simplex factor models (Bhattacharya and Dunson 2012), to allow mixed data types. For this class of models, we developed a fast parameter estimation procedure using the generalized method of moments. Our procedure extends recent moment tensor methods (Anandkumar et al. 2014b) to models with mixed data types and does not require instantiation of latent variables. We derive population moment conditions after marginalizing out the Dirichlet latent variables. We compared the two stages of our estimation procedure in different simulation settings and found that, for small-sample sizes and misspecified models, one-stage estimation had better performance. One possibility in terms of improving two-stage estimation is to apply a penalization procedure to better estimate the weight matrix.
Our results show that MELD is a promising alternative to MCMC or EM methods for producing fast and robust parameter estimates. An online method to update moment statistics when new samples arrive would allow its extension to streaming data. One limitation of our method is that the Newton-Raphson method is of order O(p2 ) using second moment functions and order O(p3 ) using third moment functions. One possible approach to speed up computation when p is large is to use stochastic gradient methods to calculate an approximate gradient in each step. Another possible extension is to adapt our model to include ordinal variables; for example, by thresholding latent variables.
Supplementary Material
Appendix
A. Proof of Theorem 1
Proof. We start with the case where yij is categorical with dj levels. The probability vector defines the mixture proportion of individual i. We assume xi ∼ Dir(α1, …, αk). Define normalizer and α = (α1, …, αk)T.
We use the standard basis for encoding. We encode yij = cj as , a binary (0/1) vector with the cjth coordinate being 1 and others being 0. Similarly, we encode the membership variable mij as a k dimensional binary vector . Consider the first moment of bij,
where Φj = (ϕj1, … ϕjk).
We consider second-order moment conditions. There are four types of second-order moments: same-variable, same-subject (type SS), same-variable, cross-subject (type SC), cross-variable, same-subject (type CS), and cross-variable, cross-subject (type CC). Of the four types, only the CS type is required to prove the theorem. The CS type second moment for bij and bit ( j ≠ t) can be written as
For a Dirichlet distributed variable,
Then, we have
| (A.1) |
We next consider third-order moment conditions. There are eight different types of third-order moments for bij. Only the moments with different variables for the same subject are required to prove the theorem.
We consider the third cross moment for bij, bit, and bis with j ≠ t ≠ s for the same subject. First, we calculate ,
Here, eh is a standard basis vector of length k with the hth coordinate equal to one. The third order moment tensor of bij ◦ bis ◦ bit can be derived as
| (A.2) |
The theorem follows from Equations (A.1) and (A.2). For noncategorical data, we let bij = yij and ϕjh is a scalar mean parameter for yij. Equations (A.1) and (A.2) still hold. □
B. Proof of Theorem 2
Proof. For notational simplicity, we suppress An in Qn(Φ; An) and A in Q0(Φ; A). Lemma 1 implies that and for . This result also implies that
| (B.1) |
| (B.2) |
On the other hand, minimizes Qn(Φ), therefore
| (B.3) |
Equations (B.1) and (B.3) imply that
Together with Equation (B.2), we get
Therefore,
| (B.4) |
follows with Q0(Φ0) = 0 and . Next, we choose a neighborhood N, which contains Φ0 in Θ. Due to the compactness of Θ, the neighborhood NC is also compact. The continuity of Q0(Φ) implies the existence of and that it is positive. Let , then we get
| (B.5) |
Therefore, , which suggests . Shrinking the neighborhood size of N, we get
□
C. Proof of Theorem 3
Proof. We approximate using a first-order Taylor expansion
Ignoring the high-order term, we left multiply both sides by . Then, we get
The fact that estimator minimizes Qn(Φ,An ) implies that the left-hand side equals zero. Therefore,
The theorem follows with and Assumptions 1 and 2. □
D. Derivatives of Moment Functions
D.1. Second Moment Matrix
The second moment matrix may be written as . The derivatives of with respect to Φj and Φt can be written as
where ⊗ indicates a Kronecker product and T is a dtk × dtk 0/1matrix that satisfies
Therefore, is a block matrix with blocks in columns corresponding to vec(Φt ) and rows corresponding to .
D.2. Third Moment Tensor
We next consider the third moment tensor. Write as . Then, only the second term involves Φ.
The derivatives of with respect to Φj can be written as
where subscript (1) indicates model-1 unfolding of a three-way tensor.
The derivatives of with respect to Φs and Φt can be calculated accordingly by introducing binary transformation matrices T (2)(1) and T (3)(1), both with size djdsdt × djdsdt, that satisfy
Then,
Conditions (1), (3), and (4) in Assumption 2 in the main text follow after we calculate the derivatives of the moment functions.
E. Derivation of Newton-Raphson Update
We denote
With the identity matrix, the two objective functions can be written as
Here, we suppress the weight matrix in the objective functions. We first consider . The terms that involve ϕjh are
where and is the hth diagonal element of . Here, we use the fact that . By letting
the gradient and Hessian can be written as
The update rule in (11) can be derived accordingly.
Then, we consider . The terms that involve ϕjh are
where and . Again we use the symmetric property of and . By organizing the terms, we get
By letting
the gradient and Hessian can be written as
The update rule in (12) follows directly.
Footnotes
Color versions of one or more of the figures in the article can be found online at www.tandfonline.com/r/JASA.
Supplementary materials for this article are available online. Please go to www.tandfonline.com/r/JASA.
Supplementary Materials
We provide supplementary materials that include (a) derivations of asymptotic optimal weight matrices; (b) calculation of MSE and computational complexity analysis; (c) two additional simulations with categorical variables and mixed data types; (d) Supplementary Tables S1–S25; (e) Supplementary Figures S1–S10. Python code implementing MELD and code used to generate results in the article are available at https://github.com/judyboon/MELD.
References
- Anandkumar A, Ge R, Hsu D, and Kakade SM (2014a), “A Tensor Approach to Learning Mixed Membership Community Models,” The Journal of Machine Learning Research, 15, 2239–2312. [Google Scholar]
- Anandkumar A, Ge R, Hsu D, Kakade SM, and Telgarsky M (2014b), “Tensor Decompositions for Learning Latent Variable Models,” The Journal of Machine Learning Research, 15, 2773–2832. [Google Scholar]
- Anandkumar A, Hsu D, and Kakade SM (2012a), “A Method of Moments for Mixture Models and Hidden Markov Models,” JMLR W&CP 23: COLT [Google Scholar]
- Anandkumar A, Liu Y, Hsu DJ, Foster DP, and Kakade SM (2012b), “A Spectral Algorithm for Latent Dirichlet Allocation,” in Advances in Neural Information Processing Systems 25, pp. 917–925. [Google Scholar]
- Anderson JC, and Gerbing DW (1988), “Structural Equation Modeling in Practice: A Review and Recommended Two-Step Approach,” Psychological Bulletin, 103, 411. [Google Scholar]
- Arora S, Ge R, and Moitra A (2012), “Learning Topic Models—Going Beyond SVD,” in Fifty-Third IEEE Annual Symposium on Foundations of Computer Science, pp. 1–10. [Google Scholar]
- Bentler PM (1983), “Some Contributions to Efficient Statistics in Structural Models: Specification and Estimation of Moment Structures,” Psychometrika, 48, 493–517. [Google Scholar]
- Bhattacharya A, and Dunson DB (2012), “Simplex Factor Models for Multivariate Unordered Categorical Data,” Journal of the American Statistical Association, 107, 362–377. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Blei DM, Ng AY, and Jordan MI (2003), “Latent Dirichlet Allocation,” The Journal of Machine Learning Research, 3, 993–1022. [Google Scholar]
- Bollen KA, Kolenikov S, and Bauldry S (2014), “Model-Implied Instrumental Variable-Generalized Method of Moments (MIIV-GMM) Estimators for Latent Variable Models,” Psychometrika, 79, 20–50. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Colombo N, and Vlassis N (2015), “FastMotif: Spectral Sequence Motif Discovery,” Bioinformatics, 31, 2623–2631. [DOI] [PubMed] [Google Scholar]
- Dunson DB (2003), “Dynamic Latent Trait Models for Multidimensional Longitudinal Data,” Journal of the American Statistical Association, 98, 555–563. [Google Scholar]
- Dunson DB, and Xing C (2009), “Nonparametric Bayes Modeling of Multivariate Categorical Data,” Journal of the American Statistical Association, 104, 1042–1051. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gallant AR, Giacomini R, and Ragusa G (2013), Generalized Method of Moments With Latent Variables, Centre for Economic Policy Research. [Google Scholar]
- Hall AR (2005), Generalized Method of Moments, New York: Oxford University Press. [Google Scholar]
- Hansen LP (1982), “Large Sample Properties of Generalized Method of Moments Estimators,” Econometrica: Journal of the Econometric Society, 50, 1029. [Google Scholar]
- Harley CB, and Reynolds RP (1987), “Analysis of E. coli Promoter Sequences,” Nucleic Acids Research, 15, 2343–2361. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hoff PD (2007), “Extending the Rank Likelihood for Semiparametric Copula Estimation,” The Annals of Applied Statistics, 1, 265–283. [Google Scholar]
- Hornik K, and Grün B (2011), “topicmodels: An R Package for Fitting Topic Models,” Journal of Statistical Software, 40, 1–30. [Google Scholar]
- Hsu D, and Kakade SM (2013), “Learning Mixtures of Spherical Gaussians: Moment Methods and Spectral Decompositions,” in Proceedings of the 4th conference on Innovations in Theoretical Computer Science, ACM, pp. 11–20. [Google Scholar]
- Jöreskog KG, and Sörbom D (1987), “New Developments in LISREL,” Paper presented at the National Symposium on Methodological Issues in Causal Modeling, University of Alabama, Tuscaloosa. [Google Scholar]
- Kiers HA (2000), “Towards a Standardized Notation and Terminology in Multiway Analysis,” Journal of Chemometrics, 14, 105–122. [Google Scholar]
- Lichman M (2013), “UCI Machine Learning Repository,” available at https://archive.ics.uci.edu/ml/index.php. [Google Scholar]
- Moustaki I, and Knott M (2000), “Generalized Latent Trait Models,” Psychometrika, 65, 391–411. [Google Scholar]
- Murray JS, Dunson DB, Carin L, and Lucas JE (2013), “Bayesian Gaussian Copula Factor Models for Mixed Data,” Journal of the American Statistical Association, 108, 656–665. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Muthén B (1984), “A General Structural Equation Model With Dichotomous, Ordered Categorical, and Continuous Latent Variable Indicators,” Psychometrika, 49, 115–132. [Google Scholar]
- Pritchard JK, Stephens M, and Donnelly P (2000a), “Inference of Population Structure Using Multilocus Genotype Data,” Genetics, 155, 945–959. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pritchard JK, Stephens M, Rosenberg NA, and Donnelly P (2000b), “Association Mapping in Structured Populations,” The American Journal of Human Genetics, 67, 170–181. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Quinn KM (2004), “Bayesian Factor Analysis for Mixed Ordinal and Continuous Responses,” Political Analysis, 12, 338–353. [Google Scholar]
- Sammel MD, Ryan LM, and Legler JM (1997), “Latent Variable Models for Mixed Discrete and Continuous Outcomes,” Journal of the Royal Statistical Society, Series B, 59, 667–678. [Google Scholar]
- Stranger BE, Montgomery SB, Dimas AS, Parts L, Stegle O, Ingle CE, Sekowska M, Smith GD, Evans D, Gutierrez-Arcelus M, Price A, Raj T, Nisbett J, Nica AC, Beazley C, Durbin R, Deloukas P, and Dermitzakis ET (2012), “Patterns of Cis-Regulatory Variation in Diverse Human Populations,” PLoS Genetics, 8, e1002639. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tung HF, and Smola AJ (2014), “Spectral Methods for Indian Buffet Process Inference,” in Advances in Neural Information Processing Systems 27, pp. 1484–1492. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
