Significance
We derive a computational framework that allows highly scalable identification of reduced Bayesian and Markov relation models, their uncertainty quantification, and inclusion of a priori physical information. It does not rely on the prior knowledge or a necessity of estimation of the full matrix of system’s relations in any step. Application to a molecular dynamics (MD) example showed that this methodology opens possibilities for a robust construction of reduced Markov state models directly from the MD data—providing ways of bridging the gap toward longer simulation times and larger systems in computational MD.
Keywords: dimension reduction, Markov state models, clustering, computer-aided diagnostics, Bayesian modeling
Abstract
The applicability of many computational approaches is dwelling on the identification of reduced models defined on a small set of collective variables (colvars). A methodology for scalable probability-preserving identification of reduced models and colvars directly from the data is derived—not relying on the availability of the full relation matrices at any stage of the resulting algorithm, allowing for a robust quantification of reduced model uncertainty and allowing us to impose a priori available physical information. We show two applications of the methodology: (i) to obtain a reduced dynamical model for a polypeptide dynamics in water and (ii) to identify diagnostic rules from a standard breast cancer dataset. For the first example, we show that the obtained reduced dynamical model can reproduce the full statistics of spatial molecular configurations—opening possibilities for a robust dimension and model reduction in molecular dynamics. For the breast cancer data, this methodology identifies a very simple diagnostics rule—free of any tuning parameters and exhibiting the same performance quality as the state of the art machine-learning applications with multiple tuning parameters reported for this problem.
Model reduction and identification of a most appropriate (small) set of collective variables are essential prerequisites for many computational methods and modeling techniques in a number of applied disciplines ranging from biophysics and bioinformatics to computational medicine and image processing. A variety of methods for the identification of collective variables can be roughly subdivided into two major groups: (i) methods that are based on some user-defined agglomeration of the original degrees of freedom into collective variables (e.g., based on the physical intuition) (1) and (ii) methods that produce/derive these agglomerations of original system’s variables based on a reduced approximation of some system-specific relation matrices. These matrices can be defined, for example, as covariance or kernel covariance matrices (2, 3), partial autocorrelation matrices of autoregressive processes (4), Gaussian distance kernel matrices (5, 6), Laplacian matrices [as in the case of spectral clustering methods for graphs (7, 8)], adjacency matrices [in community identification methods for networks (9)], or Markov transition matrices [as in spectral reduction methods for Markov processes (10, 11)]. In most of these reduction methods, the relation matrices are assumed a priori available—and this assumption is true, for example, in social sciences, network science, and many areas of biology. However, in many particular applications (e.g., in biophysics and many medical applications; examples 1 and 2 below), one first needs to estimate these matrices from available data. For systems with a large number of dimensions (for continuous data) or categories (for categorical data) and short available statistics, these matrix estimates will be subject to uncertainty and may lead to biasedness of the derived colvars. Some other reduction approaches that allow for computing the reduced representation from the data directly [e.g., the Probabilistic Latent Semantic Analysis (PLSA; used in mathematical linguistics and information retrieval for analysis and reduction of texts and documents)] (12–14) impose strong assumptions on the data and exhibit issues related to the computational cost scaling (Fig. 1 and SI Appendix, section S5 have detailed discussion), making them practically not applicable to nonsparse data in such areas as, for example, the model and data reduction in biophysics and bioinformatics. Another problem arises when trying to identify the colvars for dynamical systems while simultaneously trying to preserve some essential conservation properties (e.g., conservation of energy or probability) in the reduced representation. For example, deploying spectral methods based on Euclidean eigenvector projections [such as principal component analysis (PCA) and spectral clustering methods] to reduction of probability measures would not guarantee that the components of the projected/reduced representation will also add up to one and all be bigger than or equal to zero (i.e., the resulting reduced models may not be probability preserving).
In this paper, we present an algorithmic framework that is scalable for realistic dynamical systems and is designed for the inference and analytically computable uncertainty quantification of reduced probability-preserving Bayesian relation models directly from the data.
Methodology
Below, we will give a brief description of the methodology—detailed derivation can be found in SI Appendix, section S1. Our aim is to come out with a reduction method intending to preserve causality relations—measured in terms of the matrix of conditional probabilities between two categorical processes and . Process will serve as a reference process, meaning that it will not change when process is reduced. The terms “categorical process” and “categorical variables” mean that—in every particular case (e.g., at any given time or for any given instance in the dataset)— is taking one and only one of the possible values from categories and is taking values from one and only one of the categories . For example, in biomolecular dynamics simulations of polypeptides with amino acid residues, every peptide residue at any time can be assigned to one and only one of three Ramachandran states dependent on its current combination of torsion angle values and (SI Appendix, Fig. S1). Also, every global configuration/conformation X of the entire polypeptide molecule at any time can then be assigned to one of the categories —where every particular is defined by a vector of Ramachandran state combinations [e.g., is a category when junction 1 is being in state , junction 2 is being in state , and so on]. Efficient approaches based on the Markov state modeling (MSM) framework have been recently introduced, allowing for automated transformation of continuous-valued processes [e.g., molecular dynamics (MD) coordinates time series] into categorical time series (15, 16). Because the system cannot be in two different categories simultaneously, these categories are disjointed, and a relation between the probability for to attain a category in its instance/realization and the probabilities for can be formulated exactly via the conditional probabilities and the law of the total probability (17). Defining the column vectors of probabilities , and , we can write the exact relation between the variables and in a matrix vector form:
[1] |
where matrix elements are conditional probabilities. If known, they can be used as indicators for existence of causality relations between the variables and in the randomized studies: if for all and , the processes are then independent—meaning that information about the variable provides no additional advantage in computing the probability of the outcomes of . If for some , consequently, there exists some relation between and (18). To be able to interpret these conditional probabilities as a measure of the true causality relations in practical studies when are estimated from the available observations of and , one needs to guarantee that the data are appropriately randomized.
In a particular case, where , with being the time index and (where is a time step), the above formulation (Eq. 1) is equivalent to a so-called master equation of a Markov process [and thereby, is a particular time-discrete case of the well-known time-continuous Fokker–Planck equation (17)]. The matrix in this case will be a transpose of the Markov transition operator (19). If matrix is known, it provides full information about the relations between processes and —and can be used to predict if is available.
In many applications, the relation matrix is not known and needs to be first estimated from the available observational data and [e.g., by means of the maximum log-likelihood approach that allows us to provide the analytical estimates of the most likely parameter values and their uncertainties (lemma 1 in SI Appendix)]. However, in realistic applications (e.g., in the MD example below), the number of categories can grow exponentially with the physical dimension of the problem (“curse of dimension”)—leading to the exponential growth of overall uncertainty for the estimates when the available statistics size and a number of -categories are fixed (lemma 2 in SI Appendix). This problem also means that the uncertainty of all additional physical observables obtained from (e.g., the uncertainty of eigenvalues, eigenvectors, metastable sets, etc.) will be growing with the growing , making practical deployment of Eq. 1 problematic for realistic systems with a “large” and “small” . Therefore, if we want to reduce the dimensionality —for example, through identification of a small number of collective categorical variables that agglomerate the original categories of process into K groups/boxes—then this methodology should not rely on a direct estimation of the full Bayesian causality matrix in these situations.
To circumvent this problem, one can try to identify a latent reduced categorical process (being a reduced representation of the full categorical process ) that is defined on a reduced statistically disjoint complete set of (the yet unknown) categories with . Deploying a law of a total probability, we can establish the Bayesian relations between and on one side (by means of the conditional probabilities ) and between and on the other side (by means of the conditional probabilities ). Then, it is straightforward to validate (a detailed derivation is in SI Appendix, section S2) that an optimal probability-preserving reduced approximation of the full relation model (Eq. 1) for colvars takes a form
[2] |
where and . For every particular combination of and , defines a probability for the colvar to be in a reduced collective categorical variable when the observed original process is in a category , and therefore, it can be understood as a discrete analog of the continuous projection and reduction operators deployed in methods like PCA; is a reduced version of the matrix from the full relation model (Eq. 1). Please note that, being basically a reformulation of the exact law of the total probability, reduced model (Eq. 2) is exact in the Bayesian sense, and no additional approximations have been involved.
A similar approach to latent variable dependency modeling is used in the PLSA (13, 14) (that is, used in mathematical linguistics and information retrieval for identification of latent dependency structures in texts and documents). Deploying the definition of a conditional probability, PLSA allows one to parameterize a joint probability distribution with the help of the latent process as . To estimate the parameters, one deploys an Expectation Maximization algorithm having the computational iteration complexity of and requiring memory in a general nonsparse situation (i.e., when the underlying matrix is not assumed to be sparse a priori). However, as shown in SI Appendix, section S5, this problem requires imposing additional strong independence and stationarity assumptions on the latent variable . Moreover, as shown in Fig. 1A, the total average number of Expectation Maximization iterations for this problem grows rapidly with problem dimensions and —resulting in the overall algorithm complexity that grows polynomially in and (Fig. 1B). Applying standard statistical methods of polynomial regression fitting and discrimination (20, 21), one obtains that the statistically optimal fit of the red surface (corresponding to the PLSA) from Fig. 1B is given by a polynomial of the third degree in and . Extrapolation to the typical physical problem sizes (e.g., , ) that, for example, emerge in biophysical applications like the protein molecules indicates that such an inference procedure based on the Expectation Maximization algorithm and PLSA would require approximately 1,450 years of computations on a single laptop personal computer. Detailed methodological description of the PLSA methodology and its relation to the reduced Bayesian model reduction methods is provided in SI Appendix, section S5.
In the following section, we will suggest several computational procedures for the scalable inference of reduced Bayesian relation model parameters (Eq. 2) directly from the observed data and . The optimal parameter estimates and that maximize the observation probability (called likelihood) of the given data in Eq. 2 can be obtained by solving the following log-likelihood maximization problem subject to equality and inequality constraints:
[3] |
[4] |
[5] |
where (with being an indicator function). It is straightforward to observe that, for any fixed , the original exact log-likelihood maximization problem (Eqs. 3–5) can be decomposed into optimization problems for the columns of —and each of the column problems with optimization arguments is concave and can be solved independently from the other column problems. This observation can help in designing a convergent algorithm requiring much less memory than the Expectation Maximization [ instead of for Expectation Maximization] and with computational iteration complexity of . It can be used for identification of the reduced Bayesian relation model parameters in the situations when and are relatively small and is large (e.g., as in the medical example 2 below). Detailed derivation of this algorithm is given in SI Appendix, section S4. However, when or is large (as in a case of the MSM inference in MD, where ), this scaling would not allow us to apply this method to large realistic systems.
It turns out that substituting the function with its lower-bound approximation (which directly results from applying the Jensen’s inequality to Eq. 3) allows for providing a computational method that can solve this approximate model reduction problem with a much better scaling and allows analytically computable uncertainty estimates for the obtained reduced models.
Properties of this approximate model reduction procedure are summarized in the following theorem.
Theorem.
Given the two sets of categorical data and (where for any and ), the approximate maximum log-likelihood parameter estimates for and in the reduced model (Eq. 2) can be obtained via a maximization of the lower bound of the above log-likelihood function from Eq. 3:
[6] |
subject to the constraints (Eqs. 4 and 5). Solutions of this problem exist and are characterized by the discrete/deterministic optimal matrices that have only elements zero and one. Solutions of Eqs. 4–6 can be found in a linear time by means of the monotonically convergent Direct Bayesian Model Reduction (DBMR) Algorithm shown below, with a computational complexity of a single-iteration scaling as and requiring no more than of memory. Asymptotic posterior uncertainty of the obtained parameters (characterized in terms of the posterior parameter variance) can be computed analytically as . The least biased estimate of the ratio for the expectations of posterior parameter variances from the resulting full and reduced models equals
[7] |
DBMR Algorithm.
Choose a random (e.g., from the least biased uniform prior), set I = 0.
Set to 1 if and else to 0 for all j and k.
-
Do until becomes less than a tolerance threshold.
Step 1: set for all .
Step 2: set if and else for all .
A proof is provided in SI Appendix, section S2.
As can be seen from Fig. 1A, the average number of DBMR iterations (green surface in Fig. 1A) (computed from a large ensemble of randomly generated Bayesian model reduction problems for ) does not change with the dimensions and . It implies that also the overall computational complexity of the DBMR is scaling as . DBMR estimation of the reduced MSM for a medium-sized protein MD with and takes 33 min (as mentioned above, the extrapolated estimate of the Expectation Maximization computational time was 1,450 years under the same optimization and hardware/software settings).
Fig. 1C represents the average relative log-likelihood differences between the results of exact iterative log-likelihood optimization of Eqs. 3–5 and the DBMR results (obtained under the same conditions). It reveals that the empirical average relative log-likelihood differences between the exact and the DBMR-approximated results converge to zero exponentially in . This property implies that, for realistic high-dimensional applications, the log-likelihood difference between the reduced models obtained with the DBRM algorithm and those obtained with the optimization of the exact log-likelihood can be expected to become negligible—meaning that the reduced models obtained with the DBRM algorithm will have essentially the same posterior probability for explaining the observed full data as the exact reduced models.
The main feature of the two algorithms presented above is that they allow for obtaining the reduced model (Eq. 2) directly from the available observational data and —completely omitting a need for computation/estimation of the full relation matrix in Eq. 1. The only tunable parameter in both of the algorithmic procedures introduced above (in the direct sequential optimization of the exact log-likelihood (Eqs. 4–6) and the DBMR algorithm) is the reduced process dimension . The optimal integer value of can be obtained by performing the algorithms with different numbers of (i.e., ) and then selecting the best reduced model (Eq. 2) according to one of the standard model selection criteria [e.g., cross-validation criterion, information criteria like Akaike Information Criterion (AIC) and Bayesian Information Criterion (BIC), or approaches like L curve] (22, 23). To select an optimal for the examples below, we have used the standard L-curve method (23) that identifies the optimal as the edge point of the curve that describes a dependence of the optimal value of the maximized function (Eq. 3 or 6 in our case) from (a practical example is in SI Appendix, Fig. S2).
When dealing with real-life applications, it is also important to have an option for adjusting a set of collective variables according to a physical intuition or some prior knowledge (1). For example, one could have some prior physical information that certain dimensions of the original problem have a higher relevance for the dynamics than some other physically less relevant dimensions. In SI Appendix, section S3, we present a computationally scalable way [with computational iteration complexity of ] to impose such a priori information—cast into a form of the weighted graph—on the DBMR algorithm. The resulting DBMR graph algorithm is presented in SI Appendix, section S4, and a practical application of this information-imposing clustering method to reduced Bayesian model inference is given in the breast cancer diagnostics example 2 below.
A MATLAB library of algorithms implementing the methods introduced in this manuscript—as well as different variants of the constrained Nonnegative Matrix Factorization (12, 24) and PLSA methods (13, 14)—can be found in SI Appendix and is available as open access via a general public license from https://github.com/SusanneGerber.
Results
Example 1: Reduced Model of the 10-Alanine Dynamics in Water.
First, we consider a colvar identification for a polypeptide molecule [deca-alanine (10-ALA)] from results of the MD simulation. This dataset represents an output of the 0.5-s simulation (with a 2-fs time step) of a 10-ALA polypeptide in explicit water at the room temperature performed with the Amber99sb-ildn force field (25). These MD data were produced and provided by Frank Noe and Antonia Mey, Free University (FU) Berlin, Berlin. For additional analysis, the values of torsion angles and () inside of the molecular backbone (i.e., ignoring the two end groups and the angles) are grouped into the Ramachandran states 1–3 for every (SI Appendix, Fig. S1, Left) with a time step resolution of 100 ps, resulting in eight categorical Ramachandran time series with 5,000 time points each. Based on these eight local junction time series, we create a series of global molecular states (), where every particular combination of eight Ramachandran states is assigned to a particular category; in our case, it is a categorical series with of such eight-component combinations with time instances. As a 531D variable to be reduced, we use this set of global states; as reference variables , we choose the individual Ramachandran series of junctions (i.e., with each) at time . Thereby, we are casting the reduction problem to a setting of discrete Markov processes in time.
We start with setting and comparing the practical performance of algorithms introduced in this paper with the PLSA method (13, 14). Results of this comparison are summarized in SI Appendix, Fig. S5. As can be seen from SI Appendix, Fig. S5, methods based on optimization of Eqs. 3 and 6 provide colvars that are better in terms of the log-likelihood measure as well as in terms of the information theoretical measures, like the robust AIC and BIC (22). AIC and BIC take into account the model quality and penalize model uncertainty—for the same quality (log-likelihood), these measures would provide smaller values for the models that are less uncertain (22).
Second, we do the identification of reduced models (Eq. 2) for each of the peptide junctions (). Values of the resulting optimal solutions for reduced log-likelihoods () as functions of are shown in SI Appendix, Fig. S2. These results reveal that the reduced log-likelihood does not exhibit any nonnegligible increase for all when the number of colvars is becoming larger than three to seven, meaning that the maximal number of the nonredundant colvars is not greater than seven for this system. Next, we inspect the identified colvars for all of the . As can be seen from SI Appendix, Fig. S3 (as an example, representing a case of being the Ramachandran time series of the junction 4 for ), the three identified colvars almost perfectly—to 97%—coincide with the discretization that is solely based on this junction and disregard all other junctions in the peptide chain. In only 3% of the cases, the nonlocal information about the Ramachandran states of the peptide residues from other junctions is important. Therefore, relations in terms of temporal causality between the peptides MD dynamics can be almost (in 97% of the cases) described by a sequence of spatially independent Markov processes in each of the peptide junctions—for example, collected together in a form of the Ising model (26). To verify the validity of the obtained colvars as well as test the performance of the resulting reduced model (Eq. 2), we use these colvars to produce a long Monte Carlo time series of the reduced molecular simulation (Eq. 2) and compute statistics of the geometrical configurations for the entire molecule. As shown in Fig. 2, reduced dynamics based on just a few colvars can reliably represent the overall spatial statistics of molecular configurations in 3D—obtained from the full MD trajectory. These 3% of nonlocal dependence cases identified in SI Appendix, Fig. S3 seem to be crucial: without them, the corresponding box plot of the reduced Monte Carlo run is not capable of reproducing the true statistics of geometrical configurations from MD (SI Appendix, Fig. S4).
Example 2: Reduced Model for the Breast Cancer Diagnostics Based on the Standard BI-RADS Data.
For the second example, we consider analysis and reduction of the standard Breast Imaging Reporting and Data System (BI-RADS) dataset for breast cancer diagnostics—available as an open access data file at the University of California Irvine (UCI) Machine-Learning Repository: mlr.cs.umass.edu/ml/datasets/Mammographic+Mass. This dataset contains information about 403 healthy (benign) and 427 malignant breast cancer patients. For each of the patient entries, the age and three categorical variables obtained from the mammography images are provided together with the basic result (“cancer”/“no cancer”) obtained from the invasive analysis of the tissue—as well as assessments based on the standard noninvasive mammography diagnostics procedure called BI-RADS. The three categorical variables provide qualitative characteristics of the mammographic image features used in BI-RADS—such as the shape of the intrusion (with four categories), characteristics of the intrusion margins (with five categories), and intrusion density (with four categories). This standard categorical dataset is widely used to access the quality of various computer-aided diagnostic (CAD) tools, with the general aim of identifying such a CAD tool that would use the noninvasive information of age and mammographic image features for the precise diagnostics of breast cancer and providing lower rates of false positive and false negative diagnoses than the standard BI-RADS procedure currently used by medical doctors (27). The widely used measure of CAD performance adopted in the medical literature is called area under curve (AUC) (28). The closer the AUC value is to , the better the performance of the respective CAD tool and the lower the probability of false positive and false negative diagnoses. To compute the AUC values of different CAD tools together with the 99% confidence intervals of AUC, we use a methodology described in ref. 28 and available in the open source software library that can be downloaded at https://github.com/brian-lau/MatlabAUC. CAD tools based on the pattern recognition artificial neuronal networks (ANNs) have been reported to have the highest AUC for the breast cancer diagnostics issue (27). Training such ANNs on these data from 830 patients results in AUC of 0.85 with the 99% confidence interval of , whereas we obtain AUC of 0.82 with the 99% confidence interval of for a standard BI-RADS diagnostics (on the same data and computed deploying the same methodology from ref. 28). Therefore, despite the fact that the AUC value of ANNs is somewhat larger than the AUC of BI-RADS, their confidence intervals are largely overlapping—meaning that, from the view point of statistics, this standard dataset does not reveal an advantage of the ANNs compared with BI-RADS. In addition, ANN has many free adjustable parameters (e.g., weights and biases of neurons, transfer functions, etc.), which increase the danger of overfitting for such relatively small data. The application of the categorical reduction procedure described in this paper results in an optimal set of just two collective variables that are both completely defined by information from a single categorical variable “margin.” Very unexpectedly, obtained optimal decomposition into two colvars turns out to be completely independent from all other variables and can be summarized in a very simple diagnostic rule: if the intrusion margin on the mammography image is circumscribed, then the risk of breast cancer is low (12%), and if not, the risk of breast cancer is high (72%). Applying the same open source methodology for AUC confidence intervals on the same standard data as above, we find that this very simple rule (with no free tunable parameters at all) has the AUC value of 0.835 with the 99% confidence interval of (i.e., in terms of the AUC performance, it is not worse than the ANN with approximately 20 free adjustable parameters).
Discussion
The most important features distinguishing the methodology presented in this paper from other approaches described in the literature are that it allows highly scalable (Fig. 1) identification of reduced Bayesian relation models, their uncertainty quantification, and inclusion of a priori physical information and does not rely on the prior knowledge or a necessity of estimation of the full matrix (Eq. 1) of system’s relations in any step. It allows an identification of the colvars and the reduced relation models (Eq. 2)—as well as the MSMs—directly from the observational data.
According to the above theorem, the least biased estimate of the ratio for the expectations of posterior parameter variances from the resulting full and reduced models equals —where is the number of the original dimensions, and is the reduced dimensionality. In application examples 1 and 2 shown above, is of the order of 100—meaning that the reduced models (Eq. 2) can be estimated from much shorter data series than those required for the full model without reduction. In the context of MD and other multiscale applications, this feature can be a used to bridge the gap toward longer time scales in simulations. In particular, in both examples, we have shown how the interpretation of the obtained colvars can provide clues about the locality or nonlocality of relations in the system. Application of this methodology in both examples revealed essentially local models (i.e., models where colvars mostly coincide with only one of the original systems dimensions). In example 1, we have shown that the relation between the local geometry changes of single peptide units in time is local to 97% and that, only in 3% of the original system’s states assembled to colvar, they are nonlocal (and distinctively defined by the peptide junction configuration farther away in the chain). In example 2, the two identified colvars were completely defined through only one of the original data dimensions and are entirely independent from all other information on the system. This seemingly oversimplification of the obtained reduced models could, however, be undermined by the comparison of results and predictions obtained for these very simple reduced models (Fig. 2 or the results of AUC comparison in example 2). The proposed methodology is very simple to implement and to use—we also provide a MATLAB toolbox with all of the methods from this manuscript as open access via the https://github.com/SusanneGerber. As was shown for two application examples, obtained results are straightforward to interpret and provide insights in the underlying systems as well as situations when the system’s dimension is large (e.g., for the example 1) and standard approaches may be subject to the overfitting issues.
Supplementary Material
Acknowledgments
S.G. was supported by Forschungsinitiative Rheinland-Pfalz through the Center for Computational Sciences in Mainz. I.H. is partly funded by the Swiss Platform for Advanced Scientific Computing, Swiss National Research Foundation Grant 156398 MS-GWaves, and the German Research Foundation (Mercator Fellowship in the Collaborative Research Center 1114 Scaling Cascades in Complex Systems).
Footnotes
The authors declare no conflict of interest.
This article is a PNAS Direct Submission.
This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1612619114/-/DCSupplemental.
References
- 1.Fiorin G, Klein M, Henin J. Using collective variables to drive molecular dynamics simulations. Mol Phys. 2013;111:3345–3362. [Google Scholar]
- 2.Schölkopf B, Smola A, Müller KR. In: Kernel Principal Component Analysis. Gerstner W, Germond A, Hasler M, Nicoud J, editors. Springer; Berlin: 1997. pp. 583–588. [Google Scholar]
- 3.Jolliffe I. Principal Component Analysis. Springer; Berlin: 2002. [Google Scholar]
- 4.Schmid P. Dynamic mode decomposition of numerical and experimental data. J Fluid Mech. 2010;656:5–28. [Google Scholar]
- 5.Donoho DL, Grimes C. Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data. Proc Natl Acad Sci USA. 2003;100:5591–5596. doi: 10.1073/pnas.1031596100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Coifman R, et al. Geometric diffusions as a tool for harmonic analysis and structure definition of data: Diffusion maps. Proc Natl Acad Sci USA. 2005;102:7426–7431. doi: 10.1073/pnas.0500334102. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.von Luxburg U. A tutorial on spectral clustering. Stat Comput. 2007;17:395–416. [Google Scholar]
- 8.Liou C, Cheng W, Liou J, Liou D. Autoencoder for words. Neurocomputing. 2014;139:84–96. [Google Scholar]
- 9.Zhao Y, Levina E, Zhu J. Community extraction for social networks. Proc Natl Acad Sci USA. 2011;108:7321–7326. doi: 10.1073/pnas.1006642108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Perez-Hernández G, Paul F, Giorgino T, De Fabritiis G, Noe F. Identification of slow molecular order parameters for markov model construction. J Chem Phys. 2013;139:015102. doi: 10.1063/1.4811489. [DOI] [PubMed] [Google Scholar]
- 11.Roblitz S, Weber M. Fuzzy spectral clustering by pcca+: Application to markov state models and data classification. Adv Data Anal Classif. 2013;7:147–179. [Google Scholar]
- 12.Ding C, Li T, Peng W. Proceedings of the 21st National Conference on Artificial Intelligence. Vol 1. AAAI Press; Berkeley, CA: 2006. Nonnegative matrix factorization and probabilistic latent semantic indexing: Equivalence, chi-square statistic, and a hybrid method; pp. 342–347. [Google Scholar]
- 13.Hofmann T. Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM; New York: 1999. Probabilistic latent semantic indexing; pp. 50–57. [Google Scholar]
- 14.Hofmann T. Unsupervised learning by probabilistic latent semantic analysis. Mach Learn. 2001;42:177–196. [Google Scholar]
- 15.Prinz J, et al. Markov models of molecular kinetics: Generation and validation. J Chem Phys. 2011;134:174105. doi: 10.1063/1.3565032. [DOI] [PubMed] [Google Scholar]
- 16.Bowman G, Pande V, Noé F. 2013. An Introduction to Markov State Models and Their Application to Long Timescale Molecular Simulation, Advances in Experimental Medicine and Biology (Springer, The Netherlands)
- 17.Gardiner H. Handbook of Stochastical Methods. Springer; Berlin: 2004. [Google Scholar]
- 18.Holland P. Statistics and causal inference. J Am Stat Assoc. 1986;81:945–960. [Google Scholar]
- 19.Schütte C, Sarich M. 2013. Metastability and Markov State Models in Molecular Dynamics: Modeling, Analysis, Algorithmic Approaches, Courant Lecture Notes American Mathematical Society, New York.
- 20.Wahba G. Spline Models for Observational Data. SIAM; Philadelphia: 1990. [Google Scholar]
- 21.Nocedal J, Wright SJ. Numerical Optimization. 2nd Ed Springer; New York: 2006. [Google Scholar]
- 22.Burnham K, Anderson D. Model Selection and Multimodel Inference: A Practical Information-Theoretic Approach. Springer; Berlin: 2002. [Google Scholar]
- 23.Hansen P, OLeary D. The use of the l-curve in the regularization of discrete ill-posed problems. SIAM J Sci Comput. 1993;14:1487–1503. [Google Scholar]
- 24.Berry MW, Browne M, Langville AN, Pauca VP, Plemmons RJ. Algorithms and applications for approximate nonnegative matrix factorization. Comput Stat Data Anal. 2007;52:155–173. [Google Scholar]
- 25.Lindorff-Larsen K, et al. Improved side-chain torsion potentials for the Amber ff99SB protein force field. Proteins. 2010;78:1950–1958. doi: 10.1002/prot.22711. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Gerber S, Horenko I. On inference of causality for discrete state models in a multiscale context. Proc Natl Acad Sci USA. 2014;111:14651–14656. doi: 10.1073/pnas.1410404111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Elter M, Schulz-Wendtland R, Wittenberg T. The prediction of breast cancer biopsy outcomes using two CAD approaches that both emphasize an intelligible decision process. Med Phys. 2007;34:4164–4172. doi: 10.1118/1.2786864. [DOI] [PubMed] [Google Scholar]
- 28.Qin G, Hotilovac L. Comparison of non-parametric confidence intervals for the area under the ROC curve of a continuous-scale diagnostic test. Stat Methods Med Res. 2008;17:207–221. doi: 10.1177/0962280207087173. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.