Abstract
Many methods have been developed to predict static protein structures, however understanding the dynamics of protein structure is essential for elucidating biological function. While molecular dynamics (MD) simulations remain the in silico gold standard, its high computational cost limits scalability. We present DynaProt, a lightweight, SE(3)-invariant framework that predicts rich descriptors of protein dynamics directly from static structures. By casting the problem through the lens of multivariate Gaussians, DynaProt estimates dynamics at two complementary scales: (1) per-residue marginal anisotropy as 3 × 3 covariance matrices capturing local flexibility, and (2) joint scalar covariances encoding pairwise dynamic coupling across residues. From these dynamics outputs, DynaProt achieves high accuracy in predicting residue-level flexibility (RMSF) and, remarkably, enables reasonable reconstruction of the full covariance matrix for fast ensemble generation. Notably, it does so using orders of magnitude fewer parameters than prior methods. Our results highlight the potential of direct protein dynamics prediction as a scalable alternative to existing methods.
1. Introduction
Proteins rarely exist in static conformations. Due to interactions with ligands, other biomolecules, and external factors such as temperature and pH, protein structures continuously fluctuate. Many enzymes rely on loop motions or domain rearrangements to form catalytically active sites (Zinovjev et al., 2024), allostery often involves shifting backbone or side chain conformations propagating signals over long distances (Yu and Koshland Jr, 2001), and even membrane proteins such as GPCRs switch between inactive and active conformational states essential for signal transduction (Zhang et al., 2024). Clearly, protein structures are dynamic. Understanding these dynamics is central for mechanistic insight and, potentially, the design of functions (Guo et al., 2025; McCammon, 1984).
Capturing this ensemble-level behavior computationally has long been the domain of molecular dynamics (MD). MD simulates the time evolution of atoms under a force field, generating high-resolution conformational trajectories from which fluctuations, covariances, and time-dependent observables can be derived. MD remains the in silico gold standard for protein dynamics, offering fine-grained, physically grounded insights (Shaw et al., 2010; Hollingsworth and Dror, 2018; Childers and Daggett, 2017). However, it comes with an enormous computational cost: simulating 100 ns of dynamics for a single protein can take days or weeks on specialized hardware. This limits its scalability, especially for proteome-wide applications or tasks requiring real-time dynamics estimates.
Recent work has explored the use of deep learning to approximate and accelerate this process. Generative modeling-based methods like AlphaFlow (Jing et al., 2024) repurpose AlphaFold2 (Jumper et al., 2021) under a flow matching paradigm to sample protein conformations. Along this vein, Lewis et al. (2024) recently introduced BioEMU as a large-scale diffusion model pretrained on PDB (Burley et al., 2017) and AFDB structures, and fine-tuned on 200 ms of MD data, to efficiently generate protein conformations. Other methods, like MSA subsampling, make inference-time adjustments to the MSA input of AlphaFold2, yielding the structural ensembles (Del Alamo et al., 2022; Wayment-Steele et al., 2024; Stein and Mchaourab, 2022). Still, all of these approaches necessitate large scale PDB pretraining and suffer from inference-time computational overhead, requiring multiple stochastic forward passes to generate meaningful structural diversity. Moreover, while these ensembles can be used to approximate protein dynamics, generating them remains time intensive, and the full ensemble of diverse conformations may not always be necessary. In many practical settings, compact and interpretable representations of dynamics often suffice. This motivates the need for models that can explicitly predict such dynamics descriptors without relying on implicit dynamics learners like expensive ensemble generation methods.
Current explicit dynamics predictors, like FlexPert3D (Kouba et al., 2024), resort to predicting simple collective variables like per-residue RMSF, a scalar quantifying each residue’s positional fluctuation. RMSF is widely used due to its simplicity and interpretability, but it is fundamentally limited: it captures only the magnitude of local motion and discards directionality and residue-residue coupling. Similarly, Wayment-Steele et al. (2025) trained Dyna-1 to predict labels of µs–ms motion by cleverly exploiting missing chemical shift assignments as hidden observables in NMR ensembles, but these predictions also remain scalar and lack directionality. A different example of an explicit dynamics predictor is Normal Mode Analysis (NMA), a classic technique that approximates dynamics by identifying low frequency eigenmodes to describe the largest movements (Cui and Bahar, 2005; Skjaerven et al., 2009). NMA does not learn from data however, and instead operates solely on the input PDB structure. It can estimate the principal global directions of motion and offers insights into collective flexibility, but is sensitive to input structure quality and fails to adequately capture local anisotropy or conformational heterogeneity (Ma, 2005). This raises a natural question:
Can we design models that lie on the Pareto frontier of expressiveness and efficiency—capturing rich dynamic behavior without incurring the cost of sampling or simulation?
We introduce DynaProt, a lightweight, interpretable, and expressive framework for predicting protein dynamics through the lens of Gaussian distributions over structure (Section 2). Specifically, DynaProt predicts: (1) per-residue marginal anisotropy as covariance matrices capturing local dynamics while encompassing RMSF, and (2) joint scalar covariances encoding pairwise dynamic coupling across residues. Remarkably, while DynaProt was not explicitly trained to directly model the full joint distribution, we find that its marginal and pairwise outputs can be composed into a reasonable approximation (Section 3.4), enabling extremely fast ensemble generation in . DynaProt is trained on only ∼1,000 MD-derived proteins, without large-scale pretraining on PDB structures, and improves upon Normal Mode Analysis (NMA) in both predictive accuracy and efficiency, while remaining dramatically smaller and faster than existing ensemble generation approaches.
To our knowledge, DynaProt is the first model to explicitly learn both marginal and pairwise Gaussian representations of protein dynamics, and the first to predict the full covariance structure—akin to NMA—in a data-driven, learnable fashion, rather than relying solely on analytical approximations or less informative per-residue fluctuations.
2. Gaussian representation of dynamics
We propose a perspective for modeling protein dynamics through distributions over atomic coordinates, relying on tractable approximations such as Gaussians. Formally, we model a protein structure with residues as a random variable , where each residue contributes the three-dimensional Cartesian coordinates of its atom. While this coarse-grained representation omits side-chain flexibility, it enables scalable modeling of backbone dynamics, which is the scope of our work. We consider an ensemble to be independent samples after RMSD alignment. The distribution over conformational states is then represented as a multivariate normal distribution:
| (1) |
Here, corresponds to the average (or equilibrium) structure—typically the minimum energy conformation—and captures the full covariance across all positions, encoding both local fluctuations and long-range correlated motions. This joint covariance matrix theoretically encodes all second-order information about the protein’s dynamics: from it, one can derive a wide range of collective variables including principal components (PCs) of motion, residue-residue distance variances, and global flexibility metrics. The Gaussian formulation provides a principled way to decompose protein dynamics across different levels of granularity (Figure 2 and Table 1), depending on the modeling objective. We highlight some of these below.
Figure 2:
Overview of protein dynamics under Gaussian view.
Table 1:
Taxonomy of protein dynamics representations under a Gaussian view.
| Level | Description | Notation | Space | Captures |
|---|---|---|---|---|
| 1 | Per-residue scalar (i.e. RMSF) | Magnitude of fluctuation per residue | ||
| 2 | Per-residue full (Gaussian blob) | Anisotropic local covariance per residue | ||
| 3 | Joint scalar (pairwise coupling) | Scalar covariance across all residues | ||
| 4 | Joint full covariance | Full spatial covariance across all residues |
At the local level, the marginal distribution (Figure 2.2) for a single residue is obtained by integrating out all other residue coordinates: , where . This results in a 3D Gaussian distribution over the coordinates of residue :
| (2) |
where is the diagonal block of . These marginals can be interpreted as Gaussian blobs encoding anisotropic local fluctuations—i.e., spatial variance of where each residue may reside. Notably, this formulation allows for simple derivation of scalar flexibility metrics such as the root-mean-square fluctuation (RMSF) as . RMSF (Figure 2.1) represents a simple notion of dynamics: a single scalar per residue quantifying positional fluctuation. However, it discards directional and covariance structure captured by the full marginal.
To capture dynamics beyond residue-local fluctuations, we also consider a covariance matrix of scalar pairwise coupling (Figure 2.3). Each entry summarizes the dynamical coupling between residues and , typically computed as a scalar projection of the corresponding block in the full joint covariance: . We choose MeanPooling as the scalar projection to compute each . This compact representation enables efficient modeling of residue-residue coupling.
DynaProt focuses on levels 2 and 3 of this hierarchy—explicitly predicting both marginal Gaussians per-residue and a matrix of residue-residue couplings. As noted before, from the marginals, we can easily derive RMSF (level 1). Interestingly, utilizing both the marginals and the pairwise coupling, we can retrieve a reasonable approximation of the full joint (level 4; Section 3.4). This design strikes a balance between local interpretability and global coordination, while avoiding the intractibility of directly learning the full joint covariance.
3. Method
3.1. DynaProt Overview
DynaProt (Figure 3) consists of two models, each taking as input a protein structure but designed to explicitly capture different granularities of protein dynamics: (i) marginal Gaussian blobs per residue (Section 3.2), and (ii) pairwise covariance across residues (Section 3.3). DynaProt is given the input structure as a set of local residue frames. The frames are denoted , where each frame is parameterized by a rotation matrix and a translation vector . Simply put, captures the local orientation and position of residue . Additionally, an initial embedding layer is included to encode the amino acid sequence .
Figure 3:
DynaProt architecture.
Both models share a common architectural backbone composed of eight Invariant Point Attention (IPA) blocks from the structure module of AlphaFold2 (Jumper et al., 2021). These blocks are designed to encode geometric relationships between residues while maintaining invariance to SE(3) transformations (global transformations do not affect the learned residue-level representations). The IPA backbone processes the set of residue frames and the sequence representation, outputting a learned representation for each residue . No pair representation is given as input to the model. The two models differ only in their readout layers, which we define in the proceeding sections.
3.2. Learning marginal Gaussians
After the input sequence representations and residue frames are processed through the IPA backbone, a simple MLP readout is used for marginal prediction. Given the hidden representation for each residue, the marginal readout outputs , modeling the local position (xyz) covariance of residue . These outputs are trained to match empirical marginal distributions derived from the MD data. Note that the mean of each Gaussian is not learned. Instead, we take the input structure’s coordinate as the fixed mean of the distribution: , . This assumption is motivated by the fact that the input structure usually corresponds to the experimentally determined (or AlphaFold-predicted) minimum energy conformation, and thus serves as a natural estimator of the ensemble mean. Consequently, the marginal prediction task reduces to learning the covariance matrices alone.
Marginal dynamics module.
Recall that covariance matrices are required to be symmetric and positive definite (SPD). Predicting all 9 elements of a matrix would be overparameterized and does not guarantee SPD structure. Naively, one might consider symmetrizing an arbitrary matrix after predicting the 6 independent elements, but this only guarantees symmetry. Instead, we leverage the fact that any SPD matrix can be uniquely defined by its Cholesky factorization. Thus, we enforce SPD constraints directly by parameterizing the covariance via its Cholesky factor. Specifically, the model predicts the entries of a lower triangular matrix , enforces positivity along the diagonal with the Softplus activation function (Glorot et al., 2011), and recovers the covariance:
| (3) |
This factorization ensures that the predicted covariance matrix is SPD by construction. Since SPD matrices lie on a Riemannian manifold with non-Euclidean geometry, using loss functions that respect this structure is critical for meaningful comparison. Standard Euclidean distances (e.g., MSE or Frobenius norm) ignore the curvature of this space and can lead to unstable or distorted gradients (see ablations in Appendix A.3.2). We instead employ the log-Euclidean (or log-Frobenius) distance (Vemulapalli and Jacobs, 2015; Huang et al., 2015) that reflects the intrinsic geometry of the SPD manifold. The Bures-Wasserstein (Bhatia et al., 2019) distance can also be used, but we find the log-Frobenius distance to be more stable.
| (4) |
As Riemannian manifolds are “locally Euclidean”, this loss applies the matrix logarithm mapping the SPD matrix to its tangent space where a Euclidean metric (canonical Frobenius norm) can be utilized.
3.3. Learning pairwise dynamics
Using the output representations from the IPA backbone, the pairwise dynamics module produces a scalar-valued covariance matrix , where each entry captures the dynamical coupling between residue pairs. These scalar couplings are derived from the full joint covariance matrix via averaging per block and trained to reproduce MD-derived pairwise fluctuations.
Pairwise Dynamics Module.
To predict the global pairwise covariance structure, we first construct pairwise features for all residue pairs. For each pair , we concatenate their residue-level embedings and project them into a lower-dimensional space: .
We pass these features through a stack of AlphaFold-style pairwise attention blocks based on the Evoformer architecture (Jumper et al., 2021), which include triangle updates and residue-wise message passing. These operations are designed to model transitive and higher-order geometric dependencies across residue pairs, and have been shown to be highly effective in structure-aware tasks: . The output serves as a “learned basis”“ over the space of residue-residue covariance structure. These basis features are then mapped to scalars through an MLP head, yielding a covariance for each pair of residues: , , for . Following the same procedure as Section 3.2, we enforce SPD constraints on this covariance matrix by populating the lower-triangle entries of with the values of and applying the Softplus activation when . Finally, the pairwise covariance matrix is reconstructed via Cholesky composition and again equation 4 is used for optimization.
3.4. Learning the full joint for ensemble sampling
Joint reconstruction heuristic.
Given a predicted scalar coupling matrix and a set of per-residue marginal covariances , we propose a heuristic to reconstruct an approximate full joint covariance matrix .
Each marginal covariance is SPD by construction, and thus admits a Cholesky factorization , where . We then define a block-diagonal matrix as . By construction, is lower triangular with positive diagonal entries, since each satisfies these properties.
Drawing from the univariate identity , we define the multivariate cross-covariance block between residues and as . Here, the Cholesky factor serves as a matrix square root of the covariance , analogous to standard deviation in the univariate case. And, is a correlation matrix found by standardizing . Using the Kronecker product, we can denote this heuristic cleanly as follows,
| (5) |
Proposition 3.1 (SPD Closure). Given marginal covariances and correlation matrix to be symmetric and positive definite, then the reconstructed joint covariance is also symmetric and positive definite.
We refer the reader to Appendix A.1 for the proof. This approximation combines local anisotropic uncertainty with global correlation structure. While not exact, we find it reconstructs the joint covariance to a reasonable degree and serves as a useful tool for downstream ensemble generation.
Ensemble sampling.
Given the reconstructed joint covariance and our assumption that the mean corresponds to the coordinates of the input structure (e.g., the PDB), we have now retrieved our Gaussian distribution over conformations . To sample from this distribution, we apply a multivariate generalization of the reparameterization trick used in univariate Gaussian sampling.
Property 3.1 (Multivariate Gaussian Sampling). Given , where is SPD and is its Cholesky decomposition. Then,
Note that this sampling relies directly on the Cholesky factor (similar to a matrix square root), mirroring the scalar case (Appendix A.1). Utilizing DynaProt predictions and this heuristic, ensemble sampling becomes extremely fast with minimal computational overhead.
4. Experiments
4.1. Setup
Preprocessing.
We construct ground-truth dynamics labels from the ATLAS molecular dynamics dataset, which comprises 1,390 proteins selected based on structural diversity using the ECOD domain classification (Vander Meersche et al., 2024). Following AlphaFlow (Jing et al., 2024) for preprocessing consistency, we concatenate each of the three replicate simulations of 100 ns per protein and extract the coordinates. From each ensemble, we compute the empirical full joint covariance matrix over time and extract the relevant dynamics labels ( marginals per residue and residue coupling) as described in Section 2. We evaluate under two train/val/test split regimes. The primary matches AlphaFlow’s (1265/39/82), while comparisons to FlexPert3D use DynaProt trained on their topology-based split (1112/139/139). For naming, we refer to DynaProt-M for the model trained for marginals, DynaProt-J for the coupling predictions, and DynaProt for both.
Baselines.
For a faithful comparison, we mainly choose baseline methods that take a protein structure as input and predict dynamics descriptors either implicitly (AFMD+Templates, considered state-of-the-art) or explicitly (FlexPert3D, NMA). For NMA, we utilize the ProDy package (Zhang et al., 2021), specifically the Anisotropic Network Model instantiation. There are many methods worth noting that aim to predict ensembles or dynamics descriptors from sequence itself: standard AlphaFlow (AFMD), MSA-Subsampling, FlexPert-Seq, ESMDiff (Lu et al., 2024), ConfDiff (Wang et al., 2024), SeqDance (Hou and Shen, 2024), and SeaMoon (Lombard et al., 2024). With the advancements in structure prediction Abramson et al. (2024), we note that one could first predict a plausible structure from sequence and then predict dynamics using DynaProt. Thus, we compare against some of these sequence based methods as well (Appendix A.3.3).
4.2. Predicting residue flexibility
Since DynaProt-M is trained to predict marginal Gaussians per residue, it inherently captures residue-level flexibility, as RMSF is defined as the square root of the trace of each marginal covariance (see Section 2). To evaluate DynaProt-M’s ability to recover this flexibility, we compare against what is, to our knowledge, the only method that explicitly predicts residue flexibility: FlexPert-3D. For fair comparison, we train and evaluate DynaProt-M under the same ATLAS train/val/test split defined in Kouba et al. (2024). Despite solving the more challenging task of predicting full anisotropic marginals rather than scalar fluctuations alone, DynaProt-M achieves a substantially higher Pearson correlation with MD-derived RMSF (median , 75th percentile ) than FlexPert-3D (Table 2), while using three orders of magnitude fewer parameters (955K vs. 1.2B) and without normal mode analysis as input. This allows DynaProt-M to generalize better while being more parameter efficient. See Appendix A.4 for DynaProt-M additional RMSF plots.
Table 2:
RMSF Pearson correlation against ATLAS MD-derived RMSF (FlexPert test split). Median and 75th percentile reported.
| Method | # Params | |
|---|---|---|
| DynaProt-M | 0.865 / 0.930 | 955 K |
| FlexPert-3D | 0.830 / 0.899 | 1.2 B |
| NMA (ANM) | 0.697 / 0.784 | – |
4.3. Predicting residue full anisotropy
To assess the faithfulness of DynaProt-M’s predicted marginals, we compare against both physics-based and learned ensemble methods. In practice, only NMA (ANM) is a feasible baseline, as AlphaFlow is prohibitively slow: A single 271-length protein (7lao_A) requires ∼7000 s, compared to ∼0.02 s for DynaProt-M (Table 3). Note that DynaProt-M predicts this directly, but for AlphaFlow, we first sample 250 structures per protein and then calculate the empirical covariance to define the marginal Gaussians. With NMA, we retrieve the full joint via the normal modes and extract the marginal block diagonals.
Table 3:
Comparison of methods on anisotropic blob prediction (ATLAS test split). Runtime for a length 271 protein (7lao_A). 25th %ile / Median reported (↓ is better).
| Method | RMWD Var | Sym. KL Var | # Params | Time |
|---|---|---|---|---|
| DynaProt-M | 0.84 / 1.18 | 0.53 / 0.91 | 955 K | ∼0.02 s |
| AFMD+T | 0.87 / 1.10 | 0.37 / 0.60 | 95 M | ∼7000 s |
| NMA (ANM) | 1.14 / 1.45 | 3.03 / 4.56 | – | ∼5.37 s |
To quantify the accuracy, we compute the variance contribution of the symmetric KL divergence (see Appendix A.2) and the root mean 2-Wasserstein distance (RMWD) as described in Jing et al. (2024), compared to the ground truth marginal Gaussians computed from the ATLAS test set (AFMD split). Despite being orders of magnitude faster and smaller (955k vs. 95M parameters), DynaProt-M achieves competitive accuracy. DynaProt-M attains a median RMWD of 1.18 and symmetric KL divergence of 0.91, both substantially better than NMA (1.45 and 4.56, respectively), and comparable to AFMD+Templates’s 1.10 and 0.60. Moreover, rather than relying solely on summary statistics, we also visualize the distributions (75th percentile) of RMWD and mean symmetric KL variance contributions across test set proteins (Figure 4B). These plots compare DynaProt-M to both AFMD+T and NMA on a per-protein basis. Points below the diagonal (highlighted in pink) indicate that DynaProt-M outperformed the method in question on that particular protein. From this, we see that DynaProt-M achieves comparable performance to AFMD+T. Notably, within the 75th percentile it often outperforms AFMD+T (examples visualized in Figure 4A,B) on RMWD variance contribution. Moreover, DynaProt-M significantly outperforms NMA across both RMWD and symmetric KL. This further corroborates DynaProt-M’s ability to capture local anisotropic structure well despite being much smaller and faster than other methods.
Figure 4: DynaProt marginal Gaussian and residue coupling analysis.
A. Renderings of predicted marginal Gaussians compared to ATLAS MD constructed Gaussians (mean symmetric KL divergence and RMWD are reported). B. Joint distribution (within 75th percentile) of DynaProt performance vs. (AFMD+T, NMA). C. Band-wise Pearson correlation between predicted and ground-truth residue–residue coupling matrices as a function of sequence distance.
4.3.1. DynaProt zero-shot cryptic pocket discovery of Adenylosuccinate Synthetase
Beyond accuracy, DynaProt-M’s marginals can also provide functional insight. Many proteins are considered to be undruggable as their apo form may not display a clear binding pocket. However, the druggable pocket may only become apparent after the drug is bound (holo form)–a so called “cryptic pocket.” Identification of cryptic pockets is therefore an important task in drug discovery (Mou et al., 2025; Hollingsworth et al., 2019; Comitani and Gervasio, 2018). As a case study, we sought to investigate DynaProt’s ability in cryptic pocket identification for the enzyme adenylosuccinate synthetase. As indicated in Meller et al., this enzyme is known to exhibit a cryptic pocket and both the apo and holo forms are available in the PDB (1ADE / 1CIB).
We applied DynaProt-M to zero shot predict the marginal Gaussians on the apo form. When we look at the predictions, we notice two features (Figure 5). First, the residues with the largest variance are exactly those that encompass the binding pocket. Second, more than just the magnitude of the Gaussians, the directionality is consistent with a potential pocket opening motion (when compared with the holo form 1CIB). These early results suggest the potential of DynaProt’s utility in cryptic pocket discovery, but a systematic exploration is left for future work.
Figure 5:
DynaProt-M predicted residue Gaussians (ellipsoids) overlaid the apo form.
4.4. Predicting scalar coupling
To assess the performance of DynaProt-J in modeling residue–residue coupling, we evaluate its predicted scalar covariance matrices against those derived from NMA, a classical method for capturing global structural dynamics. While DynaProt-J directly predicts these scalar coupling matrices, we construct comparable matrices for NMA by computing full anisotropic network models (using ProDy) from each test protein, and then projecting them into scalar covariances as described in Section 2. We obtain per-residue correlation matrices by normalizing the entries to be unit diagonal and constrained to the range [−1, 1].
We observe that in the ground truth correlations, magnitude of entries diminish rapidly with distance from main diagonal, indicating weak long-range coupling. This is indicated by the dotted gray line in Figure 4C, which shows the mean absolute value of the entries from the principal diagonal up to the diagonal band. To focus on meaningful and prominent interactions, we define a diagonal band of width residues , which captures local and medium-range interactions. This essentially measures residue-residue coupling as a function of sequence distance (how distal are and along the backbone). We compute the Pearson correlation between predicted and ground-truth residue–residue correlation matrices for the entries along each diagonal band , by iteratively extracting the upper-triangular entries satisfying for . This is repeated for each of the 82 test set proteins and the median Pearson correlation is reported (each point in Figure 4C). This band-wise analysis enables us to compare the accuracy of coupling signals at increasing residue distances, and we find that DynaProt-J (peak correlation of ) strongly outperforms NMA (peak correlation of ) particularly at short to mid-range coupling distances, where the coupling is the strongest.
4.5. Ensemble generation
As described in Section 3.4, given the output marginal covariances and residue coupling from DynaProt-M and DynaProt-J respectively, we reconstruct a full joint covariance using the heuristic defined in Eq. 5. This direct access to the joint distribution enables extremely fast sampling of diverse structures. For evaluation, we sample 250 structures with DynaProt, AFMD+Templates, and NMA to form ensembles for each of the 82 test set proteins in the AFMD split. Following the evaluation protocol of Jing et al. (2024), we assess these ensembles across a variety of metrics including flexibility accuracy, distributional similarity, and the ability to reproduce complex observables. For flexibility accuracy, we measure the pairwise RMSD to ground truth MD and RMSF correlation at the global and per-target level. For distributional coverage, we measure the 2-Wasserstein distance after projecting the ensembles onto the first two principle components derived from the MD trajectory or the combined (MD+sampled) trajectory .
Table 4 summarizes the ensemble evaluation results across AFMD+Templates, DynaProt, and NMA. DynaProt achieves performance comparable to AFMD+Templates on key flexibility metrics such as pairwise RMSD and per-target RMSF correlation, while lagging slightly behind on distributional similarity and observable recovery. Some examples of where DynaProt outperforms AFMD+Templates on ensemble flexibility correlation are visualized in Figure 6 and Appendix A.4. For the visuals, all atom reconstruction is enabled by PULCHRA (Rotkiewicz and Skolnick, 2008). Moreover, DynaProt consistently outperforms NMA across nearly all evaluations—except for transient contact prediction—particularly excelling in measures of local flexibility and pairwise distance preservation. Notably, DynaProt requires only 2.86 million parameters (vs. 95 million for AFMD+Templates) and samples ensembles over 70,000× faster on average (∼0.14 s vs. ∼10,000 s), all while being trained only to predict marginal and scalar covariances. This efficiency advantage is maintained when compared against sequence-based methods (i.e. Conf Diff, BioEmu, ESMDiff, see Appendix A.3). DynaProt even outperforms them on modeling flexibility and anisotropy.
Table 4:
Comparison of ensemble evaluation metrics on ATLAS MD Dataset between AFMD+Templates, DynaProt, and NMA. Underlined is second best value.
| Metric | AFMD+Templates | DynaProt | NMA |
|---|---|---|---|
| Pairwise RMSD (=2.89) | 2.18 | 2.17 | 0.91 |
| RMSF (=1.48) | 1.17 | 1.10 | 0.52 |
| 0.91 | 0.71 | 0.54 | |
| 0.92 | 0.86 | 0.76 | |
| 1.25 | 1.74 | 1.86 | |
| 1.58 | 2.39 | 2.45 | |
| 0.62 | 0.51 | 0.43 | |
| 0.47 | 0.29 | 0.33 | |
|
| |||
| # Parameters (↓) | 95 M | 2.86 M | – |
| Ensemble sampling time (↓) | ~ 10, 000 s | ~ 0.14 s | ~ 5.69 s |
Figure 6:
Comparison of DynaProt generated ensemble vs. AFMD+T to ATLAS MD simulation (PDB 7qsu_A) overlaid on reference. RMSF Pearson correlation and sample time reported.
As a final assessment of DynaProt’s generalization to longer timescale dynamics, we compare its zero-shot ensemble of BPTI (PDB: 5PTI) to the 1ms trajectory from Shaw et al. (2010). Even with these larger conformational changes, DynaProt performs reasonably well. It achieves RMSF correlation of 0.88 (c.f. 86 on ATLAS), local anisotropy with RMWD of 0.52 Å(c.f. 1.18 Å on ATLAS), and strong recovery of transient contacts (Jaccard similarity 0.54, c.f. 0.29 on ATLAS). See appendix A.3.1 for more.
5. Conclusion
Protein dynamics is critical for understanding biological function. Existing approaches to modeling dynamics often rely on complex generative models with large-scale PDB pretraining and expensive ensemble generation. In this work, we introduce DynaProt, a lightweight and data-driven alternative akin to Normal Mode Analysis (NMA), but designed to directly predict structured dynamics descriptors in the form of per-residue and pairwise Gaussian representations. This formulation enables extreme parameter efficiency while outperforming traditional baselines on key metrics, including flexibility estimation, marginal anisotropy, and residue–residue coupling. Remarkably, DynaProt’s outputs also support ultra-fast ensemble sampling with reasonable structural fidelity—offering a compelling alternative to conventional ensemble generation methods. While further scaling may be needed to match the full capabilities of state-of-the-art generative methods, our approach highlights a promising alternative grounded in explicitly learning structured representations of dynamics.
Figure 1:
Dynamics methods information content vs. efficiency.
Acknowledgements
We would like to thank Alexander Amini, Jeffrey Li, Yuseong (Nick) Oh, Ryan Mei, Manoj Niverthi, Renzo Soatto, and Kabir Doshi for their useful ideas and discussions.
This work was supported by the National Institute of General Medical Sciences of the National Institutes of Health under award number 1R35GM141861–01.
Appendix
A. Appendix
A.1. Method details
SPD closure of joint reconstruction heuristic (restating Proposition 3.1).
Given marginal covariances and a correlation matrix that is symmetric and positive definite, then the reconstructed joint covariance is also symmetric and positive definite.
Proof. Let be the block-diagonal matrix defined as
where each is the Cholesky factor (or any valid matrix square root) of the positive definite marginal . Since each , it follows that each is full rank, lower triangular, and has positive diagonal entries. Consequently, is full rank and lower triangular with positive diagonal blocks.
Now consider the matrix . Since and , the Kronecker product as well (Kronecker product of two SPD matrices is also SPD). Finally, the product
is a congruence transformation of the SPD matrix by the full-rank matrix . Since congruence preserves positive definiteness, we conclude:
Moreover, is symmetric because it is of the form .
Multivariate Gaussian Sampling.
Let be a standard multivariate normal in , and let , be a symmetric positive definite matrix. Suppose satisfies (e.g., via Cholesky decomposition or matrix square root). Then the transformation yields a random variable .
Proof. Since Gaussian distributions are fully characterized by their first two cumulants (mean and covariance), it suffices to show that the transformed variable has the desired mean and covariance.
Mean of :
Covariance of :
A.2. Evaluation Metrics
RMWD Variance Contribution.
To evaluate the efficacy of the marginal Gaussian predictions, we adopt distributional similarity metrics used in Jing et al. (2024). The first of these is the root mean 2-Wasserstein distance (RMWD), specifically its variance contribution term. The 2-Wasserstein distance between two multivariate Gaussians has a closed-form expression as follows.
Let and be two -dimensional Gaussian distributions. The squared 2-Wasserstein distance between them is given by:
| (6) |
This expression consists of two additive components: a mean contribution and a covariance (variance) contribution. This metric is also referred to as the Bures–Wasserstein distance (Bhatia et al., 2019). Since our method predicts only the covariances, we isolate and evaluate only the second term. We define the RMWD variance contribution across residues as follows:
| (7) |
Symmetric KL Divergence Variance Contribution.
Alongside the Wasserstein-based metric, we also evaluate the discrepancy between predicted and ground-truth marginal distributions using the symmetric Kullback–Leibler (KL) divergence, defined as the mean of the two directed divergences mentioned in (Kullback and Leibler, 1951; Jeffreys, 1998):
For two -dimensional Gaussian distributions and , the divergence from to is given by:
| (8) |
This expression consists of both a mean contribution—the Mahalanobis term—and a variance contribution, comprising the trace and log-determinant terms. Since our method predicts only covariances (and uses the input structure coordinates as means), we isolate the variance terms by omitting the mean term:
| (9) |
To symmetrize the variance contribution of the divergence, we define the symmetric variance as:
| (10) |
| (11) |
A.3. Additional Experiments
A.3.1. DynaProt zero-shot ensemble generation of BPTI
We note that DynaProt was trained on the ATLAS MD dataset comprising 100 ns per replicate trajectories. In contrast, D.E. Shaw Research performed simulations of BPTI (PDB: 5PTI) at millisecond-scale revealing structurally distinct conformational states (Shaw et al., 2010). Thus, in an effort to understand DynaProt’s ability to generalize to long-timescale dynamics, we applied it to BPTI and compared to the DESRES trajectory.
Listed in Table 7b, we compute the ensemble evaluation metrics from Jing et al. (2024) and observe that DynaProt performs remarkably well: e.g., RMSF correlation of 0.88 (c.f. 86 on ATLAS), local anisotropy with RMWD of 0.52 Å (c.f. 1.18 Å on ATLAS), and strong recovery of transient contacts (Jaccard similarity 0.54, c.f. 0.29 on ATLAS). These metrics emphasize that DynaProt is able to model larger conformational changes at high fidelity.
Figure 7.
(a) Visualization of DynaProt zero shot BPTI (PDB: 5PTI) ensemble generation. (b) DynaProt zero shot ensemble generation of BPTI (PDB: 5PTI), compared to DESRES MD trajectory (Shaw et al., 2010). Note that the global RMSF Pearson correlation is omitted as there is only one protein so global = per-target.
A.3.2. DynaProt ablations
To test both the importance of DynaProt’s Riemannian aware loss (log Frobenius norm) and the SE(3) invariance from the IPA layers, we have performed the following ablations listed in Table 5. Unsurprisingly, replacing the log Frobenius norm objective with standard Mean Squared Error loss significantly degrades performance as the optimization is over the space of positive definite covariance matrices, which lies on a well-studied Riemannian manifold. Replacing the IPA blocks with standard MLPs also degrades performance, suggesting that SE(3) invariance is crucial in this low-data, low-parameter regime.
Table 5:
DynaProt-M ablations of the log Frobenius loss loss and SE(3) invariance.
| Metric | DynaProt | No LogFrob Loss | No SE(3) Invariance |
|---|---|---|---|
| RMWD Var (↓) | 1.18 | 2.70 | 1.92 |
| Sym KL Var (↓) | 0.91 | 9.26 | 4.46 |
| 0.87 | 0.38 | 0.48 |
A.3.3. Sequence input baselines
To further evaluate DynaProt’s ensemble generation, we compare against methods that take sequence as input following the AlphaFlow evaluations. DynaProt outperforms these methods on local RMSF correlation and marginal anisotropy prediction and is comparable with other distributional metrics. Moreover, the efficiency advantage is clear with DynaProt’s sub-second sample time.
Table 6:
Comparison of DynaProt generated with ensemble generation methods that take in sequence as input. ESMDiff, ESM3 entries reported from Lu et al. (2024).
| Metric | DynaProt | Conf Diff OF-r3-MD | AlphaFlow-MD | BioEmu | ESM3 (ID) | ESMDiff (ID) |
|---|---|---|---|---|---|---|
| Pairwise RMSD (=2.89) | 2.17 | 3.43 | 2.89 | 3.57 | - | - |
| RMSF (=1.48) | 1.10 | 2.21 | 1.68 | 2.47 | - | - |
| 0.71 | 0.67 | 0.60 | 0.63 | 0.19 | 0.49 | |
| 0.86 | 0.85 | 0.85 | 0.77 | 0.67 | 0.68 | |
| RMWD Var Contrib (↓) | 1.18 | 1.40 | 1.30 | 2.04 | 4.35 | 3.37 |
| 1.74 | 1.44 | 1.52 | 2.05 | 2.06 | 2.29 | |
| 2.39 | 2.25 | 2.25 | 4.22 | 5.97 | 6.32 | |
| 0.51 | 0.59 | 0.62 | 0.33 | 0.45 | 0.52 | |
| 0.29 | 0.36 | 0.41 | 0.19 | 0.26 | 0.26 | |
|
| ||||||
| # Parameters (↓) | 2.86M | 12.64M | 95M | 31M | 1.4B | 1.4B |
| Sampling time (↓) | ~0.14s | ~570s | ~10,000s | ~240s | ~70s | ~70s |
A.4. Supplementary Figures
Figure 8: DynaProt-M predicted RMSF correlations.
Visualized test set examples of predicted RMSF per residue (derived from the predicted marginal Gaussians) compared to ground truth RMSF derived from MD trajectories. Pearson correlation coefficient between predicted and ground truth RMSF is reported.
References
- Abramson Josh, Adler Jonas, Dunger Jack, Evans Richard, Green Tim, Pritzel Alexander, Ronneberger Olaf, Willmore Lindsay, Ballard Andrew J, Bambrick Joshua, et al. Accurate structure prediction of biomolecular interactions with alphafold 3. Nature, 630(8016):493–500, 2024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bhatia Rajendra, Jain Tanvi, and Lim Yongdo. On the Bures–Wasserstein distance between positive definite matrices. Expositiones Mathematicae, 37(2):165–191, 2019. [Google Scholar]
- Burley Stephen K, Berman Helen M, Kleywegt Gerard J, Markley John L, Nakamura Haruki, and Velankar Sameer. Protein data bank (pdb): the single global macromolecular structure archive. Protein crystallography: methods and protocols, pages 627–641, 2017. [Google Scholar]
- Childers Matthew Carterand Daggett Valerie. Insights from molecular dynamics simulations for computational protein design. Molecular systems design & engineering, 2(1):9–33, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Comitani Federico and Gervasio Francesco Luigi. Exploring cryptic pockets formation in targets of pharmaceutical interest with swish. Journal of chemical theory and computation, 14(6):3321–3331, 2018. [DOI] [PubMed] [Google Scholar]
- Cui Qiang and Bahar Ivet. Normal mode analysis: theory and applications to biological and chemical systems. CRC press, 2005. [Google Scholar]
- Del Alamo Diego, Sala Davide, Mchaourab Hassane S, and Meiler Jens. Sampling alternative conformational states of transporters and receptors with alphafold2. Elife, 11:e75751, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Glorot Xavier, Bordes Antoine, and Bengio Yoshua. Deep sparse rectifier neural networks. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, pages 315–323. JMLR Workshop and Conference Proceedings, 2011. [Google Scholar]
- Guo Amy B, Akpinaroglu Deniz, Stephens Christina A, Grabe Michael, Smith Colin A, Kelly Mark JS, and Kortemme Tanja. Deep learning–guided design of dynamic proteins. Science, 388(6749): eadr7094, 2025. [DOI] [PubMed] [Google Scholar]
- Hollingsworth Scott Aand Dror Ron O. Molecular dynamics simulation for all. Neuron, 99(6): 1129–1143, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hollingsworth Scott A, Kelly Brendan, Valant Celine, Michaelis Jordan Arthur, Mastromihalis Olivia, Thompson Geoff, Venkatakrishnan AJ, Hertig Samuel, Scammells Peter J, Sexton Patrick M, et al. Cryptic pocket formation underlies allosteric modulator selectivity at muscarinic gpcrs. Nature Communications, 10(1):3289, 2019. [Google Scholar]
- Hou Chao and Shen Yufeng. Seqdance: A protein language model for representing protein dynamic properties. bioRxiv, 2024. [Google Scholar]
- Huang Zhiwu, Wang Ruiping, Shan Shiguang, Li Xianqiu, and Chen Xilin. Log-euclidean metric learning on symmetric positive definite manifold with application to image set classification. In International conference on machine learning, pages 720–729. PMLR, 2015. [Google Scholar]
- Jeffreys Harold. The theory of probability. OuP Oxford, 1998. [Google Scholar]
- Jing Bowen, Berger Bonnie, and Jaakkola Tommi. AlphaFold meets flow matching for generating protein ensembles. arXiv preprint arXiv:2402.04845, 2024. [Google Scholar]
- Jumper John, Evans Richard, Pritzel Alexander, Green Tim, Figurnov Michael, Ronneberger Olaf, Tunyasuvunakool Kathryn, Bates Russ, Žídek Augustin, Potapenko Anna, et al. Highly accurate protein structure prediction with AlphaFold. Nature, 596(7873):583–589, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kouba Petr, Planas-Iglesias Joan, Damborsky Jiri, Sedlar Jiri, Mazurenko Stanislav, and Sivic Josef. Learning to engineer protein flexibility. arXiv preprint arXiv:2412.18275, 2024. [Google Scholar]
- Kullback Solomon and Leibler Richard A. On information and sufficiency. The annals of mathematical statistics, 22(1):79–86, 1951. [Google Scholar]
- Lewis Sarah, Hempel Tim, Jiménez-Luna José, Gastegger Michael, Xie Yu, Foong Andrew YK, García Satorras Victor, Abdin Osama, Veeling Bastiaan S, Zaporozhets Iryna, et al. Scalable emulation of protein equilibrium ensembles with generative deep learning. bioRxiv, pages 2024–12, 2024. [Google Scholar]
- Lombard Valentin, Timsit Dan, Grudinin Sergei, and Laine Elodie. Seamoon: Prediction of molecular motions based on language models. bioRxiv, pages 2024–09, 2024. [Google Scholar]
- Lu Jiarui, Chen Xiaoyin, Lu Stephen Zhewen, Shi Chence, Guo Hongyu, Bengio Yoshua, and Tang Jian. Structure language models for protein conformation generation. arXiv preprint arXiv:2410.18403, 2024. [Google Scholar]
- Ma Jianpeng. Usefulness and limitations of normal mode analysis in modeling dynamics of biomolecular complexes. Structure, 13(3):373–380, 2005. [DOI] [PubMed] [Google Scholar]
- McCammon JA. Protein dynamics. Reports on Progress in Physics, 47(1):1, 1984. [Google Scholar]
- Meller Artur, Ward Michael, Borowsky Jonathan, Kshirsagar Meghana, Lotthammer Jeffrey M., Oviedo Felipe, Ferres Juan Lavista, and Bowman Gregory R.. Predicting locations of cryptic pockets from single protein structures using the PocketMiner graph neural network. 14(1): 1177. ISSN 2041–1723. doi: 10.1038/s41467-023-36699-3. URL https://doi.org/10.1038/s41467-023-36699-3. [DOI] [Google Scholar]
- Mou Minyue, Yang Weicheng, Huang Guangyi, Yang Xiaoyan, Zhang Xiao, Kandegama Wasala Mudiyanselage Wishwajith Wickramabahu, Ashby Charles R Jr, Hao Gefei, and Gao Yangyang. The discovery of cryptic pockets increases the druggability of “undruggable” proteins. Medicinal Research Reviews, 2025. [Google Scholar]
- Rotkiewicz Piotr and Skolnick Jeffrey. Fast procedure for reconstruction of full-atom protein models from reduced representations. Journal of computational chemistry, 29(9):1460–1465, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Shaw David E, Maragakis Paul, Lindorff-Larsen Kresten, Piana Stefano, Dror Ron O, Eastwood Michael P, Bank Joseph A, Jumper John M, Salmon John K, Shan Yibing, et al. Atomic-level characterization of the structural dynamics of proteins. Science, 330(6002):341–346, 2010. [DOI] [PubMed] [Google Scholar]
- Skjaerven Lars, Hollup Siv M, and Reuter Nathalie. Normal mode analysis for proteins. Journal of Molecular Structure: THEOCHEM, 898(1–3):42–48, 2009. [Google Scholar]
- Stein Richard Aand Mchaourab Hassane S. Speach_af: Sampling protein ensembles and conformational heterogeneity with alphafold2. PLOS Computational Biology, 18(8):e1010483, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Meersche Yann Vander, Cretin Gabriel, Gheeraert Aria, Gelly Jean-Christophe, and Galochkina Tatiana. ATLAS: protein flexibility description from atomistic molecular dynamics simulations. Nucleic Acids Research, 52(D1):D384–D392, 2024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vemulapalli Raviteja and Jacobs David W. Riemannian metric learning for symmetric positive definite matrices. arXiv preprint arXiv:1501.02393, 2015. [Google Scholar]
- Wang Yan, Wang Lihao, Shen Yuning, Wang Yiqun, Yuan Huizhuo, Wu Yue, and Gu Quanquan. Protein conformation generation via force-guided se (3) diffusion models. arXiv preprint arXiv:2403.14088, 2024. [Google Scholar]
- Wayment-Steele Hannah K, Ojoawo Adedolapo, Otten Renee, Apitz Julia M, Pitsawong Warintra, Hömberger Marc, Ovchinnikov Sergey, Colwell Lucy, and Kern Dorothee. Predicting multiple conformations via sequence clustering and alphafold2. Nature, 625(7996):832–839, 2024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wayment-Steele Hannah K, Nesr Gina El, Hettiarachchi Ramith, Kariyawasam Hasindu, Ovchinnikov Sergey, and Kern Dorothee. Learning millisecond protein dynamics from what is missing in nmr spectra. bioRxiv, pages 2025–03, 2025. [Google Scholar]
- Yu Edward Wand Koshland Daniel E Jr. Propagating conformational changes over long (and short) distances in proteins. Proceedings of the National Academy of Sciences, 98(17):9517–9520, 2001. [Google Scholar]
- Zhang Mingyang, Chen Ting, Lu Xun, Lan Xiaobing, Chen Ziqiang, and Lu Shaoyong. G protein-coupled receptors (gpcrs): advances in structures, mechanisms and drug discovery. Signal Transduction and Targeted Therapy, 9(1):88, 2024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang She, Krieger James M, Zhang Yan, Kaya Cihan, Kaynak Burak, Mikulska-Ruminska Karolina, Doruker Pemra, Li Hongchun, and Bahar Ivet. Prody 2.0: increased scale and scope after 10 years of protein dynamics modelling with python. Bioinformatics, 37(20):3657–3659, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zinovjev Kirill, Guénon Paul, Ramos-Guzmán Carlos A, Javier Ruiz-Pernía J, Laage Damien, and Tuñón Iñaki. Activation and friction in enzymatic loop opening and closing dynamics. Nature Communications, 15(1):2490, 2024. [Google Scholar]








