Skip to main content

This is a preprint.

It has not yet been peer reviewed by a journal.

The National Library of Medicine is running a pilot to include preprints that result from research funded by NIH in PMC and PubMed.

ArXiv logoLink to ArXiv
[Preprint]. 2024 Dec 31:arXiv:2409.19838v2. [Version 2]

Using pretrained graph neural networks with token mixers as geometric featurizers for conformational dynamics

Zihan Pengmei, Chatipat Lorpaiboon, Spencer C Guo, Jonathan Weare, Aaron R Dinner
PMCID: PMC11722521  PMID: 39801625

Abstract

Identifying informative low-dimensional features that characterize dynamics in molecular simulations remains a challenge, often requiring extensive manual tuning and system-specific knowledge. Here, we introduce geom2vec, in which pretrained graph neural networks (GNNs) are used as universal geometric featurizers. By pretraining equivariant GNNs on a large dataset of molecular conformations with a self-supervised denoising objective, we obtain transferable structural representations that are useful for learning conformational dynamics without further fine-tuning. We show how the learned GNN representations can capture interpretable relationships between structural units (tokens) by combining them with expressive token mixers. Importantly, decoupling training the GNNs from training for downstream tasks enables analysis of larger molecular graphs (such as small proteins at all-atom resolution) with limited computational resources. In these ways, geom2vec eliminates the need for manual feature selection and increases the robustness of simulation analyses.

Full Text

The Full Text of this preprint is available as a PDF (7.4 MB). The Web version will be available soon.

13 pages, 8 figures, supporting information appended


Articles from ArXiv are provided here courtesy of arXiv

RESOURCES