Skip to main content

This is a preprint.

It has not yet been peer reviewed by a journal.

The National Library of Medicine is running a pilot to include preprints that result from research funded by NIH in PMC and PubMed.

ArXiv logoLink to ArXiv
[Preprint]. 2023 May 1:arXiv:2302.04313v3. Originally published 2023 Feb 8. [Version 3]

Geometry-Complete Diffusion for 3D Molecule Generation

Alex Morehead 1, Jianlin Cheng 1
PMCID: PMC9934735  PMID: 36798459

Abstract

Denoising diffusion probabilistic models (DDPMs) (Ho et al. (2020)) have recently taken the field of generative modeling by storm, pioneering new state-of-the-art results in disciplines such as computer vision and computational biology for diverse tasks ranging from text-guided image generation (Ramesh et al. (2022); Saharia et al. (2022); Rombach et al. (2022)) to structure-guided protein design (Ingraham et al. (2022); Watson et al. (2022)). Along this latter line of research, methods such as those of Hoogeboom et al. (2022) have been proposed for generating 3D molecules using equivariant graph neural networks (GNNs) within a DDPM framework. Toward this end, we propose GCDM, a geometry-complete diffusion model that achieves new state-of-the-art results for 3D molecule diffusion generation and optimization by leveraging the representation learning strengths offered by GNNs that perform geometry-complete message-passing. Our results with GCDM also offer preliminary insights into how physical inductive biases impact the generative dynamics of molecular DDPMs. The source code, data, and instructions to train new models or reproduce our results are freely available at https://github.com/BioinfoMachineLearning/Bio-Diffusion.

1. Introduction

Generative modeling has recently been experiencing a renaissance in modeling efforts driven largely by denoising diffusion probabilistic models (DDPMs). At a high level, DDPMs are trained by learning how to denoise a noisy version of an input example. For example, in the context of computer vision, Gaussian noise may be successively added to an input image with the goals of a DDPM in mind. We would then desire for a generative model of images to learn how to successfully distinguish between the original input image’s feature signal and the noise signal added to the image thereafter. If a model can achieve such outcomes, we can use the model to generate novel images by first sampling multivariate Gaussian noise and then iteratively removing, from the current state of the image, the noise predicted by our model. This classic formulation of DDPMs has achieved significant results in the space of image generation (Rombach et al. (2022)), audio synthesis (Kong et al. (2020)), and even meta-learning by learning how to conditionally generate neural network checkpoints (Peebles et al. (2022)). Furthermore, such an approach to generative modeling has expanded its reach to encompass scientific disciplines such as computational biology (Anand & Achim (2022)), computational chemistry (Xu et al. (2022)), and even computational physics (Mudur & Finkbeiner (2022)).

Concurrently, the field of geometric deep learning (GDL) (Bronstein et al. (2021)) has seen a sizeable increase in research interest lately, driven largely by theoretical advances within the discipline (Joshi et al. (2023)) as well as by novel applications of such methodology (Stärk et al. (2022)). Notably, such applications even include what is considered by many researchers to be a solution to the problem of predicting 3D protein structures from their corresponding amino acid sequences (Jumper et al. (2021)). Such an outcome arose, in part, from recent advances in sequence-based language modeling efforts (Vaswani et al. (2017)) as well as from innovations in equivariant neural network modeling (Thomas et al. (2018)).

With such diverse, successful use cases of DDPMs and GDL in mind, in this work, we explore the intersection of geometric graph representation learning and DDPMs to answer the following questions.

  • What is the impact of geometric representation learning on DDPMs designed to generate 3D molecular data?

  • What are the limitations of current equivariant graph neural networks empowering contemporary molecular DDPMs?

  • What role do physical inductive biases play within the generative denoising of molecular DDPMs?

2. Related Work

Generative Modeling.

The field of deep generative modeling (Ruthotto & Haber (2021)) has pioneered a variety of techniques by which to train deep neural networks to create new content similar to that of an existing data repository (e.g., a text dataset of English sentences). Language models such as GPT-3 and ChatGPT (Brown et al. (2020); Schulman et al. (2022)) have become known as hallmark examples of successful generative modeling of text data. In the domains of computer vision and computational biology, techniques such as latent diffusion (Rombach et al. (2022)) and equivariant graph diffusion (Luo et al. (2022)) have established some of the latest state-of-the-art results in generative modeling of images (Tang et al. (2022)) and biomolecules (Hoogeboom et al. (2022)) such as proteins (Anand & Achim (2022); Yim et al. (2023)), respectively.

Geometric Deep Learning.

Data residing in a geometric or physical space (e.g., R3) can be processed by machine learning algorithms in a plethora of ways. Subsequently, in recent years, the field of geometric deep learning has become known for its proficiency in introducing powerful new deep learning methods designed specifically to process geometric data (Cao et al. (2020)). Examples of popular GDL algorithms include convolutional neural networks designed for working with image data (LeCun et al. (1995)), recurrent neural networks for processing sequence-based data (Medsker & Jain (1999)), and graph neural networks for handling graph-structured model inputs (Zhou et al. (2020)).

Equivariant Neural Networks.

To process geometric data efficiently, however, recent GDL research (Cohen & Welling (2016); Bronstein et al. (2021); Bulusu et al. (2021)) has specifically shown that designing one’s machine learning algorithm to be equivariant to the symmetry groups the input data points naturally respect (e.g., 3D rotation symmetries) often helps such an algorithm generalize to datasets beyond those used for its cross-validation (e.g., training and testing datasets). As a particularly relevant example of a neural network that is equivariant to several important and common symmetry groups of geometric data, equivariant graph neural networks (Fuchs et al. (2020); Satorras et al. (2021b); Kofinas et al. (2021); Morehead & Cheng (2022)) that are translation and rotation equivariant to inputs residing in R3 have become known as hallmark examples of geometric deep learning algorithms that generalize remarkably well to new inputs and require notably fewer training iterations to converge.

Representation Learning of Scientific Data.

Scientific data, in particular, requires careful consideration in the context of representation learning. As much scientific data contains within it a notion of geometry or latent structure, equivariance has become a key algorithmic component for processing such inputs as well (Han et al. (2022)). Moreover, equivariant graph representation learning algorithms have recently become a de facto methodology for processing scientific data of many shapes and origins Musaelian et al. (2022); Batzner et al. (2022).

Contributions.

In this work, we connect ideas at the forefront of GDL and generative modeling to advance the state-of-the-art (SOTA) for 3D molecule generation. In detail, we provide the following contributions.

  • We introduce the Geometry-Complete Diffusion Model (GCDM) which establishes new SOTA results for unconditional and conditional 3D molecule generation on the QM9 dataset and for unconditional 3D molecule generation on the larger GEOM-Drugs dataset.

  • We investigate the impact of geometric message-passing on the behavior and performance of DDPMs trained to generate 3D molecular data.

  • Our experiments demonstrate the importance of incorporating physical inductive biases such as molecular chirality within DDPM denoising neural networks when training them on data from physical domains.

3. Methods

3.1. Problem Setting

In the context of this work, our goal is to generate new 3D molecules either ab initio or to capture a specific molecular property. We represent a molecular point cloud as a fully-connected 3D graph 𝒢=(𝒱,) with 𝒱 and representing the graph’s set of nodes and set of edges, respectively, and N=|𝒱| and E=|| representing the number of nodes and the number of edges in the graph, respectively. In addition, X=x1,x2,,xNRN×3 represents the respective Cartesian coordinates for each node (i.e., atom). Each node in 𝒢 is described by scalar features HRN×h and m vector-valued features χRN×(m×3). Likewise, each edge in 𝒢 is described by scalar features ERE×e and x vector-valued features ξRE×(x×3). Then, let =[X,H] represent the molecules our method is to generate, where [,] denotes the concatenation of two variables. Important to note is that the input features H and E are invariant to 3D rotations, reflections, and translations, whereas the input features X,χ, and ξ are equivariant to 3D rotations and reflections. In particular, we describe a denoising neural network Φ as SE(3)-equivariant (i.e., 3D rotation and translation-equivariant) if it satisfies the following constraint on its outputs (denoted by ◻′):

Definition 3.1. (SE(3) Equivariance).

Given (H,E,X,χ,ξ)=ΦH,E,X,χ,ξ, we have (H,E,QXT+g,QχT,QξT)=ΦH,E,QXT+g,QχT,QξT,QSO3,gR3×1.

3.2. Overview of GCDM

We will now introduce GCDM, a new Geometry-Complete SE(3)-Equivariant Diffusion Model. In particular, we will describe how GCDM defines a joint noising process on equivariant atom coordinates x and invariant atom types h to produce a noisy representation z=z(x),z(h) and then learns a generative denoising process using GCPNet (Morehead & Cheng (2022)). As we will show in subsequent sections, GCPNet is a desirable architecture for the task of denoising 3D graph inputs in that it contains two distinct feature channels for scalar and vector features, respectively, and supports geometry-complete and chirality-aware message-passing by embedding geometry information-complete local frames for each node (Barron (1986)). Moreover, in our subsequent experiments, we demonstrate that this enables GCPNet to learn more useful equivariant graph representations for generative modeling of 3D molecules.

As an extension of the DDPM framework (Ho et al. (2020)) outlined in Appendix A.1, GCDM is designed to generate molecules in 3D while maintaining SE(3) equivariance, in contrast to previous methods that generate molecules solely in 2D (Jin et al. (2018)) or other dimensionalities (Segler et al. (2018)). GCDM generates molecules by directly placing atoms in continuous 3D space and assigning them discrete types, which is accomplished by modeling forward and reverse diffusion processes, respectively:

q(z1:Tz0)=t=1Tq(ztzt1) (1)
pΦ(z0:T1zT)=t=1TpΦ(zt1zt) (2)

Overall, these processes describe a latent variable model pΦz0=pΦz0:Tdz1:T given a sequence of latent variables z0,z1,,zT matching the dimensionality of the data pz0. As illustrated in Figure 1, the forward process (directed from right to left) iteratively adds noise to an input, and the learned reverse process (directed from left to right) iteratively denoises a noisy input to generate new examples from the original data distribution. We will now proceed to formulate GCDM’s joint diffusion process and its remaining practical details.

Figure 1:

Figure 1:

A framework overview for our proposed Geometry-Complete Diffusion Model (GCDM). Our framework consists of (i.) a graph (topology) definition process, (ii.) a GCPNet-based graph neural network for 3D graph representation learning, (iii.) denoising of 3D input graphs using GCPNet, and (iv.) application of a trained GCPNet denoising network for 3D molecule generation. Zoom in for the best viewing experience.

3.3. Joint Molecular Diffusion

Recall that our model’s molecular graph inputs, 𝒢, associate with each node a 3D position xiR3 and a feature vector hiRh. By way of adding random noise to these model inputs at each time step t and using a fixed, Markov chain variance schedule σ12,σ22,,σT2, we can define a joint molecular diffusion process for equivariant atom coordinates x and invariant atom types h as the product of two distributions (Hoogeboom et al. (2022)):

qztzt-1=𝒩x(zt(x)αtzt-1(x),σt2I)𝒩h(zthαtzt-1h,σt2I). (3)

where the first distribution, 𝒩x, represents the noised node coordinates, the second distribution, 𝒩h, represents the noised node features, and αt=1-σt2 following the variance preserving process of Ho et al. (2020). Using 𝒩xh as concise notation to denote the product of two normal distributions, we can further simplify Eq. 3 as:

qztzt-1=𝒩xhztαtzt-1,σt2I. (4)

With αts=αt/αs and σts2=σt2-αtsσs2 for any t>s, we can directly obtain the noisy data distribution qztz0 at any time step t:

qztz0=𝒩xh(ztαt0z0,σt02I). (5)

Bayes Theorem then tells us that if we then define μtszt,z0 and σts as

μtszt,z0=αsσts2σt2z0+αtsσs2σt2ztandσts=σtsσsσt,

we have that the inverse of the noising process, the true denoising process, is given by the posterior of the transitions conditioned on z0, a process that is also Gaussian (Hoogeboom et al. (2022)):

qzszt,z0=𝒩zsμtszt,z0,σts2I. (6)

3.4. Geometry-Complete Parametrization of the Equivariant Reverse Process

Noise parametrization.

We now need to define our learned generative reverse process that denoises pure noise into realistic examples from the original data distribution. Towards this end, we can directly use the noise posteriors qzszt,z0 of Eq. 16 with z0(=[x,h]). However, to do so, we must replace the input variables x and h with the approximations xˆ and hˆ predicted by our denoising neural network Φ:

pΦzszt=𝒩xhzsμΦtszt,z˜0,σts2I, (7)

where the values for z˜0=[xˆ,hˆ] depend on zt,t, and our denoising neural network Φ.

In the context of diffusion models, many different parametrizations of μΦtszt,z˜0 are possible. Prior works have found that it is often easier to optimize a diffusion model using a noise parametrization to predict the noise ϵˆ. In this work, we use such a parametrization to predict ϵˆ=[ϵˆ(x),ϵˆ(h)], which represents the noise individually added to xˆ and hˆ. We can then use the predicted ϵˆ to derive:

z˜0=[xˆ,hˆ]=zt/αt-ϵˆtσt/αt. (8)

Invariant likelihood.

Ideally, we desire for a 3D molecular diffusion model to assign the same likelihood to a generated molecule even after arbitrarily rotating or translating it in 3D space. To ensure our model achieves this desirable property for pΦz0, we can leverage the insight that an invariant distribution composed of an equivariant transition function yields an invariant distribution (Satorras et al. (2021a); Xu et al. (2022); Hoogeboom et al. (2022)). Moreover, to address the translation invariance issue raised by Satorras et al. (2021a) in the context of handling a distribution over 3D coordinates, we adopt the zero center of gravity trick proposed by Xu et al. (2022) to define 𝒩x as a normal distribution on the subspace defined by ixi=0. In contrast, to handle node features hi that are rotation and translation-invariant, we can instead use a conventional normal distribution 𝒩. As such, if we parametrize our transition function pΦ using an SE(3)-equivariant neural network after using the zero center of gravity trick of Xu et al. (2022), our model will have achieved the desired likelihood invariance property.

Geometry-completeness.

Furthermore, in this work, we postulate that certain types of geometric neural networks serve as more effective 3D graph denoising functions for molecular DDPMs. We formalize this notion as follows.

Proposition 3.2. (Geometry-Complete Denoising).

Geometric neural networks that achieve geometry-completeness are principally more capable of denoising noisy 3D molecular graph inputs, in that geometry-complete methods encode local reference frames under which the directions of arbitrary global forces can be mapped.

This proposition comes as an extension of the definition of geometry-completeness from Morehead & Cheng (2022). An intuition for its implications on molecular diffusion models is that geometry-complete networks should be able to more effectively learn the gradients of data distributions (Ho et al. (2020)) in which a global force field is present, as is typically the case with 3D molecules (Du et al. (2022)). As a complement to understanding the theoretical benefits offered to geometry-complete networks, we support this claim through specific ablation studies in Section 4.1.

GCPNets.

Inspired by their recent success in modeling 3D molecular structures with geometry-complete message-passing, as mentioned previously, we will parametrize pΦ using an extended version of Geometry-Complete Perceptron Networks (GCPNets) as introduced by Morehead & Cheng (2022). GCPNet is a geometry-complete graph neural network that is equivariant to SE(3) transformations of its graph inputs and, as such, satisfies our SE(3) equivariance constraint (3.1) and maps nicely to the context of Proposition 3.2.

In this setting, with (hiH,χiχ,eijE,ξijξ), GCPNet consists of a composition of Geometry-Complete Graph Convolution (GCPConv) layers hil,χil,xil=GCPConv[hil-1,χil-1,(eijl-1,ξijl-1),xil-1,ij] which are defined as:

nil=ϕl(nil-1,𝒜j𝒩(i)Ωωl(nil-1,njl-1,eijl-1,ξijl-1,ij)), (9)

where nil=hil,χil;ϕl is a trainable function; l signifies the representation depth of the network; 𝒜 is a permutation-invariant aggregation function; Ωω represents a message-passing function corresponding to the ω-th GCP message-passing layer; and node i’s geometry-complete local frames are ijt=(aijt,bijt,cijt), with aijt=xit-xjtxit-xjt,bijt=xit×xjtxit×xjt, and cijt=aijt×bijt, respectively.

Lastly, if one desires to update the coordinate representations of each node in 𝒢, as we do in the context of 3D molecule generation, GCPConv provides a simple, SE(3)-equivariant method to do so using a dedicated GCP module as follows:

hpil,χpil=GCPplnil,ij (10)
xil=xil-1+χpil,whereχpilR1×3, (11)

where GCPl,ij is defined as in (Morehead & Cheng (2022)) to provide chirality-aware rotation and translation-invariant updates to hi and rotation equivariant updates to χi following centralization of the input point cloud’s coordinates X (Du et al. (2022)). The effect of using feature updates to χi to update xi is, after decentralizing X following the final GCPConv layer, that updates to xi then become SE(3)-equivariant. As such, all the transformations described above collectively satisfy the required equivariance constraint in Def. 3.1. Therefore, in adapting GCPNet as its 3D graph denoiser, GCDM achieves SE(3) equivariance, geometry-completeness, and likelihood invariance altogether. Moreover, following recent results from Du et al. (2023), GCDM is subsequently capable of encoding local geometric substructures as well as encoding equivariant transitions (e.g., messages) between geometric frames in a computationally efficient manner, which provides grounds for the theoretical soundness of the proposed generative modeling method.

3.5. Optimization Objective

Following previous works on diffusion models (Ho et al. (2020); Hoogeboom et al. (2022); Wu et al. (2022)), our noise parametrization chosen for GCDM yields the following model training objective:

t=Eϵt𝒩xh(0,1)12w(t)ϵt-ϵˆt2, (12)

where ϵˆt is our network’s noise prediction as described above and where we empirically choose to set w(t)=1 for the best possible generation results compared to w(t)=(1-SNR(t-1)/SNR(t)) with SNR(t)=αt2/σt2. Additionally, GCDM permits a negative log-likelihood computation using the same optimization terms as Hoogeboom et al. (2022), for which we refer interested readers to Appendices A.1 and A.2 for further details. Lastly, for remaining technical details regarding GCDM’s training and sampling procedures, we refer readers to Appendix A.4.

4. Experiments

4.1. Unconditional 3D Molecule Generation - QM9

The QM9 dataset (Ramakrishnan et al. (2014)) contains molecular property descriptions and 3D atom coordinates for 130k small molecules. Each molecule in QM9 can contain up to 9 heavy atoms, that is, 29 atoms when including hydrogens. For the task of 3D molecule generation, we train GCDM to unconditionally generate molecules by producing atom types (H, C, N, O, and F), integer-valued atom charges, and 3D coordinates for each of the molecules’ atoms. Following Anderson et al. (2019), we split QM9 into training, validation, and test partitions consisting of 100k, 18k, and 13k molecule examples, respectively.

Metrics.

We adopt the scoring conventions of Satorras et al. (2021a) by using the distance between atom pairs and their respective atom types to predict bond types (single, double, triple, or none) for all but one baseline method (i.e., E-NF). Subsequently, we measure the proportion of generated atoms that have the right valency (atom stability) and the proportion of generated molecules for which all atoms are stable (molecule stability). To offer additional insights into each method’s behavior for 3D molecule generation, we also report the validity of a generated molecule as determined by RDKit (Landrum et al. (2013)) and the uniqueness of the generated molecules overall.

Baselines.

Besides including a reference point for molecule quality metrics using QM9 itself (i.e., Data), we compare GCDM to three existing E(3)-equivariant models: G-Schnet (Gebauer et al. (2019)), Equivariant Normalizing Flows (E-NF) (Satorras et al. (2021a)), and Equivariant Diffusion Models (EDM) (Hoogeboom et al. (2022)). For each of these three models, we report their results as reported in Hoogeboom et al. (2022). For comparison with models for this task that are not equivariant, we also report results from Hoogeboom et al. (2022) for Graph Diffusion Models (GDM) trained with random data rotations (GDM-aug) and without them (GDM). To the best of our knowledge, the force-guided molecule generation methods of Wu et al. (2022) (i.e., Bridge and Bridge + Force) are the most recent and performant state-of-the-art open-source methods for 3D molecule generation, so we include their results for this experiment as well.

We further include two GCDM ablation models to more closely analyze the impact of certain key model components within GCDM. These two ablation models include GCDM without geometry-complete local frames ij (i.e., GCDM w/o Frames) and GCDM without scalar message attention (SMA) applied to each edge message (i.e., GCDM w/o SMA). For SMA, mij=eijmij, where mij represents the scalar messages learned by GCPNet during message-passing and eij represents a 1 if an edge exists between nodes i and j (and a 0 otherwise) via eijϕinfmij. Here, ϕinf:Re[0,1]1 resembles a linear layer followed by a sigmoid function Satorras et al. (2021b). All GCDM models train on QM9 for approximately 1,000 epochs using 9 GCPConv layers; SiLU activations (Elfwing et al. (2018)); 256 and 64 scalar node and edge hidden features, respectively; and 32 and 16 vector-valued node and edge features, respectively. All GCDM models are also trained using the AdamW optimizer (Loshchilov & Hutter (2017)) with a batch size of 64, a learning rate of 10−4, and a weight decay rate of 10−12.

Results.

In Table 1, we see that GCDM matches or outperforms all previous methods (E-NF, G-Schnet, EDM, Bridge, and Bridge + Force) as well as their non-equivariant counterparts (GDM and GDM-aug) for all metrics, with generated samples shown in Figure 2. In particular, GCDM generates the highest percentage of valid and unique molecules compared to all other methods, improving upon previous SOTA results in such measures by 3%. GCDM also advances the SOTA results in terms of negative log-likelihood (NLL) and molecule stability by 54% and 1%, respectively. Moreover, our ablation of SMA within GCDM demonstrates that GCDM heavily relies on being able to perform a lightweight version of self-attention (Vaswani et al. (2017)) in the form of fully-connected attentive message-passing to generate stable 3D molecules. This finding suggests interesting avenues for future research into the impact of different kinds of attention-based geometric message-passing (e.g., type-2 tensor message-passing) on the performance of diffusion models for 3D molecular generation tasks.

Table 1:

Comparison of GCPNet with baseline methods for 3D molecule generation. The results are reported in terms of the negative log-likelihood (NLL) − log p(x, h, N), atom stability, molecule stability, validity, and uniqueness of 10,000 samples drawn from each model, with standard deviations for each model across three runs on QM9. The top-1 (best) results for this task are in bold, and the second-best results are underlined.

Type Method NLL ↓ Atoms Stable (%) ↑ Mol Stable (%) ↑ Valid (%) ↑ Valid and Unique (%) ↑

Normalizing Flow E-NF −59.7 85.0 4.9 40.2 39.4

Graph Autoregression G-Schnet n/a 95.7 68.1 85.5 80.3

DDPM GDM −94.7 97.0 63.2 N/A N/A
GDM-aug −92.5 97.6 71.6 90.4 89.5
EDM −110.7 ± 1.5 98.7 ± 0.1 82.0 ± 0.4 91.9 ± 0.5 90.7 ± 0.6
Bridge N/A 98.7 ± 0.1 81.8 ± 0.2 N/A 90.2
Bridge + Force N/A 98.8 ± 0.1 84.6 ± 0.3 n/a 90.7

DDPM - Ours GCDM w/o Frames −162.3 ± 0.3 98.4 ± 0.0 81.7 ± 0.5 93.9 ± 0.1 92.7 ± 0.1
GCDM w/o SMA −131.3 ± 0.8 95.7 ± 0.1 51.7 ± 1.4 83.1 ± 1.7 82.8 ± 1.7
GCDM −171.0 ± 0.2 98.7 ± 0.0 85.7 ± 0.4 94.8 ± 0.2 93.3 ± 0.0

Data 99.0 95.2 97.7 97.7
Figure 2:

Figure 2:

3D molecules generated by GCDM for the QM9 dataset.

Specifically, it is interesting to note how much lower the NLL of GCDM is compared to that of EDM, the previous NLL-based SOTA method for 3D molecule generation, indicating the generative distribution that GCDM learns from QM9 likely contains much sharper peaks compared to EDM even within the context of similar diffusion modeling frameworks. A possible explanation for why GCDM can achieve such results over other equivariant methods such as EDM and Bridge is that GCDM performs geometric (and geometry-complete) message-passing over each 3D input graph to remove the noise present therein, whereas other methods learn solely using scalar features (Joshi et al. (2023)). Our ablation of geometry-complete local frames within GCDM supports this claim in that, compared to EDM, message-passing with type-1 tensor (i.e., vectors) appears to improve GCDM’s NLL over that of EDM by 47%, whereas with geometry-complete frames GCDM’s NLL improves by another 7%. In fact, with geometry-complete and chirality-sensitive frame embeddings, all of GCDM’s sample quality metrics improve to SOTA levels, providing support for Proposition 3.2.

4.2. Conditional 3D Molecule Generation - QM9

Baselines.

Towards conditional generation of 3D molecules, we compare GCDM to an existing E(3)-equivariant model, EDM (Hoogeboom et al. (2022)), as well as to two naive baselines: ”Naive (Upper-bound)” where a property classifier ϕc predicts molecular properties given a method’s generated 3D molecules and shuffled (i.e., random) property labels; and ”# Atoms” where one uses the numbers of atoms in a method’s generated 3D molecules to predict their molecular properties. For each baseline method, we report its mean absolute error in terms of molecular property prediction by an EGNN classifier ϕc Satorras et al. (2021b) as reported in Hoogeboom et al. (2022). For GCDM, we train each conditional model by conditioning it on one of six distinct molecular properties - α, gap, homo, lumo, μ, and Cv - for approximately 1,500 epochs using the QM9 validation split of Hoogeboom et al. (2022) as the model’s training dataset and the QM9 training split of Hoogeboom et al. (2022) as the corresponding EGNN classifier’s training dataset. Consequently, one can expect the gap between a method’s performance and that of ”QM9 (Lower-bound)” to decrease as the method generates molecules that more accurately model a given molecular property.

Results.

We see in Table 2 that GCDM outperforms all other methods in conditioning on a given molecular property, with conditionally-generated samples shown in Figure 3. In particular, GCDM improves upon the mean absolute error of the SOTA EDM method for all six molecular properties - α, gap, homo, lumo, μ, and Cv - by 29%, 8%, 3%, 18%, 24%, and 37%, respectively, demonstrating that, using geometry-complete message-passing, GCDM can more accurately model important molecular properties for 3D molecule generation. For interested readers, Section A.5.1 expands upon these conditional modeling results by introducing a novel means of repurposing diffusion generative models for 3D molecule optimization. Such an outcome, where we show GCDM requires only 20 denoising time steps to improve a 3D molecule’s molecular stability by as much as 6%, is to the best of our knowledge the first successful example of its kind for diffusion generative models.

Table 2:

Comparison of GCPNet with baseline methods for property-conditional 3D molecule generation. The results are reported in terms of the mean absolute error for molecular property prediction by an EGNN classifier ϕc on a QM9 subset, with results listed for GCDM-generated samples as well as two different baselines: ”Naive (Upper-bound)” and ”# Atoms”. The top-1 (best) results for this task are in bold, and the second-best results are underlined.

Task α Δϵ ϵHOMO ϵLUMO μ Cv
Units Bohr 3 meV meV meV D calmolK

Naive (Upper-bound) 9.01 1470 645 1457 1.616 6.857
# Atoms 3.86 866 426 813 1.053 1.971
EDM 2.76 655 356 584 1.111 1.101
GCDM 1.97 602 344 479 0.844 0.689
QM9 (Lower-bound) 0.10 64 39 36 0.043 0.040
Figure 3:

Figure 3:

3D molecules generated by GCDM using increasing values of α for the QM9 dataset.

4.3. Unconditional 3D Molecule Generation - GEOM-Drugs

The GEOM-Drugs dataset is a well-known source of large, 3D molecular conformers for downstream machine learning tasks. It contains 430k molecules, each with 44 atoms on average and with up to as many as 181 atoms. For this experiment, we collect the 30 lowest-energy conformers corresponding to a molecule and task each baseline method with generating new molecules with 3D positions and types for each constituent atom. Here, we also adopt the negative log-likelihood, atom stability, and molecule stability metrics as defined in Section 4.1 and train GCDM using the same hyperparameters as listed in Section 4.1 with the exception of training for approximately 75 epochs on GEOM-Drugs.

Baselines.

In this experiment, we compare GCDM to several state-of-the-art baseline methods for 3D molecule generation on GEOM-Drugs. Similar to our experiments on QM9, in addition to including a reference point for molecule quality metrics using GEOM-Drugs itself (i.e., Data), here we also compare against E-NF, GDM, GDM-aug, EDM, and Bridge with its variant Bridge + Force.

Results.

To start, Table 3 displays an interesting phenomenon: Due to the size of GEOM-Drugs’ molecules and the subsequent errors accumulated when estimating bond types based on inter-atom distances, the baseline results for the molecule stability metrics measured here (i.e., Data) are much lower than those collected for the QM9 dataset. Nonetheless, for GEOM-Drugs, GCDM improves upon SOTA negative log-likelihood results by 71% and upon SOTA atom stability results by 8%, with generated samples shown in Figure 4. Remarkably, to our best knowledge, GCDM is also the first deep learning model that can generate any stable large molecules according to the definitions of atomic and molecular stability in Section 4.1, demonstrating that GCDM can not only effectively generate large molecules but can also closely model the true distribution of stable molecules within GEOM-Drugs.

Table 3:

Comparison of GCPNet with baseline methods for 3D molecule generation. The results are reported in terms of each method’s negative log-likelihood, atom stability, and molecule stability with standard deviations across three runs on GEOM-Drugs, each drawing 10,000 samples from the model. The top-1 (best) results for this task are in bold, and the second-best results are underlined.

Type Method NLL ↓ Atoms Stable (%) ↑ Mol Stable (%) ↑

Normalizing Flow E-NF N/A 75.0 0.0

DDPM GDM −14.2 75.0 0.0
GDM-aug −58.3 77.7 0.0
EDM −137.1 81.3 0.0
Bridge N/A 81.0 ± 0.7 0.0
Bridge + Force N/A 82.4 ± 0.8 0.0

DDPM - Ours GCDM −234.3 89.0 ± 0.8 5.2 ± 1.1

Data 86.5 2.8
Figure 4:

Figure 4:

3D molecules generated by GCDM for the GEOM-Drugs dataset.

5. Conclusion

In this work, we introduced GCDM, an SE(3)-equivariant geometry-complete denoising diffusion probabilistic model for 3D molecule generation. While previous equivariant methods for this task have had difficulty establishing sizeable performance gains over non-equivariant methods for this task, GCDM establishes a clear performance advantage over all other methods, generating more realistic, stable, valid, unique, and property-specific 3D molecules compared to existing approaches. Although GCDM’s results here are promising, since the method falls into the traditional DDPM framework for generative modeling, using it to generate several thousands of large 3D molecules takes a notable amount of time (e.g., 15 minutes to generate 100 new large molecules). As such, future work in improving GCDM could involve introducing new time-efficient sampling algorithms for diffusion models (Song et al. (2020)) or even exploring other uses of GCDM in optimizing existing molecule’s geometry or chemical properties.

Acknowledgments

This work is partially supported by two NSF grants (DBI1759934 and IIS1763246), two NIH grants (R01GM093123 and R01GM146340), three DOE grants (DE-AR0001213, DE-SC0020400, and DE-SC0021303), and the computing allocation on the Summit compute cluster provided by the Oak Ridge Leadership Computing Facility, which is a DOE Office of Science User Facility supported under Contract DE-AC05– 00OR22725, granted in part by the Advanced Scientific Computing Research (ASCR) Leadership Computing Challenge (ALCC) program.

A. Appendix

A.1. Diffusion Models

Key to understanding our contributions in this work are denoising diffusion probabilistic models. As alluded to previously, once trained, DDPMs can generate new data of arbitrary shapes, sizes, formats, and geometries by learning to reverse a noising process acting on each model input. More precisely, for a given data point x, a diffusion process adds noise to x for time step t=0,1,,T to yield zt, a noisy representation of the input x at time step t. Such a process is defined by a multivariate Gaussian distribution:

qztx=𝒩ztαtxt,σt2I, (13)

where αtR+ regulates how much feature signal is retained and σt2 modulates how much feature noise is added to input x. Note that we typically model α as a function defined with smooth transitions from α0=1 to αT=0, where a special case of such a noising process, the variance preserving process (Sohl-Dickstein et al. (2015); Ho et al. (2020)), is defined by αt=1-σt2. To simplify notation, in this work, we define the feature signal-to-noise ratio as SNR(t)=αt2/σt2. Also interesting to note is that this diffusion process is Markovian in nature, indicating that we have transition distributions as follows:

qztzs=𝒩(ztαtszs,σts2I), (14)

for all t>s with αts=αt/αs and σts2=σt2-αts2σs2. In total, then, we can write the noising process as:

q(z0,z1,,zTx)=q(z0x)t=1Tq(ztzt1). (15)

If we then define μtsx,zt and σts as

μtsx,zt=αtsσs2σt2zt+αsσts2σt2xandσts=σtsσsσt,

we have that the inverse of the noising process, the true denoising process, is given by the posterior of the transitions conditioned on x, a process that is also Gaussian:

qzsx,zt=𝒩zsμtsx,zt,σtsI. (16)

The Generative Denoising Process.

In diffusion models, we define the generative process according to the true denoising process. However, for such a denoising process, we do not know the value of x a priori, so we typically approximate it as xˆ=ϕzt,t using a neural network ϕ. Doing so then lets us express the generative transition distribution pzszt as qzsxˆzt,t,zt. As a practical alternative to Eq. 16, we can represent this expression using our approximation for xˆ:

pzszt=𝒩zsμtsxˆ,zt,σts2I. (17)

If we choose to define s as s=t-1, then we can derive the variational lower bound on the log-likelihood of x given our generative model as:

logp(x)0+base+t=1Tt, (18)

where we note that 0=logpxz0 models the likelihood of the data given its noisy representation z0,base=-KLqzTxpzT models the difference between a standard normal distribution and the final latent variable qzTx, and

t=-KLqzsx,ztpzsztfort=1,2,,T.

Note that, in this formation of diffusion models, our neural network ϕ directly predicts xˆ. However, Ho et al. (2020) and others have found optimization of ϕ to be made much easier when instead predicting the Gaussian noise added to x to create xˆ. An intuition for how this changes the neural network’s learning dynamics is that, when predicting back the noise added to the model’s input, the network is being trained to more directly differentiate which part of zt corresponds to the input’s feature signal (i.e., the underlying data point x) and which part corresponds to added feature noise. In doing so, if we let zt=αtx+σtϵ, our neural network can then predict ϵˆ=ϕzt,t such that:

xˆ=1/αtzt-σt/αtϵˆ. (19)

Kingma et al. (2021) and others have since shown that, when parametrizing our denoising neural network in this way, the loss term t reduces to:

t=Eϵ𝒩(0,I)12(1-SNR(t-1)/SNR(t))ϵ-ϵˆ2 (20)

Note that, in practice, the loss term base should be close to zero when using a noising schedule defined such that αT0. Moreover, if and when α01 and x is a discrete value, we will find 0 to be close to zero as well.

A.2. Zeroth Likelihood Terms for GCDM Optimization Objective

For the zeroth likelihood terms corresponding to each type of input feature, we directly adopt the respective terms previously derived by Hoogeboom et al. (2022). Doing so enables a negative log-likelihood calculation for GCDM’s predictions. In particular, for integer node features, we adopt the zeroth likelihood term:

p(hz0(h))=h-12h+12𝒩(uz0(h),σ0)du, (21)

where we use the CDF of a standard normal distribution, Φ, to compute Eq. 21 as Φ((h+12-z0(h))/σ0)-Φ((h-12-z0(h))/σ0)1 for reasonable noise parameters α0 and σ0 (Hoogeboom et al. (2022)). For categorical node features, we instead use the zeroth likelihood term:

p(hz0(h))=Chp,p1-121+12𝒩(uz0h,σ0)du, (22)

where we normalize p to sum to one and where C is a categorical distribution (Hoogeboom et al. (2022)). Lastly, for continuous node positions, we adopt the zeroth likelihood term:

p(xz0(x))=𝒩xz0(x)/α0-σ0/α0ϵˆ0,σ02/α02I (23)

which gives rise to the log-likelihood component 0(x) as:

0(x)=Eϵ(x)𝒩x(0,I)logZ-1-12ϵx-ϕ(x)z0,02, (24)

where d=3 and the normalization constant Z=2πσ0/α0(N-1)d - in particular, its (N-1)d term - arises from the zero center of gravity trick mentioned in Section 3.4 (Hoogeboom et al. (2022)).

A.3. Diffusion Models and Equivariant Distributions

In the context of diffusion generative models of 3D data, one often desires for the marginal distribution p(x) of their denoising neural network to be an invariant distribution. Towards this end, we observe that a conditional distribution p(yx) is equivariant to the action of 3D rotations by meeting the criterion:

p(yx)=p(RyRx)forallorthogonalR. (25)

Moreover, a distribution is invariant to rotation transformations R when

p(y)=p(Ry)forallorthogonalR. (26)

As Köhler et al. (2020) and Xu et al. (2022) have collectively demonstrated, we know that if pzT is invariant and the neural network we use to parametrize pzt-1zt is equivariant, we have, as desired, that the marginal distribution p(x) of the denoising model is an invariant distribution.

A.4. Training and Sampling Procedures for GCDM

Equivariant Dynamics.

In this work, we use our previous definition of GCPNet in Section 3.4 to learn an SE(3)-equivariant dynamics function [ϵˆ(x),ϵˆ(h)]=ϕ(zt(x),zt(h),t) as:

ϵˆt(x),ϵˆt(h)=GCPNET(zt(x),[zt(h),t/T])-[zt(x),0], (27)

where we inform our denoising model of the current time step by concatenating t/T as an additional node feature and where we subtract the coordinate representation outputs of GCPNet from its coordinate representation inputs after subtracting from the coordinate representation outputs their collective center of gravity. With the parametrization in Eq. 8, GCDM subsequently achieves rotation equivariance on xˆi, thereby achieving a 3D translation and rotation-invariant marginal distribution p(x) as described in Appendix A.3.

Scaling Node Features.

In line with Hoogeboom et al. (2022), to improve the log-likelihood of our model’s generated samples, we find it useful to train and perform sampling with GCDM using scaled node feature inputs as [x,14h(categorical),110h(integer)].

Deriving The Number of Atoms.

Finally, to determine the number of atoms with which GCDM will generate a 3D molecule, we first sample Np(N), where p(N) denotes the categorical distribution of molecule sizes over GCDM’s training dataset. Then, we conclude by sampling x,hp(x,hN).

A.5. Additional Experiments

A.5.1. Property-Guided 3D Molecule Optimization - QM9

To evaluate whether molecular diffusion models can not only generate new 3D molecules but can also optimize existing molecules using molecular property guidance, we adopt the QM9 dataset for the following experiment. First, we use an unconditional diffusion model to generate 1,000 3D molecules with each baseline method, and then we provide these molecules to a separate property-conditional diffusion model for optimization of the molecules towards the conditional model’s respective property. This conditional model accepts these 3D molecules as intermediate states for 20 time steps of joint feature denoising, representing 20 time steps of property-guided optimization of the molecules’ atom types and 3D coordinates. Lastly, we repurpose our experimental setup from Section 4.2 to score these optimized molecules using an external property classifier model to evaluate (1) how much the optimized molecules’ predicted property values have been improved for the respective property (first metric) and (2) whether and how much the optimized molecules’ stability (as defined in Section 4.1) has been changed during optimization (second metric).

Baseline methods for this experiment include EDM (Hoogeboom et al. (2022)) and GCDM, where both methods use similar experimental setups for evaluation and where each generates 1,000 new molecules for optimization. Our baseline methods also include property-specificity and molecular stability measures of each method’s initial (unconditional) 3D molecules to demonstrate how much molecular diffusion models are able to modify or improve each method’s existing 3D molecules in terms of how property-specific and stable they are. As in Section 4.2, property specificity is measured in terms of the corresponding property classifier’s mean absolute error for a given molecule with a targeted property value. Molecular stability (i.e., Mol Stable (%)), here abbreviated at MS, is defined as in Section 4.1.

Table 4 showcases an interesting finding: molecular diffusion models for 3D molecule generation can effectively be repurposed as 3D molecular optimization algorithms with minimal modifications, with both baseline optimization methods offering positive refinement results. An interesting observation is that EDM-generated samples (i.e., ”EDM Samples”) seem to be easier for each baseline method to optimize in terms of molecular stability due to their initially-lower stability, while GCDM-generated samples (i.e., ”GCDM Samples”) appear to be more difficult for methods to refine as a large proportion of these molecules are already quite stable. Moreover, for groups of samples with lower average molecular stability, both baseline diffusion optimization methods seem to primarily improve molecules’ initial stability while also offering small (on average) improvements to their property specificity. In summary, Table 4 shows that GCDM achieves the best optimization results overall in both settings examined, that is, (1) for moderately-stable molecules and (2) for highly-stable molecules. In particular, when optimizing moderately-stable molecules for the molecular property μ, GCDM is simultaneously able to make the initial EDM molecules more property-specific and improve the stability of the molecules by 6% on average, demonstrating that GCDM is capable of not only 3D molecule generation but also 3D molecule optimization (i.e., refinement). Although Table 4 shows that both baseline optimization methods face difficulties in optimizing molecules that are initially highly stable, the results in this setting still show that molecular diffusion models such as GCDM and EDM can achieve success in molecular optimization of highly-stable molecules.

Table 4:

Comparison of GCPNet with baseline methods for property-guided 3D molecule optimization. The results are reported in terms of molecular stability (MS) and the mean absolute error for molecular property prediction by an EGNN classifier ϕc on a QM9 subset, with results listed for EDM and GCDM-optimized samples as well as two different molecule generation baselines (”EDM Samples” and ”GCDM Samples”). The top-1 (best) results for this task are in bold, and the second-best results are underlined.

Task α / MS Δϵ / MS ϵHOMO / MS ϵLUMO / MS μ / MS Cv / MS
Units Bohr3 / % meV / % meV / % meV / % D / % calmolK/%

EDM Samples (Moderately Stable) 4.91 / 82.9 1.24 / 82.9 0.55 / 82.9 1.23 / 82.9 1.40 / 82.9 2.84 / 82.9
EDM-Opt (on EDM Samples) 4.80 / 84.4 1.24 / 86.3 0.55 / 84.4 1.24 / 85.2 1.41 / 86.0 2.83 / 84.2
GCDM-Opt (on EDM Samples) 4.76 / 85.2 1.22 / 84.0 0.54 / 84.6 1.20 / 83.5 1.36 / 88.1 2.71 / 84.3

GCDM Samples (Highly Stable) 4.82 / 90.5 1.19 / 90.5 0.54 / 90.5 1.24 / 90.5 1.32 / 90.5 2.82 / 90.5
EDM-Opt (on GCDM Samples) 4.67 / 89.0 1.19 / 90.8 0.54 / 90.8 1.24 / 91.2 1.32 / 92.6 2.80 / 90.0
GCDM-Opt (on GCDM Samples) 4.71 / 90.1 1.18 / 91.2 0.53 / 91.0 1.23 / 89.7 1.30 / 91.3 2.81 / 90.1

We note that, in general, both baseline methods likely improve the initial molecules’ property specificities only marginally as a function of the small number of optimization steps used. Here, however, we use a small number of optimization steps with both baselines to mimic an important real-world use case of these models: rapid relaxation and optimization of generated molecules at the pace and scale of drug screening procedures in the pharmaceutical industry. To our best knowledge, the results in Table 4 demonstrate the first successful example of using diffusion models to optimize 3D molecules for molecular stability as well as for specific molecular properties, setting the stage for important future applications of these models within modern drug discovery pipelines.

References

  1. Anand Namrata and Achim Tudor. Protein structure and sequence generation with equivariant denoising diffusion probabilistic models. arXiv preprint arXiv:2205.15019, 2022. [Google Scholar]
  2. Anderson Brandon, Hy Truong Son, and Kondor Risi. Cormorant: Covariant molecular neural networks. Advances in neural information processing systems, 32, 2019. [Google Scholar]
  3. Barron LD. Symmetry and molecular chirality. Chemical Society Reviews, 15(2):189–223, 1986. [Google Scholar]
  4. Batzner Simon, Musaelian Albert, Sun Lixin, Geiger Mario, Mailoa Jonathan P, Kornbluth Mordechai, Molinari Nicola, Smidt Tess E, and Kozinsky Boris. E(3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. Nature communications, 13(1):2453, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bronstein Michael M, Bruna Joan, Cohen Taco, and Veličković Petar. Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv preprint arXiv:2104.13478, 2021. [Google Scholar]
  6. Brown Tom, Mann Benjamin, Ryder Nick, Subbiah Melanie, Kaplan Jared D, Dhariwal Prafulla, Neelakantan Arvind, Shyam Pranav, Sastry Girish, Askell Amanda, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020. [Google Scholar]
  7. Bulusu Srinath, Favoni Matteo, Ipp Andreas, Muller David I, and Schuh Daniel. Generalization¨ capabilities of translationally equivariant neural networks. Physical Review D, 104(7):074504, 2021. [Google Scholar]
  8. Cao Wenming, Yan Zhiyue, He Zhiquan, and He Zhihai. A comprehensive survey on geometric deep learning. IEEE Access, 8:35929–35949, 2020. [Google Scholar]
  9. Cohen Taco and Welling Max. Group equivariant convolutional networks. In International conference on machine learning, pp. 2990–2999. PMLR, 2016. [Google Scholar]
  10. Du Weitao, Zhang He, Du Yuanqi, Meng Qi, Chen Wei, Zheng Nanning, Shao Bin, and Liu Tie-Yan. Se(3) equivariant graph neural networks with complete local frames. In International Conference on Machine Learning, pp. 5583–5608. PMLR, 2022. [Google Scholar]
  11. Du Weitao, Du Yuanqi, Wang Limei, Feng Dieqiao, Wang Guifeng, Ji Shuiwang, Gomes Carla, and Ma Zhi-Ming. A new perspective on building efficient and expressive 3d equivariant graph neural networks. arXiv preprint arXiv:2304.04757, 2023. [Google Scholar]
  12. Elfwing Stefan, Uchibe Eiji, and Doya Kenji. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Networks, 107:3–11, 2018. [DOI] [PubMed] [Google Scholar]
  13. Fuchs Fabian, Worrall Daniel, Fischer Volker, and Welling Max. Se (3)-transformers: 3d roto-translation equivariant attention networks. Advances in Neural Information Processing Systems, 33:1970–1981, 2020. [Google Scholar]
  14. Gebauer Niklas, Gastegger Michael, and Schutt Kristof. Symmetry-adapted generation of 3d point¨ sets for the targeted discovery of molecules. Advances in neural information processing systems, 32, 2019. [Google Scholar]
  15. Han Jiaqi, Rong Yu, Xu Tingyang, and Huang Wenbing. Geometrically equivariant graph neural networks: A survey. arXiv preprint arXiv:2202.07230, 2022. [Google Scholar]
  16. Ho Jonathan, Jain Ajay, and Abbeel Pieter. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851, 2020. [Google Scholar]
  17. Hoogeboom Emiel, Satorras Vıctor Garcia, Vignac Clément, and Welling Max. Equivariant diffusion for molecule generation in 3d. In International Conference on Machine Learning, pp. 8867–8887. PMLR, 2022. [Google Scholar]
  18. Ingraham John, Baranov Max, Costello Zak, Frappier Vincent, Ismail Ahmed, Tie Shan, Wang Wujie, Xue Vincent, Obermeyer Fritz, Beam Andrew, et al. Illuminating protein space with a programmable generative model. bioRxiv, pp. 2022–12, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Jin Wengong, Barzilay Regina, and Jaakkola Tommi. Junction tree variational autoencoder for molecular graph generation. In Dy Jennifer and Krause Andreas (eds.), Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pp. 2323–2332. PMLR, 10–15 Jul 2018. URL https://proceedings.mlr.press/v80/jin18a.html. [Google Scholar]
  20. Joshi Chaitanya K, Bodnar Cristian, Mathis Simon V, Cohen Taco, and Liò Pietro. On the expressive power of geometric graph neural networks. arXiv preprint arXiv:2301.09308, 2023. [Google Scholar]
  21. Jumper John, Evans Richard, Pritzel Alexander, Green Tim, Figurnov Michael, Ronneberger Olaf, Tunyasuvunakool Kathryn, Bates Russ, Augustin Žídek Anna Potapenko, et al. Highly accurate protein structure prediction with alphafold. Nature, 596(7873):583–589, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Kingma Diederik, Salimans Tim, Poole Ben, and Ho Jonathan. Variational diffusion models. Advances in neural information processing systems, 34:21696–21707, 2021. [Google Scholar]
  23. Kofinas Miltiadis, Nagaraja Naveen, and Gavves Efstratios. Roto-translated local coordinate frames for interacting dynamical systems. Advances in Neural Information Processing Systems, 34:6417–6429, 2021. [Google Scholar]
  24. Köhler Jonas, Klein Leon, and Noé Frank. Equivariant flows: exact likelihood generative learning for symmetric densities. In International conference on machine learning, pp. 5361–5370. PMLR, 2020. [Google Scholar]
  25. Kong Zhifeng, Ping Wei, Huang Jiaji, Zhao Kexin, and Catanzaro Bryan. Diffwave: A versatile diffusion model for audio synthesis. arXiv preprint arXiv:2009.09761, 2020. [Google Scholar]
  26. Landrum Greg et al. Rdkit: A software suite for cheminformatics, computational chemistry, and predictive modeling. Greg Landrum, 8, 2013. [Google Scholar]
  27. LeCun Yann, Bengio Yoshua, et al. Convolutional networks for images, speech, and time series. The handbook of brain theory and neural networks, 3361(10):1995, 1995. [Google Scholar]
  28. Loshchilov Ilya and Hutter Frank. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101, 2017. [Google Scholar]
  29. Luo Shitong, Su Yufeng, Peng Xingang, Wang Sheng, Peng Jian, and Ma Jianzhu. Antigen-specific antibody design and optimization with diffusion-based generative models. bioRxiv, pp. 2022–07, 2022. [Google Scholar]
  30. Medsker Larry and Jain Lakhmi C. Recurrent neural networks: design and applications. CRC press, 1999. [Google Scholar]
  31. Morehead Alex and Cheng Jianlin. Geometry-complete perceptron networks for 3d molecular graphs. arXiv preprint arXiv:2211.02504, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Mudur Nayantara and Finkbeiner Douglas P. Can denoising diffusion probabilistic models generate realistic astrophysical fields? arXiv preprint arXiv:2211.12444, 2022. [Google Scholar]
  33. Musaelian Albert, Batzner Simon, Johansson Anders, Sun Lixin, Owen Cameron J, Kornbluth Mordechai, and Kozinsky Boris. Learning local equivariant representations for large-scale atomistic dynamics. arXiv preprint arXiv:2204.05249, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Peebles William, Radosavovic Ilija, Brooks Tim, Efros Alexei A, and Malik Jitendra. Learning to learn with generative models of neural network checkpoints. arXiv preprint arXiv:2209.12892, 2022. [Google Scholar]
  35. Ramakrishnan Raghunathan, Dral Pavlo O, Rupp Matthias, and Von Lilienfeld O Anatole. Quantum chemistry structures and properties of 134 kilo molecules. Scientific data, 1(1):1–7, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Ramesh Aditya, Dhariwal Prafulla, Nichol Alex, Chu Casey, and Chen Mark. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125, 2022. [Google Scholar]
  37. Rombach Robin, Blattmann Andreas, Lorenz Dominik, Esser Patrick, and Ommer Björn. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684–10695, 2022. [Google Scholar]
  38. Ruthotto Lars and Haber Eldad. An introduction to deep generative modeling. GAMM-Mitteilungen, 44(2):e202100008, 2021. [Google Scholar]
  39. Saharia Chitwan, Chan William, Saxena Saurabh, Li Lala, Whang Jay, Denton Emily, Ghasemipour Seyed Kamyar Seyed, Ayan Burcu Karagol, Mahdavi S Sara, Lopes Rapha Gontijo, et al. Photorealistic text-to-image diffusion models with deep language understanding. arXiv preprint arXiv:2205.11487, 2022. [Google Scholar]
  40. Satorras Victor Garcia, Hoogeboom Emiel, Fuchs Fabian B, Posner Ingmar, and Welling Max. E (n) equivariant normalizing flows. arXiv preprint arXiv:2105.09016, 2021a. [Google Scholar]
  41. Satorras Vıctor Garcia, Hoogeboom Emiel, and Welling Max. E (n) equivariant graph neural networks. In International conference on machine learning, pp. 9323–9332. PMLR, 2021b. [Google Scholar]
  42. Schulman J, Zoph B, Kim C, Hilton J, Menick J, Weng J, Uribe JFC, Fedus L, Metz L, Pokorny M, et al. Chatgpt: Optimizing language models for dialogue, 2022.
  43. Segler Marwin HS, Kogej Thierry, Tyrchan Christian, and Waller Mark P. Generating focused molecule libraries for drug discovery with recurrent neural networks. ACS central science, 4(1): 120–131, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Sohl-Dickstein Jascha, Weiss Eric, Maheswaranathan Niru, and Ganguli Surya. Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning, pp. 2256–2265. PMLR, 2015. [Google Scholar]
  45. Song Jiaming, Meng Chenlin, and Ermon Stefano. Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502, 2020. [Google Scholar]
  46. Stark Hannes, Ganea Octavian, Pattanaik Lagnajit, Barzilay Regina, and Jaakkola Tommi. Equibind:¨ Geometric deep learning for drug binding structure prediction. In International Conference on Machine Learning, pp. 20503–20521. PMLR, 2022. [Google Scholar]
  47. Tang Raphael, Pandey Akshat, Jiang Zhiying, Yang Gefei, Kumar Karun, Lin Jimmy, and Ture Ferhan. What the daam: Interpreting stable diffusion using cross attention. arXiv preprint arXiv:2210.04885, 2022. [Google Scholar]
  48. Thomas Nathaniel, Smidt Tess, Kearnes Steven, Yang Lusann, Li Li, Kohlhoff Kai, and Riley Patrick. Tensor field networks: Rotation-and translation-equivariant neural networks for 3d point clouds. arXiv preprint arXiv:1802.08219, 2018. [Google Scholar]
  49. Vaswani Ashish, Shazeer Noam, Parmar Niki, Uszkoreit Jakob, Jones Llion, Gomez Aidan N, Kaiser Łukasz, and Polosukhin Illia. Attention is all you need. Advances in neural information processing systems, 30, 2017. [Google Scholar]
  50. Watson Joseph L, Juergens David, Bennett Nathaniel R, Trippe Brian L, Yim Jason, Eisenach Helen E, Ahern Woody, Borst Andrew J, Ragotte Robert J, Milles Lukas F, et al. Broadly applicable and accurate protein design by integrating structure prediction networks and diffusion generative models. bioRxiv, pp. 2022–12, 2022. [Google Scholar]
  51. Wu Lemeng, Gong Chengyue, Liu Xingchao, Ye Mao, and Liu Qiang. Diffusion-based molecule generation with informative prior bridges. arXiv preprint arXiv:2209.00865, 2022. [Google Scholar]
  52. Xu Minkai, Yu Lantao, Song Yang, Shi Chence, Ermon Stefano, and Tang Jian. Geodiff: A geometric diffusion model for molecular conformation generation. arXiv preprint arXiv:2203.02923, 2022. [Google Scholar]
  53. Yim Jason, Trippe Brian L, De Bortoli Valentin, Mathieu Emile, Doucet Arnaud, Barzilay Regina, and Jaakkola Tommi. Se (3) diffusion model with application to protein backbone generation. arXiv preprint arXiv:2302.02277, 2023. [Google Scholar]
  54. Zhou Jie, Cui Ganqu, Hu Shengding, Zhang Zhengyan, Yang Cheng, Liu Zhiyuan, Wang Lifeng, Li Changcheng, and Sun Maosong. Graph neural networks: A review of methods and applications. AI open, 1:57–81, 2020. [Google Scholar]

Articles from ArXiv are provided here courtesy of arXiv

RESOURCES