Skip to main content
Bioinformatics logoLink to Bioinformatics
. 2025 Jan 9;41(3):btaf010. doi: 10.1093/bioinformatics/btaf010

FlowPacker: protein side-chain packing with torsional flow matching

Jin Sub Lee 1, Philip M Kim 2,3,
Editor: Arne Elofsson
PMCID: PMC11886813  PMID: 39786861

Abstract

Motivation

Accurate prediction of protein side-chain conformations is necessary to understand protein folding, protein–protein interactions and facilitate de novo protein design.

Results

Here, we apply torsional flow matching and equivariant graph attention to develop FlowPacker, a fast and performant model to predict protein side-chain conformations conditioned on the protein sequence and backbone. We show that FlowPacker outperforms previous state-of-the-art baselines across most metrics with improved runtime. We further show that FlowPacker can be used to inpaint missing side-chain coordinates and also for multimeric targets, and exhibits strong performance on a test set of antibody–antigen complexes.

Availability and implementation

Code is available at https://gitlab.com/mjslee0921/flowpacker.

1 Introduction

A protein’s 3D structure, determined by its primary amino acid sequence, is the main determinant of its function. The side-chain atoms heavily influences its folding and interaction with other proteins/ligands through various interatomic interactions and potentials. Therefore, to fully understand protein folding and identify protein–protein interactions, accurate models of protein side-chain packing must be developed.

Recent advances in artificial intelligence has made astounding progress in protein structure prediction (Jumper et al. 2021, Abramson et al. 2024) and design (Lee et al. 2023, Watson et al. 2023, Yim et al. 2023). Here, we focus on the problem of side-chain packing, which seeks to predict the side-chain conformations given the amino acid sequence and backbone coordinates of a protein structure. Many prior efforts rely on physics-based modeling, which uses empirical scoring functions (Alford et al. 2017), discrete rotamer libraries (Dunbrack Jr and Karplus 1993), and/or MCMC-based sampling to identify plausible rotamers. However, these methods are often ineffective due to inefficient search algorithms and the reliance of often inaccurate scoring functions that converge to suboptimal local minima. Recent deep learning-based methods have shown significant improvements in runtime and efficacy to physics-based modeling in side-chain packing (Misiura et al. 2022, McPartlon and Xu 2023, Randolph and Kuhlman 2024, Zhang et al. 2023), the most notable of which is DiffPack (Zhang et al. 2023), a torsional diffusion model that autoregressively generates the four χ torsion angles that constitute the only degrees of freedom for side-chain conformations. DiffPack presents a number of innovations, such as autoregressive generation and confidence sampling, to attain state-of-the-art performance in protein side-chain packing. In this work, we approach the side-chain packing problem using flow matching and equivariant graph attention networks to attain state-of-the-art performance.

Flow matching (Lipman et al. 2023) is a novel generative modeling paradigm that allows the training of continuous normalizing flows (CNFs) in a simulation-free manner, and have shown stronger performance, faster training convergence, and faster inference than standard diffusion models. In FlowPacker, we replace the torsional diffusion framework with torsional flow matching, derived from flow matching on Riemannian manifolds (Chen and Lipman 2024). We also apply state-of-the-art equivariant graph attention models—namely, EquiformerV2 (Liao et al. 2024)—which has been shown to improve performance over invariant message passing networks due to increased expressivity and parameter efficiency.

We observe FlowPacker outperforms other side-chain packing baselines across most evaluated metrics while being considerably faster. We show that FlowPacker can be used for partial inpainting of side-chain conformations, a feature not readily available with other packing methods. We also show that FlowPacker can be extended to multimeric complexes and specifically with antibody–antigen complexes and show significant performance improvements in CDRH3 and full variable chain (Fv) side-chain packing. Therefore, it can be easily incorporated with existing backbone generative models and sequence design tools to generate accurate full-atom structures.

2 Materials and methods

A schematic of FlowPacker is provided in Fig. 1. Here, we provide a brief overview of the theoretical background and model details.

Figure 1.

Figure 1.

FlowPacker overview. FlowPacker is an equivariant graph attention network that generates side-chain conformations of a given protein structure and sequence using torsional flow matching. The model predicts the vector field of the conditional flow along the hypertorus between a prior angle χ0 and the ground-truth angle χ1, which is used with an ODE Solver (e.g. Euler’s method) to generate a sample from the data distribution. Note we show a “virtual” side-chain conformation with three χ angles, but each amino acid can contain up to four χ angles that affect the positions of up to nine atoms.

2.1 Torsional flow matching

Continuous normalizing flows (CNFs) (Chen et al. 2018) are a class of generative models that uses a learned vector field to transform a simple prior to the desired data distribution. Specifically, in CNFs, a neural network parameterizes this invertible transformation between the prior and data by specifying an ordinary differential equation (ODE), where the output of the network can be solved using standard black-box ODE solvers. However, the maximum likelihood simulation-based training of CNFs has limited its scalability and application to complex datasets. Recently, flow matching was proposed to train continuous normalizing flows (CNFs) in a simulation-free manner. Lipman et al. (2023) show that CNFs can be trained by regressing on a conditional probability path pt(x|x1) to learn the unconditional path pt(x) that transforms a prior density p(0) to the data distribution p(1). Chen and Lipman (2024) extends the flow-matching framework to Riemannian manifolds including the high-dimensional torus, which is used in this work to define a torsional flow matching framework for side-chain conformation generation and briefly discussed below.

In flow matching, we desire to learn the time-conditioned vector field vt(x) with t[0,1] that transforms a simple prior distribution p(0) to the data distribution p(1), where the learned vector field can then be used to integrate a prior sample from t=0 to t=1 for generative modeling. However, the probability path pt(x) is intractable to compute. Lipman et al. (2023) show that we can use the conditional vector fieldvt(x|x1) that generates conditional probability pathspt(x|x1) as regression targets, which converges to the same optima as the unconditional vector field vt(x) that produces the unconditional probability path pt(x). Given a conditional flow ψt(x0|x1) that transforms the prior distribution p0(x|x1) to pt(x|x1), we observe that:

ddtψt(x)=vt(ψt(x0|x1)|x1)=x˙t (1)

which defines a conditional vector field that can serve as the regression target for simulation-free training, resulting in the conditional flow matching (CFM) loss below:

LCFM=Et,p1(x1),p0(x0)[||vt(xt)x˙t||2] (2)

In cases where the data resides in a manifold (i.e. rotation matrices, angles), applying standard Euclidean flow matching can be inefficient if the flow is not restricted to the manifold of interest—for instance, for angular data we desire to restrict the data input, conditional flow, and model output to the interval [π,π] for efficient learning. Chen and Lipman (2024) introduce Riemannian flow matching, which extends flow matching to manifolds and general geometries. They show that the conditional flows ψt(x|x1) can be constructed on simple manifolds using a linear scheduler κ(t)=1t with the geodesic distance as the premetric that concentrates all mass at x1 for t=1. The geodesic distance for simple manifolds, such as the high-dimensional torus considered in this work, can be computed in closed form with the logarithmic map between two points on the manifold, and mapped back to the manifold using the exponential map. Thus, the conditional flow is defined as:

ψt(xt|x1)=expx0(tlogx0(x1)) (3)

where the exponential and logarithmic map for high-dimensional tori are defined as:

expx0(x1)=(x0+x1)%(2π) (4)
logx0(x1)=arctan2(sin(x1x0)cos(x1x0)) (5)

This corresponds to a flow along the geodesic distance between x1 and x0 that varies linearly with time. This can be used to interpolate xt in (2) to define the torsional flow matching loss used in this work.

2.2 Equivariant graph attention using Equiformerv2

Equivariant neural networks have been shown to increase model performance and efficiency by directly integrating the symmetries of the data into the model architecture, without the need for data augmentation nor additional parameters (Joshi et al. 2023) [Although the need for equivariance has been challenged by recent works, notably AlphaFold3 (Abramson et al. 2024), since equivariant networks often are more expensive to train over optimized Transformer architectures (Dao et al. 2022).]. Equiformer (Liao and Smidt 2023) is a graph attention network that operates on equivariant irreducible representation (irreps) features and transfers information between various type-L vectors using tensor product operations, where different type-L vectors can be combined through the use of Clebsch–Gordan coefficients. Equiformer uses depthwise tensor product (DTP) blocks—where one type-L vector in the output irreps is only dependent on one type-L‘ vector in the input irreps—to reduce computational complexity of tensor product, but is still practically restricted to max L=3 vectors due to efficiency. For a more detailed discussion on tensor product operations, we refer readers to Geiger and Smidt (2022) and Fuchs et al. (2020).

Equiformerv2 (Liao et al. 2024) facilitates the scaling of the model to higher-degree features by replacing SO(3) convolutions with eSCN convolutions (Passaro and Zitnick 2023), which intuitively aligns the relative edge direction to one axis and therefore greatly sparsifies the Clebsch-Gordon coefficient matrix and simplifies to SO(2) convolutions, reducing computational complexity and allowing scaling to higher-degree tensors up to L=6 or L=8 for increased model expressivity. While the task of side-chain packing does not necessitate the use of equivariant networks since torsion angles are invariant features, in preliminary studies we observed stronger empirical performance of Equiformerv2 over invariant and equivariant baseline architectures.

2.3 FlowPacker specifications

2.3.1 Dataset curation

We train FlowPacker on two datasets: 1. BC40 dataset (release date 28 July 2020, available at https://drug.ai.tencent.com/protein/bc40/download.html), which are representative PDB structures clustered by MMseqs2 (Steinegger and Söding 2017) at 40% sequence identity, and is used by previous side-chain packing models [DiffPack (Zhang et al. 2023) and AttnPacker (McPartlon and Xu 2023)] for model training, and 2. a monomer dataset constructed from a PDB snapshot (date 28 July 2023) clustered at 40% sequence identity (hereafter referred to as PDB-S40). All results are based on the model trained using PDB-S40 unless otherwise noted. We use the CASP13, 14, and 15 targets for testing (available at https://predictioncenter.org/download_area/). We reduce data redundancy in the training set by removing any structures/clusters with 40% sequence similarity to any of the CASP13/14/15 targets using MMseqs2’s easy-search workflow. We discard structures of which >25% of residues are unknown, and remove residues with missing backbone coordinates or overlapping alpha carbon positions. We also filter for proteins with at least 40 residues. This results in 36 451 training examples for the BC40 dataset and 23 191 clusters containing 309 467 structures for the PDB-S40 dataset. The test set consists of 99 structures, with 20, 34, and 45 targets in CASP13, 14, and 15, respectively.

2.3.2 Model architecture

FlowPacker uses the Equiformerv2 model architecture as described previously. We use Lmax=3, channel dimension of 256, and 4 blocks, resulting in a total of 18.0M trainable parameters. Interestingly, we did not observe a change in performance from scaling Lmax up to 6, suggesting that higher-order features are not informative for side-chain packing.

2.3.3 Loss functions

The model is trained to predict the conditional vector field as defined in Section 2.1:

LCFM=Et,p1(χ1),p0(χ0)[1(1t)2||vt(χt)χ˙t||2] (6)

where vt(χt) corresponds to the model output, χ˙t is the exact conditional vector field efficiently computed with autograd, and 1(1t2) is a weighting factor.

We experimented with multiple variations of the loss function including using the vector field approximation logχt(χ1)/(1t), which showed negligible differences in training dynamics and performance. We also tried using a reparameterized direct χ1 prediction as with rotation matrices in FrameFlow (Yim et al. 2023), which can be used to approximate the vector field during sampling. However, this caused unstable training dynamics, most likely due to the degeneracy of χ angles.

2.3.4 Handling symmetry issues

Certain χ angles exhibit π-symmetry that may adversely affect model efficiency. We handle these cases by taking mod  π for π-symmetric χ angles and modifying Equation (4) to:

expχ0(χ1)=(χ0+χ1)%(cπ), (7)

where c=1 for π-symmetric χ angles and c=2 otherwise, restricting π-symmetric angles to [0,π).

2.3.5 Training details

In FlowPacker, we use an residue-level graph, where each residue represents a node with the position given by its idealized Cβ atom. At each training step, we sample timesteps t and generate noised torsions χt, which are used to reconstruct noised atomic coordinates Xt using idealized coordinates. The node features consist of the one-hot encoded amino acid identity, backbone torsion angles, and sinusoidal embeddings of the current timestep, while the edge features are one-hot encoded relative positional encoding clamped at [32,32], and the Euclidean distance between all idealized coordinates (using the atom14 representation) between two connected residues. The edges are defined by a k-nn graph with k=30. The model outputs the predicted vector field per torsion angle (up to 4 per residue), which is compared with the ground-truth vector fields for loss optimization. The model is trained on 4 40 GB NVIDIA A100s, effective batch size of 16, and context length of 512 residues (cropped when the length is >512) for 300 epochs, or approximately 6 days. We use the AdamW optimizer with learning rate 1.0e4 and gradient clipping at a norm value of 1.0.

2.3.6 Inference strategies

Though the model was trained using a conditional flow with a linear scheduler κ(t)=1t, we observe a similar phenomenon as described in FrameFlow where the generated samples are poor using the same linear schedule during sampling. Instead, we adopt the exponential schedule vt=c(logxt(x1)), and empirically find that c = 5 works well. We use a uniform distribution on SO(2) as the base distribution and Euler solver with t = 10 for all sampling steps, as we observed no significant increase in performance with additional timesteps. We experimented with using higher-order solvers such as Heun’s method and RK45 solvers, but did not observe a noticeable performance improvement at the cost of increased runtime. We use exponential moving average with decay of 0.999 at every training iteration.

2.3.7 Confidence model

The confidence model is trained in a similar manner to the main model, except we regress on the residue-level sidechain RMSDs rather than χ-level vector fields. We identify the best sample by simply taking the mean predicted RMSD across all residues and selecting the sample with the lowest predicted RMSD. Unless otherwise noted, we generate four samples per test case and select the highest confidence sample. The confidence model trained in this work contains 2.2M trainable parameters with max l=2 and hidden dimension of 64—we did not perform extensive hyperparameter tuning on this module.

3 Results

We compare FlowPacker to physics-based [Rosetta (Alford et al. 2017)] and deep learning-based [AttnPacker (McPartlon and Xu 2023) and DiffPack (Zhang et al. 2023)] methods across three test datasets—CASP13, CASP14, and CASP15—in Table 1. As with previous studies (McPartlon and Xu 2023, Zhang et al. 2023), we report three metrics: (i) Angle MAE corresponds to the angular mean absolute error, which returns min(χ,χmod2π), (ii) Angle Accuracy refers to the percentage of angles within 20° of the ground-truth, and (iii) Atom RMSD refers to the average root-mean squared deviation of side-chain atoms per residue, where core residues are defined as residues with at least 20 Cβ atoms within 10A˙, and surface residues are those with at most 15 Cβ within 10A˙.

Table 1.

Performance evaluation of FlowPacker and other methods on CASP targets.a

Angle MAE ° 
Angle Accuracy %
Atom RMSD Å
Dataset Method χ1 χ2 χ3 χ4 χ1 χ2 χ3 χ4 All Core Surface
CASP13 Rosetta 24.84 30.96 45.35 58.28 78.36% 67.75% 47.92% 41.23% 0.822 0.502 1.025
AttnPacker 17.82 33.41 67.31 48.89 83.32% 64.96% 32.76% 41.67% 0.745 0.531 0.874
AttnPacker-pp 16.33 26.00 51.18 49.40 83.53% 70.05% 41.83% 41.10% 0.676 0.456 0.815
DiffPack-fix 22.14 28.80 45.20 52.28 81.63% 71.57% 50.98% 48.89% 0.789 0.432 1.005
FlowPacker 16.44 24.38 42.82 52.16 85.69% 75.55% 50.31% 51.00% 0.674 0.388 0.865
+ confidence 15.48 24.44 41.83 50.80 86.58% 75.69% 51.17% 51.64% 0.660 0.382 0.840
CASP14 Rosetta 32.32 35.47 49.19 54.27 67.62% 58.48% 39.52% 42.15% 1.001 0.697 1.199
AttnPacker 27.29 39.26 67.94 49.99 70.77% 55.22% 27.44% 37.12% 0.955 0.675 1.130
AttnPacker-pp 26.06 32.75 55.06 50.59 71.03% 59.57% 34.53% 36.87% 0.900 0.607 1.085
DiffPack-fix 31.00 34.43 51.72 57.50 69.59% 61.32% 38.69% 39.35% 0.994 0.639 1.225
FlowPacker 22.17 29.52 45.33 50.48 77.37% 66.66% 44.00% 46.64% 0.830 0.503 1.053
+ confidence 21.92 29.26 45.21 49.13 77.52% 66.88% 44.21% 47.45% 0.822 0.501 1.043
CASP15 Rosetta 32.37 35.22 45.29 58.92 70.09% 61.81% 43.91% 41.93% 0.938 0.611 1.158
AttnPacker 28.16 41.90 69.90 53.22 71.99% 55.15% 30.24% 36.70% 0.925 0.696 1.099
AttnPacker-pp 26.80 32.74 56.78 53.98 72.29% 61.65% 36.04% 36.70% 0.851 0.549 1.052
DiffPack-fix 31.86 34.46 48.06 61.01 71.08% 62.93% 44.11% 41.27% 0.921 0.524 1.145
FlowPacker 23.50 29.66 43.22 54.37 78.04% 68.71% 48.69% 46.42% 0.770 0.404 0.991
+ confidence 23.41 29.07 42.69 55.46 78.19% 69.36% 48.28% 45.97% 0.764 0.390 0.985
a

The top performing model per test dataset is highlighted in bold. DiffPack-fix shows results for DiffPack using structures with idealized bond lengths and angles as input. We generate four samples for FlowPacker (displaying average metrics when confidence model is not used) and use default settings for DiffPack (four samples with confidence model selection). AttnPacker-pp corresponds to samples with post-processing applied as denoted in McPartlon and Xu (2023).

During testing, we observed a data leakage issue with DiffPack, where the generation quality was dependent on the input side-chain coordinates, everything else held constant. We noticed that during training and sampling, DiffPack rotates the ground-truth side-chain coordinates given χt, maintaining the ground-truth bond angles and bond lengths. Moreover, we discovered that if we idealize the side-chain atoms [as in AlphaFold2 (Jumper et al. 2021) and FlowPacker] and provide the idealized structures (note we still use the nonidealized ground-truth backbone coordinates) to DiffPack, the generation quality is significantly worse (see Diffpack-fix in Table 1). We hypothesize that although the χ14 angles are the primary degrees of freedom in side-chain packing, the atomic coordinates do exhibit minor deviations from various interatomic forces and potentials (Engh and Huber 2012) that may provide clues into the ground-truth side-chain conformation. Moreover, we observed that using ground-truth bond angles and lengths provides a > 0.25A RMSD advantage over using idealized ones in terms of side-chain RMSD. For further discussion, please refer to Supplementary Information S1.1. Since side-chain packing methods should require no a priori knowledge of ground-truth side-chain positions, we exclude DiffPack from ranking across metrics. We note that the numbers denoted under DiffPack-fix may not accurately reflect the true performance of DiffPack since the model is not explicitly trained with idealized coordinates—however, this requires further investigation that is outside of the scope of this work.

We observe that FlowPacker outperforms all baselines across most metrics, suggesting the superiority of flow matching and equivariant networks over diffusion and invariant models, respectively. We also analyze the number of clashes per sample across all baseline models and test datasets, and observe that samples from FlowPacker exhibits the lowest number of clashes Supplementary Table S2. Using a confidence model to select the lowest predicted RMSD sample marginally improved performance across all test datasets. However, the <0.1 RMSD decrease suggests that for high-throughput screening, the confidence model may not be necessary since it requires multiple generations (four used here), each with an extra forward pass through the confidence model that results in a >4X increase in runtime. We observe that χ1 accuracy is generally higher than the subsequent χ angles, as seen with previous methods, due to the lever effect where the errors accumulate through each χ angle and the increased flexibility of side-chain positions at longer side-chains (ex. arginines and leucines). We also analyze runtime and performance across all methods in Fig. 3, and observe that FlowPacker exhibits the best runtime and performance of all tested baselines. As shown in Supplementary Fig. S2, FlowPacker can recapitulate the χ distributions with decent accuracy, including π-symmetric ones, which lie in the interval [0, 180] due to the parameterization in Equation (7). An analysis of residue-level atom RMSDs for each unique amino acid reveals that longer and aromatic sidechains are often harder to predict accurately over shorter ones Supplementary Fig. S3. We additionally observe that the increased training data available in the PDB-S40 dataset consistently improves performance across most metrics over the widely used BC40 dataset Supplementary Table S3, suggesting the importance of careful training data curation.

Figure 3.

Figure 3.

Runtime and performance analysis of various methods. We use the CASP13 dataset to analyze the total runtime (averaged per protein) and Atom RMSD across the tested methods. FlowPacker exhibits the best RMSD-runtime tradeoff, while FlowPacker with the confidence model generates conformations with the lowest RMSD. We report the metrics of DiffPack-fix in Table 1 here, and all runtime metrics are averaged over three independent seeds for each method. All models were tested on AMD Ryzen 7 5800 (Rosetta) or NVIDIA RTX 3060 (others). Note that runtime is plotted in log scale.

Figure 2.

Figure 2.

Model architecture. The model takes in the amino acid identity, backbone torsion angles, and timestep embeddings as node features, and relative sequential positions and Euclidean distances between all atoms of two given residues as edge features. The model updates the embeddings of residue i using the equivariant graph attention module of EquiformerV2, where after many layers (four used here) the residue embeddings are fed into a final output head (not shown) to predict up to four torsional vector fields. Note that only one layer is shown for simplicity. Schematic is adapted from Fig. 1 of EquiformerV2 (Liao et al. 2024).

To test FlowPacker’s ability to inpaint missing side-chain coordinates, we mask 5%–75% randomly selected residues on the CASP15 test set and report the metrics in Table 2. As expected, we observe lower RMSDs as we provide more structural context, suggesting the utility of FlowPacker for conditional design. We expect this to be useful in common protein design cases such as motif-scaffolding or interface design, where protein backbone generative models design the backbone coordinates, and FlowPacker can be used in conjunction with sequence design tools to generate full-atom coordinates.

Table 2.

Performance on CASP15 targets with varying levels of masking.a

Angle MAE ° 
RMSD Å
Masking χ1 χ2 χ3 χ4 All
5% 21.51 29.07 50.33 54.51 0.727
10% 21.85 27.62 44.46 52.39 0.728
25% 22.37 28.05 45.03 53.34 0.737
50% 23.12 29.03 43.40 53.63 0.756
75% 23.30 29.60 43.30 54.05 0.765
100% 23.50 29.66 43.22 54.37 0.770
a

We mask varying proportions of residues randomly and calculate Angle MAE and Atom RMSD on the masked residues.

Although FlowPacker was specifically trained on single-chain proteins, we tested whether the model can be used to generate side-chain conformations of antigen-antibody complexes. We extract antibody–antigen complexes from SabDab (Dunbar et al. 2014) and filter for structures released after 1 January 2021. We also remove complexes with at least 40% antigen sequence similarity with MMseqs (Steinegger and Söding 2017) to any complex structure published prior to 1 January 2021 to remove redundancy, and crop antigens to 256 residues proximal to the antibody. This results in 104 clusters, from which we select a representative member (provided by MMseqs) from each cluster for testing. For inference on multi-chain proteins, we simply fix the relative positional encodings for edges between residues of different chains to 32 and all other input features are kept the same as in single-chain inference. In Table 3, we report the metrics for two common design tasks: CDRH3 design and full Fv design. For all cases, we provide the antigen structure as conditional inputs and pack the sidechains of the CDRH3 or full Fv region (or heavy chain only if light chain data is not available). We only use Rosetta as the baseline, as the other baseline methods are not adapted for multimeric inputs. FlowPacker outperforms Rosetta in both tasks and across most metrics despite not being explicitly trained on multimeric complexes. Therefore, FlowPacker can be used to generate more accurate full-atom structures of both monomeric and multimeric proteins, not limited to antibody–antigen complexes.

Table 3.

Test case on side-chain packing on antibody–antigen complexes.

Angle MAE ° 
RMSD Å
Target Method χ1 χ2 χ3 χ4 All
CDRH3 Rosetta 28.05 26.04 38.16 58.64 0.886
FlowPacker 20.85 25.06 42.16 51.67 0.671
Full Fv Rosetta 30.28 30.72 48.50 54.83 0.790
FlowPacker 22.79 24.32 41.71 51.03 0.649

Bold values indicate the best performing model on each metric.

4 Conclusion

In this work, we present FlowPacker, a torsional flow-matching model that exhibits superior performance and runtime efficiency over other baselines in side-chain packing. We show that FlowPacker can be used to inpaint partial residues and multi-chain inference, showcasing a test case with antibody–antigen CDR side-chain packing, outperforming other methods in angle accuracy and atom RMSD.

We see various promising avenues for future work, such as improved prediction of mutational effects using unsupervised (Luo et al. 2023) or supervised (Liu et al. 2024) learning, or alignment of generative models using preference data (Rafailov et al. 2023, Zhou et al. 2024) such as empirical force fields to increase biophysical plausibility. We also believe that FlowPacker’s performance can be improved—we did not test autoregressive sampling as in DiffPack, which may help performance. We show that the confidence model modestly improves performance, but there are other applications such as resampling of low-confidence conformations and uncertainty analysis as with pLDDT in AlphaFold that may be further explored. We also encourage the exploration of novel representations of side-chain conformations, as explicit representation in 3D space (Lee et al. 2022, Abramson et al. 2024) may outperform implicit representations since the atom RMSD using a χ angle parameterization is often largely dependent on the accuracy of the χ1 prediction. Finally, we envision accurate side-chain packing models to be increasingly necessary with the development of more powerful protein backbone generative models to provide a pipeline for end-to-end full-atom protein design.

Supplementary Material

btaf010_Supplementary_Data

Acknowledgements

We thank the many open-source codebases that this work was based on, including PyTorch, OpenFold, and e3nn. We also would like to thank Digital Research Alliance of Canada for computing resources. We also thank the anonymous reviewers for their valuable suggestions.

Contributor Information

Jin Sub Lee, Department of Molecular Genetics, University of Toronto, Toronto, Ontario, M5S 3K3, Canada.

Philip M Kim, Department of Molecular Genetics, University of Toronto, Toronto, Ontario, M5S 3K3, Canada; Department of Computer Science, University of Toronto, Toronto, Ontario, M5S 2E4, Canada.

Supplementary data

Supplementary data are available at Bioinformatics online.

Conflict of interest: P.M.K. is a co-founder and consultant to several biotechnology companies, including Fable Therapeutics, TBG Therapeutics, and Zymedi. J.S.L. reports no competing interests.

Funding

This work was supported in part by funds from Canadian Institutes of Health Research (CIHR) Project Grant [PJT-166008].

Data availability

The data used to train and evaluate the model are available in GitLab at https://gitlab.com/mjslee0921/flowpacker. Data used to reproduce the results can be provided upon request to the corresponding author.

References

  1. Abramson J, Adler J, Dunger J  et al.  Accurate structure prediction of biomolecular interactions with AlphaFold 3. Nature  2024;630:493–500. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Alford RF, Leaver-Fay A, Jeliazkov JR  et al.  The Rosetta all-atom energy function for macromolecular modeling and design. J Chem Theory Comput  2017;13:3031–48. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Chen RTQ, Lipman Y. Flow matching on general geometries. In: The Twelfth International Conference on Learning Representations. Vienna, Austria, 2024.
  4. Chen RTQ, Rubanova Y, Bettencourt J  et al. Neural ordinary differential equations. In: Advances in Neural Information Processing Systems. Montreal, Canada, 2018.
  5. Dao T, Fu DY, Ermon S  et al.  Flashattention: fast and memoryefficient exact attention with io-awareness. Adv Neural Inf Process Syst  2022;35:16344–59. [Google Scholar]
  6. Dunbar J, Krawczyk K, Leem J  et al.  SAbDab: the structural antibody database. Nucleic Acids Res  2014;42:D1140–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Dunbrack RL Jr, Karplus M.  Backbonedependent rotamer library for proteins application to sidechain prediction. J Mol Biol  1993;230:543–74. [DOI] [PubMed] [Google Scholar]
  8. Engh RA, Huber R.  Structure quality and target parameters. In: Crystallography of Biological Macromolecules. Volume F, Part 18, 2012.
  9. Fuchs F  et al.  Se (3)-transformers: 3D rototranslation equivariant attention networks. Adv Neural Inf Process Syst  2020;33:1970–81. [Google Scholar]
  10. Geiger M, Smidt T.  E3NN: Euclidean neural networks. arXiv, arXiv:2207.09453, 2022, preprint: not peer reviewed.
  11. Joshi CK, Bodnar C, Mathis S  et al. On the expressive power of geometric graph neural networks. In: International Conference on Machine Learning. PMLR, Honolulu, Hawaii, 2023, 15330–55.
  12. Jumper J, Evans R, Pritzel A  et al.  Highly accurate protein structure prediction with AlphaFold. Nature  2021;596:583–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Lee JH, Yadollahpour P, Watkins A  et al.  Equifold: protein structure prediction with a novel coarse-grained structure representation. biorXiv, 2022, preprint: not peer reviewed.
  14. Lee JS, Kim J, Kim PM.  Scorebased generative modeling for de novo protein design. Nat Comput Sci  2023;3:382–92. [DOI] [PubMed] [Google Scholar]
  15. Liao Y-L, Smidt T.  Equiformer: equivariant graph attention transformer for 3d atomistic graphs. In: The Eleventh International Conference on Learning Representations. Kigali, Rwanda, 2023.
  16. Liao Y-L, Wood B, Das A  et al. Equiformerv2: improved equivariant transformer for scaling to higherdegree representations. In: The Twelfth International Conference on Learning Representations. Vienna, Austria, 2024.
  17. Lipman Y, Chen RTQ, Ben-Hamu H et al. Flow matching for generative modeling. In: The Eleventh International Conference on Learning Representations. Kigali, Rwanda, 2023.
  18. Liu S, Zhu T, Ren M et al. Predicting mutational effects on protein–protein binding via a side-chain diffusion probabilistic model. In: Advances in Neural Information Processing Systems. Vancouver, Canada, 2024.
  19. Luo S, Su Y, Wu Z et al.  Rotamer density estimator is an unsupervised learner of the effect of mutations on protein–protein interaction. In: The Eleventh International Conference on Learning Representations. Kigali, Rwanda, 2023. [Google Scholar]
  20. McPartlon M, Xu J.  An end-to-end deep learning method for protein side-chain packing and inverse folding. Proc Natl Acad Sci USA  2023;120:e2216438120. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Misiura M, Shroff R, Thyer R  et al.  DLPacker: deep learning for prediction of amino acid side chain conformations in proteins. Proteins Struct Funct Bioinf  2022;90:1278–90. [DOI] [PubMed] [Google Scholar]
  22. Passaro S, Zitnick CL. Reducing so (3) convolutions to so (2) for efficient equivariant GNNs. In: International Conference on Machine Learning. PMLR. Honolulu, Hawaii, 2023, 27420–38.
  23. Rafailov R, Sharma A, Mitchell E et al. Direct preference optimization: your language model is secretly a reward model. In: Advances in Neural Information Processing Systems, New Orleans, Louisiana, 2023.
  24. Randolph NZ, Kuhlman B.  Invariant point message passing for protein side chain packing. Proteins Struct Funct Bioinf  2024;92:1220–33. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Steinegger M, Söding J.  MMseqs2 enables sensitive protein sequence searching for the analysis of massive data sets. Nat Biotechnol  2017;35:1026–8. [DOI] [PubMed] [Google Scholar]
  26. Watson JL, Juergens D, Bennett NR  et al.  De novo design of protein structure and function with RFdiffusion. Nature  2023;620:1089–100. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Yim J, Campbell A, Foong AYK et al. Fast protein backbone generation with SE (3) flow matching. arXiv, arXiv:2310.05297, 2023, preprint: not peer reviewed.
  28. Zhang Y, Zhang Z, Zhong B et al. Diffpack: a torsional diffusion model for autoregressive protein side-chain packing. In: Advances in Neural Information Processing Systems. New Orleans, Lousiana, 2023.
  29. Zhou X, Xue D, Chen R et al. Antigen-specific antibody design via direct energy-based preference optimization. In: Advances in Neural Information Processing Systems. Vancouver, Canada, 2024.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

btaf010_Supplementary_Data

Data Availability Statement

The data used to train and evaluate the model are available in GitLab at https://gitlab.com/mjslee0921/flowpacker. Data used to reproduce the results can be provided upon request to the corresponding author.


Articles from Bioinformatics are provided here courtesy of Oxford University Press

RESOURCES