Abstract
Computational protein structure prediction is a long-standing challenge in bioinformatics. In the process of predicting protein 3D structures, it is common that parts of an experimental structure are missing or parts of a predicted structure need to be remodeled. The process of predicting local protein structures of particular regions is called loop modeling. In this paper, five new loop modeling methods based on machine learning techniques, called NearLooper, ConLooper, ResLooper, HyLooper1 and HyLooper2 are proposed. NearLooper is based on the nearest neighbor technique; ConLooper applies deep convolutional neural networks to predict Ca atoms distance matrix as an orientation-independent representation of protein structure; ResLooper uses residual neural networks instead of deep convolutional neural networks; HyLooper1 combines the results of NearLooper and ConLooper while HyLooper2 combines NearLooper and ResLooper. Three commonly used benchmarks for loop modeling are used to compare the performance between these methods and existing state-of-the-art methods. The experiment results show promising performance in which our best method improves existing state-of-the-art methods by 28% and 54% of average RMSD on two datasets while being comparable on the other one.
Keywords: protein structure prediction, loop modeling, machine learning, deep learning, convolutional neural networks, residual neural networks
1. Introduction
KNOWLEDGE of three-dimensional (3D) structure of a protein is critical for understanding its function, mutagenesis experiments and drug development. Some experimental methods such as the X-ray crystolography or Nuclear Magnetic Resonance (NMR) can determine a good 3D structure but they are very time consuming and expensive [1]. To address those limitations, many protein structure prediction methods have been developed. Due to its ability to predict protein structure fast and accurately, template-based protein structure prediction is being used in many biological applications [2–5]. This approach is mainly based on the idea of finding good template in the Protein Data Bank (PDB) [6] with a high level of sequence similarity of a query protein [7]. In many cases, parts of the protein structure cannot be predicted due to no alignment to known structures for these parts. Thus, it is necessary to predict the structure of these missing parts to complete the final protein structure prediction. This is called the loop modeling problem. In some experimental structures, loop modeling is also needed due to lack of structural information in certain parts of the protein.
Loop modeling methods can be divided into two main approaches: ab initio and template-based approaches. The ab initio approach is based on the idea of using biological and physical properties of the loop region in the protein to determine the best loop conformation. Some ab initio methods such as LOOPY [8], PLOP [9], and MODELLER [10] can predict the loops with short length (below seven) quite well [5]. On the other hand, the template-based approaches, such as LIP [11], NGK [12], Galaxy PS1 [13], or Galaxy PS2 [14] tries to derive some the loop structure candidates and possibly combine with some energy functions or selection methods to finish the loop modeling task.
The methods mentioned above are different from one another in details, but most of them still share a general framework that starts with sampling a large number of feasible conformations.
Ab initio methods mainly depend on the performance of two factors: the conformational search algorithm and the energy (scoring) function [45]. There are many computational search algorithms that have been proposed such as simulated annealing [46,47], genetic algorithm [48,49], random tweak [50,51] and many others [45]. Similarly, many energy functions have been used for loop modeling including OPLS [52], CHARMM [53], AMBER [54], etc.
Among ab initio methods, the LOOPY algorithm is based on the random tweak algorithm [15] and carries out loop closure while avoiding steric clashes. MODELLER constructs and samples loop conformations with a bond-scaling and relaxation method. Basically, it combines conjugate gradient minimization and molecular dynamics with simulated annealing [5]. On the other hand, PLOP uses a systematic dihedral angle based build-up process to sample conformation space.
Basically, the template-based methods try to extract loop structures from experimental structural databases such as the Protein Data Bank (PDB) first. Then, further sampling and optimization methods are applied on the extracted loop structures to come up with a good set of candidate structures. Finally, a ranking method such as energy or scoring function is employed to get the best loop structure [7].
Among template-based methods, NGK performs sampling for loop conformation based on the idea of combining intensification of torsion and parameter annealing strategies. Both Galaxy PS1 and Galaxy PS2 are loop refinement methods that starts with an inaccurate loop structure. The energy of Galaxy PS1 is optimized for application to the refinement of template-based models, while Galaxy PS2 is developed for higher performance for the nearnative models as well [13].
In this paper, five novel methods are proposed for loop modeling, called NearLooper, ConLooper, HyLooper1, ResLooper and HyLooper2. The NearLooper method is based on the idea of selecting loop templates with a similar structure of surrounding environment. ConLooper uses Convolutional Neural Networks (CNN) [29–33] to predict candidate loop template. By combining the results of NearLooper and ConLooper, the HyLooper1 method takes advantages of both methods. ResLooper uses Residual Networks to predict candidate loop template. And Hy-Looper2 method is the combination of NearLooper and ResLooper. ConLooper and ResLooper use distance matrix that contains pairwise distances between two residues’ Ca atoms in the model as an orientation-independent representation of protein 3D structure. Experiment results show that these methods achieve comparable results to other state-of-the-art methods and outperform them in some cases.
This paper is organized as follows. Section II introduces major techniques and methods that we used for loop modeling. Section III presents experimental results and discussion on experimented datasets. Finally, Section IV concludes the paper.
2. Methods
The loop modeling problem is formulated as a partial protein structure construction in this paper: given an incomplete protein structure that has a gap region (also called loop) with unknown structure coordinates inside, predict the conformation (structure coordinates) of the gap region. It is indeed the missing value problem in machine learning which can be defined as: given an input N-dimensional matrix with some areas of missing values, predict an ouput N-dimensional matrix so that the areas of missing values are filled in. Chao et. al. [41] is a good example of this. As an example, Figure 2(a) shows a protein structure with a gap in the middle and 2(b) shows the result of a complete structure with the gap region predicted.
Let S = {si, 1≤ i ≤ n} be the amino acid sequence of a target protein of length n, and A = {ai, 1 ≤ i ≤ n} be its incomplete 3D structure, where ai represents the 3-D coordinates of the ith amino acid, that contains a gap region between two indices u and v, 1 < u ≤ i ≤ v < n. Suppose a predicted model of the gap region is M = {mi, u ≤ i ≤ v} and the truth structure of the gap region is D= {di, u ≤ i ≤ v}, the goal is to find an M given A so that the root-mean-square deviation (RMSD) between M and D is minimized.
To calculate RMSD between M and D, which measures the structural similarity between two 3-D structures of the same length, M is first optimally superimposed over D. Then, the root-mean-square deviation (RMSD) is calculated using the coordinates of the corresponding main chain atoms (N, Cα, C and O) in the two structures as follows:
(1) |
where k is the number of pairs of corresponding atoms.
In this section, five new methods are presented for loop modeling, i.e., finding M. They are NearLooper, ConLooper, ResLooper, Hylooper1 and HyLooper2. Figure 1 shows the framework of the five methods that share the following common pre-processing and post-processing steps:
Generate subsequence. A subsequence of the target protein with a gap region is extracted from the original sequence. This subsequence is selected so that it is longer than the gap region and contains the gap region at the center. As an example, Figure 2(c) shows an input sequence with a gap region LVKEQWIL and 2(d) shows the selected subsequence of length 50 that contains the gap region in the middle.
Generate candidate template bank. Here, the term “template” is defined as a protein with complete 3D structure (including the loop region) that was found from alignment tools. A pool of templates (candidate 3D structures) for the selected subsequence are generated using sequence-based alignment tools on a protein structure database (e.g. PDB), such as PSI-BLAST [16] and HHSearch [17]. This pool of templates will be used later to train models.
Loop modeling. The five different methods generate loop models using the pool of templates in different ways.
Superimpose the generated loop model over the target protein structure based on their overlapping regions.
The gap region in the target protein structure is filled in with coordinates from the superimposed loop model. As an example, Figure 2(b) shows the result for input 2(a).
2.1. NearLooper, a nearest neighbor algorithm for loop modeling
NearLooper first finds a pool of templates for an extended loop region (the loop region extended on both sides) using sequence-based search from a protein structure database, and select from them a template whose structure optimally matches with the target structure (A) in their overlapping region. Then, the selected template is superimposed over the original incomplete structure A based on their overlapping region (on both sides of the gap region) and the coordinates of the template in the gap region is outputted as the predicted loop model M, which can be inserted into A to make A complete.
Following the framework in Figure 1, NearLooper uses the nearest neighbor idea to select one template from the template pool in step (3). To determine the similarity between different templates and the target structure, the RMSD value of the overlapping region between a template and the target protein is used as the selection metric: the template with the lowest RMSD is selected.
2.2. ConLooper, a novel deep learning method for loop modeling
In this method, a novel idea of applying CNN on the distance matrix representation of 3-D structures is employed for predicting loop conformation instead of using the 3-D coordinates of protein directly. A protein that is rotated in space will have many different sets of values of coordinates. That leads to difficulty for machine learning methods to “learn” the data. On the other hand, distance matrix is an orientation-independent representation of protein structure that eliminates the problems of rotated proteins.
To create the distance matrix of a protein structure, only coordinates of Cα atoms are kept. A 3D model with n Cα atoms can be converted into an n by n distance matrix B with the Euclidean distance of two points in a 3D space:
(2) |
where, are the 3D coordinates of points i and j, respectively.
Given a distance matrix, a 3D model can be derived using multidimensional scaling [18]. Figure 3 shows an example of conversion between a 3D model and its distance matrix. The distance matrix is an orientation and translation independent representation of a 3D structure of protein structure. Since it is derived from protein structure which has constraints between one residue to other local surrounding residues, this type of distance matrix data does contain local relations information, which is suitable for applying the current CNN techniques to predict loop models [19].
In recently years, deep learning based on deep neural networks (DNN) has had a huge impact on computer science and achieved unprecedented performance on many machine learning problems. Deep-learning methods take raw data to automatically discover the multiple layers of representations needed for detection or classification. These layers of representation in deep learning are learned from input data. Deep convolutional neural networks (DCNN) have made breakthroughs in processing audio, image, video, and playing games. DCNNs are made up of neurons that have learnable weights and biases. Each neuron takes some inputs, perform some calculations and produce an output. DCNNs typically have layered structures including convolutional (CONV) layer, rectify linear units (RELU) layer [28], pooling (POOL) layer and fully connected (FC) layer [20,21].
The DCNN architecture of the ConLooper method is shown in Figure 4: taking the input of a distance matrix representation of a 3-D structure, the network applies several convolutional layers followed by RELU layer alternatively and finally generates a predicted distance matrix as the output. The DCNN is trained to learn a network that can predict an output distance matrix from an input distance matrix.
In this network architecture, there are no pooling or dense layers used since the network is trying to do a regression task to map input distance matrix to output distance matrix instead of trying to do a classification prediction as in other classical DCNN models. If dense layers are used at the end of DCNN network, the output would be a vector of features instead of a 2D matrix as the expected output. Thus, the final layer of DCNN should be just a convolutional layer which represented by a 2D matrix as the expected output. The key operation of our DCNN model is Convolutional layers where it convolutes the input distance matrix to map it to another output distance matrix. This type of DCNN model has been shown that it can perform the regression task efficiently given the condition of input and output are 2D matrix. A reference of this can be found from the work of Oxford Visual Geometry Group [26].
First, ConLooper trains a DCNN using the pool of templates and then predict a loop model using the trained DCNN based on the partial structure of the target protein structure, as shown in Figure 1. As an example, Figure 5 shows the input, a distance matrix with a gap region, for a DCNN to predict the gap region’s 3-D structure.
Figure 6 shows a training example of the DCNN. Given a template found by some 3rd party alignment tools, the input and output of the DCNN, both distance matrices, are derived from the template’s distance matrix. The goal is to learn the mapping from the context structure of the gap region to the structure of the gap region.
Details of ConLooper are as follow:
In training, the candidate templates are first converted into distance matrices using Equation (2) and normalized with zero mean and unit variance based on the training data. Then, each distance matrix generates a training example, a pair of input and output sub-distance matrix, in the way illustrated in Figure 6. Finally, a DCNN is trained using the train examples, i.e., learning its paramaters.
-
In prediction, coordinates of Cα atoms in the original target protein structure are extracted. Then, its sub-structure corresponding to the templates is cropped out. Note that just like the templates, this sub-structure contains the gap region, but are bigger. Next, the distance matrix of the sub-structure is calculated and normalized using the normalization parameter calculated in the training phase, while any distance involving an atom in the gap region is set to 0. Figure 5 shows an example.
The area of non-gap region is used as the input to the trained DCNN. The output of the DCNN is a distance matrix for the extended gap region (the gap region with some extra amino acid on both sides), similar to the one in Figure 6(c). The output distance matrix is unnormalized and converted into a 3-D structure (Cα atoms only) using the multidimensional scaling method. Finally, Pulchra [27] is used to generate the full atom model from the predicted Cα model.
The full atom model of the extended gap region predicted by the DCNN will be used in the step 4 of the framework in Figure 1. The model is superimposed over the target protein structure based on their overlapping region.
2.3. Res Looper, another deep learning method for loop modeling
ResLooper uses the same idea as ConLooper to apply deep learning method for predicting loop conformation. Specifically, ResLooper uses Residualdue Neural Networks [35] to predict loop conformation instead of CNN as in ConLooper.
The Residual Neural Network (ResNet) was introduced recently in the ImageNet competition and the results were quite impressive. The basic structure of this network architecture is the building block, in which there are stacked activation layers and convolutional layers. If the input of a building block is Xl, then the output of this block is Xl + Xl+1 , in which Xl+1 is the non-linear transformation of Xl. We define a function indicating the difference between the input and output of the building block, then is called residual function. This feature of the network gives it the ability to learn something different from what the input has already encoded. Also, this network handles the vanishing gradient problem well.
In our network, the pre-activation [36] configuration of the building block was used, as shown in Figure 8. We have two activation layers and two convolutional layers together with ReLU as the activation function. In Figure 8, BN stands for Batch Normalization. BN is first proposed in the work of Ioffe et al. [44]. It can normalize the training mini-batch features in a layer in the neural network to zero mean and unit variance.
The ResLooper network architecture is shown in Figure 7. The input and output format of the ResLooper is the same as ConLooper, with the input as a 42 by 42 distance matrix and output as an 18 by 18 distance matrix. The network contains 9 residual blocks, plus the first initial convolutional layer and the last fully connected layer. There are 20 weighted layers totally in our network. The first initial convolutional layer has a kernel size of 3 and the number of filters is 16. Then there are 9 residual blocks and we divided them into 3 parts, each part has 3 residual blocks. Between each part, we use stride 2 to perform downsampling, wich results in the feature map sizes {42, 21, 11}, respectively. Then number of filters are {16, 32, 64}, respectively for each part. The kernel size is 3 for all convolutions. After the final fully connected layer, we reshape the output vector to a distance matrix.
In the building block, for the shortcut connection, there are two cases. For the case that the depth dimension changes, we use 1 by 1 convolution to match the dimension. For the case that the depth dimension does not change, we use identity connection.
2.4. HyLooper1 and HyLooper2, hybrid methods for loop modeling
The HyLooper1 method combines the results of NearLooper and ConLooper as follows:
Run NearLooper to output one candidate model, T1.
Run ConLooper to output another candidate model, T2.
Select between T1 and T2 based on the RMSD value of the overlapping region between each and the target protein structure. The one with smaller RMSD is selected as the model for the extended gap region.
Continue steps 4 and 5 of the framework in Figure 1. Basically, HyLooper1 is a hybrid method, aiming to improve over the two individual methods NearLooper and ConLooper.
The HyLooper2 method is very similar to HyLooper1 method but it uses ResLooper instead of ConLooper together with NearLooper to solve the loop modeling problem. HyLooper2 also follows the 4 steps as HyLooper1 but the method ConLooper in step 2 is replaced by ResLooper.
3. Results and discussion
3.1. Dataset and experimental settings
Four datasets were used in our experiments: The first dataset, called Set CASP, is chosen from The Critical Assessment of protein Structure Prediction experiments year 2014 (CASP11) [34]. This is a biennial world-wide event in the protein structure prediction community to assess state-of-the-art protein structure prediction methods. More specifically, the protein sequence search program (HHSearch) is used to search for possible candidate templates of target protein sequences in CASP11. The tool is configured to remove native-like protein from its results. The top template with its known 3D structure from the Protein Data Bank (PDB) dataset will be added to the Set CASP dataset if it satisfies the following criteria: (1) has a gap of length between 5 and 12; and (2) both sides of the gap can be extended to have a total length of 50 residues. In total, 9 proteins were selected for this training dataset.
The two other datasets from public data of Park et al. [14] are common benchmark sets used in the loop modeling community for testing performance of loop modeling methods. It composed of two subsets of 20 8-residues loops and 20 12-residues, called Set 8Res and Set 12Res, respectively. To create these sets, the first 2-ns molecular dynamics (MD) simulations at 300 K was performed, starting from the energy-minimized crystal structures using the AMER12 package [22]. Then, the Generalized Born/Surface Area (GB/SA) implicit solvation model [24, 25] and the AMBER99SB force field [23] were used [14].
The last dataset is called TBM set. It is comprised of 22 proteins with loops in more inaccurate environment of template-based models with the loop length varies from 6–11 residues. This dataset is also derived directly from the publication of Park et. al. [14] which is the template-based model set with 23 proteins. One exception is that the protein 1DF6A is removed from this set since its length is shorter than 50 residues. It was created by running MODELLER 9.6 [37] with the templates and multiple sequence alignments taken from the SALIGN benchmark study [38,39] on the HOMSTRAD set [40]. All targets for which the template-based models have GDTTS smaller than 70 or greater than 90 are excluded. The model consensus method for detecting unreliable modeled regions were used to select the target loop regions [14].
Totally, there are 71 proteins from 4 sets of data that have been used in our experiments. The loop length varies from 6 to 12 residues. The first dataset CASP is used for model selection while the other 3 datasets including set 8Res, 12Res and TBM are used for testing.
The training and validation are done for each protein of loop modeling. That means given a protein that is needed to do loop modeling, first a set of candidate templates will be created from Protein Data Bank by using 2 alignment tools PSI-BLAST [11] and HHSearch [12] on the sequence of given protein. All native-like templates in this set of templates are removed to make sure that there is no ground-truth data included in the training phase. The number of templates varies from 250 minimum to 617 maximum and 415 on average. Each template could be seen as a candidate to fill in the loop region of the protein. Then a deep learning method will be applied to train on the template set. The training/validation ratio is 75/25 on DCNN model and 90/10 on ResNet model.
Root-mean-square deviation (RMSD) of the main chain atoms N, Cα , C and O is used as the metric to compare performances between different loop modeling methods. These atoms are used since they can be seen as the representative atoms of the protein and can be used to recover to full atom model. In our experiments, our methods are compared with existing state-of-the-art methods with the results published on Park et. al [14]. These results use N, alpha, C and O atoms for RMSD measurement. Hence, we use the same set of atoms to make it comparable between our methods and other existing methods. This set of atoms has been used widely in different publications [12,13,14] for loop modeling problems. Its formula is defined in Eq. (1).
There are 3 current state-of-the-art methods including NGK [12], Galaxy PS1 [13] and Galaxy PS2 [14] that have been used as the baseline to compare with our newly developed methods for loop modeling problem. Experimental results of NGK, Galaxy PS1 and Galaxy PS2 are acquired directly from the published results of Park et. al. [14].
The average computation time of ConLooper on each loop modeling task is 1 CPU hours on a desktop computer with Intel Core i7, memory 16GB, graphics card NVID-IA GeForce GTX 980. NearLooper takes 0.5 CPU hour while ResLooper takes 1.5 CPU hours. HyLooper1 and HyLooper2 take 1.75 CPU hours and 2 CPU hours respectively.
MatConvNet toolbox version 1.0-beta20 is used as the DCNN implementation [26]. ResNet is implemented based on Keras [42] with Tensorflow [43] backend. The length of subsequence that is used for alignment is set to 50 while the size of overlapping region is set to 5.
3.2. Determining the best configuration for ConLooper
In this experiment, Set CASP is used to find the best configuration for ConLooper method so that it can be used later on the other datasets. Specifically, ConLooper method is applied on Set CASP with different number of hidden layers and different number of epochs. For each setting of hidden layers and epochs, ConLooper ran 10 times with different random initizlization and the average RMSD and standard deviation on each of 9 proteins in Set CASP are reported.
Table 1 shows performance of ConLooper method on different number of hidden layers. Four different settings are tested including 2 layers, 6 layers, 10 layers and 14 layers. The result shows that the setting of 10 layers is the best, with average RMSD of the main chain atoms 2.87 and standard deviation 0.92. In this experiment, the number of epochs in training is set to 500.
TABLE 1.
2 Layers | 6 Layers | 10 Layers | 14 Layers | |
---|---|---|---|---|
Average RMSD |
3.03 | 3.09 | 2.87 | 3.17 |
Std. dev. | 1.18 | 0.85 | 0.92 | 1.14 |
Average RMSD and standard deviation are calculated on set CASP with different configurations.
Table 2 shows performance of ConLooper on different number of epochs with the number of hidden layers set as 10. Four different settings are tested, including 100 epochs, 500 epochs, 1000 epochs, 1500 epochs and 200 epochs. The result shows that the setting of 500 epochs is the best, with-average RMSD of the main chain atoms 2.87 and standard deviation 0.92. Other settings either underfit or overfit the training data. Increasing from 1500 epochs to 2000 epochs seems to improve the RMSD and further increasing number of epoch could help. But due to resource limitatin, the max number of epochs is set to 2000.
TABLE 2.
100 Epochs | 500 Epochs | 1000 Epochs | 1500 Epochs | 2000 Epochs | |
---|---|---|---|---|---|
Average RMSD |
3.16 | 2.87 | 3.12 | 3.22 | 3.04 |
Std. dev. | 1.05 | 0.92 | 0.83 | 1.06 | 1.06 |
Average RMSD and standard deviation are calculated on set CASP with different configurations.
Table 3 shows the final configuration of CNN in ConLooper derived from the Set CASP with 9 proteins. The number of epochs is 500 while the number of hidden layers is 10. This configuration is used by ConLooper when compared with other loop modeling methods on other datasets.
TABLE 3.
Configuration | Value |
---|---|
Learning rate | 0.002 |
Number of epochs | 500 |
Number of hidden layers | 10 |
Batch size | 10 |
Training / Validation ratio | 75% / 25% |
3.3. Determining the best configuration for ResLooper
In this experiment, Set CASP is used to find the best configuration for ResLooper method so that it can be used later on the other test datasets. Specifically, ResLooper method is applied on Set CASP with different number of hidden layers and different number of epochs. For each setting of hidden layers and epochs, ResLooper ran 10 times with different random initizlization and the average RMSD and standard deviation on each of 9 proteins in Set CASP are reported.
Table 4 shows the performance of ResLooper method on different number of hidden layers. Four different settings are tested including 8 layers, 20 layers, 32 layers and 44 layers. The result shows that the setting of 20 layers is the best, with average RMSD of the main chain atoms 3.04 and standard deviation 1.27. In this experiment, the number of epochs in training is set to 500.
TABLE 4.
8 Layers | 20 Layers | 32 Layers | 44 Layers | |
---|---|---|---|---|
Average RMSD |
3.13 | 3.04 | 3.11 | 3.07 |
Std. dev. | 1.29 | 1.27 | 1.18 | 1.22 |
Average RMSD and standard deviation are calculated on set CASP with different configurations.
Table 5 shows performance of ResLooper on different number of epochs with the number of hidden layers set as 20. Four different settings are tested, including 100 epochs, 300 epochs, 500 epochs and 700 epochs. The result shows that the setting of 500 epochs is the best, withaverage RMSD of the main chain atoms 3.04 Å and standard deviation 1.27. Other settings either underfit or overfit the training data.
TABLE 5.
100 Epochs | 300 Epochs | 500 Epochs | 700 Epochs | |
---|---|---|---|---|
Average RMSD |
3.13 | 3.09 | 3.04 | 3.1 |
Std. dev. | 1.21 | 1.21 | 1.27 | 1.2 |
Average RMSD and standard deviation are calculated on set CASP with different configurations.
Table 6 shows the final configuration of ResNet in ResLooper derived from the Set CASP with 9 proteins. The number of epochs is 500 while the number of hidden layers is 20. This configuration is used by ResLooper when compared with other loop modeling methods on other test datasets.
TABLE 6.
Configuration | Value |
---|---|
Learning rate | 0.1 -> 0.01 (250 epochs) -> 0.001 (375 epochs) |
Number of epochs | 500 |
Number of hidden layers | 20 |
Batch size | 8 |
Training / Validation ratio | 90% / 10% |
3.4. Performance comparison of loop modeling methods
In this section, the new methods are compared with 3 state-of-the-art loop modeling methods, including Nextgeneration KIC (NGK), GalaxyLoop-PS1 (Galaxy PS1) and GalaxyLoop-PS2 (Galaxy PS2) on the Set 8Res, Set 12Res and Set TBM datasets. For each protein in those datasets, ConLooper and ResLooper runs 10 times to derive the average RMSD of the main chain atoms and standard deviation.
Table 7 shows the performances of the methods on 20 proteins in Set 8Res. Among the 3 existing methods, Galaxy PS2 achieves the best performance of average RMSD 2.02 Å and standard deviation 1.87. All five new methods achieve better results than the existing methods. HyLooper2 achieves the best performance with average RMSD 1.45 Å (28% improvement) and standard deviation 1.4. Individually, it outeperforms Galaxy PS2 in 65% (13/20 proteins) cases in this dataset.
TABLE 7.
PDB | NGK | Galaxy PS1 |
Galaxy PS2 |
NearLooper | ConLooper | HyLooper1 | ResLooper | HyLooper2 |
---|---|---|---|---|---|---|---|---|
135l | 3.9 | 3.7 | 4.3 | 0.6 | 1.5 (0.21) | 0.6 | 1.7 (0.68) | 0.6 |
1alc | 1.3 | 1.4 | 1.4 | 0.2 | 0.4 (0.06) | 0.2 | 0.9 (0.32) | 0.2 |
1btl | 0.4 | 1.3 | 0.9 | 0.2 | 1.4 (0.11) | 0.2 | 0.5 (0.04) | 0.2 |
1cex | 2.1 | 2 | 1.8 | 3.3 | 3 (0.27) | 3.3 | 3.3 (0.09) | 3.3 |
1clc | 0.4 | 0.4 | 0.3 | 4.5 | 4.3 (0.17) | 4.5 | 4.6 (0.05) | 4.5 |
1ddt | 3.7 | 2 | 1.5 | 4.1 | 3.4 (0.24) | 3.3 | 3.4 (0.23) | 3.5 |
1ezm | 4.3 | 4.2 | 3.8 | 0.9 | 1.7 (0.51) | 0.9 | 0.7 (0.11) | 0.9 |
1hfc | 0.7 | 1 | 0.9 | 3.6 | 3.5 (0.26) | 3.5 | 3.6 (0.07) | 3.6 |
1iab | 1 | 2.2 | 1.8 | 0.6 | 1.9 (0.19) | 0.6 | 0.7 (0.09) | 0.7 |
1ivd | 2.7 | 3.6 | 2.2 | 0.5 | 1.5 (0.32) | 0.5 | 1.1 (0.13) | 0.5 |
1lst | 1.2 | 1.1 | 1.1 | 0.3 | 1.5 (0.49) | 0.3 | 1.5 (0.18) | 0.3 |
1nar | 1.4 | 2.1 | 1.8 | 2.4 | 2.9 (0.3) | 2.4 | 2.8 (0.24) | 2.4 |
1oyc | 1.1 | 1.6 | 1.7 | 2.3 | 2.2 (0.1) | 2.3 | 2.3 (0.02) | 2.3 |
1prn | 8.3 | 6.9 | 8.8 | 3 | 2.7 (0.16) | 3 | 3.0 (0.13) | 2.9 |
1sbp | 0.9 | 0.8 | 0.8 | 1.4 | 1.6 (0.21) | 1.4 | 1.0 (0.13) | 0.9 |
1tml | 1.1 | 1.1 | 0.6 | 0.2 | 2.3 (0.46) | 0.2 | 0.5 (0.10) | 0.2 |
2cmd | 1.9 | 4 | 2.3 | 0.2 | 1.1 (0.38) | 0.2 | 0.3 (0.13) | 0.2 |
2exo | 1.5 | 1 | 1.1 | 0.4 | 1.2 (0.46) | 0.4 | 0.8 (0.10) | 0.8 |
2sga | 1.7 | 1.2 | 1.3 | 0.5 | 1.2 (0.42) | 0.5 | 3.3 (0.12) | 0.5 |
5p21 | 1.7 | 1.9 | 1.9 | 1.6 | 0.9 (0.1) | 1 | 0.6 (0.05) | 0.6 |
Average | 2.07 | 2.18 | 2.02 | 1.54 | 2.01 | 1.47 | 1.83 | 1.45 |
Std. dev. | 1.84 | 1.57 | 1.87 | 1.46 | 1.01 | 1.39 | 1.32 | 1.4 |
RMSD scores are used to compare performances among methods. For ConLooper and ResLooper, the experiment is executed 10 times to get the average RMSD and standard deviation for each protein. All the numbers are in Å.
Table 8 shows the performance of the methods on 20 proteins in Set 12Res. Similar to Table 7, among the 3 existing methods, Galaxy PS2 achieves the best performance of average RMSD 2.15 Å and standard deviation 1.4. It is also the best among all methods. Among the new methods, HyLooper1 is the best. Comparing with Galaxy PS2, HyLooper1 is comparable and outperforms Galaxy PS2 in 45% (9/20 proteins) cases in this dataset. Even though Galaxy PS2 achieves the best performance on Set 12Res, it has the disadvantage of needing initial loop information to do refinement while our methods can predict loop region directly without initial loop conformation.
TABLE 8.
PDB | NGK | Galaxy PS1 |
Galaxy PS2 |
NearLooper | ConLooper | HyLooper1 | ResLooper | HyLooper2 |
---|---|---|---|---|---|---|---|---|
1a8d | 3.7 | 4.5 | 3.1 | 2.8 | 2.6 (0.38) | 2.8 | 2.8 (0.14) | 2.8 |
1arb | 1.7 | 2.1 | 1.9 | 0.4 | 1.2 (0.78) | 0.4 | 1.1 (0.21) | 0.4 |
1bhe | 1.7 | 2.3 | 3.5 | 4.6 | 3.7 (0.28) | 3.4 | 3.5 (0.71) | 3.3 |
1bn8 | 1.1 | 4.3 | 1.1 | 4.8 | 4.4 (0.22) | 4.8 | 5.7 (0.4) | 4.8 |
1c5e | 1.2 | 2.2 | 1.5 | 1.7 | 2 (0.71) | 1.7 | 3.6 (0.22) | 1.7 |
1cb0 | 0.9 | 5.7 | 0.9 | 6 | 5.8 (0.65) | 6 | 5.9 (0.4) | 5.9 |
1cnv | 6.3 | 6.4 | 6.5 | 1.1 | 2.4 (0.9) | 1.1 | 1.2 (0.16) | 1.1 |
1cs6 | 1.1 | 1.7 | 1.6 | 2.3 | 2.3 (0.29) | 2.3 | 2.4 (0.23) | 2.3 |
1dqz | 7.5 | 1.5 | 3.3 | 0.2 | 1.3 (0.94) | 0.2 | 1.2 (0.23) | 0.2 |
1exm | 1.1 | 3 | 1.3 | 4.2 | 3.4 (0.25) | 4.2 | 3.9 (0.3) | 3.8 |
1f46 | 2.6 | 4.5 | 3.8 | 4.4 | 3.5 (0.59) | 3 | 3.7 (0.75) | 4.4 |
1i7p | 1.9 | 2.8 | 1.7 | 4.3 | 3.4 (0.25) | 4.3 | 4.3 (0.9) | 4.3 |
1m3s | 3.2 | 4.3 | 2.7 | 0.9 | 2.5 (0.54) | 0.9 | 1.5 (0.33) | 0.9 |
1ms9 | 1.8 | 1.8 | 1.8 | 5.2 | 4.7 (0.64) | 5.2 | 5 (0.11) | 5.2 |
1my7 | 0.9 | 2.4 | 1 | 0.5 | 1.6 (0.45) | 0.5 | 0.7 (0.8) | 0.7 |
1oth | 0.8 | 1.1 | 0.9 | 1.6 | 2 (0.43) | 1.6 | 1.4 (0.16) | 1.4 |
1oyc | 0.7 | 2.7 | 1.2 | 3.1 | 3 (0.27) | 3.1 | 3 (0.3) | 3.1 |
1qlw | 6 | 2.5 | 2.9 | 4.2 | 5.1 (0.59) | 4.2 | 5.1 (0.72) | 4.2 |
1t1d | 1.3 | 2.5 | 1.5 | 0.9 | 1.8 (0.64) | 0.9 | 1.1 (0.21) | 0.9 |
2pia | 0.7 | 4.5 | 0.7 | 2.5 | 2.9 (0.3) | 2.5 | 2.1 (0.45) | 2.5 |
Average | 2.31 | 3.14 | 2.15 | 2.79 | 2.98 | 2.65 | 3.01 | 2.69 |
Std. dev. | 2.03 | 1.46 | 1.4 | 1.82 | 1.27 | 1.73 | 2.67 | 1.74 |
RMSD scores are used to compare performances among methods. For ConLooper and ResLooper, the experiment is executed 10 times to get the average RMSD and standard deviation for each protein. All the numbers are in Å.
Finally, Table 9 shows the results of loop modeling between different loop modeling methods. All five newly developed methods show good performance on this dataset over the best existing method Galaxy PS2 with average RMSD 3.73 Å and standard deviation 1.46. HyLooper2 achieves the best results with average RMSD 1.69 Å (54% improvement) and standard deviation 1.07. Individually, it outeperforms Galaxy PS2 in 77% (17/22 proteins) cases in this dataset.
TABLE 9.
PDB | NGK | Galaxy PS1 |
Galaxy PS2 |
NearLooper | ConLooper | HyLooper1 | ResLooper | HyLooper2 |
---|---|---|---|---|---|---|---|---|
1a49a_463–472 | 5 | 2.3 | 2.7 | 0.8 | 2.2 (0.43) | 0.8 | 3.2 (0.7) | 0.8 |
1a49a_86–94 | 4.6 | 3.8 | 4.2 | 0.5 | 1.8 (0.97) | 0.5 | 1.2 (0.5) | 0.5 |
1asza_313–322 | 2.9 | 2.3 | 1.9 | 0.3 | 2.3 (0.41) | 0.3 | 0.7 (0.15) | 0.3 |
1avk_20–29 | 1.2 | 1.2 | 1.5 | 2.9 | 3.4 (0.34) | 2.9 | 3.4 (0.13) | 2.9 |
1buca_61–68 | 5.3 | 3.9 | 3.2 | 4.5 | 3.3 (0.63) | 3.3 | 3.2 (0.63) | 3.2 |
1csn_216–223 | 6.5 | 8.8 | 6.1 | 2.1 | 1.7 (0.73) | 1.8 | 1.5 (0.37) | 1.3 |
1d2ka_349–354 | 1.9 | 4 | 5.7 | 1.2 | 2.4 (1.32) | 1.2 | 2.5 (0.28) | 1.2 |
1e0ca1_53–60 | 3.8 | 6.7 | 5.9 | 2 | 2.7 (0.27) | 2 | 2.0 (0.45) | 1.8 |
1esl_40–48 | 3 | 3.2 | 2.4 | 2.6 | 2.3 (0.23) | 2.2 | 3.5 (0.16) | 2.6 |
1esl_54–59 | 1.8 | 2.6 | 1.5 | 0.2 | 1.8 (0.57) | 0.2 | 1.2 (0.76) | 0.2 |
1f3g_123–131 | 4 | 3.5 | 3.3 | 2.6 | 2.4 (0.31) | 2.6 | 2.7 (0.39) | 2.6 |
1gpb_261–271 | 4.3 | 4.4 | 3 | 3.7 | 3.3 (0.66) | 3.7 | 3.8 (0.66) | 3.7 |
1iala_32–42 | 6.8 | 4.6 | 4 | 0.5 | 2.8 (0.51) | 0.5 | 2.9 (0.16) | 0.5 |
1kbt_26–31 | 3.9 | 3.4 | 3.5 | 2.1 | 2.7 (0.43) | 2.8 | 2.2 (0.5) | 2.1 |
1lxa_102–107 | 3.6 | 3.8 | 4.7 | 0.7 | 2.7 (0.53) | 0.7 | 1.9 (0.5) | 0.7 |
1qdlb_59–67 | 4.2 | 2.7 | 2.8 | 3.7 | 2 (1.67) | 2.1 | 3.4 (0.21) | 3.2 |
1qu9a_93–98 | 7 | 6.6 | 6.3 | 2.8 | 2.8 (0.27) | 2.6 | 2.6 (0.2) | 2.8 |
1rsy_74–79 | 4 | 3 | 3.1 | 1 | 2 (1.07) | 1 | 0.5 (0.13) | 0.5 |
1tml_45–50 | 5.4 | 4.7 | 5.1 | 2 | 2.3 (0.89) | 2 | 2.2 (0.11) | 2.2 |
2oata_347–353 | 2.7 | 3 | 5 | 1.4 | 2.5 (0.28) | 1.4 | 2.1 (0.23) | 1.4 |
2pola2_23–32 | 2.8 | 4.3 | 3.7 | 1.6 | 1.8 (0.96) | 1.6 | 1.3 (0.42) | 1.1 |
3fib_40–47 | 2.8 | 2.8 | 2.6 | 1.7 | 1.9 (0.52) | 1.7 | 1.8 (0.1) | 1.7 |
Average | 3.97 | 3.89 | 3.73 | 1.85 | 2.41 | 1.72 | 2.26 | 1.69 |
Std. dev. | 1.57 | 1.69 | 1.46 | 1.19 | 0.5 | 1 | 0.94 | 1.07 |
RMSD scores are used to compare performances among methods. For ConLooper and ResLooper, the experiment is executed 10 times to get the average RMSD and standard deviation for each protein. All the numbers are in Å.
Comparing performance on the loop modeling task between our five newly developed methods on test datasets, HyLooper2 is the best one in general. Between two deep learning methods, ResLooper is better than ConLooper which implies the better performance of ResNet over DCNN.
4. Conclusion
This paper presents five new loop modeling methods, in particular a new approach based on deep learning (DCNN, ResNet) and distance matrix representation of 3-D structures. From the rotation and translation independent representation of distance matrix, DCNN and ResNet are able to learn a mapping function to predict a loop model. To the best of our knowledge, this is the first attempt to combine deep learning and geometric information of distance matrix for loop modeling task.
Experiments using selected CASP models and common benchmark datasets have shown promising results. Compared to state-of-the-art loop modeling methods, our HyLooper2 achieves the best result on two test datasets and being comparable on the other test dataset. The ResNet in ResLooper is able to generate good loop candidates in many cases. The HyLooper2 method, which is the combination of ResLooper and NearLooper achieves good performance because both methods are complementary and have their own strengths.
Due to the limitation of computing power, experiments have been done on a limited sets of different network configurations. Exploring more network configurations can help improve the overall performance of loop modeling. This would be a good direction for future work.
Acknowledgment
This work was supported in part by National Institutes of Health grant R01-GM100701. The high-performance computing infrastructure is supported by the National Science Foundation under grant number CNS-1429294.
Biography
Son Phong Nguyen received the M.S. degree in Computer Science in 2014 and the Ph.D. degree in Computer Science in 2017 from the University of Missouri, USA. His current research interests focus on machine learning, deep learning, big data analytics and bioinformatics.
Zhaoyu Li received the M.S. degree in Computer Science from the University of Missouri, USA in 2014. He is currently a Ph.D. candidate in Computer Science from University of Missouri, USA. His current research interests focus on machine learning, deep learning, and bioinformatics.
Dong Xu is James C. Dowell Professor of Computer Science Department, Director of Information Technology Program, with appointments in the Christopher S. Bond Life Sciences Center and the Informatics Institute at the University of Missouri-Columbia. He obtained his Ph.D. from the University of Illinois, Urbana-Champaign in 1995 and did two years of postdoctoral work at the US National Cancer Institute. He was a Staff Scientist at Oak Ridge National Laboratory until 2003 before joining the University of Missouri, where he served as Department Chair of Computer Science during 2007–2016. He has published nearly 300 papers. He is a Fellow of American Association for the Advancement of Science (AAAS).
Yi Shang is Professor, Associate Chair and Director of Graduate Studies, Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, Missouri. He received Ph.D. in Computer Science from University of Illinois at Urbana-Champaign in 1997, M.S. from the Institute of Computing Technology, Chinese Academy of Sciences, Beijing, in 1991, and B.S. from University of Science and Technology of China, Hefei, in 1988. He has published over 160 refereed papers in the areas of nonlinear optimization, artificial intelligence, wireless sensor networks, mobile computing, and bioinformatics and was granted 6 patents. His research has been supported by NSF, NIH, Army, DARPA, Microsoft, and Raytheon. http://dslsrv1.rnet.missouri.edu/~shangy.
References
- [1].Johnson MS, Srinivasan N, Sowdhamini R, and Blundell TL, “Knowledge-based protein modeling,” Crit Rev Biochem Mol Biol, vol. 29, pp. 1–68, 1994. [DOI] [PubMed] [Google Scholar]
- [2].Petrey D, Honig B, “Protein structure prediction: inroads to biology,” Mol Cell, 20:811–819, 2005. [DOI] [PubMed] [Google Scholar]
- [3].Wang Z, Moult J, “SNPs, protein structure, and disease,” Hum Mutat, 17:263–270, 2001. [DOI] [PubMed] [Google Scholar]
- [4].Fiser A, “Protein structure modeling in the proteomics era,” Expert Rev Proteomics, 1:97–110, 2004. [DOI] [PubMed] [Google Scholar]
- [5].Zhang J, Wang Q, Barz B, He Z, Kosztin I, Shang Y, and Xu D, “MUFOLD: A new solution for protein 3D structure prediction”. Proteins, 78(5): 1137–1152, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [6].Berman HM, Westbrook J, Feng Z, Gilliland G, Bhat TN, Weissig H, Shindyalov IN, Bourne PE, “The Protein Data Bank,” Nucleic Acids Res, 28:235–242, 2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [7].Soto CS, Fasnacht M, Zhu J, Forrest L, and Honig B, “Loop modeling: Sampling, filtering, and scoring,” Proteins: Structure, Function, Bioinformatics, 2008. [DOI] [PMC free article] [PubMed]
- [8].Xiang Z, Soto CS, Honig B, “Evaluating conformational free energies: the colony energy and its application to the problem of loop prediction,” Proc Natl Acad Sci USA, 99:7432–7437, 2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Jacobson MP, Pincus DL, Rapp CS, Day TJ, Honig B, Shaw DE, Friesner RA, “A hierarchical approach to all-atom protein loop prediction,” Proteins, 55:351–367, 2004. [DOI] [PubMed] [Google Scholar]
- [10].Fiser A, Do RK, Sali A, “Modeling of loops in protein structures,” Protein Sci, 9:1753–1773, 2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [11].Michalsky E, Goede A, Preissner R, “Loops in proteins (LIP)—a comprehensive loop database for homology modelling,” Protein Eng, 16:979–985, 2003. [DOI] [PubMed] [Google Scholar]
- [12].Stein A, Kortemme T, “Improvements to robotics-inspired conformational sampling in rosetta,” PLoS One, 8: e63090, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Park H, Seok C, “Refinement of unreliable local regions in template-based protein models,” Proteins, 80: 1974–1986, 2012. [DOI] [PubMed] [Google Scholar]
- [14].Park H, Lee GR, Heo L, Seok C, “Protein loop modeling using a new hybrid energy function and its application to modeling in inaccurate structural environments,” PLoS One, 9(11): e113811, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [15].Shenkin PS, Yarmush DL, Fine RM, Wang HJ, Levinthal C, “Predicting antibody hypervariable loop conformation. I. Ensembles of random conformations for ringlike structures,” Biopolymers, 26:2053–2085, 1987. [DOI] [PubMed] [Google Scholar]
- [16].Altschul S, Madden T, Schäffer A, Zhang J, Zhang Z, Miller W, Lipman D, “Gapped BLAST and PSIBLAST: a new generation of protein database search programs,” Nucleic Acids Res, 25:3389–3402, 1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].Soding J, “Protein homology detection by HMM-HMM comparison,” Bioinformatics, 21:951–960, 2005. [DOI] [PubMed] [Google Scholar]
- [18].Borg I, Groenen P, “Modern multidimensional scaling—theory and applications,” New York: Springer-Verlag, 1997. [Google Scholar]
- [19].Nguyen SP, Shang Y, and Xu D, “DL-PRO: A novel deep learning method for protein model quality assessment,” International Joint Conference on Neural Networks, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [20].Convolutional Neural Networks for Visual Recognition Available online at: http://cs231n.github.io/convolutional-Networks/
- [21].LeCun Y and Bengio Y, “Convolutional Networks for images, speech, and time series,” The Handbook of Brain Theory and Neural Networks, 255–258. MIT Press, Cambridge, Massachusetts, 1995. [Google Scholar]
- [22].Case DA, Darden TA, Cheatham TE, Simmerling CL, Wang J, et al. AMBER 12 University of California, San Francisco: 2012 [Google Scholar]
- [23].Hornak V, Abel R, Okur A, Strockbine B, Roitberg A, et al. , “Comparison of multiple amber force fields and development of improved protein backbone parameters,” Proteins, 65: 712–725, 2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [24].Still WC, Tempczyk A, Hawley RC, Hendrickson T, “Semianalytical treatment of solvation for molecular mechanics and dynamics,” Journal of the American Chemical Society, 112: 6127–6129, 1990. [Google Scholar]
- [25].Qiu D, Shenkin PS, Hollinger FP, Still WC, “The GB/SA continuum model for solvation. A fast analytical method for the calculation of approximate Born radii,” Journal of Physical Chemistry A, 101: 3005–3014, 1997. [Google Scholar]
- [26].Convolutional Neural Networks toolbox Available online at: http://www.vlfeat.org/matconvnet/
- [27].Rotkiewicz P, Skolnick J, “Fast procedure for reconstruction of full-atom protein models from reduced representations,” Journal of computational chemistry, 29(9):1460–5, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [28].Nair V, Hinton GE, “Rectified linear units improve restricted Boltzmann machines,” International Conference on Machine Learning, 807–814, 2010. [Google Scholar]
- [29].LeCun Y, Boser B, Denker JS, Henderson D, Howard RE, Hubbard W, Jackel LD, “Backpropagation applied to handwritten zip code recognition,” Neural computation, 541–551, 1989.
- [30].Krizhevsky A, Sutskever I, Hinton GE, “ImageNet classification with deep convolutional neural networks,” Advances in Neural Information Processing Systems, 1097–1105, 2012.
- [31].Ouyang W, Luo P, Zeng X, Qiu S, Tian Y, Li H, Yang S, Wang Z, Xiong Y, Qian C, Zhu Z, Wang R, Loy CC, Wang X, Tang X, “Deepid-net: multi-stage and deformable deep convolutional neural networks for object detection,” arXiv preprint arXiv:14093505, 2014. [Google Scholar]
- [32].Zhang N, Donahue J, Girshick R, Darrell T, “Part-based RCNNs for fine-grained category detection,” European Conference on Computer Vision, 834–849, 2014. [Google Scholar]
- [33].Sun Y, Chen Y, Wang X, Tang X, “Deep learning face representation by joint identification-verification,” Advances in Neural Information Processing Systems, 1988–1996, 2014.
- [34].Available online at: http://www.predictioncenter.org/casp11/index.cgi
- [35].He K, Zhang X, Ren S, Sun J, “Deep residual learning for image recognition,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016. [Google Scholar]
- [36].He K, Zhang X, Ren S, Sun J, “Identity mappings in deep residual networks,” European Conference on Computer Vision Springer International Publishing, 2016. [Google Scholar]
- [37].Sali A, Blundell TL, “Comparative protein modeling by satisfaction of spatial restraints,” Journal of Molecular Biology, 234:779–815, 1993. [DOI] [PubMed] [Google Scholar]
- [38].Marti-Renom MA, Madhusudhan MS, Sali A, “Alignment of protein sequences by their profiles,” Protein Science, 13:1071–1087, 2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [39].Braberg H, Webb BM, Tjioe E, Pieper U, Sali A, et al. , “SALIGN: a web server for alignment of multiple protein sequences and structures,” Bioinformatics, 28:2072–2073, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [40].Mizuguchi K, Deane CM, Blundell TL, Overington JP, “HOM-STRAD: a database of protein structure alignments for homologous families,” Protein Science, 7:2469–2471, 1998. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [41].Yang C, Lu X, Lin Z, Shechtman E, Wang O, Li H, “High-resolution image inpainting using multi-scale neural patch synthesis,” arXiv, 2016
- [42].Chollet F, Keras, GitHub repository: https://github.com/fchollet/keras
- [43].Abadi Martín, et al. “Tensorflow: Large-scale machine learning on heterogeneous distributed systems.” arXiv, 160304467, 2016. [Google Scholar]
- [44].Ioffe Sergey, and Szegedy Christian. “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” International Conference on Machine Learning, 2015. [Google Scholar]
- [45].Arnautova YA, Abagyan RA, Totrov M, “Development of a new physics-based internal coordinate mechanics force field (ICMFF) and its application to protein loop modeling,” Proteins, 2012. [DOI] [PMC free article] [PubMed]
- [46].Rapp CS, Friesner RA, “Prediction of loop geometries using a generalized born model of solvation effects,” Proteins, 1999. [PubMed]
- [47].Zhang H , Lai L, Wang L, Han Y, Tang Y, “A fast and efficient program for modeling protein loops,” Biopolymers, 1999.
- [48].McGarrah DB, Judson RS, “Analysis of the genetic algorithm method of molecular conformation determination,” Journal of computational chemistry, 1993.
- [49].Ring CS, Cohen FE, “Conformational sampling of loop structures using genetic algorithms,” Israeli journal of chemistry, 1994.
- [50].Shenkin PS, Yarmush DL, Fine RM, Wang HJ, Levinthal C, “Predicting antibody hypervariable loop conformation. I. Ensembles of random conformations for ringlike structures,” Biopolymers, 1987. [DOI] [PubMed]
- [51].Wedemeyer W, Scheraga HA, “Exact analytical loop closure in proteins using polynomial equations,” Journal of computational chemistry, 1999. [DOI] [PubMed]
- [52].Soto CS, Fasnacht M, Zhu J, Forrest L, Honig B, “Loop modeling: Sampling, filtering, and scoring,” Proteins, 2008. [DOI] [PMC free article] [PubMed]
- [53].Olson MA, Feig M, Brooks CL 3rd, “Prediction of protein loop conformations using multiscale modeling methods with physical energy scoring functions,” Journal of Computational Chemistry, 2008. [DOI] [PubMed]
- [54].de Baker PI, DePristo MA, Burke DF, Blundell TL, “Ab initio construction of polypeptide fragments: Accuracy of loop decoy discrimination by an all-atom statistical potential and the AMBER force field with the Generalized Born solvation model,” Proteins, 2003. [DOI] [PubMed]