Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2019 Dec 19.
Published in final edited form as: J Phys Chem Lett. 2019 Jul 30;10(16):4558–4565. doi: 10.1021/acs.jpclett.9b01570

A Multi-Resolution 3D-DenseNet for Chemical Shift Prediction in NMR Crystallography

Shuai Liu , Jie Li , Kochise C Bennett , Brad Ganoe , Tim Stauch , Martin Head-Gordon , Alexander Hexemer , Daniela Ushizima , Teresa Head-Gordon §
PMCID: PMC6922005  NIHMSID: NIHMS1062653  PMID: 31305081

Abstract

We have developed a deep learning algorithm for chemical shift prediction for atoms in molecular crystals that utilizes an atom-centered Gaussian density model for the 3D data representation of a molecule. We define multiple channels that describe different spatial resolutions for each atom type that utilizes cropping, pooling, and concatenation to create a multi-resolution 3D-DenseNet architecture (MR-3D-DenseNet). Because the training and testing time scale linearly with the number of samples, the MR-3D-DenseNet can exploit data augmentation that takes into account the property of rotational invariance of the chemical shifts, thereby also increasing the size of the training dataset by an order of magnitude without additional cost. We obtain very good agreement for 13C, 15N, and 17O chemical shifts when compared to ab initio quantum chemistry methods, with the highest accuracy found for 1H chemical shifts that is comparable to the error between the ab initio results and experimental measurements. Principal Component analysis (PCA) is used to both understand these greatly improved predictions for 1H, as well as indicating that chemical shift prediction for 13C, 15N, and 17O, which have far fewer training environments than the 1H atom type, will improve once more unique training samples are made available to exploit the deep network architecture.

Graphical Abstract

graphic file with name nihms-1062653-f0001.jpg


Nuclear magnetic resonance (NMR) crystallography is an experimental technique to determine the structure of complex materials,1,2 biomolecules such as proteins,3,4 as well as small molecules and pharmaceuticals57 in the solid state. In practice, NMR crystallography is a structural model building procedure that depends on a number of NMR data types, of which chemical shifts in particular play a prominent role. A strength of NMR chemical shift data is its excellent sensitivity to hydrogen8 which, given the importance of hydrogen-bonding in most molecular systems, makes it very complementary to X-ray diffraction techniques.

In the case where little is known about the chemical bonding of an unknown structure, the experimental measurements for chemical shifts are compared to the results of ab initio methods based on density functional theory (DFT), typically using Gauge-Including Projector-Augmented Waves (GIPAW) methods.9 However, because of the cubic computational complexity scaling with the number of atoms (O(N3)), alternative methods are being actively investigated to mitigate its large computational cost, especially for large systems. Many of these more inexpensive approaches are focused on fragment models that incorporate the long-range many-body polarization effects of the lattice environment via electrostatic embedding, such as the self-consistent reproduction of the Madelung potential (SCRMP),10 which has yielded very high quality results when combined with hybrid DFT functionals.

An alternative approach is to apply machine learning methods to predict the experimental and/or DFT results for systems ranging from proteins in solution1115 to solid-state materials.1,2,16 Cuny et al. reported a fully connected shallow neural network to predict the quadrupolar couplings and chemical shifts in silica materials for 17O and 29Si using symmetric functions of the Cartesian coordinates as the input.17 Paruzzo et al. applied the kernel ridge regression (KRR) using a smooth overlap of atomic positions (SOAP) kernel, that also directly incorporates rotational invariance of the chemical shift value to applied magnetic field, for molecular crystal systems.18 However, the KRR approach requires O(N2) complexity for calculating the similarity kernel matrix, and quadratic-to-cubic complexity for kernel matrix inversion, which is ultimately not tenable for large training and testing datasets.

Convolutional neural networks (CNNs) have been applied to several problems in chemistry and biology, such as enzyme classification,19 molecular representation,20 amino acid environment similarity,21 and potential energy prediction.22 They have not to the best of our knowledge been applied to NMR crystallography property prediction. There are a number of deep network variants that have been developed to address important deficiencies of a vanilla CNN, which are hard to train because of the vanishing (or exploding) gradient problem. This is because the repeated application of non-linear activation functions cause later outputs in the deep layers to flatten out, and back-propagated gradients are then diminished.

Residual networks (ResNets) were developed to precondition the network to learn the residual of a non-linear mapping by referencing it to an identity mapping, which is easier to train due to the presence in the network architecture of “identity shortcut connections”.23 Because these network connections skip layers, there is more direct information flow from the loss function to correct the weight parameters of earlier layers. DenseNets build on these ideas by also utilizing skipped connections for better gradient flow, while at the same time also performing concatenation of feature maps that permits greater propagation and reuse of features in what is termed “deeper supervised learning”.24

Here we report a deep learning approach to predict chemical shifts in the solid-state for hydrogen (1H), carbon (13C), nitrogen (15N) and oxygen (17O) that outperforms KRR while also allowing for chemical interpretation of the results using principal component analysis (PCA). The deep learning approach is based on a multi-resolution (MR) spatial data representation, where each resolution level and atom type is formulated as an independent channel of a deep learning 3D-DenseNet architecture, augmented with special concatenation of pooling layers (at reduced resolution) with cropping of feature maps (retaining high resolution features with reduced size) of the transformed spatial data. The resulting MR-3D-DenseNet removes the restrictions imposed by KRR, i.e. the need to build in rotational invariance of chemical shifts as well as the limitations to small data sets,18 in order to take advantage of a data augmentation procedure in which we rotate the chemical environment for each atom in a sample, thereby increasing the data set size by close to an order of magnitude with little computational expense.

Using the greater capacity of the MR-3D-DenseNet deep network, we obtain significant improvements for 13C, 15N, and 17O chemical shifts over KRR, with excellent agreement for 1H chemical shift prediction with RMSE error of 0.37 ppm, which is the same level of error between ab initio calculations and experimental measurements. The PCA allows us to both understand these improvements, as well as interpreting the remaining deficiencies, for chemical shift prediction for all atom types. In particular the PCA suggests that chemical shift prediction for 13C, 15N, and 17O, which have far fewer training environments than the 1H atom type, will improve once more unique training samples are made available to exploit the deep network architecture. Given the far better computational scaling of the multi-resolution 3D-DenseNet, we can afford to address this deficiency with much larger data sets than currently available in future studies.

Results

Data Representation.

The molecular crystal structures are from the Cambridge Structural Database (CSD),25 comprising 2,000 crystal structures in the training dataset and 500 crystal structures in the testing dataset. The coordinates of atoms in the unit cell, and the corresponding calculated chemical shieldings, are as given in the reported literature by Paruzzo and co-workers.18 In that paper, the training data was generated by conducting farthest data sampling to yield 2,000 crystal structures from which to extract training examples, whereas for testing they utilized uniform sampling to yield 500 crystal structures from which to extract testing examples, although there are 61,000 available structures in the CSD. This is reasonable given the cost of the underlying DFT calculations for chemical shift values, and also because of limitations due to the unfavorable scaling of the KRR which we discuss later. This resulted in the number of unaugmented 3D samples for training and testing for each of the atom types as given in Table 1. No further data selection or cleaning procedures are applied to the original dataset, except that 0.05% outliers (chemical shielding < 0 or > 40) in the 1H-NMR training dataset were removed.

Table 1:

The number of samples in training and testing datasets with and without data augmentation.

Atom Type Number of Samples
Training Dataset Testing Dataset
w/o Augmentation w/ Augmentation w/o Augmentation w/ Augmentation
1H 76,174 609,392 29,913 239,304
13C 58,148 465,184 26,607 212,856
15N 27,814 222,512 2,713 21,704
17O 25,924 207,392 5,404 43,232

Given the limited number of examples in the training dataset, we apply a physically motivated data augmentation method to improve the prediction performance of the MR-3D-DenseNet model. Since the chemical shift is invariant under rotational operations, we augment the data by rotating the Cartesian coordinates of atoms randomly with the Euler angles uniformly distributed between [π2,π2] along each of x, y and z axis. During the training phase, both the original data and augmented data are included in the training dataset. During the testing phase, we average the prediction results among 8 different rotation configurations. The final number of training and testing examples after this augmentation are given in Table 1.

The input data representation to the MR-3D-DenseNet assumes that chemical shifts are sensitive to the electron density distribution of atoms in molecules. Hence a molecule is represented on a 3D grid in which each atom takes on a radial Gaussian density. The 3D image is a bounded box with 16 × 16 × 16 voxels, with the density D(r) at each voxel taken as a sum of Gaussian distributions from all of the atoms

D(r)=rAexp(rr2σ2) (1)

where the summation runs over atoms of a given atom type A and the r′ are the corresponding atomic centers. The coordinate r = (x, y, z) at the center of voxel (with index (i, j, k)) is calculated as

r=(x,y,z)=((i152)d15,(j152)d15,(k152)d15) (2)

where d is the grid resolution. Unlike the Gaussian smearing method reported in literature,20 we calculate the density at the center of the voxel numerically using 16-bit floating point numbers. We also considered additional electron density representations including Slater orbitals and calculated from the inverse Fourier transform of the atomic form factor, but found that they performed worse than the Gaussian representation that can be explained by their heavy tails (see Supplementary Information).

The atom whose chemical shift is being evaluated is placed at the center of the 3D grid, and its chemical environment is represented by calculating the density under different grid sizes, where d = 4 Å, 6 Å, 8 Å, 10 Å, and 14 Å, each of which is represented by a dedicated channel in the MR-3D-DenseNet model. Under each grid size, we divide the density based on the atom types into 4 different channels for 1H, 13C, 15N, 17O, respectively, resulting in a total of 20 separate channels in the MR-3D-DenseNet network. Figure 1 shows a visualization of the carbon channels of the molecule (Z)-2-Hydroxy-3,3’,4’-trimethoxystilbene (reported by Stomberg et al.27) at two different grid size resolutions.

Figure 1:

Figure 1:

Visualization of the Gaussian densities of atoms on different grid sizes. Representative example is shown for carbon channels on (a) 4 Å and (b) 10 Å grid. The densities are visualized through Mayavi package.26

Machine Learning Models.

In this study, we start from a variant of the DenseNet architecture that omits skip connections going directly to the Dense block output. In addition, we designed a modification to the DenseNet that is motivated by the hypothesis that the importance of a given voxel increases as the distance between it and the investigated atom decreases, which is represented by multi-resolution channels. A schematic of the MR-3D-DenseNet architecture is shown in Figure 2, and is comprised of a regular 3 × 3 × 3 convolutional layer followed by two DenseNet blocks with a 1×1×1 transition convolutional layer in between them. The flattened output from the last DenseNet block is then fully connected to a layer with 256 units which is fully connected to a 128 unit layer, which is then fully connected to the output layer.

Figure 2:

Figure 2:

Illustration of the overall architecture of the MR-3D-DenseNet model. (a) Flowchart of the network (b) Illustration of 3 × 3 × 3 convolution layer prior to the first dense block (c) Illustration of the repeating unit in DenseNet block that contains two 1 × 1 × 1 convolution layers followed by a 3 × 3 × 3 convolution layer (d) Illustration of the cropping layer from the center of the feature map.

Each DenseNet block has four repeating units: each repeating unit has two 1 × 1 × 1 bottleneck convolutional layers with 256 and 64 channels followed by a 3 × 3 × 3 convolution layer with 64 channels. The MR-3D-DenseNet utilizes cropping and pooling such that at the end of each block, we concatenate the 2 × 2 × 2 average pooling layer and the cropping of the center segment of the feature map with the same size (l2, where l is the current feature map size). This retains low and high resolution features throughout the deep layers. After the concatenation of pooling and cropping, there are two 1 × 1 × 1 convolutional layers with 256 and 64 channels, respectively, to process the information and retain the channel size to 64 before entering the next block. Using this network architecture, we describe the detailed training protocol and hyperparameters in Methods.

Prediction Performance.

The performance on chemical shift predictions for all atoms using MR-3D-DenseNet compared to KRR is summarized in Table 2. We emphasize that cross-validation was applied to the training data only, and the testing results are “out of sample” predictions. The testing RMSEs of chemical shifts for 1H, 13C, 15N, and 17O using the MR-3D-DenseNet architecture is found to be 0.37 ppm, 3.3 ppm, 10.2 ppm and 15.3 ppm, which are 24%, 23%, 23% and 14% lower than the RMSEs given by a KRR method.18 Among the four atom types, the error of 1H prediction is comparable to the error between high standard ab initio calculations and experiment, i.e. GIPAW/PBE and SCRMP/PBE0 which have chemical shift accuracy RMSEs of 0.43 ppm and 0.33 ppm, respectively.10,28 Although the predictions on the other atom types are very good, we attribute their lessened performance with respect to ab initio models as a lack of unique data compared to that available for 1H (Table 1), a point to which we return to later.

Table 2:

Testing RMSEs (ppm) using MR-3D-DenseNet. We also report the improvement of RMSE in percentage compared to KRR18 and the R2 values using MR-3D-DenseNet.

Atom Type MR-3D-DenseNet R2
H 0.37 (24%) 0.9856
C 3.3 (23%) 0.9957
N 10.2 (23%) 0.9916
O 15.3 (14%) 0.9933

In a separate publication, we will present a full study of different deep learning architectures, but here we contrast the best MR-3D-DenseNet model to the KRR machine learning method for which results are available on the same chemical shift problem.18 We can attribute the success of the MR-3D-DenseNet approach based on three factors: (1) the greater flexibility in input representation of individual atom types and spatial resolution, and the advantages of concatenation of the pooling and cropping operations in the architecture, (2) the dependence on the size and quality of the training set, and (3) the ability to learn chemical bonding features, all of which are unique to chemical shift prediction using the MR-3D-DenseNet architecture.

In regards the first point we decompose the MR-3D-DenseNet result based on its multi-resolution input representation with no special concatenation of pooling and cropping operations (MR-NoConcat) versus the network architecture that utilizes concatenation of pooling and cropping but takes in only a single resolution input representation (SR-Concat). It is evident that the input and architecture features trained in isolation of each other offer significant improvements in performance over KRR, with further benefit being realized by their combined use in MR-3D-DenseNet (Table 3).

Table 3:

Testing RMSEs (ppm) for KRR and using different features of the MR-3D-DenseNet model for each atom type: SR-Concat, MR-NoConcat, and MR-3D-DenseNet. For the single-resolution input, the SR-Concat model is sensitive to the grid size for a given atom type and an optimized value must be determined (parentheses).

Atom Type KRR SR-Concat MR-NoConcat 3D-MR-DenseNet std
H 0.49 0.38 (10 Å) 0.38 0.37 < 0.01
C 4.3 3.5 (6 Å) 3.5 3.3 < 0.1
N 13.3 10.2 (8 Å) 10.3 10.2 0.2
O 17.7 16.3 (6 Å) 15.6 15.3 0.5

The main limitation of the SR-Concat model is that the 3D-grid size of the single resolution input depends on the atom under consideration, and even when the grid size is optimized it gives only limited improvement for the oxygen chemical shift. Hence the multi-resolution input representation is undoubtedly very important, and the novel feature of our approach is its greater flexibility compared to KRR which must determine the weights for mixed-scale kernels that use different cut-off sizes.18 Although the concatenation of pooling and cropping operations appears to play a more limited role for some of the atom types (MR-NoConcat), it does improve the prediction for carbon as judged by the standard deviation.

Furthermore, the success of a deep learning network model will be highly dependent on the size, variety and quality of the training dataset. To understand the effect of the training data size, we examine the 1H chemical shift testing RMSE for KRR and the MR-3D-DenseNet model as a function of increasing number of training examples (Figure 3a). Without data augmentation, the prediction performance of MR-3D-DenseNet improves over KRR after being presented ~ 1500 training points. However, the MR-3D-DenseNet has the capacity to exploit the augmented data to outperform the KRR model even with only a hundred training examples.

Figure 3:

Figure 3:

Testing RMSEs and timings for 1H chemical shift for different numbers of samples using the MR-3D-DenseNet. (a) using no augmentation (red), with training dataset 8-fold augmentation (green), using both training and testing dataset with 8-fold augmentation (blue), and compared to the testing error reported previously for KRR on the same dataset18 (black). The models are trained under the same number of batches to obtain a fair comparison; for example, when the data is augmented by 8-fold, the number of training epochs decrease to 1/8. (b) Training (8-fold) time of MR-3D-DenseNet model for the 1H chemical shift under the same network architecture and number of epochs. The testing time (1-fold) of 1H chemical shift is about 4–5 minutes for 500 preprocessed testing structures and is independent on the number of training structures. The training and testing time are benchmarked on Nvidia Tesla P100 GPU.

Although it might be argued that the KRR model has no need for the augmented data, since rotational invariance is built directly into the kernel, the data augmentation is clearly doing something more than invoking the rotational invariance feature of the chemical shift (i.e. the performance would be the same otherwise). In addition, augmentation of the testing dataset can be seen as equivalent to an ensemble averaging prediction without the need to retrain many networks to realize the same benefit, lowering the testing RMSE further to realize the best MR-3D-DenseNet performance (Figure 3a).

How to discover what the data augmentation is providing can’t be easily calculated with KRR due its unfavorable computational scaling for kernel matrix computation and kernel matrix inversion. By contrast the training time for the MR-3D-DenseNet model scales linearly with the number of training samples (Figure 3b). More importantly, the prediction time of MR-3D-DenseNet with a trained model does not scale with the number of training examples, whereas the testing time for KRR scales linearly because the similarity kernel has to be calculated using all of the training samples.

In totality, the MR-3D-DenseNet architecture with data augmentation yields a much tighter prediction error across the unique data across all atom types relative to KRR as seen in Figure 4. We found that further increasing the data augmentation to 16-fold rotations or adding the effects of small vibrational smearing of atom positions had a neutral effect on the prediction performance. Instead Figure 4 emphasizes that creating more unique data for the heavy atoms will certainly improve the MR-3D-DenseNet performance relative to ab initio models, as the number of heavy atom samples are limited compared to 1H samples in the current dataset. This may also explain why the prediction performance was limited for 15N and 17O when using MR-NoConcat (Table 3) because there was insufficient data to exploit the deep network architecture design.

Figure 4:

Figure 4:

Histogram of testing error distribution comparing MR-3D-DenseNet and KRR for (a) 1H, (b) 13C, (c) 15N and (d) 17O.

Finally, we interpret the MR-3D-DenseNet model using PCA to extract chemical bonding and hydrogen-bonding information derived from the transformed data in the last fully connected layer from the 1H chemical shift prediction, and projecting the first 3 principal components into a reduced 3D space as shown in Figure 5. Even though no explicit bonding information was provided as input to the MR-3D-DenseNet network, the model is capable of separating the C-H from the N-H and O-H chemical bond clusters (Figure 5a), although there is less well-defined separability of the different types of chemical and hydrogen-bonding environments for the N-H and O-H bonds (Figure 5a and Figure 5b). The relative lack of clean separability for the N-H and O-H bonding clusters exhibited by the PCA analysis would support the conclusion that there is a lack of unique chemical environments for oxygen and nitrogen, since there are more C-H than N-H or O-H examples in the training dataset. We caution that it is possible that the chemical or hydrogen-bonded clusters for the O-H and N-H data may well be separated in a higher dimensional space, however the three principal components shown here can explain > 85% of the variances of the data (Figure S2 in the Supplementary Information).

Figure 5:

Figure 5:

Visualization of the data in the last fully connected layer by projecting the data into 3D space using principal component analysis (PCA). It shows the clustering of different (a) chemical bonds and (b) hydrogen bonds.

Conclusion

We have presented a 3D-DenseNet deep learning model that exploits a multi-resolution input representation for NMR chemical shift prediction of atoms in molecular crystals. A unique feature of our deep learning model is the use of multi-resolution spatial input data organized into individualized channels for each atom type, that provides a learning framework for different tasks which are sensitive to the density representation with different cut-off sizes. Furthermore, the multi-resolution architecture combines the benefits of both pooling and the high-resolution feature map close to the investigated atom, which can potentially be applied to the prediction of other chemical properties sensitive to different length scales.

In addition to its greater flexibility in representing spatial density distributions, the MR-3D-DenseNet can more efficiently handle much larger data sets. In comparison to the KRR approach, the MR-3D-DenseNet method has the capacity and favorable scaling characteristics that allowed us to increase the training data by an order of magnitude through rotation of the input samples to predict chemical shift values based in part on the rotational invariance property of chemical shifts. As a result, the totality of our deep learning approach can predict the chemical shifts more accurately, especially for 1H chemical shifts. The accurate chemical shift prediction of 1H is important for the structure characterization of many solid-state chemistry and biological systems as NMR crystallography is one of the most powerful techniques to study the structure and dynamics of hydrogen atoms in solid-state under natural abundance.

Although our chemical shift prediction for 13C, 15N and 17O were significantly improved with respect to KRR, it does not reach the same level of accuracy as compared to high quality ab initio methods. This is almost certainly due to the more limited amount of training examples available for these atom types that prevents us from exploiting the capacity of the MR-3D-DenseNet, and highlights the importance of the size, diversity, and uniqueness of the training datasets. Finally, when interpreting the MR-3D-DenseNet with PCA, we found that we can extract relevant chemical intuition for its performance on chemical shift predictions, similar to deep network interpretation for structural properties in proteins,1115 while also yielding insight into characterizing data sufficiency that can guide future improvements in chemical shift prediction.

Methods

The neural network is implemented using Keras29 with Tensorflow30 as the backend. The density generation step is accelerated using PyTorch31 with GPU implementation. The neural network architecture and hyperparameters are optimized through a 4-fold cross-validation on the training dataset. The training and testing are performed on Tesla P100 GPU. We trained four dedicated models for the chemical shift prediction of 1H, 13C, 15N and 17O-NMR separately. To speed up the data preprocessing process, we calculate the density using 320 nearest neighbor atoms. To accelerate the convergence, we subtract the mean of the chemical shielding values and divide them by 1, 10, 30, 40 for 1H, 13C, 15N and 17O-NMR respectively during the training phase. The mean and scaling factors are applied back during the testing phase. Each layer is followed by a batch normalization (BatchNorm32) layer, and a rectified linear units (ReLU33) layer. There are two dropout layers added after each fully connected layers with rate 0.1. The L2 regularizer with λ = 3 × 10−5 is applied to all the weights in the neural network. The training epochs used are 12, 15, 24 and 24, and the decay rates α are 0.6, 0.5, 0.25 and 0.25 for 1H, 13C, 15N and 17O, respectively. The batch size is fixed to 128. The learning rate starts with 10−3 and decays exponentially. In epoch i, the learning rate is decayed to 10−3 × exp(−). The hyperparameter details are summarized in Table 4. The testing RMSEs are reported by averaging the results from at least three experiments in which the models are initialized with different random seed and the training data are shuffled in random order in each training epoch. The implementation, trained models and detailed instructions for reproducibility are available online: https://thglab.berkeley.edu/software-and-data/

Table 4:

Number of epochs and learning rate decay.

Atom Type Number of Training Epoch Decay Rate α
1H 12 0.6
13C 15 0.5
15N 24 0.25
17O 24 0.25

Supplementary Material

SI

Acknowledgments

JL, KCB, MHG and THG thank the National Institutes of Health for support under Grant No. 5U01GM121667. KCB thanks the California Alliance Postdoctoral Fellowship for early support of his work on this project. BG is a Berkeley Chancellor’s Fellow and is also supported by the National Science Foundation Graduate Research Fellowship under Grant No. DGE 110640. TS gratefully acknowledges funding from the Deutsche Forschungsgemeinschaft (Grant Nos. STA 1526/1-1 and STA 1526/2-1). This research used the computational resources of the National Energy Research Scientific Computing Center, a DOE Office of Science User Facility supported by the Office of Science of the U.S. Department of Energy under Contract No. DE-AC02-05CH11231.

Footnotes

Supporting Information

The study of different density representation and the plot of explained ratio of PCA result.

References

  • (1).Martineau C NMR crystallography: Applications to inorganic materials. Solid State Nuclear Magnetic Resonance 2014, 63–64, 1 – 12. [DOI] [PubMed] [Google Scholar]
  • (2).Bryce DL NMR crystallography: structure and properties of materials from solid-state nuclear magnetic resonance observables. IUCrJ 2017, 4, 350–359. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (3).Shahid SA; Bardiaux B; Franks WT; Krabben L; Habeck M; van Rossum B-J; Linke D Membrane-protein structure determination by solid-state NMR spectroscopy of microcrystals. Nature methods 2012, 9, 1212. [DOI] [PubMed] [Google Scholar]
  • (4).Macholl S; Tietze D; Buntkowsky G NMR crystallography of amides, peptides and protein–ligand complexes. CrystEngComm 2013, 15, 8627–8638. [Google Scholar]
  • (5).Abraham A; Apperley DC; Byard SJ; Ilott AJ; Robbins AJ; Zorin V; Harris RK; Hodgkinson P Characterising the role of water in sildenafil citrate by NMR crystallography. CrystEngComm 2016, 18, 1054–1063. [Google Scholar]
  • (6).Skotnicki M; Apperley DC; Aguilar JA; Milanowski B; Pyda M; Hodgkinson P Characterization of two distinct amorphous forms of valsartan by solid-state NMR. Molecular pharmaceutics 2015, 13, 211–222. [DOI] [PubMed] [Google Scholar]
  • (7).Baias M; Widdifield CM; Dumez J-N; Thompson HP; Cooper TG; Salager E; Bassil S; Stein RS; Lesage A; Day GM et al. Powder crystallography of pharmaceutical materials by combined crystal structure prediction and solid-state 1H NMR spectroscopy. Physical Chemistry Chemical Physics 2013, 15, 8069–8080. [DOI] [PubMed] [Google Scholar]
  • (8).Cui J; Olmsted DL; Mehta AK; Asta M; Hayes SE NMR Crystallography: Evaluation of Hydrogen Positions in Hydromagnesite by 13C {1H} REDOR Solid-State NMR and Density Functional Theory Calculation of Chemical Shielding Tensors. Angewandte Chemie International Edition 2019, 58, 4210–4216. [DOI] [PubMed] [Google Scholar]
  • (9).Pickard CJ; Mauri F All-electron magnetic response with pseudopotentials: NMR chemical shifts. Physical Review B 2001, 63, 245101. [Google Scholar]
  • (10).Hartman JD; Balaji A; Beran GJO Improved Electrostatic Embedding for Fragment-Based Chemical Shift Calculations in Molecular Crystals. Journal of Chemical Theory and Computation 2017, 13, 6043–6051. [DOI] [PubMed] [Google Scholar]
  • (11).Shen Y; Bax A Protein backbone chemical shifts predicted from searching a database for torsion angle and sequence homology. Journal of biomolecular NMR 2007, 38, 289–302. [DOI] [PubMed] [Google Scholar]
  • (12).Shen Y; Bax A SPARTA+: a modest improvement in empirical NMR chemical shift prediction by means of an artificial neural network. Journal of biomolecular NMR 2010, 48, 13–22. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (13).Neal S; Nip AM; Zhang H; Wishart DS Rapid and accurate calculation of protein 1 H, 13 C and 15 N chemical shifts. Journal of biomolecular NMR 2003, 26, 215–240. [DOI] [PubMed] [Google Scholar]
  • (14).Han B; Liu Y; Ginzinger SW; Wishart DS SHIFTX2: significantly improved protein chemical shift prediction. Journal of biomolecular NMR 2011, 50, 43. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (15).Kohlhoff KJ; Robustelli P; Cavalli A; Salvatella X; Vendruscolo M Fast and accurate predictions of protein NMR chemical shifts from interatomic distances. Journal of the American Chemical Society 2009, 131, 13894–13895. [DOI] [PubMed] [Google Scholar]
  • (16).Leclaire J; Poisson G; Ziarelli F; Pepe G; Fotiadu F; Paruzzo FM; Rossini AJ; Dumez J-N; Elena-Herrmann B; Emsley L Structure elucidation of a complex CO 2-based organic framework material by NMR crystallography. Chemical science 2016, 7, 4379–4390. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (17).Cuny J; Xie Y; Pickard CJ; Hassanali AA Ab initio quality NMR parameters in solid-state materials using a high-dimensional neural-network representation. Journal of chemical theory and computation 2016, 12, 765–773. [DOI] [PubMed] [Google Scholar]
  • (18).Paruzzo FM; Hofstetter A; Musil F; De S; Ceriotti M; Emsley L Chemical shifts in molecular solids by machine learning. Nature communications 2018, 9, 4501. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (19).Amidi A; Amidi S; Vlachakis D; Megalooikonomou V; Paragios N; Zacharaki EI EnzyNet: enzyme classification using 3D convolutional neural networks on spatial representation. PeerJ 2018, 6, e4750. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (20).Kuzminykh D; Polykovskiy D; Kadurin A; Zhebrak A; Baskov I; Nikolenko S; Shayakhmetov R; Zhavoronkov A 3D Molecular Representations Based on the Wave Transform for Convolutional Neural Networks. Molecular pharmaceutics 2018, 15, 4378–4385. [DOI] [PubMed] [Google Scholar]
  • (21).Torng W; Altman RB 3D deep convolutional neural networks for amino acid environment similarity analysis. BMC bioinformatics 2017, 18, 302. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (22).Ryczko K; Mills K; Luchak I; Homenick C; Tamblyn I Convolutional neural networks for atomistic systems. Computational Materials Science 2018, 149, 134–142. [Google Scholar]
  • (23).He K; Zhang X; Ren S; Sun J Identity mappings in deep residual networks. European conference on computer vision 2016; pp 630–645. [Google Scholar]
  • (24).Huang G; Liu Z; Van Der Maaten L; Weinberger KQ Densely connected convolutional networks. Proceedings of the IEEE conference on computer vision and pattern recognition 2017; pp 4700–4708. [Google Scholar]
  • (25).Groom CR; Bruno IJ; Lightfoot MP; Ward SC The Cambridge Structural Database. Acta Crystallographica Section B 2016, 72, 171–179. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (26).Ramachandran P; Varoquaux G Mayavi: 3D visualization of scientific data. Computing in Science & Engineering 2011, 13, 40–51. [Google Scholar]
  • (27).Stomberg R; Li S; Lundquist K; Albinsson B Investigation of three stilbene derivatives by X-ray crystallography and NMR spectroscopy. Acta Crystallographica Section C: Crystal Structure Communications 1998, 54, 1929–1934. [Google Scholar]
  • (28).Hartman JD; Kudla RA; Day GM; Mueller LJ; Beran GJO Benchmark fragment-based 1H, 13C, 15N and 17O chemical shift predictions in molecular crystals. Phys. Chem. Chem. Phys 2016, 18, 21686–21709. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • (29).Chollet F Keras. https://keras.io, 2015.
  • (30).Abadi M; Agarwal A; Barham P; Brevdo E; Chen Z; Citro C; Corrado GS; Davis A; Dean J; Devin M et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015; https://www.tensorflow.org/, Software available from tensorflow.org.
  • (31).Paszke A; Gross S; Chintala S; Chanan G; Yang E; DeVito Z; Lin Z; Desmaison A; Antiga L; Lerer A Automatic differentiation in PyTorch. 2017,
  • (32).Ioffe S; Szegedy C Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167 2015,
  • (33).Nair V; Hinton GE Rectified linear units improve restricted boltzmann machines. Proceedings of the 27th international conference on machine learning (ICML-10) 2010; pp 807–814. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

SI

RESOURCES