Skip to main content
Patterns logoLink to Patterns
. 2021 Aug 12;2(9):100329. doi: 10.1016/j.patter.2021.100329

Accurate prediction of B-form/A-form DNA conformation propensity from primary sequence: A machine learning and free energy handshake

Abhijit Gupta 1, Mandar Kulkarni 2, Arnab Mukherjee 1,3,
PMCID: PMC8441556  PMID: 34553171

Summary

DNA carries the genetic code of life, with different conformations associated with different biological functions. Predicting the conformation of DNA from its primary sequence, although desirable, is a challenging problem owing to the polymorphic nature of DNA. We have deployed a host of machine learning algorithms, including the popular state-of-the-art LightGBM (a gradient boosting model), for building prediction models. We used the nested cross-validation strategy to address the issues of “overfitting” and selection bias. This simultaneously provides an unbiased estimate of the generalization performance of a machine learning algorithm and allows us to tune the hyperparameters optimally. Furthermore, we built a secondary model based on SHAP (SHapley Additive exPlanations) that offers crucial insight into model interpretability. Our detailed model-building strategy and robust statistical validation protocols tackle the formidable challenge of working on small datasets, which is often the case in biological and medical data.

Keywords: machine learning, LightGBM, DNA sequence, DNA conformation, nested cross-validation, genome

Highlights

  • A robust machine learning model to predict A- or B-DNA conformation

  • Outcome of machine learning model is explained with free energy values

  • Our approach works well under class imbalance and limited data constraints

The bigger picture

The sequence in the genome of an organism encodes all the information of life. We combine a data-driven approach using machine learning (ML) and the results of free energy calculations to offer a fresh perspective on this long-standing problem of prediction of DNA conformation (A or B) from the sequence. We trained our ML model using sophisticated state-of-the art algorithms such as LightGBM along with a nested cross-validation strategy to overcome the common problems associated with data bias and overfitting when constrained by limited data size. Our study will serve the broader interest of researchers who are not only seeking accurate and reliable predictive models but also want to understand the physical and chemical origins behind the predictions.


We have developed a robust predictive machine learning-based predictive model to accurately predict DNA conformation (A or B) from just the DNA sequence. Unlike a black-box model, our approach offers key chemical and thermodynamic insights into predictions made by our models.

Introduction

The prediction of a DNA conformation from the mere knowledge of its sequence presents an opportunity to presume its role in specific biological processes. The biological processes, such as direct and indirect readout mechanisms in protein-DNA interactions, exploit the conformational flexibility exhibited by DNA. The A-DNA conformation is shorter and more compact than that of B-DNA. During B → A transition, the phosphate groups protrude outward and the minor groove becomes broad and shallow, forming more water bridges in accordance with the theory of economy of hydration proposed by Saenger et al.1

The protein molecules such as transposase, endonuclease, and polymerase interact with B-DNA locally and convert a few dinucleotide steps to A form in a whole DNA.2 A-philic DNA segments exhibit low energy cost for deformation, and thus proteins bind to such hotspots during indirect recognition mechanism to commence the transcription process.2 The A form also participates in the protection of bacterial cells under extreme UV exposure.3 Whelan and coworkers have shown fully reversible B-DNA → A-DNA transition in living bacterial cells on desiccation and rehydration using Fourier transform infrared spectroscopy.4 Extremophiles such as SIRV2 virus (Sulfolobus islandicus rod-shaped virus 2) survives at extreme temperatures of 80°C and acidity of pH 3 by adopting complete DNA in the A form and aids protein to encapsidate DNA.5

Thus, it has become clear of late that A-DNA is merely not a non-functional conformation of DNA; it is an essential adaptation of DNA to survive harsh conditions. It is, therefore, intriguing to predict the sequence-structure relationship in DNA. Moreover, an understanding of sequence specificity of B-form → A-form transition and an a priori detection of the A-philic segment in the genome will unveil the possible hotspots of certain biological processes in specific genes of organisms.

Only a few studies have attempted prediction of DNA conformation from its sequence. Basham and coworkers derived A-DNA propensity energy (APE)6 based on the solvation free energy of trinucleotide steps to determine DNA structural preferences. However, APEs are unavailable for specific trinucleotide steps, thereby making this method inapplicable in general across a genomic DNA sequence. In a different approach, Tolstorukov and coworkers7 formulated free energy models for all ten unique dinucleotide steps (D-12 model) and 32 individual trinucleotide steps (T-32 model) from experimental data of midpoints in B-DNA → A-DNA transition studied earlier by others.8,9 The T-32 model was found to be more accurate than the D-12 model. It inherently considers stereochemical effects present along the B → A transition as it is based on three consecutive DNA base steps. However, the absence of the TAA/TTA free energy values limits the application of this dataset for a DNA structure prediction. The comparison between these free energy methods and present machine learning (ML) model could be interesting. However, the lack of free energy values for all unique 32 trinucleotide steps in these models limits us from comparing these methods. For example, TAA/TTA step and CAG/CTG step (apart from many other steps) are absent in Tolstorukov's trimeric T-32 and Basham's APE model, respectively. This refrained us from using at least 2–4 sequences in each train and test fold and compare accuracies of these methods with our ML model.

Schneider and coworkers developed an automated workflow10, 11, 12 to analyze DNA local conformations. They classified DNA dinucleotide steps based on local backbone conformations. It was observed that DNA structure exhibits mixed A-form/B-form traits in the backbone torsional space, even though the overall structure appears as either A form or B form. Their work demonstrated a high-resolution atlas of local DNA conformations.

In our approach, we have focused on the development of a general and more accurate method based on an ML approach that considers occurrences of all ten unique dinucleotide steps to predict the conformational preference of a given DNA sequence (see Figure 1). In an ML-based approach, the inference is drawn based on observation alone. Therefore, although ML methods are suitable for prediction, the molecular or thermodynamic origin behind the prediction remains unknown. To address this issue, we have also built an explanatory model based on SHAP (SHapley Additive exPlanations) values13 for interpreting and explaining our model output. This method also incorporates the information obtained from free energy values that we obtained earlier to explain the molecular and thermodynamic basis of the prediction made by our ML model.

Figure 1.

Figure 1

Schematic illustration of feature extraction

Results

We describe here the results of the nested cross-validation (CV) performance of the LightGBM algorithm across different metrics used for model assessment (see Figure 2). We observed that the LightGBM algorithm gave the best overall classification results across all five test sets in the nested CV. Figure 3 shows receiver operator characteristic (ROC) curves and precision-recall (PR) curves plotted across all five different test sets (folds). Table 1 shows performance metrics across test sets. We obtained a mean ROC area under the curve (AUC) score of 0.97±0.03, a mean Matthews correlation coefficient (MCC) score of 0.83, a mean accuracy score of 92.7%, a mean F1 score of 0.881, a mean AUC PR of 0.956, and a mean average precision (average PR) of 0.957 on the test sets. The overall performance of our classifier summarized across different thresholds is given by the ROC AUC. Similar to the ROC curve, the PR curve can be used to test all the possible positive predictive values and sensitivities obtained through a binary classification. They are especially valuable for assessing how well an ML model performs on the positive class (A-DNA samples). A high area under the PR curve represents both high recall and high precision. Table 2 displays the per-class performance across all test sets. We observe both high precision and high recall values for each class label. Table 3 shows performance comparison between different ML algorithms. The weighted average returns the average score considering the proportion for each label in the dataset, whereas the macro average returns the average without considering the proportion for each label in the dataset. Furthermore, to ensure reproducibility, we also provide the values of tuned hyperparameters for each model and for all datasets in section C and Data S2 of supplemental information. Detailed results of other approaches are presented in section B of supplemental information (see random forest [Figure S4 and Table S2], support vector machine [SVM] classifier [Figure S5 and Table S3], logistic regression [Figure S6 and Table S4], and naive Bayes classifier [Figure S7 and Table S5]). Section G of supplemental information compares the results on repeated (k = 2) stratified nested CV (Figure S9), and Table S6 shows a comparison between using different methodologies for adjusting class imbalance.

Figure 2.

Figure 2

Schematic display of nested 5-fold stratified cross-validation

A set of n observations is randomly split into five non-overlapping groups in the outer loop. Each group contains approximately the same percentage of samples of each target class as the complete set (stratification). In the inner loop, each training fold is divided again for another round of cross-validation (k = 3) to determine optimal hyperparameters for the classifier.

Figure 3.

Figure 3

Nested stratified 5-fold cross-validation performance of the LightGBM model

(A) Receiver operator characteristic (ROC) AUC curves.

(B) Precision-recall (PR) curves. AUC PR, area under curve of PR curves.

Table 1.

Classification performance of LightGBM algorithm with tuned hyperparameters (see section C in supplemental information) across different test folds

Average PR AUC PR ROC AUC Accuracy F1 MCC
Test fold 1 0.954 0.952 0.969 0.923 0.857 0.822
Test fold 2 0.946 0.944 0.973 0.923 0.880 0.825
Test fold 3 0.987 0.986 0.994 0.947 0.917 0.878
Test fold 4 0.994 0.993 0.997 0.947 0.917 0.878
Test fold 5 0.906 0.904 0.939 0.895 0.833 0.756
Mean 0.957 0.956 0.974 0.927 0.881 0.832

Table 2.

Detailed model evaluation report across different test folds

Precision Recall F1 score Support
Test fold 1 B-DNA 0.90 1.00 0.90 27
A-DNA 1.00 0.75 0.86 12
macro average 0.95 0.88 0.90 39
weighted average 0.93 0.92 0.92 39
Test fold 2 B-DNA 0.89 0.96 0.93 26
A-DNA 0.91 0.77 0.83 13
macro average 0.90 0.87 0.88 39
weighted average 0.90 0.90 0.90 39
Test fold 3 B-DNA 0.96 0.96 0.96 26
A-DNA 0.92 0.92 0.92 12
macro average 0.94 0.94 0.94 38
weighted average 0.95 0.95 0.95 38
Test fold 4 B-DNA 0.96 0.96 0.96 26
A-DNA 0.92 0.92 0.92 12
macro average 0.94 0.94 0.94 38
weighted average 0.95 0.95 0.95 38
Test fold 5 B-DNA 0.96 0.92 0.94 26
A-DNA 0.85 0.92 0.88 12
macro average 0.90 0.92 0.91 38
weighted average 0.92 0.92 0.92 38

Table 3.

Comparison between different ML algorithms

Average PR AUC PR ROC AUC Accuracy F1 MCC
Random forest 0.888 0.885 0.945 0.881 0.829 0.747
SVM classifier 0.912 0.908 0.943 0.886 0.822 0.741
Logistic regression 0.910 0.906 0.945 0.896 0.821 0.758
Naive Bayes classifier 0.919 0.918 0.952 0.891 0.841 0.765
LightGBM 0.957 0.956 0.974 0.927 0.881 0.832

In LightGBM, boosting helps in reducing bias and variance in ensemble-based models, which is particularly useful for controlling overfitting. It builds trees in a stage-wise forward manner, where weak learners (trees) are added to address the shortcomings of existing weak learners. As the end result, the model is able to achieve high accuracy by increasing the importance of “difficult” observations (samples that have a complex non-linear decision boundary). As more trees are added, they rectify the misclassification error committed by existing learners. To control overfitting, we use the optimal value of regularization parameters: L1, L2 regularization, bagging fraction and frequency, number of leaves, and feature fraction (see section C in supplemental information). Another benefit of using gradient boosting is that after the boosted trees are constructed, it is relatively straightforward to retrieve importance scores for each attribute.14 To understand how individual dinucleotide steps affect the propensity of a sequence to assume a given conformation, we have used SHAP,13 a unified approach for explaining the output of any ML model. SHAP connects game theory with local explanations, uniting several previous methods, and representing the only possible consistent and locally accurate additive feature attribution method based on expectations.13 This explanation model uses simplified inputs, which toggle features on and off, rather than raw inputs to the original model. Figure 4 shows the schematic models of SHAP, where data are processed using the original model and using the SHAP criteria as mentioned above. g(z)is a linear function of binary variables (ON or OFF), which determines the role of individual inputs of features in the prediction. SHAP builds model explanations by asking the same question for every prediction and feature: “How does prediction i change when feature j is removed from the model?” as mentioned above.

Figure 4.

Figure 4

Schematics of SHAP (SHapley Additive exPlanations) model

To interpret and relate these SHAP values with the thermodynamics, we describe the concept of the absolute free energy values (see section E in supplemental information and Kulkarni and Mukherjee15 for further details). Thermodynamically, the conformation of a particular structure depends on the free energetic stability. Therefore, the propensity of a sequence to adopt a particular conformation should depend on the overall free energy of the sequence in that conformation. Keeping that in mind, we had earlier calculated the free energy cost (Table 3) for the formation of the A form of each of the ten dinucleotide steps, as discussed below.15

To obtain an idea about which features are most important for our model, we have plotted the SHAP values of each dinucleotide step (feature) for every sample. Figure 5 shows the SHAP summary plot, which sorts features by the sum of SHAP value magnitudes over all samples and uses these SHAP values to show the distribution of the impacts of each feature on the model output. The summary plot combines feature importance with feature effects. Each point on the summary plot is a Shapley value for a feature and an instance. The position on the y axis is determined by the feature and on the x axis by the Shapley value. The color represents the value of the feature from low to high (red means high impact, blue means low impact). Overlapping points are jittered in the y-axis direction, so we get a sense of the distribution of the Shapley values per feature. The features are ordered according to their importance. Traditionally AA/TT and GG/CC are considered to be the most B-philic and A-philic dinucleotide steps. We see that (AA/TT), a B-promoting dinucleotide step, and GG/CC, an A-promoting dinucleotide step, have the highest impact on our model prediction. The AA/TT step has the highest negative SHAP value, which corresponds to its highest contribution in predicting B-promoting DNA sequences. Similarly, the GG/CC and GC/GC have the highest positive SHAP value, which corresponds to their highest contribution in predicting A-promoting DNA sequences. GG/CC has lower junction free energy ((ΔGJ) compared with AA/TT and thus, lower mean cooperative length for B-to-A transition. High positive SHAP values for GG/CC support this free energy-based observation.

Figure 5.

Figure 5

SHAP summary

The plot sorts features by the sum of SHAP value magnitudes over all samples and uses SHAP values to show the distribution of the impacts of each feature on the model output. The color represents the feature value (red = high, blue = low). The horizontal scale represents the SHAP values, with the left side indicating B-DNA region (negative values) and the right side indicating A-DNA region (positive values). The absolute free energy value of each dinucleotide step is mentioned adjacent to its label.

CG/CG and AT/AT dinucleotide steps have highest negative SHAP values after AA/TT step, thus suggesting these steps as B-philic steps and supported by absolute free energy values. This tells us that they have a high negative impact (B-DNA prone) on the model prediction. Similarly, GC/GC and AC/GT are observed to be A-philic and B/A intermediate step, respectively, based on both positive SHAP values and low free energy values.

Absolute energy (ΔGa) values indicate AG/CT and GA/TC as A-philic steps, whereas SHAP value distribution indicate these steps as B-philic steps. Dinucleotide steps AG and GA steps prefer B-form conformation over A form as observed in previous studies (Svozil et al.;10 Marathe et al.16; our review17). Earlier studies observed that structurally TA/TA and CA/TG steps exhibit high roll angle similar to A-form steps but still maintain flexible B form (Hassan and Calladine (1998) J. Mol. Biol. 282, 331–343). The flexible B form allows DNA bending or kinking during protein interactions. ΔGa values predicted TA and CA steps as neutral (neither A-philic or B-philic) and B-philic step, respectively.

For CA and TA steps, overall SHAP values are distributed close to the abscissa and neutral feature values (violet color) with negative SHAP values possibly indicate the intermediate nature of these steps. Thus, we can see that SHAP values have correctly predicted conformation preference in comparison with free energy values, demonstrating its classification efficiency over the free energy method.

It is interesting to note that there is good agreement between these inferences drawn from our ML model with the absolute free energy values, except for TA and CA steps (Table 4). Figure 6 shows the standard bar plot obtained by taking the mean absolute value of the SHAP values for each feature. This plot shows how each dinucleotide step (feature) contributes to the prediction of the propensity of the A/B-promoting DNA sequence.

Table 4.

List of absolute energy values (ΔGa) and mean of absolute SHAP values for all ten possible dinucleotide steps

Dinucleotide steps ΔGa(kcal/mol)
AA/TT 2.34
GG/CC 0.86
AC/GT 1.91
CA/TG 2.40
AT/AT 2.29
TA/TA 1.59
AG/CT 0.67
GA/TC 0.84
CG/CG 3.06
GC/GC 1.33

Note that ΔGJ values were calculated only for homonucleotide steps and not heteronucleotide steps. ΔGJ is 1.59 kcal/mol for AA/TT and 0.52 kcal/mol for GG/CC.

Figure 6.

Figure 6

Mean of absolute SHAP values show the average impact of each dinucleotide step in predicting whether a given sequence will attain A or B conformation

Discussion

In our approach, we have trained different ML algorithms using a set of known A-DNA/B-DNA sequences. The best ML approach (LightGBM) provides prediction with correctness of ∼93% and an MCC score of 0.832. As it turns out, our model is able to capture the complex relationship between the feature vectors (dinucleotide steps) that attribute to the final conformation assumed by a DNA sequence. Understanding why a model makes a specific prediction can be as important as the prediction's accuracy in many applications. This is crucial when we want to understand how each fundamental dinucleotide step contributes toward the conformation attained by a sequence. The highest accuracy for large modern datasets is often achieved by complex models that are difficult to interpret, such as an ensemble of several models or deep learning models. LightGBM18 is an implementation of a gradient boosting decision tree technique that offers a balanced tradeoff between accuracy and interpretability. For gaining further insight into the interpretability of our model, SHAP analysis was employed with which we could come up with a consistent and locally accurate additive feature attribution method based on expectations. Our study thus indicates that the conformational preference of a DNA lies in the fundamental free energetic driving force at a local dinucleotide level. Most of the DNA sequences used here, however, are short. Therefore, the cooperative effect may play a role in the case of longer DNA sequences, and an effort is under way to understand this. Our training set contains some hexamer or octamer A-DNA sequences. Such short oligonucleotides are affected by crystal packing forces.19,20 At the same time, we also find nuclear magnetic resonance (NMR) structures with A form. We would also like to point out that crystal packing may have a role in producing A form for certain sequences. However, this influence is limited to only fa specific set of sequences. We believe that this has to do with the inherent propensities of these sequences to adopt the A form. For this reason, we do not find all short sequences adopting the A form. There are some short AT-rich sequences (PDB: 4U9M, 2G1Z) which are crystallized as B-DNA as opposed to crystal packing-derived A-DNA conformation. There are many 8- or 10-mer sequences that are in B form, just as not all A-DNA sequences are 8- or 10-mers. Therefore, it is not obvious that the length of the DNA sequence would dictate a particular conformation. The high predictive performance of our model indicates that there must be some inherent tendencies of these sequences to adopt the A form, and the objective of the present work is to capture that.

AA steps are highly B-philic due to the steric hindrance of their antisense counterpart TT step. A severe steric hindrance between protruding methyl groups of thymine base exists if it undergoes B → A transition and thus enhances the free energetic cost of the process. It is surprising to see that ML models can predict the AA step as the most B-philic step without the knowledge of the structure and interactions between the stacking base steps.

The GG step is well known to adopt or induce the A form in DNA sequences. Again, it is encouraging to note that the ML model can predict GG and GC as the most A-philic steps without any structural information. The eukaryotic and prokaryotic genomes contain DNA segments that can be easily converted to A form (A-DNA promoter sequences [APS]). These A-form segments can then be specifically recognized by DNA binding proteins during the indirect readout mechanism. Such APSs allow binding of certain transcription factor (TF) binding proteins21 and could play a role in protein-DNA binding mechanisms. Recently, Li et al. incorporated DNA sequence and shape as features along with information from X-ray and simulated structures to determine TF binding regions in DNA sequences.22

Whitley and coworkers21 used Basham's trinucleotide solvation free energy method of A/B DNA structure prediction6 to find A-DNA promoters in the Xenopus tropicalis genome. Owing to the limited applicability of the aforementioned method, we believe that our proposed ML model can be implemented on other genomes to find unknown A-DNA promoter DNA steps a priori. Further study is under way to explore eukaryotic genome analysis as well as the genome of organisms that survive under stringent conditions using the A form of DNA.

Limitations of the study

We would now like to discuss the limitations of the present study. The DNA structures considered here are assigned as B-DNA or A-DNA because these structures do not contain mixed A-form/B-form dinucleotide steps. We assume that even with mixed A/B traits at the local level, based on the definition of recent studies the whole DNA structure appears as B or A due to prominent conformational preference of each dinucleotide step of the DNA. The cooperative effects of these dinucleotide steps contribute to the overall conformational preference in DNA oligonucleotides.

Finally, the classification of a sequence to A or B is based on the Nucleic Acid Database (NDB) data. Therefore, our goal was to apply the method to a given sequence and predict the A/B classification in conformity with the NDB (global) structural classification. As mentioned earlier, we tried to include in our curated dataset the sequences whose structures were obtained under similar experimental conditions so as to minimize the effect of varying experimental conditions.

At the moment, we are restricted by the paucity of a sufficient number of labeled DNA sequences. Out of 192 curated DNA sequences in the NDB dataset, 61 are A-DNA sequences and 131 are B-DNA sequences (supplemental information, section F [Data S1]). The lack of data is one of the significant challenges in any ML model. Furthermore, the severe class imbalance between A- and B-DNA is another limitation, although we have adopted several measures to overcome these limitations in the present study. Also, the present study focuses only on the canonical A-DNA or B-DNA conformation, with the objective of developing a method to understand the tendency of short DNA segments in long oligonucleotides to adopt these conformations. Thus, we have not considered non-canonical DNA structures. We also acknowledge that there are subclasses of this broad classification23—different A-form conformers, conformers bridging A to B form and vice versa, a separate Z form, subdivision of B conformations into BI and BII form24—which we could not categorize owing to the paucity of data in the NDB database.

Experimental procedures

Resource availability

Lead contact

The lead contact for this work is Arnab Mukherjee, arnab.mukherjee@iiserpune.ac.in.

Materials availability

The study did not generate new unique materials or reagents.

Methods

Data curation

The first step in an ML approach is data curation. Since we use a supervised learning approach, we collected A- and B-DNA structures from the Nucleic Acid Database (NDB repository).25,26 The corresponding sequences were retrieved from the RCSB PDB27 database by a parser written by us. We filtered out all redundant sequences along with all those sequences which had anything in addition to A, C, G, and T. Furthermore, we considered only the unbound double-stranded DNA structures. We removed all DNA sequences less than 5 base pairs long from our analysis as they are too short to be deciding a particular conformation. While selecting sequences for our study, we looked at the different experimental conditions under which different DNA structures were crystallized/labeled, namely "Crystallization Method," "Temperature (K)," "pH," "Crystal Growth Procedure," "R-free Values," and "Percent Solvent Content." In particular, for X-ray structures we selected those sequences that corresponded to structures with high Rfree values and resolution. For NMR-based structures, we considered the “Sample Temperature,” “Sample pH Values,” “Solvent System,” “Ionic Strength,” and other relevant parameters. We have presented the distribution of different experimental conditions under which different structures were obtained in section A of supplemental information. To minimize the effect of the influence of varying experimental conditions, we tried to select the sequences obtained under similar conditions. We also checked for outlier samples using a skewness adjusted interquartile range method28(see section A in supplemental information), which takes into consideration the skewness in the distribution for robust outlier detection. This helped us in obtaining those sequences for which the experimental conditions were similar, irrespective of the class label. Section A of supplemental information shows kernel density estimation plots of each experimental condition for both A- and B-DNA samples that we included in our dataset.

We also performed sequence similarity analysis across all sequences in a given class. We used the alignment-free sequence comparison approach that is based on the frequencies of k-mers (subsequences or words of length k).29 This considers the “Euclidean distance” between k-mers frequency profiles of two sequences as a measure of the dissimilarity between them. The pairwise distance matrix hence obtained is normalized between 0 and 1. Unlike alignment-based methods, the alignment-free method does not assume the contiguity of homologous regions. They are also less dependent on substitution/evolutionary models and are comparatively computationally inexpensive. The choice of k depends on the nature and the length of the sequences. Smaller k-mers should be used when sequences are obviously different (e.g., they are not related), whereas longer k-mers can be used for very similar sequences.30,31 For nucleotide sequences, k is usually set to 4–10 for smaller sequences, and k = 8 or 10 is typically used for comparing longer sequences.30,32 We considered k={4,5,6} for comparing sequence similarity. The mean sequence similarity is 31.9% for A-DNA samples and 28.7% for B-DNA samples in our curated dataset. In our dataset, the smallest sequences are of length 6 and hence is the upper bound on the choice of k.

Our curated dataset contained 192 samples, of which 61 are A-DNA sequences and 131 are B-DNA sequences. The list of curated DNA sequences along with resolution (Å), R value, Rfree (for crystallographic structures), and other relevant experimental conditions are mentioned in section F of supplemental information (Data S1).

Feature extraction

Feature extraction or “feature design” is an essential step in any ML approach. The characteristics of any object are called features. In a DNA sequence, relevant features could be the length of the DNA, the number and types of dinucleotide steps, or the number and types of tetranucleotide steps. In this study, we have considered the count of all ten unique dinucleotide steps in the given DNA sequence as our feature vectors (Figure 1). There are two main rationales behind our choice: (1) the dinucleotide step represents the smallest possible building block for DNA conformation;33,34 (2) we have used the absolute free energy values for each dinucleotide step15 in model interpretation, explaining how a particular conformation can be attributed to structural and chemical aspects associated with each dinucleotide step.

We wish to mention that the lack of data precludes us from building a model that considers relative positions of the different dinucleotide steps in a sequence. Such a model, although desirable, would require a large number of training samples for training. Our approach, on the other hand, offers a viable compromise.

Pre-processing and adjusting the class imbalance

Data pre-processing involves the transformations that are applied to the data before feeding them to our ML models. For this classification problem, we have encoded the A-DNA samples as the positive class with the label “1” and the B-DNA samples as the negative class with the label “0.” Some ML models such as support vector machines with radial basis function as the kernel35 and models that use L1 and L2 regularization assume that all features are centered around 0 and have variance in the same order.36 We, therefore, standardized the features by removing the mean and scaling to unit variance. The standard score of a sample x is calculated as

z=(xu)/s,

where u is the mean of the training samples and s is the standard deviation of the training samples. Centering and scaling happen independently on each feature by computing the relevant statistics on the samples in the training set.

We also observed a significant class imbalance (32% A-DNA versus 68% B-DNA curated, non-redundant sequences) that became apparent during the preliminary analysis. To address the class imbalance issue, in which training data belonging to one class outnumber the examples in the other, we tried two different strategies during the training stage. First, we adjusted the class weight. Due to the imbalanced number of positive (A-DNA) and negative (B-DNA) samples, the class weight option imposes a heavier penalty for errors in the minority class. Class weights are inversely proportional to class frequencies in the training data. The second strategy employed the SMOTE + TOMEK method,37 which is a combination of oversampling and undersampling. SMOTE is an oversampling method that synthesizes new plausible examples in the majority class. Tomek-Links refers to a method for identifying pairs of nearest neighbors in a dataset that have different classes. Removing one or both examples in these pairs has the effect of making the decision boundary in the training dataset less noisy or ambiguous. Despite the differences between the two approaches, they deliver similar improvements.

Model building

In this stage, we considered different ML algorithms for our problem. Classification of a sequence into A-/B-DNA is a binary classification problem. We tried LightGBM18 (based on gradient boosting decision tree), SVM classifier with “RBF” and linear kernel,35 random forest classifier, naive Bayes classifier, and logistic regression.35 Each model outputs the probability p(Ck|x) of a class, Ck={0,1}, given a sequence x (0 represents B-DNA and 1 represents A-DNA). We then used an optimal threshold for converting this probability into class labels. When selecting a classification algorithm for a particular problem, one has to simultaneously select the best algorithm for that dataset and the best set of hyperparameters for the chosen model. These hyperparameters are intrinsic to each algorithm and define the model architecture. The accuracy of a model on unseen data is critically dependent on the choice of suitable values for the hyperparameters. The search for optimal values for the hyperparameters is a process known as model selection. ML models such as LightGBM have several hyperparameters. These are the threshold parameter scale_pos_weight for adjusting the threshold for an imbalanced dataset, regularization parameters L1 and L2, number of leaves (for controlling the complexity of the model), number of iterations, learning rate, bagging fraction, and bagging frequency. Even fairly simple generalized additive models such as logistic regression have hyperparameters such as regularization, class_weight, or threshold. Most of these models would perform poorly on the unseen data if one were to use the default set of hyperparameters. Hyperparameter optimization can be accomplished in several ways: one can exhaustively consider all parameter combinations using grid search, use randomized search strategy to sample a given number of candidates from a parameter space with a specified distribution, or optimize the criterion of expected improvement (EI) using a Gaussian process/tree-structured Parzen estimator approach (TPE). We chose to use the optimization of EI criterion because it is intuitive and has been shown to work well in a wide variety of settings.38 To tune the hyperparameters of our models, we used the TPE approach implemented in the Optuna framework.39

We have used Intel Distribution for Python and Python API for Intel Data Analytics Acceleration Library (Intel DAAL)—named PyDAAL40—to boost ML and data analytics performance. Using the advantage of optimized Scikit-learn (Scikit-learn with Intel DAAL) that comes with it, we were able to achieve faster training time and accurate results for the prediction problem.

Training and evaluation

In an ideal situation, we would have a large dataset to be able to train and validate our models (training samples) and have separate data for assessing the quality of our model (test samples). However, such data-rich situations are, more often than not, rare in the life sciences. In many practical applications, we seldom have the luxury of having a sufficiently large test set, which would provide an unbiased estimate of the generalization performance of our models. If we reserve too many data for training, this results in unreliable and biased estimates of the generalization performance; setting aside too many data for testing results in too few data for training, which in turn hurts model performance. For such situations where the dataset is small and reserving data for independent test sets is not feasible, the nested CV41,42 procedure offers a viable alternative. Nested CV can be used for choosing an appropriate classifier (model) and optimizing its hyperparameters to obtain a reliable and unbiased estimate of generalization performance.41,43 Model selection without nested CV uses the same data to tune model parameters and evaluate model performance. Information may thus “leak” into the model and overfit the data, leading to a phenomenon called “overfitting in model selection.”43 We compared the performance of the machine learning algorithms, referred to as ML algorithms hereafter, by performing nested 5-fold stratified nested CV. This process consists of two nested CV loops which are often referred to as inner (internal) and outer (external) CV loops. We perform the model selection in the inner loop, and in the outer loop we estimate the generalization performance (see Figure 2 for a schematic overview of nested CV). In the outer loop, our dataset is randomly split into five non-overlapping groups. Stratification is used to preserve the percentage of samples for each class. In each group, these two disjoint subsets are referred to as the training and the test set. In each group, the test set is exclusively used for model assessment. In the inner loop, the training set is used for model building and model selection. In each iteration of the inner loop, the incoming training set is repeatedly split into inner training and validation datasets by a stratified 3-fold CV approach. The inner training folds are used to derive different models by varying the hyperparameters (tuning parameters) of the model family at hand, whereas the validation sets are used to estimate the models' performance. The hyperparameters corresponding to the model with the lowest CV error across the inner folds are chosen for training the outer loop model. Along with tuning of hyperparameters, we also choose the optimal threshold via a threshold-moving technique on the validation data. This involves choosing the threshold that corresponds to the maximum score on a chosen evaluation metric. For this purpose, we have chosen the F1 score metric. This tries to find the balance between precision and recall, which is extremely useful in scenarios when we are working with imbalanced datasets. Finally, in each iteration of the outer loop, we initialize the model with the tuned hyperparameters and threshold and use the test set to obtain an unbiased estimate of the selected model. We present below the pseudocode for the nested CV algorithm.

  • Fori=1 to K1 splits do://(outer loop)
    • Split D into Ditrain,Ditest for the ith split
    • Forj=1 to K2 splits do://(inner loop)
      • Split Ditrain into Djinnertrain,Djvalidation for the jth split
      • sample parameter space (Psets) using random search and TPE to get Pj
        • Initialize and train model M on Djinnertrain with hyperparameter set Pj
        • Tune hyperparameters to get Pj and compute validation error Ejvalidation for M with Djvalidation
    • Select optimal hyperparameter set P from Psets, where Ejvalidation is the least
    • Train M with Ditrain, using Pas hyperparameters
    • Compute test error metrics Eitest for M with Ditest

For assessment of the performance of our classification model, we have chosen accuracy, F1 score, MCC, ROC curve, and PR curves as our primary evaluation metrics. When there is a class imbalance, the accuracy alone cannot give an accurate assessment of the performance of a classification model. A classifier may proclaim all data points as belonging to the majority class and obtain a high-accuracy score while performing poorly on the prediction of minority class samples. Therefore, using accuracy as the sole criterion for model evaluation can lead to overoptimistic inflated results, especially on imbalanced datasets. ROC represents a probability curve, and the AUC of the ROC curve represents the measure of separability between the two classes. The higher the AUC-ROC score, the better the model is at distinguishing between A- and B-DNA samples. Precision is defined as the ratio of true positives and the sum of true positives and false positives. False positives are outcomes the model incorrectly labels as positive that are actually negative. In our example, false positives are B-DNA that the model classifies as A-DNA. In contrast, recall expresses the number of true positives divided by the sum of true positives and false negatives. In most problems pertaining to classification, one could give a higher priority to maximizing precision, or recall, depending on the problem one is trying to solve. However, in general, there exists a more straightforward metric that takes into consideration both precision and recall. This metric is known as the F1 score, the harmonic mean of precision and recall. Notably, the MCC coefficient considers true and false positives and negatives and is generally regarded as a balanced measure that can be used when there is a class imbalance.44 It produces a more informative and truthful score in evaluating binary classifications than accuracy and F1 score. The formulas of these metrics are mentioned in section D of supplemental information. Section F of supplemental information contains the list of all samples used for training and testing for each iteration of the outer loop.

Acknowledgments

We thank Dr. Leelavati Narlikar, National Chemical Laboratory, India, for extensive discussions. We acknowledge funding from the Department of Science and Technology, Science and Engineering Board, Government of India (grant EMR/2016/001069). The study is also partially supported by Department of Biotechnology, India (BT/PR34215/AI/133/22/2019).

Author contributions

A.M. conceptualized the problem. A.G. designed and developed the methodology, software, formal analysis, data curation, and investigation. M.K. provided key inputs about the A-/B-DNA transition and its association with the free energy. A.M. supervised the project. The manuscript was written through the contributions of all authors.

Declaration of interests

The authors declare no competing interests.

Published: August 12, 2021

Footnotes

Supplemental information can be found online at https://doi.org/10.1016/j.patter.2021.100329.

Supplemental information

Document S1. Supplemental experimental procedures, Figures S1–S9, Tables S1–S6, and Data S1
mmc1.pdf (2.2MB, pdf)
Data S2. Tuned hyperparameters of different Machine Learning algorithms
mmc2.xlsx (25.6KB, xlsx)
Document S2. Article plus supplemental information
mmc3.pdf (3.9MB, pdf)

Data and code availability

Data and code can be accessed at the following link: https://github.com/abhijitmjj/DNA-structure-prediction (code author and repo maintainer: Abhijit Gupta [github username: abhijitmjj]).

DNA structure prediction from its sequence code is available in the GitHub repository (https://github.com/abhijitmjj/DNA-structure-prediction). We intend to build a webserver for our program soon, where the user can provide raw sequences as the input and obtain the probabilities for them to attain A/B form conformation.

References

  • 1.Saenger W., Hunter W.N., Kennard O. DNA conformation is determined by economics in the hydration of phosphate groups. Nature. 1986;324:385–388. doi: 10.1038/324385a0. [DOI] [PubMed] [Google Scholar]
  • 2.Lu X.-J., Shakked Z., Olson W.K. A-form conformational motifs in ligand-bound DNA structures. J. Mol. Biol. 2000;300:819–840. doi: 10.1006/jmbi.2000.3690. [DOI] [PubMed] [Google Scholar]
  • 3.Mohr S.C., Sokolov N.V., He C.M., Setlow P. Binding of small acid-soluble spore proteins from Bacillus subtilis changes the conformation of DNA from B to A. Proc. Natl. Acad. Sci. U S A. 1991;88:77–81. doi: 10.1073/pnas.88.1.77. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Whelan D.R., Hiscox T.J., Rood J.I., Bambery K.R., McNaughton D., Wood B.R. Detection of an en masse and reversible B- to A-DNA conformational transition in prokaryotes in response to desiccation. J. R. Soc. Interface. 2014;11:20140454. doi: 10.1098/rsif.2014.0454. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.DiMaio F., Yu X., Rensen E., Krupovic M., Prangishvili D., Egelman E.H. A virus that infects a hyperthermophile encapsidates A-form DNA. Science. 2015;348:914–917. doi: 10.1126/science.aaa4181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Basham B., Schroth G.P., Ho P.S. An A-DNA triplet code: thermodynamic rules for predicting A- and B-DNA. Proc. Natl. Acad. Sci. U S A. 1995;92:6464–6468. doi: 10.1073/pnas.92.14.6464. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Tolstorukov M.Y., Ivanov V.I., Malenkov G.G., Jernigan R.L., Zhurkin V.B. Sequence-dependent B↔A transition in DNA evaluated with dimeric and trimeric scales. Biophys. J. 2001;81:3409–3421. doi: 10.1016/S0006-3495(01)75973-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Minchenkova L.E., Schyolkina A.K., Chernov B.K., Ivanov V.I. CC/GG contacts facilitate the B to A transition of DMA in solution. J. Biomol. Struct. Dyn. 1986;4:463–476. doi: 10.1080/07391102.1986.10506362. [DOI] [PubMed] [Google Scholar]
  • 9.Ivanov V.I., Minchenkova L.E., Minyat E.E., Schyolkina A.K. Cooperative transitions in DNA with no separation of strands. Cold Spring Harb. Symp. Quant. Biol. 1983;47:243–250. doi: 10.1101/sqb.1983.047.01.029. [DOI] [PubMed] [Google Scholar]
  • 10.Svozil D., Kalina J., Omelka M., Schneider B. DNA conformations and their sequence preferences. Nucleic Acids Res. 2008;36:3690–3706. doi: 10.1093/nar/gkn260. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Čech P., Kukal J., Černý J., Schneider B., Svozil D. Automatic workflow for the classification of local DNA conformations. BMC Bioinformatics. 2013;14:205. doi: 10.1186/1471-2105-14-205. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Schneider B., Božíková P., Čech P., Svozil D., Černý J. A DNA structural alphabet distinguishes structural features of DNA bound to regulatory proteins and in the nucleosome core particle. Genes. 2017;8:278. doi: 10.3390/genes8100278. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Lundberg S.M., Lee S.-I. In: Advances in Neural Information Processing Systems 30. Guyon I., Luxburg U.V., Bengio S., Wallach H., Fergus R., Vishwanathan S., Garnett R., editors. Curran Associates, Inc.); 2017. A unified approach to interpreting model predictions; pp. 4765–4774. [Google Scholar]
  • 14.Mayr A., Binder H., Gefeller O., Schmid M. The evolution of boosting algorithms-from machine learning to statistical modelling. arXiv. 2014 doi: 10.3414/ME13-01-0122. 1403.1452. [DOI] [PubMed] [Google Scholar]
  • 15.Kulkarni M., Mukherjee A. Computational approach to explore the B/A junction free energy in DNA. ChemPhysChem. 2016;17:147–154. doi: 10.1002/cphc.201500690. [DOI] [PubMed] [Google Scholar]
  • 16.Marathe A., Karandur D., Bansal M. Small local variations in B-form DNA lead to a large variety of global geometries which can accommodate most DNA-binding protein motifs. BMC structural biology. 2009;9(1):1–26. doi: 10.1186/1472-6807-9-24. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Kulkarni M., Mukherjee A. Understanding B-DNA to A-DNA transition in the right-handed DNA helix: Perspective from a local to global transition. Progress in biophysics and molecular biology. 2017;128:63–73. doi: 10.1016/j.pbiomolbio.2017.05.009. [DOI] [PubMed] [Google Scholar]
  • 18.Ke G., Meng Q., Finley T., Wang T., Chen W., Ma W., Ye Q., Liu T.-Y. In: Advances in Neural Information Processing Systems 30. Guyon I., Luxburg U.V., Bengio S., Wallach H., Fergus R., Vishwanathan S., Garnett R., editors. Curran Associates, Inc.); 2017. LightGBM: a highly efficient gradient boosting decision tree; pp. 3146–3154. [Google Scholar]
  • 19.Ramakrishnan B., Sundaralingam M. Evidence for crystal environment dominating base sequence effects on DNA conformation: crystal structures of the orthorhombic and hexagonal polymorphs of the A-DNA decamer d (GCGGGCCCGC) and comparison with their isomorphous crystal structures. Biochemistry. 1993;32:11458–11468. doi: 10.1021/bi00093a025. [DOI] [PubMed] [Google Scholar]
  • 20.Shakked Z., Guerstein-Guzikevich G., Eisenstein M., Frolow F., Rabinovich D. The conformation of the DNA double helix in the crystal is dependent on its environment. Nature. 1989;342:456–460. doi: 10.1038/342456a0. [DOI] [PubMed] [Google Scholar]
  • 21.Whitley D.C., Runfola V., Cary P., Nazlamova L., Guille M., Scarlett G. APTE: identification of indirect read-out A-DNA promoter elements in genomes. BMC Bioinformatics. 2014;15:288. doi: 10.1186/1471-2105-15-288. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Li J., Sagendorf J.M., Chiu T.-P., Pasi M., Perez A., Rohs R. Expanding the repertoire of DNA shape features for genome-scale studies of transcription factor binding. Nucleic Acids Res. 2017;45:12877–12887. doi: 10.1093/nar/gkx1145. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Schneider B., Božíková P., Nečasová I., Čech P., Svozil D., Černý J. A DNA structural alphabet provides new insight into DNA flexibility. Acta Crystallogr. D Biol. Crystallogr. 2018;74:52–64. doi: 10.1107/S2059798318000050. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Hartmann B., Piazzola D., Lavery R. B I-B II transitions in B-DNA. Nucleic Acids Res. 1993;21:561–568. doi: 10.1093/nar/21.3.561. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Berman H.M., Olson W.K., Beveridge D.L., Westbrook J., Gelbin A., Demeny T., Hsieh S.H., Srinivasan A.R., Schneider B. The nucleic acid database. A comprehensive relational database of three-dimensional structures of nucleic acids. Biophys. J. 1992;63:751–759. doi: 10.1016/S0006-3495(92)81649-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Coimbatore Narayanan B., Westbrook J., Ghosh S., Petrov A.I., Sweeney B., Zirbel C.L., Leontis N.B., Berman H.M. The Nucleic Acid Database: new features and capabilities. Nucleic Acids Res. 2014;42:D114–D122. doi: 10.1093/nar/gkt980. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Deshpande N., Addess K.J., Bluhm W.F., Merino-Ott J.C., Townsend-Merino W., Zhang Q., Knezevich C., Xie L., Chen L., Feng Z. The RCSB Protein Data Bank: a redesigned query system and relational database based on the mmCIF schema. Nucleic Acids Res. 2005;33:D233–D237. doi: 10.1093/nar/gki057. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Hubert M., Vandervieren E. An adjusted boxplot for skewed distributions. Comput. Stat. Data Anal. 2008;52:5186–5201. [Google Scholar]
  • 29.Zielezinski A., Vinga S., Almeida J., Karlowski W.M. Alignment-free sequence comparison: benefits, applications, and tools. Genome Biol. 2017;18:186. doi: 10.1186/s13059-017-1319-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Sims G.E., Jun S.-R., Wu G.A., Kim S.-H. Alignment-free genome comparison with feature frequency profiles (FFP) and optimal resolutions. Proc. Natl. Acad. Sci. U S A. 2009;106:2677–2682. doi: 10.1073/pnas.0813249106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Wu T.-J., Huang Y.-H., Li L.-A. Optimal word sizes for dissimilarity measures and estimation of the degree of dissimilarity between DNA sequences. Bioinformatics. 2005;21:4125–4132. doi: 10.1093/bioinformatics/bti658. [DOI] [PubMed] [Google Scholar]
  • 32.Chan C.X., Bernard G., Poirion O., Hogan J.M., Ragan M.A. Inferring phylogenies of evolving sequences without multiple sequence alignment. Sci. Rep. 2014;4:6504. doi: 10.1038/srep06504. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Dickerson R.E. Definitions and nomenclature of nucleic acid structure components. Nucleic Acids Res. 1989;17:1797–1803. doi: 10.1093/nar/17.5.1797. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.El Hassan M., Calladine C. Conformational characteristics of DNA: empirical classifications and a hypothesis for the conformational behaviour of dinucleotide steps. Philos. Trans. R. Soc. Lond. A. 1997;355:43–100. [Google Scholar]
  • 35.Bishop C.M. Springer; 2006. Pattern Recognition and Machine Learning. [Google Scholar]
  • 36.Friedman J., Hastie T., Tibshirani R. Springer; 2001. The Elements of Statistical Learning. [Google Scholar]
  • 37.Batista G.E.A.P.A., Prati R.C., Monard M.C. A study of the behavior of several methods for balancing machine learning training data. SIGKDD Explor. Newsl. 2004;6:20–29. [Google Scholar]
  • 38.Bergstra J.S., Bardenet R., Bengio Y., Kégl B. Proceedings of the 24th International Conference on Neural Information Processing Systems. 2011. Algorithms for hyper-parameter optimization; pp. 2546–2554. [Google Scholar]
  • 39.Akiba T., Sano S., Yanase T., Ohta T., Koyama M. Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining KDD ’19. Association for Computing Machinery); 2019. Optuna: a next-generation hyperparameter optimization framework; pp. 2623–2631. [Google Scholar]
  • 40.Malakhov A., Liu D., Gorshkov A., Wilmarth T. Proceedings of the 17th Python in Science Conference. 2018. Composable multi-threading and multi-processing for numeric libraries. [DOI] [Google Scholar]
  • 41.Varma S., Simon R. Bias in error estimation when using cross-validation for model selection. BMC Bioinformatics. 2006;7:91. doi: 10.1186/1471-2105-7-91. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Kohavi R. Vol. 2. 1995. A study of cross-validation and bootstrap for accuracy estimation and model selection; pp. 1137–1145. (Proceedings of the 14th International Joint Conference on Artificial Intelligence). [Google Scholar]
  • 43.Cawley G.C., Talbot N.L.C. On over-fitting in model selection and subsequent selection bias in performance evaluation. J. Mach. Learn. Res. 2010;11:2079–2107. [Google Scholar]
  • 44.Baldi P., Brunak S., Chauvin Y., Andersen C.A.F., Nielsen H. Assessing the accuracy of prediction algorithms for classification: an overview. Bioinformatics. 2000;16:412–424. doi: 10.1093/bioinformatics/16.5.412. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Document S1. Supplemental experimental procedures, Figures S1–S9, Tables S1–S6, and Data S1
mmc1.pdf (2.2MB, pdf)
Data S2. Tuned hyperparameters of different Machine Learning algorithms
mmc2.xlsx (25.6KB, xlsx)
Document S2. Article plus supplemental information
mmc3.pdf (3.9MB, pdf)

Data Availability Statement

Data and code can be accessed at the following link: https://github.com/abhijitmjj/DNA-structure-prediction (code author and repo maintainer: Abhijit Gupta [github username: abhijitmjj]).

DNA structure prediction from its sequence code is available in the GitHub repository (https://github.com/abhijitmjj/DNA-structure-prediction). We intend to build a webserver for our program soon, where the user can provide raw sequences as the input and obtain the probabilities for them to attain A/B form conformation.


Articles from Patterns are provided here courtesy of Elsevier

RESOURCES