Skip to main content
Frontiers in Bioengineering and Biotechnology logoLink to Frontiers in Bioengineering and Biotechnology
. 2020 Jul 14;8:808. doi: 10.3389/fbioe.2020.00808

Riboflow: Using Deep Learning to Classify Riboswitches With ∼99% Accuracy

Keshav Aditya R Premkumar 1,, Ramit Bharanikumar 2,, Ashok Palaniappan 3,*
PMCID: PMC7371854  PMID: 32760712

Abstract

Riboswitches are cis-regulatory genetic elements that use an aptamer to control gene expression. Specificity to cognate ligand and diversity of such ligands have expanded the functional repertoire of riboswitches to mediate mounting apt responses to sudden metabolic demands and signal changes in environmental conditions. Given their critical role in microbial life, riboswitch characterisation remains a challenging computational problem. Here we have addressed the issue with advanced deep learning frameworks, namely convolutional neural networks (CNN), and bidirectional recurrent neural networks (RNN) with Long Short-Term Memory (LSTM). Using a comprehensive dataset of 32 ligand classes and a stratified train-validate-test approach, we demonstrated the accurate performance of both the deep learning models (CNN and RNN) relative to conventional hyperparameter-optimized machine learning classifiers on all key performance metrics, including the ROC curve analysis. In particular, the bidirectional LSTM RNN emerged as the best-performing learning method for identifying the ligand-specificity of riboswitches with an accuracy >0.99 and macro-averaged F-score of 0.96. An additional attraction is that the deep learning models do not require prior feature engineering. A dynamic update functionality is built into the models to factor for the constant discovery of new riboswitches, and extend the predictive modeling to new classes. Our work would enable the design of genetic circuits with custom-tuned riboswitch aptamers that would effect precise translational control in synthetic biology. The associated software is available as an open-source Python package and standalone resource for use in genome annotation, synthetic biology, and biotechnology workflows.

Keywords: riboswitch family, synthetic biology, machine learning, convolutional neural network, recurrent neural network, hyperparameter optimization, multiclass ROC, clustering

Availability:

PyPi package: riboflow @ https://pypi.org/project/riboflow

Repository with Standalone suite of tools: https://github.com/RiboswitchClassifier

Language: Python 3.6 with numpy, keras, and tensorflow libraries.

License: MIT.

Introduction

Riboswitches are ubiquitous and critical metabolite-sensing gene expression regulators in bacteria that are capable of folding into at least two alternative conformations of 5′UTR mRNA secondary structure, which functionally switch gene expression between on and off states (Mandal et al., 2003; Roth and Breaker, 2009; Serganov and Nudler, 2013). The selection of conformation is dictated by the binding of ligand cognate to the aptamer domain of a given riboswitch (Gelfand et al., 1999; Winkler et al., 2002, 2004). Cognate ligands are key metabolites that mediate responses to metabolic or external stimuli. Consequent to conformational changes, riboswitches weaken transcriptional termination or occlude the ribosome binding site thereby inhibiting translation initiation of associated genes (Yanofsky, 1981; Mandal and Breaker, 2004). Riboswitches provide an intriguing window into the ‘RNA world’ biology (Stormo and Ji, 2001; Brantl, 2004; Breaker et al., 2006; Strobel and Cochrane, 2007) and there is evidence of their wider distribution in complex genomes (Sudarsan et al., 2003; Barrick and Breaker, 2007; Bocobza and Aharoni, 2014; McCown et al., 2017). The modular properties of riboswitches have engendered the possibility of synthetic control of gene expression (Tucker and Breaker, 2005), and combined with the ability to engineer binding to an ad hoc ligand, riboswitches have turned out to be a valuable addition to the synthetic biologist’s toolkit (Wieland and Hartig, 2008; Wittmann and Suess, 2012). In addition to orthogonal gene control they are useful in a variety of applications, notably metabolic engineering (Zhou and Zeng, 2015), biosensor design (Yang et al., 2013; Meyer et al., 2015) and genetic electronics (Villa et al., 2018). Riboswitches have been used as basic computing units of a complex biocomputation network, where the concentration of the ligand of interest is titrated into measurable gene expression (Beisel and Smolke, 2009; Domin et al., 2017). Riboswitches have also been directly used as posttranscriptional and translational checkpoints in genetic circuits (Chang et al., 2012). Their key functional roles in infectious agents but absence in host genomes make them attractive targets for the design of cognate inhibitors (Blount and Breaker, 2006; Deigan and Ferré-D’Amaré, 2011; Wang et al., 2017). Characterisation of riboswitches would expand the repertoire of translational control options in synthetic biology and bioengineering. In turn, this would facilitate the reliable construction of precise genetic circuits. In view of their myriad applications, robust computational methods for the accurate characterisation of novel riboswitch sequences would be of great value.

Since the discovery of riboswitches (Mironov et al., 2002; Nahvi et al., 2002), many computational efforts have been advanced for their characterisation, notably Infernal (Nawrocki and Eddy, 2013), Riboswitch finder (Bengert and Dandekar, 2004), RibEx (Abreu-Goodger and Merino, 2005), RiboSW (Chang et al., 2009) and DRD (Havill et al., 2014), and reviewed in Clote (2015) and Antunes et al. (2017). These methods used probabilistic models of known classes with or without secondary structure information to infer or predict the riboswitch class. Singh and Singh explored featuring mono-nucleotide and di-nucleotide frequencies in a supervised machine learning framework to classify different riboswitch sequences, and concluded that the multi-layer perceptron was optimal (Singh and Singh, 2016). Their work achieved modest performance (F-score of 0.35 on 16 different riboswitch classes). None of the above methods were shown to generalize effectively to unseen riboswitches. Our remedy was to explore the use of deep learning models for riboswitch classification. Deep networks are relatively recent neural network-based frameworks that use a type of learning known as representation learning (Bengio et al., 2013). Convolutional neural networks are one type of deep learning, known for hierarchical information extraction. Such architectures with alternating convolutional and pooling layers have been earlier used to extract structural and functional information from genome sequences (Alipanahi et al., 2015; Sønderby et al., 2015; Zhou and Troyanskaya, 2015; Kelley et al., 2016). Recurrent neural networks are counterparts to CNNs and specialize in extracting features from time-series data (Che et al., 2018). RNNs with Long Short-Term Memory (termed LSTM) incorporate recurrent connections to model long-run dependencies in sequential information (Hochreiter and Schmidhuber, 1997), such as in speech and video (Graves and Schmidhuber, 2005). This feature of LSTM RNNs immediately suggests their use in character-level modeling of biological sequence data (Lipton, 2015; Lo Bosco and Di Gangi, 2017). Bidirectional LSTM RNN have been shown to be especially effective, given that they combine the outputs of two LSTM RNNs, one processing the sequence from left to right, the other one from right to left, together enabling the capture of dynamic temporal or spatial behavior (Sundermeyer et al., 2014). Bidirectional LSTM RNNs are a particularly powerful abstraction for modeling nucleic acid sequences whose spatial secondary structure determines function (Lee et al., 2015). Two recent successes of deep learning methods in RNA biology have been: (i) prediction of RNA secondary structure (Singh et al., 2019), and (ii) dynamic range improvement in riboswitch devices (Groher et al., 2019). Here we have evaluated the relative merits of a spectrum of state-of-the-art learning methods for resolving the ligand-specificity of riboswitches from sequence. It is demonstrated that the deep learning models vastly outperformed other machine learning models with respect to the classification of riboswitches belonging to 32 different families.

Materials and Methods

Dataset and Pre-processing

We searched the Rfam database of RNA families (Kalvari et al., 2018) with the term “Riboswitch AND family” and the corresponding hit sequences were obtained in FASTA format from the Rfam ftp server (Rfam v13 accessed on July 6, 2019). Each riboswitch is represented by the coding strand sequence, with uracil replaced by thymine, thereby conforming to the nucleotide alphabet ‘ACGT.’ Each sequence was scanned for non-standard letters (i.e., other than the alphabet) and such occurrences were corrected using the character mapping defined in Table 1. The feature vectors for machine learning were extracted from the sequences. For each sequence, 20 features were computed, comprising four mononucleotide frequencies (A,C,G,T) and 16 dinucleotide frequencies. To address possible skew in distribution, all the frequency features were normalized to zero mean and unit variance. Deep models, namely convolutional neural networks (CNNs) and bidirectional recurrent neural networks with LSTM (hereafter simply referred as RNNs) are capable of using the sequences directly as the feature space, obviating any need for feature engineering. We used the first 250 bases of the riboswitch sequence as the input, with the proviso that shorter sequences (which is usually the case; Table 2) were padded for the extra spaces. Python scripts used to create the final dataset are available in the repository for this project1. The dataset consists of the riboswitch sequence, four 1-mer frequencies, 16 2-mer frequencies, and class, for each instance, which could be appropriately subsetted for training the base and deep models.

TABLE 1.

Non-standard nucleotide mapping.

S. No. Original letter Mapped character #Occurrences in dataset
1 R G 6
2 Y T 8
3 K G 1
4 S G 3
5 W A 2
6 H A 2

Rare occurrences of non-standard nucleotides in the sequences were converted using this mapping key.

TABLE 2.

A summary of the riboswitch dataset used in our study.

Class no. Rfam ID Class Name Class size Avg. length
1 RF00504 Glycine riboswitch 4592 100
2 RF01786 Cyclic di-GMP-II riboswitch 661 86
3 RF01750 ZMP/ZTP riboswitch 1674 92
4 RF00059 Thiamine pyrophosphate riboswitch 12559 110
5 RF01057 SAH Riboswitch 832 92
6 RF01725 SAM -1 -4 Variant riboswitch 793 104
7 RF00162 SAM - 1 Riboswitch 6027 113
8 RF00174 Cobalamin riboswitch 14212 189
9 RF01055 Molybdenum Co-factor riboswitch 1221 134
10 RF01727 SAM-SAH Riboswitch 240 50
11 RF01482 AdoCbl riboswitch 182 137
12 RF03057 nhaA-I RNA 559 56
13 RF01734 Fluoride Riboswitch 2018 70
14 RF00167 Purine Riboswitch 2632 101
15 RF00234 Glucosamine-6-phosphate riboswitch 936 175
16 RF01739 Glutamine riboswitch 1103 64
17 RF03072 raiA RNA 736 219
18 RF03058 sul1 RNA 344 56
19 RF00380 Ykok riboswitch (Magnesium sensing) 1059 170
20 RF00168 Lysine Riboswitch 2240 180
21 RF03071 DUF1646 265 53
22 RF01689 AdoCbl variant RNA 212 125
23 RF00379 ydaO/yuaA leader 3918 164
24 RF00634 SAM - 4 Riboswitch 1245 116
25 RF01767 SAM - 3 Riboswitch 195 90
26 RF00080 yybP-ykoY manganese riboswitch 833 158
27 RF02683 NiCo riboswitch (Nickel or Cobalt sensing) 235 97
28 RF00442 Guanidine-I riboswitch 902 109
29 RF00522 Pre-queosine riboswitch -1 533 45
30 RF00050 Flavin Mononucleotide Riboswitch 4080 142
31 RF01831 THF riboswitch 663 102
32 RF00521 SAM - 2 Riboswitch 819 78

The dataset includes a mixture of metabolite/ion/cofactor/amino-acid/nucleotide/vitamin/signaling-molecule aptamer ligands. ‘Class no.’ corresponds to the response classes to be learnt in machine learning. The average length of all sequences in a given class is also given.

Predictive Modeling

The machine learning problem is simply stated as: given the riboswitch sequence, predict the ligand class of the riboswitch. Toward this, a battery of eight supervised machine learning and deep classifiers were studied and evaluated in the present work (Table 3). Classifiers derived from implementations in the Python scikit-learn machine learning library (Pedregosa et al., 2011)2 are referred to as base models and include the Decision Tree, K-nearest Neighbors, Gaussian Naive Bayes, the ensemble classifiers AdaBoost and Random Forest and the Multi-layer Perceptron. The deep classifiers namely CNN and RNN derived from implementations in the Python keras library3 on tensorflow (Abadi et al., 2015). Three scripts in the repository, namely baseModels.py, rnnApp.py, and cnnApp.py, implement the base models, RNN, and CNN, respectively. For both the base and deep classifiers, the dataset was split into 0.9:0.1 training:test sets. Multi-class modeling is fraught with overfitting to particular classes (especially pronounced in cases of extreme class skew). To address this issue, two strategies were adopted: (i) the splitting process was stratified on the class, which ensured that each class was proportionately and sufficiently distributed in both the training and test sets, and (ii) hyperparameter optimisation, discussed below.

TABLE 3.

Description of the base model and deep classifiers.

Classifier Features used Hyperparameters of interest ML Library
Decision Tree 1- and 2-mer frequencies Maximum features, Minimum sample split, Minimum sample leaf, Random state, Max depth Sklearn
Gaussian Naïve Bayes 1- and 2-mer frequencies Priors Sklearn
k-Nearest Neighbors 1- and 2-mer frequencies Number of neighbors Leaf size, Weights, Algorithm Sklearn
Adaptive Boosting 1- and 2-mer frequencies Number of estimators, Learning rate, Algorithm Sklearn
Random Forest 1- and 2-mer frequencies Number of estimators, Max depth, Impurity criterion Sklearn
Multi-layer Perceptron 1- and 2-mer frequencies Activation, Solver, Alpha (regularization term), Learning rate, epochs, hidden layers, nodes per hidden layer Sklearn
CNN Riboswitch sequence Number of filters, Kernel size, Activation function, Pooling method, number of Conv1D layers, Dropout ratio, Optimiser, #Epochs TensorFlow (Keras)
Bidirectional RNN with LSTM Riboswitch sequence Activation function, number of LSTM nodes, number of Bidirectional layers, Dropout ratio, Optimiser, #Epochs TensorFlow (Keras)

Hyperparameters noted for each classifier are meant to be representative. For the deep models, any long riboswitch genome sequences were truncated to 250 nucleotides, which is an adjustable parameter (max_len) set to much larger than the average length of any riboswitch class. The Python3 library used for implementation of machine learning model is noted.

Hyperparameter Optimisation

Hyperparameter fine tuning for each classifier was effected by discrete combinatorial grid search on the hyperparameters associated with that classifier. The grid search was evaluated with 10-fold cross-validation of the training set. This yielded the optimal hyperparameters for each classifier. The scripts for hyperparameter optimisation of the base models are available in the repository. In the case of the deep models, we used a train-validate-test approach to model optimisation with keras/TensorFlow, by setting the ‘validation’ flag to 0.1.

Evaluation Metrics

The performance of each classifier was evaluated on the test set using the receiver operating characteristic (ROC) analysis in addition to standard metrics such as the precision, recall, accuracy and F-score (harmonic mean of precision and recall) (van Rijsbergen, 1975). The ROC curve was obtained by plotting the TP rate vs. the FP rate, i.e., sensitivity vs. (1 – specificity), and the area under the ROC curve (AUROC) could be estimated to rate the model’s performance. AUROC represents the probability that a given classifier would rank a randomly chosen positive instance higher than a randomly chosen negative one. ROC analysis is robust to class imbalance, typical of the machine learning problem at hand, however, a multi-class adaptation of the binary ROC is necessary. For each classifier, this is achieved by computing classwise binary AUROC values in a one-vs.-all manner, followed by aggregating the classwise AUROC values into a single multi-class AUROC measure (Yang, 1999; Manning et al., 2008). Aggregation of the classwise AUROC values could be done in at least two ways:

  • 1.

    Micro-average AUROC, where each instance is given equal weight.

  • 2.

    Macro-average AUROC, where each class is given equal weight.

In micro-averaging, all the instances from different classes are treated equally, to arrive at the final metrics. In particular, the microaverage AUROC is given by the area under the overall TP rate vs. overall FP rate curve.

On the other hand, the macro-average of a given metric for a multi-class prediction problem is estimated by the average of the metric for the individual classes. For example, the macro-average AUROC is given by:

Macro-averageAUROC=(AUROC+1AUROC+2+AUROC)32/32

where AUROCi is the binary AUROC for the ith class.

It is clear that the micro-average AUROC would be dominated by the larger classes, while the macro-average AUROC is a more balanced measure. Both the micro-average and macro-average AUROC measures were used to evaluate the performance of all our classifiers. A python script, multiclass.py available in the repository, generates all the performance metrics and plots.

Dynamic Extension of the Models

Genome sequencing of diverse, exotic prokaryotes is likely to yield new regulatory surprises mediated by riboswitches (Breaker, 2011). A model that could classify a fixed set of 32 classes remains static in the wake of exponentially growing number of known genomes. To address the challenge of extending the model to new classes, we have devised a strategy for dynamically updating the model. This process is initiated by feeding the sequences corresponding to the new class(es) to an updater script, which revises the dataset and then trains a new model. The automation of modeling would ensure a user-friendly pipeline for learning any number of riboswitch classes along with the production of performance metrics of such models. The script dynamic.py, available in the repository, implements the dynamic functionality of our modeling effort.

Results

Our Rfam query retrieved 39 riboswitch class hits, however, seven of these classes had a membership of less than 100 sequences and were filtered out. Subsequently, our dataset consisted of 32 riboswitch classes with a total of 68,520 sequences. A summary of this dataset is presented in Table 2. The largest classes include the cobalamin and thiamine pyrophosphate classes, with >10,000 riboswitches in each, accounting for considerable diversity within classes. Classes with >4,000 members include Flavin mononucleotide (FMN) and glycine riboswitches. Other notable classes with at least 1,000 members include the lysine, purine, fluoride and glutamine switches. The riboswitch sequences were inspected for the standard alphabet (Table 1) and the final pre-processed comma-separated values (csv) datafile with each instance containing the sequence, 20 features and riboswitch class was prepared (available at4).

Table 3 recapitulates the key properties of the classifier models used in this study. We noticed poor performance of the base models on the test set with default model parameters, which could be traced to persistent overfitting (dominated by the larger classes), despite stratified sampling. Hyperparameter optimisation of the default parameters is one solution to address this problem and was carried out on the base models. The exercise is summarized in Appendix I (Supplementary File S1), which includes the final configuration of the hyperparameter space for all the base and deep classifiers. The optimized CNN and RNN architectures are illustrated in Figure 1. In the CNN, two convolutional layers were used followed by a pooling layer and dropout layer before flattening to a fully connected output layer. The RNN employed two sophisticated bidirectional LSTM units sandwiched by dropout layers before flattening to a fully connected output layer. The number of training epochs necessary for each deep model was determined based on the convergence of the error function (shown in Figure 2).

FIGURE 1.

FIGURE 1

Deep learning frameworks used in the study. (A), CNN architecture, optimized for two 1-dimensional convolutional layers; and (B), Bidirectional RNN with LSTM, optimized for two bidirectional layers. Two dropout layers are used in the RNN.

FIGURE 2.

FIGURE 2

Epoch tuning curves for the CNN (A) and RNN (B). The CNN converges faster with respect to the number of epochs, however, the RNN learns better, as seen with the continuously decreasing loss function.

With the optimized classifiers, the performance of the predictive modeling was evaluated on the unseen testing set. Figure 3 shows the resultant classifier performance by an array of metrics including accuracy, and F-score. It is abundantly clear that the deep models vastly outperformed the base classifiers in all metrics across all classes. Figures 4, 5 show the ROC curves along with the micro-averaged and macro-averaged AUROC for the base models and the deep models, respectively. The AUROC is indicative of the quality of the overall model. It is seen that the AUROC is 1.00 for all classes for the RNN. Table 4 summarizes the performance of the classifiers, with the detailed classwise F-score of each classifier available in the Supplementary Table S2 and the classwise break-up of the AUROC of all classifiers in the Supplementary Table S3.

FIGURE 3.

FIGURE 3

Standard performance metrics. Clockwise from top left, Accuracy; Precision; F-score; and Recall. The overall precision, recall and F-score were computed by macro-averaging the classwise scores. The deep models emerged as vastly superior alternatives to the base machine learning models on all performance metrics.

FIGURE 4.

FIGURE 4

AUROC for the base models. (A) Decision Tree, (B) Gaussian NB, (C) kNN, D: AdaBoost, (E) Random Forest, and (F) Multi-layer Perceptron. Gray lines denote classwise AUROCs of all 32 classes, from which it is clear that not all classes are equally learnt.

FIGURE 5.

FIGURE 5

AUROC for the deep models. (A) CNN, and (B) RNN. Gray lines denote classwise AUROCs of all 32 classes. It clear that RNN achieves learning perfection at both the macro and classwise levels.

TABLE 4.

Performance metrics for all classifiers.

Model Accuracy Precision Recall F-score Micro AUROC Macro AUROC
Decision Tree 0.54 0.49 0.39 0.42 0.88 0.81
Gaussian NaïveBayes 0.37 0.39 0.46 0.38 0.9 0.89
K-neighbors 0.67 0.75 0.55 0.61 0.94 0.88
AdaBoost 0.47 0.42 0.36 0.36 0.89 0.92
Random Forest 0.71 0.86 0.58 0.65 0.98 0.97
Multi-layer perceptron 0.74 0.75 0.67 0.70 0.99 0.98
CNN 0.97 0.98 0.91 0.93 1 1
RNN 0.99 0.97 0.96 0.96 1 1

The macro-averaged values of precision, recall and F-score are shown. Micro AUROC, micro-average AUROC; Macro AUROC, macro-average AUROC.

Discussion

The RNN model marginally (but clearly) outperformed the CNN model, and both of them significantly outperformed all the base models on all key metrics, notably accuracy and F-score. The best-performing among the base models was the Multi-layer Perceptron. It is noteworthy that the AUROC approached unity and near-perfection for both the deep models, especially the bidirectional RNN with LSTM. This implied that the use of k-mer features masked long-range information whereas the deep models were able to capture such correlations directly from the full sequence. These results affirmed that RNNs could be used to effectively simulate the interactions in biological sequences.

The F-score (a measure of balanced accuracy) is a more unforgiving metric than AUROC in the case of multi-class problems (Table 4). While the CNN and RNN had macro-averaged F-scores of 0.93 and 0.96, respectively, none of the base models exceeded 0.70 including the multilayer perceptron. Supplementary Table S2 provides the classwise F-scores of all classifiers. All the base models struggled to classify the largest riboswitch classes, namely TPP and Cobalamin. This is a consequence of the diversity of such large riboswitch classes, making the ‘outlier’ members harder to classify correctly. Both the RNN and CNN are mapping the sequence input to its corresponding riboswitch family. In such a case, the sequence similarity within a family and sequence dissimilarity across families represent plausible discriminating features that the models are learning. Higher order features such as sequence context and base dependencies dictated by RNA secondary structure also constitute learnable information. Table 5 shows the results of a sequence-based clustering analysis of the riboswitch families using the cd-hit algorithm (Li and Godzik, 2006). It is seen that there are >7000 and >10,000 singleton clusters in the TPP and Cobalamin classes, respectively. Here we introduce a diversity metric for riboswitch families, defined as the ratio of the number of clusters at 90% sequence identity to the total size of the family. Compared to the overall diversity score of 0.7, both TPP and Cobalamin classes had above-average diversities (0.71 and 0.82, respectively). However, these diverse classes did not pose any problems for the deep models. On the contrary, the AdoCbl and AdoCbl variant riboswitch classes posed significant learning challenges to both base and deep models. AdoCbl in particular is the smallest riboswitch family considered here, but is also unnaturally diverse (with a score of 0.83). This frustrates learning, because the limited number of instances do not adequately represent the class diversity, and result in class outliers. Consequently this emerged as the most challenging for all classifiers, reflected in the classwise F-scores (Supplementary Table S2). Four of the base classifiers managed a zero F-score, while the CNN and RNN achieved F-scores of 0.38 and 0.67, respectively, by far their lowest for any class. Even here, the consistency of the RNN model is remarkable. The other classes that were notably challenging to the base models but not to the deep models included Cyclic di-GMP-II, ZMP/ZTP, SAM 1–4 variant, Molybdenum co-factor, Glucosamine-6-phosphate and Guanidine-I riboswitch classes. Of these, the Glucosamine-6-phosphate, Cyclic di-GMP-II, and Molybdenum co-factor riboswitch classes are among the most diverse riboswitch families, with diversity scores ≥0.80. It is seen that the classification problems arise either with diverse classes or at the extremes of class sizes. Too large the class, the diversity is challenging, whereas too small and the learning itself is incomplete and challenged. The deep models – RNN and CNN – consistently performed well across all classes, independent of the size of the class. It could be inferred in this case that using direct features (i.e., sequences) rather than engineered features (i.e., k-mer frequencies) led to more robust models. From Table 5, it is also clear that most of the learnt classes (especially the large ones) are diverse, thereby elevating the classifier performance to robustness against adversarial attacks – that is, changing a few nucleotides in the input sequence would be unlikely to drastically alter the class prediction.

TABLE 5.

Clustering analysis of riboswitch families at 90% sequence identity.

Rfam ID #Clusters at 90% Diversity #Singleton clusters #Redundant sequences #Clusters in the rest
RF00050 2484 0.61 1866 46 45208
RF00059 8954 0.71 7366 272 38738
RF00080 611 0.73 493 29 47081
RF00162 3736 0.62 2722 141 43956
RF00167 2122 0.81 1818 59 45570
RF00168 1905 0.85 1707 45 45787
RF00174 11620 0.82 10130 289 36077
RF00234 814 0.87 731 11 46878
RF00379 2583 0.66 2044 97 45109
RF00380 441 0.42 265 35 47251
RF00442 600 0.67 449 46 47092
RF00504 3010 0.66 2345 66 44682
RF00521 495 0.61 379 20 47197
RF00522 214 0.40 113 28 47478
RF00634 501 0.40 317 12 47191
RF01055 978 0.80 865 73 46714
RF01057 629 0.76 519 13 47063
RF01482 150 0.83 132 12 47546
RF01689 131 0.62 101 2 47562
RF01725 435 0.55 326 25 47257
RF01727 149 0.62 105 16 47543
RF01734 1616 0.80 1389 127 46076
RF01739 344 0.31 215 56 47348
RF01750 1074 0.64 821 94 46618
RF01767 122 0.63 88 12 47570
RF01786 555 0.84 480 49 47137
RF01831 450 0.68 327 57 47242
RF02683 159 0.68 122 31 47533
RF03057 359 0.64 266 70 47333
RF03058 97 0.28 57 17 47595
RF03071 86 0.33 48 29 47606
RF03072 273 0.37 141 72 47419
ALL 47686 0.70 38871 1951 N-

Singleton clusters indicate clusters with only one representative sequence. Redundant sequences are 100% identical to another member in the riboswitch family. The full set of riboswitch sequences (ALL) and ALL minus the riboswitch family of interest were also clustered.

These results might be put in perspective by benchmarking against the existing literature. Guillen-Ramirez and Martinez-Perez (Guillén-Ramírez and Martínez-Pérez, 2018) extended the k-mer features logic and arrived at an optimal combination of 5460 k-mer features. Using a limited dataset of 16 classes, they used state-of-the-art machine learning to achieve accuracies in the high nineties, however, their results did not generalize equally to riboswitches with remote homology. For e.g., their best-performing classifier (Multi-layer Perceptron) misclassified 6 out of the 225 instances of Lysine riboswitch as cobalamin-gated. The source code for the features and modeling used in their work is not readily available for new applications. To make the workflow described in our study easily reproducible and user-friendly, we have developed a Python package riboflow5 mirroring the best RNN model. In addition to predicting the most probable class of a given riboswitch sequence, riboflow provides an option to predict the complete vector of class probabilities, which could be helpful in disambiguating any class confusion. It would also inform the design of synthetic orthogonal riboswitches for biotechnology applications. The implementation and usage details are provided in Appendix II (Supplementary File S1).

An interesting benchmark is afforded by the Riboswitch Scanner (Mukherjee and Sengupta, 2016), which used profile HMMs (Eddy, 2011) of riboswitch classes to detect riboswitches in genomic sequences. While our method addresses inter-class discrimination of riboswitch sequences, Riboswitch Scanner is a web-server that essentially performs riboswitch vs. not-riboswitch discrimination for user-given riboswitch classes. The absence of F-score metrics does not allow for direct comparisons, however, the sensitivity and specificity seemed consistently comparable for most classes, with noticeable variations with respect to the Glycine, THF and SAM I-IV variant riboswitch classes. It must be indicated that their method is validated with Rfam seed sequences, without consideration for the proliferation of riboswitch sequences. Performance evaluation on limited data could inflate performance estimates and complicate their interpretation. We further note that it is possible to extend our method to the ‘riboswitch-or-not’ classification problem by calibrating the prediction probability thresholds. In any case, our work enables the ranking of riboswitches using the strength of the predicted probabilities, which would aid the selection of the best riboswitch sequence design.

It must be noted that riboswitches are precisely specific to cognate ligands. For e.g., the AdoCbl riboswitch would not tolerate a methyl-substituted cobalamin (Nahvi, 2004) nor does the TPP riboswitch interact with thiamine or thiamine monophosphate (Lang et al., 2007). At the same time, these two riboswitches are very diverse in their phylogenomic distribution and actual sequences. The key to effective learning lies in treading a fine line between the intra-class diversity and inter-class specificity. It is remarkable the bidirectional LSTM RNN was able to achieve exactly this tradeoff. The roots of such performance of the deep models in general has been explained recently to be related to the lottery ticket hypothesis (Frankle and Carbin, 2019) as well as learning the intrinsic dimension of the problem (Li et al., 2018), here the classification of riboswitches.

To extend the functionality of our work, we have introduced a dynamic component to all our models, both base and deep. With the exponential growth in genome sequencing, the room for riboswitch discovery is enormous (Yu and Breaker, 2020). Our models could accommodate new riboswitch class definitions by way of dataset augmentation, thereby making them general and more robust. This work used the dynamic functionality to extend a preliminary 24-class model to the present 32 classes with sustained performance. The implementation and usage details of the dynamic functionality and other utilities provided in the repository are given in Appendix II (Supplementary File S1). It is noted that the deep learning models could be adapted to new classes and related problems by the technique of transfer learning (Weiss et al., 2016). Addition of new data to existing models presents data quality issues, which remain contentious (Leonelli, 2019), and could be partially addressed using the tools employed in this study to assess the canonical Rfam dataset.

In summary, we present riboflow, a python package (see foot note 5) as well as standalone suite of tools, that have been validated and thoroughly tested on 32 riboswitch classes. By using large and complete datasets, the variance of our modeling procedure has been optimized, and this ensures the generality and applicability of our models on new instances/classes without compromise of performance. riboflow is an off-the-shelf solution that is ready for programmatic incorporation of the RNN model into automatic annotation and design pipelines. All of our trained models are available to the interested user as ‘pickled’ models from https://github.com/RiboswitchClassifier. Riboswitches are a cornerstone of progress in synthetic biology, representing key checkpoints for translation activation. Our work presents an intuitive general-purpose extensible platform for the effortless characterization of new riboswitch sequences and classes, which would accelerate bacterial genome annotation, synthetic biology, and biotechnology, including the rapid design of novel genetic circuits with exquisite specificity. The predicted probabilities of class membership could be used as a proxy of aptamer binding strength with cognate signaling molecule, and this paves the way for the design of effective riboswitches for any stimulus or set of stimuli. In addition to being indispensable workhorses in synthetic biology, riboswitches represent novel and exciting targets for the development of new class of antibiotics (Penchovsky et al., 2020), and our work would also help toward the design of riboswitch inhibitors to combat emerging and multi-drug resistant pathogens. In addition, our work opens up the applications of deep learning methods, including advanced relatives like stacked Bi-LSTM and attention models (Vaswani et al., 2017), for addressing related and unrelated problems in the biological realm.

Conclusion

We have demonstrated that CNN and RNN, without needing prior feature extraction, are capable of robust multi-class learning of ligand specificity from riboswitch sequence, with the RNN posting an F-score of ∼0.96. The confidence of classification could be obtained from an inspection of the predicted classwise probabilities. The bidirectional LSTM RNN model has been packaged into riboflow to enable embedding into genome annotation pipelines, genetic circuit-design automation, and biotechnology workflows. The CNN shows the best tradeoff between the time-cost of training the model and overall performance and could be applied to the task of learning new riboswitch classes using the provided dynamic update option that is provided. All the code used in our study is freely available for any use and further improvement by the scientific community as well as in the larger interest of reproducible research. Our study has highlighted the use of macro-averaged F-score as a disciminating objective metric of classifier performance on multi-class data. Our work reaffirms the competitive advantages of bidirectional LSTM RNNs over conventional machine learning and hidden markov profiles in modeling data sequences, and opens up their applications for modeling other non-coding RNA elements.

Data Availability Statement

All datasets generated for this study are included in the article/Supplementary Material.

Author Contributions

AP conceived and designed the work and wrote the manuscript. KP, RB, and AP performed the experiments and analyzed the data. All authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We are grateful to the School of Chemical and Biotechnology, SASTRA Deemed University for infrastructure and computing support. A preliminary version of the manuscript has been released as a Pre-Print at bioRxiv (Premkumar et al., 2019).

Funding. This work was funded in part by DST-SERB grant EMR/2017/000470 and the SASTRA TRR grant (to AP).

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fbioe.2020.00808/full#supplementary-material

References

  1. Abadi M., Agarwal A., Barham P., Brevdo E., Chen Z., Citro C., et al. (2015). TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. Available online at: www.tensorflow.org (accessed May 10, 2018). [Google Scholar]
  2. Abreu-Goodger C., Merino E. (2005). RibEx: a web server for locating riboswitches and other conserved bacterial regulatory elements. Nucleic Acids Res. 33(Suppl. 2), W690–W692. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Alipanahi B., Delong A., Weirauch M. T., Frey B. J. (2015). Predicting the sequence specificities of DNA- and RNA-binding proteins by deep learning. Nat. Biotechnol. 33 831–838. 10.1038/nbt.3300 [DOI] [PubMed] [Google Scholar]
  4. Antunes D., Jorge N., Caffarena E. R., Passetti F. (2017). Using RNA sequence and structure for the prediction of riboswitch aptamer: a comprehensive review of available software and tools. Front. Genet. 8:231. 10.3389/fgene.2017.00231 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Barrick J. E., Breaker R. R. (2007). The distributions, mechanisms, and structures of metabolite-binding riboswitches. Genome Biol. 8:R239. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Beisel C. L., Smolke C. D. (2009). Design principles for riboswitch function. PLoS Comput. Biol. 5:e1000363. 10.1371/journal.pcbi.1000363 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bengert P., Dandekar T. (2004). Riboswitch finder—a tool for identification of riboswitch RNAs. Nucleic Acids Res. 32(Suppl. 2), W154–W159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bengio Y., Courville A., Vincent P. (2013). Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intelligence 35 1798–1828. 10.1109/tpami.2013.50 [DOI] [PubMed] [Google Scholar]
  9. Blount K. F., Breaker R. R. (2006). Riboswitches as antibacterial drug targets. Nat. Biotechnol. 24 1558–1564. 10.1038/nbt1268 [DOI] [PubMed] [Google Scholar]
  10. Bocobza S. E., Aharoni A. (2014). Small molecules that interact with RNA: riboswitch-based gene control and its involvement in metabolic regulation in plants and algae. Plant J. 79 693–703. 10.1111/tpj.12540 [DOI] [PubMed] [Google Scholar]
  11. Brantl S. (2004). Bacterial gene regulation: from transcription attenuation to riboswitches and ribozymes. Trends Microbiol. 12 473–475. 10.1016/j.tim.2004.09.008 [DOI] [PubMed] [Google Scholar]
  12. Breaker R. R. (2011). Prospects for riboswitch discovery and analysis. Mol. Cell 43 867–879. 10.1016/j.molcel.2011.08.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Breaker R. R., Gesteland R. F., Cech T. R., Atkins J. F. (2006). The RNA World. New York, NY: Cold Spring Harbor Laboratory Press. [Google Scholar]
  14. Chang C. L., Lei Q., Lucks J. B., Segall-Shapiro T. H., Wang D., Mutalik V. K. (2012). An adaptor from translational to transcriptional control enables predictable assembly of complex regulation. Nat. Methods 9 1088–1094. 10.1038/nmeth.2184 [DOI] [PubMed] [Google Scholar]
  15. Chang T.-H., Huang H.-D., Wu L.-C., Yeh C.-T., Liu B.-J., Horng J.-T. (2009). Computational identification of riboswitches based on RNA conserved functional sequences and conformations. RNA 15 1426–1430. 10.1261/rna.1623809 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Che Z., Purushotham S., Cho K., Sontag D., Liu Y. (2018). Recurrent neural networks for multivariate time series with missing values. Sci. Rep. 8:6085. 10.1038/s41598-018-24271-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Clote P. (2015). Computational prediction of riboswitches. Methods Enzymol. 553 287–312. 10.1016/BS.MIE.2014.10.063 [DOI] [PubMed] [Google Scholar]
  18. Deigan K. E., Ferré-D’Amaré A. R. (2011). Riboswitches: discovery of drugs that target bacterial gene-regulatory RNAs. Acc. Chem. Res. 44 1329–1338. 10.1021/ar200039b [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Domin G., Findeiß S., Wachsmuth M., Will S., Stadler P. F., Mörl M. (2017). Applicability of a computational design approach for synthetic riboswitches. Nucleic Acids Res. 45 4108–4119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Eddy S. R. (2011). Accelerated profile HMM searches. PLoS Comput. Biol. 7:e1002195. 10.1371/journal.pcbi.1002195 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Frankle J., Carbin M. (2019). The lottery ticket hypothesis: finding sparse, trainable neural networks. arXiv [Preprint]. Available online at: https://arxiv.org/abs/1803.03635 (accessed March 4, 2019). [Google Scholar]
  22. Gelfand M. S., Mironov A. A., Jomantas J., Kozlov Y. I., Perumov D. A. (1999). A conserved RNA structure element involved in the regulation of bacterial riboflavin synthesis genes. Trends Genet. 15 439–442. 10.1016/s0168-9525(99)01856-9 [DOI] [PubMed] [Google Scholar]
  23. Graves A., Schmidhuber J. (2005). Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Net. 18 602–610. 10.1016/j.neunet.2005.06.042 [DOI] [PubMed] [Google Scholar]
  24. Groher A. C., Jager S., Schneider C., Groher F., Hamacher K., Suess B. (2019). Tuning the performance of synthetic riboswitches using machine learning. ACS Synth. Biol. 8 34–44. 10.1021/acssynbio.8b00207 [DOI] [PubMed] [Google Scholar]
  25. Guillén-Ramírez H. A., Martínez-Pérez I. M. (2018). Classification of riboswitch sequences using k-mer frequencies. Biosystems 174 63–76. 10.1016/j.biosystems.2018.09.001 [DOI] [PubMed] [Google Scholar]
  26. Havill J. T., Bhatiya C., Johnson S. M., Sheets J. D., Thompson J. S. (2014). A new approach for detecting riboswitches in DNA sequences. Bioinformatics 30 3012–3019. 10.1093/bioinformatics/btu479 [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Hochreiter S., Schmidhuber J. (1997). Long short-term memory. Neural Comput. 9 1735–1780. 10.1162/neco.1997.9.8.1735 [DOI] [PubMed] [Google Scholar]
  28. Kalvari I., Argasinska J., Quinones-Olvera N., Nawrocki E. P., Rivas E., Eddy S. R., et al. (2018). Rfam 13.0: shifting to a genome-centric resource for non-coding RNA families. Nucleic Acids Res. 46 D335–D342. 10.1093/nar/gkx1038 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Kelley D. R., Snoek J., Rinn J. (2016). Basset: learning the regulatory code of the accessible genome with deep convolutional neural networks. Genome Res. 26 990–999. 10.1101/gr.200535.115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Lang K., Rieder R., Micura R. (2007). Ligand-induced folding of the thiM TPP riboswitch investigated by a structure-based fluorescence spectroscopic approach. Nucleic Acids Res. 35 5370–5378. 10.1093/nar/gkm580 [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Lee B., Lee T., Na B., Yoon S. (2015). DNA-level splice junction prediction using deep recurrent neural networks. arXiv [Preprint]. Available online at: arXiv.org>cs>arXiv:1512.05135 (accessed August 25, 2019). [Google Scholar]
  32. Leonelli S. (2019). Philosophy of biology: the challenges of big data biology. eLife 8:e47381. 10.7554/eLife.47381 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Li C., Farkhoor H., Liu R., Yosinski J. (2018). Measuring the intrinsic dimension of objective landscapes. arXiv [Preprint]. Available online at: https://arxiv.org/pdf/1804.08838.pdf (accessed July 3, 2019). [Google Scholar]
  34. Li W., Godzik A. (2006). Cd-hit: a fast program for clustering and comparing large sets of protein or nucleotide sequences. Bioinformatics 22 1658–1659. 10.1093/bioinformatics/btl158 [DOI] [PubMed] [Google Scholar]
  35. Lipton Z. C. (2015). A critical review of recurrent neural networks for sequence learning. arXiv [Preprint]. Available online at: https://arxiv.org/abs/arXiv:1506.00019 (accessed May 29, 2015). [Google Scholar]
  36. Lo Bosco G., Di Gangi M. (2017). Deep learning architectures for DNA sequence classification. Lecture Notes Comput. Sci. 10147 162–171. 10.1007/978-3-319-52962-2_14 [DOI] [Google Scholar]
  37. Mandal M., Boese B., Barrick J. E., Winkler W. C., Breaker R. R. (2003). Riboswitches control fundamental biochemical pathways in Bacillus subtilis and other bacteria. Cell 113 577–586. 10.1016/s0092-8674(03)00391-x [DOI] [PubMed] [Google Scholar]
  38. Mandal M., Breaker R. R. (2004). Gene regulation by riboswitches. Nat. Rev. Mol. Cell Biol. 5 451–463. [DOI] [PubMed] [Google Scholar]
  39. Manning C. D., Raghavan P., Schütze H. (2008). Introduction to Information Retrieval. Cambridge: Cambridge University Press. [Google Scholar]
  40. McCown P. J., Corbino K. A., Stav S., Sherlock M. E., Breaker R. R. (2017). Riboswitch diversity and distribution. RNA 23 995–1011. 10.1261/rna.061234.117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Meyer A., Pellaux R., Potot S., Becker K., Hohmann H. P., Panke S., et al. (2015). Optimization of a whole-cell biocatalyst by employing genetically encoded product sensors inside nanolitre reactors. Nat. Chem. 7 673–678. 10.1038/nchem.2301 [DOI] [PubMed] [Google Scholar]
  42. Mironov A. S., Gusarov I., Rafikov R., Lopez L. E., Shatalin K., Kreneva R. A., et al. (2002). Sensing small molecules by nascent RNA: a mechanism to control transcription in bacteria. Cell 111 747–756. 10.1016/S0092-8674(02)01134-0 [DOI] [PubMed] [Google Scholar]
  43. Mukherjee S., Sengupta S. (2016). Riboswitch scanner: an efficient pHMM-based web-server to detect riboswitches in genomic sequences. Bioinformatics 32 776–778. 10.1093/bioinformatics/btv640 [DOI] [PubMed] [Google Scholar]
  44. Nahvi A. (2004). Coenzyme B12 riboswitches are widespread genetic control elements in prokaryotes. Nucleic Acids Res. 32 143–150. 10.1093/nar/gkh167 [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Nahvi A., Sudarsan N., Ebert M. S., Zou X., Brown K. L., Breaker R. R. (2002). Genetic control by a metabolite binding mRNA. Chem. Biol. 9 1043–1049. 10.1016/S1074-5521(02)00224-7 [DOI] [PubMed] [Google Scholar]
  46. Nawrocki E. P., Eddy S. R. (2013). Infernal 1.1: 100-fold faster RNA homology searches. Bioinformatics 29 2933–2935. 10.1093/bioinformatics/btt509 [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Pedregosa F., Varoquaux G., Gramfort A., Michel V., Thirion B., Grisel O., et al. (2011). Scikit-learn: machine learning in python. J. Mach. Learn. Res. 12 2825–2830. [Google Scholar]
  48. Penchovsky R., Pavlova N., Kaloudas D. (2020). RSwitch: a novel bioinformatics database on riboswitches as antibacterial drug targets. IEEE/ACM Trans. Comput. Biol. Bioinform. 99:1. 10.1109/TCBB.2020.2983922 [DOI] [PubMed] [Google Scholar]
  49. Premkumar K. A. R., Bharanikumar R., Palaniappan A. (2019). Riboflow: using deep learning to classify riboswitches with ∼99% accuracy. bioRxiv [Preprint]. 10.1101/868695 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Roth A., Breaker R. R. (2009). The structural and functional diversity of metabolite-binding riboswitches. Annu. Rev. Biochem. 78 305–334. 10.1146/annurev.biochem.78.070507.135656 [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Serganov A., Nudler E. (2013). A decade of riboswitches. Cell 152 17–24. 10.1016/j.cell.2012.12.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Singh J., Hanson J., Paliwal K., Zhou Y. (2019). RNA secondary structure prediction using an ensemble of two-dimensional deep neural networks and transfer learning. Nat. Commun. 10:5407. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Singh S., Singh R. (2016). Application of supervised machine learning algorithms for the classification of regulatory RNA riboswitches. Brief. Funct. Genomics 16 99–105. 10.1093/bfgp/elw005 [DOI] [PubMed] [Google Scholar]
  54. Sønderby S. K., Sønderby C. K., Nielsen H., Winther O. (2015). Convolutional LSTM networks for subcellular localization of proteins. ArXiv [Preprint]. 10.1007/978-3-319-21233-3_6 [DOI] [Google Scholar]
  55. Stormo G. D., Ji Y. (2001). Do mRNAs act as direct sensors of small molecules to control their expression? Proc. Natl. Acad. Sci. U.S.A. 98 9465–9467. 10.1073/pnas.181334498 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Strobel S. A., Cochrane J. C. (2007). RNA catalysis: ribozymes, ribosomes, and riboswitches. Curr. Opin. Chem. Biol. 11 636–643. 10.1016/j.cbpa.2007.09.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Sudarsan N., Barrick J. E., Breaker R. R. (2003). Metabolite-binding RNA domains are present in the genes of eukaryotes. RNA 9 644–647. 10.1261/rna.5090103 [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Sundermeyer M., Alkhouli T., Wuebker J., Ney H. (2014). “Translation modeling with bidirectional recurrent neural networks,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, 14–25. [Google Scholar]
  59. Tucker B. J., Breaker R. R. (2005). Riboswitches as versatile gene control elements. Curr. Opin. Struct. Biol. 15 342–348. 10.1016/j.sbi.2005.05.003 [DOI] [PubMed] [Google Scholar]
  60. van Rijsbergen C. J. (1975). Information Retrieval. London: Butterworths. [Google Scholar]
  61. Vaswani A., Jones L., Shazeer N., Parmar N., Gomez A. N., Uszkoreit J., et al. (2017). “Attention is all you need,” in Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA. [Google Scholar]
  62. Villa J. K., Su Y., Contreras L. M., Hammond M. C. (2018). Synthetic biology of small RNAs and riboswitches. Microbiol. Spectr. 6:10.1128/microbiolspec.RWR-0007-2017. 10.1128/microbiolspec.rwr-0007-2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Wang H., Mann P. A., Xiao L., Gill C., Galgoci A. M., Howe J. A., et al. (2017). Dual-targeting small-molecule inhibitors of the Staphylococcus aureus FMN riboswitch disrupt riboflavin homeostasis in an infectious setting. Cell Chem. Biol. 24 576–588. 10.1016/j.chembiol.2017.03.014 [DOI] [PubMed] [Google Scholar]
  64. Weiss K., Khoshgoftaar T. M., Wang D. (2016). A survey of transfer learning. J. Big Data 3:9. [Google Scholar]
  65. Wieland M., Hartig J. S. (2008). Artificial riboswitches: synthetic mRNA-based regulators of gene expression. Chembiochem 9 1873–1878. 10.1002/cbic.200800154 [DOI] [PubMed] [Google Scholar]
  66. Winkler W., Nahvi A., Breaker R. R. (2002). Thiamine derivatives bind messenger RNAs directly to regulate bacterial gene expression. Nature 419 952–956. 10.1038/nature01145 [DOI] [PubMed] [Google Scholar]
  67. Winkler W. C., Nahvi A., Roth A., Collins J. A., Breaker R. R. (2004). Control of gene expression by a natural metabolite-responsive ribozyme. Nature 428 281–286. [DOI] [PubMed] [Google Scholar]
  68. Wittmann A., Suess B. (2012). Engineered riboswitches: expanding researchers’ toolbox with synthetic RNA regulators. FEBS Lett. 586 2076–2083. 10.1016/j.febslet.2012.02.038 [DOI] [PubMed] [Google Scholar]
  69. Yang J., Seo S. W., Jang S., Shin S. I., Lim C. H., Roh T. Y., et al. (2013). Synthetic RNA devices to expedite the evolution of metabolite-producing microbes. Nat. Commun. 4:1413. 10.1038/ncomms2404 [DOI] [PubMed] [Google Scholar]
  70. Yang Y. (1999). An evaluation of statistical approaches to text categorization. J. Inf. Retr. 1 67–88. [Google Scholar]
  71. Yanofsky C. (1981). Attenuation in the control of expression of bacterial operons. Nature 289 751–758. 10.1038/289751a0 [DOI] [PubMed] [Google Scholar]
  72. Yu D., Breaker R. R. (2020). A bacterial riboswitch class senses xanthine and uric acid to regulate genes associated with purine oxidation. RNA 10.1261/rna.075218.120 [Epub ahead of print]. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Zhou J., Troyanskaya O. G. (2015). Predicting effects of noncoding variants with deep learning- based sequence model. Nat. Methods 12 931–934. 10.1038/nmeth.3547 [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Zhou L. B., Zeng A. P. (2015). Engineering a lysine-ON riboswitch for metabolic control of lysine production in Corynebacterium glutamicum. ACS Synth. Biol. 4 1335–1340. 10.1021/acssynbio.5b00075 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data Availability Statement

All datasets generated for this study are included in the article/Supplementary Material.


Articles from Frontiers in Bioengineering and Biotechnology are provided here courtesy of Frontiers Media SA

RESOURCES