Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2014 Feb 24;4(5):468–481. doi: 10.1002/wcms.1183

Machine learning methods in chemoinformatics

John B O Mitchell 1,*
PMCID: PMC4180928  EMSID: EMS57556  PMID: 25285160

Abstract

Machine learning algorithms are generally developed in computer science or adjacent disciplines and find their way into chemical modeling by a process of diffusion. Though particular machine learning methods are popular in chemoinformatics and quantitative structure–activity relationships (QSAR), many others exist in the technical literature. This discussion is methods-based and focused on some algorithms that chemoinformatics researchers frequently use. It makes no claim to be exhaustive. We concentrate on methods for supervised learning, predicting the unknown property values of a test set of instances, usually molecules, based on the known values for a training set. Particularly relevant approaches include Artificial Neural Networks, Random Forest, Support Vector Machine, k-Nearest Neighbors and naïve Bayes classifiers. WIREs Comput Mol Sci 2014, 4:468–481.

How to cite this article: WIREs Comput Mol Sci 2014, 4:468–481. doi:10.1002/wcms.1183

INTRODUCTION

The field known as chemoinformatics, or sometimes cheminformatics, can be considered as that part of computational chemistry whose models are not based on reproducing the real physics and chemistry by which the world works at the molecular scale. Unlike quantum chemistry or molecular simulation, which are designed to model physical reality, chemoinformatics is intended simply to produce useful models that can predict chemical and biological properties of compounds given the two-dimensional (or sometimes three, see Box 1 chemical structure of a molecule.

REPRESENTING MOLECULES: TWO OR THREE-DIMENSIONAL?

In chemoinformatics, the researcher is presented with a fundamental dilemma—should the molecules be described with two- or three-dimensional representations? A two-dimensional representation is essentially a molecular graph with the atoms as nodes and the bonds as edges. Onto this may be added extra information, such as bond orders, and the stereochemistry about double bonds and at chiral centers. Such a representation of chemical structure is essentially a digitized form of the structural diagrams familiar to chemists, and lacks the explicit spatial coordinates of the atoms.

An alternative approach is to generate a three-dimensional structure. This can be done from the molecular graph or connection table using a program such as CORINA,7,8 from a crystal structure, or from a quantum chemical calculation. Although a three-dimensional structure carries additional information, the difficulty is that molecules generally exist as an equilibrium between multiple conformers. Even if our structure correctly represents the lowest energy, and hence most abundant, conformer, alternative conformations may be critical for biological functions such as protein binding. Nonetheless, using three-dimensional structure opens up possibilities like scaffold hopping in drug design, where molecules with diverse two-dimensional structures but similar three-dimensional shapes may bind the same target. Sheridan and Kearsley's review9 and an article by the Ritchie group10 are two of the numerous papers discussing the relative merits of two and three-dimensional molecular representations.

The history of chemoinformatics began with local models, typically for quantitative structure–activity relationships (QSAR) or quantitative structure–property relationships (QSPR). Popular versions of this history usually begin with Hammett or Hansch,1,2 though Borman has followed the trail of QSAR back into the 19th century.3 Early models were generally based on linear, and later multilinear, regression. These were typically built using only a very few features, and were valid only for a small series of closely related compounds. Interestingly, machine learning and pattern recognition methods have an association with chemistry going back more than four decades, with methods like the linear learning machine being applied to problems such as the interpretation of spectroscopic data, as discussed in an early review by Kowalski.4

In contrast to the very small applicability domains of early QSAR studies, much recent work has concentrated on global models, by which we mean models trained on and hence valid for a wide range of organic or drug-like compounds. A number of factors, most notably the availability of data for molecules spanning a much wider chemical space, the use of a large and diverse selection of descriptors, and the development of sophisticated nonlinear machine-learning algorithms have increased the use of such global models in recent years.

CHEMOINFORMATICS

From Molecules to Features to Properties

Although to some extent a postrationalization, it is helpful to consider chemoinformatics model building as a two-part process.5 Firstly a molecular structure, typically represented as a molecular graph or connection table, is converted into a vector of features (which are also known as descriptors and represented generically by the symbol x). This first stage may sometimes also use three-dimensional information (see Box 1), and can be referred to as the encoding. Numerous recipes, some freely available and many commercial, exist for encoding a compound as a feature vector.6 The second part of building a machine-learning model for chemoinformatics is the mapping (using Lusci et al.'s terminology4). This involves empirically discovering a function that maps between the feature vectors and the property of interest, represented by the symbol y. It is this mapping that is most often learnt by the machine-learning algorithm. The two-part process is illustrated in Figure 1.

Figure 1.

Figure 1

We can conceive of chemoinformatics as a two-part problem: encoding chemical structure as features, and mapping the features to the output property. The second of these is most often the province of machine learning.

Feature Selection

The descriptors chosen to represent the molecules have no a priori reason either all to be relevant for describing the property to be predicted or all to be independent of one another. Some machine-learning algorithms are robust against the inclusion of irrelevant or of mutually correlated features, others less so. It is fairly common, as part of the training phase of the algorithm, to choose a subset of the original features that are helpful for building a predictive model and not strongly correlated with one another.1113 An alternative approach is principal component analysis (PCA),14 a statistical procedure that transforms mutually correlated variables, which should first be scaled, into mutually uncorrelated combinations called principal components. The process maximizes the variance of the first principal component; the components are ordered from first downwards by decreasing variance. PCA provides a way of explaining most of the variance in the output variable with a small number of orthogonal components. Each principal component is a linear combination of the original features, so although the linear combinations do not have such clear meanings as the original features, the user can at least see which features are contributing to the model.

Similar Property Principle

Two compounds that are ‘similar’ to one another will have feature vectors that, when considered as position vectors in the chemical space spanned by the descriptors, are close to each other. If the mapping function varies reasonably slowly and smoothly across chemical space, then we expect similar molecules to have similar values of the relevant chemical (or biological) property. This is the basis of the similar property principle: ‘Similar molecules have similar properties’. While this may be considered a central principle of chemoinformatics, it is far from universally valid. In the case of ‘activity cliffs’ for instance, the mapping function varies dramatically over a small distance in chemical space, corresponding perhaps to a change of one functional group which might prevent a ligand from binding effectively to a protein, and apparently similar molecules can have very different bioactivities.15

What Properties Can We Model?

For many properties, especially those of isolated molecules, chemoinformatics and machine learning would be a poor choice of methodology. If we want to calculate dipole moments, polarizabilities or vibrational frequencies, we would be better off using quantum chemistry. On the other hand, where a complex biological or condensed phase system cannot easily be directly modeled by physics-based methods, then chemoinformatics becomes a sensible option. Given the extent to which the development of chemoinformatics has been intertwined with drug discovery and the pharmaceutical industry, it is hardly surprising that bioactivity and ADMET (Absorption, Distribution, Metabolism, Excretion and Toxicity) properties rank high on the list of those addressed by informatics approaches. There are also numerous physicochemical properties that are hard to obtain from theoretical chemical methods such as density functional theory or molecular dynamics, and hence are often modeled by chemoinformatics. Amongst such properties, aqueous solubility and logP (the logarithm of the octanol:water partition coefficient) are directly relevant to drug discovery, and melting point is indirectly so, due to its correlation with solubility.16 Properties such as solubility17,18 and sublimation energy1921 are potentially amenable to modeling by either chemoinformatics or theoretical chemistry approaches.

MACHINE-LEARNING METHODS

Artificial Neural Networks

An artificial neural network (ANN), often simply called a neural network where confusion with biology is unlikely, is a mathematical model used for pattern recognition and machine learning. The network's architecture is based on connected neurons in an input layer, a hidden layer or layers, and an output layer. In a typical design, each connection between neurons carries a weight. The weights are varied during the training phase as the network learns how to connect input and output data, before being tested on unseen instances. While the ANN is inspired by the structure and function of the human brain, it is massively simpler in design and in no way simulates higher brain function. In fact, a typical ANN will be smaller than the minimal 302 neuron brain of the nematode Caenorhabditis elegans.22

ANNs have been used for a wide range of chemoinformatics applications. Amongst studies seeking to predict bioactivity are Li et al.'s work on estrogen receptor agonists,23 and So et al.'s study of steroids,24 while neural networks were amongst a number of methods used by Briem et al. to identify possible kinase inhibitors.25 As with other machine-learning methods, ANNs are often used to predict toxicological, pharmacological, and physicochemical properties, such as hERG blockade,26 aquatic toxicity,27 drug clearance,28 pKa,29 melting point,30,31 and solubility.17,32,33 However, ANNs suffer from vulnerability to overfitting, with a danger of learning the noise as well as the signal from the training set and hence being less able to generalize to the unseen test data.34 Addressing this requires careful study design, so that the training process can be stopped close to the optimal time.

Deep Learning

Deep learning, a concept closely associated with ANNs, is in principle the learning of layered concepts. Thus, a model could describe higher and lower-level concepts at different layers of its structure. Lusci et al.5 use a multilayer ANN, which they describe as a deep learning technique, to go from molecular graphs to a set of descriptors and then, via a suitable output function, to predict aqueous solubility, as shown in Figure 1. Overall, they are able to generate good predictions of solubility, competitive with other sophisticated machine-learning methods. Their model is in fact an ensemble of 20 different ANNs, each with slightly different architectures.

The Wisdom of Crowds and Ensembles of Predictors

The wisdom of crowds is a well-known expression of the benefit of utilizing multiple independent predictors.35 For example, a fairground competition might involve members of the public guessing the weight of a cow. No doubt, some guesses would be far too high and others much too low. However, the wisdom of crowds concept holds that the ensemble of estimates is capable of making an accurate prediction, as observed by Galton more than a century ago.36 In order to avoid the excessive effect of one or two absurd guesses on the mean, it is preferable to use the median of the estimates as the best prediction in the context of a public guessing game.

This idea was exploited by Bhat et al.,29 who used an ensemble of neural networks to predict the melting points of organic compounds. Each network has a different, randomly assigned, set of initial weights. The authors found a significant improvement in prediction accuracy as a result of using the ensemble approach, with the ensemble prediction being better not just than that of a typical single network, but better than the best performing single ANN. They achieved substantial improvement upon adding the first few additional ANNs, but there was little further effect in going beyond a few tens of networks. The authors chose to use 50 ANNs in their final model, though the rapid training time (around 9 seconds per network) meant that they could have afforded more if required. While Bhat et al. adapted an ANN approach to benefit from using an ensemble of predictors, we will see that the use of multiple independent models is also fundamental to Random Forest (RF). The key to benefitting from the wisdom of crowds is to design an algorithm that can produce multiple independent predictors, even though their predictions must be based on essentially the same pool of data. Interestingly, the idea of combining predictors of different kinds into an ensemble has been much less explored in chemoinformatics than in postdock scoring, where consensus scores constitute a well-established method.37

Random Forest

RF38,39 is a technique for classification based on an ensemble, or forest, of decision trees. The large number of independent trees allows RF to benefit from the wisdom of crowds effect. The trees are built using training data consisting of multiple features for each of a training set of objects. As we are discussing chemical applications, we will assume that these objects or instances are molecules. Each tree is generated by stochastic recursive partitioning and is randomized in two ways. Firstly, the tree is randomized by allowing it to use, at each node, only a stochastically chosen subset of the features. As the training instances progress through the tree, they are partitioned into increasingly homogeneous groups, so that each terminal node of the decision tree is associated with a group of molecules with similar values of the property to be predicted. Each split within a tree is created based on the best partitioning that is possible according to the Gini criterion,40 using any single-valued attribute from a randomly chosen subset of descriptors. The number of descriptors in this random subset is the parameter mtry, and the subset is freshly chosen for each node. The tree building continues until all training instances have been assigned to a terminal leaf node, see Figure 2.

Figure 2.

Figure 2

Five illustrative decision trees forming a (very small) Random Forest for classification. The terminal leaf nodes are shown as squares and colored red or green according to class. The path taken through each tree by a query instance is shown in orange. Trees A, B, C, and E predict that the instance belongs to the red class, tree D dissenting, so that the Random Forest will assign it to the red class by a 4–1 majority vote.

Secondly, each tree is randomized by basing it on a bootstrap sample of the training data. From a pool consisting of N distinct objects, a sample of N objects is chosen with replacement, so that each object may be chosen zero, one, two, or occasionally more times. The probability of a given molecule not being chosen for a given tree's bootstrap sample is (1 – 1/N)N, which tends to a limit of 1/e, or approximately 0.37, as N becomes large. Thus, for each tree, approximately 37% of the training set molecules do not appear in that tree's bootstrap sample, and constitute the so-called out-of-bag data; conversely, every molecule is out-of-bag for about 37% of the trees. The out-of-bag sample can be used as an internal validation set for each tree; the performance on the different out-of-bag samples of each tree provides a fair test of the predictivity of the RF.

The RF consists of ntree stochastically different trees, each built from its own bootstrap sample of the training data. The trained forest is then used to predict unseen test data. In the case of predicting a binary or a multiclass categorical variable, the classifications are determined by majority vote amongst the trees. The proportion of votes cast for a class may provide an indication of the probability of a label being correctly assigned, or of confidence in a prediction, but this should be considered an informal estimate only.41 Similarly, a RF of regression trees can be used for predicting numerical quantities, with the predictions from the different trees being averaged. For classification, the default value of mtry recommended by Svetnik et al.39 is the square root of the total number of descriptors; for regression, they advise using mtry equal to one third of the number of descriptors. If mtry were increased to equal the full number of descriptors, RF would become equivalent to another machine-learning method, the bootstrap aggregating technique known as bagging. Alternatively, mtry could be treated as an optimizable parameter. A typical default value of ntree is 500, though there is a case for using an odd number of trees in binary prediction to avoid ties (which would be resolved randomly). Often, RF calculations are relatively cheap and a larger number of trees could be afforded; however, the improvement in prediction accuracy with additional trees is small. RF is generally considered relatively robust against overfitting. Svetnik et al. demonstrate that, unlike ANN, the test set error of RF does not increase but converges to a limiting asymptotic value as the training error is reduced toward zero, one of many interesting observations contained in an excellent paper that describes RF in full detail.39

RF has proven a very successful method in chemoinformatics and has been used in many different contexts. These include prediction of athletic performance enhancement,42 QSAR,4345 mutagenicity,46 phospholipidosis,47 hERG blockade,48 and skin sensitization.49 Applications in physicochemical properties include discovery of new crystalline solvates50 and solubility,51 the prediction of which has also been systematically compared with those of melting point and logP.52 RF has also found applications in the area of postdock scoring functions and predicting protein–ligand binding affinity.5356

Support Vector Machine

Support Vector Machine (SVM)57 maps the data into a high-dimensional space, using a kernel function that is typically nonlinear. The SVM seeks to find an optimal separation between two classes, such that each in their entirety lie on opposite sides of a separating hyperplane. This is achieved by maximizing the margin between the closest points, known as support vectors, and the hyperplane. SVM can be adapted to either multiclass classification58,59 or to regression. SVMs are one of the most popular machine-learning methods in chemoinformatics. Uses in bioactivity prediction include drug repurposing,60,61 kinase inhibition,25 estrogen receptor agonists,23 and opioid activity.62 The SVM is often used to predict toxicity-related properties such as hERG blockade,47,63,64 mutagenic toxicity,65 toxicity classification,66 and phospholipidosis.47,67 Applications in physicochemical property prediction include solubility,33,52,68 pKa,29 logP, and melting point.52 The interested reader is referred to Noble's instructive article for further discussion of SVMs.57

k-Nearest Neighbors

The k-nearest neighbors (kNN) algorithm is one of the simplest machine-learning methods to understand and explain, the principle being that an instance is classified by a majority vote of its neighbors, see Figure 3. Each test instance is predicted to belong to the class most commonly found amongst its k closest neighbors, where k is a positive integer. Most often, k is chosen to be small; if k = 1, the instance is simply assigned to the same class as its nearest neighbor in a feature space. The instances, which in chemical applications are typically molecules, are described as position vectors in the feature space, which is usually of high dimensionality. It is helpful to scale the features so that distances measured in different directions in the space are comparable. Neighbors are identified on the basis of distance in the feature space. This is usually taken to be the Euclidean distance, though other metrics such as the Jaccard distance could be used. In binary classification problems, it is helpful to choose k to be an odd number as this reduces the risk of tied votes, though depending on the granularity of the space multiple neighbors may share the same distance. Once the labeled instances and their positions in the feature space are available, no explicit training phase is required. Because of this, kNN is considered a ‘lazy learning’ algorithm.

Figure 3.

Figure 3

Illustration of a kNN classification model. For k = 1, the model will classify the blue query instance as a member of the red class; for k = 3, it will again be assigned to the red class, this time by a 2–1 vote; however, since the fourth and fifth nearest neighbors are both green, a k = 5 model would classify it as part of the green class by a 3–2 majority.

The same method can be used for regression. This is most simply achieved by assigning the property for the test instance to take the mean value calculated from its k closest neighbors. However, it can be helpful to weight the contributions of the neighbors, the closest neighbor contributing most to the average and the kth neighbor contributing least; a procedure for doing this was published by Nigsch et al.69 This effectively smooths the transition between neighboring instances being counted and non-neighboring instances being ignored.

The kNN algorithm is sensitive to the local structure of the data. Thus it is ideal for calculating properties with strong locality, as is the case with protein function prediction.70 If a single neighbor with, say, 90% identity to the test sequence is found, it is highly likely that the functional label can be safely transferred to the query sequence; a dozen neighbors each with 25% identity would be much less useful. Despite its locality, kNN can still give global coverage, so long as the instances are distributed throughout the feature space. In principle, the distance to the neighbors, or the proportions of neighbors having a given label, could be used to measure prediction confidence, though this is rarely done in practice.

Many studies use some kind of internal validation to optimize the value of k, with the optimum value dependent upon the dataset at hand. kNN has been used in bioactivity studies of anticonvulsants and dopamine D1 agonists71 of kinase inhibition,25 cannabinoid psychoactivity,72 steroids, anti-inflammatories and anti-cancer drugs,73 athletic performance enhancement,42 and estrogen receptor agonists.23 Studies of toxicological and pharmacological relevance have looked at drug clearance,28 mutagenic potency,74 and percutaneous drug absorption.75 kNN has been used to predict the odor characteristics of compounds.76,77 Studies using kNN to investigate physicochemical properties have considered melting point,52,69 boiling point,78 logP,52,78 aqueous solubility,52,79 and the analysis of mixtures.80

Naïve Bayes

The naïve Bayes classifier provides a rather different kind of algorithm, one based on estimating the probabilities of class membership. Application of Bayes' theorem, together with the assumption that the features xi are conditionally independent of one another given the output class y, leads to the formula

graphic file with name wcms0004-0468-m1.jpg

The decision rule is to assign a test instance to the class with the highest estimated probability. Undoubtedly, the assumption of conditional independence of the features is not strictly valid. Nonetheless, naïve Bayes often performs well enough to be competitive with other machine-learning methods, and has the advantage of conceptual simplicity compared to most. As early as 1974, Cramer et al. published a method of computing the conditional probability of a molecule being bioactive given the fragments it contained.81 Naïve Bayes classifiers are now frequently used in chemoinformatics, usually for predicting biological rather than physicochemical properties, naïve Bayes often being used alongside and compared against other classifiers. This has been done in studies of athletic performance enhancement,42 toxicity,66 the mechanism of phospholipidosis,82 and also for protein target prediction and bioactivity classification for drug-like molecules.8385 It is in principle possible to use naïve Bayes for regression,86 but this is rarely seen in chemoinformatics (Table1).

Table 1.

Some Other Machine-Learning Methods Used in Chemoinformatics

Algorithm Description
Ant Colony87 Uses virtual pheromones based on ant behavior for optimization
Relevance Vector Machine (RVM)88 Sparse probabilistic binary classifier related to SVM; gives probabilities rather than all-or-nothing classification
Parzen-Rosenblatt Window82,83,89 Kernel density estimation method that allows molecular similarities to be transformed into probabilities of class membership
Fuzzy Logic90 Designed to give interpretable rules based on descriptor values
Rough Sets91 Rule-based method designed to give interpretable rules
Support Vector Inductive Logic Programming (SVILP)84 Rule-based method incorporating SVM ideas
Winnow47,85,92,93 For every class, Winnow learns a vector of weights for each feature. Test instances are compared with these using score thresholds
Decision Tree23,76,94,95 Like one tree from a Random Forest, but without randomization
Linear Discriminant Analysis (LDA)96,97 Models statistical differences between classes in order to make a classification
kScore98 Analogous to a weighted kNN scheme in which the weights are optimized by Leave-One-Out cross-validation
Projection to Latent Structures (PLS)29,52,68 Obtains a linear regression by projecting x and y variables to a new space. Also called Partial Least Squares

VALIDATION

Study Design

Test Sets

Chemoinformatics models are only useful if they are predictive. It is not sufficient simply to fit known data, a useful model must be able to generalize to unknown data, and thus must be validated.99 The traditional way of doing this is to have the total dataset divided into two parts, the training set and the test set. The training set is used to build the model, and its property values are known to the algorithm. The model is then tested on the test set, whose property values are not given to the machine-learning algorithm. However, many machine-learning approaches produce models sufficiently complex to contain internal variable parameters. Often it is helpful to optimize these parameters by holding back part of the training set as an internal validation set, which is used to find the parameter values giving the best predictivity.

This approach to validation is a good one, though it requires that the test set falls within the applicability domain of the model.100 This means that the training and test data should span the same region of chemical space. A test instance is unlikely to be predictable if there is no instance like it in the training set. ‘Real life’ usage of QSAR and other chemoinformatics models may well involve training and model building at one time, followed by testing on newly available data at some later date. In such cases, it is important to ensure that the new test data are within the applicability domain of the model.

Cross-Validation

One common and effective approach is cross-validation. In n-fold cross-validation, the data are distributed, either randomly or in a stratified way, into n separate folds, with one fold being the initial test set. If relevant, a second fold is used for internal validation. The remaining folds are the initial training set. The identities of the folds are then cyclically permuted, such that every fold is the test fold once—and hence each instance is predicted exactly once,99 see Figure 4. Thus n separate models are generated, and critically each instance is predicted from a model built without knowledge of that instance's property value. This requirement means that any feature selection needs to be carried out separately for each of the n models, ensuring that no information about the test instances finds its way directly or indirectly into the model building process. Typically, fivefold or 10-fold cross validation will be carried out. It is also fairly common to use Leave-One-Out (LOO) cross-validation, in which a separate model is built to predict each instance, trained on the remaining (n − 1) instances. An alternative is to use Monte Carlo cross-validation,68,69,101 in which a number of different training-test set splits are chosen randomly. Bootstrap resampling99 is a related approach to randomizing datasets for cross-validation, which is similar to assessing RF models by their predictivity for out-of-bag data,39 as discussed above. Cross-validation, Monte-Carlo or otherwise, has the advantage that it can in most study designs be repeated many times with randomly different fold definitions, with the results being averaged, leading to a more robust conclusion.

Figure 4.

Figure 4

Design of a cross-validation exercise, here shown for eight-fold cross-validation. The identities of the six training, one test, and one internal validation folds are cyclically permuted.

y-Randomization

One powerful test of a machine-learning model is y-randomization, also known as y-scrambling.99,102 The real model is compared with alternative models, which are generated from datasets in which the property values y are repeatedly randomly reassigned amongst the instances. The process of randomization breaks the true chemical link between the features x and the output property y, so that there is no meaningful signal left to model. If the machine learning method is still able to produce good validation statistics for the randomized models then we should be highly suspicious, as we know that it must be modeling noise rather than signal.

Measuring Success

Measuring the success of a classification model is not as straightforward as it might initially seem. For a binary classification exercise, predictions can be classed as true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN). These are combined into a confusion matrix of actual against predicted classes, the diagonal elements being the numbers of TP and TN, and the off-diagonal ones the numbers of FP and FN. For a multi-class classification problem, the confusion matrix is analogous to this, with correct predictions again being on the diagonal and incorrect ones off-diagonal, but it has higher dimensionality. There are numerous recipes for extracting single valued measures of prediction success from the multiple numbers in a confusion matrix, as discussed in several excellent articles.103105 For regression models, the square of the Pearson correlation coefficient R2, together with its cross-validated counterpart Q2, are often used, as discussed by Consonni et al.106 Most often, the root mean squared error (RMSE) is used as the numerical measure of the prediction accuracy of a regression model, as it naturally accounts for errors of either sign. The average absolute error (AAE) is sometimes used instead and often gives substantially lower numerical values by avoiding the quadratic contributions from poorly modeled instances, as can be seen from papers such as16 which tabulate both measures. It is also possible to assess the number of ‘correct’ predictions if an arbitrary threshold is defined; for example, the Solubility Challenge107 defined any prediction of logS with an absolute error within 0.5 log10 units as successful.

Interpretability

When we have identified suitable descriptors, built a chemoinformatics predictor from them, and assessed its predictive accuracy, how far can we interpret the resulting model? The model tells us that descriptor x is correlated with property y. For instance, Ploemen et al.108 found that induction of phospholipidosis is correlated with a molecule's acid dissociation constant pKa. Knowing that x can predict y does not tell us how or why molecules with feature x exhibit property y. There are probably several possible explanations. Maybe induction of phospholipidosis involves acid–base chemistry. Maybe it involves molecules obtaining a particular charge state, perhaps to help dissolve in an aqueous phase, or perhaps to bind to some receptor, or possibly to sequester ligands of the opposite charge? The correlation revealed by the model doesn't on its own prove one particular mechanistic hypothesis. This is what we mean when we say that QSAR, QSPR, and other chemoinformatics models ‘reveal correlation, not causation’.

So is a QSAR model useless for understanding what is going on? Clearly not. The absence of an expected correlation may allow us to reject a mechanistic hypothesis, while the presence of unanticipated correlations can suggest new hypotheses. These will need proper testing, either by direct experiment or by more sophisticated computational studies leveraging existing mechanism-relevant experimental data, as performed by Lowe et al. for phospholipidosis.82 So a QSAR can be a step along the road to a mechanistic understanding of a biological or chemical phenomenon, but never the final step.

Chemoinformatics models are sometimes described as ‘black boxes’. The archetypal black box would be a (virtual) machine that predicted properties excellently, but that offered no clue as to how or why they occurred. In practice, a chemoinformatics model is unlikely ever to be a completely black box. Some methods, such as multilinear regression, immediately tell us what descriptors contribute to the model; RF, at least in its implementation in R,109 allows this information to be extracted very easily via calculation of descriptor importance. Even for methods without an inherent simple measure of importance, we could (though possibly at some computational cost) build a set of models in which each descriptor is successively randomized, just as is done when a RF is ‘noised up’.39 Such a procedure will work best, maximizing the effect of descriptor randomization and minimizing the number of models to be built, if we first remove correlated descriptors. The fundamental point is that for any machine-learning model, even a neural network, we can examine the effect that removing input information has on the final model. Thus, we can measure and extract the importance values of descriptors from any model if we are prepared to try hard enough, as Carlsson et al. demonstrated in two cleverly designed studies that allowed them to extract chemical substructure contributions to toxicity and bioactivity from SVM and RF models.110,111

Conclusion

Although numerous articles cited herein have compared performances of the various machine-learning algorithms used in chemoinformatics, there is no single best method for all problems. The relative abilities of methods will depend on the size and distribution in chemical space of the dataset, the linearity or otherwise of the chemical problem to be solved, the nature and internal correlation of the descriptor set available, and the relevance of nonlocal data, amongst several other factors. For linear problems, simple linear regression may be as effective as complex machine learning algorithms.18 For nonlocal problems, SVM and RF are probably at least as good as any other algorithm and often perform similarly when compared.47,52 The frequency with which these algorithms have been used and discussed in chemoinformatics, together with their easy availability via platforms such as R,109 makes SVM and RF good starting points for chemists beginning to incorporate machine learning into their work. For problems where only local data are likely to be relevant, kNN is an excellent and simple approach.69,70

Validation is seen to be a critical part of any machine-learning project and the design of the in silico experiment is crucial to the robustness of the study. A traditional training-test split of the data is a good validation strategy, provided that the two sets span the same regions of chemical space. Cross-validation is also a popular strategy, and still allows models to be tested on data unseen in their generation. y-randomization provides a useful test of a chemoinformatics model; a predictor that can find a signal in random data is not to be trusted. Various different metrics are used for measuring success, with RMSE and R2 typically being used for regression studies. Several different metrics for binary and multi-class classification exist, all being derived from the confusion matrix. QSAR and QSPR models reveal correlation between descriptors and properties, but do not by themselves prove mechanistic hypotheses.

Acknowledgments

J. B. O. M. is grateful to the Scottish Universities Life Sciences Alliance (SULSA) for financial support and he thanks the BBSRC for funding his research through grant BB/I00596X/1. The author thanks Dr Luna De Ferrari (University of St Andrews) and Dr Lazaros Mavridis (Queen Mary University of London) for their helpful comments on the manuscript.

REFERENCES

  • 1.Hammett LP. Reaction rates and indicator acidities. Chem Rev. 1935;16:67–79. [Google Scholar]
  • 2.Hansch C, Fujita T. p-σ-π Analysis. A method for the correlation of biological activity and chemical structure. J Am Chem Soc. 1964;86:1616–1626. [Google Scholar]
  • 3.Borman S. New QSAR techniques eyed for environmental assessments. Chem Eng News. 1990;19:20–23. [Google Scholar]
  • 4.Kowalski BR. Pattern recognition in chemical research. In: Klopfenstein CE, Wilkins CL, editors. Computers in Chemical and Biochemical Research. Vol. 2. Academic Press: New York; 1974. pp. 1–76. [Google Scholar]
  • 5.Lusci A, Pollastri G, Baldi P. Deep architectures and deep learning in chemoinformatics: the prediction of aqueous solubility for drug-like molecules. J Chem Inf Model. 2013;53:1563–1575. doi: 10.1021/ci400187y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Leach AR, Gillet VJ. An Introduction to Chemoinformatics. Dordrecht, The Netherlands: Springer; 2007. [Google Scholar]
  • 7.Gasteiger J, Rudolph C, Sadowski J. Automatic generation of 3D-atomic coordinates for organic molecules. Tetrahedron Comput Methodol. 1990;3:537–547. [Google Scholar]
  • 8.Sadowski J, Gasteiger J, Klebe G. Comparison of automatic three-dimensional model builders using 639 X-ray structures. J Chem Inf Comput Sci. 1994;34:1000–1008. [Google Scholar]
  • 9.Sheridan RP, Kearsley SK. Why do we need so many chemical similarity search methods? Drug Discov Today. 2002;7:903–911. doi: 10.1016/s1359-6446(02)02411-x. [DOI] [PubMed] [Google Scholar]
  • 10.Venkatraman V, Perez-Nueno VI, Mavridis L, Ritchie DW. Comprehensive comparison of ligand-based virtual screening tools against the DUD data set reveals limitations of current 3D methods. J Chem Inf Model. 2010;50:2079–2093. doi: 10.1021/ci100263p. [DOI] [PubMed] [Google Scholar]
  • 11.Smialowski P, Frishman D, Kramer S. Pitfalls of supervised feature selection. Bioinformatics. 2010;26:440–443. doi: 10.1093/bioinformatics/btp621. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Zheng W, Tropsha A. Novel variable selection quantitative structure-property relationship approach based on the k-nearest-neighbor principle. J Chem Inf Model. 2000;40:185–194. doi: 10.1021/ci980033m. [DOI] [PubMed] [Google Scholar]
  • 13.O'Boyle NM, Palmer DS, Nigsch F, Mitchell JBO. Simultaneous feature selection and parameter optimisation using an artificial ant colony: case study of melting point prediction. Chem Cent J. 2008;2:21. doi: 10.1186/1752-153X-2-21. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Ringner M. What is principal component analysis? Nat Biotech. 2008;26:303–304. doi: 10.1038/nbt0308-303. [DOI] [PubMed] [Google Scholar]
  • 15.Maggiora GM. On outliers and activity cliffs—why QSAR often disappoints. J Chem Inf Model. 2006;46:1535. doi: 10.1021/ci060117s. [DOI] [PubMed] [Google Scholar]
  • 16.Ran Y, Yalkowsky SH. Prediction of drug solubility by the general solubility equation (GSE) J Chem Inf Comput Sci. 2001;41:354–357. doi: 10.1021/ci000338c. [DOI] [PubMed] [Google Scholar]
  • 17.Palmer DS, Llinas A, Morao I, Day GM, Goodman JM, Glen RC, Mitchell JBO. Predicting intrinsic aqueous solubility by a thermodynamic cycle. Mol Pharm. 2008;5:266–279. doi: 10.1021/mp7000878. [DOI] [PubMed] [Google Scholar]
  • 18.Hewitt M, Cronin MTD, Enoch SJ, Madden JC, Roberts DW, Dearden JC. In silico prediction of aqueous solubility: the solubility challenge. J Chem Inf Model. 2009;49:2572–2587. doi: 10.1021/ci900286s. [DOI] [PubMed] [Google Scholar]
  • 19.Charlton MH, Docherty R, Hutchings MG. Quantitative structure-sublimation enthalpy relationship studied by neural networks, theoretical crystal packing calculations and nonlinear regression analysis. J Chem Soc Perkin Trans. 1995;2:2023–2030. [Google Scholar]
  • 20.Ouvrard C, Mitchell JBO. Can we predict lattice energy from molecular structure? Acta Cryst B. 2003;59:676–685. doi: 10.1107/s0108768103019025. [DOI] [PubMed] [Google Scholar]
  • 21.Salahinejad M, Le TC, Winkler DA. Capturing the crystal: prediction of enthalpy of sublimation, crystal lattice energy, and melting points of organic compounds. J Chem Inf Model. 2013;53:223–229. doi: 10.1021/ci3005012. [DOI] [PubMed] [Google Scholar]
  • 22.Connors BW, Long MA. Electrical synapses in the mammalian brain. Annu Rev Neurosci. 2004;27:393–418. doi: 10.1146/annurev.neuro.26.041002.131128. [DOI] [PubMed] [Google Scholar]
  • 23.Li H, Ung C, Yap C, Xue Y, Li Z, Chen Y. Prediction of estrogen receptor agonists and characterization of associated molecular descriptors by statistical learning methods. J Mol Graph Model. 2006;25:313–323. doi: 10.1016/j.jmgm.2006.01.007. [DOI] [PubMed] [Google Scholar]
  • 24.So SS, Karplus M. Three-dimensional quantitative structure-activity relationships from molecular similarity matrices and genetic neural networks. 1. Method and validations. J Med Chem. 1997;40:4347–4359. doi: 10.1021/jm970487v. [DOI] [PubMed] [Google Scholar]
  • 25.Briem H, Günther J. Classifying “kinase inhibitor-likeness” by using machine-learning methods. ChemBioChem Eur J Chem Biol. 2005;6:558–566. doi: 10.1002/cbic.200400109. [DOI] [PubMed] [Google Scholar]
  • 26.Thai KMM, Ecker GF. Classification models for hERG inhibitors by counter-propagation neural networks. Chem Biol Drug Des. 2008;72:279–289. doi: 10.1111/j.1747-0285.2008.00705.x. [DOI] [PubMed] [Google Scholar]
  • 27.Basak SC, Grunwald GD, Gute BD, Balasubramanian K, Opitz D. Use of statistical and neural net approaches in predicting toxicity of chemicals. J Chem Inf Comput Sci. 2000;40:885–890. doi: 10.1021/ci9901136. [DOI] [PubMed] [Google Scholar]
  • 28.Yap CW, Li ZR, Chen YZ. Quantitative structure-pharmacokinetic relationships for drug clearance by using statistical learning methods. J Mol Graph Model. 2006;24:383–395. doi: 10.1016/j.jmgm.2005.10.004. [DOI] [PubMed] [Google Scholar]
  • 29.Harding AP, Wedge DC, Popelier PLA. pKa Prediction from quantum chemical topology descriptors. J Chem Inf Model. 2009;49:1914–1924. doi: 10.1021/ci900172h. [DOI] [PubMed] [Google Scholar]
  • 30.Bhat AU, Merchant SS, Bhagwat SS. Prediction of melting points of organic compounds using extreme learning machines. Ind Eng Chem Res. 2008;47:920–925. [Google Scholar]
  • 31.Godavarthy SS, Robinson RL, Gasem KAM. An improved structure-property model for predicting melting-point temperatures. Ind Eng Chem Res. 2006;45:5117–5126. [Google Scholar]
  • 32.Erić S, Kalinić M, Popović A, Zloh M, Kuzmanovski I. Prediction of aqueous solubility of drug-like molecules using a novel algorithm for automatic adjustment of relative importance of descriptors implemented in counter-propagation artificial neural networks. Int J Pharm. 2012;437:232–241. doi: 10.1016/j.ijpharm.2012.08.022. [DOI] [PubMed] [Google Scholar]
  • 33.Louis B, Agrawal VK, Khadikar PV. Prediction of intrinsic solubility of generic drugs using MLR, ANN and SVM analyses. Eur J Med Chem. 2010;45:4018–4025. doi: 10.1016/j.ejmech.2010.05.059. [DOI] [PubMed] [Google Scholar]
  • 34.Hinton GE, Srivastava N, Krizhevsky A, Sutskever I, Salakhutdinov RR. Improving neural networks by preventing co-adaptation of feature detectors. 2012. Available at: http://arxiv.org/abs/1207.0580.
  • 35.Surowiecki J. The Wisdom of Crowds: Why the Many Are Smarter than the Few and How Collective Wisdom Shapes Business, Economies, Societies, and Nations. New York: Doubleday; 2004. [Google Scholar]
  • 36.Galton F. Vox populi. Nature. 1907;75:450–451. [Google Scholar]
  • 37.Charifson PS, Corkery JJ, Murcko MA, Walters WP. Consensus scoring: a method for obtaining improved hit rates from docking databases of three-dimensional structures into proteins. J Med Chem. 1999;42:5100–5109. doi: 10.1021/jm990352k. [DOI] [PubMed] [Google Scholar]
  • 38.Breiman L. Random Forests. Mach Learn. 2001;45:5–32. [Google Scholar]
  • 39.Svetnik V, Liaw A, Tong C, Culberson JC, Sheridan RP, Feuston BP. Random forest: a classification and regression tool for compound classification and QSAR modeling. J Chem Inf Comput Sci. 2003;43:1947–1958. doi: 10.1021/ci034160g. [DOI] [PubMed] [Google Scholar]
  • 40.Raileanu LE, Stoffel K. Theoretical comparison between the Gini Index and Information Gain criteria. Ann Math Artif Intell. 2004;41:77–93. [Google Scholar]
  • 41.Cannon EO, Nigsch F, Mitchell JBO. A novel hybrid ultrafast shape descriptor method for use in virtual screening. Chem Cent J. 2008;2:3. doi: 10.1186/1752-153X-2-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Cannon EO, Bender A, Palmer DS, Mitchell JBO. Chemoinformatics-based classification of prohibited substances employed for doping in sport. J Chem Inf Model. 2006;46:2369–2380. doi: 10.1021/ci0601160. [DOI] [PubMed] [Google Scholar]
  • 43.Kuz'min VE, Polishchuk PG, Artemenko AG, Andronati SA. Interpretation of QSAR models based on random forest methods. Mol Inf. 2011;30:593–603. doi: 10.1002/minf.201000173. [DOI] [PubMed] [Google Scholar]
  • 44.Bruce CL, Melville JL, Pickett SD, Hirst JD. Contemporary QSAR classifiers compared. J Chem Inf Model. 2007;47:219–227. doi: 10.1021/ci600332j. [DOI] [PubMed] [Google Scholar]
  • 45.Fukunishi H, Teramoto R, Shimada J. Hidden active information in a random compound library: extraction using a pseudo-structure—activity relationship model. J Chem Inf Model. 2008;48:575–582. doi: 10.1021/ci7003384. [DOI] [PubMed] [Google Scholar]
  • 46.McCarren P, Springer C, Whitehead L. An investigation into pharmaceutically relevant mutagenicity data and the influence on Ames predictive potential. J. Cheminformatics. 2011;3:51. doi: 10.1186/1758-2946-3-51. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Lowe R, Glen RC, Mitchell JBO. Predicting phospholipidosis using machine learning. Mol Pharm. 2010;7:1708–1718. doi: 10.1021/mp100103e. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Marchese Robinson RL, Glen RC, Mitchell JBO. Development and comparison of hERG Blocker classifiers: assessment on different datasets yields markedly different results. Mol Inf. 2011;30:443–458. doi: 10.1002/minf.201000159. [DOI] [PubMed] [Google Scholar]
  • 49.Li S, Fedorowicz A, Singh H, Soderholm SC. Application of the random forest method in studies of local lymph node assay based skin sensitization data. J Chem Inf Model. 2005;45:952–964. doi: 10.1021/ci050049u. [DOI] [PubMed] [Google Scholar]
  • 50.Johnston A, Johnston BF, Kennedy AR, Florence AJ. Targeted crystallisation of novel carbamazepine solvates based on a retrospective random forest classification. CrystEngComm. 2008;10:23–25. [Google Scholar]
  • 51.Palmer DS, O'Boyle NM, Glen RC, Mitchell JBO. Random forest models to predict aqueous solubility. J Chem Inf Model. 2007;47:150–158. doi: 10.1021/ci060164k. [DOI] [PubMed] [Google Scholar]
  • 52.Hughes LD, Palmer DS, Nigsch F, Mitchell JBO. Why are some properties more difficult to predict than others? A study of QSPR models of solubility, melting point, and log P. J Chem Inf Model. 2008;48:220–232. doi: 10.1021/ci700307p. [DOI] [PubMed] [Google Scholar]
  • 53.Ballester PJ, Mitchell JBO. A machine learning approach to predicting protein-ligand binding affinity with applications to molecular docking. Bioinformatics. 2010;26:1169–1175. doi: 10.1093/bioinformatics/btq112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Springer C, Adalsteinsson H, Young MM, Kegelmeyer PW, Roe DC. PostDOCK:  a structural, empirical approach to scoring protein ligand complexes. J Med Chem. 2005;48:6821–6831. doi: 10.1021/jm0493360. [DOI] [PubMed] [Google Scholar]
  • 55.Sato T, Honma T, Yokoyama S. Combining machine learning and pharmacophore-based interaction fingerprint for in silico screening. J Chem Inf Model. 2010;50:170–185. doi: 10.1021/ci900382e. [DOI] [PubMed] [Google Scholar]
  • 56.Zilian D, Sotriffer CA. SFCscoreRF: a random forest-based scoring function for improved affinity prediction of protein–ligand complexes. J Chem Inf Model. 2013;53:1923–1933. doi: 10.1021/ci400120b. [DOI] [PubMed] [Google Scholar]
  • 57.Noble WS. What is a support vector machine? Nat Biotech. 2006;24:1565–1567. doi: 10.1038/nbt1206-1565. [DOI] [PubMed] [Google Scholar]
  • 58.Joachims T, Finley T, Yu CN. Cutting-plane training of structural SVMs. Mach Learn. 2009;77:27–59. [Google Scholar]
  • 59.Hsu CW, Lin CJ. A comparison of methods for multiclass support vector machines. IEEE Trans Neural Netw. 2002;13:415–425. doi: 10.1109/72.991427. [DOI] [PubMed] [Google Scholar]
  • 60.Napolitano F, Zhao Y, Moreira V, Tagliaferri R, Kere J, D'Amato M, Greco D. Drug repositioning: a machine-learning approach through data integration. J Cheminf. 2013;5:30. doi: 10.1186/1758-2946-5-30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Kinnings SL, Liu N, Tonge PJ, Jackson RM, Xie L, Bourne PE. A machine learning-based method to improve docking scoring functions and its application to drug repurposing. J Chem Inf Model. 2011;51:408–419. doi: 10.1021/ci100369f. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Dong N, Lu WC, Chen NY, Zhu YC, Chen KX. Using support vector classification for SAR of fentanyl derivatives. Acta Pharmacol Sin. 2005;26:107–112. doi: 10.1111/j.1745-7254.2005.00014.x. [DOI] [PubMed] [Google Scholar]
  • 63.Li Q, Jørgensen FS, Oprea T, Brunak S, Taboureau O. hERG classification model based on a combination of support vector machine method and GRIND descriptors. Mol Pharm. 2008;5:117–127. doi: 10.1021/mp700124e. [DOI] [PubMed] [Google Scholar]
  • 64.Doddareddy MR, Klaasse EC, Shagufta, IJzerman AP, Bender A. Prospective validation of a comprehensive in silico hERG model and its applications to commercial compound and drug databases. ChemMedChem. 2010;5:716–729. doi: 10.1002/cmdc.201000024. [DOI] [PubMed] [Google Scholar]
  • 65.Liao Q, Yao J, Yuan S. Prediction of mutagenic toxicity by combination of recursive partitioning and support vector machines. Mol Divers. 2007;11:59–72. doi: 10.1007/s11030-007-9057-5. [DOI] [PubMed] [Google Scholar]
  • 66.von Korff M, Sander T. Toxicity-indicating structural patterns. J Chem Inf Model. 2006;46:536–544. doi: 10.1021/ci050358k. [DOI] [PubMed] [Google Scholar]
  • 67.Sun H, Shahane S, Xia M, Austin CP, Huang R. Structure based model for the prediction of phospholipidosis induction potential of small molecules. J Chem Inf Model. 2012;52:1798–1805. doi: 10.1021/ci3001875. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Cao DS, Xu QS, Liang YZ, Chen X, Li HD. Prediction of aqueous solubility of druglike organic compounds using partial least squares, back-propagation network and support vector machine. J Chemometrics. 2010;24:584–595. [Google Scholar]
  • 69.Nigsch F, Bender A, van Buuren B, Tissen J, Nigsch E, Mitchell JBO. Melting point prediction employing k-nearest neighbor algorithms and genetic parameter optimization. J Chem Inf Model. 2006;46:2412–2422. doi: 10.1021/ci060149f. [DOI] [PubMed] [Google Scholar]
  • 70.De Ferrari L, Aitken S, van Hemert J, Goryanin I. EnzML: multi-label prediction of enzyme classes using InterPro signatures. BMC Bioinf. 2012;13:61. doi: 10.1186/1471-2105-13-61. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Itskowitz P, Tropsha A. k nearest neighbors QSAR modeling as a variational problem:  theory and applications. J Chem Inf Model. 2005;45:777–785. doi: 10.1021/ci049628+. [DOI] [PubMed] [Google Scholar]
  • 72.Honório KM, da Silva AB. A study on the influence of molecular properties in the psychoactivity of cannabinoid compounds. J Mol Model. 2005;11:200–209. doi: 10.1007/s00894-005-0243-z. [DOI] [PubMed] [Google Scholar]
  • 73.Ajmani S, Jadhav K, Kulkarni SA. Three-dimensional QSAR using the k-nearest neighbor method and its interpretation. J Chem Inf Model. 2006;46:24–31. doi: 10.1021/ci0501286. [DOI] [PubMed] [Google Scholar]
  • 74.Basak SC, Grunwald GD. Predicting mutagenicity of chemicals using topological and quantum chemical parameters: a similarity based study. Chemosphere. 1995;31:2529–2546. doi: 10.1016/0045-6535(95)00122-o. [DOI] [PubMed] [Google Scholar]
  • 75.Neumann D, Kohlbacher O, Merkwirth C, Lengauer T. A fully computational model for predicting percutaneous drug absorption. J Chem Inf Model. 2006;46:424–429. doi: 10.1021/ci050332t. [DOI] [PubMed] [Google Scholar]
  • 76.Kovatcheva A, Golbraikh A, Oloff S, Xiao YD, Zheng W, Wolschann P, Buchbauer G, Tropsha A. Combinatorial QSAR of ambergris fragrance compounds. J Chem Inf Comput Sci. 2004;44:582–595. doi: 10.1021/ci034203t. [DOI] [PubMed] [Google Scholar]
  • 77.Zakarya D, Chastrette M, Tollabi M, Fkih-Tetouani S. Structure-camphor odour relationships using the generation and selection of pertinent descriptors approach. Chemometrics Intell Lab Syst. 1999;48:35–46. [Google Scholar]
  • 78.Gute BD, Basak SC, Mills D, Hawkins DM. Tailored similarity spaces for the prediction of physicochemical properties. Internet Electr J Mol Des. 2002;1:374–387. [Google Scholar]
  • 79.Kuhne R, Ebert RU, Schuurmann G. Model selection based on structural similarity-method description and application to water solubility prediction. J Chem Inf Model. 2006;46:636–641. doi: 10.1021/ci0503762. [DOI] [PubMed] [Google Scholar]
  • 80.Ajmani S, Rogers SC, Barley MH, Livingstone DJ. Application of QSPR to mixtures. J Chem Inf Model. 2006;46:2043–2055. doi: 10.1021/ci050559o. [DOI] [PubMed] [Google Scholar]
  • 81.Cramer RD, Redl G, Berkhoff CE. Substructural analysis. A novel approach to the problem of drug design. J Med Chem. 1974;17:533–535. doi: 10.1021/jm00251a014. [DOI] [PubMed] [Google Scholar]
  • 82.Lowe R, Mussa HY, Nigsch F, Glen RC, Mitchell JBO. Predicting the mechanism of phospholipidosis. J Cheminformatics. 2012;4:2. doi: 10.1186/1758-2946-4-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Koutsoukas A, Lowe R, KalantarMotamedi Y, Mussa HY, Klaffke W, Mitchell JBO, Glen R, Bender A. In silico target predictions: defining a benchmarking dataset and comparison of performance of the multiclass Naïve Bayes and Parzen-Rosenblatt Window. J Chem Inf Model. 2013;53:1957–1966. doi: 10.1021/ci300435j. [DOI] [PubMed] [Google Scholar]
  • 84.Cannon EO, Amini A, Bender A, Sternberg MJE, Muggleton SH, Glen RC, Mitchell JBO. Support vector inductive logic programming outperforms the naive Bayes classifier and inductive logic programming for the classification of bioactive chemical compounds. J Comput Aided Mol Des. 2007;21:269–280. doi: 10.1007/s10822-007-9113-3. [DOI] [PubMed] [Google Scholar]
  • 85.Nigsch F, Bender A, Jenkins JL, Mitchell JBO. Ligand-target prediction using Winnow and naive Bayesian algorithms and the implications of overall performance statistics. J Chem Inf Model. 2008;48:2313–2325. doi: 10.1021/ci800079x. [DOI] [PubMed] [Google Scholar]
  • 86.Frank E, Trigg L, Holmes G, Witten IH. Technical note: naive Bayes for regression. Mach Learn. 2000;41:5–25. [Google Scholar]
  • 87.Korb O. Efficient ant colony optimization algorithms for structure- and ligand-based drug design. Chem Cent J. 2009;3:O10. [Google Scholar]
  • 88.Lowe R, Mussa HY, Mitchell JBO, Glen RC. Classifying molecules using a sparse probabilistic kernel binary classifier. J Chem Inf Model. 2011;51:1539–1544. doi: 10.1021/ci200128w. [DOI] [PubMed] [Google Scholar]
  • 89.Mavridis L, Mitchell JBO. Predicting the protein targets for athletic performance-enhancing substances. J Cheminformatics. 2013;5:31. doi: 10.1186/1758-2946-5-31. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Mikut R, Hilpert K. Interpretable features for the activity prediction of short antimicrobial peptides using fuzzy logic. Int J Pept Res Ther. 2009;15:129–137. [Google Scholar]
  • 91.Strömbergsson H, Prusis P, Midelfart H, Lapinsh M, Wikberg JE, Komorowski J. Rough set-based proteochemometrics modeling of G-protein-coupled receptor-ligand interactions. Proteins. 2006;63:24–34. doi: 10.1002/prot.20777. [DOI] [PubMed] [Google Scholar]
  • 92.Nigsch F, Mitchell JBO. How to winnow actives from inactives: introducing molecular orthogonal sparse bigrams (MOSBs) and multiclass Winnow. J Chem Inf Model. 2008;48:306–318. doi: 10.1021/ci700350n. [DOI] [PubMed] [Google Scholar]
  • 93.Nigsch F, Mitchell JBO. Toxicological relationships between proteins obtained from protein target predictions of large toxicity databases. Toxicol Appl Pharmacol. 2008;231:225–234. doi: 10.1016/j.taap.2008.05.007. [DOI] [PubMed] [Google Scholar]
  • 94.Dubus E, Ijjaali I, Petitet F, Michel A. In silico classification of hERG channel blockers: a knowledge-based strategy. ChemMedChem. 2006;1:622–630. doi: 10.1002/cmdc.200500099. [DOI] [PubMed] [Google Scholar]
  • 95.Asikainen A, Kolehmainen M, Ruuskanen J, Tuppurainen K. Structure-based classification of active and inactive estrogenic compounds by decision tree, LVQ and kNN methods. Chemosphere. 2006;62:658–673. doi: 10.1016/j.chemosphere.2005.04.115. [DOI] [PubMed] [Google Scholar]
  • 96.Karakoc E, Sahinalp SC, Cherkasov A. Comparative QSAR- and fragments distribution analysis of drugs, druglikes, metabolic substances, and antimicrobial compounds. J Chem Inf Model. 2006;46:2167–2182. doi: 10.1021/ci0601517. [DOI] [PubMed] [Google Scholar]
  • 97.Beltran N, Duartemermoud M, Bustos M, Salah S, Loyola E, Penaneira A, Jalocha J. Feature extraction and classification of Chilean wines. J Food Eng. 2006;75:1–10. [Google Scholar]
  • 98.Oloff S, Muegge I. kScore: a novel machine learning approach that is not dependent on the data structure of the training set. J Comput Aided Mol Des. 2007;21:87–95. doi: 10.1007/s10822-007-9108-0. [DOI] [PubMed] [Google Scholar]
  • 99.Tropsha A, Gramatica P, Gombar VK. The importance of being earnest: validation is the absolute essential for successful application and interpretation of QSPR models. QSAR Comb Sci. 2003;2:69–77. [Google Scholar]
  • 100.Varnek A, Baskin I. Machine learning methods for property prediction in chemoinformatics: quo vadis? J Chem Inf Model. 2012;52:1413–1437. doi: 10.1021/ci200409x. [DOI] [PubMed] [Google Scholar]
  • 101.Xu QS, Liang YZ, Du YP. Monte Carlo cross-validation for selecting a model and estimating the prediction error in multivariate calibration. J Chemometrics. 2004;18:112–120. [Google Scholar]
  • 102.Rücker C, G R, Meringer M. y-Randomization and its variants in QSPR/QSAR. J Chem Inf Model. 2007;47:2345–2357. doi: 10.1021/ci700157b. [DOI] [PubMed] [Google Scholar]
  • 103.Sokolova M, Lapalme G. A systematic analysis of performance measures for classification tasks. Inf Process Manage. 2009;45:427–437. [Google Scholar]
  • 104.Baldi P, Brunak S, Chauvin Y, Andersen CAF, Nielsen H. Assessing the accuracy of prediction algorithms for classification: an overview. Bioinformatics. 2000;16:412–424. doi: 10.1093/bioinformatics/16.5.412. [DOI] [PubMed] [Google Scholar]
  • 105.Plewczynski D, Spieser SAH, Koch U. Assessing different classification methods for virtual screening. J Chem Inf Model. 2006;46:1098–1106. doi: 10.1021/ci050519k. [DOI] [PubMed] [Google Scholar]
  • 106.Consonni V, Ballabio D, Todeschini R. Comments on the definition of the Q2 parameter for QSAR validation. J Chem Inf Model. 2009;49:1669–1678. doi: 10.1021/ci900115y. [DOI] [PubMed] [Google Scholar]
  • 107.Hopfinger AJ, Esposito EX, Llinàs A, Glen RC, Goodman JM. Findings of the challenge to predict aqueous solubility. J Chem Inf Model. 2009;49:1–5. doi: 10.1021/ci800436c. [DOI] [PubMed] [Google Scholar]
  • 108.Ploemen JPHTM, Kelder J, Hafmans T, van de Sandt H, van Burgsteden JA, Saleminki PJ, van Esch E. Use of physicochemical calculation of pKa and CLogP to predict phospholipidosis-inducing potential: a case study with structurally related piperazines. Exp Toxicol Pathol. 2004;55:347–355. doi: 10.1078/0940-2993-00338. [DOI] [PubMed] [Google Scholar]
  • 109.R Development Core Team. R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing. 2006. ISBN 3-900051-07-0.
  • 110.Carlsson L, Helgee EA, Boyer S. Interpretation of nonlinear QSAR models applied to Ames mutagenicity data. J Chem Inf Model. 2009;49:2551–2558. doi: 10.1021/ci9002206. [DOI] [PubMed] [Google Scholar]
  • 111.Chen H, Carlsson L, Eriksson M, Varkonyi P, Norinder U, Nilsson I. Beyond the scope of Free-Wilson analysis: building interpretable QSAR models with machine learning algorithms. J Chem Inf Model. 2013;53:1324–1336. doi: 10.1021/ci4001376. [DOI] [PubMed] [Google Scholar]

Articles from Wiley Interdisciplinary Reviews. Computational Molecular Science are provided here courtesy of Wiley

RESOURCES