Abstract
Recent progress in theoretical systems biology, applied mathematics and computational statistics allows us to compare the performance of different candidate models at describing a particular biological system quantitatively. Model selection has been applied with great success to problems where a small number—typically less than 10—of models are compared, but recent studies have started to consider thousands and even millions of candidate models. Often, however, we are left with sets of models that are compatible with the data, and then we can use ensembles of models to make predictions. These ensembles can have very desirable characteristics, but as I show here are not guaranteed to improve on individual estimators or predictors. I will show in the cases of model selection and network inference when we can trust ensembles, and when we should be cautious. The analyses suggest that the careful construction of an ensemble—choosing good predictors—is of paramount importance, more than had perhaps been realized before: merely adding different methods does not suffice. The success of ensemble network inference methods is also shown to rest on their ability to suppress false-positive results. A Jupyter notebook which allows carrying out an assessment of ensemble estimators is provided.
Keywords: ensemble estimator, network inference, model averaging, model selection, statistical physics
1. Introduction
In physics, simple and elegant symmetry relationships have often led the way to theoretical models [1]. Most importantly Emmy Noether’s theorem has been pivotal in establishing the correspondence between continuous symmetries and conservation laws in physics [2]; it has been instrumental in the derivation of physical laws of nature. Biology has not been able to draw on such fundamental principles [3], to a large degree because most processes are intrinsically dissipative (meaning energy is produced and consumed) and hence the conditions where Noether’s theorem holds simply do not apply. Instead, biological models have often had a heuristic element or described the rates of change in (often macroscopic) observables (e.g. species of plants or animals, or molecular species).
Writing down the set of equations is an important starting point in modelling as it forces us to express our assumptions in a precise form. Which form to choose is, however, not unambiguously obvious. Instead, we often rely on data to decide between the different options. Statistical model selection [4] provides the tools to make such decisions, balancing the ability of a model to fit the data with the model’s complexity. As larger and larger models, even models for whole cells [5–8], are being considered model selection problems will presumably become the norm, especially when models are constructed exhaustively or automatically [9–16].
In some, probably many, situations model selection will not be able to decide on a single best model. Instead, many models may have comparable support. In such a situation, we may then base analysis or predictions on the ensemble of models that are supported by the data [17]. Each model’s contribution to the prediction etc. is weighted by the relative support it has received. Such estimates or predictions based on ensembles have been referred to as exploiting the ‘wisdom of crowds’ [18]. This refers to the notion that groups of individuals/models are more likely to be better than those based on a single individual/model. This concept, however, also relates to much earlier work, Charles Mackay’s ‘Extraordinary Popular Delusions and the Madness of Crowds’ [19], a 19th century account of how popular opinion can support quite extraordinary and plainly wrong opinions and concepts.
Ensemble estimators have a surprisingly long history, outlined in [20], and aspects such as bagging, boosting and stacking [21] are firmly established in the statistical learning literature; see, for example, [22] for a comprehensive treatment. There has been interest in evolutionary biology [23]; and following [18], there have been further developments in systems and computational biology, e.g. [24,25]. But in the context of network inference, combining different network reconstruction methods has sometimes been viewed as necessarily optimal [26]. Below, we show that this is not automatically the case. In turn, I will show that model averaging and ensemble estimation are susceptible to poorly defined sets of candidate models; and that the behaviour of ensemble approaches to network reconstruction depends strongly on the composition of the ensemble. For very good ensembles, the advantage comes mainly from reducing the number of false-positive edges. Both problems share a dependence on the quality of the ensemble, and we map out and quantify these influences; we also provide self-contained Julia code for further in silico experimentation and analysis of ensemble prediction methods.
2. Model selection and multi-model inference
We assume that we have a universe of models, ,
that are potential mechanisms, by which some data, , have been generated. For simplicity, we consider a finite number of models, N. Furthermore, for each model, Mi, we assume that we have a data generating function, fi(θi), parametrized by a parameter vector θi which is chosen from some suitable continuous parameter space,
Coping with the size of the parameter space is one of the essential challenges of parameter estimation and model selection.
We start from a Bayesian framework [27], where we seek to determine the posterior distribution over parameters,
2.1 |
where is the likelihood, πi(θi) the prior over the parameters for model Mi, and (here, we make the dependence on the choice of model explicit through an index) is known as the evidence.
In the Bayesian framework model selection is a (relatively) straightforward extension, and the model posterior is given by
2.2 |
where analogously to equation (2.1), we have the model posterior, , and model prior, π(Mi). The denominator terms in equations (2.1) and (2.2) are notoriously hard to evaluate for all but the simplest cases, and a large amount of ingenuity and work has been invested into computational schemes [27,28], including Markov chain Monte Carlo, sequential Monte Carlo and related approaches. Often even the likelihood is prohibitively expensive to evaluate and so-called approximate Bayesian computation schemes have been devised to make Bayesian statistical inference possible [29].
Alternatives to the Bayesian framework, such as likelihood-based inference and optimization of cost functions [30], result in point estimates for the parameters, e.g. the value of θ ′ that maximizes the probability of observing the data,
2.3 |
Similarly, we can determine the maximum a posteriori estimate by finding the mode of the posterior distribution [27],
2.4 |
Compared to analysis of the posterior distribution, such local estimates lose a lot of relevant information, but some characteristics can still be recovered by considering the local curvature of the likelihood, i.e. the information matrix, or cost-function surface around the (local) extremum identified in this manner [31,32].
Model selection frameworks that are based on likelihood inference rely on criteria to find the optimal model among a set of candidate models. The Akaike information criterion (AIC) [4,33] for model Mi is given by
2.5 |
with given by equation (2.3), and ni the number of parameters of model Mi. The AIC is probably the most widely used model selection criterion, despite the fact that it is biased in favour of overly complicated models as the amount of available data increases. The Bayesian information criterion does not suffer in the same way; it is defined as
2.6 |
with ν representing the size of the data or number of samples. Several other information criteria exist (discussed e.g. in [4,34]), but they all share in common the purpose of balancing model complexity with model fit. The BIC can be derived as an approximation to the full Bayesian model posterior probability, which achieves this balance implicitly.
If model selection cannot pick out a clear winner, then either (i) further analysis should be used to design better, more informative experiments that can discriminate among these models [35–37]; or (ii) these models should be considered as an ensemble [4,25]. The former has definite attractions as it will lead to an increase in our understanding if we can discard some of the models/mechanistic hypotheses.
The latter approach, basing analysis and especially predictions on an ensemble of models has become a popular concept in machine learning. Most notably, in the context of biological network inference the ‘wisdom of crowds’ concept [18] has been important in popularizing inference based on several models. Here, we are considering model averaging where contributions from different models may be weighted by their relative explanatory performance. In the Bayesian framework, we can use the posterior probability directly. In the context of an information criterion Ii for model i, we define [4]
2.7 |
and then determine the model weight as
2.8 |
The model weights (e.g. the Akaike weight if Ii is the AIC) provide the relative probability of model Mi to be the correct model conditional on the data and the set of models, , considered. Model averaging in this framework can serve as a variance reduction technique [4,21].
3. Statistical physics of model selection and ensemble estimation
In order to simplify the discussion, we define a relationship between the model probability (always implicitly understood to be either a posterior or relative model probability), pi, and a cost or energy, εi, [8] as
3.1 |
with the normalization constant Z (the partition function) given by
With this in hand, we can straightforwardly consider different model selection/averaging frameworks in a similar manner.
In general, the true data-generating model (a natural system) will not be represented perfectly by any of the models in ; we denote this true model by . But if we are interested in finding out whether has a certain characteristic ϕ we would obtain this from the appropriate ensemble average
3.2 |
where is the conventional indicator function, i.e.
Equation (3.2) is based on three, potentially strong assumptions.
-
1.
The model universe, , is ‘complete’ in the sense that we expect to contain any model, Mk, which we might expect to be a reasonable description of (always remembering that ).
-
2.
It is decidable if ϕ is part of .
-
3.
ϕ has played no part in the construction of the model probabilities, pi, equation (3.1).
The first assumption is arguably the strongest assumption. In principle, we can update by adding additional models in light of data; specifying new priors π ′ for the models will require some care. One important condition that (and ) must fulfil is that models must not appear more than once in the set. This is important to keep in mind as we are increasingly relying on automated generation or exhaustive enumeration of models.
With fixed and equation (3.2) is, however, a good approach for ensemble-based prediction and estimation. It also encompasses Bayesian model averaging and multi-model inference based on information criteria [4].
3.1. The effective model space
We first analyse a simple case where all models in our universe have associated costs drawn from a suitably well behaved probability distribution, q(η) [8], with positive support and associated density, fq(ε), over the model energies, εi, such that
3.3 |
Because εi is a random variable, the relative weight, ωi = exp (−εi), will also be a random variable, and we can obtain the probability density function, f(ω), via change of variables as
3.4 |
For different choices of q(ε), we can now investigate the distribution over model weights. For example, if εi ∼ Gamma(α, θ) (where α and θ denote the shape and scale parameters, respectively), then
3.5 |
with Γ( · ) the Gamma function. The Gamma distribution is a flexible and generic distribution and is chosen for its generality rather than any particular property and our discussion here does not depend on its specifics. Some representative distributions over ε and the corresponding distributions over ω are shown in figure 1a,b.
The distribution over model costs, ε, affects the distributions over model weights, ω. This is important to realize when deciding on how to triage model sets for further analysis and prediction [38]: generally, inference based on all models weighted by ωi is neither practical nor desirable, as many models with low weight will mask the information available in the good models. If, for example, we only include models with ω ∈ [0.9, 1.0] then the average model cost
(because ε = −log (ω)) for these sets will be 0.005 (blue), 0.0015 (green), 6 × 10−4 (purple) and 9 × 10−5 (orange), see figure 1c. The (unknown) distribution over costs can affect multi-model inference quite profoundly. But for model universes that are enriched for good models (i.e. many models Mi with low values of εi) selecting a subset of models based on even a fairly conservative threshold for the model weights ωi can result in a sufficiently large model sets for further prediction.
3.2. A simple test case for multi-model inference
Here, we study a very simplistic scenario in which we have three types of models, borrowing and adapting Box’s [39] terminology:
Useful Models which capture important aspects of and which have an associated cost ε1.
Useless Models which are qualitatively and quantitative poor descriptions of reality and have an associated cost ε2 ≫ ε1.
Nuisance Models which are qualitatively different from reality, but which can quantitatively capture aspects of by chance; their costs are random variables, η.
Nuisance models are here assumed to be models where the quantitative agreement with data is unrelated to the data generating mechanism . Purely machine-learning-based models are one way in which we can realize such nuisance models [40]; for small datasets , poorly designed experiments, or simply lack of prior knowledge, there are many ways in which model fit may only be a poor reflection of reality, and this can also give rise to nuisance models [41].
For concreteness let these three different model classes have sizes ν1, ν2 and ν3 = N − ν1 − ν2, and assume that i.e. nuisance models are at worst as bad as useless models. Then the number of nuisance models that have lower associated costs than the useful models is given by
The relative influence of nuisance models can be studied by contrasting three features, ϕ1, ϕ2, and ϕ3, with the following properties:
-
ϕ1:
equally represented with frequency ξ among models of all classes.
-
ϕ2:
only represented among useful models.
-
ϕ3:
only represented among ‘nuisance’ models.
With equation (3.2) we can obtain for any property with frequencies ξi in the classes i = 1, 2, 3,
3.6 |
First, for ϕ1, we trivially obtain
3.7 |
For the more interesting probability for ϕ2 under the model averaging framework, we obtain
3.8 |
and finally, for a characteristic shared by and confined to the set of nuisance models, we obtain
3.9 |
From equations (3.8) and (3.9), we see that multi-model averaging is prone to propagate wrong results if nuisance models are frequent and some receive good quantitative support (i.e. low model costs, ε). Equally worrying, the same scenario can make it hard for true aspects of to receive sufficient support through equation (3.8) if there are many nuisance and useless models included in the data.
To illustrate this further we can consider the case where ξ2 = 1 and ξ3 = 1 (meaning every useful model exhibits characteristic ϕ2, and every nuisance model characteristic ϕ3) and ask when is ? We obtain
3.10 |
Thus, if useful models are sufficiently rare in the model set (say ν1 < 0.1) the nuisance models’ characteristics will have high weight in the ensemble average; see also figure 2. None of the parameters ν1, ν2, ν3, ε1, ε2 are, of course, known, and we cannot know which class a model belongs to a priori. Thus model averaging is not a panacea and requires careful consideration of which models are included in the prediction set.
4. Ensemble estimation and network inference
Network inference can also benefit from ensemble methods [18,24,42] but here, too, potential pitfalls arise. We are considering directed networks, with V nodes, and L edges; the adjacency matrix, A, is a convenient way to specify such networks, if we indicate the presences and absences of edges, by one and zero, respectively, in its entries,
In network inference, we seek to determine whether the data suggest statistical relationships among pairs of nodes vi, vj that may indicate a functional relationship, represented by an edge connecting the nodes. We consider k different algorithms, , which predict the presence or absence of a link from i to j. If we have for the false positive and false-negative probabilities for method κ,
4.1 |
and
4.2 |
we can assess how beneficial the ensemble estimators are for the quality of the inferred networks.
4.1. Ensembles of identical estimators
The simplest case, which is already instructive as a baseline, is where all methods have identical performance, and , . If the performance of the inference methods is statistically independent, then the number of agreeing inference methods is a binomial random variable. If we set a threshold on the minimum number of concordances among inference methods, k0, we observe for the overall probability of scoring a true edge from the ensemble,
4.3 |
while the probability of a false negative is
4.4 |
To illustrate the outcome of such a simple estimation procedure, we assume a network loosely based on expected Homo sapiens network sizes [43] (22 000 nodes and 750 000 interactions). In figure 3,we consider 10 ensembles and base the ensemble estimator on the minimum number, k0, of methods that have to predict an edge. If k0 is too small, then there will be too many positives as is the cases here; a majority-vote rule, i.e. here k0 = 6, would do an acceptable job in terms of precision, recall and F1 statistics [21], but this does depend on the false positive ratio in particular (as biological networks are sparse), as well as the size of the ensemble of inference algorithms, .
4.2. Ensemble estimators can be worse than individual estimators
We are interested in ensemble estimators because we know that individual estimators are far from perfect. But ensembles are not guaranteed to always improve on individual estimators. We compare an ensemble of equally well performing estimators with a single estimator. The ensemble false negative probability, T, is given by
4.5 |
where 2F1 is the hypergeometric function [47] (see appendix B for an approximation for small arguments). From this, we can determine when the ensemble false negative rate, T, will be greater than t, i.e.
Equally, we obtain for the ensemble false positive rate, S,
4.6 |
For low thresholds, k0, the ensemble error rate can be greater than that of the individual estimator; this is because the stringency of the ensemble prediction is then reduced, as the cumulative probability of a sufficiently small number of estimators to ‘accidentally’ agree is greater than the error rate of the individual estimator. We show this for two false negative rates, t = 0.1 and t = 0.05, in figure 4a,b.
4.3. Heterogeneous ensembles of network inference methods
We next focus on the case of a small set of predictors, k = 10, and two classes of methods: a set of good methods with error rates t1 and s1; and a set of poor methods with error rates t2 > t1 and s2 > s1. In figure 5, we show, again for a case modelled on a likely human gene regulation network, the likely true- and false-positive predictions arising from ensembles with different numbers of good versus poor edge prediction methods. The basic lesson is that good methods have to outnumber bad methods; otherwise, especially the precision will suffer. Here, we have chosen a simple majority-vote criterion. To bring precision and F1 statistic up to a satisfying level (say in excess of 0.7) requires essentially purging the ensemble of the weakly performing methods (i.e. ). This only points to the extent to which poor methods can compromise the performance of ensemble estimators (and the accompanying Jupyter notebook can be used to explore this further).
In sparse networks especially, poor estimators will result in inflation of false-positive results, and lead to overall poor performance: in a directed network there are N × (N − 1) possible interactions, but this is still vastly greater than the number of existing edges. For example, in the case of the human network we would have of the order of 4.8 × 108 potential interactions of which only some 750 000 are expected to exist. So even for s2 = 0.5 there would be about 470 000 cases where ten such methods would agree and score an edge, and 1.8 × 108 false predictions, if a simple majority vote rule were applied.
5. Discussion
We have shown that ensemble estimators are not as robust as has sometimes been claimed [26], or incorrectly surmised from the success of community average results in the DREAM competition [18]; there, of course, it had already been shown that certain, carefully selected, subsets of estimators give better results than others [24].
For the analysis of multi-model inference from mechanistic models, we can distill two points: (i) ensembles of mechanistic models that are reasonably defined [48] (i.e. their construction incorporates any available mechanistic insights; duplicate models are avoided; the model is predictive and can be used to generate data that can be compared with real observations/data) can be combined with the aid of model selection criteria or Bayesian posterior model probabilities with relative ease and safety; (ii) the inclusion of ‘nuisance models’ can hamper ensemble approaches if they come to predominate the model universe . Such situations could become more likely as model spaces are explored exhaustively [16] or automatically [12]. Because of the formalism connecting different model selection criteria, equation (3.1), these are general results, and do not depend on the particular model averaging procedure chosen (as also clear from the analysis in [20]). So, in essence, the construction of the models in [3,49] will determine the robustness of model averaging or ensemble approaches to prediction and analysis. Little is to be gained by increasing the size of beyond the (already quite large) set of reasonable models.
In the context of network inference, the situation is similar. We find that the poor performance of some methods can drag down the performance of an ensemble estimator or network predictor. So like in the construction of the model universe before, the make-up of the ensemble of network inference methods, , does matter considerably (as was also found in the empirical study of [18]). Majority vote will typically be a sound criterion for a set of reasonable estimators, though not necessarily optimal from the perspective of precision and F1 measures. This is because biological networks are sparse and false-positives will predominate the inferred networks unless they are carefully controlled. For a set of statistically similar powerful inference methods, a conservative criterion for scoring edges will improve on the overall performance of the individual estimators, however.
The problem of network inference has long been known to be challenging. One reason for this (in addition to the large-scale testing problem) is that we do not have a fair way to score and compare the performance of different network inference approaches. The most promising existing approaches are typically computationally expensive and rigorous in silico assessment of performance as well as the factors influencing performance is often seen as computationally prohibitively expensive. There is also a danger of biasing simulated data in favour of a given method and the DREAM competition has aimed—with some success we feel—to avoid this, and other approaches have followed suite. Clearly, more effort in this domain is needed and computational cost should not preclude rigorous assessment [50]. This situation is mirrored in other areas of genomics and systems biology, e.g. pseudo-temporal ordering approaches have until recently [51] rarely been rigorously tested. But what is also needed are approaches which allow us to assign confidence to inferred networks, or, more specifically, predicted interactions without recourse to a gold-standard [52]. Here, measures based on biological expectations/knowledge offer promising routes for filtering out poor methods [53] (see figure 6).
One of the potential initial attractions of using a panel of network inference algorithms is that different methods may capture different aspects of the data and in concert may provide a more complete representation of a ‘true’ network of functional relationships among the genes in an organism under a given scenario. While appealing this notion needs to be viewed with caution [56]. Combining the most powerful methods by leveraging their individual strengths is possible in principle [20,22,54,55], but requires us to characterize each method Oi reliably and independently.
6. Conclusion
In summary, unless we know the constituency of the model universe, , or the ensemble of predictors, , we have limited ways of telling whether we are dealing with a madness of crowds or a wisdom of crowds scenario. However, the present analysis shows that ensemble procedures will be robust as long as the ensembles are carefully constructed. In the context of biological network inference, reduction in false-positives is the primary cause of their success. Without a robust and transferable way of assessing the strengths and weaknesses of different methods, we cannot (yet) use tools from decision theory that pool these strengths for diverse and heterogeneous ensembles [54,55]. Currently, the best advice, in light of the analysis carried out here, is to be ruthless in weeding out poorly performing methods for network inference, or models with low weight for multi-model averaging. So there is no need, for example, to include correlation networks, even though they are cheap to calculate: their performance is simply too poor to warrant inclusion in an ensemble. Quality is more important than quantity.
Appendix A. Assessing the performance of network inference methods
Real biological networks are sparse [57]; this means that a predictor which scores each candidate edge to be absent can have high performance if the number of false-positives is heavily influencing how we quantity performance of a network inference method. We therefore focus on precision and recall and a derived statistic, the F1 statistic [21,57]. We denote the numbers of true and false positive inferences (i.e. scored edges in the context of network inference) by TP and FP, and the true and false negatives by TN and FN. Then the precision, P, is given by
and the recall by
The F1 statistic is given by
These statistics are confined to the [0, 1] range, with larger values indicating better performance, and the F1 statistic in particular becomes maximal at the point where the curves for P and R intersect.
Appendix B. Approximating the hypergeometric function
The hypergeometric function [47], 2F1, appearing in equation (4.5) can be unwieldy to evaluate for large k and k0. For sufficiently small values of t, we can Taylor-expand it around t = 0 and then obtain (if we restrict the expansion of the hypergeometric function to third order)
From this, we can also determine when the ensemble false-negative rate T will be greater than t by solving
Data accessibility
A Jupyter Notebook containing the Julia code to reproduce all the computational results here, and to explore the effects of the madness of crowds in network inference and model averaging is available at https://github.com/MichaelPHStumpf/Madness-Of-Crowds.
Competing interests
I declare I have no competing interest.
Funding
This work has been supported by the University of Melbourne Driving Research Momentum Program. Funding has been received from the Volkswagen Stiftung through a ‘Life?’ Program grant.
References
- 1.Thorne KS, Blandford RD. 2017. Modern classical physics: optics, fluids, plasmas, elasticity, relativity, and statistical physics. Princeton, NJ: Princeton University Press. [Google Scholar]
- 2.Neuenschwander DE. 2017. Emmy Noether’s wonderful theorem. Baltimore, MD: JHU Press. [Google Scholar]
- 3.May RM. 2004. Uses and abuses of mathematics in biology. Science 303, 790–793. ( 10.1126/science.1094442) [DOI] [PubMed] [Google Scholar]
- 4.Burnham KP, Anderson DR. 2013. Model selection and inference: a practical information-theoretic approach. New York, NY: Springer Science & Business Media. [Google Scholar]
- 5.Karr JR, Sanghvi JC, Macklin DN, Gutschow MV, Jacobs JM, Bolival B, Assad-Garcia N, Glass JI, Covert MW. 2012. A whole-cell computational model predicts phenotype from genotype. Cell 150, 389–401. ( 10.1016/j.cell.2012.05.044) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Karr JR. et al. 2015. Summary of the DREAM8 parameter estimation challenge: toward parameter identification for whole-cell models. PLoS Comput. Biol. 11, e1004096 ( 10.1371/journal.pcbi.1004096) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Lang M, Stelling J. 2016. Modular parameter identification of biomolecular networks. SIAM J. Sci. Comput. 38, B988–B1008. ( 10.1137/15M103306X) [DOI] [Google Scholar]
- 8.Babtie AC, Stumpf MPH. 2017. How to deal with parameters for whole-cell modelling. J. R. Soc. Interface 14, 20170237 ( 10.1098/rsif.2017.0237) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Ma W, Trusina A, El-Samad H, Lim WA, Tang C. 2009. Defining network topologies that can achieve biochemical adaptation. Cell 138, 760–773. ( 10.1016/j.cell.2009.06.013) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Barnes CP, Silk D, Sheng X, Stumpf MPH. 2011. Bayesian design of synthetic biological systems. Proc. Natl Acad. Sci. USA 108, 15190–15195. ( 10.1073/pnas.1017972108) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Szederkényi G, Banga JR, Alonso AA. 2011. Inference of complex biological networks: distinguishability issues and optimization-based solutions. BMC Syst. Biol. 5, 177 ( 10.1186/1752-0509-5-177) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Sunnåker M, Zamora-Sillero E, López García de Lomana A, Rudroff F, Sauer U, Stelling Jor, Wagner A. 2014. Topological augmentation to infer hidden processes in biological systems. Bioinformatics 30, 221–227. ( 10.1093/bioinformatics/btt638) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Babtie AC, Kirk P, Stumpf MPH. 2014. Topological sensitivity analysis for systems biology. Proc. Natl Acad. Sci. USA 111, 18507–18512. ( 10.1073/pnas.1414026112) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Leon M, Woods ML, Fedorec AJH, Barnes CP. 2016. A computational method for the investigation of multistable systems and its application to genetic switches. BMC Syst. Biol. 10, 130–112 ( 10.1186/s12918-016-0375-z) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Gerardin J, Reddy NR, Lim WA. 2019. The design principles of biochemical timers: circuits that discriminate between transient and sustained stimulation. Cell Syst. 102, 100651 ( 10.1016/j.cels.2019.07.008) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Scholes NS, Schnoerr D, Isalan M, Stumpf MPH. 2019. A comprehensive network atlas reveals that turing patterns are common but not robust. Cell Syst. 9, 243–257.e4. ( 10.1016/j.cels.2019.07.007) [DOI] [PubMed] [Google Scholar]
- 17.Toni T, Ozaki Yi, Kirk PDW, Kuroda S, Stumpf MPH. 2012. Elucidating the in vivo phosphorylation dynamics of the ERK MAP kinase using quantitative proteomics data and Bayesian model selection. Mol. Biosyst. 8, 1921–1929. ( 10.1039/c2mb05493k) [DOI] [PubMed] [Google Scholar]
- 18.Marbach D. et al. 2012. Wisdom of crowds for robust gene network inference. Nat. Biotech. 9, 796–804. ( 10.1038/nmeth.2016) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Mackay C. 1841. Extraordinary popular delusions and the madness of crowds. London, UK: Harriman House Limited. [Google Scholar]
- 20.Laan A, Madirolas G, de Polavieja GG. 2017. Rescuing collective wisdom when the average group opinion is wrong. Front. Robot. AI 4, 358 ( 10.3389/frobt.2017.00056) [DOI] [Google Scholar]
- 21.Murphy KP. 2012. Machine learning: a probabilistic perspective. Cambridge, MA: MIT Press. [Google Scholar]
- 22.Kuncheva LI. 2004. Combining pattern classifiers: methods and algorithms. Hoboken, NJ: J. Wiley. [Google Scholar]
- 23.Strimmer K, Rambaut A. 2002. Inferring confidence sets of possibly misspecified gene trees. Proc. R. Soc. B: Biol. Sci. 269, 137–142. ( 10.1098/rspb.2001.1862) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Saez-Rodriguez J, Costello JC, Friend SH, Kellen MR, Mangravite L, Meyer P, Norman T, Stolovitzky GA. 2016. Crowdsourcing biomedical research: leveraging communities as innovation engines. Nat. Rev. Genet. 17, 470–486. ( 10.1038/nrg.2016.69) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Thorne TW, Stumpf MPH. 2012. Graph spectral analysis of protein interaction network evolution. J. R. Soc. Interface 9, 2653–2666. ( 10.1098/rsif.2012.0220) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Le Novère N. 2015. Quantitative and logic modelling of molecular and gene networks. Nat. Rev. Genet. 16, 146–158. ( 10.1038/nrg3885) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Robert CP. 2007. The Bayesian choice: from decision-theoretic foundations to computational implementation. Berlin, Germany: Springer Verlag. [Google Scholar]
- 28.Green PJ, Latuszynski K, Pereyra M, Robert CP. 2015. Bayesian computation: a summary of the current state, and samples backwards and forwards. Stat. Comput. 25, 835–862. ( 10.1007/s11222-015-9574-5) [DOI] [Google Scholar]
- 29.Toni T, Stumpf MPH. 2010. Simulation-based model selection for dynamical systems in systems and population biology. Bioinformatics 26, 104–110. ( 10.1093/bioinformatics/btp619) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Raue A, Karlsson J, Saccomani MP, Jirstrand M, Timmer J. 2014. Comparison of approaches for parameter identifiability analysis of biological systems. Bioinformatics 30, 1440–1448. ( 10.1093/bioinformatics/btu006) [DOI] [PubMed] [Google Scholar]
- 31.Rand DA. 2008. Mapping global sensitivity of cellular network dynamics: sensitivity heat maps and a global summation law. J. R. Soc. Interface 5, S59–S69. ( 10.1098/rsif.2008.0084.focus) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Erguler K, Stumpf MPH. 2011. Practical limits for reverse engineering of dynamical systems: a statistical analysis of sensitivity and parameter inferability in systems biology models. Mol. Biosyst. 7, 1593–1602. ( 10.1039/c0mb00107d) [DOI] [PubMed] [Google Scholar]
- 33.Akaike H. 1974. A new look at the statistical model identification. In Selected papers of Hirotugu Akaike (eds E Parzen, K Tanabe, G Kitagawa), pp. 215–222. New York, NY: Springer ( 10.1007/978-1-4612-1694-0_16) [DOI]
- 34.Kirk PDW, Kirk P, Thorne TW, Stumpf MPH. 2013. Model selection in systems and synthetic biology. Curr. Opin. Biotechnol. 24, 767–774. ( 10.1016/j.copbio.2013.03.012) [DOI] [PubMed] [Google Scholar]
- 35.Liepe J, Filippi S, Komorowski M, Stumpf MPH. 2013. Maximizing the information content of experiments in systems biology. PLoS Comput. Biol. 9, e1002888 ( 10.1371/journal.pcbi.1002888) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Busetto AG, Hauser A, Krummenacher G, Sunnåker M, Dimopoulos S, Ong CS, Stelling J, Buhmann JM. 2013. Near-optimal experimental design for model selection in systems biology. Bioinformatics 29, 2625–2632. ( 10.1093/bioinformatics/btt436) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Silk D, Kirk PDW, Barnes CP, Toni T, Stumpf MPH. 2014. Model selection in systems biology depends on experimental design. PLoS Comput. Biol. 10, e1003650 ( 10.1371/journal.pcbi.1003650) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Cade BS. 2015. Model averaging and muddled multimodel inferences. Ecology 96, 2370–2382. ( 10.1890/14-1639.1) [DOI] [PubMed] [Google Scholar]
- 39.Box GEP. 1976. Science and statistics. J. Am. Stat. Assoc. 71, 791–799. ( 10.1080/01621459.1976.10480949) [DOI] [Google Scholar]
- 40.Camacho DM, Collins KM, Powers RK, Costello JC, Collins JJ. 2018. Next-generation machine learning for biological networks. Cell 173, 1581–1592. ( 10.1016/j.cell.2018.05.015) [DOI] [PubMed] [Google Scholar]
- 41.Baker RE, Peña JM, Jayamohan J, Jérusalem A. 2018. Mechanistic models versus machine learning, a fight worth fighting for the biological community? Biol. Lett. 14, 20170660 ( 10.1098/rsbl.2017.0660) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Meyer P. et al. 2014. Network topology and parameter estimation: from experimental design methods to gene regulatory network kinetics using a community based approach. BMC Syst. Biol. 8, 13 ( 10.1186/1752-0509-8-13) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Stumpf MPH, Thorne TW, de Silva E, Stewart R, An HJ, Lappe M, Wiuf C. 2008. Estimating the size of the human interactome. Proc. Natl Acad. Sci. USA 105, 6959–6964. ( 10.1073/pnas.0708078105) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Stumpf MPH, Wiuf C. 2010. Incomplete and noisy network data as a percolation process. J. R. Soc. Interface 7, 1411–1419. ( 10.1098/rsif.2010.0044) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Penfold CA, Wild DL. 2011. How to infer gene networks from expression profiles, revisited. Interface Focus 1, 857–870. ( 10.1098/rsfs.2011.0053) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Mc Mahon SS, Sim A, Filippi S, Johnson R, Liepe J, Smith D, Stumpf MPH. 2014. Information theory and signal transduction systems: from molecular information processing to network inference. Semin. Cell Dev. Biol. 35, 98–108. ( 10.1016/j.semcdb.2014.06.011) [DOI] [PubMed] [Google Scholar]
- 47.Arfken G, Weber HJ, Harris F. 2013. Mathematical methods for physicists. New York, NY: Academic Press. [Google Scholar]
- 48.Aijo T, Bonneau R. 2016. Biophysically motivated regulatory network inference: progress and prospects. Hum. Hered. 81, 62–77. ( 10.1159/000446614) [DOI] [PubMed] [Google Scholar]
- 49.Kirk PDW, Babtie AC, Stumpf MPH. 2015. Systems biology (un)certainties. Science 350, 386–388. ( 10.1126/science.aac9505) [DOI] [PubMed] [Google Scholar]
- 50.Chan TE, Stumpf MPH, Babtie AC. 2017. Gene regulatory network inference from single-cell data using multivariate information measures. Cell Syst. 5, 251–267.e3. ( 10.1016/j.cels.2017.08.014) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Saelens W, Cannoodt R, Todorov H, Saeys Y. 2019. A comparison of single-cell trajectory inference methods. Nat. Biotechnol. 37, 547–554. ( 10.1038/s41587-019-0071-9) [DOI] [PubMed] [Google Scholar]
- 52.Pratapa A, Jalihal AP, Law JN, Bharadwaj A, Murali TM. 2020. Benchmarking algorithms for gene regulatory network inference from single-cell transcriptomic data. Nat. Methods 17, 147–154. ( 10.1038/s41592-019-0690-6) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Diaz LP, Stumpf MPH. 2020. Gaining confidence in inferred networks. bioRxiv ( 10.1101/2020.09.19.304980) [DOI]
- 54.Madirolas G, de Polavieja GG. 2015. Improving collective estimations using resistance to social influence. PLoS Comput. Biol. 11, e1004594 ( 10.1371/journal.pcbi.1004594) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Díez-Pastor JF, Rodríguez JJ, García-Osorio CI, Kuncheva LI. 2015. Diversity techniques improve the performance of the best imbalance learning ensembles. Inf. Sci. 325, 98–117. ( 10.1016/j.ins.2015.07.025) [DOI] [Google Scholar]
- 56.Kirk PDW, Silk D, Stumpf MPH. 2015. Reverse engineering under uncertainty. In Uncertainty in biology (eds L Geris, D Gomez-Cabrero), pp. 15–32. Cham, Switzerland: Springer ( 10.1007/978-3-319-21296-8_2) [DOI]
- 57.Lèbre S, Becq J, Devaux F, Stumpf MPH, Lelandais G. 2010. Statistical inference of the time-varying structure of gene-regulation networks. BMC Syst. Biol. 4, 130 ( 10.1186/1752-0509-4-130) [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
A Jupyter Notebook containing the Julia code to reproduce all the computational results here, and to explore the effects of the madness of crowds in network inference and model averaging is available at https://github.com/MichaelPHStumpf/Madness-Of-Crowds.