Skip to main content
PLOS One logoLink to PLOS One
. 2014 Dec 31;9(12):e115806. doi: 10.1371/journal.pone.0115806

A Full Bayesian Approach for Boolean Genetic Network Inference

Shengtong Han 1, Raymond K W Wong 2, Thomas C M Lee 3, Linghao Shen 4, Shuo-Yen R Li 4,5, Xiaodan Fan 1,*
Editor: Xiaodong Cai6
PMCID: PMC4281059  PMID: 25551820

Abstract

Boolean networks are a simple but efficient model for describing gene regulatory systems. A number of algorithms have been proposed to infer Boolean networks. However, these methods do not take full consideration of the effects of noise and model uncertainty. In this paper, we propose a full Bayesian approach to infer Boolean genetic networks. Markov chain Monte Carlo algorithms are used to obtain the posterior samples of both the network structure and the related parameters. In addition to regular link addition and removal moves, which can guarantee the irreducibility of the Markov chain for traversing the whole network space, carefully constructed mixture proposals are used to improve the Markov chain Monte Carlo convergence. Both simulations and a real application on cell-cycle data show that our method is more powerful than existing methods for the inference of both the topology and logic relations of the Boolean network from observed data.

Introduction

A central focus in genomic research is to infer how genes are related to each other. Due to the complexity of real biological systems, it is essential to learn genetic networks in a holistic rather than an atomistic manner [1]. Various network models have been proposed to describe gene regulatory mechanisms, such as deterministic Boolean networks, random Boolean networks [2], probabilistic Boolean networks [3], probabilistic gene regulatory networks [4], Bayesian networks [5], [6], etc. For a review of methods for reconstructing genetic networks, see [7]. Each model has its own advantages and drawbacks. Boolean networks have the appealing characteristics of model simplicity, dynamic complexity and robustness to the noisy data. Moreover, recent research indicates that many realistic biological questions can be answered by the simple Boolean formulation, which essentially emphasizes fundamental and generic principles rather than quantitative biochemical details [8]. Biologists also traditionally prefer using ON and OFF to describe gene expression status. However, Boolean networks suffer the risk of losing useful information because of the two-state simplification for the continuous gene expression values. A detailed discussion of the prospects and limitations of Boolean genetic network models can be found in [9].

A number of algorithms have been proposed to infer Boolean genetic networks from observed data sets; [10] provided a good review of these algorithms. In [11], two popular algorithms, REVEAL [12] and Best-Fit Extension (BFE) [13], are implemented in a R package called BoolNet. REVEAL is based on exhaustive mutual information comparison, but it essentially assumes a deterministic Boolean network model. Thus it is not always able to reconstruct networks in the presence of noisy and inconsistent measurements in the input data. BFE accommodates noisy input data by minimizing the number of misclassifications. Its optimization is performed for each output node separately instead of for the whole network jointly. More recently, [14] proposed a likelihood-based approach to reconstruct Time Delay Boolean Networks (TDBN) from noisy data, but again the likelihood is maximized for each output node separately. To achieve better inference efficiency and accuracy, there is a need of new network reconstruction methods which use the optimization of a proper objective function simultaneously for the whole network. In this paper, we developed a full Bayesian Inference approach for a Boolean Network (BIBN), which is based on maximizing the joint posterior probability over the whole network. We show the new BIBN method outperforms REVEAL [12], BFE [13] and TDBN [14] through simulation. We also applied BIBN on the yeast cell-cycle data.

Materials and Methods

Model

Our method uses a probabilistic Boolean network model, where each node represents a gene with binary expression values. More specifically, we model the relations among the Inline graphic genes under study as a directed acyclic graph denoted by a set of components Inline graphic, where Inline graphic represents the set of nodes Inline graphic, Inline graphic denotes a set of Boolean functions Inline graphic, and Inline graphic represents the topology of the network, i.e., the input-output connectivity information. Here Inline graphic denotes both the node corresponding to the Inline graphic-th gene and its gene expression values. Suppose we have Inline graphic observations of the network, then Inline graphic. Each value Inline graphic is a binary variable, taking values from Inline graphic. The binary formulation corresponds to the simplification of the gene activity to either an active (ON) or inactive (OFF) state. The set of input nodes of the node Inline graphic, denoted as its parent set Inline graphic, is the set of genes which may directly affect the gene expression Inline graphic. The information about Inline graphic is derived from the topology Inline graphic. The Boolean function Inline graphic is composed of four commonly used logic operators: Inline graphic (representing AND, OR, exclusive-OR respectively) and the logic Inline graphic operation (the Inline graphic operation on Inline graphic is denoted by Inline graphic).

If Inline graphic is an empty set, it means the Inline graphic-th gene is not regulated by any other genes in the network. In this case, we call Inline graphic as a root node, and assume an independent Bernoulli distribution for it, i.e., Inline graphic and Inline graphic.

If Inline graphic is non-empty, we assume that Inline graphic is determined by Inline graphic through Inline graphic and an independent and identically distributed (i.i.d.) additive noise Inline graphic, which follows a Bernoulli distribution, i.e.:

graphic file with name pone.0115806.e035.jpg (1)

If the Inline graphic observations of the network are independent from each other, Inline graphic is determined by the Inline graphic-th observation of its parent set Inline graphic. If the Inline graphic observations of the Inline graphic genes form a synchronized time series, Inline graphic shall be determined by the Inline graphic-th observation of its parent set Inline graphic. In either case, the noise term Inline graphic of Inline graphic is assumed to be independent and identically distributed (i.i.d.) with Inline graphic and Inline graphic. For presentation convenience, we will stick to the notations as if the Inline graphic observations are independent, although our algorithm suits both cases.

Assume the network contains Inline graphic root nodes and, for notation convenience, assume the root nodes are Inline graphic. Denote Inline graphic as the set of the noise parameter Inline graphic and all of the Inline graphic root node parameters Inline graphic. We can then write down the full likelihood of the model as:

graphic file with name pone.0115806.e056.jpg (2)

Here Inline graphic represents the number of non-zero data points Inline graphic of the root node Inline graphic, and Inline graphic represents the total number of non-root data points Inline graphic which is not equal to Inline graphic. That is, Inline graphic counts the number of times that Inline graphic is equal to 1. The full likelihood is consisted of two parts. The first part is contributed by the noise and the second part is from all root nodes.

The number of input nodes of Inline graphic is referred to as the in-degree of Inline graphic. The computing complexity will inevitably increase if the in-degree increases, although the principle of our algorithm suits networks with any in-degree. Similar to BFE [13] and TDBN [14], we will focus on the case where the maximum in-degree of all nodes in the network is bounded by 2. Therefore, both the number of valid network topologies (defined now as all directed acyclic graphs of Inline graphic nodes where every node has no more than 2 input nodes) and the number of possible Boolean function types for Inline graphic are also bounded. Although this in-degree constraint is rooted in the computing scalability, it actually has biological justifications because most genes in the cell are regulated by only a very small number of genes [15][17]. It is believed that most Boolean functions require few essential variables [18] and networks where most nodes have many parents will offer little scientific insight [19].

In this paper, we are interested in inferring the network topology Inline graphic and Boolean functions Inline graphic based on Inline graphic, i.e., Inline graphic observations of the Inline graphic concerned genes.

Algorithm

To fit the above models to input data sets, we use a full Bayesian approach to take advantage of the conditionally independent nature of some random variables in the network model, to take account of the estimation uncertainty and to provide a convenient way to incorporate prior knowledge. Markov chain Monte Carlo (MCMC) algorithms will be developed to sample from the joint posterior distribution of the network topology and Boolean functions, which will provide both a point estimate and an uncertainty measure for these unknown variables.

Prior Distributions

For Bayesian inference, we need to specify the prior distributions for Inline graphic, Inline graphic and Inline graphic. If we have some prior knowledge about these unknown variables, it is an advantage of the Bayesian approach to seamlessly integrate this knowledge into the inference result. If we do not have any prior knowledge, specifying a flat prior will result in a posterior inference which is equivalent to the maximum likelihood estimation. Overall, we assume Inline graphic is independent of Inline graphic and Inline graphic in the prior distribution, i.e., Inline graphic.

For all Inline graphic and Inline graphic in Inline graphic, we assume that they follow independent Beta distributions as in [20][23]. More specifically, we assume that the noise parameter Inline graphic is sampled from Inline graphic, and all root parameters Inline graphic are independently sampled from Inline graphic. The parametric form of Beta distribution will make the computation more convenient since it is the conjugate prior for the likelihood. The hyper-parameters Inline graphic and Inline graphic are chosen constants. Since we know little about Inline graphic, we can set Inline graphic and Inline graphic as 1, which will result in a flat prior distribution. As the noise rate Inline graphic should not be too big, we set Inline graphic to be smaller than Inline graphic.

Let Inline graphic denote the total number of valid network topology as defined before. We use uniform prior for Inline graphic, i.e., Inline graphic.

As for Inline graphic, the actual number of possible function forms for Inline graphic is dependent on the topology Inline graphic and is no more than 16 if the maximum in-degree is 2. For its prior, we assume that Inline graphic are independent of each other conditional on Inline graphic and Inline graphic is sampled uniformly from all possible non-degenerative Boolean functions of Inline graphic. For example, if Inline graphic is the set Inline graphic, Inline graphic can be either Inline graphic or Inline graphic; if Inline graphic is the set Inline graphic, Inline graphic has 10 non-degenerative choices: Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic.

Posterior Distributions

From the above prior distributions and the full likelihood, it is straightforward to derive the following joint posterior distribution:

graphic file with name pone.0115806.e124.jpg
graphic file with name pone.0115806.e125.jpg (3)

Since the number of root nodes is unknown and is determined by the topology Inline graphic, the dimension of Inline graphic may change once we change the topology. Thus, if we use an MCMC algorithm to directly sample from the above joint posterior distribution, we have to deal with the trans-dimensional problem. Although theoretically some algorithms, such as reversible jump MCMC [24], can be used to handle this problem, the convergence speed of such MCMC algorithms is still problematic. To circumvent this problem, we analytically integrate out all Inline graphic and Inline graphic from the above posterior distribution, which results in the following collapsed version of the posterior distribution:

graphic file with name pone.0115806.e130.jpg
graphic file with name pone.0115806.e131.jpg (4)

We have designed an MCMC algorithm to sample from Inline graphic, which avoids the dimension change caused by Inline graphic. More specifically, we update Inline graphic iteratively for all Inline graphic with Metropolis-Hastings (MH) algorithms. If we are also interested in estimating Inline graphic, we can subsequently estimate Inline graphic from Inline graphic after we obtain the posterior estimates Inline graphic and Inline graphic.

Constructing Efficient Proposal Distributions for MH algorithms

One major concern of using MCMC algorithms to sample from complicated distributions, such as the posterior network topology space, is the convergence rate, which will determine the computing time to achieve a stationary sample of a desired effective sample size. For the MH algorithm which we will use to sample from Inline graphic, a good proposal distribution is the key for its sampling efficiency. We will first use the Inline graphic goodness-of-fit test to pick out well-fitted parent sets and corresponding functions for each node as preferential candidates, then construct a node-specific proposal distribution as a mixture of random-walk and weighted sampling from the preferential candidates. These proposal distributions will not change the stationary distribution of the MCMC chain, but it will improve the mixing of the Markov chain by placing more effect on more likely regions of the parameter space.

The Inline graphic goodness-of-fit test to check how well a combination Inline graphic fits the data of Inline graphic goes as follow. Without loss of generality, considering the two-parent case with Inline graphic and the OR function Inline graphic. There are 4 possible values for Inline graphic, i.e., Inline graphic, Inline graphic, Inline graphic, Inline graphic. Denote the probabilities of the 4 values as Inline graphic, which satisfy Inline graphic. According to the model in Equation 1, the probabilities of the 8 possible values of Inline graphic are listed in Table 1, where all unknown parameters will be estimated from the data of Inline graphic. The Inline graphic goodness-of-fit test is then used to test whether the observed frequencies of the 8 possible values fit the distribution in Table 1. If fitting, the combination Inline graphic is called a preferential candidate for Inline graphic. The reciprocal of the noise level estimate Inline graphic will be used to weigh the preferential candidate.

Table 1. The theoretical distribution of Inline graphic for the relation Inline graphic.
Inline graphic 0 0 0 0 1 1 1 1
Inline graphic 0 0 1 1 0 0 1 1
Inline graphic 0 1 0 1 0 1 0 1
Probability Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic

There are Inline graphic and Inline graphic possible choices for Inline graphic, 2 and 10 possible choices for Inline graphic, in the case of one parent and two parents, respectively. All possible parent and function combinations are tested in the similar way one by one. The resulted preferential candidates and their associated weights are used to construct two multinomial distributions, one for the one-parent case and one for the two-parent case, which are called the preferential distributions of the node Inline graphic. Two uniform distributions are constructed for the node Inline graphic by assigning equal weights to all of its possible parent and function combinations in the case of one parent and two parents, separately. The proposal distribution for updating Inline graphic in the case of a given number of parents is the mixture distribution of the corresponding preferential distribution and the corresponding uniform distribution of the node Inline graphic, with the mixing proportion of preferential distribution gradually reducing from one to a selected percentage. This proposal constructing procedure is applied to each node.

The MCMC Algorithm

The general MCMC framework will be the Metropolis-within-Gibbs algorithm, which starts with initial values of Inline graphic and Inline graphic, and iteratively updates them from their conditional posterior distributions until the chain is converged.

Updating network topology refers to link addition and removal between nodes, which is equivalent to changing nodes' parent sets. There are three types of MCMC moves to update the parent sets: adding parent(s), removing parent(s) and swapping parent(s). We call one move as legal if it results in a valid network topology as defined previously. For instance, for a node currently without any input node, there may be 2 legal moves, i.e., adding one parent and adding two parents. But if adding parent(s) leads to a cyclic graph, that specific move is illegal.

Once the topology Inline graphic changes, the associated Boolean function Inline graphic will also have to change. We sequentially and iteratively update each node's parent set Inline graphic and associated function Inline graphic through a MH algorithm using the proposal distributions constructed in the previous subsection.

Results

Simulation Studies

Simulation studies are performed to validate our method and compare with existing methods. We synthesized data sets for networks with 20 nodes. For each data set, we first randomly generated a valid network topology Inline graphic. This step proceeds as follows. For each node, we selected the number of its parent from Inline graphic, with probabilities with sum of 1. Once this number is determined, we chose the parents from the remaining nodes at random. This operation is applied to each node, which results in a full network candidate. Finally we checked the validity of the resulting network by checking whether there are directed loops. This network is used in the subsequent step if it passes the validity checking. Otherwise we repeated this process till a valid network topology is obtained. Once Inline graphic is known, we then randomly assigned a Boolean function to each node from all possible candidate functions, depending on its parent set. Thus we generated Inline graphic. For Inline graphic, we randomly sampled these probability parameters from their prior distributions. Finally, with the generated Inline graphic, we applied Equation 1 to generate Inline graphic observations of the network Inline graphic. Since our model covers all possible boolean relationships with in-degree up to 2, the simulated data should be general enough for a fair comparison among BFE, REVEAL and TDBN.

To measure the inference accuracy, we define the correct rate (CR) as the percentage of the Inline graphic nodes whose parent sets and associated functions are both correctly identified as compared to the truth. Hence CR = 1 if and only if the inferred network indexed by Inline graphic is the same as the true model.

To test BIBN, we synthesized different data sets with varying settings. The sample sizes tested include 50, 100, 300, and 500. The noise levels at 0.1 and 0.2 are considered. For each sample size and noise level combination, 20 different data sets corresponding to 20 different networks are generated. For each data set, a Markov chain is run with a total of 20,000 iterations. The first 15,000 iterations are treated as burn-in and the last 5,000 iterations are collected to calculate the average accuracy for a single chain. We averaged the 20 accuracies to obtain the final average accuracy for a specific sample size and noise level combination.

For comparison, we chose REVEAL and BFE which are two popular inference algorithms for Boolean network inference, and TDBN which is a recently developed method for reconstructing Boolean networks. Both REVEAL and BFE are implemented in the R package BoolNet [11]. The code of TDBN is from the author of [14]. The same data sets are inputted into REVEAL, BFE, TDBN and BIBN to obtain their inference accuracies. The results are summarized in Table 2. REVEAL is not listed in this table because its performance is very poor due to its low capability to handle nondeterministic network models. BFE and TDBN have a better tolerance of noise compared to REVEAL, but they are poor in pursuing the global optimization of the full network, thus resulting in lower correct rates. Obviously, Table 2 shows that our method outperformed all other methods for all settings. Generally speaking, when fixing the sample size, increasing noise level will deteriorate the inference accuracy. One can improve the accuracy by increasing the sample size when the noise level can not be reduced.

Table 2. Average accuracy comparisons on the synthesized data.

Inline graphic Inline graphic
Sample Size BIBN BFE TDBN BIBN BFE TDBN
10 0.1827 0.1725 0.1750 0.0809 0.0375 0.1425
50 0.8599 0.6975 0.4175 0.6858 0.5575 0.3300
100 0.9565 0.7425 0.4900 0.8864 0.7375 0.4375
300 0.9951 0.8575 0.7700 0.9358 0.8350 0.6800
500 1.0000 0.8775 0.8125 0.9975 0.8725 0.7825

It should be noted that TDBN calculated p values for all possible transition relations. We selected their most likely one to calculate the correct rate for comparison.

To further evaluate the proposed method, we also checked the prediction power of BIBN, with the results summarized in Table 3. In each scenario, we generated an observed sample as described before. Then we randomly chose 2/3 of the sample to perform the inference as we presented before, and the remaining 1/3 of the sample to test the prediction accuracy. More specifically, for each inferred network, we predicted the value of each child node using the observed values of its parents, then checked whether the predicted and observed values of the child are the same. The percentage of correct prediction over the 1/3 sample is treated as the prediction accuracy of this child node. The average prediction accuracy over all child nodes is treated as the prediction accuracy of whole network. This is done for the inferred network at each iteration after the burn-in period. The average prediction accuracy of these networks is treated as the prediction accuracy for this chain. This procedure is repeated independently for ten times for each scenario on Table 3. The correct prediction rate reported under each scenario in Table 3 is the average over the ten repetitions. It shows that BIBN has good prediction accuracy. Given the sample size, the correct prediction rate decreases as the noise level increases. With the noise level fixed, the correct prediction rate is improving as the sample size grows, which is as expected.

Table 3. Correct prediction rate of BIBN under difference scenarios.

Sample Size Inline graphic Inline graphic
75 0.8877 0.7813
150 0.8929 0.7982
450 0.8977 0.8050
750 0.9149 0.8073

Real Data Analysis

Cell-Cycle Gene Expression Data

The cell cycle is the biological process by which one cell grows and divides into two daughter cells. Due to its fundamental importance in cell biology, it has been studied extensively in various model organisms [25][28]. But due to its complexity, the complete composition and regulatory mechanisms of the cell-cycle gene network is still unclear for most eukaryotes.

Some studies indicate that components may vary over a long evolutionary distance [29]. However, most key components and their interactions are conserved [30][32]. With the cumulated gene expression data for yeast, we target at inferring the relationships among the key genes in yeast cell cycle.

Similar to the cell-cycle network used in [27], we study 14 key cell-cycle genes, including Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic, Inline graphic and Inline graphic. The real gene expression data can be downloaded from Inline graphic. It contains the normalized data from 500 yeast microarray experiments under various conditions, including stress responses, cell-cycle synchronization, sporulation, etc. Missing values in the downloaded data are deleted since our current method only handles complete data. To transform the data into binary values, values that are higher than the corresponding gene's mean value are set to 1. Otherwise they are set to 0.

Network Inference Result

For the transformed binary data set of 14 genes, we ran three independent Markov chains using three different initial networks, which include the empty network without any links, one randomly generated valid network and a valid network constructed from the preferential candidates. Each chain is run for 14,000 iterations. The trace plot of the unnormalized log-posterior probabilities for these three chains are displayed in Fig. 1. It shows that the chains converged after about 10,000 iterations. Thus, the network samples within the last 4,000 iterations are used for posterior inference.

Figure 1. Trace plots of the unnormalized log-posterior probability of the Markov chain for real cell-cycle data.

Figure 1

Each line represents an independent Markov chain. Each chain is run for 14,000 iterations.

It turns out that the last 4,000 iterations contain 43 unique network models. A total of 12.82% of the links in the reference yeast cell cycle network reported in [27] are identified in 100% of the posterior samples. For instance, the relation Inline graphic has a probability of over Inline graphic of being inferred correctly. The “coupled” gene pairs in [27], such as Inline graphic, Inline graphic and Inline graphic, are correctly linked together in most of the posterior samples. Other correctly inferred relations also have a high show-up frequency in the posterior samples.

We also applied REVEAL, BFE and TDBN to this real data. The read data is too noisy for REVEAL and BFE to produce anything. While the accuracy of TDBN is 5.13%, which is much lower than that of BIBN. This comparison on real biological data clearly showed the advantage of our method, but it has be to admitted that we still need to improve our method in order to meet the accuracy requirement of real gene network inference. Future works shall check whether the boolean formulation is sufficient and whether the number of parents is not small for real biological network.

Discussion

In this paper, we propose a new method for inferring the Boolean network from noisy data using a probabilistic model and an MCMC algorithm. Our inference focuses not only on the network structure but also on the transition functions associated with the network of interest. Compared to other inference algorithms, our method has the advantage of taking both random noise and model uncertainty into consideration, which is verified by the consistently higher inference accuracy for networks with varying sample size and noise levels in the simulation study. Furthermore, a data-based proposal is constructed using a Inline graphic goodness-of-fit test for guiding the proposal of new local topology and function relations. Since the search space of networks is so large, especially for networks with many nodes, the use of carefully chosen proposals greatly improves the inference efficiency in terms of the fewer iterations needed to reach the convergence of the chain. Currently our algorithm, which is implemented in R and run on a 2.66 GHz CPU, takes about 1.6 hours to run 20,000 iterations when the sample size is 50, and 1.9 hours when the sample size is 500.

It should be noted that our method also has some limitations. One is the assumption that each node has at most two parents, which may limit its wide application in practice. In principle, the method can be extended to deal with networks with more than 2 parents for each node without further technical difficulties. However, the computational requirements of the method would increase significantly and there is a danger to overfit the data. Another shortcoming of our method is to assume the model to be a directed acyclic graph in order to use the Bayesian network framework [33]. Regulatory networks are known to contain feedback loops, thus our inference shall be considered as a preliminary step. Future research can extend our model on the line of dynamic Bayesian network in order to model loops [34]. Also, since our method is based on Boolean values, genes with more than two expressing status or gene relations may not be correctly modeled here. The method for discretizing gene expression values is also a very important issue and deserves the exploration of a separate paper [35]. In terms of future enhancement, techniques for MCMC algorithms to avoid trapping in local modes can be added.

Acknowledgments

We thank three anonymous reviewers and the academic editor for their very helpful comments.

Data Availability

The authors confirm that all data underlying the findings are fully available without restriction. Data have been deposited to Figshare (http://dx.doi.org/10.6084/m9.figshare.1255005).

Funding Statement

This research is partially supported by a grant from the Research Grants Council of the Hong Kong SAR (Project no. CUHK 400913), a CUHK direct grant (Project no. CUHK 2060419), a grant from the University Grants Committee of the Hong Kong Special Administrative Region, China (Project No. AoE/E-02/08), three grants from the National Science Foundation of USA (Grant No. 1007520, 1209226, and 1209232), and a grant from the National Basic Research Program of China (973 Program, No. 2012CB315901, No. 2012CB315904). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Lähdesmäki H, Shmulevich I, Yli-Harja O (2003) On learning gene regulatory networks under the Boolean network model. Machine Learning 52:147–167. [Google Scholar]
  • 2. Kauffman SA (1969) Metabolic stability and epigenesis in randomly constructed genetic nets. Journal of Theoretical Biology 22:437–467. [DOI] [PubMed] [Google Scholar]
  • 3. Shmulevich I, Dougherty ER, Kim S, Zhang W (2002) Probabilistic Boolean networks: a rule-based uncertainty model for gene regulatory networks. Bioinformatics 18:261–274. [DOI] [PubMed] [Google Scholar]
  • 4. Zhou X, Wang X, Pal R, Ivanov I, Bittner M, et al. (2004) A Bayesian connectivity-based approach to constructing probabilistic gene regulatory networks. Bioinformatics 20:2918–2927. [DOI] [PubMed] [Google Scholar]
  • 5.Murphy K, Mian S (1999) Modelling gene expression data using dynamic Bayesian networks. Technical report, Berkeley.
  • 6.Hartemink AJ, Gifford DK, Jaakkola TS, Young RA (2001) Using graphical models and genomic expression data to statistically validate models of genetic regulatory networks. Pac Symp Biocomput: 422–433. [DOI] [PubMed]
  • 7. Markowetz F, Spang R (2007) Inferring cellular networks - a review. BMC Bioinformatics 8:S5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Huang S (1999) Gene expression profiling, genetic networks, and cellular states: an integrating concept for tumorigenesis and drug discovery. J Mol Med (Berl) 77:469–480. [DOI] [PubMed] [Google Scholar]
  • 9. Bornholdt S (2008) Boolean network models of cellular regulation: prospects and limitations. Journal of The Royal Society Interface 5:S85–S94. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Akutsu T, Hayashida M, Tamura T (2008) Algorithms for inference, analysis and control of Boolean networks. In: Proceedings of the 3rd international conference on Algebraic Biology. Berlin, Heidelberg: Springer-Verlag, AB '08, pp.1–15.
  • 11. Müssel C, Hopfensitz M, Kestler HA (2010) BoolNet – an R package for generation, reconstruction and analysis of Boolean networks. Bioinformatics 26:1378–1380. [DOI] [PubMed] [Google Scholar]
  • 12.Liang S, Fuhrman S, Somogyi R (1998) Reveal, a general reverse engineering algorithm for inference of genetic network architectures. Pac Symp Biocomput: 18–29. [PubMed]
  • 13. Boros E, Ibaraki T, Makino K, Makino K (1998) Error-free and best-fit extensions of partially defined Boolean functions. Information and Computation 140:254–283. [Google Scholar]
  • 14. Chueh TH, Lu HHS (2012) Inference of biological pathway from gene expression profiles by time delay Boolean networks. PLoS ONE 7:e42095. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Luscombe NM, Qian J, Zhang Z, Johnson T, Gerstein M (2002) The dominance of the population by a selected few: power-law behaviour applies to a wide variety of genomic properties. Genome Biol 3:RESEARCH0040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Rzhetsky A, Gomez SM (2001) Birth of scale-free molecular networks and the number of distinct dna and protein domains per genome. Bioinformatics 17:988–996. [DOI] [PubMed] [Google Scholar]
  • 17. Wagner A (2002) Estimating coarse gene network structure from large-scale gene perturbation data. Genome Research 12:309–315. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Kauffman SA (1993) The Origins of Order: Self-Organization and Selection in Evolution. New York: Oxford University Press.
  • 19. Ellis B, Wong WH (2008) Learning causal Bayesian network structures from experimental data. Journal of the American Statistical Association 103:778–789. [Google Scholar]
  • 20.Buntine W (1991) Theory refinement on Bayesian networks. San Mateo, CA: Morgan Kaufmann, pp.52–60.
  • 21. Cooper GF, Herskovits E (1992) A Bayesian method for the induction of probabilistic networks from data. Machine Learning 9:309–347. [Google Scholar]
  • 22. Dawid AP, Lauritzen SL (1993) Hyper Markov laws in the statistical analysis of decomposable graphical models. The Annals of Statistics 21:1272–1317. [Google Scholar]
  • 23. Heckerman D, Chickering DM (1995) Learning Bayesian networks: The combination of knowledge and statistical data. Machine Learning 20:197–243. [Google Scholar]
  • 24. Green PJ (1995) Reversible jump Markov chain Monte Carlo computation and Bayesian model determination. Biometrika 82:711–732. [Google Scholar]
  • 25. Cross FR, Archambault V, Miller M, Klovstad M (2002) Testing a mathematical model of the yeast cell cycle. Mol Biol Cell 13:52–70. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Koch C, Moll T, Neuberg M, Ahorn H, Nasmyth K (1993) A role for the transcription factors Mbp1 and Swi4 in progression from g1 to s phase. Science 261:1551–1557. [DOI] [PubMed] [Google Scholar]
  • 27. Li F, Long T, Lu Y, Ouyang Q, Tang C (2004) The yeast cell-cycle network is robustly designed. Proceedings of the National Academy of Sciences of the United States of America 101:4781–4786. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Stoll G, Rougemont J, Naef F (2006) Few crucial links assure checkpoint efficiency in the yeast cell-cycle network. Bioinformatics 22:2539–2546. [DOI] [PubMed] [Google Scholar]
  • 29. Bähler J (2005) Cell-cycle control of gene expression in budding and fission yeast. Annu Rev Genet 39:69–94. [DOI] [PubMed] [Google Scholar]
  • 30. Rustici G, Mata J, Kivinen K, Li P, Penkett CJ, et al. (2004) Periodic gene expression program of the fission yeast cell cycle. Nat Genet 36:809–817. [DOI] [PubMed] [Google Scholar]
  • 31. Peng X, Karuturi RKM, Miller LD, Lin K, Jia Y, et al. (2005) Identification of cell cycle-regulated genes in fission yeast. Mol Biol Cell 16:1026–1042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Oliva A, Rosebrock A, Ferrezuelo F, Pyne S, Chen H, et al. (2005) The cell cycle-regulated genes of Schizosaccharomyces pombe. PLoS Biol 3:e225. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Markowetz F, Spang R (2007) Inferring cellular networks - a review. BMC Bioinformatics 8:S5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Shojaie A, Jauhiainen A, Kallitsis M, Michailidis G (2014) Inferring regulatory networks by combining perturbation screens and steady state gene expression profiles. PLoS ONE 9:e82393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Berestovsky N, Nakhleh L (2013) An evaluation of methods for inferring Boolean networks from time-series data. PLoS ONE 8:e66031. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The authors confirm that all data underlying the findings are fully available without restriction. Data have been deposited to Figshare (http://dx.doi.org/10.6084/m9.figshare.1255005).


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES