Skip to main content
Springer logoLink to Springer
. 2016 Jul 14;78:1520–1545. doi: 10.1007/s11538-016-0193-x

Logical Reduction of Biological Networks to Their Most Determinative Components

Mihaela T Matache 1,, Valentin Matache 1
PMCID: PMC4993808  PMID: 27417985

Abstract

Boolean networks have been widely used as models for gene regulatory networks, signal transduction networks, or neural networks, among many others. One of the main difficulties in analyzing the dynamics of a Boolean network and its sensitivity to perturbations or mutations is the fact that it grows exponentially with the number of nodes. Therefore, various approaches for simplifying the computations and reducing the network to a subset of relevant nodes have been proposed in the past few years. We consider a recently introduced method for reducing a Boolean network to its most determinative nodes that yield the highest information gain. The determinative power of a node is obtained by a summation of all mutual information quantities over all nodes having the chosen node as a common input, thus representing a measure of information gain obtained by the knowledge of the node under consideration. The determinative power of nodes has been considered in the literature under the assumption that the inputs are independent in which case one can use the Bahadur orthonormal basis. In this article, we relax that assumption and use a standard orthonormal basis instead. We use techniques of Hilbert space operators and harmonic analysis to generate formulas for the sensitivity to perturbations of nodes, quantified by the notions of influence, average sensitivity, and strength. Since we work on finite-dimensional spaces, our formulas and estimates can be and are formulated in plain matrix algebra terminology. We analyze the determinative power of nodes for a Boolean model of a signal transduction network of a generic fibroblast cell. We also show the similarities and differences induced by the alternative complete orthonormal basis used. Among the similarities, we mention the fact that the knowledge of the states of the most determinative nodes reduces the entropy or uncertainty of the overall network significantly. In a special case, we obtain a stronger result than in previous works, showing that a large information gain from a set of input nodes generates increased sensitivity to perturbations of those inputs.

Electronic supplementary material

The online version of this article (doi:10.1007/s11538-016-0193-x) contains supplementary material, which is available to authorized users.

Keywords: Boolean networks, Biological information theory, Mutual information, Sensitivity, Network reduction, Linear operators, Numerical simulations

Introduction

The past few decades have generated a large influx of data and information regarding a variety of real or artificial networks. The necessity to understand and use the data in a meaningful way has lead to various modeling approaches. In particular, Boolean network (BN) models introduced by Kauffman (1969) have acquired a significant importance in modeling networks where the node activity can be described by two states, 1 and 0, “ON and OFF,” “active and nonactive,” and where each node is updated based on logical relationships with other nodes. These models incorporate Boolean functions that are relevant to particular types of applications, such as signal transduction in cells (e.g., Helikar et al. 2008), genetic regulatory networks or other biological networks (e.g., Kauffman 1993; Shmulevich and Kauffman 2004; Shmulevich et al. 2002; Klemm and Bornholdt 2000; Albert and Othmer 2003), or neural networks (e.g., Huepe and Aldana 2002).

However, a simplification of the reality to binary states of the nodes does not ease the difficulty of studying large, complex networks for which the existing data may offer only partial information on the real interactions in the network and for which the dynamics are hard to study even under a deterministic approach. As a matter of fact, even smaller networks of only a few hundred nodes or less can pose serious difficulties in assessing the dynamics, given the exponential dependence of the state space on the number of nodes. Consequently, a number of approaches aiming at simplifying the computational difficulty of analyzing the dynamics have been proposed in recent years. For example, Goles et al. (2013) reduce the network and the number of updates needed to analyze the dynamics by generating sub-configurations of states that remain fixed regardless of the values of the other nodes and by identifying sets of updating schedules which have the same dynamics. They show that such networks are minimal in the sense that no edge can be deleted because every one of them represents a real interaction in the respective network. Various methods for reducing the network to a fairly small subset of nodes that are relevant for the long-term dynamics have been proposed. The definitions of “relevant” and “irrelevant” nodes differ depending on the actual approach. Some methods are related to eliminating stable nodes that end up in an attractor after a transient period and thus considered irrelevant. This may be paired with removing leaf nodes that do not contribute to the evolution of any other node, that is, with zero out-degree (outputs) like in Bilke and Sjunnesson (2001) or Richardson (2004), or with merging or collapsing mediator nodes with one in-degree (input) and one out-degree in Saadatpour et al. (2013). Other methods are based on eliminating irrelevant nodes that are frozen at the same value on every attractor together with nodes whose outputs go only to irrelevant nodes in Socolar and Kauffman (2003), Kaufman et al. (2005), or Kaufman and Drossel (2006). The basis of these methods is to reduce the “structure” of the network using some rules on the Boolean functions and then prove that such a reduction simplifies the identification of attractors. As expected, they may carry an intrinsic numerical burden. Furthermore, alternative methods for eliminating the need for complete enumeration of the states have been considered. For example,  Devloo et al. (2003) propose another formalism which permits calculation of the steady states as solutions of a system of steady-state equations, via an image function which identifies a state by its image under a certain function. The authors use constraint programming to solve the derived system of equations in an efficient way from a computational point of view.

In addition to the previously mentioned methods for network reduction, the entropy of the relevant components of the network which are comprised of relevant nodes that eventually influence each other’s state is used as a measure of uncertainty of the future behavior of a random state of the network by Krawitz and Shmulevich (2007a, (2007b). The entropy is a measure of uncertainty that has been used also by Ribeiro et al. (2008) to find the average mutual information of a random Boolean model of regulatory network as a way to quantify the efficiency of information propagation through the entire network. In this context, one needs to consider pairs of connected nodes and the intrinsic Boolean functions that govern the node updates, as opposed to evolving the networks in order to identify the attractors. Further research by some of the authors of Ribeiro et al. (2008), in particular Lloyd-Price et al. (2012), uses mutual information to test for a relationship between the robustness to perturbations of an attractor in a random BN and the amount of information propagated within the network when in that attractor. They found that there is a trade-off between robustness and information propagation and that at the edge of chaos, robustness is not correlated with information propagation.

On the other hand, the notions of entropy and mutual information have been long used as measures of complexity of dynamical systems, such as BNs, as described, for example, by Feldman and Crutchfield (1998) or by Sole and Luque (1997). Luque and Ferrera are concerned with the mutual information contained in random BNs and its behavior as the networks undergo their order–disorder phase transition, showing that the mutual information stored in the network has a maximum at the transition point.

Only recently the mutual information has been used as a method for identifying the most powerful and therefore relevant nodes in a BN, thus offering an efficient alternative approach to network reduction to a relevant subset of nodes Heckel et al. (2013). The mutual information, as a basic concept in information theory, allows one to represent the reduction in the uncertainty or entropy of the state of a node due to the knowledge of any of its inputs. A summation of all mutual information quantities over all nodes having a common input can be viewed as the determinative power of that input node. The more powerful the node, the more the information gain provided by the knowledge of its state. In Heckel et al. (2013), the authors use harmonic analysis to compare the determinative power of a set of inputs to the sensitivity to perturbations to those inputs showing that an input with large sensitivity need not have a large determinative power. On the other hand, large information gain from a set of inputs generates large sensitivity to perturbations of those inputs. Moreover, by considering the feedforward regulatory network of E. coli, it is shown that the knowledge of the states of the most determinative nodes reduces the uncertainty of the overall network significantly. Thus, one could focus on the dynamics of the reduced network of the nodes with the most determinative power.

In Heckel et al. (2013), the mutual information formula is obtained in terms of Fourier coefficients expressed in the Bahadur basis which assumes independence of the inputs of a Boolean function. In a subsequent paper by Klotz et al. (2014), it is shown that canalizing Boolean functions maximize the mutual information under the same assumption as in Heckel et al. (2013). This assumption is strong, since in a BN, there are correlations between inputs that build up as the network evolves in time. Our goal is to relax this assumption and allow dependence of inputs, while exploring the impact of a (necessarily) different basis on the results regarding the mutual information and the sensitivity to perturbations. We notice that some results still hold; however, not all are independent of the basis. At the same time we are interested to see the impact of our approach on the network reduction based on most determinative nodes of a specific biological network. In particular, we use a Boolean model of the signal transduction network of a generic fibroblast cell and we obtain results similar to Heckel et al. (2013).

In Sect. 2, we provide the basic definitions, the mathematical setup, and we use elements of operator theory to generate formulas for finding the sensitivity to perturbations of the nodes of the network, quantified by the concepts of influence, average sensitivity, and strength (to be defined in that section). We also discuss the computational aspects of using those formulas in applications. In Sect. 3, we provide formulas and estimates for mutual information, determinative power, and strength, paired with simulations, and estimates that link the mutual information and the sensitivity to perturbations. We also consider a special case that allows us to compare our analytical results to those in Heckel et al. (2013). Conclusions and further directions of research are in Sect. 4.

Influence, Sensitivity, and Strength

In this section, we provide analytical formulas for the sensitivity to perturbations using a complete orthonormal basis that does not assume independence of the Boolean inputs. We pair this with some computational aspects regarding the application of the formulas to an actual biological network.

Analytical Approach

Let Ωn={0,1}n and the random vector X valued in Ωn. If P denotes the probability measure on the domain of definition of X, then denote PX-1 the (cumulative) distribution of X. A BN is modeled as the set [n]:={1,2,,n} of n nodes, each node being ON (that is in state 1) or OFF (that is in state 0). Then any ωΩn is a possible state of the network. Each node i[n] has an associated Boolean function fi:ΩnΩ that governs the dynamics of the node. We are usually interested in how the network evolves by iterating the map F=(f1,f2,,fn) a large number of times. Although the measures used in this paper are discrete, we use notation typical for measure theoretical arguments to write shorter and more elegant proofs.

Given a node i with Boolean function fi:ΩnΩ, the influence of the jth input on fi has been formulated in various ways in the literature. Following the authors of Shmulevich and Kauffman (2004), Kahn et al. (1988), and Ghanbarnejad and Klemm (2012), we recall the following:

Definition 1

The influence, Ij(fi), of variable xj on the Boolean function fi, or equivalently, the activity of node j on node i, is defined as follows:

Ij(fi):=P(fi(X)fi(Xej)) 1

where Xej is the random vector obtained by flipping the jth slot of X from 1 to 0 or viceversa. The average sensitivity, avs(fi), of fi is the sum of its incoming activities

avs(fi):=j=1nIj(fi). 2

The strength, σ(fi), of fi is the sum of the outgoing activities

σ(fi):=j=1nIi(fj). 3

Alternatively, Ij(fi) can be regarded as the average of the Boolean partial derivative (j)fi(X)=δfi(X),fi(Xej) with respect to the probability measure P as specified in Ghanbarnejad and Klemm (2012). Here δ is Kronecker’s delta function which is equal to one if the two variables are equal and zero otherwise. The definition is originally introduced in the context of assuming the state of the BN as a random vector X=(X1,,Xn) with independently distributed coordinates, but this property plays no role in it. However, the aforementioned property plays an essential role in papers like Heckel et al. (2013), Klotz et al. (2014), or Kahn et al. (1988) where it is essential in obtaining formulas for Ij(fi) (and other related quantities relevant in the study of BNs) in terms of the Fourier coefficients of fi [see, e.g., (Kahn et al. 1988, Lemma 4.1)]. The Hilbert space where those formulas are obtained is L2(Ωn,dPX-1). The complete orthonormal basis used in Kahn et al. (1988) is the, so-called, Bahadur basis (see Heckel et al. 2013). In order for that family of functions to form a complete orthonormal basis of L2(Ωn,dPX-1) it is necessary that X1,,Xn be independently distributed. However, the nodes of a BN may mutually influence each other, so independence is a restrictive assumption.

A composition operator is an operator on a linear space S of functions defined on a set E. For any fixed self-map φ of E, the operator

Cφf:=fφfS

is called the composition operator with symbol φ or induced by φ. It is necessarily linear. We will use such operators on S=L2(Ωn,dPX-1).

Let φj be the jth slot flip map. This means that, for all ωΩn,φj(ω) is the Boolean vector in Ωn obtained by flipping the jth coordinate of ω. Observe that φj are obviously self-inverse and hence so are the composition operators they induce, a fact which will be used without further comments throughout this paper. In the following, , denotes the usual inner product of L2(Ωn,dPX-1) and the norm induced by that inner product. Also, T is the notation used for the adjoint of any operator T, whereas I denotes the identity operator. With these notations we prove:

Proposition 1

For all Boolean functions fi, the following formulas hold:

Ij(fi)=(I-Cφj)(I-Cφj)fi,fij=1,,n 4
avs(fi)=Tfi,fiwhereT=j=1n(I-Cφj)(I-Cφj) 5
σ(fi)=j=1nTifj,fjwhereTi=(I-Cφi)(I-Cφi) 6

Proof

Using a well-known change in measure formula, for any Boolean function f (we drop the index for simplicity of notation) one can write

Ij(f)=P({f(X)f(Xej)})=|f(X)-f(Xej)|2dP=|f(X)-fφj(X)|2dP=Ωn|f-fφj|2dPX-1=(I-Cφj)f2=(I-Cφj)(I-Cφj)f,f.

Proposition 2

For all j=1,2,,n, let Λj denote the largest eigenvalue of Tj=(I-Cφj)(I-Cφj), respectively, whereas Λ is the largest eigenvalue of T. The following estimates hold:

Ij(f)ΛjE[f(X)] 7
avs(f)ΛE[f(X)]. 8

Hence:

σ(fi)ΛiE[F(X)],i[n], 9

where F=j=1nfj.

Proof

Indeed, both the operators Tj and T are nonnegative operators, and hence, their numerical range is equal to the line interval with endpoints the least, respectively, the largest eigenvalue. On the other hand, formulas (4) and (5) show that Ij(f)/f2 and avs(f)/f2 belong to the numerical range of the operator Tj, respectively, T. Combining all that with the fact that f, being a Boolean function, satisfies condition E[f(X)]=f2, proves (7) and (8).

For arbitrary fixed i[n] one has by (7) that

σ(fi)Λij=1nE[fj(X)].

Given that obviously j=1nE[fj(X)]=E[F(X)], estimate (9) follows.

The space L2(Ωn,dPX-1) has a simple and natural complete orthonormal basis, namely B={eω=χω/pω:ωΩn}, where

χω(x)=δω,xxΩn

and

pω=P(X=ω)ωΩn.

We assume all states are possible, so pω>0.

Checking the fact that B is a complete orthonormal basis of L2(Ωn,dPX-1), whether X1,,Xn are independently distributed or not is left to the reader.

Proposition 3

For all j=1,,n, the operator Tj=(I-Cφj)(I-Cφj) has a matrix with respect to B whose entries are

aω,η(Tj)=δω,η-pφj(ω)pωδφj(ω),η-pφj(η)pηδφj(η),ω+pφj(ω)pφj(η)pωpηδφj(ω),φj(η)ω,ηΩn. 10

Hence, the entries in the matrix of the operator T are:

aω,η(T)=j=1naω,η(Tj)ωΩn. 11

Proof

Given ω,ηΩn, the entry aω,η(Tj) in the matrix of (I-Cφj)(I-Cφj) is

aω,η(Tj)=(I-Cφj)(I-Cφj)eη,eω=(I-Cφj)eη,(I-Cφj)eω. 12

Note that

Cφjχν=χφj(ν)νΩn,

and hence

Cφjeν=pφj(ν)pνeννΩn. 13

Equalities (12) and (13) combine into establishing by a straightforward computation Eq. (10).

Therefore, one can state the following:

Corollary 1

Given a Boolean function f, the following practical formulas hold:

Ij(f)=ω,ηΩnaω,η(Tj)f(ω)f(η)pωpη 14

and

avs(f)=ω,ηΩnaω,η(T)f(ω)f(η)pωpη. 15

Hence:

σ(fi)=j=1nω,ηΩnaω,η(Ti)fj(ω)fj(η)pωpη,i[n]. 16

Indeed, (14) and (15) are immediate consequences of the matricial description of operators Tj and T combined with the following computation of the Fourier coefficients cω,ωΩn of f relative to B:

cω=f,eω=1pωΩnfχωdPX-1=1pωf(ω)P(X=ω)=f(ω)pω. 17

Example 1

In this example, we show that formula (14) agrees with the definition of Ij(f). To this goal, assume a product distribution, which is the basic assumption of Heckel et al. (2013), such that every state of the network is equally likely. Thus, the probability of any state is 1/2n. Then

Ij(f)=ω,ηΩnaω,η(Tj)f(ω)f(η)pωpη=12nω,ηΩnaω,η(Tj)f(ω)f(η)=12nω,ηsuppfδω,η-δφj(ω),η-δφj(η),ω+δφj(ω),φj(η)

where suppf is the support f-1(1) of the function f.

Let f be the Boolean function with support suppf={(0,1,1),(1,1,1)}. Then obviously a flip of x1 does not generate a flip in the output, so I1(f)=0. Similarly, a flip in x2 generates a flip of the output only for (x1,x2,x3) = (0,1,1), (0,0,1), (1,0,1) or (1,1,1), so, by definition, I2(f)=4/23=1/2. By symmetry, I3(f)=1/2. On the other hand,

I1(f)=18[a(0,1,1),(0,1,1)(T1)+a(0,1,1),(1,1,1)(T1)+a(1,1,1),(0,1,1)(T1)+a(1,1,1),(1,1,1)(T1)]=18[2+(-2)+(-2)+2]=0

and similarly

I2(f)=18[2+0+0+2]=12

so one can see that formula (14) agrees with the definition of Ij(f).

In order to compute the influence of each node in the network on all its output nodes (those to which the node under consideration is an input), we generate MATLAB codes that involve nested “for loops.” Note that, if we denote by ki the actual number of inputs to node i (its connectivity), for formula (14) there are i×ki×22ki such loops. This exponential number of loops can easily make the computations prohibitive. As a matter of fact, in the actual simulations, even a single connectivity of at least 12 nodes turned out to be excessive for the capabilities of MATLAB. Thus, one has to rely either on easier estimates, such as the upper bounds of Proposition 2, or find alternative exact formulas. However, a quick analysis indicates that to compute the upper bounds one would still need exponentially many “for loops,” since the procedure would require again the construction of the matrices Tj and T. For this reason, in the sequel we use the following equivalent formula which follows from the definition.

Recall that, for any Boolean function f,suppf denotes the support of f, that is, suppf=f-1(1).

Remark 1

The following formula holds

Ij(f)=ωsuppf\φj(suppf)pω+pφj(ω), 18

where the sum above is considered 0 if suppf\φj(suppf)=.

Hence:

Corollary 2

Let fi,i[n] be the Boolean update function of node i. Then

avs(fi)=j=1nωsuppfi\φj(suppfi)pω+pφj(ω)i[n] 19
σ(fi)=j=1nωsuppfj\φi(suppfj)pω+pφi(ω)i[n]. 20

Observe that:

Remark 2

Formula (18) is computationally efficient because it identifies the influential nodes, namely those in the support of the given Boolean function which get mapped outside of support by that function. In particular, the variable xj has null influence on the Boolean function f, that is Ij(f)=0, if and only if φj(suppf)suppf.

Indeed, φj(suppf)suppf implies

suppf=φj(φj(suppf))φj(suppf)suppf.

This agrees perfectly with the definition (1) of the influence as the probability of a change in the output of f when its jth input is flipped.

To understand the computational efficiency of formula (18), or equivalently the definition of the influence, we run MATLAB codes for both formula (14) and (18) on the same network, and compare the processing times. The results are included in Supplementary Material, Section 1. We present one more example.

Example 2

We call a BN: “Dominant states network (DSN)” if the update Boolean functions of nodes are characteristic functions of distinct states (called dominant states), that is if there is a set of n states S={ω1,,ωn}, so that f1=χω1,,fn=χωn.

Our previous considerations and formulas show by straightforward computations that, in the case of a DSN, one has that:

σ(fi)=j=1npωj+pφi(ωj)i[n]

and

avs(fi)=j=1npωi+pφj(ωi)i[n].

Determinative Power and Strength

In this section, we are comparing the impact of node strength to the so-called determinative power of nodes defined and explored in Heckel et al. (2013) under the assumption of a BN with product distribution of states. We recall the main definitions and concepts from Heckel et al. (2013) and Cover and Thomas (2006). These include the notion of entropy of random variables, which is a measure of uncertainty, and the mutual information which is a measure of dependence between two random variables and is defined in terms of the entropy.

Definition 2

Let X and Y be discrete random variables. The entropy of X is defined as

H(X)=-xpxlog2px=-E[log2(X)]

which in binary reduces to the function

h(p)=-plog2(p)-(1-p)log2(1-p),p=P(X=1).

The conditional entropy of Y conditional on the knowledge of X is

H(Y|X)=xpxH(Y|X=x)=-E[log2P(Y|X)].

The mutual information (MI) is the reduction in uncertainty of the random variable Y due to the knowledge of X. That is

MI(Y;X)=H(Y)-H(Y|X).

In principle, the mutual information is a measure of the “gain of information,” or the determinative power (DP) of X over Y. The authors of Heckel et al. (2013) use the MI to construct the DP of a node j over the states of a BN, namely

DP(j)=i=1nMI(fi(X);Xj) 21

which represents a summation of all “information gains” obtained from node j over its outputs (i.e., nodes that have j as an input). Besides providing a number of related mathematical results to which we will refer below, the authors identify the nodes with the largest determinative power in a feedforward E. coli network, assuming a product distribution of the input states. The goal is to be able to reduce the network to a smaller sub-network whose knowledge can provide sufficient information about the entire network; in other words the entropy of the network conditional on the knowledge of this sub-network is small enough. The authors show that by considering the nodes with most DP one can reduce the network to less than a half of its original size, so that for larger sub-networks the entropy does not improve significantly once an approximate (threshold) network size is reached.

As specified in Introduction, network reduction is an important topic in the literature, since many real networks have sizes that lead to prohibitive computations and manipulations as we can see in Supplementary Material with the computation of the influence. For instance, signal transduction networks such as that of a generic fibroblast cell, which we will use as an example, can have thousands of nodes that are interconnected. The analysis and dynamical study of such networks becomes prohibitive due to the computational burden despite the advances in technology and data science. Thus, finding meaningful ways to reduce the network to a significant “core” or “relevant component” has been of interest for a number of authors, and a number of procedures have been proposed. Ultimately, all of them generate a clear trade-off between accuracy and computational burden.

We are interested in comparing the effect of network reduction applied to the fibroblast network, by considering the nodes with largest DP on the one hand, and the nodes with largest strength values on the other hand. Before we compare them, let us focus on some theoretical results that supplement some of the formulas in Heckel et al. (2013) for the less restrictive case we consider in this paper, that is, not requiring a product distribution of the input values.

If X is the state of the network with values in Ωn, let XA taking values in Ω|A| be the collection of states of the nodes in set A[n]. So X can be written as (XA,XAc), where Ac=[n]\A. Let pω|ωA=P(X=ω|XA=ωA).

Theorem 1

The following formula for conditional entropy holds

H(f(X)|XA)=EAhωsuppfpω|XA 22

where EA refers to expected value with respect to the marginal distribution of XA.

Proof

By the definition of the conditional entropy

H(f(X)|XA)=ωAΩ|A|P(XA=ωA)H(f(X)|XA=ωA)

which in our binary case reduces to

H(f(X)|XA)=ωAΩ|A|P(XA=ωA)h(P(f(X)=1|XA=ωA))

and the obvious equality P(f(X)=1|XA=ωA)=E[f(X)|XA=ωA] implies

H(f(X)|XA)=ωAΩ|A|P(XA=ωA)h(E[f(X)|XA=ωA])=EA[h(E[f(X)|XA])]. 23

But

E[f(X)|XA=ωA]=ωΩnf(ω)P(X=ω|XA=ωA)=ωsuppfpω|ωA

which is a number in [0, 1] and we can substitute it in (23) to get formula (22).

Formula (22) is exactly the analog of Theorem 1 of Heckel et al. (2013) where it is written for a system with states -1 and 1 as opposed to 0 and 1.

Proposition 4

The mutual information formula MI(f(X);XA) can be written as

MI(f(X);XA)=hωsuppfpω-EAhωsuppfpω|XA. 24

Proof

The formula for MI follows again directly from the definition of the mutual information

MI(f(X);XA)=H(f(X))-H(f(X)|XA)=h(E[f(X)])-H(f(X)|XA).

But

h(E[f(X)])=hωΩnf(ω)P(X=ω)=hωsuppfpω

with the argument of h being in [0, 1] as needed. Substituting this and formula (22) in the definition of the mutual information we obtain formula (24).

Now we focus on two special (extreme) cases considered in Heckel et al. (2013), to see if we can identify an analog of the results in that paper, where the authors use the additional assumption of independence paired with the Bahadur basis (a family of functions which form a complete orthonormal basis of L2(Ωn,dPX-1) provided that X1,X2,,Xn are independent).

Special cases:

  • A=[n]. Then Ac= and
    MI(f(X);X)=hωsuppfpω
    which is maximized for ωsuppfpω=1/2, in other words if E[f(X)]=P(f(X)=1)=1/2. Hence, we are dealing with a nonbiased function. So the closer the function f is to a nonbiased function, the larger the MI between its output and all of its inputs. This is similar to what was observed in Heckel et al. (2013).
  • A={i} where i is a fixed input/node. Thus, ωA=ωi,XA=Xi. The mutual information can be written as
    MI(f(X);Xi)=hωsuppfpω-P(Xi=1)hωsuppfpω|1-P(Xi=0)hωsuppfpω|0
    In comparison with Heckel et al. (2013), this formula does not allow for a simplification to a small subset of Fourier coefficients in order to find the MI. In their formula, the authors of Heckel et al. (2013) identify three coefficients that act as independent variables. Based on them they manage to obtain some information on the behavior of the MI, which is not the case for our approach. The reason for considering this special case is that the quantity MI(f(X);Xi) is also known as information gain, or informativeness, and is common in information theory. Thus, we note that a change in the underlying basis can induce different results and situations.

Now that we have a deeper understanding of the MI as used in formula (21), let us turn to the network under consideration, namely the signal transduction network of a generic fibroblast cell which consists of several main signaling pathways, including the receptor tyrosine kinase, the G-protein coupled receptor, and the integrin signaling pathway. A Boolean representation of this network has been provided in Helikar et al. (2008), and has been studied further in Kochi and Matache (2012) and Kochi et al. (2014). The fully annotated signal transduction model is freely available for simulations and/or download via the Cell Collective software from www.thecellcollective.org Helikar et al. (2012) and Helikar et al. (2013). Each node in the model represents a signaling molecule (mainly protein). The network has 130 nodes with connectivities that vary between 1 and 14 nodes.

Using formula (18), one can quickly determine the average sensitivity and the strength of all nodes in the fibroblast network. In Fig. 1, we plot them against the nodes (in alphabetic order), together with two more plots on the connectivity and the bias, i.e., P(f(X)=1), for an overall view of the main numerical characteristics of the fibroblast nodes.

Fig. 1.

Fig. 1

(Color Figure online) Main numerical characteristics of the nodes of the fibroblast network as specified in each subplot

We compute the DP of the nodes in the fibroblast network and compare them with the node strength σ. The results are shown in Fig. 2. For the network under consideration, the strength values seem to be slightly larger than the DP values. We have conducted a statistical analysis related to DP and σ values for the fibroblast network. In summary, there is enough statistical evidence that the average DP-σ is negative with a p value of basically zero. The paired test gives an upper bound of -0.14208 for a 95 % confidence interval for the difference DP-σ. On the other hand, a linear regression analysis indicates a fairly strong linear relationship between the two variables with a 75.1 % coefficient of determination (COD), and a higher COD of 82.4 % for the linear relationship between the average σ and the number of outlinks corresponding to the nodes. The average values are computed over all nodes with a given number of outlinks. This relationship is weaker for average DP versus number of outlinks with a COD of 60.3 %. We also note that the outliers occur mostly for nodes with a larger number of outlinks. In other words, fewer outlinks generate a stronger correlation between the DP or σ and the number of outlinks. For example, there is one particular node in the network, namely EGFR, that generates the maximum DP and is the only node with 13 outlinks. If we eliminate this node from the correlation analysis, the COD for average DP versus outlinks increases from 60.3 to 81.3 %. Notably, mutations of the EGFR, epidermal growth factor receptor, are known to be related to lung cancer, interfering with the signaling pathways within the cell triggered to promote cell growth and division (proliferation) and cell survival. The second node in the order of DP is ASK1, apoptosis signal-regulating kinase 1, and plays important roles in many stress-related diseases, including cancer, diabetes, cardiovascular, and neurodegenerative diseases. The third node is Src, proto-oncogene tyrosine-protein kinase, which is involved in the control of many functions, including cell adhesion, growth, movement, and differentiation. The fourth node is PIP3_345, phosphatidylinositol (3,4,5)-trisphosphate that functions to activate downstream signaling components, while the fifth node is PKC, protein kinase C, involved in receptor desensitization, in modulating membrane structure events, in regulating transcription, in mediating immune responses, in regulating cell growth, and in learning and memory. The DP procedure managed to capture the importance of these nodes in relationship to the rest of the network. Four of the top five DP nodes are also among the five strongest nodes which are: Src, PIP3_345, PKC, PIP2_45, and EGFR. Thus, the strength also captures biologically important nodes. Moreover, higher DP and strength values are correlated with a larger number of outlinks as seen from the figures, which means that this procedure can identify hubs in the network. It is also apparent from the figures that the COD increases when considering smaller DP and σ values. We have included relevant figures in Supplementary Material, Sect. 2

Fig. 2.

Fig. 2

(Color Figure online) Comparison of DP and σ by nodes of the fibroblast network. The nodes are sorted by names in the top panels, while in the bottom panels they are ordered according to increasing DP, and σ, respectively, as indicated in the graphs. Note that the strength values seem to be slightly larger than the DP values

We would also like to point out at this time that the MI has been used as way to identify relevant pairs of genes in genetic expression data sets by Butte and Kohane (2000, (2003), and Jiang et al. (2009). Those authors identify relevance networks by selecting pairs of genes whose MI surpasses a given threshold. For example, in Butte and Kohane (2000) it is shown that the relevance networks are of the following types: those that link identical genes, those linking genes with similar functions, those linking genes in the same biological pathway, and combinations of these. In our work, the MI is directed from the node toward its outputs via the outlinks, so we do not involve the bidirectional aspect of the previous studies mentioned here. Moreover, we do not use genetic expression data sets, but actual Boolean functions. Nevertheless, it will be of interest to explore in the future the main types or classes of nodes/sub-networks identified by the DP procedure in a variety of cellular networks.

Next, let us compute the network entropy generated by sub-networks chosen based on top DP or strength values of nodes. If we denote by Al the collection of the top l nodes in order of DP or σ, then we compute

H(X|XAl)i=1nH(Xi|XAl),forl=1,2,3,,n. 25

We plot the values of the larger quantity in (25) against l and obtain the graph in Fig. 3, where we note that the entropy decreases with increased sub-network size l, and that the two curves are very close, with a slightly better result for σ which provides lower entropy values for most values of l, as seen from the bottom panel where we plot also the differences of entropy values by l. As the sub-network size increases, the strength provides a somewhat tighter upper bound for the entropy of the network. Note that the actual entropy will be smaller than the upper bound graphed in Fig. 3, and that sub-networks of sizes 60 or more (with approximation), do not yield a significant improvement of the entropy. Thus, it suffices to consider less than half of the original network to be able to predict the overall network behavior with fairly low uncertainty/entropy levels.

Fig. 3.

Fig. 3

(Color Figure online) Values of the upper bound in (25) for sub-networks chosen based on the top l values of DP and σ, respectively. The bottom panel shows the differences in the entropy that favor mostly σ for l>20 with approximation

Knowing that the DP method allows a network reduction without a significant increase in entropy, one could use a tested approach for node elimination that preserves essential dynamical properties to reduce the network to the sub-network of the top DP nodes. For example, in Naldi et al. (2009b) the authors introduce a general method for eliminating nodes one by one by basically connecting directly the inputs of a removed node to its output nodes. Of course, one needs to either keep or consider some extra decisions on autoregulated nodes, that is nodes that are or may become self-inputs upon elimination of other nodes. One also needs to understand the impact of the order in which nodes are removed. In our case, the natural order is provided by the sorted DP values. In Naldi et al. (2009b), it is shown that with their approach the attractors and stable states are preserved. We note here that there are nodes whose removal may have no impact on the dynamics, like those with no outputs. Moreover, the authors of Naldi et al. (2009b) have developed a Java software Naldi et al. (2009a) that allows one to apply the reduction algorithm and analyze attractors. Alternative methods have been proposed by Veliz-Cuba and collaborators in Veliz-Cuba (2011); Veliz-Cuba et al. (2015) that could be used in conjunction with the DP method. Future work will explore those methods.

One could actually go one step further and provide the following estimates for the conditional entropy H(f(X)|XA), which are the analog of Theorem 2 of Heckel et al. (2013).

Theorem 2

The following estimates of the conditional entropy hold:

LBH(f(X)|XA)LB1/ln4 26

where

LB=4E[f(X)]-EA[E[f(X)|XA]2]=4ωsuppfpω-ωAΩ|A|pωAωsuppfpω|ωA2 27

Proof

We use the following inequality (found in Topsoe 2001 and used in Heckel et al. 2013) that provides lower and upper bounds on the binary entropy function h(p)

4p(1-p)h(p)[4p(1-p)]1/ln4.

If in formula (22) we denote

q(XA)=ωsuppfpω|XA

then LB=EA[4q(XA)(1-q(XA))]. For the upper bound, we use the fact that

EA[(4q(XA)(1-q(XA)))1/ln4]EA[4q(XA)(1-q(XA))]1/ln4.

Then the double inequality (26) is immediate. Now, LB can be expressed as follows

LB=EA[4q(XA)(1-q(XA))]=4EAωsuppfpω|XA1-ωsuppfpω|XA=4EAωsuppfpω|XA-ωsuppfpω|XA2=4ωAΩ|A|pωAωsuppfpω|ωA-ωAΩ|A|pωAωsuppfpω|ωA2=4ωsuppfpω-ωAΩ|A|pωAωsuppfpω|ωA2

where the first term is obtained by the law of total probability.

We are interested in identifying possible relationships or inequalities between the mutual information and the influence of a subset of nodes of the network. In Heckel et al. (2013), the authors show that for a given collection of nodes A[n], the following holds

IA(f)miniA1σi2MI(f(X);XA)-Ψ(Var(f(X))) 28

where Ψ(x)=x(1/ln4)-x takes positive values that are very close to zero, and σi2=Var(Xi), under the assumption of independence of the random variables. Let us explore an alternative inequality using the results of this paper. For this purpose, we consider a special case that allows us to compare directly our results with inequality (28). In this special case, we consider a uniform distribution of inputs, which is actually a case of product distributions or independent inputs.

Special Case

Consider a network with n nodes, such that for each ωΩn, we have pω=12n, so that we consider a uniform distribution of the inputs. Let |suppf|=K and A[n]. Then

pωA=P(XA=ωA)=ωAcΩn-|A|12n=12|A|ωAΩ|A|.

Using formula (24), we can write the following for the MI,

MI(f(X);XA)=hωsuppfpω-EAhωsuppfpω|XA=hK2n-ωAΩ|A|pωAhωsuppfpω|ωA=hK2n-12|A|ωAΩ|A|hωsuppfpω|ωA

Note that

pω|ωA=P(X=ω|XA=ωA)=P(X=ω,XA=ωA)P(XA=ωA)=P(X=ω)P(XA=ωA)=12n-|A|.

If we let KωA=|suppfPrA-1(ω)|, where PrA is the projection of ω on A, then

MI(f(X);XA)=hK2n-12|A|ωAΩ|A|hKωA2n-|A|. 29

We immediately notice that 0KωAK for all ωAΩ|A|, and that ωAΩ|A|KωA=K, so that we create a partition of suppf. Therefore, in the sum of (29), some of the terms could be zero, since not all ωAΩ|A| need to be represented in suppf, leading to KωA=0, which in turn leads to h(0)=0.

Let us focus on IA(f)=jAIj(f). By formula (18) we obtain

Ij(f)=ωsuppf\φj(suppf)pω+pφj(ω)=ωsuppf\φj(suppf)12n-1=|suppf\φj(suppf)|2n-1.

If mj=|suppfφj(suppf)| then

Ij(f)=K-mj2n-1

and consequently

IA(f)=jAK-mj2n-1=K|A|-jAmj2n-1. 30

Notice that 0mjK and that mj is an even number since φj is its own inverse.

Observe that if in formula (28) of Heckel et al. (2013) we consider a uniform distribution of random variables, then all σi2=1, and thus the influence bounds above the MI minus a small positive quantity. We would like to check a similar inequality for our case. To this end, we first generate some simulations in which we plot both MI(f(X);XA) and IA(f) versus K for various values of |A| in Fig. 4, using formulas (29) and (30). Not only is IA(f) an upper estimate of MI(f(X);XA), but also its values are significantly larger. So in our case, the inequality becomes

MI(f(X);XA)IA(f) 31

which is stronger than the corresponding inequality of  Heckel et al. (2013). Besides, the quantity MI(f(X);XA)-Ψ(Var(f(X))) used in Heckel et al. (2013) takes on also negative values. For example, one can easily check that this is true for n=8,K=1,|A|=3,4,5 under the assumptions of the uniform distribution (which leads to independence of random variables). Then the inequality becomes superfluous.

Fig. 4.

Fig. 4

(Color Figure online) Graphs of IA(f) and MI(f(X);XA) computed with formulas (29) and (30) versus K for a network with n=8 nodes and |A|=1,2,,n as specified in the subplots. The MI curve is very close to zero for most values of |A|, while the influence has much larger values in all cases. Recall also that MI is always a number in [0, 1]. The actual values increase with the increase in |A|. Notice also the expected symmetry as K crosses from values less than to values greater than 2n/2

Observe that in this case, if A={j}, and fi is the Boolean function associated with node i of the network, then the inequality (31) becomes MI(fi(X);Xj)Ij(fi) which implies i=1nMI(fi(X);Xj)i=1nIj(fi), or in other words DP(j)σ(fj). However, when looking at the relationship between the DP and strength values in Fig. 2, we observe that this inequality does not hold for all nodes of the fibroblast network. That is most likely due to the actual dependencies between the states of the nodes. Thus, dependent inputs may lead to different results. However, we notice in Fig. 2, top right graph, that the magnitudes of the positive differences DP-σ are generally smaller than the magnitudes for negative differences. Thus, it may be possible that a version of inequality (28) is still valid. We have not been able to find such an alternative inequality so far. On the other hand, all the examples involving dependent inputs we have looked at, support inequality (31). We provide one such example before returning to the Special Case 3.1.

Example 3

Consider the Boolean function f(x) where x=(x1,x2,x3), represented by the truth table shown below and the corresponding probabilities of states. It is easy to check that this is not a product distribution, so the variables are dependent; for example, P(X1=0,X2=0)=3/10 while P(X1=0)P(X2=0)=1/4. One can also check that I1(f)=3/5 and I2(f)=I3(f)=1/2. In this case, for any A[3] of cardinal at least two, we get automatically that MI(f(X);XA)1IA(f). On the other hand, for A={1} we obtain via formula (24), that MI(f(X);X{1})=h(1/2)-(1/2)h(3/5)-(1/2)h(2/5)=.029<0.6=I1(f). Also, MI(f(X);X{2})=MI(f(X);X{3})=h(1/2)-(1/2)h(7/10)-(1/2)h(3/10)=.1187<0.5=I2(f)=I3(f). So in all cases, MI(f(X);XA)IA(f).

(x1,x2,x3) f(x1,x2,x3) P(x1,x2,x3)
(0, 0, 0) 0 3/20
(0, 0, 1) 1 3/20
(0, 1, 0) 0 1/20
(0, 1, 1) 1 3/20
(1, 0, 0) 1 3/20
(1, 0, 1) 1 1/20
(1, 1, 0) 0 3/20
(1, 1, 1) 0 3/20

Now we return to our Special Case 3.1 and we conjecture that the following inequality is true for all choices of parameters

hK2n-12|A|ωAΩ|A|hKωA2n-|A|K|A|-jAmj2n-1. 32

A general proof of this inequality seems to be very technical and intricate. We present briefly a few particular cases whose proofs can be found in Supplementary Material, Sect. 3 Note that the extreme cases of K=0 and K=2n are trivially satisfied since they lead to null quantities on both sides of inequality (32).

Case 1: The support is a singleton.

The inequality (32) takes on the particular form

h12n-12kh12n-kk2n-1

where |A|=k.

One obtains the following consequence, which is valid no matter the cardinality of suppf.

Corollary 3

Inequality (32) holds if |PrA(suppf)|=1.

Case 2:suppf={τ,η}, τη

The inequality (32) becomes

h22n-12kωAΩkhKωA2n-k2k-jAmj2n-1.

Case 3: The support is a subgroup of Ωn and A=[n].

What is meant here is that we identify {0,1} to Z2, the additive group of equivalence classes modulo 2, and Ωn to the product group Z2n. For any fixed j[n], denote by δj the Boolean vector in Ωn whose entries are all null, except entry j. Under the previously described identification, one easily sees that, given a Boolean function f, the quantities

mj=|suppfφj(suppf)|j=1,,n,

can be calculated with the alternative formula

mj=|suppf(δj+suppf)|j=1,,n,

where the kind of addition used is addition modulo 2. Finally, recall that the order of a subgroup of Ωn must be a divisor of 2n; hence, it will have the form 2k for some nonnegative integer kn. Keeping all the above in mind, we state and prove the following:

Lemma 1

Let f be a Boolean function, S its support, and S, the subgroup of Ωn generated by S. Then, the following inequality holds:

j=1nmjk2k

where 2k=|S|.

Now, observe that if A=[n], the inequality we wish to prove has the form

hK2nKn-j=1nmj2n-1.

and if S=suppf is a subgroup of Ωn of order 2k, then the inequality becomes

h12n-k2kn-j=1nmj2n-1.

This one holds since one can write

h12n-k(n-k)2n-k-1=n2k-k2k2n-12kn-j=1nmj2n-1.

Remark 3

If f is a Boolean function having support S,A=[n], and S{δ1,,δn}=, then (32) holds.

Final Comments

The main conclusions of this work are that operator theory can offer computationally efficient ways to find or estimate important quantities used in assessing the sensitivity to perturbations of BNs, and to quantify the relevance of nodes using elements of information theory, in particular MI. We conclude that MI is an excellent tool for identifying a subset of relevant nodes in the network that offer the most information gain and whose knowledge reduces the entropy of the whole network significantly. Moreover, the MI provides a lower estimate for the influence of nodes in various scenarios.

It would be of interest to continue this exploration under various scenarios of dependent nodes in the network, as well as to refine further some of the results of this paper. For example, could one strengthen inequality (32) and prove it in general or for different scenarios?

On the other hand, in Klotz et al. (2014) it is shown that MI is maximized for canalizing functions. However, real networks do not consist of a single type of Boolean function. Therefore, it would be of interest to explore a possible hierarchy of various types of functions regarding the information gain they provide. That could offer more information regarding the inequality between the MI and the influence. Besides, most of the functions in real networks, such as cellular networks, need not be strictly canalizing as considered in Klotz et al. (2014) (i.e., one value of one of the inputs forces the output to take on a certain fixed value regardless of the other inputs). In reality, functions may be only partially canalizing, allowing for multiple, but not necessarily all inputs, to be canalizing in a cascading fashion as discussed, for example, in Layne et al. (2012) or Dimitrova et al. (2015). Partially nested canalizing functions have been considered recently as a more realistic alternative to canalizing functions. Another type of function that is common in applications is a threshold function that turns a node ON provided that a sufficient number of inputs are ON. This type of function is typical for neural networks. This opens the door for a variety of directions of research that stem from the work in this paper.

Furthermore, as specified before, it is our intention to explore the Java software Naldi et al. (2009a) to actually perform network reduction preserving attractors and stability and use it to analyze dynamics of various signal transduction networks found in Cell Collective Helikar et al. (2012) and Helikar et al. (2013).

Finally, it would be of great interest to look further into other real networks to identify the possibility of reducing them to the most determinative nodes. For example, it would be interesting to compare the results of the fibroblast network to other networks such as the Boolean model of the influenza–host interactions during infection of an epithelial cell Madrahimov et al. (2013), to identify possible similarities and differences that may occur. At the same time, identifying possible classes of biological nodes/sub-networks obtained with the DP sub-network procedure in a variety of other networks could bring further clarifications on the advantages of the DP method. Even more, it is important to assess the degree to which the reduced network provides accurate information on specific tasks typical for the whole network, such as pattern recognition or decision making. At the same time, exploring the impact of considering DP versus strength might provide new insights.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Contributor Information

Mihaela T. Matache, Email: dmatache@unomaha.edu

Valentin Matache, Email: vmatache@unomaha.edu.

References

  1. Albert R, Othmer H. The topology of the regulatory interactions predicts the expression pattern of the segment polarity genes in Drosophila melanogaster. J Theor Biol. 2003;223:1–18. doi: 10.1016/S0022-5193(03)00035-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Bilke S, Sjunnesson F. Stability of the Kauffman model. Phys Rev E. 2001;65:016129. doi: 10.1103/PhysRevE.65.016129. [DOI] [PubMed] [Google Scholar]
  3. Butte AJ, Kohane IS. Mutual information relevance networks: functional genomic clustering using pairwise entropy measurements. Pac Symp Biocomput. 2000;5:415–426. doi: 10.1142/9789814447331_0040. [DOI] [PubMed] [Google Scholar]
  4. Butte AJ, Kohane IS (2003) Relevance networks: a first step toward finding genetic regulatory networks within microarray data. In: Parmigiani G, Garett ES, Irizarry RA, Zeger SL (eds) The analysis of gene expression data. Part of the series statistics for biology and health. Springer, Berlin, pp 428–446
  5. Cover TM, Thomas JA (2006) Elements of information theory. John Wiley & Sons, Inc., Hoboken, New Jersey
  6. Devloo V, Hansen P, Labbé M. Identification of all steady states in large networks by logical analysis. Bull Math Biol. 2003;65(6):10251051. doi: 10.1016/S0092-8240(03)00061-2. [DOI] [PubMed] [Google Scholar]
  7. Dimitrova ES, Yordanov OI, Matache MT. Difference equation for tracking perturbations in systems of boolean nested canalyzing functions. Phys Rev E. 2015;91:062812. doi: 10.1103/PhysRevE.91.062812. [DOI] [PubMed] [Google Scholar]
  8. Feldman DP, Crutchfield JP. Measures of statistical complexity: why? Phys Lett A. 1998;238(45):244252. [Google Scholar]
  9. Ghanbarnejad F, Klemm K. Impact of individual nodes in boolean network dynamics. Europhys Lett. 2012;99(5):58006. doi: 10.1209/0295-5075/99/58006. [DOI] [Google Scholar]
  10. Goles E, Montalva M, Ruz GA. Deconstruction and dynamical robustness of regulatory networks: application to the yeast cell cycle networks. Bull Math Biol. 2013;75(6):939–966. doi: 10.1007/s11538-012-9794-1. [DOI] [PubMed] [Google Scholar]
  11. Heckel R, Schober S, Bossert M. Harmonic analysis of boolean networks: determinative power and perturbations. EURASIP J Bioinform Syst Biol. 2013;2013:1–13. doi: 10.1186/1687-4153-2013-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Helikar T, Konvalina J, Heidel J, Rogers JA. Emergent decision-making in biological signal transduction networks. PNAS. 2008;105(6):1913–1918. doi: 10.1073/pnas.0705088105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Helikar T, Kowal B, McClenathan S, Bruckner M, Rowley T, Wicks B, Shrestha M, Limbu K, Rogers JA. The cell collective: toward an open and collaborative approach to systems biology. BMC Syst Biol. 2012;6:96. doi: 10.1186/1752-0509-6-96. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Helikar T, Kowal B, Rogers JA. A cell simulator platform: the cell collective. Clin Pharmacol Ther. 2013;93:393–395. doi: 10.1038/clpt.2013.41. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Huepe C, Aldana-González M. Dynamical phase transition in a neural network model with noise: an exact solution. J Stat Phys. 2002;108(3–4):527–540. doi: 10.1023/A:1015777824097. [DOI] [Google Scholar]
  16. Jiang W, Zhang L, Na B, Wang L, Xu J, Li X, Wang Y, Rao S. Mapping and characterization of two relevance networks from SNP and gene levels. Prog Nat Sci. 2009;19(5):653657. doi: 10.1016/j.pnsc.2008.05.038. [DOI] [Google Scholar]
  17. Kahn J, Kalai G, Linial N (1988) The influence of variables on Boolean functions. In: Proceedings of the 29th annual symposium on foundations of computer science. IEEE, New York, pp 68–80
  18. Kauffman SA. Metabolic stability and epigenesis in randomly constructed genetic nets. J Theor Biol. 1969;22(3):437–467. doi: 10.1016/0022-5193(69)90015-0. [DOI] [PubMed] [Google Scholar]
  19. Kauffman SA. The origins of order. Oxford: Oxford University Press; 1993. [Google Scholar]
  20. Kaufman V, Mihaljev T, Drossel B. Scaling in critical random boolean networks. Phys Rev E. 2005;72(4):046124. doi: 10.1103/PhysRevE.72.046124. [DOI] [PubMed] [Google Scholar]
  21. Kaufman V, Drossel B. Relevant components in critical random boolean networks. New J Phys. 2006;8(10):228. doi: 10.1088/1367-2630/8/10/228. [DOI] [Google Scholar]
  22. Klemm K, Bornholdt S. Stable and unstable attractors in boolean networks. Phys Rev E. 2000;72:055101. doi: 10.1103/PhysRevE.72.055101. [DOI] [PubMed] [Google Scholar]
  23. Klotz JG, Kracht D, Bossert M, Schober S. Canalizing boolean functions maximize the mutual information. IEEE Trans Inform Theory. 2014;60(4):2139–2147. doi: 10.1109/TIT.2014.2304952. [DOI] [Google Scholar]
  24. Kochi N, Helikar T, Allen L, Rogers JA, Wang Z, Matache MT. Sensitivity analysis of biological boolean networks using information fusion based on nonadditive set functions. BMC Syst Biol. 2014;8:92. doi: 10.1186/s12918-014-0092-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Kochi N, Matache MT. Mean-field boolean network model of a signal transduction network. Biosystems. 2012;108:14–27. doi: 10.1016/j.biosystems.2011.12.001. [DOI] [PubMed] [Google Scholar]
  26. Krawitz P, Shmulevich I. Basin entropy in boolean network ensembles. Phys Rev Lett. 2007;98:158701. doi: 10.1103/PhysRevLett.98.158701. [DOI] [PubMed] [Google Scholar]
  27. Krawitz P, Shmulevich I. Entropy of complex relevant components of boolean networks. Phys Rev E. 2007;76:036115. doi: 10.1103/PhysRevE.76.036115. [DOI] [PubMed] [Google Scholar]
  28. Layne L, Dimitrova E, Macauley M. Nested canalyzing depth and network stability. Bull Math Biol. 2012;74(2):422–433. doi: 10.1007/s11538-011-9692-y. [DOI] [PubMed] [Google Scholar]
  29. Lloyd-Price J, Gupta A, Ribeiro AS. Robustness and information propagation in attractors of random boolean networks. PLoS One. 2012;7(7):e42018. doi: 10.1371/journal.pone.0042018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Luque B, Ferrera A. Measuring mutual information in random boolean networks. Complex Syst. 2000;12(2):241. [Google Scholar]
  31. Madrahimov A, Helikar T, Kowal B, Lu G, Rogers J. Dynamics of influenza virus and human host interactions during infection and replication cycle. Bull Math Biol. 2013;75(6):988–1011. doi: 10.1007/s11538-012-9777-2. [DOI] [PubMed] [Google Scholar]
  32. Naldi A, Berenguier D, Fauré A, Lopez F, Thieffry D, Chaouiya C (2009a) Logical modelling of regulatory networks with GINsim 2.3. Biosystems 97(2):134–139 [DOI] [PubMed]
  33. Naldi A, Remy E, Thieffry D, Chaouiya C (2009b) A reduction of logical regulatory graphs preserving essential dynamical. In: Degano P, Gorrieri R (eds) Computational methods in systems biology. Lecture notes in computer science, vol 5688. Springer-Verlag Berlin Heidelberg, pp 266–280
  34. Ribeiro AS, Kauffman SA, Lloyd-Price J, Samuelsson B, Socolar JES. Mutual information in random boolean models of regulatory networks. Phys Rev E. 2008;77:011901. doi: 10.1103/PhysRevE.77.011901. [DOI] [PubMed] [Google Scholar]
  35. Richardson KA. Simplifying boolean networks. Adv Complex Syst. 2004;8(4):365381. [Google Scholar]
  36. Saadatpour A, Albert R, Reluga TC. A reduction method for boolean network models proven to conserve attractors. SIAM J Appl Dyn Syst. 2013;12:1997–2011. doi: 10.1137/13090537X. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Shmulevich I, Dougherty ER, Zhang W. From boolean to probabilistic boolean networks as models for genetic regulatory networks. Proc IEEE. 2002;90(11):1778–1792. doi: 10.1109/JPROC.2002.804686. [DOI] [Google Scholar]
  38. Shmulevich I, Kauffman SA. Activities and sensitivities in boolean network models. Phys Rev Lett. 2004;93(4):048701. doi: 10.1103/PhysRevLett.93.048701. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Socolar JES, Kauffman SA. Scaling in ordered and critical random boolean networks. Phys Rev Lett. 2003;90:068702. doi: 10.1103/PhysRevLett.90.068702. [DOI] [PubMed] [Google Scholar]
  40. Sole RV, Luque B (1997) Statistical measures of complexity for strongly interacting systems. Santa Fe Institute working paper 97-11-083
  41. Topsoe F. A bounds for entropy and divergence for distributions over a two-element set. J Inequal Pure Appl Math. 2001;2(2):25. [Google Scholar]
  42. Veliz-Cuba A. Reduction of boolean network models. J Theor Biol. 2011;289:167172. doi: 10.1016/j.jtbi.2011.08.042. [DOI] [PubMed] [Google Scholar]
  43. Veliz-Cuba A, Laubenbacher R, Aguilar B. Dimension reduction of large sparse AND–NOT network models. Electron Notes Theor Comput Sci. 2015;316:8395. doi: 10.1016/j.entcs.2015.06.012. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Bulletin of Mathematical Biology are provided here courtesy of Springer

RESOURCES