Skip to main content
PLOS One logoLink to PLOS One
. 2012 Feb 15;7(2):e31395. doi: 10.1371/journal.pone.0031395

Recent Developments in Quantitative Graph Theory: Information Inequalities for Networks

Matthias Dehmer 1,*, Lavanya Sivakumar 1
Editor: Frank Emmert-Streib2
PMCID: PMC3280299  PMID: 22355362

Abstract

In this article, we tackle a challenging problem in quantitative graph theory. We establish relations between graph entropy measures representing the structural information content of networks. In particular, we prove formal relations between quantitative network measures based on Shannon's entropy to study the relatedness of those measures. In order to establish such information inequalities for graphs, we focus on graph entropy measures based on information functionals. To prove such relations, we use known graph classes whose instances have been proven useful in various scientific areas. Our results extend the foregoing work on information inequalities for graphs.

Introduction

Complexity is an intricate and versatile concept that is associated with the design and configuration of any system [1], [2]. For example, complexity can be measured and characterized by quantitative measures often called indices [3][5]. When studying the concept of complexity, information theory has been playing a pioneering and leading role. Prominent examples are the theory of communication and applied physics where the famous Shannon entropy [6] has extensively been used. To study issues of complexity in natural sciences and, in particular, the influence and use of information theory, see [7].

In this paper, we deal with an important aspect when studying the complexity of network-based systems. In particular, we establish relations between information-theoretic complexity measures [3], [8][11]. Recall that such entropic measures have been used to quantify the information content of the underlying networks [8], [12]. Generally, this relates to exploring the complexity of a graph by taking its structural features into account. Note that numerous measures have been developed to study the structural complexity of graphs [5], [8], [13][22]. Further, the use and ability of the measures has been demonstrated by solving interdisciplinary problems. As a result, such studies have led to a vast number of contributions dealing with the analysis of complex systems by means of information-theoretic measures, see, e.g., [8], [13][22]. Figure 1 shows a classification scheme of quantitative network measures exemplarily.

Figure 1. A classification of quantitative network measures.

Figure 1

The main contribution of this paper is to study relations between entropy measures. We will tackle this problem by means of inequalities involving network information measures. In particular, we study so-called implicit information inequalities which have been introduced by Dehmer et al. [23], [24] for studying graph entropies using information functionals. Generally, an implicit information inequality involves information measures which are present on either side of the inequality. It is important to emphasize that relatively little work has been done to investigate relations between network measures. A classical contribution in this area is due to Bonchev et al. [25]. Here, the relatedness between information-theoretic network measures has been investigated to detect branching in chemical networks. Further, implicit information inequalities have been studied for hierarchical graphs which turned out to be useful in network biology [26].

We first present closed form expressions of graph entropies using the graph classes, stars and path graphs. Further, we infer novel information inequalities for the measures based on the Inline graphic-sphere functional. The section “Implicit Information Inequalities” presents our main results on novel implicit inequalities for networks. We conclude the paper with a summary and some open problems. Before discussing our results, we will first present the information-theoretic measures that we want to investigate in this paper.

Methods

In this section, we briefly state the concrete definitions of the information-theoretic complexity measures that are used for characterizing complex network structures [3], [6], [9], [27]. Here we state measures based on two major classifications namely partition-based and partition-independent measures and deal mainly with the latter.

Given a simple, undirected graph Inline graphic, let Inline graphic denote the distance between two vertices Inline graphic and Inline graphic, and let Inline graphic. Let Inline graphic denote the Inline graphic-sphere of a vertex Inline graphic defined as Inline graphic. Throughout this article, a graph Inline graphic represents a simple undirected graph.

Definition 1 Let Inline graphic be a graph on Inline graphic vertices and let Inline graphic be a graph invariant of Inline graphic. Let Inline graphic be an equivalence relation that partitions Inline graphic into Inline graphic subsets Inline graphic, with cardinality Inline graphic for Inline graphic. The total structural information content of Inline graphic is given by

graphic file with name pone.0031395.e023.jpg (1)

Definition 2 Let Inline graphic be a graph on Inline graphic vertices and let Inline graphic, for Inline graphic be the probability value for each partition. The mean information content of Inline graphic is

graphic file with name pone.0031395.e029.jpg (2)

In the context of theory of communication, the above equation is called as Shannon equation of information [28].

Definition 3 Let Inline graphic be a graph on Inline graphic vertices. The quantity

graphic file with name pone.0031395.e032.jpg (3)

is a probability value of Inline graphic. Inline graphic is an arbitrary information functional that maps a set of vertices to the non-negative real numbers.

Remark 1 Observe that, Inline graphic defines a probability distribution over the set of vertices as it satisfies Inline graphic, for every vertex Inline graphic, Inline graphic and Inline graphic.

Using the resulting probability distribution associated with Inline graphic leads to families of network information measures [3], [9].

Definition 4 The graph entropy of Inline graphic given representing its structural information content:

graphic file with name pone.0031395.e042.jpg (4)

In order to define concrete graph entropies, we reproduce the definitions of some information functionals based on metrical properties of graphs [3], [9], [27].

Definition 5 Parameterized exponential information functional using Inline graphic -spheres:

graphic file with name pone.0031395.e044.jpg (5)

where Inline graphic and Inline graphic for Inline graphic.

Definition 6 Parameterized linear information functional using Inline graphic -spheres:

graphic file with name pone.0031395.e049.jpg (6)

where Inline graphic for Inline graphic.

Remark 2 Observe that, when either Inline graphic or the Inline graphic are all equal, the functional Inline graphic and Inline graphic becomes a constant function and, hence, the probability on all the vertices are equal. That is Inline graphic, for Inline graphic. Thus, the value of the entropy attains its maximum value, Inline graphic. Thus, in all our proofs, we only consider the non-trivial case, namely Inline graphic and/or at least for two coefficients holds Inline graphic.

Next, we will define the local information graph to use local centrality measures from [9]. Let Inline graphic be the subgraph induced by the shortest path starting from the vertex Inline graphic to all the vertices at distance Inline graphic in Inline graphic. Then, Inline graphic is called the local information graph regarding Inline graphic with respect to Inline graphic, see [9]. A local centrality measure that can be applied to determine the structural information content of a network [9] is then defined as follows.

Definition 7 The closeness centrality of the local information graph is defined by

graphic file with name pone.0031395.e068.jpg (7)

Remark 3 Note that centrality is an important concept that has been introduced for analyzing social networks [29], [30] . Many centrality measures have been contributed [30] , and in particular, various definitions for closeness centrality [30][32] . We remark that the above definition has been firstly defined by Sabidussi [31] for arbitrary graphs. However, we use the measure as a local invariant defined on the subgraphs induced by the local information graph [9] .

Similar to the Inline graphic-sphere functionals, we define further functionals based on the local centrality measure as follows.

Definition 8 Parameterized exponential information functional using local centrality measure:

graphic file with name pone.0031395.e070.jpg (8)

where Inline graphic, Inline graphic for Inline graphic.

Definition 9 Parameterized linear information functional using local centrality measure:

graphic file with name pone.0031395.e074.jpg (9)

where Inline graphic, for Inline graphic.

Note that the coefficients Inline graphic can be chosen arbitrarily. However, the functionals become more meaningful when we choose the coefficients to emphasize certain structural characteristics of the underlying graphs. Also, this remark implies that the notion of graph entropy is not unique because each measure takes different structural features into account. Further, this can be understood by the fact that a vast number of entropy measures have been developed so far. Importantly, we point out that the measures we explore in this paper are notably different to the notion of graph entropy introduced by Körner [21]. The graph entropy due to Körner [21] is rooted in information theory and based on the known stable set problem. To study more related work, survey papers on graph entropy measures have been authored by Dehmer et al. [3] and Simonyi [33].

Results and Discussion

Closed Form Expressions and Explicit Information Inequalities

When calculating the structural information content of graphs, it is evident that the determination of closed form expressions using arbitrary networks is critical. In this section, we consider simple graphs namely trees with smallest and largest diameter and compute the measures defined in the previous section. By using arbitrary connected graphs, we also derive explicit information inequalities using the measures based on information functionals (stated in the previous section).

Stars

Star graphs, Inline graphic, have been of considerable interest because they represent trees with smallest possible diameter (Inline graphic) among all trees on Inline graphic vertices.

Now, we present closed form expressions for the graph entropy by using star graphs. For this, we apply the information-theoretic measures based on information functionals defined in the preliminaries section.

Theorem 4 Let Inline graphic be a star on Inline graphic vertices. Let Inline graphic be the information functionals as defined before. The information measure is given by

graphic file with name pone.0031395.e084.jpg (10)

where Inline graphic is the probability of the central vertex of Inline graphic :

graphic file with name pone.0031395.e087.jpg (11)

if Inline graphic.

graphic file with name pone.0031395.e089.jpg (12)

if Inline graphic.

graphic file with name pone.0031395.e091.jpg (13)

if Inline graphic.

graphic file with name pone.0031395.e093.jpg (14)

if Inline graphic.

Proof:

  • Consider Inline graphic, where Inline graphic and Inline graphic for Inline graphic.

We get,

graphic file with name pone.0031395.e099.jpg (15)

Therefore,

graphic file with name pone.0031395.e100.jpg (16)

Hence,

graphic file with name pone.0031395.e101.jpg (17)

By substituting the value of Inline graphic in Inline graphic and simplifying, we get

graphic file with name pone.0031395.e104.jpg
  • Consider Inline graphic, where Inline graphic for Inline graphic.

We get,

graphic file with name pone.0031395.e108.jpg (18)

Therefore,

graphic file with name pone.0031395.e109.jpg (19)

Hence,

graphic file with name pone.0031395.e110.jpg (20)

By substituting the value of Inline graphic in Inline graphic and simplifying, we get

graphic file with name pone.0031395.e113.jpg
  • Consider the case Inline graphic, where Inline graphic, Inline graphic for Inline graphic.

graphic file with name pone.0031395.e118.jpg (21)

denotes the closeness centrality measure.

Then, we yield

graphic file with name pone.0031395.e119.jpg (22)

Therefore,

graphic file with name pone.0031395.e120.jpg (23)

Hence,

graphic file with name pone.0031395.e121.jpg (24)

By substituting the value of Inline graphic in Inline graphic and simplifying, we obtain

graphic file with name pone.0031395.e124.jpg

where Inline graphic.

  • Consider Inline graphic, where Inline graphic for Inline graphic. Inline graphic is defined via Equation (18). We get,
    graphic file with name pone.0031395.e130.jpg (25)
    Therefore,
    graphic file with name pone.0031395.e131.jpg (26)
    Thus,
    graphic file with name pone.0031395.e132.jpg (27)
    By substituting the value of Inline graphic in Inline graphic and simplifying, we get
    graphic file with name pone.0031395.e135.jpg (28)
    where Inline graphic. Inline graphic

By choosing particular values for the parameters involved, we get concrete measures using the above stated functionals. For example, consider the functional Inline graphic and set

graphic file with name pone.0031395.e139.jpg (29)

If we plug in those values in Equations (10) and (11), we easily derive

graphic file with name pone.0031395.e140.jpg (30)

Paths

Let Inline graphic be the path graph on Inline graphic vertices. Path graphs are the only trees with maximum diameter among all the trees on Inline graphic vertices, i.e., Inline graphic. We remark that to compute a closed form expression even for path graphs, is not always simple. To illustrate this, we present the concrete information measure Inline graphic by choosing particular values for its coefficients.

Lemma 5 Let Inline graphic be a path graph and consider the functional Inline graphic defined by Equation (6) . We set Inline graphic, Inline graphic. We yield

graphic file with name pone.0031395.e150.jpg (31)

Proof: Let Inline graphic be a path graph trivially labeled by Inline graphic, Inline graphic (from left to right).

Given Inline graphic with Inline graphic for Inline graphic.

By computing Inline graphic, when Inline graphic, for Inline graphic, we infer

graphic file with name pone.0031395.e160.jpg (32)
graphic file with name pone.0031395.e161.jpg (33)
graphic file with name pone.0031395.e162.jpg (34)

Therefore,

graphic file with name pone.0031395.e163.jpg (35)

and, hence,

graphic file with name pone.0031395.e164.jpg (36)

where Inline graphic, for Inline graphic. By substituting these quantities into Inline graphic yields the desired result.

Note that when using the same measure with arbitrary coefficients, its computation is intricate. In this regard, we present explicit bounds or information inequalities for any connected graph if the measure is based on the information functional using Inline graphic-spheres. That is, either Inline graphic or Inline graphic.

General connected graphs

Theorem 6 Given any connected graph Inline graphic on Inline graphic vertices and let Inline graphic given by Equation (5) . Then, we infer the following bounds:

graphic file with name pone.0031395.e175.jpg (37)
graphic file with name pone.0031395.e176.jpg (38)
graphic file with name pone.0031395.e177.jpg (38)
graphic file with name pone.0031395.e178.jpg (40)
graphic file with name pone.0031395.e179.jpg (41)

Proof: Consider Inline graphic, where Inline graphic and Inline graphic for Inline graphic. Let Inline graphic and Inline graphic. Recall (see Remark (2)) that, when either Inline graphic or when all the coefficients (Inline graphic) are equal, the information functional becomes constant and, hence, the value of Inline graphic equals Inline graphic. In the following, we will discuss the cases Inline graphic and Inline graphic, and we also assume that not all Inline graphic are equal.

Case 1: Inline graphic: We first construct the bounds for Inline graphic as shown below:

graphic file with name pone.0031395.e195.jpg (42)
graphic file with name pone.0031395.e196.jpg (43)

Similarly,

graphic file with name pone.0031395.e197.jpg (44)

Therefore, from the Equations (43) and (44), we get

graphic file with name pone.0031395.e198.jpg (45)

Hence,

graphic file with name pone.0031395.e199.jpg (46)

Let Inline graphic. Then, the last inequality can be rewritten as,

graphic file with name pone.0031395.e201.jpg (47)

Upper bound for Inline graphic:

Since Inline graphic and Inline graphic, we have Inline graphic. Hence, we have Inline graphic and Inline graphic. Thus we get,

graphic file with name pone.0031395.e208.jpg (48)

By adding over all the vertices of Inline graphic, we obtain

graphic file with name pone.0031395.e210.jpg (49)

Lower bound for Inline graphic:

We have to distinguish two cases, either Inline graphic or Inline graphic.

Case 1.1: Inline graphic. We yield Inline graphic. Therefore,

graphic file with name pone.0031395.e216.jpg (50)

By adding over all the vertices of Inline graphic, we get

graphic file with name pone.0031395.e218.jpg (51)

Case 1.2: Inline graphic.

In this case, we obtain Inline graphic and Inline graphic. Therefore, by using these bounds in Equation (4), we infer Inline graphic.

Case 2: Inline graphic:

Consider Equation (42). We get the following bounds for Inline graphic:

graphic file with name pone.0031395.e225.jpg (52)

Therefore,

graphic file with name pone.0031395.e226.jpg (53)

Hence,

graphic file with name pone.0031395.e227.jpg (54)

Set Inline graphic. Then, the last inequality can be rewritten as,

graphic file with name pone.0031395.e229.jpg (55)

Upper bound for Inline graphic:

Since Inline graphic and Inline graphic, we have Inline graphic. Hence, we have Inline graphic and Inline graphic. Thus, we obtain,

graphic file with name pone.0031395.e236.jpg (56)

By adding over all the vertices of Inline graphic, we get

graphic file with name pone.0031395.e238.jpg (57)

Lower bound for Inline graphic:

Again, we consider two cases, either Inline graphic or Inline graphic.

Case 2.1: Inline graphic.

In this case, we have Inline graphic and Inline graphic. Therefore, by substituting these bounds in the Equation (4), we obtain Inline graphic.

Case 2.2: Inline graphic.

We have Inline graphic. Therefore,

graphic file with name pone.0031395.e248.jpg (58)

By adding over all the vertices of Inline graphic, we get

graphic file with name pone.0031395.e250.jpg (59)

Hence, the theorem follows.

In the next theorem, we obtain explicit bounds when using the information functional given by Equation (6).

Theorem 7 Given any connected graph Inline graphic on Inline graphic vertices and let Inline graphic be given as in Equation (6) . We yield

graphic file with name pone.0031395.e255.jpg (60)
graphic file with name pone.0031395.e256.jpg (61)
graphic file with name pone.0031395.e257.jpg (62)
graphic file with name pone.0031395.e258.jpg (63)

Proof: Consider Inline graphic, where Inline graphic for Inline graphic. Let Inline graphic and Inline graphic. We have,

graphic file with name pone.0031395.e264.jpg (64)

Similarly,

graphic file with name pone.0031395.e265.jpg (65)

Therefore, from the Equations (64) and (65), we get

graphic file with name pone.0031395.e266.jpg (66)

Hence,

graphic file with name pone.0031395.e267.jpg (67)

Upper bound for Inline graphic:

Since Inline graphic, we have Inline graphic and Inline graphic. Hence,

graphic file with name pone.0031395.e272.jpg (68)

By adding over all the vertices of Inline graphic, we obtain

graphic file with name pone.0031395.e274.jpg (69)

Lower bound for Inline graphic:

Let us distinguish two cases:

Case 1: Inline graphic.

We have Inline graphic and Inline graphic. Therefore, by applying these bounds to Equation (4), we obtain Inline graphic.

Case 2: Inline graphic.

In this case, we have Inline graphic. Therefore,

graphic file with name pone.0031395.e282.jpg (70)

By adding over all the vertices of Inline graphic, we obtain the lower bound for Inline graphic given by

graphic file with name pone.0031395.e285.jpg (71)

Hence, the theorem follows.

Implicit Information Inequalities

Information inequalities describe relations between information measures for graphs. An implicit information inequality is a special type of an information inequality where the entropy of the graph is estimated by a quantity that contains another graph entropy expression. In this section, we will present some implicit information inequalities for entropy measures based on information functionals. In this direction, a first attempt has been done by Dehmer et al. [23], [24], [26]. Note that Dehmer et al. [23], [26] started from certain conditions on the probabilities when two different information functionals Inline graphic and Inline graphic are given. In contrast, we start from certain assumptions which the functionals themselves should satisfy and, finally, derive novel implicit inequalities. Now, given any graph Inline graphic. Let Inline graphic and Inline graphic be two mean information measures of Inline graphic defined using the information functionals Inline graphic and Inline graphic respectively. Let us further define another functional Inline graphic, Inline graphic. In the following, we will study the relation between the information measure Inline graphic and the measures Inline graphic and Inline graphic.

Theorem 8 Suppose Inline graphic, for all Inline graphic, then the information measure Inline graphic can be bounded by Inline graphic and Inline graphic as follows:

graphic file with name pone.0031395.e304.jpg (72)
graphic file with name pone.0031395.e305.jpg (73)

where Inline graphic, Inline graphic, and Inline graphic.

Proof: Given Inline graphic. Let Inline graphic and Inline graphic. Therefore Inline graphic. The information measures of Inline graphic with respect to Inline graphic and Inline graphic are given by

graphic file with name pone.0031395.e316.jpg (74)

where Inline graphic

graphic file with name pone.0031395.e318.jpg (75)

where Inline graphic.

Now consider the probabilities,

graphic file with name pone.0031395.e320.jpg (76)
graphic file with name pone.0031395.e321.jpg (77)
graphic file with name pone.0031395.e322.jpg (78)

Using Equation (77) and based on the fact that Inline graphic, we get

graphic file with name pone.0031395.e324.jpg (79)

Thus,

graphic file with name pone.0031395.e325.jpg (80)

and

graphic file with name pone.0031395.e326.jpg (81)

Since the last term in the above inequality is positive, we get

graphic file with name pone.0031395.e327.jpg (82)

By adding up the above inequalities over all the vertices of Inline graphic, we get the desired upper bound. From Equation (77), we also get a lower bound for Inline graphic, given by

graphic file with name pone.0031395.e330.jpg (83)

Now proceeding as before with the above inequality for Inline graphic, we obtain

graphic file with name pone.0031395.e332.jpg (84)
graphic file with name pone.0031395.e333.jpg (85)

By using the concavity property of the logarithm, that is, Inline graphic, we yield

graphic file with name pone.0031395.e335.jpg (86)

By adding the above inequality over all the vertices of Inline graphic, we get the desired lower bound. This proves the theorem.

Corollary 9 The information measure Inline graphic, for Inline graphic, is bounded by Inline graphic and Inline graphic as follows:

graphic file with name pone.0031395.e342.jpg (87)
graphic file with name pone.0031395.e343.jpg (88)

Proof: Set Inline graphic in Theorem (8), then the corollary follows.

Corollary 10 Given two information functionals, Inline graphic, Inline graphic such that Inline graphic, Inline graphic. Then

graphic file with name pone.0031395.e350.jpg (89)

Proof: Follows from Corollary (9).

The next theorem gives another bound for Inline graphic in terms of both Inline graphic and Inline graphic by using the concavity property of the logarithmic function.

Theorem 11 Let Inline graphic and Inline graphic be two arbitrary functionals defined on a graph Inline graphic. If Inline graphic for all Inline graphic, we infer

graphic file with name pone.0031395.e360.jpg (90)

and

graphic file with name pone.0031395.e361.jpg (91)

where Inline graphic, Inline graphic and Inline graphic.

Proof: Starting from the quantities for Inline graphic based on Equation (77), we obtain

graphic file with name pone.0031395.e366.jpg (92)
graphic file with name pone.0031395.e367.jpg (93)
graphic file with name pone.0031395.e368.jpg (94)
graphic file with name pone.0031395.e369.jpg (95)

Since each of the last two terms in Equation (95) is positive, we get a lower bound for Inline graphic, given by

graphic file with name pone.0031395.e371.jpg (96)

Applying the last inequality to Equation (4), we get the upper bound as given in Equation (91). By further applying the inequality Inline graphic to Equation (95), we get an upper bound for Inline graphic, given by

graphic file with name pone.0031395.e374.jpg (97)

Therefore,

graphic file with name pone.0031395.e375.jpg (98)

Finally, we now apply this inequality to Equation (4) and get the lower bound as given in Equation (90).

The next theorem is a straightforward extension of the previous statement. Here, an information functional is expressed as a linear combination of Inline graphic arbitrary information functionals.

Theorem 12 Let Inline graphic and Inline graphic be arbitrary functionals defined on a graph Inline graphic. Inline graphic are the corresponding information contents. If Inline graphic for all Inline graphic, we infer

graphic file with name pone.0031395.e383.jpg (99)

and

graphic file with name pone.0031395.e384.jpg (100)

where Inline graphic, Inline graphic for Inline graphic. Inline graphic

Union of Graphs

In this section, we determine the entropy of the union of two graphs. Let Inline graphic and Inline graphic be two arbitrary connected graphs on Inline graphic and Inline graphic vertices, respectively. Let Inline graphic be an information functional defined on these graphs denoted by Inline graphic, Inline graphic and let Inline graphic and Inline graphic be the information measures on Inline graphic and Inline graphic respectively.

Theorem 13 Let Inline graphic be the disjoint union of the graphs Inline graphic and Inline graphic. Let Inline graphic be an arbitrary information functional. The information measure Inline graphic can be expressed in terms of Inline graphic and Inline graphic as follows:

graphic file with name pone.0031395.e407.jpg (101)

where Inline graphic with Inline graphic and Inline graphic.

Proof: Let Inline graphic be the given information functional. Let Inline graphic and Inline graphic. The information measures of Inline graphic and Inline graphic are given as follows:

graphic file with name pone.0031395.e416.jpg (102)

where Inline graphic, and

graphic file with name pone.0031395.e418.jpg (103)
graphic file with name pone.0031395.e419.jpg
graphic file with name pone.0031395.e420.jpg (104)

Hence,

graphic file with name pone.0031395.e421.jpg (105)
graphic file with name pone.0031395.e422.jpg (106)
graphic file with name pone.0031395.e423.jpg (107)

Using these quantities to determine Inline graphic, we obtain

graphic file with name pone.0031395.e425.jpg (108)

and

graphic file with name pone.0031395.e426.jpg (109)

Upon simplification, we get the desired result.

Also, we immediately obtain a generalization of the previous theorem by taking Inline graphic-disjoint graphs into account.

Theorem 14 Let Inline graphic, Inline graphic be Inline graphic arbitrary connected graphs on Inline graphic vertices, respectively. Let Inline graphic be an information functional defined on these graphs denoted by Inline graphic. Let Inline graphic be the disjoint union of the graphs Inline graphic for Inline graphic. The information measure Inline graphic can be expressed in terms of Inline graphic, Inline graphic as follows:

graphic file with name pone.0031395.e441.jpg (110)

where Inline graphic with Inline graphic for Inline graphic.

Join of Graphs

Let Inline graphic and Inline graphic be two arbitrary connected graphs on Inline graphic and Inline graphic vertices, respectively. The join of the graphs Inline graphic is defined as the graph Inline graphic with vertex set Inline graphic and the edge set Inline graphic. Let Inline graphic be the information functional (given by Equation (5)) based on the Inline graphic-sphere functional (exponential) defined on these graphs and denoted by Inline graphic, Inline graphic. Let Inline graphic and Inline graphic be the information measures on Inline graphic and Inline graphic respectively.

Theorem 15 Let Inline graphic be the join of the graphs Inline graphic and Inline graphic with Inline graphic vertices. The information measure Inline graphic can then be expressed in terms of Inline graphic and Inline graphic as follows:

graphic file with name pone.0031395.e469.jpg (111)

where Inline graphic for Inline graphic, Inline graphic and Inline graphic with Inline graphic and Inline graphic.

Proof: Let Inline graphic be the join of two connected graphs Inline graphic and Inline graphic. Here, Inline graphic. Let Inline graphic be the information functional defined by using the Inline graphic-sphere functional on Inline graphic. Let Inline graphic and Inline graphic. The information measures of Inline graphic and Inline graphic are given as follows:

graphic file with name pone.0031395.e487.jpg (112)
graphic file with name pone.0031395.e488.jpg
graphic file with name pone.0031395.e489.jpg (113)
graphic file with name pone.0031395.e490.jpg
graphic file with name pone.0031395.e491.jpg (114)
graphic file with name pone.0031395.e492.jpg (115)

Hence,

graphic file with name pone.0031395.e493.jpg (116)
graphic file with name pone.0031395.e494.jpg (117)
graphic file with name pone.0031395.e495.jpg (118)

Using those entities to determine Inline graphic, we infer

graphic file with name pone.0031395.e497.jpg (119)

and

graphic file with name pone.0031395.e498.jpg (120)

Upon simplification, we get the desired result.

If we consider the linear Inline graphic-sphere functional Inline graphic (see Equation (6)), to infer an exact expression for the join of two graphs as in Theorem (15) is an intricate problem. By Theorem (16) and Theorem (17), we will now present different bounds in terms of Inline graphic and Inline graphic.

Theorem 16 Let Inline graphic be the join of the graphs Inline graphic and Inline graphic on Inline graphic vertices. Then, we yield

graphic file with name pone.0031395.e508.jpg (121)

where Inline graphic for Inline graphic, Inline graphic and Inline graphic with Inline graphic and Inline graphic.

Proof: Let Inline graphic and Inline graphic. The information measures of Inline graphic and Inline graphic are given as follows:

graphic file with name pone.0031395.e519.jpg (122)
graphic file with name pone.0031395.e520.jpg
graphic file with name pone.0031395.e521.jpg (123)
graphic file with name pone.0031395.e523.jpg (124)
graphic file with name pone.0031395.e524.jpg (125)

Hence,

graphic file with name pone.0031395.e525.jpg (126)
graphic file with name pone.0031395.e526.jpg (127)
graphic file with name pone.0031395.e527.jpg (128)

Since Inline graphic and Inline graphic are positive, we get a lower bound for Inline graphic given as

graphic file with name pone.0031395.e531.jpg (129)

To infer a lower bound for the information measure Inline graphic, we start from the Equations (128), (129) and obtain

graphic file with name pone.0031395.e533.jpg (130)
graphic file with name pone.0031395.e534.jpg (131)
graphic file with name pone.0031395.e535.jpg (132)

By using the inequality Inline graphic and performing simplification steps, we get,

graphic file with name pone.0031395.e537.jpg (133)

By adding up the above inequality system (across all the vertices of Inline graphic) and by simplifying, we get the desired lower bound.

Further, an alternate set of bounds can be achieved as follows.

Theorem 17 Let Inline graphic be the join of the graphs Inline graphic and Inline graphic on Inline graphic vertices. Then, we infer

graphic file with name pone.0031395.e543.jpg (134)

and

graphic file with name pone.0031395.e544.jpg (135)

where Inline graphic for Inline graphic, Inline graphic and Inline graphic with Inline graphic and Inline graphic.

Proof: Starting from Theorem (16), consider the value of Inline graphic given by Equation (128). By using the quantities for Inline graphic to calculate Inline graphic, we get

graphic file with name pone.0031395.e554.jpg (136)

and

graphic file with name pone.0031395.e555.jpg (137)

By simplifying and performing summation, we get

graphic file with name pone.0031395.e556.jpg (138)

An upper bound for the measure Inline graphic can be derived as follows:

graphic file with name pone.0031395.e558.jpg (139)

since each of the remaining terms in Equation (138) is positive. Finally, we infer the lower bound for Inline graphic as follows. By applying inequality Inline graphic to Equation (138), we get

graphic file with name pone.0031395.e561.jpg (140)

Upon simplification, we get

graphic file with name pone.0031395.e562.jpg (141)

Putting Inequality (139) and Inequality (141) together finishes the proof of the theorem.

Summary and Conclusion

In this article, we have investigated a challenging problem in quantitative graph theory namely to establish relations between graph entropy measures. Among the existing graph entropy measures, we have considered those entropies which are based on information functionals. It turned out that these measures have widely been applicable and useful when measuring the complexity of networks [3].

In general, to find relations between quantitative network measures is a daunting problem. The results could be used in various branches of science including mathematics, statistics, information theory, biology, chemistry and social sciences. Further, the determination of analytical relations between measures is of great practical importance when dealing with large scale networks. Also, relations involving quantitative network measures could be fruitful when determining the information content of large complex networks.

Note that our proof technique follows the one proposed in [23]. It is based on three main steps: Firstly, we compute the information functionals and in turn, we calculate the probability values for every vertex of the graph in question. Secondly, we start with certain conditions for the computed functionals and arrive at a system of inequalities. Thirdly, by adding up the corresponding inequality system, we obtain the desired implicit information inequality. Using this approach, we have inferred novel bounds by assuming certain information functionals. It is evident that further bounds could be inferred by taking novel information functionals into account. Further, we explored relations between the involved information measures for general connected graphs and for special classes of graphs such as stars, path graphs, union and join of graphs.

At this juncture, it is also relevant to compare the results proved in this paper with those proved in [23]. While we derived the implicit information inequalities by assuming certain properties for the functionals, the implicit information inequalities derived in [23] are based on certain conditions for the calculated vertex probabilities. Interestingly, note that by using Theorem (11) and Theorem (17), the range of the corresponding bounds is very small. We inferred that the difference between the upper and lower bound equals Inline graphic.

As noted earlier, relations between entropy-based measures for graphs have not been extensively explored so far. Apart from the results we have gained in this paper, we therefore state a few open problems as future work:

  • To find relations between Inline graphic and Inline graphic, when Inline graphic is an induced subgraph of Inline graphic and Inline graphic is an arbitrary information functional.

  • To find relations between Inline graphic and Inline graphic, where Inline graphic, Inline graphic are so-called generalized trees, see [34]. Note that it is always possible to decompose an arbitrary, undirected graph into a set of generalized trees [34].

  • To find relations between measures based on information functionals and the other classical graph measures.

  • To derive information inequalities for graph entropy measures using random graphs.

  • To derive statements to judge the quality of information inequalities.

Footnotes

Competing Interests: The authors have declared that no competing interests exist.

Funding: Matthias Dehmer and Lavanya Sivakumar thank the Austrian Science Fund (Project No. PN22029-N13) for supporting this work. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Basak SC. Devillers J, Balaban AT, editors. Information-theoretic indices of neighborhood complexity and their applications. Topological Indices and Related Descriptors in QSAR and QSPAR, Gordon and Breach Science Publishers. 1999. pp. 563–595. Amsterdam, The Netherlands.
  • 2.Wang J, Provan G. Zhou J, editor. Characterizing the structural complexity of real-world complex networks. 2009. pp. 1178–1189. Complex Sciences, Springer, Berlin/Heidelberg, Germany, volume 4 of Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering.
  • 3.Dehmer M, Mowshowitz A. A history of graph entropy measures. Information Sciences. 2011;181:57–78. [Google Scholar]
  • 4.Li M, Vitányi P. An Introduction to Kolmogorov Complexity and Its Applications. Springer; 1997. [Google Scholar]
  • 5.Mowshowitz A. Entropy and the complexity of graphs: I. an index of the relative complexity of a graph. Bulletin of Mathematical Biophysics. 1968;30:175–204. doi: 10.1007/BF02476948. [DOI] [PubMed] [Google Scholar]
  • 6.Shannon CE. A mathematical theory of communication. Bell System Technical Journal. 1948;27:379–423 and 623–656. [Google Scholar]
  • 7.Bonchev D, Rouvray DH. Complexity in chemistry: Introduction and Fundamentals. Mathematical and Computational Chemistry 7. New York: CRC Press; 2003. [Google Scholar]
  • 8.Bonchev D. Information Theoretic Indices for Characterization of Chemical Structures. 1983. Research Studies Press, Chichester.
  • 9.Dehmer M. Information processing in complex networks: graph entropy and information functionals. Appl Math Comput. 2008;201:82–94. [Google Scholar]
  • 10.Emmert-Streib F, Dehmer M. Information theoretic measures of UHG graphs with low computational complexity. Appl Math Comput. 2007;190:1783–1794. [Google Scholar]
  • 11.Mehler A, Weiß P, Lücking A. A network model of interpersonal alignment. Entropy. 2010;12:1440–1483. [Google Scholar]
  • 12.Bonchev D. Complexity in Chemistry. 2003. Introduction and Fundamentals. Taylor and Francis. Boca Raton, FL, USA.
  • 13.Anand K, Bianconi G. Entropy measures for networks: Toward an information theory of complex topologies. Phys Rev E. 2009;80:045102. doi: 10.1103/PhysRevE.80.045102. [DOI] [PubMed] [Google Scholar]
  • 14.Costa LdF, Rodrigues FA, Travieso G, Boas PRV. Characterization of complex networks: A survey of measurements. Advances in Physics. 2007;56:167–242. [Google Scholar]
  • 15.Kim J, Wilhelm T. What is a complex graph? Physica A: Statistical Mechanics and its Applications. 2008;387:2637–2652. [Google Scholar]
  • 16.Balaban AT, Balaban TS. New vertex invariants and topological indices of chemical graphs based on information on distances. Journal of Mathematical Chemistry. 1991;8:383–397. [Google Scholar]
  • 17.Bertz SH. King R, editor. A mathematical model of complexity. 1983. pp. 206–221. Chemical applications of topology and graph theory, Elsevier, Amsterdam.
  • 18.Basak SC, Magnuson VR, Niemi GJ, Regal RR. Determining structural similarity of chemicals using graph-theoretic indices. Discrete Applied Mathematics. 1988;19:17–44. [Google Scholar]
  • 19.Bonchev D, Rouvray DH. Complexity in chemistry, biology, and ecology. Mathematical and Computational Chemistry. New York: Springer; 2005. pp xx+344. doi: http://dx.doi.org/10.1007/b136300. [Google Scholar]
  • 20.Claussen JC. Offdiagonal complexity: A computationally quick complexity measure for graphs and networks. Physica A: Statistical Mechanics and its Applications. 2007;375:365–373. [Google Scholar]
  • 21.Körner J. Coding of an information source having ambiguous alphabet and the entropy of graphs. 1973. pp. 411–425. Trans 6th Prague Conference on Information Theory.
  • 22.Butts C. The complexity of social networks: Theoretical and empirical findings. Social Networks. 2001;23:31–71. [Google Scholar]
  • 23.Dehmer M, Mowshowitz A. Inequalities for entropy-based measures of network information content. Applied Mathematics and Computation. 2010;215:4263–4271. [Google Scholar]
  • 24.Dehmer M, Mowshowitz A, Emmert-Streib F. Connections between classical and parametric network entropies. PLoS ONE. 2011;6:e15733. doi: 10.1371/journal.pone.0015733. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Bonchev D, Trinajstic N. Information theory, distance matrix, and molecular branching. The Journal of Chemical Physics. 1977;67:4517–4533. [Google Scholar]
  • 26.Dehmer M, Borgert S, Emmert-Streib F. Entropy bounds for hierarchical molecular networks. PLoS ONE. 2008;3:e3079. doi: 10.1371/journal.pone.0003079. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Skorobogatov VA, Dobrynin AA. Metrical analysis of graphs. MATCH Commun Math Comp Chem. 1988;23:105–155. [Google Scholar]
  • 28.Shannon C, Weaver W. The Mathematical Theory of Communication. 1997. University of Illinois Press, Urbana, IL, USA. [PubMed]
  • 29.Freeman LC. A set of measures of centrality based on betweenness. Sociometry. 1977;40:35–41. [Google Scholar]
  • 30.Freeman LC. Centrality in social networks conceptual clarification. Social Networks. 1978;1:215–239. [Google Scholar]
  • 31.Sabidussi G. The centrality index of a graph. Psychometrika. 1966;31:581–603. doi: 10.1007/BF02289527. [DOI] [PubMed] [Google Scholar]
  • 32.Emmert-Streib F, Dehmer M. Networks for systems biology: Conceptual connection of data and function. IET Systems Biology. 2011;5:185–207. doi: 10.1049/iet-syb.2010.0025. [DOI] [PubMed] [Google Scholar]
  • 33.Simonyi G. Cook W, Lovász L, Seymour P, editors. Graph entropy: A survey. 1995. pp. 399–441. Combinatorial Optimization, DIMACS Series in Discrete Mathematics and Theoretical Computer Science, volume 20.
  • 34.Emmert-Streib F, Dehmer M, Kilian J et al HRA, editor. Classification of large graphs by a local tree decomposition. 2006. pp. 477–482. Proceedings of DMIN'05, International Conference on Data Mining, Las Vegas, USA.

Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES