Abstract
Information theory provides robust measures of multivariable interdependence, but classically does little to characterize the multivariable relationships it detects. The Partial Information Decomposition (PID) characterizes the mutual information between variables by decomposing it into unique, redundant, and synergistic components. This has been usefully applied, particularly in neuroscience, but there is currently no generally accepted method for its computation. Independently, the Information Delta framework characterizes non-pairwise dependencies in genetic datasets. This framework has developed an intuitive geometric interpretation for how discrete functions encode information, but lacks some important generalizations. This paper shows that the PID and Delta frameworks are largely equivalent. We equate their key expressions, allowing for results in one framework to apply towards open questions in the other. For example, we find that the approach of Bertschinger et al. is useful for the open Information Delta question of how to deal with linkage disequilibrium. We also show how PID solutions can be mapped onto the space of delta measures. Using Bertschinger et al. as an example solution, we identify a specific plane in delta-space on which this approach’s optimization is constrained, and compute it for all possible three-variable discrete functions of a three-letter alphabet. This yields a clear geometric picture of how a given solution decomposes information.
Keywords: partial information decomposition, information delta, synergy, co-information, non-pairwise dependence
1. Introduction
The variables in complex biological data frequently have nonlinear and non-pairwise dependency relationships. Understanding the functions and/or dysfunctions of biological systems requires understanding these complex interactions. How can we reliably detect interdependence within a set of variables, and how can we distinguish simple, pairwise dependencies from those which are fundamentally multivariable?
An analytical approach formulated by Williams and Beer frames these questions in terms of the Partial Information Decomposition (PID) [1]. The PID proposes to decompose the mutual information between a pair of source variables X and Y and a target variable Z, , into four non-negative components:
(1) |
The constituent terms proposed are: the unique informations, and , which represent the amounts of information about Z encoded by X alone and by Y alone; the redundant information, R, which is the information about Z encoded redundantly by both X and Y; and the synergistic information S, which is the information about Z contained in neither X or Y individually, but encoded by X and Y taken together. An illustration of this decomposition, the associated governing equations, and examples characterizing each type of information are all shown in Figure 1. It was shown that PID components can distinguish between dyadic and triadic relationships which no conventional Shannon information measure can distinguish [2].
The problem with this approach is that its governing equations form an underdetermined system, with only three equations relating the four components. To actually calculate the decomposition, an additional assumption must be made to provide an additional equation. Williams and Beer proposed a method for the calculation of R in their original paper, but this has since been shown to have some undesirable properties [1]. Much of the subsequent work in this domain consisted of attempts to define new relationships or formulae to calculate the components, as well as critiques of these proposed measures [3]. These proposed measures include (as an incomplete list): a measure based on information geometry [4]; an intersection information based on the Gács-Körner common information [5]; the minimum mutual information [6]; the pointwise common change in surprisal [7]; and the extractable shared information [8].
Another noteworthy putative solution is that of [9], which requires solving an optimization problem over a space Q of probability distributions, but is rigorous in that it directly follows from reasonable assumptions about the unique information. However, it is unclear how to sensibly generalize this approach to larger numbers of variables. Nonetheless, there has been considerable interest in using the PID approach to gather insights from real datasets, particularly within the neuroscience community [10,11,12,13,14,15].
Independently, an alternative approach to many of these same questions has been formulated focusing on devising new information theory-based measures of multivariable dependency. In genetics, non-pairwise epistatic effects are often crucially important in determining complex phenotypes, but traditional methods are sensitive only to pairwise relationships; thus there is particular interest in methods to identify the existence of synergistic dependencies within genetic datasets. Galas et al. [16,17] quantified the non-pairwise information between genetic loci and phenotype data with the Delta measure, . Briefly, given a set of variables , quantifies the change in co-information when considering the variables as opposed to only (we hereafter denote as , as , and so on). In its simplest application, the magnitudes of can be used to detect and quantify non-pairwise interactions [16,17].
Recent work showed that the delta values encode considerable additional information about the dependency. Sakhanenko et al. [18] defined the normalized delta measures , which define an “information space”, and considered the -values of all possible discrete functions . Fully mapping the specific example set of functions where are all discrete variables with 3 possible values, they found that the 19,683 possible functional relationships mapped onto a highly structured plane in the space of normalized deltas (as shown in Figure 2). Different regions of this plane corresponded to qualitatively different types of functional relationships; in particular, completely pairwise functions such as and completely non-pairwise functions such as were mapped onto the extremes of the plane (see Figure 2; note that this paper defines “XOR” for a ternary alphabet as ). Since discrete variables such as these occur naturally in genetics, this suggests that relationships between genetic variables may be usefully characterized by their -coordinates, with useful intuitive value. The difficulty of this in practice is that the coordinates are constrained to this plane only when X and Y are statistically independent, which is not the case in many real datasets, e.g., in genetic datasets in the presence of linkage disequilibrium.
In this paper, we show that the Partial Information Decomposition approach and Information Delta approach are largely equivalent, since their component variables can be directly related. The -coordinates can be written explicitly in terms of PID components, which leads us to an intuitive understanding of how -space encodes PID information by casting them into a geometric context. We then apply our results to two different approaches to solving the PID problem, first one from Bertschinger et al. [9] and then from Finn and Lizier [19]. We show that the sets of probability distributions, Q, used by Bertschinger can be mapped onto low-dimensional manifolds in -space, which intersect with the -plane of Figure 2. This approach is theoretically useful for the Delta information framework, since it factors out dependence in the data, thereby accounting for linkage disequilibrium between genetic variables. We suggest an approach for the analysis of genetic datasets which would return both the closest discrete function underlying the data and its PID in the Bertschinger solution, and which would require no further optimization after the initial construction of a solution library. This realization thus yields a low-dimensional geometric interpretation of this optimization problem, and we compute the solution for all possible three-variable discrete functions of alphabet size three. For these same functions, we then compute the PID components using the Pointwise PID approach of Finn and Lizier [19]. This visualization yields an immediate comparison of how each solution decomposes information. Since our derived relationship between the frameworks is general, it could be similarly applied to any putative PID solution as demonstrated here. Code to replicate these computations and the associated figures is freely available [20].
2. Background
2.1. Interaction Information and Multi-Information
An important body of background work, which served as a foundation for both the Information Decomposition and Information Delta approaches, involves the Interaction Information, . can be thought of as a multivariable extension of the mutual information [21]. Unlike the mutual information, however, the interaction information can assume negative values. What does it mean for the interaction information to be negative? It was once common to interpret as implying a synergistic interaction, and as implying a redundant interaction between the variables. As detailed in [1] and discussed in the following sections, this interpretation is mistaken. Interactions can be both partly synergistic and partly redundant, and the interaction information indicates the balance of these components.
For a set of variables , can be defined as [22]:
(2) |
where is the total number of variables in the set, and the sum is over all possible subsets (where is the total number of variables in each subset). is the joint entropy between the variables in subset . The interaction information, , is very similar to a measure called the co-information, [23]. These measures differ only by their sign: for an even number of variables they are identical (e.g., ), and for an odd number of variables they are of opposite sign.
(3) |
An additional, useful measure is the “multi-information”, , introduced by Watanabe [24], sometimes called the “total correlation”, which represents the sum of all dependencies of variables and is zero only if all variables are independent. For a set of n variables it is defined as:
(4) |
2.2. Information Decomposition
Consider a pair of “source variables” X,Y which determine the value of a “target variable” Z. Assume that we can measure the mutual information each source carries about a target, and (which we abbreviate as and ), as well as the mutual information between the joint distribution of and Z, (which we abbreviate as ). These mutual informations can be written in terms of the entropies (which we abbreviate using subscripts, e.g., ):
(5) |
These mutual informations can be decomposed into components which measure how much of each “type” of information they contain, as follows:
(6) |
where and are the unique informations, R is the redundant information, and S is the synergistic information, as described previously in Section 1. This is an underdetermined system which requires an additional equation for the variables to render it solvable. Many of the current and previous efforts to define such an equation (for example, several proposals on how to directly compute the value of R from data), as well as the limitations of those efforts, have been nicely summarized in [3].
2.3. Solution from Bertschinger et al.
One solution to this problem came from Bertschinger et al. [9], who proposed that the unique information be approximated as:
(7) |
Let be the set of all joint probability distributions of X, Y, and Z. Then we define Q as the set of all distributions, q, which have the same marginal probability distributions and as our dataset. That is,
(8) |
Please note that in [9], this set of probability distributions is denoted as , which we change here to Q to avoid notational confusion with the information deltas. Similarly, its elements are indicated by Q in the original paper. Here we indicate the distributions, elements of the set Q, by a lowercase q for consistency with our notation for probability distributions.
Put another way, we consider all possible probability distributions that maintain the marginals and implied by our data. The relationship between X and Y (, and consequently the joint distribution ) is allowed to vary. The minimization criterion is perhaps more intuitive when written, equivalently, as:
(9) |
Thus, the unique information can be thought of as the smallest possible increase in the interaction information when the variable X is added to the set . For example, if there exists a probability distribution in Q for which , then the addition of X adds no unique information about Z and . The core assumption of this approach is that the unique and redundant informations depend simply upon the marginal distributions and . This solution is rigorous in the sense that the result follows directly from this assumption without any ad-hoc assumptions for how the components are related.
2.4. Information Deltas and Their Geometry
Consider a set of three variables . Using Equation (3), we can write the co-information in terms of the entropies:
(10) |
The differential interaction information is the change in the interaction information when a given variable Z is added to the set. This can be written in terms of and then the conditional mutual information:
(11) |
These measures can be normalized by the multi-information for the three variables, (which we abbreviate as ), which by Equation (4) we can write in terms of the entropies as:
(12) |
The normalized measures are then:
(13) |
If Z is a function of X and Y, and if X and Y are i.i.d., then lies within a highly structured plane, where different regions of the plane correspond to qualitatively different types of interactions. Figure 2 shows the mapping of all possible functions onto this highly structured plane.
The normalized deltas can be expressed as:
(14) |
The normalized deltas can also be written in terms of joint mutual informations, as follows:
(15) |
We can write all normalized deltas in this form:
(16) |
By inverting previous equations, we can then write:
(17a) |
(17b) |
(17c) |
Specifically, Equation Set (16) can be inverted to yield Equation (17a), and Equation Set (14) can be inverted to yield Equations (17b) and (17c).
3. PID Mapped into Information Deltas
3.1. Information Decomposition in Terms of Deltas
With Equations (6) and (17), we can equate the expressions for the mutual informations in their delta and information decomposition forms:
(18) |
From the above relations we can derive:
(19) |
In other words, the difference between the synergy and the redundancy increases as we get farther from the origin in -space. Also:
(20) |
so the distance from the diagonal in the -plane is proportional to the difference between the unique informations. These striking relationships are visualized in Figure 3.
3.2. Relationship between Diagonal and Interaction Information
Considering again Equation (19) and using Equation (13), we can write:
(21) |
where is the interaction information between the variables. This replicates the important result that from the original Williams and Beer paper [1].
3.3. The Function Plane
When the variables are related by a discrete function (as defined in [18]), and X and Y are i.i.d., the function will lie on a plane defined by:
(22) |
Thus, the distance d of a coordinate above the plane is given by
(23) |
And so from Equation (18):
(24) |
4. Solving the PID on the Function Plane
4.1. Transforming Probability Tensors within Q
As noted previously, there is no generally accepted solution for completing and computing the set of PID equations. Our results connecting the PID to the information deltas have therefore, up to this point, been agnostic on this question. All equations in the previous section follow from the basic PID formulation, and the delta coordinate equations. This means they are true for any putative solution, but also brings us no closer to an actual solution to the PID problem; we can still only compute the differences between PID components. We therefore now extend our analysis by using the solution of Bertschinger et al. [9] to fully compute the PID for the functions in Figure 2. We wish to emphasize, however, that the following approach could be used equally well to gain a geometric interpretation of any alternate solution to the PID.
Consider a probability tensor for an alphabet size of N:
(25) |
where we use the notation . What transformations are permissible that will preserve the distribution within the set Q (as defined in Equation (8))? Please note that we can obtain the marginal distributions simply by summing over the appropriate tensor index. For example, summing along the first index yields the marginal distribution . To stay in Q, then, we require that the sums along the first and second indices both remain constant.
For an alphabet size of , we can parameterize the set of all possible transformations quite simply:
(26) |
All possible changes to each layer of the tensor can be captured with a single parameter. For example, increasing will require that and be decreased by the same amount, as the row and column sums must remain constant (which, in turn, determines ). Each layer can be modified independently, and thus the second layer has an independent parameter.
For a given probability tensor with , then, the probability tensor for any distribution in Q can be fully parameterized with two parameters, and thus the corresponding coordinates in delta-space are at most two-dimensional. In practice, we find that functions have delta-coordinates that are restricted to a one-dimensional manifold.
Consider, for example, the AND function:
(27) |
We can describe all possible perturbations which remain in Q by the parameterization:
(28) |
However, it can be seen that we must have , as all probabilities must remain in the range . The parameter , on the other hand, can fall within the range . Since all possible perturbations can be captured by varying a single parameter, Q must therefore be mapped to a one-dimensional manifold in -space.
The layers of a probability tensor become significantly harder to parameterize for . Consider a single layer of a probability tensor:
(29) |
The permissible transformations to this layer can be parameterized by:
(30) |
subject to the constraints that:
(31) |
Clearly, these relations are too complicated to lend any immediate insight into the problem. However, it is a simple matter to use the above inequalities to calculate permissible values of parameters and to plot out the corresponding delta coordinates. This is done for randomly generated sample functions in Figure 4. In this case, the delta coordinates have a complex distribution but are nonetheless restricted to a plane in delta-space (the vertically oriented red plane in Figure 5).
4.2. -Coordinates in Q Are Always Restricted to a Plane
In the case, delta-coordinates were parameterized by a single variable such that they must be restricted onto a line. In the example, they are restricted onto a plane. Will larger alphabets map Q onto a three-dimensional volume? If not, is it possible to get a non-planar two-dimensional manifold, or are coordinates always restricted to a plane? We will now prove that Q is always constrained to a plane, regardless of the alphabet size.
Lemma 1.
In any set Q as defined previously, the following entropies remain constant: all individual entropies , and ; the joint, 2-variable entropies containing Z, namely and . The only entropies which vary within a particular Q then are and .
Proof.
The definition of Q preserves the marginal distributions by construction. and being constant is a trivial consequence of holding and constant, which is the condition defining Q. From these constant marginal distributions, we can calculate the distributions , and , which are therefore also constant, as are their corresponding entropies. □
Only two entropic quantities vary between the different distributions in Q. By considering just their effect on the delta coordinates, we can now show the following:
Theorem 1.
In any set Q of distributions with equal marginal distributions and , the delta-coordinates will be restricted to a plane. This is true for any alphabet size.
Proof.
We begin by making several notational definitions to simplify the algebra which follows, first from the joint entropies which vary within Q:
We then define quantities which collect the constant entropy terms:
In terms of these quantities we can now write the normalized delta coordinates as follows:
Solving for d in the and equations yields:
And the equation allows us to solve for h:
Plugging this into the equation above yields an equation which simplifies to:
(32) Since and are all constant over Q, this defines a plane in space. □
Equation (32) not only shows that the points in Q are bound to a plane, but it also implies that this plane always contains the line defined by and . Therefore for any function in Figure 2, we can trivially compute the plane in which the corresponding Q is contained.
4.3. PID Calculation for All Functions
For the set of probability distributions Q, Bertschinger et al. [9] provide the following estimators for the PID components:
(33) |
If we numerically compute the set Q for a given function f (i.e., by generating a distribution such as the one shown in Figure 4 via the parameterization of Equation (30)), these estimators are trivially consistent. Figure 6 shows the computed values of the PID components for all of the functions shown in Figure 2. There is a clear geometric interpretation here: Functions in the lower left/right corners consist almost entirely of and , respectively. Functions approaching the top corner become increasingly synergistic with a higher proportion of S. Functions are most redundant towards the lower center of the plane, though no single function is primarily R.
4.4. Alternate Solutions: Pointwise PID
The Pointwise Partial Information Decomposition (PPID) of Finn and Lizier [19] is an alternate approach to solving the PID problem. It is motivated by the fact that the entropy and mutual information can be expressed as the expectation value of pointwise quantities, which measure the information content of a single event. For example, the event has the associated pointwise mutual information:
(34) |
and the overall mutual information between the two variables is the expectation value of this pointwise quantity, taken over all possible events. It is important to note that while the overall mutual information is non-negative, the pointwise mutual information can be negative. Finn and Lizier decompose this pointwise quantity into two non-negative components, the “specificity” and “ambiguity” , and argue that:
(35) |
They similarly decompose the redundancy R into a pointwise specific redundancy and pointwise specific ambiguity , and argue for the following definitions:
(36) |
where are the values of each of the source variables in a particular realization (e.g., if we have two source variables predicting Z, then the event has ). The expectation value of the difference of these quantities then yields the redundancy:
(37) |
from which the rest of the PID components follow. See [19] for a full discussion of motivations and Axioms which these definitions satisfy (including a discussion of the relationship between this formulation and that of Bertschinger et al. [9], and how the many aspects of [19] are arguably pointwise adaptations of the assumptions in [9]).
One consequence of this approach is that the PID components are no longer non-negative. There is extensive discussion of the interpretation of this in [19], but one example, RdnErr, is particularly informative. In our probability tensor notation, we can write this function as:
(38) |
This can be interpreted as follows: X is always equal to Z. Y is usually equal to Z, but occasionally (with probability ) makes an error. What should we expect the PID components to be, in this case? The PPID yields , which implies the following interpretation: the information about Z is encoded redundantly by both X and Y, but Y carries unique misinformation about Z due to its tendency to make errors. If all components were strictly positive, we would likely draw a different conclusion: both X and Y encode some information about Z, with X encoding additional unique information. In this way, different solutions will lead to slightly different interpretations about the nature of the relationship between the variables.
In Figure 7, we compute the PPID for all functions and map them onto -space, just as we did in Figure 6. Comparing Figure 6 and Figure 7 immediately highlights key differences in how each method decomposes information. For example: in Figure 6, the top corner is purely synergistic, the lower-left corner has information solely in X, and the lower-right corner has information solely in Y; in Figure 7, the top corner has zero redundancy, the lower-left corner has misinformation in Y, and the lower-right corner has misinformation in X.
It is not our goal here to argue which result is more correct. Instead, we wish to highlight how comparing Figure 6 and Figure 7 readily yields subtle insights into how the two approaches differ in decomposing information. It also yields immediate insights into the subtleties of how we might interpret coordinates in -space.
5. Conclusions
The key overall result of this paper is that the PID problem can be mapped directly into the the previously defined “information landscape” represented by the “delta space” of [18]. This theoretical framework is simple and has a geometric interpretation which was well worked out previously. The simple set of relations between the frameworks, as explicated in Equation (18) and visualized in Figure 3, anticipates a much deeper set of geometric constraints.
We build upon this general relationship using the solution of Bertschinger et al. [9]. Using this solution, we parameterize the permissible transformations to a discrete function to numerically generate the distribution set Q, and prove in Theorem 1 that this set is mapped onto a plane in delta-space. The optimization problem defined by this approach is cast in terms of our variables in Equation (33), and the various extrema can be extracted directly from our parameterization and mapping procedure. Code which replicates these computations and generates the figures within this paper is freely available[20].
These results suggest the following approach for computation of the PID components, if using the solution from [9], and given the added assumption that there is some function which best approximates the relationship between variables. The steps are these:
Construct a library (set) of distributions for all functions, . Specifically, record the -coordinates spanned by each distribution (e.g., as plotted in Figure 4) along with the corresponding function and its PID component values.
For a set of variables in data for which we wish to find the decomposition, compute its -coordinates and then match them to the closest . This will then immediately yield the corresponding function and PID components.
If this approach proves to be practical, it would have several clear advantages. First, the computational cost of the library construction would only need to be done once, and not need to be repeated for any subsequent analysis. The cost of the library construction is itself quite tractable (for example, exactly this computation was done to generate Figure 6). Second, this solves an open problem in the use of Information Deltas for which the source variables are not independent, for example, in applications to genetics in the presence of linkage disequilibrium. Specifically, this approach relaxes the common assumption in [18] that X and Y must be statistically independent.
The practical application of this approach to data analysis requires further development, which is beyond the scope of this paper. Specifically, the actual data will contain noise such that the computed -coordinates will not lie perfectly within any distribution of Q set. The naïve approach of simply taking the closest may therefore be insufficient in general. Future work will characterize the response of -coordinates to various levels of noise within the data, to enable the computation of (i.e., the probability that variables belong to the set given their observed coordinates and some noise level ).
Future work will extend the approach of to larger sets of variables, so as to fully characterize a higher-dimensional -space and its relationship to the PID. Much of the complexity of each framework is contained in considering these higher-order relationships. Future work will also consider additional solutions to the PID problem beyond the solutions of [9,19] considered here. All equations in Section 3 are general and agnostic to the precise solution used for the actual PID computation, and it should be straightforward to generate figures similar to Figure 6 for different solutions to show how they differ in mapping information components onto the function plane. This will provide interpretable geometric comparisons between solutions and also immediately highlight all functions for which results offer differing interpretations, as seen in Figure 6 and Figure 7. We anticipate that this direct comparison of how different solutions map the information content of discrete functions will provide a powerful visual tool for understanding the differing consequences of putative solutions, and thus our unification of these frameworks will be useful in resolving the open question of how best to compute the PID.
Acknowledgments
We wish to acknowledge support from the Pacific Northwest Research Institute.
Abbreviations
The following abbreviations are used in this manuscript:
PID | Partial Information Decomposition |
II | Interaction Information |
CI | Co-Information |
Unique Information in X | |
Unique Information in Y | |
R | Redundant Information |
S | Synergistic Information |
PPID | Pointwise Partial Information Decomposition |
Author Contributions
J.K.-G., N.S., and D.G. conceived of and designed the project; J.K.-G. performed the computations and formal analysis; N.S. and D.G. supervised and validated the results; J.K.-G. visualized the results; J.K.-G., N.S. and D.G. wrote the paper. All authors have read and agreed to the published version of the manuscript.
Funding
Research reported in this publication was supported by the National Heart, Lung, And Blood Institute of the National Institutes of Health under Award Number U01HL126496. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
Conflicts of Interest
The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Williams P.L., Beer R.D. Nonnegative decomposition of multivariate information. arXiv. 20101004.2515 [Google Scholar]
- 2.James R.G., Crutchfield J.P. Multivariate dependence beyond Shannon information. Entropy. 2017;19:531. doi: 10.3390/e19100531. [DOI] [Google Scholar]
- 3.Lizier J.T., Bertschinger N., Jost J., Wibral M. Information Decomposition of Target Effects from Multi-Source Interactions: Perspectives on Previous, Current and Future Work. Entropy. 2018;20:307. doi: 10.3390/e20040307. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Lizier J.T., Bertschinger N., Jost J., Wibral M. Bivariate measure of redundant information. Phys. Rev. E. 2013;87:012130. doi: 10.1103/PhysRevE.87.012130. [DOI] [PubMed] [Google Scholar]
- 5.Griffith V., Chong E.K., James R.G., Ellison C.J., Crutchfield J.P. Intersection information based on common randomness. Entropy. 2014;16:1985–2000. doi: 10.3390/e16041985. [DOI] [Google Scholar]
- 6.Barrett A.B. Exploration of synergistic and redundant information sharing in static and dynamical Gaussian systems. Phys. Rev. E. 2015;91:052802. doi: 10.1103/PhysRevE.91.052802. [DOI] [PubMed] [Google Scholar]
- 7.Ince R.A. Measuring multivariate redundant information with pointwise common change in surprisal. Entropy. 2017;19:318. doi: 10.3390/e19070318. [DOI] [Google Scholar]
- 8.Rauh J., Banerjee P.K., Olbrich E., Jost J., Bertschinger N. On extractable shared information. Entropy. 2017;19:328. doi: 10.3390/e19070328. [DOI] [Google Scholar]
- 9.Bertschinger N., Rauh J., Olbrich E., Jost J., Ay N. Quantifying unique information. Entropy. 2014;16:2161–2183. doi: 10.3390/e16042161. [DOI] [Google Scholar]
- 10.Timme N., Alford W., Flecker B., Beggs J.M. Synergy, redundancy, and multivariate information measures: An experimentalist’s perspective. J. Comput. Neurosci. 2014;36:119–140. doi: 10.1007/s10827-013-0458-4. [DOI] [PubMed] [Google Scholar]
- 11.Stramaglia S., Cortes J.M., Marinazzo D. Synergy and redundancy in the Granger causal analysis of dynamical networks. New J. Phys. 2014;16:105003. doi: 10.1088/1367-2630/16/10/105003. [DOI] [Google Scholar]
- 12.Timme N.M., Ito S., Myroshnychenko M., Nigam S., Shimono M., Yeh F.C., Hottowy P., Litke A.M., Beggs J.M. High-degree neurons feed cortical computations. PLoS Comput. Biol. 2016;12:e1004858. doi: 10.1371/journal.pcbi.1004858. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Wibral M., Priesemann V., Kay J.W., Lizier J.T., Phillips W.A. Partial information decomposition as a unified approach to the specification of neural goal functions. Brain Cogn. 2017;112:25–38. doi: 10.1016/j.bandc.2015.09.004. [DOI] [PubMed] [Google Scholar]
- 14.Wibral M., Finn C., Wollstadt P., Lizier J.T., Priesemann V. Quantifying information modification in developing neural networks via partial information decomposition. Entropy. 2017;19:494. doi: 10.3390/e19090494. [DOI] [Google Scholar]
- 15.Kay J.W., Ince R.A., Dering B., Phillips W.A. Partial and entropic information decompositions of a neuronal modulatory interaction. Entropy. 2017;19:560. doi: 10.3390/e19110560. [DOI] [Google Scholar]
- 16.Galas D.J., Sakhanenko N.A., Skupin A., Ignac T. Describing the complexity of systems: Multivariable “set complexity” and the information basis of systems biology. J. Comput. Biol. 2014;21:118–140. doi: 10.1089/cmb.2013.0039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Sakhanenko N.A., Galas D.J. Biological data analysis as an information theory problem: Multivariable dependence measures and the Shadows algorithm. J. Comput. Biol. 2015;22:1005–1024. doi: 10.1089/cmb.2015.0051. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Sakhanenko N., Kunert-Graf J., Galas D. The Information Content of Discrete Functions and Their Application in Genetic Data Analysis. J. Comp. Biol. 2017;24:1153–1178. doi: 10.1089/cmb.2017.0143. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Finn C., Lizier J.T. Pointwise partial information decomposition using the specificity and ambiguity lattices. Entropy. 2018;20:297. doi: 10.3390/e20040297. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Kunert-Graf J. kunert/deltaPID: Initial Release (Version v1.0.0) Zenodo. 2020 doi: 10.5281/zenodo.4287515. [DOI] [Google Scholar]
- 21.McGill W. Multivariate information transmission. Trans. IRE Prof. Group Inf. Theory. 1954;4:93–111. doi: 10.1109/TIT.1954.1057469. [DOI] [Google Scholar]
- 22.Jakulin A., Bratko I. Quantifying and Visualizing Attribute Interactions: An Approach Based on Entropy. arXiv. 2003cs/0308002 [Google Scholar]
- 23.Bell A.J. The co-information lattice; Proceedings of the Fifth International Workshop on Independent Component Analysis and Blind Signal Separation, ICA, Citeseer; Granada, Spain. 22–24 September 2003; [Google Scholar]
- 24.Watanabe S. Information theoretical analysis of multivariate correlation. IBM J. Res. Dev. 1960;4:66–82. doi: 10.1147/rd.41.0066. [DOI] [Google Scholar]