Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2020 Dec 21;16(12):e1008289. doi: 10.1371/journal.pcbi.1008289

Reconciling emergences: An information-theoretic approach to identify causal emergence in multivariate data

Fernando E Rosas 1,2,3,*,#, Pedro A M Mediano 4,*,#, Henrik J Jensen 3,5,6, Anil K Seth 7,8, Adam B Barrett 7,9, Robin L Carhart-Harris 1, Daniel Bor 4
Editor: Daniele Marinazzo10
PMCID: PMC7833221  PMID: 33347467

Abstract

The broad concept of emergence is instrumental in various of the most challenging open scientific questions—yet, few quantitative theories of what constitutes emergent phenomena have been proposed. This article introduces a formal theory of causal emergence in multivariate systems, which studies the relationship between the dynamics of parts of a system and macroscopic features of interest. Our theory provides a quantitative definition of downward causation, and introduces a complementary modality of emergent behaviour—which we refer to as causal decoupling. Moreover, the theory allows practical criteria that can be efficiently calculated in large systems, making our framework applicable in a range of scenarios of practical interest. We illustrate our findings in a number of case studies, including Conway’s Game of Life, Reynolds’ flocking model, and neural activity as measured by electrocorticography.

Author summary

Many scientific domains exhibit phenomena that seem to be “more than the sum of their parts”; for example, flocks seem to be more than a mere collection of birds, and consciousness seems more than electric impulses between neurons. But what does it mean for a physical system to exhibit emergence? The literature on this topic contains various conflicting approaches, many of which are unable to provide quantitative, falsifiable statements. Having a rigorous, quantitative theory of emergence could allow us to discover the exact conditions that allow a flock to be more than individual birds, and to better understand how the mind emerges from the brain. Here we provide exactly that: a formal theory of what constitutes causal emergence, how to measure it, and what different “types” of emergence exist. To do this, we leverage recent developments in information dynamics—the study of how information flows through and is modified by dynamical systems. As part of this framework, we provide a mathematical definition of causal emergence, and also practical formulae for analysing empirical data. Using these, we are able to confirm emergence in the iconic Conway’s Game of Life, in certain flocking patterns, and in representations of motor movements in the monkey’s brain.

Introduction

While most of our representations of the physical world are hierarchical, there is still no agreement on how the co-existing “layers” of this hierarchy interact. On the one hand, reductionism claims that all levels can always be explained based on sufficient knowledge of the lowest scale and, consequently—taking an intentionally extreme example—that a sufficiently accurate theory of elementary particles should be able to predict the existence of social phenomena like communism. On the other hand, emergentism argues that there can be autonomy between layers, i.e. that some phenomena at macroscopic layers might only be accountable in terms of other macroscopic phenomena. While emergentism might seem to better serve our intuition, it is not entirely clear how a rigorous theory of emergence could be formulated within our modern scientific worldview, which tends to be dominated by reductionist principles.

Emergent phenomena are usually characterised as either strong or weak [1]. Strong emergence corresponds to the somewhat paradoxical case of supervenient properties with irreducible causal power [2]; i.e. properties that are fully determined by microscopic levels but can nevertheless exert causal influences that are not entirely accountable from microscopic considerations (the case of strong emergence most commonly argued in the literature is the one of conscious experiences with respect to their corresponding physical substrate [3, 4]). Strong emergence has been as much a cause of wonder as a perennial source of philosophical headaches, being described as “uncomfortably like magic” while accused of being logically inconsistent [2] and sustained on illegitimate metaphysics [5]. Weak emergence has been proposed as a more docile alternative to strong emergence, where macroscopic features have irreducible causal power in practice but not in principle. A popular formulation of weak emergence is due to Bedau [5], and corresponds to properties generated by elements at microscopic levels in such complicated ways that they cannot be derived via explanatory shortcuts, but only by exhaustive simulation. While this formulation is usually accepted by the scientific community, it is not well-suited to address mereological questions about emergence in scenarios where parts-whole relationships are the primary interest.

Part of the difficulty in building a deeper understanding of strong emergence is the absence of simple but clear analytical models that can serve the community to guide discussions and mature theories. Efforts have been made to introduce quantitative metrics of weak emergence [6], which enable fine-grained data-driven alternatives to traditional all-or-none classifications. In this vein, an attractive alternative comes from the work on causal emergence introduced in Ref. [7] and later developed in Refs. [8, 9], which showed that macroscopic observables can sometimes exhibit more causal power (as understood within the framework of Pearl’s do-calculus [10]) than microscopic variables. However, this framework relies on strong assumptions that are rarely satisfied in practice, which severely hinders its applicability (this point is further elaborated in Section Relationship with other quantitative theories of emergence).

Inspired by Refs. [6, 7], here we introduce a practically useful and philosophically innocent framework to study causal emergence in multivariate data. Building on previous work [11], we take the perspective of an experimentalist who has no prior knowledge of the underlying phenomenon of interest, but has sufficient data of all relevant variables that allows an accurate statistical description of the phenomenon. In this context, we put forward a formal definition of causal emergence that doesn’t rely on coarse-graining functions as Ref. [7], but addresses the “paradoxical” properties of strong emergence based on the laws of information flow in multivariate systems.

The main contribution of this work is to enable a rigorous, quantitative definition of downward causation, and introduce a novel notion of causal decoupling as a complementary modality of causal emergence. Another contribution is to extend the domain of applicability of causal emergence analyses to include cases of observational data, in which case causality ought to be understood in the Granger sense, i.e. as predictive ability [12]. Furthermore, our framework yields practical criteria that can be effectively applied to large systems, bypassing prohibitive estimation issues that severely restrict previous approaches.

The rest of this paper is structured as follows. First, Section Fundamental intuitions discusses minimal examples of emergence. Then, Section A formal theory of causal emergence presents the core of our theory, and Section Measuring emergence discusses practical methods to quantify emergence from experimental data. Our framework is then illustrated on a number of case studies, presented in Section Case studies. Finally, the Section Discussion concludes the paper with a discussion of some of the implications of our findings.

Fundamental intuitions

To ground our intuitions, let us introduce minimal examples that embody a few key notions of causally emergent behaviour. Throughout this section, we consider systems composed of n parts described by a binary vector Xt=(Xt1,,Xtn){0,1}n, which undergo Markovian stochastic dynamics following a transition probability pXt+1|Xt. For simplicity, we assume that at time t the system is found in an entirely random configuration (i.e. pXt(xt)=2-n). From there, we consider three evolution rules.

Example 1. Consider a temporal evolution where the parity of Xt is preserved with probability γ ∈ (0, 1). Mathematically,

pXt+1|Xt(xt+1|xt)={γ2n-1ifj=1nxt+1j=j=1nxtj,1-γ2n-1otherwise,

for all tN, where j=1naj1 if j=1naj is even and zero otherwise. Put simply: xt+1 is a random sample from the set of all strings with the same parity as xt with probability γ; and is a sample from the strings with opposite parity with probability 1 − γ.

This evolution rule has a number of interesting properties. First, the system has a non-trivial causal structure, since some properties of the future state (its parity) can be predicted from the past state. However, this structure is noticeable only at the collective level, as no individual variable has any predictive power over the evolution of itself or any other variable (see Fig 1). Furthermore, even the complete past of the system Xt has no predictive power over any individual future Xt+1j. This case shows an extreme kind of causal emergence that we call “causal decoupling,” in which the parity predicts its own evolution but no element (or subset of elements) predicts the evolution of any other element.

Fig 1. Minimal examples of causally emergent dynamics.

Fig 1

In Example 1 (left) the system’s parity tends to be preserved while no interactions occur between low-level elements, which is an example of causal decoupling. In Example 2 (right) the system’s parity determines one element only, corresponding to downward causation.

Example 2. Consider now a system where the parity of Xt determines Xt+11 (i.e. Xt+11=i=1nXti), and Xt+1j for j ≠ 1 is a fair coin flip independent of Xt (see Fig 1). In this scenario Xt predicts Xt+11 with perfect accuracy, while it can be verified that XtiXt+11 for all i ∈ {1, …, n}. Therefore, under this evolution rule the whole system has a causal effect over a particular element, although this effect cannot be attributed to any individual part (for a related discussion, see Ref. [13]), being a minimal example of downward causation.

Example 3. Let us now study an evolution rule that includes the mechanisms of both Examples 1 and 2. Concretely, consider

pXt+1|Xt(xt+1|xt)={0ifxt+11j=1nxtj,γ2n-2ifxt+11=j=1nxtjandj=1nxt+1j=j=1nxtj,1-γ2n-2otherwise.

As in Example 1, the parity of Xt is transfered to Xt+1 with probability γ; additionally, it is guaranteed that Xt+11=i=1nXti. Hence, in this case not only is there a macroscopic effect that cannot be explained from the parts, but at the same time there is another effect going from the whole to one of the parts. Importantly, both effects co-exist independently of each other.

The above are minimal examples of dynamical laws that cannot be traced from the interactions between their elementary components: Example 1 shows how a collective property can propagate without interacting with its underlying substrate; Example 2 how a collective property can influence the evolution of specific parts; and Example 3 how these two kinds of phenomena take place in the same system. All these issues are formalised by the theory developed in the next section.

A formal theory of causal emergence

This section presents the main body of our theory of causal emergence. To fix ideas, we consider a scientist measuring a system composed of n parts. The scientist is assumed to measure the system regularly over time, and the results of those measurements are denoted by Xt=(Xt1,,Xtn), with XtiXi corresponding to the state of the ith part at time tN with phase space Xi. When referring to a collection of parts, we use the notation Xtα=(Xti1,,XtiK) for α = {i1, …, iK}⊂{1, …, n}. We also use the shorthand notation [n] ≔ {1, …, n}.

Supervenience

Our analysis considers two time points of the evolution of the system, denoted as t and t′, with t < t′. The corresponding dynamics are encoded in the transition probability pXt|Xt(xt|xt). We consider features VtV generated via a conditional probability pVt|Xt that are supervenient on the underlying system; i.e. that does not provide any predictive power for future states at times t′ > t if the complete state of the system at time t is known with perfect precision. We formalise this in the following definition.

Definition 1. A stochastic process Vt is said to be supervenient over Xt if VtXtXt′′ form a Markov chain for all t′′ ≠ t.

The above condition is equivalent to require Vt to be statistically independent of Xt′′ when Xt is given. The relationship between supervenient features and the underlying system is illustrated in Fig 2.

Fig 2. Diagram of causally emergent relationships.

Fig 2

Causally emergent features have predictive power beyond individual components. Downward causation takes place when that predictive power refers to individual elements; causal decoupling when it refers to itself or other high-order features.

This formalisation of supervenience characterises features Vt that are fully determined by the state of the system at a given time t, but also allows the feature to be noisy—which is not critical for our results, but is useful for extending their domain of applicability to practical scenarios. In effect, Definition 1 includes as particular cases deterministic functions F:j=1nXjV such that Vt = F(Xt), as well as features calculated under observational noise—e.g. Vt = F(Xt) + νt, where νt is independent of Xt for all t. In contrast, features that are computed using the values of Xt at multiple timepoints (e.g. the Fourier transform of Xt) generally fail to be supervenient.

Partial information decomposition

Our theory is based on the Partial Information Decomposition (PID) framework [14], which provides powerful tools to reason about information in multivariate systems. In a nutshell, PID decomposes the information that n sources X = (X1, …, Xn) provide about a target variable Y in terms of information atoms as follows:

I(X;Y)=αAIα(X;Y), (1)

with A={{α1,,αL}:αi[n],αi¬αji,j} being the set of antichain collections [14]. Intuitively, Iα for α = {α1, …, αL} represents the information that the collection of variables Xα1,,XαL provide redundantly, but their sub-collections don’t. For example, for n = 2 source variables, α = {{1}{2}} corresponds to the information about Y that is provided by both of them, α = {{i}} to the information provided uniquely by Xi, and, most interestingly, α = {{12}} corresponds to the information provided by both sources jointly but not separately—commonly referred to as informational synergy.

One of the drawbacks of PID is that the number of atoms (i.e. the cardinality of A) grows super-exponentially with the number of sources, and hence it is useful to coarse-grain the decomposition according to specific criteria. Here we introduce the notion of kth-order synergy between n variables, which is calculated as

Syn(k)(X;Y)αS(k)Iα(X;Y),

with S(k)={{α1,,αL}A:minj|αj|>k}. Intuitively,

Syn

(k)(X;Y) corresponds to the information about the target that is provided by the whole X but is not contained in any set of k or less parts when considered separately from the rest. Accordingly, S(k) only contains collections with groups of more than k sources.

Similarly, we introduce the unique information of Xβ with β ⊂ [n] with respect to sets of at most k other variables, which is calculated as

Un(k)(Xβ;Y|X-β)αU(k)(β)Iα(X;Y).

Above, U(k)(β)={αA:βα,αβα,α[n]\β,|α|>k}, and Xβ being all the variables in X whose indices are not in β. Put simply, Un(k)(Xtβ;Y|X-β) represents the information carried by Xβ about Y that no group of k or less variables within Xβ has on its own. Note that these coarse-grained terms can be used to build a general decomposition of I(X, Y) described in S1 Appendix (Section 1), the properties of which are proven in S1 Appendix (Section 2).

One peculiarity of PID is that it postulates the structure of information atoms and the relations between them, but it does not prescribe a particular functional form to compute Iα. In fact, only one of the information atoms must be specified to determine the whole PID—usually the redundancy between all individual elements [14]. There have been multiple proposals for specific functional forms of Iα in the PID literature; see e.g. Refs. [1518]. A particular method for fully computing the information atoms based on a recent PID [19] is discussed in Section Measuring emergence via synergistic channels.

Conveniently, our theory doesn’t rely on a specific functional form of PID, but only on a few basic properties that are precisely formulated in S1 Appendix (Section 2). Therefore, the theory can be instantiated using any PID—as long as those properties are satisfied. Importantly, as shown in Section Practical criteria for large systems, the theory allows the derivation of practical metrics that are valid independently of the PID chosen.

Defining causal emergence

With the tools of PID at hand, now we introduce our formal definition of causal emergence.

Definition 2. For a system described by Xt, a supervenient feature Vt is said to exhibit causal emergence of order k if

Un(k)(Vt;Xt|Xt)>0. (2)

Accordingly, causal emergence takes place when a supervenient feature Vt has irreducible causal power, i.e. when it exerts causal influence that is not mediated by any of the parts of the system. In other words, Vt represents some emergent collective property of the system if: 1) contains information that is dynamically relevant (in the sense that it predicts the future evolution of the system); and 2) this information is beyond what is given by the groups of k parts in the system when considered separately.

To better understand the implications of this definition, let us study some of its basic properties.

Lemma 1. Consider a feature Vt that exhibits causal emergence of order 1 over Xt. Then,

  1. The dimensionality of the system satisfies n ≥ 2.

  2. There exists no deterministic function g(⋅) such that Vt=g(Xtj) for any j = 1, …, n.

Proof. See S1 Appendix, Section 3.

These two properties establish causal emergence as a fundamentally collective phenomenon. In effect, property (i) states that causal emergence is a property of multivariate systems, and property (ii) that Vt cannot have emergent behaviour if it can be perfectly predicted from a single variable.

In order to use Definition 2, one needs a candidate feature Vt to be tested. However, in some cases there are no obvious candidates for an emergent feature, for which Definition 2 might seem problematic. Our next result provides a criterion for the existence of emergent features based solely on the system’s dynamics.

Theorem 1. A system Xt has a causally emergent feature of order k if and only if

Syn(k)(Xt;Xt)>0. (3)

Proof. See S1 Appendix, Section 2.

Corollary 1. The following bound holds for any supervenient feature Vt: Un(k)(Vt;Xt|Xt)Syn(k)(Xt;Xt).

This result shows that the capability of exhibiting emergence is closely related to how synergistic the system components are with respect to their future evolution. Importantly, this result enables us to determine whether or not the system admits any emergent features by just inspecting the synergy between its parts—without knowing what those features might be. Conversely, this result also allows us to discard the existence of causal emergence by checking a single condition: the lack of dynamical synergy. Furthermore, Corollary 1 implies that the quantity Syn(k)(Xt;Xt) serves as a measure of the emergence capacity of the system, as it upper-bounds the unique information of all possible supervenient features.

Theorem 1 establishes a direct link between causal emergence and the system’s statistics, avoiding the need for the observer to propose a particular feature of interest. It is important to remark that the emergence capacity of a system depends on the system’s partition into microscopic elements—in fact, it is plausible that a system might have emergence capacity under one microscopic representation, but not with respect to another after a change of variables. Therefore, emergence in the context of our theory always refers to “emergence with respect to a given microscopic partition.”

A taxonomy of emergence

Our theory, so far, is able to detect whether there is emergence taking place; the next step is to be able to characterise which kind of emergence it is. For this purpose, we combine our feature-agnostic criterion of emergence presented in Theorem 1 with Integrated Information Decomposition, ΦID, a recent extension of PID to multi-target settings [20].

Using ΦID, one can decompose a PID atom as

Iα(Xt;Xt)=βAIαβ(Xt;Xt). (4)

For example, if n = 2 then I{1}{2}{1}{2} represents the information shared by both time series at both timesteps (for example, when Xt1,Xt2,Xt1,Xt2 are all copies of each other); and I{12}{1} corresponds to the synergistic causes in Xt that have a unique effect on Xt1 (for example, when Xt1=Xt1Xt2). More details and intuitions on ΦID can be found in Ref. [20].

With the fine-grained decomposition provided by ΦID one can discriminate between different kinds of synergies. In particular, we introduce the downward causation and causal decoupling indices of order k, denoted by D(k) and G(k) respectively, as

D(k)(Xt;Xt)αS(k)βA\S(k)Iαβ(Xt;Xt), (5)
G(k)(Xt;Xt)α,βS(k)Iαβ(Xt;Xt). (6)

From these definitions and Eq (4), one can verify that

Syn(k)(Xt;Xt)=G(k)(Xt;Xt)+D(k)(Xt;Xt). (7)

Therefore, the emergence capacity of a system naturally decomposes in two different components: information about k-plets of future variables, and information about future collective properties beyond k-plets. The ΦID atoms that belong to these two terms are illustrated within the ΦID lattice for two time series in Fig 3. The rest of this section shows that D(k) and G(k) are natural metrics of downward causation and causal decoupling, respectively.

Fig 3. Integrated information decomposition (ΦID).

Fig 3

ΦID lattice for n = 2 time series [20], with downward (D) causation and causal decoupling (G) terms highlighted.

Downward causation

Intuitively, downward causation occurs when collective properties have irreducible causal power over individual parts. More formally:

Definition 3. A supervenient feature Vt exhibits downward causation of order k if, for some α with |α| = k:

Un(k)(Vt;Xtα|Xt)>0. (8)

Note that, in contrast with Definition 2, downward causation requires the feature Vt to have unique predictive power over the evolution of specific subsets of the whole system. In particular, an emergent feature Vt that has predictive power over e.g. Xtj is said to exert downward causation, as it predicts something about Xtj that could not be predicted from any particular Xti for i ∈ [n]. Put differently, in a system with downward causation the whole has an effect on the parts that cannot be reduced to low-level interactions. A minimal case of this is provided by Example 2 in Section Fundamental intuitions.

Our next result formally relates downward causation with the index D(k) introduced in Eq (5).

Theorem 2. A system Xt admits features that exert downward causation of order k iff D(k)(Xt;Xt)>0.

Proof. See S1 Appendix, Section 3.

Causal decoupling

In addition to downward causation, causal decoupling takes place when collective properties have irreducible causal power over other collective properties. In technical terms:

Definition 4. A supervenient feature Vt is said to exhibit causal decoupling of order k if

Un(k)(Vt;Vt|Xt,Xt)>0. (9)

Furthermore, Vt is said to have pure causal decoupling if Un(k)(Vt;Xt|Xt)>0 and Un(k)(Vt;Xtα|Xt)=0 for all α ⊂ [n] with |α| = k. Finally, a system is said to be perfectly decoupled if all the emergent features exhibit pure causal decoupling.

Above, the term Un(k)(Vt;Vt|Xt,Xt) refers to information that Vt and Vt share that cannot be found in any microscopic element, either at time t or t′ (note that Un(k)(Vt;Vt|Xt,Xt) is information shared between Vt and Vt that no combination of k or less variables from Xt or Xt has in its own).

Features that exhibit causal decoupling could still exert influence over the evolution of individual elements, while features that exhibit pure decoupling cannot. In effect, the condition Un(Vt;Xtj|Xt)=0 implies that the high-order causal effect does not affect any particular part – only the system as a whole. Interestingly, a feature that exhibits pure causal decoupling can be thought of as having “a life of its own;” a sort of statistical ghost, that perpetuates itself over time without any individual part of the system influencing or being influenced by it. The system’s parity, in the first example of Section Fundamental intuitions, constitutes a simple example of perfect causal decoupling. Importantly, the case studies presented in Section Case studies show that causal decoupling can take place not only in toy models but also in diverse scenarios of practical relevance.

We close this section by formally establishing the connection between causal decoupling and the index G(k) introduced in Eq (6).

Theorem 3. A system possesses features that exhibit causal decoupling if and only if G(k)(Xt;Xt)>0. Additionally, the system is perfectly decoupled if G(k)(Xt;Xt)>0 and D(k)(Xt;Xt)=0.

Proof. See S1 Appendix, Section 3.

Measuring emergence

This section explores methods to operationalise the framework presented in the previous section. We discuss two approaches: first, Section Practical criteria for large systems introduces sufficiency criteria that are practical for use in large systems; then, Section Measuring emergence via synergistic channels illustrates how further considerations can be made if one adopts a specific method of computing ΦID atoms. The latter approach provides accurate discrimination at the cost of being data-intensive and hence only applicable to small systems; the former can be computed in large systems and its results hold independently of the chosen PID, but is vulnerable to misdetections (i.e. false negatives).

Practical criteria for large systems

While theoretically appealing, our proposed framework suffers from the challenge of estimating joint probability distributions over many random variables, and the computation of the ΦID atoms themselves. As an alternative, we consider approximation techniques that do not require the adoption of any particular PID or ΦID function and are data-efficient, since they are based on pairwise distributions only.

As practical criteria to measure causal emergence of order k, we introduce the quantities Ψt,t(k), Δt,t(k), and Γt,t(k). For simplicity, we write here the special case k = 1, and provide full formulae for arbitrary k and accompanying proofs in S1 Appendix, Section 4:

Ψt,t(1)(V)I(Vt;Vt)-jI(Xtj;Vt), (10a)
Δt,t(1)(V)maxj(I(Vt;Xtj)-iI(Xti;Xtj)), (10b)
Γt,t(1)(V)maxjI(Vt;Xtj). (10c)

Our next result links these quantities with the formal definitions in Section A formal theory of causal emergence, showing their value as practical criteria to detect causal emergence.

Proposition 1. Ψt,t(k)(V)>0 is a sufficient condition for Vt to be causally emergent. Similarly, Δt,t(k)(V)>0 is a sufficient condition for Vt to exhibit downward causation. Finally, Ψt,t(k)(V)>0 and Γt,t(k)(V)=0 is sufficient for causal decoupling.

Proof. See S1 Appendix, Section 4.

Although calculating whether a system has emergent features via Proposition 1 may be computationally challenging, if one has a candidate feature V one believes may be emergent, one can compute the simple quantities in Eq (10) which depend only on standard mutual information and bivariate marginals, and scales linearly with system size (for k = 1). These quantities are easy to compute and test for significance using standard information-theoretic tools [21, 22]. Moreover, the outcome of these measures is valid for any choice of PID and ΦID that is compatible with the properties specified in S1 Appendix, Section 2.

In a broader context, Ψt,t(k) and Δt,t(k) belong to the same whole-minus-sum family of measures as the interaction information [14, 23], the redundancy-synergy index [24] and, more recently, the O-information Ω [25]—which cannot measure synergy by itself, but only the balance between synergy and redundancy. In practice, this means that if there is redundancy in the system it will be harder to detect emergence, since redundancy will drive Ψt,t(k) and Δt,t(k) more negative. Furthermore, by summing all marginal mutual informations (e.g. I(Xtj;Vt) in the case of Ψt,t(1)), these measures effectively double-count redundancy up to n times, further penalising the criteria. This problem of double-counting can be avoided if one is willing to commit to a particular PID or ΦID function, as we show next.

It is worth noticing that the value of k can be tuned to explore emergence with respect to different “scales.” For example, k = 1 corresponds to emergence with respect to individual microscopic elements, while k = 2 refers to emergence with respect to all couples—i.e. individual elements and their pairwise interactions. Accordingly, the criteria in Proposition 1 are, in general, harder to satisfy for larger values of k. In addition, from a practical perspective, considering large values of k requires estimating information-theoretic quantities in high-dimensional distributions, which usually requires exponentially larger amounts of data.

Measuring emergence via synergistic channels

This section leverages recent work on information decomposition reported in Ref. [19], and presents a way of directly measuring the emergence capacity and the indices of downward causation and causal decoupling. The key takeaway of this section is that if one adopts a particular ΦID, then it is possible to evaluate D(k) and G(k) directly, providing a direct route to detect emergence without double-counting redundancy, as the methods introduced in Section Practical criteria for large systems do. Moreover, additional properties may become available due to the characteristics of the particular ΦID chosen.

Let us first introduce the notion of k-synergistic channels: mappings pV|X that convey information about X but not about any of the parts Xα for all |α| = k. The set of all k-synergistic channels is denoted by

Ck(X)={pV|X|VXα,α[n],|α|=k}. (11)

A variable V generated via a k-synergistic channel is said to be a k-synergistic observable.

With this definition, we can consider the kth-order synergy to be the maximum information extractable from a k-synergistic channel:

Syn(k)(Xt;Xt)suppV|XtCk(Xt):V-Xt-XtI(V;Xt). (12)

This idea can be naturally extended to the case of causal decoupling by requiring synergistic channels at both sides, i.e.

G(k)(Xt;Xt)suppV|XtCk(Xt),pU|XtCk(Xt):V-Xt-Xt-UI(V;U). (13)

Finally, the downward causation index can be computed from the difference

D(k)(Xt;Xt)Syn(k)(Xt;Xt)-G(k)(Xt;Xt). (14)

Note that Syn(k)G(k), which is a direct consequence of the data processing inequality applied on VXtXtU, and therefore D(k),G(k)0.

By exploiting the properties of this specific way of measuring synergy, one can prove the following result. For this, let us say that a feature Vt is auto-correlated if I(Vt;Vt)>0.

Proposition 2. If Xt is stationary, all auto-correlated k-synergistic observables are kth-order emergent.

Proof. See S1 Appendix, Section 4.

In summary, D(k) and G(k) provide data-driven tools to test—and possibly reject—hypotheses about emergence in scenarios of interest. Efficient algorithms to compute these quantities are discussed in Ref. [26]. Although current implementations allow only relatively small systems, this line of thinking shows that future advances in PID might make the computation of emergence indices more scalable, avoiding the limitations of Eq (10).

Case studies

Let us summarise our results so far. We began by formulating a rigorous definition of emergent features based on PID (Section Defining causal emergence), and then used ΦID to break down the emergence capacity into the causal decoupling and downward causation indices (Section A taxonomy of emergence). Although these are not straightforward to compute, the ΦID framework allows us to formulate readily computable sufficiency conditions (Section Practical criteria for large systems). This section illustrates the usage of those conditions in various case studies. Code to compute all emergence criteria in Eq (10) is provided in an online open-source repository (https://github.com/pmediano/ReconcilingEmergences).

Canonical examples of putative emergence

Here we present an evaluation of our practical criteria for emergence (Proposition 1) in two well-known systems: Conway’s Game of Life (GoL) [27], and Reynolds’ flocking boids model [28]. Both are widely regarded as paradigmatic examples of emergent behaviour, and have been thoughtfully studied in the complexity and artificial life literature [29]. Accordingly, we use these models as test cases for our methods. Technical details of the simulations are provided in S1 Appendix, Section 5.

Conway’s Game of Life

A well-known feature of GoL is the presence of particles: coherent, self-sustaining structures known to be responsible for information transfer and modification [30]. These particles have been the object of extensive study, and detailed taxonomies and classifications exist [29, 31].

To test the emergent properties of particles, we simulate the evolution of 15x15 square cell arrays, which we regard as a binary vector Xt ∈ {0, 1}n with n = 225. As initial condition, we consider configurations that correspond to a “particle collider” setting, with two particles of known type facing each other (Fig 4). In each trial, the system is randomised by changing the position, type, and relative displacement of the particles. After an intial configuration has been selected, the well-known GoL evolution rule [27] is applied 1000 times, leading to a final state Xt. Simulations showed that this interval is enough for the system to settle in a stable state after the collision.

Fig 4. Causal emergence in Conway’s Game of Life.

Fig 4

The system is initialised in a “particle collider” setting, and run until a stable configuration is reached after the collision. Using particle type as a supervenient feature V, we find the system meets our practical criterion for causal emergence.

To use the criteria from Eq 10, we need to choose a candidate emergent feature Vt. In this case, we consider a symbolic, discrete-valued vector that encodes the type of particle(s) present in the board. Specifically, we consider Vt=(Vt1,,VtL), where Vtj=1 iff there is a particle of type j at time t—regardless of its position or orientation.

With these variables, we compute the quantities in Eq 10 using Bayesian estimators of mutual information [32]. The result is that, as expected, the criterion for causal emergence is met with Ψt,t(1)(V)=0.58±0.02. Furthermore, we found that Γt,t(1)(V)=0.009±0.0002, which is orders of magnitude smaller than I(Vt;Vt) = 0.99±0.02. Errors represent the standard deviation over surrogate data, as described in S1 Appendix, Section 5. Using Proposition 1, these two results suggest that particle dynamics in GoL may not only be emergent, but causally decoupled with respect to their substrate.

Reynolds’ flocking model

As a second test case, we consider Reynolds’ model of flocking behaviour. This model is composed by boids (bird-oid objects), with each boid represented by three numbers: its position in 2D space and its heading angle. As candidate feature for emergence, we use the 2D coordinates of the center of mass of the flock, following Seth [6].

In this model boids interact with one another following three rules, each regulated by a scalar parameter [6]:

  • aggregation (a1), as they fly towards the center of the flock;

  • avoidance (a2), as they fly away from their closest neighbour; and

  • alignment (a3), as they align their flight direction to that of their neighbours.

Following Ref. [6], we study small flocks of N = 10 boids with different parameter settings to showcase some properties of our practical criterion of emergence. Note that this study is meant as an illustration of the proposed theory, and not as a thorough exploration of the flocking model, for which a vast literature exists (see e.g. the work of Vicsek [33] and references therein).

Fig 5 shows the results of a parameter sweep over the avoidance parameter, a2, while keeping a1 and a3 fixed. When there is no avoidance, boids orbit around a slowly-moving center of mass, in what could be called an ordered regime. Conversely, for high values of a2 neighbour repulsion is too strong for lasting flocks to form, and isolated boids spread across the space avoiding one another. For intermediate values, the center of mass traces a smooth trajectory, as flocks form and disintegrate. In line with the findings of Seth [6], our criterion indicates that the flock exhibits causally emergent behaviour in this intermediate range.

Fig 5. Causal emergence in the flocking boids model.

Fig 5

As the avoidance parameter is increased, the flock transitions from an attractive regime (in which all boids orbit regularly around a stable center of mass), to a repulsive one (in which boids spread across space and no flocking is visible). a) Our criterion Ψ detects causal emergence in an intermediate range of the avoidance parameter (error bars represent the standard deviation estimated over surrogate data). b) Sample trajectories of boids (grey) and their center of mass (red).

By studying separately the two terms that make up Ψ we found that the criterion of emergence fails for both low and high a2, but for different reasons (see Fig 5). In effect, for high a2 the self-predictability of the center of mass (i.e. I(Vt;Vt)) is low; while for low a2 it is high, yet lower than the mutual information from individual boids (i.e. iI(Xti;Vt)). These results suggest that the low-avoidance scenario is dominated not by a reduction in synergy, but by an increase in redundancy, which effectively increases the synergy threshold needed to detect emergence. However, note that, due to the limitations of the criterion, the fact that Ψt,t(1)<0 is inconclusive and does not rule out the possibility of emergence. This is a common limitation of whole-minus-sum estimators like Ψ; further refinements may provide bounds that are less susceptible to these issues and perform accurately in these scenarios.

Mind from matter: Emergence, behaviour, and neural dynamics

A tantalising outcome of having a formal theory of emergence is the capability of bringing a quantitative angle to the archetype of emergence: the mind-matter relationship [35, 36]. As a first step in this direction, we conclude this section with an application of our emergence criteria to neurophysiological data.

We study simultaneous electrocorticogram (ECoG) and motion capture (MoCap) data of Japanese macaques performing a reaching task [34], obtained from the online Neurotycho database. Note that the MoCap data cannot be assumed to be a supervenient feature of the available ECoG data, since it doesn’t satisfy the conditional independence conditions required by our definition of supervenience (see Section A formal theory of causal emergence). This is likely to be the case, because the neural system is only partially observed—i.e. the ECoG does not capture every source of relevant activity in the macaque’s cortex. Note that non-supervenient features are of limited interest within our framework, as they can satisfy Proposition 1 in trivial ways (e.g. time series which are independent of the underlying system satisty Ψ > 0 if they are auto-correlated). Instead, we focus on the portion of neural activity encoded in the ECoG signal that is relevant to predict the macaque’s behaviour, and conjecture this information to be an emergent feature of the underlying neural activity (Fig 6).

Fig 6. Causal emergence in motor behaviour of an awake macaque monkey.

Fig 6

a) Position of electrocorticogram (ECoG) electrodes used in the recording (in blue) overlaid on an image of the macaque’s left hemisphere (front of the brain towards the top of the page). b) Sample time series from the 64-channel ECoG recordings used, which correspond to the system of interest XtR64. c) 3D position of the macaques’s wrist, as measured by motion capture (blue) and as predicted by the regression model (orange), taken as a supervenient feature VtR3. d) Our emergence criterion yields Ψt,t(1)(V)>0, detecting causal emergence of the behaviour with respect to the ECoG sources. Original data and image from Ref. [34] and the Neurotycho database.

To test this hypothesis, we take the neural activity (as measured by 64 ECoG channels distributed across the left hemisphere) to be the system of interest, and consider a memoryless predictor of the 3D coordinates of the macaque’s right wrist based on the ECoG signal. Therefore, in this scenario XtR64 and Vt=F(Xt)R3. To build Vt, we used Partial Least Squares (PLS) and a Support Vector Machine (SVM) regressor, the details of which can be found in S1 Appendix, Section 6.

After training the decoder and evaluating on a held-out test set, results show that Ψ > 0, confirming our conjecture that the motor-related information is an emergent feature of the macaque’s cortical activity. For short timescales (t′ − t = 8 ms), we find Γt,t(1)(V)=0.049±0.002, which is orders of magnitude smaller than Ψt,t(1)(V)=1.275±0.002, suggesting that the behaviour may have an important component decoupled from individual ECoG channels (errors are standard deviations estimated over time-shuffled data). Furthermore, the emergence criterion is met for multiple timescales t′ − t of up to ≈0.2s, beyond which the predictive power in Vt and individual electrodes decrease and become nearly identical.

As a control, we performed a surrogate data test to confirm the results in Fig 6 were not driven by the autocorrelation in the ECoG time series. To this end, we re-run the analysis (including training and testing the PLS-SVM) using the same ECoG data, but time-shuffling the wrist position—resulting in a Vt that does not extract any meaningful information from the ECoG, but has the same properties induced by autocorrelation, filtering and regularisation. As expected, the resulting surrogate Ψt,t(1) is significantly lower than the one using the un-shuffled wrist position, confirming the measured Ψt,t(1) is positive and higher than what would be expected from a similar, random projection of the ECoG (details in S1 Appendix, Section 6).

This analysis, while just a proof of concept, helps us quantify how and to what extent behaviour emerges from collective neural activity; and opens the door to further tests and quantitative empirical explorations of the mind-matter relationship.

Discussion

A large fraction of the modern scientific literature considers strong emergence to be impossible or ill-defined. This judgement is not fully unfounded: a property that is simultaneously supervenient (i.e. that can be computed from the state of the system) and that has irreducible causal power (i.e. that “tells us something” that the parts don’t) can indeed seem to be an oxymoron [5]. Nonetheless, by linking supervenience to static relationships and causal power to dynamical properties, our framework shows that these two phenomena are perfectly compatible within the—admittedly counterintuitive – laws of multivariate information dynamics [20], providing a tentative solution to this paradox.

Our theory of causal emergence is about predictive power, not “explicability” [3], and therefore is not related to views on strong emergence such as Chalmers’ [3]. Nevertheless, our framework embraces aspects that are commonly associated with strong emergence—such as downward causation—and renders them quantifiable. Our framework also does not satisfy conventional definitions of weak emergence (systems studied in Section Fundamental intuitions are not weakly emergent in the sense of Bedau [5], being simple and susceptible to explanatory shortcuts) but is compatible with more general notions of weak emergence, e.g. the one introduced by Seth (see Section Relationship with other quantitative theories of emergence). Hence, our theory can be seen as an attempt at reconciling these approaches [36], showing how “strong” a “weak” framework can be.

An important consequence of our theory is the fundamental connection established between causal emergence and statistical synergy: the system’s capacity to host emergent features was found to be determined by how synergistic its elements are with respect to their future evolution. Although previous ideas about synergy have been loosely linked to emergence in the past [37], this is (to the best of our knowledge) the first time such ideas have been formally laid out and quantified using recent advances in multivariate information theory.

Next, we examine a few caveats regarding the applicability of the proposed theory, its relation with prior work, and some open problems.

Scope of the theory

Our theory focuses on synchronic [38] aspects of emergence, analysing the interactions between the elements of dynamical systems and collective properties of them as they jointly evolve over time. As such, our theory directly applies to any system with well-defined dynamics, including systems described by deterministic dynamical systems with random initial conditions [11] and stochastic systems described by Fokker-Planck equations [39]. In contrast, the application of our theory to systems in thermodynamic equilibrium may not be straightforward, as their dynamics are often not uniquely specified by the corresponding Gibbs distributions (for an explicit example, when considering the Ising model, Kawasaki and Glauber dynamics are known to behave differently even when the system is in equilibrium [40]; and thus may provide quite different values of the measures described in Section Measuring emergence). Finding principled approaches to guide the application of our theory to those cases is an interesting challenge for future studies.

In addition, given the breadth of the concept of “emergence,” there are a number of other theories leaning more towards philosophy that are orthogonal to our framework. This includes, for example, theories of emergence as radical novelty (in the sense of features not previously observed in the system) [41], most prominently encapsulated in the aphorism “more is different” by Anderson—see Refs. [42, 43], particularly his approach to emergence in biology (note that some of Anderson’s views, particularly the ones related to rigidity, are nevertheless closely related to the approach developed by our framework), and also articulated in the work of Kauffman [44, 45]. Also, contextual emergence emphasises a role for macro-level contexts that cannot be described at the micro-level, but which impose constraints on the micro-level for the emergence of the macro [46, 47]. These are valuable philosophical positions, which have been studied from a statistical mechanics perspective in Ref. [46, 48]. Future work shall attempt to unify these other approaches with our proposed framework.

Causality

The de facto way to assess the causal structure of a system is to analyse its response to controlled interventions or to build intervention models (causal graphs) based on expert knowledge, which leads to the well-known do-calculus spearheaded by Judea Pearl [10]. This approach is, unfortunately, not applicable in many scenarios of interest, as interventions may incur prohibitive costs or even be impossible, and expert knowledge may not be available. These scenarios can still be assessed via the Wiener-Granger theory of statistical causation, which studies the blueprint of predictive power across the system of interest by accounting non-mediated correlations between past and future events [12]. Both frameworks provide similar results when all the relevant variables have been measured, but can neverthelss differ radically when there are unobserved interacting variables [10]. The debate between the Wiener-Granger and the Pearl schools has been discussed in other related contexts—see e.g. Refs. [49, 50] for a discussion regarding Integrated Information Theory (IIT), and Ref. [51] for a discussion about effective and functional connectivity in the context of neuroimaging time series analysis (in a nutshell, effective connectivity aims to uncover the minimal physical causal mechanism underlying the observed data, while functional connectivity describes directed or undirected statistical dependences [51]).

In our theory, the main object of analysis is Shannon’s mutual information, I(Xt;Xt), which depends on the joint probability distribution pXt,Xt. The origin of this distribution (whether it was obtained by passive observation or by active intervention) will change the interpretation of the quantities presented above, and will speak differently to the Pearl and the Wiener-Granger schools of thought; some of the implications of these differences are addressed when discussing Ref. [7] below. Nonetheless, since both methods of obtaining pXt,Xt allow synergy to take place, our results are in principle applicable in both frameworks—which allows us to formulate our theory of causal emergence without taking a rigid stance on a theory of causality itself.

Relationship with other quantitative theories of emergence

This work is part of a broader movement towards formalising theories of complexity through information theory. In particular, our framework is most directly inspired by the work of Seth [6] and Hoel et al. [7], and also related to recent work by Chang et al. [52]. This section gives a brief account of these theories, and discusses how they differ from our proposal.

Seth [6] proposes that a process Vt is Granger-emergent (or G-emergent) with respect to Xt if two conditions are met: (i) Vt is autonomous with respect to Xt (i.e. I(Vt;Vt|Xt)>0), and (ii) Vt is G-caused by Xt (i.e. I(Xt;Vt|Vt)>0). The latter condition is employed to guarantee a relationship between Xt and Vt; in our framework an equivalent role is taken by the requirement of supervenience. The condition of autonomy is certainly related with our notion of causal decoupling. However, as shown in Ref. [14], the conditional mutual information conflates unique and synergistic information, which can give rise to undesirable situations: for example, it could be that I(Vt;Vt|Xt)>0 while, at the same time, I(V;Vt) = 0, meaning that the dynamics of the feature Vt are only visible when considering it together with the full system Xt, but not on its own. Our framework avoids this problem by refining this criterion via PID, and uses only the unique information for the definition of emergence.

Our work is also strongly influenced by the framework put forward by Hoel and colleagues in Ref. [7]. Their approach is based on a coarse-graining function F(⋅) relating a feature of interest to the system, Vt = F(Xt), which is a particular case of our more general definition of supervenience. Emergence is declared when the dependency between Vt and Vt is “stronger” than the one between Xt and Xt. Note that VtXtXtVt is a Markov chain, and hence I(Vt;Vt)≤I(Xt;Xt) due to the data processing inequality; therefore, a direct usage of Shannon’s mutual information would make the above criterion impossible to fulfill. Instead, this framework focuses on the transition probabilities pVt|Vt and pXt|Xt, and hence the mutual information terms are computed using maximum entropy distributions instead of the stationary marginals. By doing this, Hoel et al. account not for what the system actually does, but for all the potential transitions the system could do. However, in our view this approach is not well-suited to assess dynamical systems, as it might account for transitions that are never actually explored. The difference between stationary and maximum entropy distributions can be particularly dramatic in non-ergodic systems with multiple attractors—for a related discussion in the context of Integrated Information Theory, see Ref. [50]. Additionally, this framework relies on having exact knowledge about the microscopic transitions as encoded by pXt|Xt, which is not possible to obtain in most applications.

Finally, Chang et al. [4] consider supervenient variables that are “non-trivially informationally closed” (NTIC) to their corresponding microscopic substrate. NTIC is based on a division of Xt into a subsystem of interest, Xtα, and its “environment” given by Xt-α. Interestingly, a system being NTIC requires Vt to be supervenient only with respect to Xtα (i.e. Vt=F(Xtα)), as well as information flow from the environment to the feature (i.e. I(Xt-α;Vt)>0) mediated by the feature itself, so that XtVtVt is a Markov chain. Hence, NTIC requires features that are sufficient statistics for their own dynamics, which is akin to our concept of causal decoupling but focused on the interaction between a macroscopic feature, an agent, and its environment. Extending our framework to agent-environment systems involved in active inference is part of our future work.

Limitations and open problems

The framework presented in this paper focuses on features from fully observable systems with Markovian dynamics. These assumptions, however, often do not hold when dealing with experimental data—especially in biological and social systems. As an important extension, future work should investigate the effect of unobserved variables on our measures. This could be done, for example, leveraging Takens’ embedding theorem [5355] or other methods [56].

An interesting feature of our framework is that, although it depends on the choice of PID and ΦID, its practical application via the criteria discussed in Section Practical criteria for large systems is agnostic to those choices. However, they incur the cost of a limited sensitivity to detect emergence due to an overestimation of the microscopic redundancy; so they can detect emergence when it is substantial, but might miss it in more subtle cases. Additionally, these criteria are unable to rule out emergence, as they are sufficient but not necessary conditions. Therefore, another avenue of future work should search for improved practical criteria for detecting emergence from data. One interesting line of research is providing scalable approximations for Syn(k) and G(k) as introduced in Section Measuring emergence via synergistic channels, which could be computed in large systems.

Another open question is how the emergence capacity is affected by changes in the microscopic partition of the system (c.f. Section Defining causal emergence). Interesting applications of this includes scenarios where elements of interest have been subject to a mixing process, such as the case of electroencephalography where each electrode detects a mixture of brain sources. Other interesting questions include studying systems with non-zero emergence capacity for all reasonable microscopic partitions, which may correspond to a stronger type of emergence.

Conclusion

This paper introduces a quantitative definition of causal emergence, which addresses the apparent paradox of supervenient macroscopic features with irreducible causal power using principles of multivariate statistics. We provide a formal, quantitative theory that embodies many of the principles attributed to strong emergence, while being measurable and compatible with the established scientific worldview. Perhaps the most important contribution of this work is to bring the discussion of emergence closer to the realm of quantitative, empirical scientific investigation, complementing the ongoing philosophical inquiries on the subject.

Mathematically, the theory is based on the Partial Information Decomposition (PID) framework [14], and on a recent extension, Integrated Information Decomposition (ΦID) [20]. The theory allows the derivation of sufficiency criteria for the detection of emergence that are scalable, easy to compute from data, and based only on Shannon’s mutual information. We illustrated the use of these practical criteria in three case studies, and concluded that: i) particle collisions are an emergent feature in Conway’s Game of Life, ii) flock dynamics are an emergent feature of simulated birds; and iii) the representation of motor behaviour in the cortex is emergent from neural activity. Our theory, together with these practical criteria, enables novel data-driven tools for scientifically addressing conjectures about emergence in a wide range of systems of interest.

Our original aim in developing this theory, beyond the contribution to complexity theory, is to help bridge the gap between the mental and the physical, and ultimately understand how mind emerges from matter. This paper provides formal principles to explore the idea that psychological phenomena could emerge from collective neural patterns, and interact with each other dynamically in a causally decoupled fashion—perhaps akin to the “statistical ghosts” mentioned in Section Causal decoupling. Put simply: just as particles in the Game of Life have their own collision rules, we wonder if thought patterns could have their own emergent dynamical laws, operating at a larger scale with respect to their underlying neural substrate (similar ideas have been recently explored by Kent [57]). Importantly, the theory presented in this paper not only provides conceptual tools to frame this conjecture rigorously, but also provides practical tools to test it from data. The exploration of this conjecture is left as an exciting avenue for future research.

Supporting information

S1 Appendix. Provides the mathematical proofs of our results, and further details about simulations and preprocessing pipelines.

(PDF)

Acknowledgments

The authors thank Martin Biehl, Shamil Chandaria, Acer Chang, and Matthew Crosby for insightful discussions, the creators of the Neurotycho database for opening to the public such a valuable resource, and Yike Guo for supporting this research. F.R. is supported by the Ad Astra Chandaria foundation. P.M. and D.B. are funded by the Wellcome Trust (grant no. 210920/Z/18/Z). A.K.S. and A.B.B. are grateful to the Dr. Mortimer and Theresa Sackler Foundation, which supports the Sackler Centre for Consciousness Science. R.C-H. is supported by Ad Astra Trust, Tim Ferriss, The Nikean Foundation, and The Tamas Family.

Data Availability

All relevant data are within the manuscript and its Supporting information files.

Funding Statement

F.R. is supported by the Ad Astra Chandaria foundation. P.M. and D.B. are funded by the 887 Wellcome Trust (grant no. 210920/Z/18/Z). A.K.S. and A.B.B. are grateful to the Dr. Mortimer and Theresa Sackler Foundation, which supports the Sackler Centre for Consciousness Science. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Gibb S, Findlay RH, Lancaster T. The Routledge Handbook of Emergence. Routledge; 2019. [Google Scholar]
  • 2. Bedau M. Downward causation and the autonomy of weak emergence. Principia: An International Journal of Epistemology. 2002;6(1):5–50. [Google Scholar]
  • 3. Chalmers DJ. Strong and Weak Emergence. Oxford University Press; 2006. p. 244–256. [Google Scholar]
  • 4.Chang AY, Biehl M, Yu Y, Kanai R. Information Closure Theory of Consciousness. arXiv preprint arXiv:190913045. 2019;. [DOI] [PMC free article] [PubMed]
  • 5. Bedau MA. Weak emergence. Noûs. 1997;31:375–399. [Google Scholar]
  • 6. Seth AK. Measuring autonomy and emergence via Granger causality. Artificial Life. 2010;16(2):179–196. 10.1162/artl.2010.16.2.16204 [DOI] [PubMed] [Google Scholar]
  • 7. Hoel EP, Albantakis L, Tononi G. Quantifying causal emergence shows that macro can beat micro. Proceedings of the National Academy of Sciences. 2013;110(49):19790–19795. 10.1073/pnas.1314922110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Hoel EP. When the map is better than the territory. Entropy. 2017;19(5):188 10.3390/e19050188 [DOI] [Google Scholar]
  • 9. Klein B, Hoel E. The emergence of informative higher scales in complex networks. Complexity. 2020;2020 10.1155/2020/8932526 [DOI] [Google Scholar]
  • 10. Pearl J. Causality: Models, Reasoning and Inference. Cambridge University Press; 2000. [Google Scholar]
  • 11. Rosas F, Mediano PAM, Ugarte M, Jensen HJ. An information-theoretic approach to self-organisation: Emergence of complex interdependencies in coupled dynamical systems. Entropy. 2018;20(10). 10.3390/e20100793 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Bressler SL, Seth AK. Wiener–Granger causality: A well established methodology. Neuroimage. 2011;58(2):323–329. 10.1016/j.neuroimage.2010.02.059 [DOI] [PubMed] [Google Scholar]
  • 13.James RG, Ayala BDM, Zakirov B, Crutchfield JP. Modes of information flow. arXiv preprint arXiv:180806723. 2018;.
  • 14.Williams PL, Beer RD. Nonnegative decomposition of multivariate information. arXiv preprint arXiv:10042515. 2010;.
  • 15.Ay N, Polani D, Virgo N. Information decomposition based on cooperative game theory. arXiv preprint arXiv:191005979. 2019;.
  • 16. Lizier JT, Bertschinger N, Jost J, Wibral M. Information decomposition of target effects from multi-source interactions: Perspectives on previous, current and future work. Entropy. 2018;20(4). 10.3390/e20040307 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. James R, Emenheiser J, Crutchfield J. Unique information via dependency constraints. Journal of Physics A: Mathematical and Theoretical. 2018;. [Google Scholar]
  • 18. Ince RA. Measuring multivariate redundant information with pointwise common change in surprisal. Entropy. 2017;19(7):318 10.3390/e19070318 [DOI] [Google Scholar]
  • 19.Rosas F, Mediano P, Rassouli B, Barrett A. An operational information decomposition via synergistic disclosure. arXiv preprint arXiv:200110387. 2020;.
  • 20.Mediano PA, Rosas F, Carhart-Harris RL, Seth AK, Barrett AB. Beyond integrated information: A taxonomy of information dynamics phenomena. arXiv preprint arXiv:190902297. 2019;.
  • 21. Kraskov A, Stögbauer H, Grassberger P. Estimating mutual information. Physical Review E. 2004;69(6):066138 10.1103/PhysRevE.69.066138 [DOI] [PubMed] [Google Scholar]
  • 22. Lizier JT. JIDT: An information-theoretic toolkit for studying the dynamics of complex systems. Frontiers in Robotics and AI. 2014;1:37 10.3389/frobt.2014.00011 [DOI] [Google Scholar]
  • 23. McGill WJ. Multivariate information transmission. Psychometrika. 1954;19(2):97–116. 10.1007/BF02289159 [DOI] [Google Scholar]
  • 24. Timme N, Alford W, Flecker B, Beggs JM. Synergy, redundancy, and multivariate information measures: An experimentalist’s perspective. Journal of Computational Neuroscience. 2014;36(2):119–140. 10.1007/s10827-013-0458-4 [DOI] [PubMed] [Google Scholar]
  • 25. Rosas FE, Mediano PAM, Gastpar M, Jensen HJ. Quantifying high-order interdependencies via multivariate extensions of the mutual information. Physical Review E. 2019;100:032305 10.1103/PhysRevE.100.032305 [DOI] [PubMed] [Google Scholar]
  • 26. Rassouli B, Rosas FE, Gündüz D. Data Disclosure under Perfect Sample Privacy IEEE Transactions on Information Forensics and Security. 2019;. [Google Scholar]
  • 27. Conway J. The game of life. Scientific American. 1970;223(4):4. [Google Scholar]
  • 28. Reynolds CW. Flocks, Herds and Schools: A Distributed Behavioral Model. vol. 21 ACM; 1987. [Google Scholar]
  • 29. Adamatzky A, Durand-Lose J. Collision-based computing. Springer; 2012. [Google Scholar]
  • 30. Lizier J. The local information dynamics of distributed computation in complex systems. University of Sydney; 2010. [Google Scholar]
  • 31. Wolfram S. A new kind of science. Wolfram Media; 2002. [Google Scholar]
  • 32. Archer E, Park I, Pillow J. Bayesian and quasi-Bayesian estimators for mutual information from discrete data. Entropy. 2013;15(5):1738–1755. 10.3390/e15051738 [DOI] [Google Scholar]
  • 33.Vicsek T. Universal patterns of collective motion from minimal models of flocking. In: 2008 IEEE Conference on Self-Adaptive and Self-Organizing Systems. IEEE; 2008. p. 3–11.
  • 34. Chao Z, Nagasaka Y, Fujii N. Long-term asynchronous decoding of arm motion using electrocorticographic signals in monkey. Frontiers in Neuroengineering. 2010;3:3 10.3389/fneng.2010.00003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Dehaene S. Consciousness and the brain: Deciphering how the brain codes our thoughts. Penguin; 2014. [Google Scholar]
  • 36. Turkheimer FE, Hellyer P, Kehagia AA, Expert P, Lord LD, Vohryzek J, et al. Conflicting emergences. Weak vs. strong emergence for the modelling of brain function. Neuroscience & Biobehavioral Reviews. 2019;99:3–10. 10.1016/j.neubiorev.2019.01.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Corning PA. The synergism hypothesis: On the concept of synergy and its role in the evolution of complex systems. Journal of Social and Evolutionary Systems. 1998;21(2):133–172. 10.1016/S1061-7361(00)80003-X [DOI] [Google Scholar]
  • 38. Rueger A. Physical emergence, diachronic and synchronic. Synthese. 2000;124(3):297–322. 10.1023/A:1005249907425 [DOI] [Google Scholar]
  • 39. Breuer HP, Petruccione F, et al. The Theory of Open Quantum Systems. Oxford University Press; 2002. [Google Scholar]
  • 40. Smith TH, Vasilyev O, Abraham DB, Maciołek A, Schmidt M. Interfaces in confined Ising models: Kawasaki, Glauber and sheared dynamics. Journal of Physics: Condensed Matter. 2008;20(49):494237. [Google Scholar]
  • 41. Corning PA. The re-emergence of “emergence”: A venerable concept in search of a theory. Complexity. 2002;7(6):18–30. 10.1002/cplx.10043 [DOI] [Google Scholar]
  • 42. Anderson PW. More is different. Science. 1972;177(4047):393–396. 10.1126/science.177.4047.393 [DOI] [PubMed] [Google Scholar]
  • 43. Anderson PW. Basic notions of condensed matter physics. CRC Press; 2018. [Google Scholar]
  • 44. Kauffman S, Clayton P. On emergence, agency, and organization. Biology and Philosophy. 2006;21(4):501–521. 10.1007/s10539-005-9003-9 [DOI] [Google Scholar]
  • 45. Kauffman SA. A World Beyond Physics: The Emergence and Evolution of Life. Oxford University Press; 2019. [Google Scholar]
  • 46. Bishop RC, Atmanspacher H. Contextual emergence in the description of properties. Foundations of Physics. 2006;36(12):1753–1777. 10.1007/s10701-006-9082-8 [DOI] [Google Scholar]
  • 47. Atmanspacher H, beim Graben P. Contextual emergence. Scholarpedia. 2009;4(3):7997 10.4249/scholarpedia.7997 [DOI] [PubMed] [Google Scholar]
  • 48. Jensen HJ, Pazuki RH, Pruessner G, Tempesta P. Statistical mechanics of exploding phase spaces: Ontic open systems. Journal of Physics A: Mathematical and Theoretical. 2018;51(37):375002 10.1088/1751-8121/aad57b [DOI] [Google Scholar]
  • 49. Mediano PAM, Seth AK, Barrett AB. Measuring Integrated Information: Comparison of Candidate Measures in Theory and Simulation. Entropy. 2018;21(1). 10.3390/e21010017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Barrett A, Mediano P. The Phi measure of integrated information is not well-defined for general physical systems. Journal of Consciousness Studies. 2019;26(1-2):11–20. [Google Scholar]
  • 51. Seth AK, Barrett AB, Barnett L. Granger causality analysis in neuroscience and neuroimaging. Journal of Neuroscience. 2015;35(8):3293–3297. 10.1523/JNEUROSCI.4399-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Chang AYC, Biehl M, Yu Y, Kanai R. Information closure theory of consciousness. Frontiers in Psychology. 2020;11:1504 10.3389/fpsyg.2020.01504 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Takens F. Detecting strange attractors in turbulence In: Dynamical Systems and Turbulence. Springer; 1981. p. 366–381. [Google Scholar]
  • 54. Cliff OM, Prokopenko M, Fitch R. An information criterion for inferring coupling of distributed dynamical systems. Frontiers in Robotics and AI. 2016;3:71 10.3389/frobt.2016.00071 [DOI] [Google Scholar]
  • 55. Tajima S, Kanai R. Integrated information and dimensionality in continuous attractor dynamics. Neuroscience of consciousness. 2017;2017(1):nix011 10.1093/nc/nix011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Wilting J, Priesemann V. Inferring collective dynamical states from widely unobserved systems. Nature Communications. 2018;9(1):1–7. 10.1038/s41467-018-04725-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Kent A. Toy Models of Top Down Causation. arXiv preprint arXiv:190912739. 2019;.
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008289.r001

Decision Letter 0

Daniele Marinazzo

20 Jul 2020

Dear Dr. Rosas,

Thank you very much for submitting your manuscript "Reconciling emergences: An information-theoretic approach to identify causal emergence in multivariate data" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers.

The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations, and to the notes below.

The paper has the cut and the style for a more technical/physical journal, but we see the reasons for your submissions to PLOS CB, and agree that the paper can provide some new technical and conceptual fuel to the debate on the mechanisms of consciousness.

Still, it would be appreciated if, keeping the technical part intact, you could insist a bit more on the biological aspect, and discuss the generalizability to other protocols and datasets.

Furthermore, even if the data are from a public repository, and the calculations are done with a freeware tool, please make sure that you provide on a third party repository some scripts to reproduce the figures. 

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. 

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: What can I say - I re-read this piece a few times and I cannot really find any fault. It deals with what, in my mind, is the most challenging piece of mathematical modeling - the modeling of emergence. The authors achieve this in a very accomplished manner both theoretically and- very importantly, practically.

I guess once in a while the only thing left to a reviewer is to congratulate the authors for a masterful submission - and this is the time to do that

Reviewer #2: The authors present a new theoretical framework for thinking about the issue of emergence in complex systems. By considering features extracted from a multivariate dynamical system they define two different types of evolving relationship: causal decoupling and downward causation. They provide specific measure based on PID to quantify these properties, and courser level measures to demonstrate their existence with quantities that are easier to compute in practise.

For a highly technical subject matter the manuscript is very clearly presented and was a pleasure to read. They situate their approach well in comparison to existing techniques from complex systems and the study of consciousness.

Major comments

The examples are very nice but I think ECoG data need a null comparison. There is a strong effect of timescale on the measures considered, but neural signals can be strongly autocorrelated with high power in low frequencies and this autocorrelation can change the bias properties of information theoretic measures as a function of frequency / timescale. I think a full treatment of the bias properties of these measures is certainly beyond the scope of this paper, but a simple shuffled control would be enough to demonstrate whether the profiles seen might be influenced by autocorrelation / filtering parameters. For example, the whole analysis procedure could be repeated with the same filtering and cross-validation pipeline, but at the start all the wrist position timecourses used for training should be permuted (e.g. random circular shift or shuffling across trials). This would lead to a classifier Vt that does not extract any meaningful information from the ECoG, but has the same properties induced by autocorrelation, filtering and regularisation. The plots in panel d) could be shown for 1 such random permutation to see if there is any similar dependence on timescale. (or ideally run many times and mean + spread reported, but understand that may be computationally demanding).

Minor comments

It would be nice to have a bit more discussion and motivation of the definition of supervenient feature, as it seems this is a key area where this framework differs from others (discussed nicely P10-11). In discussion P9 supervenient is described as a property “that can be computed from the state of the system”. But the definition P3 is stronger than that. Would be nice to have some more motivation where the definition is introduced, perhaps with some intuitive examples of features that would be supervenient, and those that would not meet the definition.

P3 “information that is provided independently (and hence redundantly) by both of them”. Not sure I agree that {1}{2} PID term should be called independent information. In the two-bit copy I think its fair to say each predictor provides information about the target independently, but they are not redundant. Maybe could just say “information that is provided by both of them”.

P5 Eqs 5 and 6. Could be clearer for the reader to put the term for D(k) and G(k) directly by the definition, ie “we introduce the downward causation, denoted D(k), and the causal decoupling indices, denoted G(k)” and switch the order of equations to match the order they are presented in the text.

P5. Eq 9. Definition of Un(k) conditioning on multiple terms wasn’t clear to me (ie how to relate to definition on p3)

P8, +/- uncertainty are reported on the numerical results, but not told what these are (s.d., s.e.m., confidence interval etc.) Similarly Fig 5. Error bars not specified.

P9 “linking supervenience to static and causal power to dynamic properties” sentence unclear, add commas or repeat the word properties?

P16: typo “pecifically”.

P17 “calculated using … JIDT”. More details of which method and associated options? (JIDT implements a wide range of estimators)

Reviewer #3: Review of Rosas et al. “Reconciling emergences: An information-theoretic approach to identify causal emergence in multivariate data”

This study proposed possible information theoretic formulations of causal emergence. Specifically, the study distinguished downward causation and causal decoupling and showed critical conditions for the existence of them both theoretically and practically. What is particularly exciting about this study is the application of PID and Phi_ID in this context, because they clarify and disentangle various existing ideas surrounding the notion of emergence. I’m particularly impressed by the theorems presented here, because they allow us to detect causal emergence when we do not know exact ways to construct features. I do not have major concerns regarding the contents of the paper.

I have one general question about causal decoupling. I understand that decoupling exists (i.e. is defined) mathematically in the parity dynamics example. But I wonder whether this can occur in physical interactions which should cover much smaller part of all possible dynamics. So I’m curious to know whether causal decoupling is possible for a dynamics where the dynamics is determined by direct interactions among micro elements (i.e., the state of one element is determined by the states of other elements). Intuitively, if I consider physical implementations of the parity case, I would think we need an additional physical element to directly store the parity of the current state. If this is indeed required, the process itself relies on something like ghost (i.e., not existent, but is used for computation). In other words, what I’m suggesting is that decoupling exists only in mathematics, but not in physics. So I was curious how Gamma behaves in application cases. But the Gamma was not zero, so it may be difficult to find an example of decoupling in real/simulation cases. Could the authors comment on this if they had any thoughts?

For practical applications, the authors focused on k=1 cases. I understand that this is a pragmatic choice, but I wonder how results may different if we consider k>1 cases. Do we need to consider them to be more precise or is there anything that we may miss and be cautious about?

Very minor points:

Page 3: A formal definition of a supervenient feature would help clarify what kind of functions are assumed.

Page 5. The order of eq 5 an eq6 should be reversed for consistency with the text.

Page 11: “Taken’s” should be “Takens’” or “Takens’s” as the name is “Takens”. Relatedly, Satohiro Tajima had a few papers motivated by Takens/Sugihara embedding to compute integrated information using delay embedding.

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

Reviewer #2: No: ECoG data are from a pubilc repository. No reference to availability of simulated data or processed data, and data underlying graphs is not provided in spreadsheet form.

Reviewer #3: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, PLOS recommends that you deposit laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see http://journals.plos.org/ploscompbiol/s/submission-guidelines#loc-materials-and-methods

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008289.r003

Decision Letter 1

Daniele Marinazzo

25 Aug 2020

Dear Dr. Rosas,

We are pleased to inform you that your manuscript 'Reconciling emergences: An information-theoretic approach to identify causal emergence in multivariate data' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

***********************************************************

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1008289.r004

Acceptance letter

Daniele Marinazzo

4 Dec 2020

PCOMPBIOL-D-20-01042R1

Reconciling emergences: An information-theoretic approach to identify causal emergence in multivariate data

Dear Dr Rosas,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Nicola Davies

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. Provides the mathematical proofs of our results, and further details about simulations and preprocessing pipelines.

    (PDF)

    Attachment

    Submitted filename: response.pdf

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting information files.


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES