Skip to main content
Sage Choice logoLink to Sage Choice
. 2021 Feb 16;28(4):382–399. doi: 10.1177/1073858421994784

A Complex Systems Perspective on Neuroimaging Studies of Behavior and Its Disorders

Federico E Turkheimer 1,, Fernando E Rosas 2,3,4, Ottavia Dipasquale 1, Daniel Martins 1, Erik D Fagerholm 1, Paul Expert 5, František Váša 1, Louis-David Lord 6, Robert Leech 1
PMCID: PMC9344570  PMID: 33593120

Abstract

The study of complex systems deals with emergent behavior that arises as a result of nonlinear spatiotemporal interactions between a large number of components both within the system, as well as between the system and its environment. There is a strong case to be made that neural systems as well as their emergent behavior and disorders can be studied within the framework of complexity science. In particular, the field of neuroimaging has begun to apply both theoretical and experimental procedures originating in complexity science—usually in parallel with traditional methodologies. Here, we illustrate the basic properties that characterize complex systems and evaluate how they relate to what we have learned about brain structure and function from neuroimaging experiments. We then argue in favor of adopting a complex systems-based methodology in the study of neuroimaging, alongside appropriate experimental paradigms, and with minimal influences from noncomplex system approaches. Our exposition includes a review of the fundamental mathematical concepts, combined with practical examples and a compilation of results from the literature.

Keywords: neuroimaging, complexity science, emergence, multi-scale, self-organization, evolution

Introduction

One of the central aims of contemporary neuroimaging is to unify the models of neural structure and function in order to shed light, not just on the evolutionary organization of the brain, but also on its output, emergent behavior, as well as on its disorders. Intuitively, one would expect the etiology of brain disorders to be mainly driven by a few dominant anatomical regions. However, evidence accumulated in recent years has unraveled a different picture. In fact, most studies indicate that the functional and structural variance in the human brain—during normal functioning, development, ageing, and disease—tends to be widely distributed and can hardly be attributed to a set of specific regions (Ecker and others 2010; Thompson and others 2001; Zhuo and others 2019). These findings have influenced the paradigms used to interpret neuroimaging data, which are gradually evolving from phrenological perspectives—that aimed to find associations between the variance in the data and parsimonious sets of regions (see Donaldson 2004; Uttal 2001, for two contrasting views on the subject)—to embrace distributed analyses (Sporns 2011a) that map the data into regional regulatory networks. This latter approach has been borrowed from the rich literature of complexity science whose main aim is to progress our understanding of the essential features of complex systems. These two perspectives currently coexist in neuroimaging analyses, and their tools are used, sometimes interchangeably, on similar datasets or within the same study (Brauchli and others 2019; Hebart and Baker 2018; Pillet and others 2019).

In this work, we combine a number of mathematical concepts with practical examples and observations from the literature to build a case in favor of the use of complexity science analytics to study the brain through neuroimaging techniques. We aim to demonstrate that if the brain is a complex system, and there is strong evidence that it is, then the use of traditional models may not fully describe the outcomes of imaging experiments. This issue, we believe, is one of the key drivers behind the lack of robustness, replicability, and, ultimately, poor interpretability of a significant portion of contemporary neuroimaging studies. As a solution, we suggest that when a model of complexity is adopted, the choice of concepts and tools should be embraced together with appropriate experimental designs

The Brain as a Complex System

While the exact definition of a complex system is still under debate (Ladyman and others 2013), there is consensus on four properties that all systems characterized as “complex” necessarily share. These properties are as follows (Jensen 1998):

  1. Multiplicity. A complex system is made up of a large number of components, each of which interacts with its neighboring components in relatively simple ways.

  2. Nonlinearity. The interactions between components are nonlinear, for example, changes in the output are not always proportional to the corresponding changes of the input.

  3. Self-organization. The system self-organizes spontaneously in the absence of any form of centralized control.

  4. Emergence. The system exhibits emergent behavior, that is, the macroscopic behavior cannot be understood purely in terms of the microscopic interactions (Turkheimer and others 2019). Rather, the emergent behavior may result from (a) multiscale 1 organization; (b) information processing capability; (c) dynamical spatiotemporal patterns; and (d) evolution.

Of these four properties, multiplicity and nonlinearity can be seen as the fundamental enablers that—in the right circumstances—allow self-organization and emergence to happen.

This article analyses the idea of the brain as a complex system, exploring each of these four properties under the light of recent evidence provided by contemporary neuroscience. It is to be noted that these properties have been studied in depth in the neuroscience literature (Bassett and Gazzaniga 2011; Bullmore and others 2009; Deco and others 2011; Friston 2009; Kelso 1995; Sporns 2011b; Telesford and others 2011); of these references, the ones that are of practical use (Bassett and Gazzaniga 2011; Kelso 1995; Sporns 2011a) do not have general but rather specific content, while those with more ambitious and general scope (Friston 2009; Tononi and others 1994; Wolfram 2002) still have limited practical value. Here we wish to present a unified and simplified account that looks at the four properties above with their interrelationships and consequences in experimental neuroimaging. An introductory brief for each of the four properties is presented in Boxes 1 to 4.

Box 1. Multiplicity.

A dynamical system that consists of a large number of interacting components can arrange itself in a variety of configurations—so-called “microstates”—and the total number of these microstates is known as its “multiplicity.” We can, for instance, consider a simple dynamical system that consists of only three components that can each take two states: blue or red. One possible state of this system is shown on the right.Inline graphicIn general, the multiplicity Ω of any system is given by the following relationship:

Ω=(no.ofstates)(no.ofcomponents)

or in words: the multiplicity is the number of possible states each element is able to take, to the power of the number of components comprising the system. For our three-component toy model, the multiplicity is

Ω=23=8

Inline graphic Given the simplicity of this system, we can show all of the possible combinations on the left. However, for any real system composed of millions of interacting components this would not be feasible, which is when one needs to treat the system in question via statistical tools. One of the central quantities in statistical physics is the Boltzmann entropy S:

SlogΩ

or in words: the entropy of a system is proportional to the logarithm of its multiplicity. Thanks to the properties of logarithms, one can find that the Boltzmann entropy satisfies the following property:

logΩ=log[statescomp]=comp×log[states]

which implies that the entropy of a system grows linearly (see Box 2) with the number of its constituent components.

Here we provide evidence that (A) brain activity, even during simple functional tasks, is associated with very large and distributed networks that (B) dynamically interact and form transient states.Inline graphic

Box 4. Emergence.

During the process of self-organization, the interaction among elements generates an organized system with properties referred to as being “emergent.” These properties are unique to complex systems and can be summarized by the following three points:

1. Emergent properties cannot be explained solely by the physical attributes of the components which comprise the system.

2. Once generated via interactions between constituent components, the emergent properties operate and interact with the environment via laws that are not reducible to those of the interacting elements.

3. The emergent properties supersede previously dominant properties of the system and may become the most important factor in determining the behavior of the interacting elements—as indicated by the large downward-pointing arrows on the right of the diagram below.

The brain is organized in a hierarchy of spatial scales from the cellular milieu (e.g., blood vessels, neurons, and glia) to brain tissues, nuclei, and cortical layers (Box 3), which, ultimately, are then further arranged into cyto-architectonic regions. Within this hierarchy, each layer exhibits a function that seems to “emerge” from the function of layers at smaller scales. It follows that the top level of this hierarchy, human behavior, emerges from interactions within and between these different spatial scales.

Box 4. Emergence

Box 2. Nonlinearity.

A good way to describe a nonlinear system is to describe what it is not—that is, linear. Linearity means that whatever a system receives as input, it will return in proportional amounts as an output, see top two rows of the figure below.

Linear systems are easy to deal with due to their mathematical simplicity and associated computational expediency when used in numerical applications. In contrast, a nonlinear system does not respond in such a simple way, in that it will produce an output that is not directly proportional to its input. Using a similar example as above, the now nonlinear system in the bottom row of the figure below receives the same input as the linear response in the middle row, but its output is not proportional to the magnitude of the input anymore.

Most natural phenomena are nonlinear and are therefore better described by similarly nonlinear equations. However, we pay the price for the complexity in these expressions by the associated increase in computation time.

In the case of neural systems, the couplings between stimuli and measurements are generally nonlinear, where these measurements are taken from the small-scale electrophysiological recordings to the large-scale functional imaging data. For |computational feasibility, linear analytical models are at times used as an approximation of the underlying nonlinear dynamics.

        Inline graphic

Box 3. Self-Organization.

It is sometimes possible for a coherent organization to emerge within a dynamical system that is initially in a disordered state. This organization manifests itself at large spatial scales by virtue of local interactions among the system’s constituent components. This phenomenon is known as self-organization and it relies upon the dynamical system in question possessing four essential attributes:

1. There must be a balance between exploitation and exploration. This feature of any self-organizing system means that its interacting elements are able to strike an equilibrium between their tendency to (a) remain within their current region and exchange energy with their environment (exploitation) and (b) venture into new territories within the system (exploration).  Inline graphic  As seen in the diagram on the left, these two attributes can balance to allow the system to form an ordered state.

2. The system must be nonlinear (see Box 2). Briefly, this means that the strength of an input into the system is not proportional to the strength of its output.

3. There must be a large number of small-scale (local) interactions. The contribution of each of these interactions is vanishingly small (right diagram), but together they are able to form a large-scale system with the appearance of organization.              Inline graphic

4. The system must be able to resist the tendency toward disorder (via the second law of thermodynamics) by the availability of sufficient energy to maintain self-organization.

Brain tissue is currently thought to present self-organization at every scale from the local cortical unit (A) to the larger integration of cortical areas (B) and its organization principles are similar at every scale.

Box 3. Self-Organization

From Phrenology to AN Extended View of Neuroimaging Models

Let us start by considering the nature of the various parts that make the brain, which is related to the first property of complex systems. One of the longest-standing challenges in neuroscience is to decipher how variation in neural systems—both structural and functional—maps onto variation in behavioral phenotypes. The initial approach used in the analysis of imaging data was heavily influenced by a highly localized perspective relating neural and phenotypic variability (Uttal 2001). Its use stemmed from the success of neurology/neuropsychology at making differential diagnoses, relating dissociable cognitive/perceptual disorders to apparent localized damage. Even though the localized damage was just a partial view of the underlying pathology, it facilitated the easy intuition that there is a clear mapping between form and function. As a consequence, it focused on the use of mass univariate testing to map brain correlates of such phenotypic variation (Friston and others 1991; Worsley and others 1996). This paradigm triggered a long and enduring controversy, facing those with the view that it is paramount to protect results from false positives by using stringent significance thresholds (Eklund and others 2016), against those concerned about the loss of information due to high rates of false negatives (Lieberman and Cunningham 2009). To a large extent, this discussion stems from conceptual differences in the a priori expectation of the distribution of the brain’s signals associated with a certain phenotype. Whereas the former position assumes that variability is highly concentrated in very few brain loci, the latter expects wider associations across brain networks.

Interestingly, this debate is reminiscent of a similar discussion that took place in the early 1900s among geneticists, who were divided between the Mendelians—those inspired by Mendel’s work on pea genetics and saw phenotypes as the result of discrete, monogenic variants—and the biometricians—who did not accept that the inheritance of a continuous distribution of traits (e.g., height, intelligence, etc.) could be related to individual genes. At the time, R. A. Fisher resolved the issue by presenting his seminal work on the infinitesimal model (Fisher 1918), in which the inheritance of continuous traits could be explained by large numbers of variants, each of infinitesimal effect. In this theoretical work, Fisher conjectured that a large number of loci should be associated with what, at the time, were termed “complex traits.” Only about 50 years have passed since the first genetic markers allowed for the detection of variants that have major effects, and only 20 years since single-nucleotide polymorphism technology provided dense markers throughout the genome. The combination of technological advancements, cost reductions, and increased sample sizes has allowed for the gradual expansion of models describing the genetic foundations of complex traits. It is now generally accepted that common single-nucleotide polymorphisms with small individual effect sizes—generally below the statistical significance levels of experimental designs—account for most of the variability of common genetic traits (Goldstein 2009; Shi and others 2016; Yang and others 2010). Equally, complex disorders that are defined over a clinical spectrum demonstrate similarly broad genomic patterns (Smoller and others 2019). For example, recent work on the genetic architectures of schizophrenia has revealed that more than 70% of 1-Mb genomic regions contain variants influencing the risk of the disorder (Loh and others 2015). In fact, a more general “omnigenic” view of complex traits and their heritability and evolution is now driving most of the genetic contemporary discourse (Boyle and others 2017; Sella and Barton 2019).

Moving from the genotype to the intermediate phenotype, there is increasing evidence that some of the remarkable computational capabilities of the brain, such as object recognition, depend on the ability of neural systems to encode simple inputs into redundant widespread representations (DiCarlo and others 2012) and that this distributed architecture supports executive function (Steinmetz and others 2019). The conceptual models used in neuroimaging studies have evolved in the same direction. The field has been gradually shifting its focus from the analysis of variance to the analysis of covariance, for example, moving from localized effects to distributed networks (Bullmore and Sporns 2009; Friston and others 1993; Greicius 2008; McIntosh and others 1996; McKeown and others 1998). However, a conscious, argumentative debate between these approaches has been lacking, and the two methods still coexist. In particular, there remains the unanswered question of whether complex behavior is supported by activity concentrated in few loci or by widely distributed networks. The latter view is supported by substantial evidence: Gonzalez-Castillo and colleagues used very large sample sizes or very extended single subject designs to reveal brain-wide functional support for simple tasks in functional magnetic resonance imaging (fMRI) studies (Gonzalez-Castillo and others 2012; Gonzalez-Castillo and others 2015). Similarly, Haxby and colleagues (Haxby and others 2001) demonstrated that the representations of faces and objects in ventral temporal cortex are widely distributed and overlapping; moreover Chu and colleagues (Chu and others 2012), using structural data, demonstrated that feature selection (e.g., regional variation) was as effective as total brain volume at predicting the progression from minimally cognitively impaired cohorts to Alzheimer’s disease.

A complex system perspective on neural function results in predictions for the distribution of measured neural activity across the brain, which differs from predictions due to the localized perspective. For discussing some of these predictions, here we consider fMRI data from the Individual Brain Charting (IBC) project (Pinho and others 2018), which consists of a curated fMRI dataset, acquired in a fixed environment from a permanent cohort of the same 12 participants during the performance of a dozen tasks, encompassing a wide range of motor and psychological domains. The results are presented in Figure 1, following the format in Boyle and others (2017) where the z-scores obtained from the testing of each task-versus-control contrast were sorted in descending order and plotted versus the log of their rank (only positive z-scores were considered). Figure 1A and B depicts simpler (mostly motor) and higher mental processing tasks, respectively.

Figure 1.

Figure 1.

The panels illustrate the vectors of z-scores calculated for each voxel by contrasting the activity of task-versus-baseline (positive values only). Maps were obtained from fMRI studies of simple/motor tasks (A) or more complex activities such as calculation, language comprehension and social cognition (B). The x-axis represents the logarithm of the z-score ranks. The red lines represent the same z-scores for the normal null distribution. The blue lines represent the James-Stein boundaries, that is, the minimal decay of a z-score distribution that would justify the use of statistical thresholds for the estimation of the z-scored vector for each contrast (see Equation 1). The boundary curve has been arbitrarily scaled to the maximum z-score value for the group of tasks for which the theory suggests the use of shrinkage estimators for all vectors with smoother decays.

These plots are to be contrasted with the expected null distribution (red) and the decay (blue):

zk=Ck12, (1)

where zk represents the z-score of rank k ; C is an arbitrary constant (which in the two plots in Figure 1 was fitted to the max z-score in each panel); and the exponent −1/2 is a minimum bound on the rate of decay of the z-scored vector—the estimation of which warrants the use of a statistical threshold, whereas vectors with a slower rate of decay require approaches based on weighted averages of the whole distribution (Donoho and Johnstone 1994; Turkheimer and others 2003).

It may be seen that the curves in Figure 1 decay in a smooth manner—even in the simpler tasks, in which one would expect the signals to be described by a small number of coefficients with steeper decay rates—and do not satisfy the minimum requirement for the use of thresholding. In other words, the boundary in Equation (1) is a way of formally stating that the use of thresholds to separate signal from noise in a multivariate vector is unwarranted when no such clear separation exists. In fact, any small change in experimental setting, imaging technology, preprocessing, as well as interindividual variability (both neural activation, morphological, or nonneural) could easily swap the ranks of the z-scores, thus limiting the reproducibility of the results (Turkheimer and others 2003). Furthermore, these plots confirm that the activity supporting the execution of a very large variety of tasks is distributed across much if not the entire neural system and that the brain therefore satisfies the first requirement in being classified as a complex system. Note that, as each voxel is itself the numerical reflection of the compounded activity of the unit volume, the observations made for the time-space scales of fMRI studies may well be valid for measurements at higher resolutions made with other technologies.

Interaction and Nonlinearity

Let us consider now nonlinearity, the second key aspect of complexity. Nonlinearities within a complex system arise as a result of interactions between its elements. The brain offers several examples of nonlinear relationships between its physiological parameters. For example, there is the way in which V1 neurons respond to inputs (Williams and Shapley 2007)—where object recognition is performed by nonlinear operations on distributed, largely redundant representations (Williams and Shapley 2007), or the response of evoked deep tissue pain (Loggia and others 2012)—where the response is linearly related to pain intensity in certain regions, but is nonlinear in others.

However, the point we want to make in this section is that nonlinearities in any system can be generated by the interactions among elements when changes are applied to the organic whole. To illustrate this perspective, let us start with a simple theoretical illustration which, although applied to a small number of elements, is quite effective to show how—in biological scenarios—simple environmental constraints create nonlinear interactions among the elements of a group. 2 We will then show how these inferences might apply to a real scenario. The simple example considers a population of rabbits that reproduce regularly with a certain birth-rate R. The relationship between the number of rabbits at time t, denoted by Nt, and the corresponding quantity at time t + 1 can be expressed as

Nt+1=RNt. (2)

In particular, if one considers the case R = 2 and runs three iterations of the model, starting from one rabbit at t = 0 we obtain eight rabbits at t = 3, while if we start with two rabbits the final total is 16. This confirms that this system is linear and that the totality of the system is equal to the sum of the parts, as two rabbits produce twice the offspring of a single one.

The linear model above can be readily turned into a nonlinear complex system with interacting elements if we introduce a constraint wherein offspring need to compete for resources. Then, the model in (2) turns into the simple, but effective Verhulst model (Verhulst 1838):

Nt+1=R(NtNt2Pmax), (3)

where Pmax represents the maximum capacity of the environment to keep the offspring alive. Setting Pmax=10 , we run 3 iterations of this algorithm with an initial state of one rabbit to obtain four offspring. Running the same model with three iterations and starting with two rabbits we obtain five offspring. In this case, the offspring of two rabbits is not two-times the off-spring of one rabbit (see example illustrated in Fig. 2).

Figure 2.

Figure 2.

The panel above illustrates a simple linear growth model whereby the offspring of two rabbits is always twice the number of offspring of a single rabbit, whatever the number of iterations. In the logistic model below, a resource constraint makes the rabbits compete for resources and, this time, the offspring of two rabbits is not twice the offspring of one.

The example in Figure 2 demonstrates three important phenomena. First, a simple resource constraint on the system introduces a highly nontrivial interaction between elements, so that the totality is not anymore equal to the sum of the parts. Second, this interaction translates into a nonlinear behavior observed within the system. Third, the reproduction rate in (3) depends on the initial number of rabbits Nt.

The insights developed over the simple example can be used to understand more complicated issues related to brain development, brain ageing, and their disorders. For example, it has proven difficult for neuroimaging studies in autism to replicate both structural and functional results (He and others 2020; Lotspeich and others 2004), whereas the most consistent finding is the heterogeneity of imaging results (Lenroot and Yeung 2013). One particularly interesting finding is related to brain development of subjects within the autistic spectrum: these populations exhibit increased brain volumes in young children, whereas the rate of growth decreases subsequently during adolescence to meet the normal developmental curve between 10 and 15 years of age, with the volume of many structures finally declining atypically into adulthood (Ha and others 2015; Lange and others 2015; Wolff and others 2018). These findings are paradoxical from the perspective of a linear system; however, it is likely that a suitable nonlinear model of growth could explain why a more numerous initial state would experience a strong bias downwards in its developmental rate, especially when the whole system is placed under exogenous pressure.

More generally, it is often noticeable in neuroimaging that individual developmental trajectories seem to provide much better resolution to categorize psychiatric disorders than cross-sectional data (Liberg and others 2016). This phenomenon is known in the literature of complexity science and dynamical systems as nonergodicity, which characterizes dynamical systems in which the statistics computed over single trajectories are not necessarily equal to the statistics across the population (Griffith 1963). This is equivalent to saying that trajectories that start from different initial conditions may exhibit different dynamics. Nonergodicity has dramatic consequences in the context of ageing and age-related disorders, where imaging studies have shown that cross-sectional analyses produce findings that are often in discrepancy with longitudinal analyses on the same subjects (Nyberg and others 2010). In particular, longitudinal analyses unveil complex nonlinear trajectories of structural and functional data of anatomically parcellated regions, which can have radically different dynamics challenging most available explanatory models (Fjell and others 2013; Fox and Schott 2004; Good and others 2001; Vinke and others 2018). Notably, during normal ageing, some brain areas may exhibit little age effects on regional cerebral volume, while others show a complex, nonlinear relationship with age. This is mirrored by the significant intrasubject variability of cognitive decline—see, for example, the findings of the Nun Study (Snowdon and Nun 2003), in which cognitive domains within the same individual deteriorate at very different rates (Salthouse 2019).

When it comes to connecting the intermediate phenotype (i.e., the brain) with environmental modifiers on the one hand and genetic factors on the other, recent analysis of very large datasets have demonstrated that individual variability of MRI measures across ageing is spread across a large number of regional networks, or modes—each associated with overlapping environmental factors and genes (Smith and others 2020). Interestingly, the biological and genetic factors involved in ageing accrue over three fundamental processes: metabolism, stress (particularly its effects on cellular senescence), and immunity. However, each of these processes is in itself a complex outcome of interactions among multiple genes and environmental factors, such as nutrition, environmental/psychosocial stress, and infections (Kim 2007; Shmookler and others 2009).

Motivated by the aforementioned evidence, we propose here that the complex developmental and neurodegenerative patterns observed in structural and functional data are the result of nonlinear interactions generated by systemic pressures, which stem from metabolic and immune challenges punctuated across time (Amad and others 2020; Lord and others 2012; Turkheimer and others 2020). For example, metabolism is a fundamental determinant of neural function, as the brain accounts for ~2% of total body mass, yet burns ~20% of its energy, which is obtained through oxidative glucose metabolism, thus rendering the brain in constant demand of oxygen supply from cerebral blood flow (Clarke and Sokoloff 1999). In fact, the human brain is, from a metabolic perspective, a scalable version of a primate brain, as the energy budget of the whole brain per neuron is approximately fixed across species and brain sizes (Herculano-Houzel 2011) and hence its development has been strictly controlled by metabolic constraints (Fonseca-Azevedo and Herculano-Houzel 2012). Brain metabolism is mostly devoted to sustaining synaptic activity (Sibson and others 1998) and metabolic rates influence neural circuitry and activity patterns by exerting selective pressure toward metabolically efficient wiring patterns (Chen and others 2006), gray/white matter segregation (Wen and Chklovskii 2005), neuronal morphology (Wen and Chklovskii 2008), and neural codes (Balasubramanian and others 2001; Levy and Baxter 1996). The metabolic demands of neural circuitry render the system highly susceptible to mitochondrial impairments, oxidative stress, and deficits in glucose metabolism and perfusion. Therefore, the resulting global metabolic pressure stemming from noxious metabolic events translates into complex, nonlinear interactions among the system’s constituent parts.

Following the same principles, we propose that immunity-based secondary processes, that are recurrently associated to either developmental or neurodegenerative disorders, may generate system-wide pressures that then translate into complex outcomes. In fact, either maternal infections during gestation, or activation of the innate or adaptive immune system caused by stressful events, trauma, or infections across a lifespan, have demonstrated to be the cause of psychiatric or degenerative conditions (Gate and others 2020; Labzin and others 2018; Meltzer and Van de Water 2017; Meyer 2019; Miller and Raison 2016; Richetto and others 2017).

Self-Organization and Criticality

The third key property of complex systems is self-organization. Self-organization takes place where some form of global order arises as a consequence of the local nonlinear interactions between parts of an initially disordered system (Bak 1996). Importantly, the organization should emerge in a spontaneous manner, provided energy is available, and not as a result of the actions of a centralized controller or an external guidance. Therefore, the self-structuring events are often triggered by seemingly random fluctuations, which are dynamically amplified by some form of feedback enabled by the underlying nonlinearities. The resulting organization is wholly decentralized, being distributed over all the components of the system, making it robust and able to endure (i.e., self-repair) or recover from substantial perturbations (Kelso 1995). The process of self-organization of a dynamical system such as the brain can be visualized as a trajectory in n-dimensional phase space (where n is the number of its elements) that moves across states that can be described in terms of steady states. The evolution of the system is constrained to remain on these states and this constraint implies some form of coordination between its constituent elements (Ashby 1947; Rosas and others 2018). A substantial amount of evidence, based on both human and animal experiments, indicates the important role of neural plasticity in the maturation of perceptual and cognitive processes. In other words, neural structures underlying these functions require sensory input for their maturation, and genetic instructions are not sufficient to specify neuronal connections with significant precision (Singer 1986; Tetzlaff and others 2010).

Importantly, the brain (as many natural complex systems) displays macroscopic patterns of activity with spatial and/or temporal scale invariance. The latter is a property which results in dynamical systems, for example, those exhibiting turbulent flow (Bohr 1998; Deco and Kringelbach 2020), evolving in such a way that (except at the spatial or temporal extremes) it is not possible to distinguish the scale at which the system is being observed. This self-similar (or “fractal”) character suggests that the self-organizing process that is ongoing during the development of the central nervous system (CNS) drives the system toward a state called self-organized criticality (SOC; Bak 1996; Jensen 1998). SOC was originally proposed as an explanatory mechanism for the origin of complexity; systems displaying SOC possess long-range correlations both in time and space (Expert and others 2011). The philosophy underlying the SOC paradigm can be intuitively summarized as follows: the current state of a system is a result of all its past states. Per Bak used the sand pile model as an exploration of SOC (Bak and Paczuski 1995). A sand pile can be organized by a simple rule that sees the grains falling randomly on a grid and accumulating until they topple sideways once they have reached a certain height (Pruessner 2012). Once the grid starts filling up, avalanches of various sizes start cascading with an overall size-probability distribution which demonstrates the characteristic power law behavior observed in self-similar dynamical systems (Fig. 3).

Figure 3.

Figure 3.

A simple sand-pile experiment simulated on a 20 × 20 grid (left); each pixel represents a column where sand piles up and colors indicate the number of grains (blue = 0 to yellow = max number of grains = 4). Sand is dropped randomly on the grid, and it topples sideways, as well off the grid, when the set limit for the height of a sand column is reached. The right panel shows the total sand in the grid through time, as well as the size of the avalanches falling off it.

Many natural systems tend toward such power law distributions, as these distributions allow for maximal adaptivity and other functional benefits (Shew and Plenz 2013). As in the case of the sand pile model, any perturbation to the system can be passed across scales and dissipated quickly—this endows the system with stability (Bak 1987). Another example of this phenomenon is the energy cascade in turbulent hydrodynamic systems, in which kinetic energy is passed between eddies of decreasing size until it is distributed, at the smallest scales, to the viscosity of the liquid (Kolmogorov 1941).

One important requirement for self-organization is the abundance of available energy. In the same way that the sand pile needs a constant supply of grains to sustain itself in a critical state, the brain devotes up to 80% of its energy budget to maintaining its intrinsic activity (i.e., brain activity in the absence of an overt motor, perceptual or cognitive task), whereas evoked activity in the brain does not use more than 1% of its metabolic reservoir (Raichle 2015; Raichle and Mintun 2006). This feature has been dubbed the Brain’s Dark Energy (Raichle 2006).

Self-organization rests upon four basic requirements (NBonabeau and others 1999): (1) interactions between multiple constituent elements; (2) strong nonlinearity, possibly involving positive and negative feedback; (3) a dynamic behavior striking a balance between exploitation and exploration; and (4) availability of energy, which allows to overcome the natural tendency to increase entropy (e.g., disorder).

The above-mentioned requirements can be readily observed in the organization of cortical structures. As summarized in Turkheimer and others (2015), the brain (primarily the cortex) is made up of a multitude of elementary units, each of which consists of two neurons—a pyramidal excitatory unit coupled with a gamma aminobutyric acid (GABA) interneuron that, via a delayed negative feedback, transforms the firing of the unit into an oscillation in the gamma frequency (Borgers and Kopell 2003), the so-called pyramidal interneuronal network gamma (PING) network. These elementary units are weakly connected via further GABA interneuron projections into and across the individual neural columns (Helmstaedter and others 2009). Therefore, local oscillations in the gamma frequency are the binding force that generates rapid metastable dynamics, moving between synchronization and de-synchronization, which spread across the cortex. These elementary computational motifs have been observed across various brain regions, as well as across scales (Turkheimer and others 2015), reaching the long-range synchrony of neural activity (Agrawal and others 2019; Fagerholm and others 2020; Meshulam and others 2019). This supports the view of the brain as exhibiting SOC (Cabral and others 2011; Cabral and others 2014; Roberts and others 2019; Tognoli and Kelso 2014), with a distinctive fractal signature both in its structure and function (Expert and others 2011; Squarcina and others 2015; Turkheimer and others 2015). The fractal distribution of neural output, also known as the 1/f noise, can also be observed in the behavioral repertoire, from the motor regions to the higher cognitive areas, capturing the range between basic physiological responses and creative abilities in the human brain (Levitin and others 2012; Palva and others 2013). Conversely, evidence suggests that this scale-invariance in brain output may be lost during anesthesia-induced loss of consciousness (Scott and others 2014). Interestingly, self-organized systems may produce top-down control, or at least the illusion of it: for example, once started by bottom-up synchronization, the oscillation of large neuronal masses may take control of the activity of smaller primary cortices and generate those phenomena related to attention or vigilance (Turkheimer and others 2015); these will be shaped by environmental interactions and, ultimately, by evolution. This is different from all those proposals where top-down controls are somehow hard-wired in the brain (see Grossi [2017] for a general discussion on the topic). These ideas lead straight to another fundamental property of complex systems—that of emergence.

Emergence

The fourth property of complex systems is that they are characterized by emergent behaviors. Complex systems, that is, organisms or environments made up of very large numbers of elementary agents (e.g., ants in colonies, birds in a flock, or arrays of molecules or weather elements), can produce an impressive array of sophisticated behaviors that cannot readily be explained solely by the physical properties of their constituent components. As a matter of fact, emergence—although still lacking a precise definition—has come to be recognized as a key ingredient of any system studied under the umbrella of complexity science (Turkheimer and others 2019).

The concept of emergence is credited to the physiologist and philosopher George H. Lewes in his book Problems of Life and Mind (Lewes 1879). At the time, Lewes was concerned with reductionism, that is, why all science does ultimately reduce to physics. Victorian England in times of Lewes saw important advances in mechanics; however, chemical reactions (e.g., the production of water from oxygen and hydrogen) seemed to require explanations that were qualitatively different from those that contemporary physics could provide. Building on this idea, Lewes drew a difference between mechanical effects and chemical effects—which he called “emergents.” In his words,

The emergent is unlike its components in so far as these are incommensurable, and it cannot be reduced to their sum or their difference. (Lewes 1879, p. 413)

The notion of emergence was first brought into the neurosciences by Roger W Perry, the 1981 Nobel laureate for physiology and medicine, in the context of the mind-body problem. According to Perry, emergence

is the idea, in brief, that conscious phenomena as emergent functional properties of brain processing exert an active control role as causal detents in shaping the flow patterns of cerebral excitation. Once generated from neural events, the higher order mental patterns and programs have their own subjective qualities and progress, operate and interact by their own causal laws and principles which are different from and cannot be reduced to those of neurophysiology. (Sperry 1980, p. 201)

In other words, the behavior of the whole cannot be explained purely by focusing only on local spatiotemporal interactions between its elements in isolation—new theories are needed to explain these phenomena. Similarly, higher order cognitive function operates via laws that cannot be reduced to those of neurophysiology—instead they operate in the domain of psychology.

Despite their popularity, strong emergent approaches have been the source of numerous controversies. From a philosophical perspective, strong emergence has been accused of being logically inconsistent (Bedau 2002) and sustained on illegitimate metaphysics (Bedau 1997). Additionally, diverse concerns with respect to strongly emergent approaches in neuroscience have been raised, including arguments related to their plausibility in regard to empirical evidence (Turner 2019), or to potential limits of their falsifiability (Colombo and Wright 2018). Finally, the fact that consciousness is usually the only proposed example of strong emergence in the natural world (see, e.g., Chalmers 2006) presumes an exceptionalism of consciousness that could be hard to justify.

These difficulties are avoided by “weak emergent” approaches, which rely on computational models that start from biologically plausible elementary units and build higher order outputs, using tools and concepts borrowed by complexity science (Turkheimer and others 2019). These models are defined as “weakly” emergent, are not limited by ontological faults, have been proven to be scalable, and able to realistically simulate the hierarchies of brain output and have now reached a level of maturity that enables predictions in the clinical realm (Deco and Kringelbach 2014; Fagerholm and others 2015; Lord and others 2017; Proix and others 2017; Turkheimer and others 2015; Turner 2019; Zimmermann and others 2018).

More technically, weak emergence can be broadly categorized into computational irreducible and causal emergence approaches. The former, introduced by Bedau (1997) based on ideas from Wolfram (2002), posits that emergence relates to limitations that hold in practice but not in principle; specifically, emergent properties would be generated by elements at microscopic levels in such complicated ways that they cannot be derived via explanatory shortcuts, but only by exhaustive simulation. While theoretically attractive, these approaches currently lack ways of operationalizing their definitions, and hence their practical value is limited. In contrast, causal emergent approaches identify formal methods to assess the causal power exerted by macroscopic properties of a system. Causal emergence has been operationalized using Granger Causality (Seth 2010), Pearl’s do() calculus and effective information (Chicharro and Panzeri 2014), and most recently by Integrated Information Decomposition (Oizumi and others 2014).

The Tools

Having examined neural function in terms of the four main properties associated with complex systems, we now address the question of how the obtained insights can be operationalized through the following scheme.

From the Analysis of Variance to the Analysis of Covariance

Realizing the large number of degrees of freedom that characterize the brain, one suggested approach is to shift the focus from the analysis of variance to the analysis of covariance. In other words, inference should be driven by the way the elements of the systems interact, and less so on how much they vary individually. The statistical literature provides numerous multivariate methods to study data covariance, which have been already been adapted for use in neuroimaging. A first approach involves using standard linear methods such as principal component analysis (PCA; Friston and others 1993), independent component analysis (ICA; McKeown and others 1998), and partial least squares (McIntosh and others 1996). Although not aimed at complexity analysis per se, these tools can be used to explore patterns in the data, as well as to derive approximate measures of complexity. For example, the Morgera Information Complexity is a measure of the extent to which covariance is distributed across principal components (Morgera 1985).

Capturing Covariance through Networks

A further approach is to represent covariance as networks, which are a popular tool to capture and explore covariance structures (Sporns 2011a). Networks can be built on multiple metrics, ranging from standard correlation coefficients to mutual information or model-based indices that are generally better suited to capture nonlinear interactions (Song and others 2012). Nonlinear metrics are sometimes preferred, as they may avoid spurious patterns that can be caused by unjustified linear assumptions (Bright and Murphy 2015).

Exploring the Macro- and Meso-Scale of Brain Networks

Once the web of interactions has been modelled as vertices and edges, the topology of the structure can be further explored, both at the macro-scale level—using graph theoretical measures that summarize the network properties (Bullmore and Sporns 2009)—and at the meso-scale level where the topological stability of the structure can be further quantified (Expert and others 2019; Petri and others 2014).

Toward High-Order Statistical Properties via Multivariate Correlation

A third approach to analyze covariant structures is to explore multivariate correlations, which capture high-order statistical properties that go beyond pairwise phenomena (Baudot and others 2019; Mediano and others 2019; Rosas and others 2019; Tononi and others 1994). High-order interactions can also be used to build hypergraphs, which are networks were links can connect more than two nodes (Bretto 2013). Hypergraph analysis is an active area of ongoing research, which is opening novel avenues for neuroimaging research (see, e.g., Gatica and others 2020).

Designs: From Cross-Sectional to Longitudinal and Life-Trajectories

Importantly, the brain is a nonlinear and nonergodic system. The discussions above have clarified the importance of experimental designs that move beyond stationary cross-sectional data paradigms and instead attempt, whenever practical, the acquisition of individual trajectories. This also enables the analysis of initial conditions; we have already shown how the artificial splitting of life-trajectories in development and ageing may result in faulted inferences. The consideration of trajectories is also relevant at smaller timescales, including the duration of a BOLD MRI study where, for example, higher resting-state functional activity before a stimulus can result in less activation or more deactivation (He 2013).

Using Dynamics to Extract Brain States

It is often useful to characterize dynamics in terms of “states” or attractors, that is, regional patterns of activity between which the system oscillates (Kelso 1995). The number of these states and their stability fully define such dynamics and can be quantified using standard analytics (Calhoun and others 2014; Gonzalez-Castillo and Bandettini 2018; Gu and others 2017; Hutchison and others 2013; Lord and others 2019; see Fig. 4), which have been used successfully in recent work on a broad spectrum of psychiatric disorders (de Lacy and Calhoun 2019; de Lacy and others 2017; Jin and others 2017; Kaiser and others 2016; Lurie and others 2020; Rabany and others 2019; Sakoglu and others 2010).

Figure 4.

Figure 4.

A brain state consists in a set of regions with synchronized activity during a time interval. The three-dimensional plot represents the energy levels of each state and the red line with arrow on the energy landscape the dynamic trajectory of the brain moving from state to state, whereas “energy” here is a term derived from information criteria and not calculated from metabolic expenditure. Brain states are more or less likely to be visited by the dynamics over time based on their relative energy/stability (adapted from Gu and others 2018).

From Euclidean to Fractal Geometry

Furthermore, if one accepts SOC as a model for the brain, then the use of fractal geometry can be justified in studying brain structural data (Expert and others 2011; Squarcina and others 2015) and its deviations from the SOC model. Fractal principles enable parsimonious methods to explain rich structures, holding great potential to explain brain structure and function. SOC approaches, however, face a number of important challenges (Beggs and Timme 2012), including the analysis of applicability of principles developed to study homogeneous systems (made, e.g., by equivalent spins) over heterogeneous systems such as the brain. Also, there is a need to operationalize the beneficial properties of criticality, such as efficiency and robustness, so that they can render verifiable predictions. It must be clarified that these challenges do not put in question the core contribution of SOC into neuroscience, namely, that by neglecting the right details, neuroscience might be capable of finding general underlying principles (Chialvo 2018; Cofré and others 2019).

Diversity and Emergence

Finally, the study of complex systems requires the quantification of two key properties: diversity and emergence. The first one can be described by a measure of entropy, for example, Shannon’s entropy (e.g., Σipilogpi where pi is the probability of state i), or variations of it including the excess entropy (Crutchfield and Feldman 2003; Grassberger 1986) and the entropy rate—this being closely related to the Lempel-Ziv complexity (Schartner and others 2017) and the statistical complexity of epsilon machines (Muñoz and others 2020; Shalizi and Crutchfield 2001). The second one, emergence, is a unique attribution of complex systems; the availability of one such measure would be a key tool in understanding the information flows between the various hierarchies/layers of neural models. So far, attempts to operationalize the various models of emergence have had limited applicability (Hoel 2017; Hoel and Klein 2020; Hoel and others 2013; Seth 2010). This work is, however, fundamental to the progress of complexity science in general and its application to the neurosciences in particular, and novel modelling approaches and effective measures are appearing in the recent literature (Rosas and others 2020). These quantitative approaches allow researchers to frame hypotheses about emergence in a formal, rigorous manner, and test these hypotheses on fine-grained data-driven alternatives to traditional all-or-none classifications, opening a broad range of possibilities for applications.

Conclusions

We reviewed neural systems from the perspective of complexity science and matched the required properties of complex systems with evidence from contemporary neuroscience. The use of complexity science is not new to studies of the brain. However, its use has often been intertwined with standard analytical approaches that stem from older conceptual frameworks. What is now required is the acceptance of these new frameworks both in terms of experimentation and analysis. Complexity science is itself experiencing ongoing progress, and some of the required tools are perhaps yet to be discovered. Nonetheless, the potential insights to be gained by this paradigm are sufficiently promising to render this a very worthwhile endeavor.

Acknowledgments

We would like to thank the three anonymous reviewers for their critical appraisal and very helpful comments.

1.

Also sometimes referred to as hierarchical organization, see discussion in Simon (1962).

2.

The example is taken from the Complexity Explorer course of the Santa Fe Institute (URL: https://www.complexityexplorer.org).

Footnotes

Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Funding: The author(s) received no financial support for the research, authorship, and/or publication of this article.

ORCID iD: Federico E. Turkheimer Inline graphic https://orcid.org/0000-0002-3766-3815

References

  1. Agrawal V, Chakraborty S, Knopfel T, Shew WL. 2019. Scale-change symmetry in the rules governing neural systems. iScience 12:121–31. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Amad A, Expert P, Lord LD, Fovet T, Geoffroy PA. 2020. Plastic adaptation to pathology in psychiatry: are patients with psychiatric disorders pathological experts? Neuroscientist 26(3):208–23. [DOI] [PubMed] [Google Scholar]
  3. Ashby WR. 1947. Principles of the self-organizing dynamic system. J Gen Psychol 37(2):125–8. [DOI] [PubMed] [Google Scholar]
  4. Bak P. 1987. Self-organized criticality: an explanation of 1/f noise. Phys Rev Lett 59(4):381–4. [DOI] [PubMed] [Google Scholar]
  5. Bak P. 1996. How nature works: the science of self-organized criticality. Copernicus. [Google Scholar]
  6. Bak P, Paczuski M. 1995. Complexity, contingency, and criticality. Proc Natl Acad Sci U S A 92(15):6689–96. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Balasubramanian V, Kimber D, Berry MJ, 2nd. 2001. Metabolically efficient information processing. Neural Comput 13(4):799–815. [DOI] [PubMed] [Google Scholar]
  8. Bassett DS, Gazzaniga MS. 2011. Understanding complexity in the human brain. Trends Cogn Sci 15(5):200–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Baudot P, Tapia M, Bennequin D, Goaillard J. 2019. Topological information data analysis. Entropy 21(9):869. [Google Scholar]
  10. Bedau M. 1997. Weak emergence. Noûs 31:375–99. [Google Scholar]
  11. Bedau M. 2002. Downward causation and the autonomy of weak emergence. Principia 6(1):5–50. [Google Scholar]
  12. Beggs JM, Timme N. 2012. Being critical of criticality in the brain. Front Physiol 3:163. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Bohr T. 1998. Dynamical systems approach to turbulence. Cambridge University Press. [Google Scholar]
  14. Borgers C, Kopell N. 2003. Synchronization in networks of excitatory and inhibitory neurons with sparse, random connectivity. Neural Comput 15(3):509–38. [DOI] [PubMed] [Google Scholar]
  15. Boyle EA, Li YI, Pritchard JK. 2017. An expanded view of complex traits: from polygenic to omnigenic. Cell 169(7):1177–86. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Brauchli C, Leipold S, Jancke L. 2019. Univariate and multivariate analyses of functional networks in absolute pitch. Neuroimage 189:241–7. [DOI] [PubMed] [Google Scholar]
  17. Bretto A. 2013. Hypergraph theory. An introduction: Springer. [Google Scholar]
  18. Bright MG, Murphy K. 2015. Is fMRI “noise” really noise? Resting state nuisance regressors remove variance with network structure. Neuroimage 114:158–69. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Bullmore E, Barnes A, Bassett DS, Fornito A, Kitzbichler M, Meunier D, and others. 2009. Generic aspects of complexity in brain imaging data and other biological systems. Neuroimage 47(3):1125–34. [DOI] [PubMed] [Google Scholar]
  20. Bullmore E, Sporns O. 2009. Complex brain networks: graph theoretical analysis of structural and functional systems. Nat Rev Neurosci 10(3):186–98. [DOI] [PubMed] [Google Scholar]
  21. Cabral J, Hugues E, Sporns O, Deco G. 2011. Role of local network oscillations in resting-state functional connectivity. Neuroimage 57(1):130–9. [DOI] [PubMed] [Google Scholar]
  22. Cabral J, Luckhoo H, Woolrich M, Joensson M, Mohseni H, Baker A, and others. 2014. Exploring mechanisms of spontaneous functional connectivity in MEG: how delayed network interactions lead to structured amplitude envelopes of band-pass filtered oscillations. Neuroimage 90:423–35. [DOI] [PubMed] [Google Scholar]
  23. Calhoun VD, Miller R, Pearlson G, Adali T. 2014. The chronnectome: time-varying connectivity networks as the next frontier in fMRI data discovery. Neuron 84(2):262–74. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Chalmers DJ. 2006. Strong and weak emergence. In: Davies PCaP, editor. The re-emergence of emergence: Oxford University Press. p. 244–56. [Google Scholar]
  25. Chen BL, Hall DH, Chklovskii DB. 2006. Wiring optimization can relate neuronal structure and function. Proc Natl Acad Sci U S A 103(12):4723–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Chialvo DR. 2018. Life at the edge: complexity and criticality in biological function. arXiv. https://arxiv.org/abs/1810.11737
  27. Chicharro D, Panzeri S. 2014. Algorithms of causal inference for the analysis of effective connectivity among brain regions. Front Neuroinform 8:64. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Chu C, Hsu AL, Chou KH, Bandettini P, Lin C. 2012. Does feature selection improve classification accuracy? Impact of sample size and feature selection on classification using anatomical magnetic resonance images. Neuroimage 60(1): 59–70. [DOI] [PubMed] [Google Scholar]
  29. Clarke DD, Sokoloff L. 1999. Regulation of cerebral metabolic rate. In: Siegel GJ, Agranoff BW, Albers RW, editors. Basic neurochemistry: molecular, cellular and medical aspects. 6th ed. Lippincott-Raven. [Google Scholar]
  30. Cofré R, Herzog R, Corcoran D, Rosas FE. 2019. A comparison of the maximum entropy principle across biological spatial scales. Entropy 21(10):1009. [Google Scholar]
  31. Colombo M, Wright C. 2018. First principles in the life sciences: the free-energy principle, organicism, and mechanism. Synthese. 10.1007/s11229-018-01932-w [DOI]
  32. Crutchfield JP, Feldman DP. 2003. Regularities unseen, randomness observed: levels of entropy convergence. Chaos 13(1):25–54. [DOI] [PubMed] [Google Scholar]
  33. de Lacy N, Calhoun VD. 2019. Dynamic connectivity and the effects of maturation in youth with attention deficit hyperactivity disorder. Netw Neurosci 3(1):195–216. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. de Lacy N, Doherty D, King BH, Rachakonda S, Calhoun VD. 2017. Disruption to control network function correlates with altered dynamic connectivity in the wider autism spectrum. Neuroimage Clin 15:513–24. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Deco G, Jirsa VK, McIntosh AR. 2011. Emerging concepts for the dynamical organization of resting-state activity in the brain. Nat Rev Neurosci 12(1):43–56. [DOI] [PubMed] [Google Scholar]
  36. Deco G, Kringelbach ML. 2014. Great expectations: using whole-brain computational connectomics for understanding neuropsychiatric disorders. Neuron 84(5):892–905. [DOI] [PubMed] [Google Scholar]
  37. Deco G, Kringelbach ML. 2020. Turbulence in the human brain: Discovering the homogeneous isotropic functional core organisation of the human brain. Cell 33(10):108471. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. DiCarlo JJ, Zoccolan D, Rust NC. 2012. How does the brain solve visual object recognition? Neuron 73(3):415–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Donaldson DI. 2004. Parsing brain activity with fMRI and mixed designs: what kind of a state is neuroimaging in? Trends Neurosci 27(8):442–4. [DOI] [PubMed] [Google Scholar]
  40. Donoho DH, Johnstone IM. 1994. Minimax risk over lp-balls for lq-error. Probab Theory Relat Fields 99:277–303. [Google Scholar]
  41. Ecker C, Rocha-Rego V, Johnston P, Mourao-Miranda J, Marquand A, Daly EM, and others. 2010. Investigating the predictive value of whole-brain structural MR scans in autism: a pattern classification approach. Neuroimage 49(1):44–56. [DOI] [PubMed] [Google Scholar]
  42. Eklund A, Nichols TE, Knutsson H. 2016. Cluster failure: Why fMRI inferences for spatial extent have inflated false-positive rates. Proc Natl Acad Sci U S A 113(28):7900–5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Expert P, Lambiotte R, Chialvo DR, Christensen K, Jensen HJ, Sharp DJ, and others. 2011. Self-similar correlation function in brain resting-state functional magnetic resonance imaging. J R Soc Interface 8(57):472–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Expert P, Lord LD, Kringelbach ML, Petri G. 2019. Editorial: Topological neuroscience. Netw Neurosci 3(3):653–55. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Fagerholm ED, Foulkes WMC, Gallero-Salas Y, Helmchen F, Friston KJ, Moran RJ, and others. 2020. Conservation laws by virtue of scale symmetries in neural systems. PLoS Comput Biol 16(5):e1007865. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Fagerholm ED, Hellyer PJ, Scott G, Leech R, Sharp DJ. 2015. Disconnection of network hubs and cognitive impairment after traumatic brain injury. Brain 138(Pt 6):1696–709. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Fisher RA. 1918. The correlation between relatives on the supposition of Mendelian inheritance. Trans R Soc Edinb 52:399–433. [Google Scholar]
  48. Fjell AM, Westlye LT, Grydeland H, Amlien I, Espeseth T, Reinvang I, and others. 2013. Critical ages in the life course of the adult brain: nonlinear subcortical aging. Neurobiol Aging 34(10):2239–47. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Fonseca-Azevedo K, Herculano-Houzel S. 2012. Metabolic constraint imposes tradeoff between body size and number of brain neurons in human evolution. Proc Natl Acad Sci U S A 109(45):18571–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Fox NC, Schott JM. 2004. Imaging cerebral atrophy: normal ageing to Alzheimer’s disease. Lancet 363(9406):392–4. [DOI] [PubMed] [Google Scholar]
  51. Friston K. 2009. The free-energy principle: a rough guide to the brain? Trends Cogn Sci 13(7):293–301. [DOI] [PubMed] [Google Scholar]
  52. Friston KJ, Frith CD, Liddle PF, Frackowiak RS. 1991. Comparing functional (PET) images: the assessment of significant change. J Cereb Blood Flow Metab 11(4):690–9. [DOI] [PubMed] [Google Scholar]
  53. Friston KJ, Frith CD, Liddle PF, Frackowiak RS. 1993. Functional connectivity: the principal-component analysis of large (PET) data sets. J Cereb Blood Flow Metab 13(1):5–14. [DOI] [PubMed] [Google Scholar]
  54. Gate D, Saligrama N, Leventhal O, Yang AC, Unger MS, Middeldorp J, and others. 2020. Clonally expanded CD8 T cells patrol the cerebrospinal fluid in Alzheimer’s disease. Nature 577(7790):399–404. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Gatica M, Cofré R, Mediano PAM, Rosas FE, Orio P, Diez I, and others. 2020. High-order interdependencies in the aging brain. bioRxiv. 10.1101/2020.03.17.995886 [DOI] [PubMed]
  56. Goldstein DB. 2009. Common genetic variation and human traits. N Engl J Med 360(17):1696–8. [DOI] [PubMed] [Google Scholar]
  57. Gonzalez-Castillo J, Bandettini PA. 2018. Task-based dynamic functional connectivity: recent findings and open questions. Neuroimage 180(Pt B):526–33. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Gonzalez-Castillo J, Hoy CW, Handwerker DA, Roopchansingh V, Inati SJ, Saad ZS, and others. 2015. Task dependence, tissue specificity, and spatial distribution of widespread activations in large single-subject functional MRI datasets at 7T. Cereb Cortex 25(12):4667–77. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Gonzalez-Castillo J, Saad ZS, Handwerker DA, Inati SJ, Brenowitz N, Bandettini PA. 2012. Whole-brain, time-locked activation with simple tasks revealed using massive averaging and model-free analysis. Proc Natl Acad Sci U S A 109(14):5487–92. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Good CD, Johnsrude IS, Ashburner J, Henson RN, Friston KJ, Frackowiak RS. 2001. A voxel-based morphometric study of ageing in 465 normal adult human brains. Neuroimage 14(1 Pt 1):21–36. [DOI] [PubMed] [Google Scholar]
  61. Grassberger P. 1986. Toward a quantitative theory of self-generated complexity. Int J Theor Phys 25(9):907–38. [Google Scholar]
  62. Greicius M. 2008. Resting-state functional connectivity in neuropsychiatric disorders. Curr Opin Neurol 21(4):424–30. [DOI] [PubMed] [Google Scholar]
  63. Griffith JS. 1963. On the stability of brain-like structures. Biophys J 3:299–308. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Grossi G. 2017. Hardwiring: innateness in the age of the brain. Biol Philos 32:1047–82. [Google Scholar]
  65. Gu S, Betzel RF, Mattar MG, Cieslak M, Delio PR, Grafton ST, and others. 2017. Optimal trajectories of brain state transitions. Neuroimage 148:305–17. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Gu S, Cieslak M, Baird B, Muldoon SF, Grafton ST, Pasqualetti F, and others. 2018. The energy landscape of neurophysiological activity implicit in brain network structure. Sci Rep 8(1):2507. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Ha S, Sohn IJ, Kim N, Sim HJ, Cheon KA. 2015. Characteristics of brains in autism spectrum disorder: structure, function and connectivity across the lifespan. Exp Neurobiol 24(4):273–84. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Haxby JV, Gobbini MI, Furey ML, Ishai A, Schouten JL, Pietrini P. 2001. Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science 293(5539):2425–30. [DOI] [PubMed] [Google Scholar]
  69. He BJ. 2013. Spontaneous and task-evoked brain activity negatively interact. J Neurosci 33(11):4672–82. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. He Y, Byrge L, Kennedy DP. 2020. Nonreplication of functional connectivity differences in autism spectrum disorder across multiple sites and denoising strategies. Hum Brain Mapp 41(5):1334–50. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Hebart MN, Baker CI. 2018. Deconstructing multivariate decoding for the study of brain function. Neuroimage 180(Pt A):4–18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Helmstaedter M, Sakmann B, Feldmeyer D. 2009. Neuronal correlates of local, lateral, and translaminar inhibition with reference to cortical columns. Cereb Cortex 19(4):926–37. [DOI] [PubMed] [Google Scholar]
  73. Herculano-Houzel S. 2011. Scaling of brain metabolism with a fixed energy budget per neuron: implications for neuronal activity, plasticity and evolution. PLoS One 6(3):e17514. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Hoel EP. 2017. When the map is better than the territory. Entropy 19(5):188. [Google Scholar]
  75. Hoel EP, Albantakis L, Tononi G. 2013. Quantifying causal emergence shows that macro can beat micro. Proc Natl Acad Sci U S A 110(49):19790–5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Hoel EP, Klein B. 2020. The emergence of informative higher scales in complex networks. Complexity 2020:8932526. [Google Scholar]
  77. Hutchison RM, Womelsdorf T, Allen EA, Bandettini PA, Calhoun VD, Corbetta M, and others. 2013. Dynamic functional connectivity: promise, issues, and interpretations. Neuroimage 80:360–78. [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Jensen HJ. 1998. Self-organized criticality: emergent complex behavior in physical and biological systems. Cambridge University Press. [Google Scholar]
  79. Jin C, Jia H, Lanka P, Rangaprakash D, Li L, Liu T, and others. 2017. Dynamic brain connectivity is a better predictor of PTSD than static connectivity. Hum Brain Mapp 38(9):4479–96. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Kaiser RH, Whitfield-Gabrieli S, Dillon DG, Goer F, Beltzer M, Minkel J, and others. 2016. Dynamic resting-state functional connectivity in major depression. Neuropsychopharmacology 41(7):1822–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Kelso JAS. 1995. Dynamic patterns: the self-organization of brain and behavior. MIT Press. [Google Scholar]
  82. Kim SK. 2007. Common aging pathways in worms, flies, mice and humans. J Exp Biol 210(Pt 9):1607–12. [DOI] [PubMed] [Google Scholar]
  83. Kolmogorov AN. 1941. The local structure of isotropic turbulence in an incompressible viscous fluid. Dokl Akad Nauk SSSR 30:301–5. [Google Scholar]
  84. Labzin LI, Heneka MT, Latz E. 2018. Innate immunity and neurodegeneration. Annu Rev Med 69:437–49. [DOI] [PubMed] [Google Scholar]
  85. Ladyman J, Lambert J, Wiesner K. 2013. What is a complex system? Euro J Philos Sci 3:33–67. [Google Scholar]
  86. Lange N, Travers BG, Bigler ED, Prigge MB, Froehlich AL, Nielsen JA, and others. 2015. Longitudinal volumetric brain changes in autism spectrum disorder ages 6–35 years. Autism Res 8(1):82–93. [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Lenroot RK, Yeung PK. 2013. Heterogeneity within autism spectrum disorders: what have we learned from neuroimaging studies? Front Hum Neurosci 7:733. [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Levitin DJ, Chordia P, Menon V. 2012. Musical rhythm spectra from Bach to Joplin obey a 1/f power law. Proc Natl Acad Sci U S A 109(10):3716–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Levy WB, Baxter RA. 1996. Energy efficient neural codes. Neural Comput 8(3):531–43. [DOI] [PubMed] [Google Scholar]
  90. Lewes GH. 1879. Problems of life and mind. Osgood and company. [Google Scholar]
  91. Liberg B, Rahm C, Panayiotou A, Pantelis C. 2016. Brain change trajectories that differentiate the major psychoses. Eur J Clin Invest 46(7):658–74. [DOI] [PubMed] [Google Scholar]
  92. Lieberman MD, Cunningham WA. 2009. Type I and type II error concerns in fMRI research: re-balancing the scale. Soc Cogn Affect Neurosci 4(4):423–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  93. Loggia ML, Edwards RR, Kim J, Vangel MG, Wasan AD, Gollub RL, and others. 2012. Disentangling linear and nonlinear brain responses to evoked deep tissue pain. Pain 153(10):2140–51. [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Loh PR, Bhatia G, Gusev A, Finucane HK, Bulik-Sullivan BK, Pollack SJ, and others. 2015. Contrasting genetic architectures of schizophrenia and other complex diseases using fast variance-components analysis. Nat Genet 47(12):1385–92. [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. Lord LD, Allen P, Expert P, Howes O, Broome M, Lambiotte R, and others. 2012. Functional brain networks before the onset of psychosis: a prospective fMRI study with graph theoretical analysis. Neuroimage Clin 1(1):91–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Lord LD, Expert P, Atasoy S, Roseman L, Rapuano K, Lambiotte R, and others. 2019. Dynamical exploration of the repertoire of brain networks at rest is modulated by psilocybin. Neuroimage 199:127–42. [DOI] [PubMed] [Google Scholar]
  97. Lord LD, Stevner AB, Deco G, Kringelbach ML. 2017. Understanding principles of integration and segregation using whole-brain computational connectomics: implications for neuropsychiatric disorders. Philos Trans A Math Phys Eng Sci 375(2096):20160283. [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Lotspeich LJ, Kwon H, Schumann CM, Fryer SL, Goodlin-Jones BL, Buonocore MH, and others. 2004. Investigation of neuroanatomical differences between autism and Asperger syndrome. Arch Gen Psychiatry 61(3):291–8. [DOI] [PubMed] [Google Scholar]
  99. Lurie DJ, Kessler D, Bassett DS, Betzel RF, Breakspear M, Kheilholz S, and others. 2020. Questions and controversies in the study of time-varying functional connectivity in resting fMRI. Netw Neurosci 4(1):30–69. [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. McIntosh AR, Bookstein FL, Haxby JV, Grady CL. 1996. Spatial pattern analysis of functional brain images using partial least squares. Neuroimage 3(3 Pt 1):143–57. [DOI] [PubMed] [Google Scholar]
  101. McKeown MJ, Makeig S, Brown GG, Jung TP, Kindermann SS, Bell AJ, and others. 1998. Analysis of fMRI data by blind separation into independent spatial components. Hum Brain Mapp 6(3):160–88. [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Mediano PAM, Rosas FE, Carhart-Harris R, Seth AK, Barrett AB. 2019. Beyond integrated information: a taxonomy of information dynamics phenomena. arXiv. https://arxiv.org/abs/1909.02297
  103. Meltzer A, Van de Water J. 2017. The role of the immune system in autism spectrum disorder. Neuropsychopharmacology 42(1):284–98. [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Meshulam L, Gauthier JL, Brody CD, Tank DW, Bialek W. 2019. Coarse graining, fixed points, and scaling in a large population of neurons. Phys Rev Lett 123(17):178103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Meyer U. 2019. Neurodevelopmental resilience and susceptibility to maternal immune activation. Trends Neurosci 42(11):793–806. [DOI] [PubMed] [Google Scholar]
  106. Miller AH, Raison CL. 2016. The role of inflammation in depression: from evolutionary imperative to modern treatment target. Nat Rev Immunol 16(1):22–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  107. Morgera SD. 1985. Information theoretic covariance complexity and its relation to pattern recognition. IEEE Trans Syst Man Cybernet 15(5):608–19. [Google Scholar]
  108. Muñoz RN, Leung A, Zecevik A, Pollock FA, Cohen D, van Swinderen B, and others. 2020. General anesthesia reduces complexity and temporal asymmetry of the informational structures derived from neural recordings in Drosophila. Phys Rev Res 2(2):023219. [Google Scholar]
  109. NBonabeau E, Dorigo M, Theraulaz G. 1999. Swarm intelligence: from natural to artificial systems. Oxford University Press. [Google Scholar]
  110. Nyberg L, Salami A, Andersson M, Eriksson J, Kalpouzos G, Kauppi K, and others. 2010. Longitudinal evidence for diminished frontal cortex function in aging. Proc Natl Acad Sci U S A 107(52):22682–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Oizumi M, Albantakis L, Tononi G. 2014. From the phenomenology to the mechanisms of consciousness: integrated information theory 3.0. PLoS Comput Biol 10(5):e1003588. [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Palva JM, Zhigalov A, Hirvonen J, Korhonen O, Linkenkaer-Hansen K, Palva S. 2013. Neuronal long-range temporal correlations and avalanche dynamics are correlated with behavioral scaling laws. Proc Natl Acad Sci U S A 110(9):3585–90. [DOI] [PMC free article] [PubMed] [Google Scholar]
  113. Petri G, Expert P, Turkheimer F, Carhart-Harris R, Nutt D, Hellyer PJ, and others. 2014. Homological scaffolds of brain functional networks. J R Soc Interface 11(101):20140873. [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Pillet I, Op de Beeck H, Lee Masson H. 2019. A comparison of functional networks derived from representational similarity, functional connectivity, and univariate analyses. Front Neurosci 13:1348. [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Pinho AL, Amadon A, Ruest T, Fabre M, Dohmatob E, Denghien I, and others. 2018. Individual brain charting, a high-resolution fMRI dataset for cognitive mapping. Sci Data 5:180105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  116. Proix T, Bartolomei F, Guye M, Jirsa VK. 2017. Individual brain structure and modelling predict seizure propagation. Brain 140(3):641–54. [DOI] [PMC free article] [PubMed] [Google Scholar]
  117. Pruessner G. 2012. Self-organised criticality; theory, models and characterisation. Cambridge University Press. [Google Scholar]
  118. Rabany L, Brocke S, Calhoun VD, Pittman B, Corbera S, Wexler BE, and others. 2019. Dynamic functional connectivity in schizophrenia and autism spectrum disorder: convergence, divergence and classification. Neuroimage Clin 24:101966. [DOI] [PMC free article] [PubMed] [Google Scholar]
  119. Raichle ME. 2006. Neuroscience. The brain’s dark energy. Science 314(5803):1249–50. [PubMed] [Google Scholar]
  120. Raichle ME. 2015. The restless brain: how intrinsic activity organizes brain function. Philos Trans R Soc Lond B Biol Sci 370(1668):20140172. [DOI] [PMC free article] [PubMed] [Google Scholar]
  121. Raichle ME, Mintun MA. 2006. Brain work and brain imaging. Annu Rev Neurosci 29:449–76. [DOI] [PubMed] [Google Scholar]
  122. Richetto J, Chesters R, Cattaneo A, Labouesse MA, Gutierrez AMC, Wood TC, and others. 2017. Genome-wide transcriptional profiling and structural magnetic resonance imaging in the maternal immune activation model of neurodevelopmental disorders. Cereb Cortex 27(6):3397–413. [DOI] [PubMed] [Google Scholar]
  123. Roberts JA, Gollo LL, Abeysuriya RG, Roberts G, Mitchell PB, Woolrich MW, and others. 2019. Metastable brain waves. Nat Commun 10(1):1056. [DOI] [PMC free article] [PubMed] [Google Scholar]
  124. Rosas FE, Mediano PAM, Gastpar M, Jensen HJ. 2019. Quantifying high-order interdependencies via multivariate extensions of the mutual information. Phys Rev E 100(3–1):032305. [DOI] [PubMed] [Google Scholar]
  125. Rosas FE, Mediano PAM, Jensen HJ, Seth AK, Barrett AB, Carhart-Harris R, and others. 2020. Reconciling emergences: an information-theoretic approach to identify causal emergence in multivariate data. arXiv>q-bio. https://arxiv.org/abs/2004.08220 [DOI] [PMC free article] [PubMed]
  126. Rosas FE, Mediano PAM, Ugarte M, Jensen HJ. 2018. An information-theoretic approach to self-organisation: emergence of complex interdependencies in coupled dynamical systems. Entropy 20(10):793. [DOI] [PMC free article] [PubMed] [Google Scholar]
  127. Sakoglu U, Pearlson GD, Kiehl KA, Wang YM, Michael AM, Calhoun VD. 2010. A method for evaluating dynamic functional network connectivity and task-modulation: application to schizophrenia. MAGMA 23(5–6):351–66. [DOI] [PMC free article] [PubMed] [Google Scholar]
  128. Salthouse TA. 2019. Trajectories of normal cognitive aging. Psychol Aging 34(1):17–24. [DOI] [PMC free article] [PubMed] [Google Scholar]
  129. Schartner MM, Carhart-Harris RL, Barrett AB, Seth AK, Muthukumaraswamy SD. 2017. Increased spontaneous MEG signal diversity for psychoactive doses of ketamine, LSD and psilocybin. Sci Rep 7:46421. [DOI] [PMC free article] [PubMed] [Google Scholar]
  130. Scott G, Fagerholm ED, Mutoh H, Leech R, Sharp DJ, Shew WL, and others. 2014. Voltage imaging of waking mouse cortex reveals emergence of critical neuronal dynamics. J Neurosci 34(50):16611–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  131. Sella G, Barton NH. 2019. Thinking about the evolution of complex traits in the era of genome-wide association studies. Annu Rev Genomics Hum Genet 20:461–93. [DOI] [PubMed] [Google Scholar]
  132. Seth AK. 2010. Measuring autonomy and emergence via Granger causality. Artif Life 16(2):179–96. [DOI] [PubMed] [Google Scholar]
  133. Shalizi CR, Crutchfield JP. 2001. Computational mechanics: pattern and prediction, structure and simplicity. J Stat Phys 104(3–4):817–79. [Google Scholar]
  134. Shew WL, Plenz D. 2013. The functional benefits of criticality in the cortex. Neuroscientist 19(1):88–100. [DOI] [PubMed] [Google Scholar]
  135. Shi H, Kichaev G, Pasaniuc B. 2016. Contrasting the genetic architecture of 30 complex traits from summary association data. Am J Hum Genet 99(1):139–53. [DOI] [PMC free article] [PubMed] [Google Scholar]
  136. Shmookler Reis RJ, Bharill P, Tazearslan C, Ayyadevara S. 2009. Extreme-longevity mutations orchestrate silencing of multiple signaling pathways. Biochim Biophys Acta 1790(10):1075–83. [DOI] [PMC free article] [PubMed] [Google Scholar]
  137. Sibson NR, Dhankhar A, Mason GF, Rothman DL, Behar KL, Shulman RG. 1998. Stoichiometric coupling of brain glucose metabolism and glutamatergic neuronal activity. Proc Natl Acad Sci U S A 95(1):316–21. [DOI] [PMC free article] [PubMed] [Google Scholar]
  138. Simon HA. 1962. The architecture of complexity. Proc Am Phil Soc 106:467–82. [Google Scholar]
  139. Singer W. 1986. The brain as a self-organizing system. Eur Arch Psychiatry Neurol Sci 236(1):4–9. [DOI] [PubMed] [Google Scholar]
  140. Smith SM, Elliott LT, Alfaro-Almagro F, McCarthy P, Nichols TE, Douaud G, and others. 2020. Brain aging comprises many modes of structural and functional change with distinct genetic and biophysical associations. Elife 9:252677. [DOI] [PMC free article] [PubMed] [Google Scholar]
  141. Smoller JW, Andreassen OA, Edenberg HJ, Faraone SV, Glatt SJ, Kendler KS. 2019. Psychiatric genetics and the structure of psychopathology. Mol Psychiatry 24(3):409–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  142. Snowdon DA, Nun S. 2003. Healthy aging and dementia: findings from the Nun Study. Ann Intern Med 139(5 Pt 2):450–4. [DOI] [PubMed] [Google Scholar]
  143. Song L, Langfelder P, Horvath S. 2012. Comparison of co-expression measures: mutual information, correlation, and model based indices. BMC Bioinformatics 13:328. [DOI] [PMC free article] [PubMed] [Google Scholar]
  144. Sperry RW. 1980. Mind-brain interaction: mentalism, yes; dualism, no. Neuroscience 5(2):195–206. [DOI] [PubMed] [Google Scholar]
  145. Sporns O. 2011. a. Networks of the brain. MIT Press. [Google Scholar]
  146. Sporns O. 2011. b. The non-random brain: efficiency, economy, and complex dynamics. Front Comput Neurosci 5:5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  147. Squarcina L, De Luca A, Bellani M, Brambilla P, Turkheimer FE, Bertoldo A. 2015. Fractal analysis of MRI data for the characterization of patients with schizophrenia and bipolar disorder. Phys Med Biol 60(4):1697–716. [DOI] [PubMed] [Google Scholar]
  148. Steinmetz NA, Zatka-Haas P, Carandini M, Harris KD. 2019. Distributed coding of choice, action and engagement across the mouse brain. Nature 576(7786):266–73. [DOI] [PMC free article] [PubMed] [Google Scholar]
  149. Telesford QK, Simpson SL, Burdette JH, Hayasaka S, Laurienti PJ. 2011. The brain as a complex system: using network science as a tool for understanding the brain. Brain Connect 1(4):295–308. [DOI] [PMC free article] [PubMed] [Google Scholar]
  150. Tetzlaff C, Okujeni S, Egert U, Worgotter F, Butz M. 2010. Self-organized criticality in developing neuronal networks. PLoS Comput Biol 6(12):e1001013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  151. Thompson PM, Vidal C, Giedd JN, Gochman P, Blumenthal J, Nicolson R, and others. 2001. Mapping adolescent brain change reveals dynamic wave of accelerated gray matter loss in very early-onset schizophrenia. Proc Natl Acad Sci U S A 98(20):11650–5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  152. Tognoli E, Kelso JA. 2014. The metastable brain. Neuron 81(1):35–48. [DOI] [PMC free article] [PubMed] [Google Scholar]
  153. Tononi G, Sporns O, Edelman GM. 1994. A measure for brain complexity: relating functional segregation and integration in the nervous system. Proc Natl Acad Sci U S A 91(11):5033–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  154. Turkheimer FE, Aston JA, Banati RB, Riddell C, Cunningham VJ. 2003. A linear wavelet filter for parametric imaging with dynamic PET. IEEE Trans Med Imaging 22(3):289–301. [DOI] [PubMed] [Google Scholar]
  155. Turkheimer FE, Hellyer P, Kehagia AA, Expert P, Lord LD, Vohryzek J, and others. 2019. Conflicting emergences. Weak vs. strong emergence for the modelling of brain function. Neurosci Biobehav Rev 99:3–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  156. Turkheimer FE, Leech R, Expert P, Lord LD, Vernon AC. 2015. The brain’s code and its canonical computational motifs. From sensory cortex to the default mode network: a multi-scale model of brain function in health and disease. Neurosci Biobehav Rev 55:211–22. [DOI] [PubMed] [Google Scholar]
  157. Turkheimer FE, Selvaggi P, Mehta MA, Veronese M, Zelaya F, Dazzan P, and others. 2020. Normalizing the abnormal: do antipsychotic drugs push the cortex into an unsustainable metabolic envelope? Schizophr Bull 46(3):484–95. [DOI] [PMC free article] [PubMed] [Google Scholar]
  158. Turner R. 2019. Myelin and modeling: bootstrapping cortical microcircuits. Front Neural Circuits 13:34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  159. Uttal WR. 2001. The new phrenology: the limits of localizing cognitive processes in the brain. MIT Press. [Google Scholar]
  160. Verhulst PF. 1838. Notice sur la loi que la population poursuit dans son accroisse- ment. Corresp Math Phys 10:113–21. [Google Scholar]
  161. Vinke EJ, de Groot M, Venkatraghavan V, Klein S, Niessen WJ, Ikram MA, and others. 2018. Trajectories of imaging markers in brain aging: the Rotterdam Study. Neurobiol Aging 71:32–40. [DOI] [PubMed] [Google Scholar]
  162. Wen Q, Chklovskii DB. 2005. Segregation of the brain into gray and white matter: a design minimizing conduction delays. PLoS Comput Biol 1(7):e78. [DOI] [PMC free article] [PubMed] [Google Scholar]
  163. Wen Q, Chklovskii DB. 2008. A cost-benefit analysis of neuronal morphology. J Neurophysiol 99(5):2320–8. [DOI] [PubMed] [Google Scholar]
  164. Williams PE, Shapley RM. 2007. A dynamic nonlinearity and spatial phase specificity in macaque V1 neurons. J Neurosci 27(21):5706–18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  165. Wolff JJ, Jacob S, Elison JT. 2018. The journey to autism: Insights from neuroimaging studies of infants and toddlers. Dev Psychopathol 30(2):479–95. [DOI] [PMC free article] [PubMed] [Google Scholar]
  166. Wolfram S. 2002. A new kind of science. Wolfram media. [Google Scholar]
  167. Worsley KJ, Marrett S, Neelin P, Vandal AC, Friston KJ, Evans AC. 1996. A unified statistical approach for determining significant signals in images of cerebral activation. Hum Brain Mapp 4(1):58–73. [DOI] [PubMed] [Google Scholar]
  168. Yang J, Benyamin B, McEvoy BP, Gordon S, Henders AK, Nyholt DR, and others. 2010. Common SNPs explain a large proportion of the heritability for human height. Nat Genet 42(7):565–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  169. Zhuo C, Li G, Lin X, Jiang D, Xu Y, Tian H, and others. 2019. The rise and fall of MRI studies in major depressive disorder. Transl Psychiatry 9(1):335. [DOI] [PMC free article] [PubMed] [Google Scholar]
  170. Zimmermann J, Perry A, Breakspear M, Schirner M, Sachdev P, Wen W, and others. 2018. Differentiation of Alzheimer’s disease based on local and global parameters in personalized virtual brain models. Neuroimage Clin 19:240–51. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from The Neuroscientist are provided here courtesy of SAGE Publications

RESOURCES