Abstract
Understanding neurocognitive computations will require not just localizing cognitive information distributed throughout the brain but also determining how that information got there. We review recent advances in linking empirical and simulated brain network organization with cognitive information processing. Building on these advances, we offer a new framework for understanding the role of connectivity in cognition – network coding (encoding/decoding) models. These models utilize connectivity to specify the transfer of information via neural activity flow processes, successfully predicting the formation of cognitive representations in empirical neural data. The success of these models supports the possibility that localized neural functions mechanistically emerge (are computed) from distributed activity flow processes that are specified primarily by connectivity patterns.
Keywords: connectivity, machine learning, representations, neural encoding/decoding, connectome, neural networks, artificial intelligence
Placing Brain Network Organization Within A Computational Framework
A central goal of neuroscience is to understand how neural entities (such as brain regions) interact to compute cognitive functions. Historically, cognitive neuroscientists have tried to understand neural systems by mapping cognitive processes to neural entities. However, a comprehensive functional map of neural entities would still be insufficient to explain how cognition emerges through the collective interaction among these components. What could facilitate such mechanistic inferences? To answer this question, cognitive neuroscientists have begun to incorporate descriptions of brain network architecture (i.e. how neural entities are connected to each other) with these functional maps. Indeed, recent theoretical work suggests that mechanistic understanding of a brain system depends on characterizing the causal relationships between the system’s components [1–3]. This is analogous to how we understand other mechanical systems; for example, understanding how a car moves upon pressing its accelerator requires knowledge of the causal relationships between the system’s components (e.g., the pedal, the transmission, the engine, and the wheels). Similarly, mapping brain network connectivity allows for a mechanistic understanding by revealing the causal relations among neural components.
Facilitating this effort, recent advances in network neuroscience (see Glossary) are providing ever more detailed descriptions of brain network architecture [4]. Yet these descriptions have so far provided only limited insights into the neural computations underlying cognitive functions. Simultaneously, connectionist artificial neural network research has revealed the theoretical importance of network connectivity in computing cognitive functions (Box 1). Yet it remains unclear whether these theoretical insights emerging primarily from simulations apply to empirical neural computations. Thus, connectionism has provided an abundance of theoretical insights with limited translation to empirical evidence, while empirical neuroscience now has an abundance of empirical observations with limited theoretical insights. We propose that combining connectionist theory with empirical brain connectivity data will advance understanding of the computational and cognitive relevance of brain network organization.
Box 1. The evolution of connectionism in relation to neuroscience.
Despite the biological and neural origins of connectionism [38,55,67,69], there are stark differences in current connectionist (deep learning) and neuroscientific research. One of the most influential advances in connectionism was the discovery of backpropagation, an algorithm that enabled the training of multilayer artificial neural networks. Pioneered by Paul Werbos in 1974 and popularized by other connectionist researchers [51,69,95], backpropagation provided researchers with the tools to train networks to perform increasingly complex tasks. This discovery strongly influenced the direction of connectionist research away from its neurobiological roots and towards the development of learning algorithms. Until recently, the development of these learning algorithms has been the primary focus of many connectionist and deep learning researchers.
Despite this focus on learning algorithms, learning in artificial systems is in many ways incomparable to learning in biological systems. One reason for this is the biological implausibility of the backpropagation learning algorithm, though biologically-plausible equivalents exist [96]. A more fundamental reason is that biological systems are born with inductive biases (i.e., innate structure) that shape how they interact with their environments early in life [93,97]. A recent opinion paper argued that the inductive biases afforded to a biological system are the result of learning during evolution (rather than within a single lifespan), and is likely embedded in an organism’s DNA [97]. Thus, learning in artificial systems is more appropriately compared to learning on evolutionary timescales (which often takes place outside the context of neuroscience research), since artificial neural networks are typically “born” with a blank slate.
The network coding model approach advocated here is consistent with recent work focused on integrating inductive biases into neural network models to understand neurocomputational principles. Such biologically-based inductive biases, such as the incorporation of highly structured network architectures, have been especially useful in understanding core object recognition in the ventral visual stream [47,48,50,75,98]. Other inductive biases, such as the incorporation of cortical heterogeneity and local biophysical properties, have been especially useful in characterizing simulated large-scale network dynamics [62,99–101]. Here we focus primarily on the incorporation of empirical brain network architecture to provide principled inductive biases to predictive (encoding/decoding) brain models. We expect insights from these network coding models to positively impact both neuroscience and connectionist research.
We begin by discussing efforts to model neural computations using empirical brain connectivity. We then draw on conceptual insights from the connectionist literature – including from recent research involving “deep” neural network architectures – to help identify the computational and cognitive relevance of the distributed neural processes reported in recent empirical studies. Here we focus on a subset of studies that have successfully linked distributed brain connectivity with neurocognitive computations. This research is framed in a new data analytic perspective – network coding models – that suggests a way toward unifying empirical brain network analysis with a rich theoretical connectionist foundation.
From Mapping Localized Functions to Building Network Coding Models
Network coding models facilitate understanding of the function of localized neural entities (such as brain regions) by clarifying how they send and receive information: connectivity. This also has the advantage of clarifying the role of each brain connection in computing cognitive functions. In contrast to network coding models, cognitive neuroscience has primarily mapped tasks and stimuli to activity in neurons and neural populations – function-structure mappings (Figure 1A, Key Figure) [5]. Some examples of this general strategy include spike rate changes in single- or multi-unit recordings [6], general linear modeling with functional MRI (fMRI) [7], and event-related potentials with electroencephalography [8]. This strategy has been tremendously useful for characterizing the functions of spatially localized neural populations [9].
Encoding and Decoding: Predictive Approaches for Mapping Localized Functions
More recent methods have used a predictive framework to allow for more complex function-structure mappings. By predicting independent data, these approaches avoid overestimation of effects from increased model complexity (e.g., overfitting to noise), while providing more robust function-structure mappings [10,11]. It is commonly underappreciated how problematic overfitting is in standard data analysis approaches (e.g., general linear models or event-related potentials) that do not test for generalization in independent data using cross-validation [10,11]. Among these predictive approaches, an encoding model uses the task/stimulus condition to predict neural activity [12–14]. In contrast, a decoding model supports the opposite inference, using activity in a neural population to predict the task/stimulus condition [12–14]. These predictive models facilitate interpretation of neural activity in terms of task-related information content, which is an important step towards understanding the brain as an information processing system [15,16]. Notably, we have conceptualized network coding models within the framework of encoding and decoding models, since they inherit many of the benefits of encoding and decoding models while incorporating connectivity as a mechanistic constraint (as illustrated in Figure 1).
The Role of Connectivity in Functional Localization
Despite the utility of these approaches, a comprehensive function-structure mapping would not provide insight into how local functionality is specified mechanistically. As an example of mechanistic understanding, consider work in visual neuroscience suggesting that receptive field properties in V1 can be explained by connectivity patterns with the lateral geniculate nucleus [17]. More recent work has extended this insight to suggest that localized functionality in visual cortex might result from intrinsic connectivity architecture [18,19], with one study indicating that face and word selectivity in visual areas emerge from connectivity and specific computational processes [20].
Thus, one possibility to gain mechanistic insight is to map relationships between neural entities – connectivity mapping – based on the well-supported hypothesis that cognitive/behavioral functionality emerges due to neural interactions [21–24] (Figure 1B). Conceptually, this goal is similar to constructing a causal map among the functional components of a car (e.g., pedal -> engine -> transmission -> wheels). Various approaches to estimating connectivity could potentially be useful here, such as resting-state functional connectivity [25–27], task-state functional connectivity [28–34], or structural connectivity [35,36]. However, connectivity mapping can only facilitate mechanistic understanding to the extent that it can provide insight into the causal relations among neural entities [1,3]. Nonetheless, progress can likely be made even with limited causal inferences (e.g., structural/anatomical connectivity indicating possible rather than actual causal influence) by constraining the likelihood of possible causal models [1]. However, similar to the limitations of function-structure mappings, a complete connectivity map of neural populations (i.e., a connectome) would still not explain the emergence of any cognitive/behavioral functionality in a neural population [3,37]. Critically, unlike function-structure mappings, a connectivity mapping has no reference to cognitive/behavioral functionality. Without grounding in cognitive/behavioral functionality, a connectivity mapping merely describes potential routes of signaling among localized neural populations with no reference to the relevant cognitive/behavioral information content.
Therefore, we propose the importance of combining function-structure mapping and connectivity mapping to understand how function emerges from network connectivity in the form of network coding models (Figure 1C). The use of network coding models will facilitate a shift from purely functional models towards mechanistic models of brain function [15]. In the following section, we provide theoretical and empirical support for brain network organization as the basis for cognitive computations, highlighting converging ideas from network neuroscience and connectionist theory. We review instances of empirical studies that have bridged this gap, demonstrating how connectionist principles can be evaluated in models estimated from empirical neural data. We will conclude with a detailed description of network coding models (foreshadowed in Figure 1C) as a particularly useful approach to characterizing the role of brain connectivity patterns in cognitive computations. Ultimately, we expect these emerging areas of research to unify the major theoretical perspectives and methodological approaches in neuroscience: localized functions and information (reflected in task-evoked neural activity), and network neuroscience.
Brain Network Organization Shapes Cognitive Computations
Connectionist Architectures and Cognitive Computations
Decades of neuroscience have focused on function-structure mapping. Why might incorporating brain network connectivity (and, by extension, network coding models) address this strategy’s limitations? Evidence from multiple sources suggests connectivity can provide a mechanistic explanation of how function emerges in neural networks (Box 1). First, decades of “connectionist” work with artificial neural network models have demonstrated the plausibility of distributed connectivity-based processes driving various complex cognitive functions [38–41]. Second, the standard model of neuroscience, as proposed by Ramón y Cajal [42] and solidified by Hodgkin and Huxley [43] and others [17], provides a prominent role for connectivity among discrete neural units in determining localized functionality. Third, there is increasing empirical evidence that the fundamental unit of functionality in the brain is not single neurons but rather populations of neurons [44–46].
In particular, neural network models have primarily utilized inter-unit connectivity to define the architecture for cognitive computations [47–49] (Figure 2A). This includes recently developed deep neural networks that improve model performance by including additional neural units with structured connectivity as “hidden” layers between input and output [47,49,50]. Thus, decades of modeling work demonstrates that connectivity architectures can support dozens (or hundreds) of complex cognitive processes, with more recent deep learning work indicating that additional performance gains are possible through refinement of connectivity architecture.
These considerations support the conclusion that the brain’s network organization is a major contributor to the computational architecture underlying its functionality, leading to a more focused question: how much function (in terms of both neural activity variance and cognitive/behavioral variance) can connectivity patterns explain? This is an important question since there are various alternatives to distributed connectivity in determining the functionality of single neurons and neural populations. For instance, relevant local neural population properties include population firing thresholds [43,55], neuron types [56,57], excitatory-inhibitory balance [58,59], local glial cell properties [60], and local recurrent connectivity [61,62]. Also highly relevant are neurotransmitter types (including large-scale neuromodulatory effects) [63,64] as well as systemic delivery of neuromodulators (e.g., hormones) [65].
Testing Connectionist Principles in Empirical Data
One way to evaluate the relative contribution of local versus distributed factors in determining the functionality of neural populations is to test the plausibility of each in computational models. We recently built on the fundamental algorithm underlying most computational simulations of large-scale neural processing – activity flow (Figure 2A) – to develop a procedure for testing local vs. distributed contributions to functionality [53,54,66] (Figure 2B). Activity flow is the movement of neural activity (e.g., spike rates or fMRI signal amplitudes in biological contexts) between neural entities. Quantitatively, the activity flow approach involves simulating a single neural network computation (e.g., a forward iteration in a feedforward neural network): ah = f(ΣiϵIaiwhi), where a unit ah’s activity is a linear combination of all other units activity (ΣiϵIai) weighted by their connectivity (whi) to ah before passing through a transfer function f, such as a sigmoid. Within the connectionism framework, this formalization is a combination of the propagation and activation rules [51]. Other names given to more specific instantiations of this algorithm (e.g., with a particular form of nonlinearity) include the McCulloch-Pitts neuron [55], the perceptron [67], the divisive normalization model of neural activity [68], adaptive linear element [69], and spreading activation [70]. Critically, this algorithm was adapted for use with empirical data (e.g., fMRI activity and functional connectivity estimates) to parameterize empirically-derived models that make quantitative predictions of the spread of estimated activity over brain network connections (Figure 2B).
In neural network simulations we found that activity flow mapping was only effective in network architectures with relatively large effects of distributed (relative to recurrent/local) connectivity (Figure 2C). This is consistent with previous findings, where large effects of inter-regional synaptic coupling (relative to local coupling) were important for predicting functional from structural connectivity [71]. Thus, when we applied this approach successfully in empirical fMRI data across a variety of diverse tasks (see Figure 2D for an example) we were able to conclude that distributed connectivity plays a substantial role in determining localized functionality [76]. This is highly compatible with the notion that each localized population has a “connectivity fingerprint” that largely determines its functionality [72,73]. These results are also in line with the observation that large-scale propagation of neural activity in animal models tends to conform to large-scale anatomical connectivity patterns [74].
Other studies have further demonstrated the role of brain connectivity in mediating distributed neural activity. One study found that decodable cognitive information transfer between pairs of brain regions can be mediated via activity flow processes [54]. This illustrated that information in a source brain region can be sent (encoded) through network connectivity to be received (decoded) by a target brain region (Figure 2E). In another study, we found that the task activations of healthy older adults can be transformed into dysfunctional task activations of unhealthy older adults via the latter’s connectivity patterns [66]. This was accomplished by estimating activity flow through disrupted intrinsic functional networks of subjects at increased risk for developing Alzheimer’s disease. This suggests that changes in brain network organization underlie “unhealthy” cognitive computations [66].
Using Connectionist Models to Understand Empirical Data
A general principle of connectionist theory is that the ability of a neural network model to perform an experimental task is dependent upon its connectivity patterns [46]. This suggests an alternative approach to study the importance of connectivity in producing cognitive computations: training an artificial neural network model to perform an experimental cognitive task. Despite the connectivity patterns of these models being specified more by task training than biological constraints, many of these models are able to accurately predict empirical neural responses [40,41,46,75]. Moreover, both biological and artificial systems can be characterized by the same computational mechanism: activity flow processes (Figure 2A) [52]. This again supports the conclusion that connectivity patterns – even those derived primarily from task training constraints – are central to neural computations underlying cognitive processing.
Despite these promising results, much work remains to determine the role of the brain’s network architecture in influencing cognitive functionality. For instance, the particular cognitive tasks used are only a small sample of the wide array of tasks humans are capable of [76,77]. Thus, much work remains to verify the role of connectivity patterns in determining the computations across diverse tasks and stimuli. Notably, understanding how many different tasks can be encoded through distributed network processes will likely facilitate the design of generalized models of cognition that can adaptively perform novel tasks [13,76,77]. Additionally, it will be important to assess how activity flow computations are altered across different levels of functional organization, from finer-grained activity and connectivity patterns to large-scale functional brain regions and networks. Finally, it will be important to determine the role of features other than connectivity – such as local activation functions (incorporating nonlinearities, such as compressive summation in visual cortex) [78,79] and operating timescales/dynamics [71,80] – in specifying the neural computations that facilitate cognitive processes.
Additional Approaches for Mapping Cognitive Function with Connectivity
Given the strong evidence that connectivity is central to neural computation, any methods that link cognitive function with connectivity are likely to provide useful theoretical insight. In this section we focus on efforts that characterize information in distributed networks, as well as efforts that quantify how cognitive information representations quantitatively change between brain areas. In the subsequent section we will focus on network coding models, which provide a mechanistic characterization of information processing, due to the incorporation of connectivity estimates as constraints. For instance, there have been significant advances in characterizing how the synchronization of neural time series during tasks – task-state functional connectivity – relates to ongoing cognitive processes. While links between resting-state functional connectivity and cognitive ability (estimated via individual difference correlations) have been widely reported [21,22,81–83], changes to the underlying functional network organization from resting state to task state can shed light on how those network changes contribute to task-related cognitive processes [30,34].
Recent work has shown that the functional network organization during resting and task states are highly similar [29,31], with the functional network organization at rest accounting for up to 80% of the variance in whole-brain network organization during tasks. However, studies have reported systematic task-related changes in functional network organization that reflect ongoing cognitive processes [28,33,84]. Moreover, transient changes in task-state functional connectivity have been shown to predict task performance [84]. Thus, measures of task-state functional connectivity can provide insight into which brain region interactions are involved in a cognitive process.
Despite the insights offered by standard task-state functional connectivity analyses, correlation-based measures of the BOLD signal between pairs of regions limit the identification of what kind of information is transmitted between brain regions and how this information might be transmitted. Several recent approaches have gone beyond standard functional connectivity measures, providing multivariate measures of temporal and spatial dependence between brain areas [85]. For example, one study measured the time-varying information content (i.e., decodability) during task states, and correlated the informational time series with other brain regions [86]. This technique goes beyond asking whether two brain regions are synchronized and addresses whether pairs of regions contain task-related information at the same time.
Other approaches estimate the neural transformation of cognitive representations between brain regions, clarifying the mathematical relationship between distinct units. One approach maps spatial activation patterns between brain areas using a nonlinear transformation, capturing the optimal or “normative” computational transformation required to project one brain region’s information into another brain region’s geometry [87]. Another similar approach estimated the optimal linear transformation required to project activation patterns in early visual cortex to regions further along the ventral visual stream, such as the fusiform face complex [88]. By identifying a simple linear transformation, the authors were able to investigate the computational properties of the linear transformation matrix, such as whether the mapping projected to a lower dimensional space (e.g., information compression) as the information was mapped from early visual areas to the fusiform face complex. Thus, by characterizing the representational mappings between brain regions during cognitive tasks, these studies go beyond what standard task-state functional connectivity approaches offer to characterize inter-unit computational relationships. However, such models fail to constrain their predictions with separate estimates of brain connectivity; a key aspect of network coding models.
Network Coding Models: Computing Cognitive Information in Neural Networks
We and others have made the case that a particularly powerful framework for characterizing the functionality of brain regions is to use encoding and decoding models [12–14]. However, most uses of encoding and decoding models are designed to characterize information of interest to the experimenter, and are inconsistent with how neural entities likely encode and decode task information biophysically [89,90]. An example of this are function-structure mappings that map high-level, human annotated information (such as semantic information) onto brain activity [91]. Though these data-driven functional models can be useful, neural entities likely encode and decode complex task features through network connectivity, rather than the direct task stimulus-to-neural response associations composing traditional encoding and decoding models (Figure 3A). The absence of mechanistic constraints (such as network connectivity) in these models limit the causal relevance of experimenter-extracted neural representations [89,90].
A potential solution to this problem is to incorporate network connectivity estimates in conjunction with traditional encoding and decoding models – a network coding model (Figures 1C, 3B, and 3C). Figure 3 formalizes the distinction between standard encoding/decoding models and network coding models. The latter approach would add mechanistic constraints – brain network connectivity – to encoding and decoding models. Constraining models via connectivity, as performed in activity flow mapping (Figure 2), enables a stronger mechanistic interpretation of how neural entities receive (encode) and send (decode) information within the brain. Additionally, this approach explicitly implements activity flow processes in empirical data (Figure 2A). This increases the biological relevance of these models by simulating the physical process by which neural signals are relayed through network connectivity [53,54,66]. Ultimately, this framework strengthens causal inferences made about the transmission of task information within the brain, beyond traditional descriptive statistics [3,54].
Network coding models provide a biologically plausible link between encoding/decoding models and connectionist theory when studying the propagation of activity during cognitive tasks. Several recent methods have applied this general framework. The first is to incorporate estimates of functional connectivity directly with traditional, experimenter-based encoding and decoding models (Figure 3B) [54]. This approach tests whether decodable information in one brain area can be re-encoded through connectivity patterns and decoded in a downstream target brain area, providing a biophysically plausible model of information transfer. We recently demonstrated that task information in a set of brain regions could be used to predict the same task features in downstream regions through activity flow processes using functional connectivity estimated using resting-state fMRI (Figure 2E) [54]. These findings illustrate that task information in a brain area not only can be encoded/decoded by the experimenter, but is also used by other brain areas through distributed network connectivity.
The second approach uses structured connectionist models (neural network models with architectural constraints) to study the emergence of localized functionality from connectivity (Figure 3C). Recent technological advances in the training of structured connectionist models, such as biologically-inspired deep neural networks and recurrent neural networks, have enabled the study of the encoding and decoding of cognitive information via activity flow through optimized connectivity architectures. For example, a recent study showed that a recurrent neural network was able to represent an array of different inputs, such as different task states and stimuli (encoding), and map those inputs to different motor outputs (decoding) for accurate task performance [41]. Importantly, the task features/stimuli that this neural network encoded (its inputs) were qualitatively different from the features that it decoded (motor outputs), offering insight into how information might be transformed across neural entities via network connectivity [49]. Thus, while traditional experimenter-based encoding and decoding models typically address what a neural entity might be encoding and decoding, network coding models can also address how that information might be computed and/or subsequently processed.
Although unconstrained/unstructured neural network models are universal function approximators [92], employing biological priors in neural network models can aid in producing mechanistic models of neural computations (Figure 3C) [93]. Such constraints (model assumptions) are essential to discovering the model features that are involved in the neural computations underlying cognitive functions. For instance, connectivity architecture and task performance constraints were implemented to demonstrate how images can be encoded and decoded to identify objects in neural network models in both human [50] (Figure 4) and non-human primates [47,48]. These types of models map visual images onto specific object codes via a nonlinear computation mediated by connectivity (Figure 4A). While any neural network model can be theoretically trained to accurately perform this computation, Wen and colleagues demonstrated that when adding biological constraints into this neural network (e.g., number of layers and number of units per layer to match those of the ventral visual stream) the network exhibited similar neuronal responses to empirical fMRI data obtained in humans (Figure 4B) [50]. Moreover, they identified a face-selective unit in their model and showed that its activity was highly correlated with the fusiform face area in empirical fMRI data when dynamic naturalistic stimuli was shown to both human and model (Figure 4C). This suggests that like the brain, localized functionality can emerge in neural network models even when primarily optimized for task performance. Critically, incorporating realistic biological constraints can improve the fidelity with which artificial neural networks mimic the brain in representing cognitive information.
Network coding models can therefore provide directly testable hypotheses in empirical data. For example, if a neural network model contains enough biological constraints, a mapping between the network model and empirical neural data can be made (Figure 3C). In a recent study, a biologically-constrained artificial neural network of the ventral visual stream provided insight into how to control and activate targeted neural populations in area V4 [75]. By dissecting the receptive fields of specific units in their biologically-constrained artificial neural network model, the authors generated novel visual stimuli that could theoretically target a subset of biological units in the primate V4. Indeed, the authors showed that these stimuli could activate and control corresponding neural populations empirically in the primate brain, uncovering the receptive fields (and localized function) of those biological units through model analysis. This approach provides a powerful framework to test how localized functionality emerges from network organization in both artificial and biophysical neural networks. We suggest that including additional biological constraints, such as empirically-derived connectivity estimates, would likely lend additional mechanistic insight into how neural information processing is carried out in the brain [15].
Concluding Remarks
The recent proliferation of large neural data sets with rich task features has created a wealth of opportunities in functional brain mapping. However, data-driven approaches to mapping task features to neural responses largely disregard the biological mechanisms of neural information processing: distributed cognitive processing through brain network connectivity. Since the early days of connectionism, the functionality of neural entities has long been hypothesized to be embedded in its patterns of network connectivity [51,72,73]. Here we propose the incorporation of brain network connectivity as a biological constraint underlying the emergence of localized functionality and distributed cognitive processing.
Though mounting evidence suggests that connectivity undoubtedly plays a role in cognition, the precise contribution of distributed versus local processing remains an open question (see Outstanding Questions). Several advances are needed to answer this question. First, it will be important to advance current connectivity estimates, given that current methods may only indicate associational (and not causal) relations [3,94]. Second, it will be important to integrate additional biological properties that shape local processing properties (such as nonlinear transfer functions) and incorporate them into network coding models [20]. Lastly, it will be important to evaluate how these principles might generalize across levels of organization and spatial scales using the wide variety of data types available to neuroscientists. Such insight can provide more mechanistic insight into how local functionality emerges from the interplay between distributed (e.g., connectivity) and local (e.g., nonlinear) processes. Taken together, we can begin to reconstruct how the collective interaction of functionally-specific localized processes work together to compute diverse cognitive functions.
Outstanding questions.
How do other local properties, such as the intrinsic timescale and local transfer function contribute to the localized functions of neural entities? What are the relative contributions of each neural entity’s global connectivity pattern vs. its local properties?
What are the best functional connectivity methods that incorporate the nonlinear relationships between neural entities that are thought to be critical for neural computation?
Do activity flow mapping procedures generalize to other data types (outside of just fMRI), such as whole-brain electrophysiology data (electroencephalography/magnetoencephalography), multi-unit spiking data (obtained from distributed brain areas), or wide-field calcium imaging data?
How might network coding models be integrated across different sensory, motor, and association systems to achieve an integrative model of cognition and behavior?
Highlights.
Recent results suggest that localized functions in brain areas are specified primarily by their distributed global connectivity patterns
Recent approaches go beyond stimulus/task brain mapping (and encoding/decoding) to characterize the role of brain connectivity in neural information processing
We introduce network coding models as a framework encompassing neural network models optimized for task performance and those optimized for biological realism
Biological and task performance constraints are complementary for aiding the search for accurate models of empirical brain function
The activity flow algorithm is a core computational mechanism underlying network coding models, linking brain activity and connectivity in a biologically-plausible mechanistic framework
Acknowledgements
The authors would like to thank Ruben Sanchez-Romero for assistance with preparing this article. The authors acknowledge the support of the US National Institutes of Health under awards R01 AG055556 and R01 MH109520, and the Behavioral and Neural Sciences Graduate Program at Rutgers, The State University of New Jersey. The content is solely the responsibility of the authors and does not necessarily represent the official views of any of the funding agencies.
Glossary
- Activity flow
A fundamental computation (see Figure 2A) describing the movement of activity between neural units as a function of their connectivity (e.g., propagating spikes over axons in the brain). This is equivalent to the activation and propagation rules used in connectionist research
- Connectionist/Connectionism
A subfield within cognitive science that focuses on implementing cognitive and mental phenomena computationally through artificial neural networks
- Connectivity mapping
The quantification of the relationship between two or more neural entities (e.g., brain regions) using either statistical dependencies of the neural time series (functional connectivity) or estimates of structural/anatomical pathways (structural connectivity)
- Connectome
A complete connectivity map of all units in a neural system
- Constraint
A limitation placed on a model in an optimization context. We focus primarily on empirical constraints, especially biological constraints such as brain network organization. We also focus on task performance constraints, which are typically emphasized in connectionist modeling
- Decoding model
A statistical model that predicts a task stimulus or condition as a function of a set of neural responses
- Deep neural network models
Neural network models with more than one hidden layer, which have been shown to boost task performance in many cases relative to traditional neural network models
- Encoding model
A statistical model that predicts a neural response as a function of a task stimulus or condition
- Experimenter-based encoding/decoding models
In contrast to network coding models, encoding/decoding models that focus on how the experimenter encodes/decodes information from a neural entity, rather than how other neural entities in the brain encode/decode that information
- Function-structure mapping
The association between a particular neural entity and its functionality, such as what task stimuli a neural entity activates or responds to
- Network coding models
Models of brain function that simulate encoding and decoding processes through network connectivity (typically via the activity flow algorithm) to predict empirical brain activity
- Network neuroscience
A subfield of neuroscience concerned with understanding the network principles of neural phenomena by exploiting tools from network science
- Neural network models
Computational models consisting of a network of interconnected units that are optimized to match biological features (biological constraints) and task performance (normative task performance constraints) to varying extents
- Recurrent neural network model
Neural network models that feedforward through time, rather than through different spatial layers. Recurrent connections indicate directed connections to other units within the network that propagate activity through time
- Structured connectionist models
Neural network models with built-in architectures (i.e., innate structure), such as many deep neural networks with a fixed number of layers
Footnotes
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
References
- 1.Reid AT et al. (In Press) Advancing functional connectivity research from association to causation. Nat. Neurosci. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Craver C (2007) Explaining the brain, Oxford University Press. [Google Scholar]
- 3.Mill RD et al. (2017) From connectome to cognition: The search for mechanism in human functional brain networks. NeuroImage DOI: 10.1016/j.neuroimage.2017.01.060 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Bassett DS and Sporns O (2017) Network neuroscience. Nat. Neurosci 20, 353. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Henson R (2005) What can functional neuroimaging tell the experimental psychologist? Q. J. Exp. Psychol. Sect. A 58, 193–233 [DOI] [PubMed] [Google Scholar]
- 6.Wallis JD (2018) Decoding Cognitive Processes from Neural Ensembles. Trends Cogn. Sci DOI: 10.1016/j-tics.2018.09.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Poldrack RA et al. (2011) Handbook of functional MRI data analysis, Cambridge University Press. [Google Scholar]
- 8.Luck SJ (2014) An introduction to the event-related potential technique, MIT press. [Google Scholar]
- 9.Genon S et al. (2018) How to Characterize the Function of a Brain Region. Trends Cogn. Sci 22, 350–364 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Kriegeskorte N et al. (2009) Circular analysis in systems neuroscience: the dangers of double dipping. Nat. Neurosci 12, 535–540 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Varoquaux G and Poldrack RA (2019) Predictive models avoid excessive reductionism in cognitive neuroimaging. Curr. Opin. Neurobiol 55, 1–6 [DOI] [PubMed] [Google Scholar]
- 12.Gallant J et al. (2010) System identification, encoding models, and decoding models, a powerful new approach to fMRI research. Underst. Vis. Popul. Codes [Google Scholar]
- 13.Naselaris T et al. (2011) Encoding and decoding in fMRI. NeuroImage 56, 400–410 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Kriegeskorte N and Douglas PK (2018) Interpreting Encoding and Decoding Models. ArXiv181200278 Q-Bio at <http://arxiv.org/abs/1812.00278> [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Kay KN (2018) Principles for models of neural information processing. NeuroImage 180, 101–109 [DOI] [PubMed] [Google Scholar]
- 16.Kriegeskorte N and Douglas PK (2018) Cognitive computational neuroscience. Nat. Neurosci 21, 1148–1160 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Hubel DH and Wiesel TN (1962) Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J. Physiol 160, 106–154 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Haak KV et al. (2013) Connective field modeling. NeuroImage 66, 376–384 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Arcaro MJ et al. (2015) Widespread correlation patterns of fMRI signal across visual cortex reflect eccentricity organization. eLife 4, [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Kay K and Yeatman J (2016) Bottom-up and top-down computations in high-level visual cortex. bioRxiv DOI: 10.1101/053595 [DOI] [Google Scholar]
- 21.Cole MW et al. (2012) Global connectivity of prefrontal cortex predicts cognitive control and intelligence. J. Neurosci. Off. J. Soc. Neurosci 32, 8988–99 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Finn ES et al. (2015) Functional connectome fingerprinting: identifying individuals using patterns of brain connectivity. Nat. Neurosci 18, 1664–1671 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Smith SM et al. (2015) A positive-negative mode of population covariation links brain connectivity, demographics and behavior. Nat. Neurosci 18, 1–7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Li J et al. (2019) Global signal regression strengthens association between resting- state functional connectivity and behavior. NeuroImage 196, 126–141 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Biswal B et al. (1995) Functional connectivity in the motor cortex of resting human brain using echo-planar MRI. Magn. Reson. Med. 34, 537–541 [DOI] [PubMed] [Google Scholar]
- 26.Power JD et al. (2011) Functional Network Organization of the Human Brain. Neuron 72, 665–678 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Yeo BT et al. (2011) The organization of the human cerebral cortex estimated by intrinsic functional connectivity. J. Neurophysiol 106, 1125–1165 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Cole MW et al. (2013) Multi-task connectivity reveals flexible hubs for adaptive task control. Nat. Neurosci. 16, 1348–1355 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Cole MW et al. (2014) Intrinsic and task-evoked network architectures of the human brain. Neuron 83, 238–251 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Cole MW et al. (2019) Task activations produce spurious but systematic inflation of task functional connectivity estimates. NeuroImage 189, 1–18 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Krienen FM et al. (2014) Reconfigurable task-dependent functional coupling modes cluster around a core functional architecture. Philos. Trans. R. Soc. B Biol. Sci 369, 20130526–20130526 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Hearne L et al. (2015) Interactions between default mode and control networks as a function of increasing cognitive reasoning complexity. Hum. Brain Mapp 36, 2719–2731 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Hearne LJ et al. (2017) Reconfiguration of Brain Network Architectures between Resting-State and Complexity-Dependent Cognitive Reasoning. J. Neurosa 37, 8399–8411 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Ito T et al. (2019) Task-evoked activity quenches neural correlations and variability in large-scale brain systems. bioRxiv DOI: 10.1101/560730 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Honey CJ et al. (2007) Network structure of cerebral cortex shapes functional connectivity on multiple time scales. Proc. Natl. Acad. Sci. U. S. A 104, 10240–5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Hagmann P et al. (2008) Mapping the structural core of human cerebral cortex. PLoS Biol. 6, 1479–1493 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Bargmann C.l. and Marder E (2013) From the connectome to brain function. Nat. Methods 10, 483–490 [DOI] [PubMed] [Google Scholar]
- 38.Rumelhart DE and McClelland JL (1986) Parallel distributed processing: explorations in the microstructure of cognition. Volume 1 Foundations; at <https://www.osti.gov/biblio/5838709> [DOI] [PubMed] [Google Scholar]
- 39.Cohen JD et al. (1990) On the control of automatic processes: a parallel distributed processing account of the Stroop effect. Psychol. Rev 97, 332–61 [DOI] [PubMed] [Google Scholar]
- 40.Mante V et al. (2013) Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature 503, 78–84 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Yang GR et al. (2019) Task representations in neural networks trained to perform many cognitive tasks. Nat. Neurosai DOI: 10.1038/s41593-018-0310-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Ramón y Cajal S (1888) Estructura de los centros nerviosos de las aves, [Google Scholar]
- 43.Hodgkin AL and Huxley AF (1952) A quantitative description of membrane current and its application to conduction and excitation in nerve. J. Physiol 117, 500–44 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Rigotti M et al. (2013) The importance of mixed selectivity in complex cognitive tasks. Nature 497, 585–90 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Saxena S and Cunningham JP (2019) Towards the neural population doctrine. Curr. Opin. Neurobiol 55, 103–111 [DOI] [PubMed] [Google Scholar]
- 46.Williamson RC et al. (2019) Bridging large-scale neuronal recordings and large-scale network models using dimensionality reduction. Curr Opin. Neurobiol 55, 40–47 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Khaligh-Razavi SM and Kriegeskorte N (2014) Deep Supervised, but Not Unsupervised, Models May Explain IT Cortical Representation. PLoS Comput. Biol 10, [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Yamins DLK et al. (2014) Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proc. Natl. Acad. Sci 111, 8619–8624 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Yamins DLK and DiCarlo JJ (2016) Using goal-driven deep learning models to understand sensory cortex. Nat. Neurosci 19, 356. [DOI] [PubMed] [Google Scholar]
- 50.Wen H et al. (2018) Neural Encoding and Decoding with Deep Learning for Dynamic Natural Vision. Cereb. Cortex 28, 4136–4160 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Rumelhart DE et al. (1986) A general framework for parallel distributed processing. Parallel Distrib. Process. Explor. Microstruct. Cogn 1,45–76 [Google Scholar]
- 52.McClelland JL and Rogers TT (2003) The parallel distributed processing approach to semantic cognition. Nat. Rev. Neurosci 4, 310–322 [DOI] [PubMed] [Google Scholar]
- 53.Cole MW et al. (2016) Activity flow over resting-state networks shapes cognitive task activations. Nat. Neurosci DOI: 10.1038/nn.4406 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Ito T et al. (2017) Cognitive task information is transferred between brain regions via resting-state network topology. Nat. Commun 8, [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.McCulloch WS and Pitts W (1943) A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys 5, 115–133 [PubMed] [Google Scholar]
- 56.Markram H et al. (2004) Interneurons of the neocortical inhibitory system. Nat. Rev. Neurosci 5, 793. [DOI] [PubMed] [Google Scholar]
- 57.Gouwens NW et al. (2019) Classification of electrophysiological and morphological neuron types in the mouse visual cortex. Nat. Neurosci DOI: 10.1038/s41593-019-0417-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Barrai J and Reyes AD (2016) Synaptic scaling rule preserves excitatory-inhibitory balance and salient neuronal network dynamics. Nat. Neurosci DOI: 10.1038/nn.4415 [DOI] [PubMed] [Google Scholar]
- 59.Hennequin G et al. (2017) Inhibitory Plasticity: Balance, Control, and Codependence. Annu. Rev. Neurosci 40, 557–579 [DOI] [PubMed] [Google Scholar]
- 60.De Pittà M and Berry H (2019) Computational Glioscience, Springer. [Google Scholar]
- 61.Demirtaş M et al. (2019) Hierarchical Heterogeneity across Human Cortex Shapes Large-Scale Neural Dynamics. Neuron DOI: 10.1016/j.neuron.2019.01.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Wang P et al. (2019) Inversion of a large-scale circuit model reveals a cortical hierarchy in the dynamic resting human brain. Sci. Adv 5, eaat7854. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Servan-Schreiber D et al. (1990) A network model of catecholamine effects: gain, signal-to-noise ratio, and behavior. Science 249, 892–895 [DOI] [PubMed] [Google Scholar]
- 64.Shine JM (2019) Neuromodulatory Influences on Integration and Segregation in the Brain. Trends Cogn. Sci 0, [DOI] [PubMed] [Google Scholar]
- 65.Izquierdo I and Medina JH (1997) The biochemistry of memory formation and its regulation by hormones and neuromodulators. Psychobiology 25, 1–9 [Google Scholar]
- 66.Mill RD et al. (2019) Predicting dysfunctional age-related task activations from resting-state network alterations. bioRxiv DOI: 10.1101/678086 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Rosenblatt F (1958) The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev 65, 386–408 [DOI] [PubMed] [Google Scholar]
- 68.Carandini M and Heeger DJ (2012) Normalization as a canonical neural computation. Nat. Rev. Neurosci 13, 51–62 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Widrow B and Lehr MA (1990) 30 Years of Adaptive Neural Networks: Perceptron, Madaline, and Backpropagation. Proc. IEEE 78, 1415–1442 [Google Scholar]
- 70.Collins AM and Loftus EF (1975) A spreading-activation theory of semantic processing. Psychol. Rev. 82, 407 [Google Scholar]
- 71.Deco G et al. (2013) Resting-state functional connectivity emerges from structurally and dynamically shaped slow linear fluctuations. J. Neurosci. Off. J. Soc. Neurosci 33, 11239–52 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Passingham RE et al. (2002) The anatomical basis of functional localization in the cortex. Nat. Rev. Neurosci 3, 606–616 [DOI] [PubMed] [Google Scholar]
- 73.Mars RB et al. (2018) Connectivity Fingerprints: From Areal Descriptions to Abstract Spaces. Trends Cogn. Sci 22, 1026–1037 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Kotter R and Sommer FT (2000) Global relationship between anatomical connectivity and activity propagation in the cerebral cortex. Philos. Trans. R. Soc. Lond. B. Biol. Sci 355, 127–134 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Bashivan P et al. (2019) Neural population control via deep image synthesis. Science 364,eaav9436. [DOI] [PubMed] [Google Scholar]
- 76.Cole MW et al. (2013) Rapid instructed task learning: A new window into the human brain’s unique capacity for flexible cognitive control. Cogn. Affect. Behav. Neurosci 13, 1–22 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Yang GR et al. (2019) How to study the neural mechanisms of multiple tasks. Curr. Opin. Behav. Sci 29, 134–143 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Reynolds JH and Heeger DJ (2009) The normalization model of attention. Neuron 61, 168–185 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Kay KN et al. (2013) Compressive spatial summation in human visual cortex. J. Neurophysiol 110, 481–494 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Murray JD et al. (2014) A hierarchy of intrinsic timescales across primate cortex. Nat. Neurosci 17, 1661–1663 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Schultz DH and Cole MW (2016) Higher Intelligence Is Associated with Less Task-Related Brain Network Reconfiguration. J. Neurosci 36, 8551–8561 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Gratton C et al. (2018) Functional Brain Networks Are Dominated by Stable Group and Individual Factors, Not Cognitive or Daily Variation. Neuron 98, 439–452.e5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Kashyap R et al. (2019) Individual-specific fMRI-Subspaces improve functional connectivity prediction of behavior. NeuroImage 189, 804–812 [DOI] [PubMed] [Google Scholar]
- 84.Sadaghiani S et al. (2015) Ongoing dynamics in large-scale functional connectivity predict perception. Proc. Natl. Acad. Sci. U. S. A 112, 8463–8468 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Anzellotti S and Coutanche MN (2017) Beyond Functional Connectivity: Investigating Networks of Multivariate Representations. Trends Cogn. Sci xx, 1–12 [DOI] [PubMed] [Google Scholar]
- 86.Coutanche MN and Thompson-Schill SL (2013) Informational connectivity: identifying synchronized discriminability of multi-voxel patterns across the brain. Front. Hum. Neurosci 7, 15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Anzellotti S et al. (2016) Measuring and Modeling Transformations of Information Between Brain Regions with fMRI. bioRxiv DOI: 10.1101/074856 [DOI] [Google Scholar]
- 88.Basti A et al. (2019) Analysing linear multivariate pattern transformations in neuroimaging data. bioRxiv DOI: 10.1101/497180 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89.De-Wit L et al. (2016) Is neuroimaging measuring information in the brain? Psychon. Bull. Rev DOI: 10.3758/s13423-016-1002-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90.Brette R (2018) Is coding a relevant metaphor for the brain? bioRxiv DOI: 10.1101/168237 [DOI] [PubMed] [Google Scholar]
- 91.Huth AG et al. (2016) Natural speech reveals the semantic maps that tile human cerebral cortex. Nature 532, 453–458 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Hartman EJ et al. (1990) Layered Neural Networks with Gaussian Hidden Units as Universal Approximations. Neural Comput 2, 210–215 [Google Scholar]
- 93.Sinz FH et al. (2019) Engineering a Less Artificial Intelligence. Neuron 103, 967–979 [DOI] [PubMed] [Google Scholar]
- 94.Power JD et al. (2012) Spurious but systematic correlations in functional connectivity MRI networks arise from subject motion. NeuroImage 59, 2142–2154 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Werbos P (1974) Beyond Regression:” New Tools for Prediction and Analysis in the Behavioral Sciences. Ph Diss. Harv. Univ [Google Scholar]
- 96.O’Reilly RC (1996) Biologically plausible error-driven learning using local activation differences: The generalized recirculation algorithm. Neural Comput. 8, 895–938 [Google Scholar]
- 97.Zador AM (2019) A critique of pure learning and what artificial neural networks can learn from animal brains. Nat. Commun 10, 1–7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Kar K et al. (2019) Evidence that recurrent circuits are critical to the ventral stream’s execution of core object recognition behavior. Nat. Neurosci 22, 974–983 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Singh MF et al. (2019) Individualized Dynamic Brain Models: Estimation and Validation with Resting-State fMRI. bioRxiv DOI: 10.1101/678243 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Seeliger K et al. (2019) Neural System Identification with Neural Information Flow. bioRxiv DOI: 10.1101/553255 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Demirtas M et al. (2018) Hierarchical heterogeneity across human cortex shapes large-scale neural dynamics. bioRxiv DOI: 10.1101/341966 [DOI] [PMC free article] [PubMed] [Google Scholar]