Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Jul 8.
Published in final edited form as: Annu Rev Neurosci. 2020 Jul 8;43:249–275. doi: 10.1146/annurev-neuro-092619-094115

Computation Through Neural Population Dynamics

Saurabh Vyas 1,3, Matthew D Golub 2,3, David Sussillo 2,3,4, Krishna V Shenoy 1,2,3,5
PMCID: PMC7402639  NIHMSID: NIHMS1612314  PMID: 32640928

Abstract

Significant experimental, computational, and theoretical work has identified rich structure within the coordinated activity of interconnected neural populations. An emerging challenge now is to uncover the nature of the associated computations, how they are implemented, and what role they play in driving behavior. We term this computation through neural population dynamics. If successful, this framework will reveal general motifs of neural population activity and quantitatively describe how neural population dynamics implement computations necessary for driving goal-directed behavior. Here, we start with a mathematical primer on dynamical systems theory and analytical tools necessary to apply this perspective to experimental data. Next, we highlight some recent discoveries resulting from successful application of dynamical systems. We focus on studies spanning motor control, timing, decision-making, and working memory. Finally, we briefly discuss promising recent lines of investigation and future directions for the computation through neural population dynamics framework.

Keywords: neural computation, neural population dynamics, dynamical systems, state spaces

1. INTRODUCTION

The human brain is composed of many billions of interconnected neurons and gives rise to our mental lives and behavior. Understanding how neurons take in information, process it, and produce behavior is the focus of considerable experimental, computational, and theoretical neuroscience research. In this review, we focus on how computations are performed by populations of neurons. Many insights have been gained in recent years, and the field is now ripe for substantial progress toward a more comprehensive understanding. This is due to a confluence of advances in large-scale neural measurements and perturbations, computational models, and theory (Paninski & Cunningham 2018).

A neuron’s inputs and cellular biophysics determine when it generates an action potential (Hodgkin & Huxley 1952). This is commonly summarized as a neuron’s action potential emission rate or firing rate [i.e., the firing rate x of neuron i at time t is xi(t)]. This rate, especially in sensory areas, relates to various parameters [i.e., xi(t) = ri param1(t), param2(t), param3(t), … where r is some function and parami is a parameter such as the angle of a visual stimulus] (e.g., Hubel & Wiesel 1962).

While modeling individual neurons has provided foundational insights, in recent years there has been considerable interest in understanding how populations of neurons encode and process information (e.g., Georgopoulos et al. 1982, Saxena & Cunningham 2019), including the interaction between neurons (Cohen & Maunsell 2009, Mitchell et al. 2009). Most recently, models have embraced an even more general and potentially powerful conceptual framework that is central in engineering and physical sciences: a dynamical systems framework (e.g., Luenberger 1979, Shenoy et al. 2013). This approach has helped us to understand cortical responses that are dominated by intrinsic neural dynamics, as opposed to sensory input dynamics (Seely et al. 2016), and has been explored over the past 15 years in a variety of contexts (Briggman et al. 2005, 2006; Broome et al. 2006; Churchland et al. 2006c; Mazor & Laurent 2005; Rabinovich et al. 2008; Stopfer et al. 2003; Sussillo & Abbott 2009; Yu et al. 2005). In this framework, a neural population constitutes a dynamical system that, through its temporal evolution, performs a computation to, for example, generate and control movement (Churchland et al. 2006a, 2010, 2012; Pandarinath et al. 2018a,b; Shenoy et al. 2013; Williamson et al. 2019).

A prediction of this dynamical systems perspective is that dynamical features should be present in the neural population response. In its general and deterministic form, but setting aside synaptic plasticity and hidden neurons/states that can influence dynamics, a dynamical system can be expressed as

dxdt=f(x(t),u(t)), 1.

where x is now an N-dimensional (ND) vector describing the firing rates of all neurons recorded simultaneously, which we call the neural population state, and is evaluated around time t. The vector dx/dt is its temporal derivative, f is a potentially nonlinear function, and u is a vector describing an external input to the neural circuit (a U-dimensional vector). In this formulation, neural population responses reflect underlying dynamics resulting from the intracellular dynamics, circuitry connecting neurons to one another (captured by f), and inputs to the circuit (captured by u).

To help build intuition for what a dynamical feature might look like in a neural population response, let us consider a physical example that is low dimensional such as a pendulum (Figure 1a). A pendulum is a two-dimensional (2D) dynamical system whose state variables are position and velocity. The pendulum is shown in one of two initial states, which are the initial conditions of the pendulum dynamics. When the pendulum is released from its initial condition such that its equation of motion [xsin(x)] evolves through time, the initial conditions trace out two different position and velocity trajectories (Figure 1b). Plotting the state variables of position and velocity against each other in a 2D plot results in a state space plot (Figure 1c). The flow field shows what the pendulum would do if it started in any given position in its state space; thus, the state space plot yields a convenient summary of the overall dynamics.

Figure 1.

Figure 1

Example dynamics and population state of a frictionless pendulum. (a–c) A pendulum is a 2D dynamical system that has a simple relationship between the changes of the state of the pendulum (1D position and 1D velocity) and its current state. (a) Two example initial conditions are shown, p1 (gray; position is maximally negative and velocity is zero) and p2 (blue; position is maximally positive and velocity is zero). (b) As the pendulum moves through time, its state vector (position, top; velocity, bottom) is updated. (c) These trajectories can also be visualized by plotting one state variable against the other (position against velocity), beginning at the initial condition (circles) and moving along the flow field (green arrows), yielding the state space plot. The flow field shows the dynamical evolution of the pendulum for any set of initial conditions. (d–f) A pendulum with input perturbations. (d) One may perturb the pendulum during its motion, for example, by bumping it with a finger (brief period of the bumps appears in red). (e,f) This leads to a change in the pendulum state through time (red dashed lines) such that both the time series (e) and the state space trajectories (f) no longer follow the autonomous dynamics (as shown in panel e) while the input is active. After the perturbation is finished, the pendulum returns to its autonomous dynamics again (gray lines after red dashed lines). Figure adapted with permission from Pandarinath et al. (2018b).

We can elaborate on this example by considering the pendulum to also occasionally experience an external input [i.e., u(t) in Equation 1]. For example, we might briefly push it with our finger (Figure 1d). That perturbative input results in a change in the dynamical trajectory, shown in Figure 1e. When the dynamics are perturbed by pushing the pendulum, the autonomous dynamics are modified such that state trajectories (Figures 1e,f) no longer follow autonomous dynamics (Figure 1f). Thus, an input can modify the dynamics in a time-dependent way. This can be used in two ways in neural circuits: to add, for example, sensory input to a computation, or to contextualize a computation by changing the dynamical regime in which the circuit is operating. To summarize the pendulum example, the dynamical features are the initial conditions of the two trajectories (Figure 1a), the pendulum’s dynamical equation (shown in graphical form in Figure 1c and captured in the function f), and inputs that, for example, set the pendulum’s initial condition or briefly perturb its dynamics (Figure 1d).

Now let us return to a neural system, where if we have N neurons we can construct an ND neural population state by first counting spikes within some short time period (e.g., 10 ms), as in Figure 2a. We can then plot these spike counts in a state space plot, as shown in Figure 2a,b. Here, the vector of spike counts for neurons 1 through N (i.e., the neural population state) takes the place of position and velocity (i.e., physical states) from the pendulum example. We can similarly plot flow fields for these neural population dynamics, as shown in Figure 2c (in analogy to Figure 1c). The dynamics govern the evolution of the trajectory in state space; different initial conditions (i.e., starting states) lead to different trajectory evolutions (Figure 2d). Note that although we have described the neural population state in terms of binned spike counts, it is common to instead choose to work with the abstraction of continuous-valued underlying firing rates. These firing rates are a latent state, since they cannot be directly observed but rather must be inferred, typically through an inhomogeneous Poisson process over the observed spike counts.

Figure 2.

Figure 2

Neural population state and neural dynamics. (a) Three neurons spike through time (left), and these spikes are binned and counted (center). These spike counts are plotted in a three-dimensional state space (right) representing the three neurons (n1, n2, n3), with a yellow square representing the state of the neural system at that time point. (b) Continued binning of spikes through time (left) leads to a neural population state that progresses in a sequence (blue, cyan, green, yellow), which is plotted in the neural state space plot (right) to yield an intuitive picture of the neural population dynamics. (c) The neural population dynamics can be represented by a flow field (green arrows) that describes how the neural population state evolves through time. (d) Two different neural trajectories (magenta and purple curves) can result from following the neural population dynamics from different initial conditions (magenta and purple squares).

Figure 2 describes a three-neuron population. However, most data sets involve many neurons, so we must contend with this additional complexity. In many areas of the brain, it is possible to reduce high-dimensional neural data (e.g., a hundred dimensions if measuring N= 100 neurons simultaneously) to lower-dimensional data (e.g., ten to twenty dimensions) without much loss in neural data variance captured (e.g., 5% variance lost) using a variety of different dimensionality reduction techniques (e.g., Cunningham & Yu 2014, Brunton et al. 2016, Kobak et al. 2016, Williams et al. 2018). While plotting only the first two or three dimensions (e.g., x1 and x2, whose dimensions capture the most and second-most data variance) ignores important data variance, the 2D to three-dimensional (3D) plot remains instructive, as it provides a sense of where the neural population state is and how it moves through time. However, we emphasize that any quantitative analysis of the population data should be done in a dimension high enough to explain all or nearly all of the neural variance.

The central question underlying many of the studies we review is how a neural circuit (captured by f) and its input (captured by u) create the population dynamics (depicted in Figure 2c,d; captured by Equation 1) that cause the neural population state to evolve through time to perform sensory, cognitive, or motor computations that result in perception, memory, and action. We term this general framework computation through neural population dynamics, or more succinctly, computation through dynamics (CTD).

2. COMPUTATION THROUGH NEURAL POPULATION DYNAMICS

Here we give an atypically lengthy mathematical introduction providing the series of concepts and procedures needed to better understand complex, high-dimensional neural systems. Readers familiar with these concepts may wish to skip to Section 3.

2.1. Deep Learning for Modeling Dynamical Systems

Deep learning (DL) is a catch-all phrase for a set of artificial neural network architectures along with supporting training algorithms, software libraries, and hardware platforms. DL is used to efficiently learn complex input-output relationships. In DL, a task is defined by a large training set of inputs and the desired outputs that should result from those inputs. To train a network, a loss function is defined relative to this training set, and that function is optimized by modifying the network parameters, typically via a gradient-descent algorithm. Successful optimization thus produces a network that solves the provided task.

For this review, we are primarily interested in using DL to model dynamical systems, i.e., f in Equation 1. The first approach is called data modeling, where we record the internal state of f [e.g., neural recordings represented by x(t)] and wish to identify f from those recordings. The second approach is task-based modeling, where we have a set of input-output pairs from a behavioral task (e.g., visually instructed arm reaches) and wish to identify a dynamical system, f, capable of transforming the input into the output. Both data modeling and task-based modeling are systems identification problems, and the DL approach to solving it is to use a parameterized dynamical system, called a recurrent neural network (RNN), denoted by

dxdt=Rθ(x(t),u(t)), 2.

where Rθ is the RNN with parameters θ. Here the state variables, x(t), take on continuous values, and thus Equation 2 is often referred to as a rate network (as opposed to a spiking network). RNNs typically have an explicitly defined linear readout, expressed as

y(t)=Cx(t), 3.

which states that the output, y(t), is a linear transformation of the state variables, x(t), through the readout matrix C. RNNs are universal approximators, which means they can approximate any f with high accuracy, whether f is defined by a large set of neural recordings or by task-based input-output pairs.

For example, imagine we wish to learn the dynamics of the pendulum, so f again is the equation of motion for the pendulum. To accomplish this, we collect a large set of initial conditions (the initial positions and velocities and a set of 2D inputs) and corresponding time series of the pendulum in operation (position and velocity state vectors through time) as training examples. We define the loss function to be mean squared error, averaged across time, of the RNN’s readouts y(t) relative to those measured positions and velocities. After training, Rθ should have dynamics such that y(t) very closely matches the measured position and velocity traces, and thus Rθ approximates f.

For neuroscientists, f is a biological neural circuit. This means f is likely expensive to sample, hard to run quickly, and challenging to run with novel inputs or under novel conditions. Furthermore, we do not know the precise equations and parameters for f, so a pure biophysical modeling approach is likely impractical. Training an RNN to learn f provides a model of the dynamics that can be simulated quickly, with novel inputs and conditions, and which has equations we can study as a proxy for f. This is true for either case where f models data or a task.

2.2. Understanding High-Dimensional Nonlinear Systems Via Learned Models

A key goal of systems neuroscience is to understand how neural computation gives rise to behavior, so it is not sufficient to simply learn data or task models that are inspired by experiments in neuroscience. We also want to understand how such models operate, which poses serious mathematical challenges, as the biological circuits (f) we study are usually nonlinear and high dimensional. Our approach to understanding f is to first model it, for example, with an RNN, and then study that model.

One ubiquitous approach is to ignore the nonlinearity of f and model the system with a linear dynamical system (LDS), L, defined as

dxdt=L(x(t),u(t))=Ax(t)+Bu(t), 4.

where A and B are matrices of learned parameters that map state and input, respectively, to the state update. Modeling f via an LDS is popular because linear models have a rich history of unreasonably good performance across many quantitative fields and because an LDS can be readily understood through well-established linear systems theory. An LDS can only have a few qualitatively different behaviors. It can exhibit decaying dynamics (Figure 3a), expansive dynamics (Figure 3b), oscillations (Figure 3df), or combinations of all of these dynamics in different dimensions (Figure 3c). An LDS can also have marginally stable dynamics if it is finely tuned. For example, an LDS can generate precisely balanced oscillations (Figure 3f) or act as a perfect integrator if an input to the system aligns with the integrating mode (Figure 3g). If one cannot precisely fine-tune the LDS, then the balance of the oscillator or quality of the integration degrades accordingly.

Figure 3.

Figure 3

Linear dynamics and linear approximations of nonlinear systems around fixed points. (a–g) Examples of linearized dynamics around fixed points in a nonlinear system, showing fixed points (large dots) and flow fields (arrows). Panels show (a) an attractor, which is stable in all dimensions; (b) a repeller, which is unstable in all dimensions; and (c) a saddle point, which is a fixed point with both stable and unstable dimensions. Fixed points can have oscillatory dynamics that are both (d) stable and (e) unstable. Linear systems that are finely tuned may have marginally stable oscillations, which neither decay nor expand, or (f) exhibit line attractor dynamics, (g) which are stable (do not move) along the line attractor (blue line) but otherwise show stable attracting dynamics to the line attractor (yellow planes and arrows). (h) A nonlinear system such as a Duffing oscillator is well approximated by linear systems at the three fixed points: The two orange points are well approximated by oscillatory linear dynamics, while the green point is a saddle point, whose dynamics are well approximated by linear dynamics that are both stable and unstable. Note the flow fields given by the red line are not well approximated by linearization, highlighting a failure mode of linearization around fixed points in that good approximations tend to be local.

Linear systems are also useful for understanding RNNs. We do this by employing multiple LDSs, i.e., multiple linearizations of Rθ, tiling the RNN’s state space with LDSs. Each LDS is an approximately correct model of the RNN in its particular region of state space. This approach decomposes understanding nonlinear neural dynamics into three additional steps: identifying (a) where in state space to linearize, (b) the local LDSs, and (c) when to switch between these LDSs (e.g., Sussillo & Barak 2013). Through these linear approximations we understand Rθ, and through it f. Models that switch between multiple LDSs (SLDS) have a rich history and also may serve as stand-alone models of f (e.g., Linderman et al. 2017, Petreska et al. 2011).

The most natural place to linearize in state space is around a fixed point, which we denote with x*. For simplicity, we define fixed points while ignoring inputs, i.e., dx/dt = Rθ (x(t)), which again approximates dx/dt = f (x(t)). Fixed points are points in state space where repeated iteration of Rθ does not change the state, i.e., dx/dt = Rθ (x*(t)) = 0 (see Figure 3a). Fixed points are interesting for two reasons. First, stable fixed points are the simplest model of memory, as in Figure 3a or Figure 3d, since at those points the state is unchanged by repeated iteration of the dynamical equations, and perturbations away from the fixed points collapse back to the fixed points. Second, and more importantly for this review, the dynamics around fixed points are known to be approximately linear. The fixed points and corresponding linearizations of an RNN’s dynamics can be identified via numerical optimizations (Golub & Sussillo 2018, Sussillo & Barak 2013). Empirically, linearizing around fixed points is proving an effective means to interpret local dynamics of RNNs engaged in computation through neural population dynamics (Barak 2017).

Linearizing around fixed points yields LDSs of the form

d(δx)dt=A(x*)δx(t), 5.

where δx(t) = x(t) − x*, which simply states that the linear system and state dynamics are now centered around the fixed point, x*. Furthermore, the linear dynamics, as expressed by the matrix A, are now a function of x* and are defined as the first derivative of Rθ with respect to x, evaluated at the fixed point, i.e., A(x*) = ∂Rθ /∂x(x*). If there are multiple fixed points in state space, typically one chooses to expand around the fixed point closest to the dynamics of interest. Both the utility and limitations of linearization for understanding a nonlinear system can be seen by considering the Duffing oscillator (Figure 3h).

In the general case, we must also consider the input u(t) and how to linearize around it as well. Accounting for inputs, we end up with the equation

d(δx)dt=A(x*,u*)δx(t)+B(x*,u*)δu(t), 6.

where u* is the linearization point for the input and δu(t) = u(t) − u*. Careful choices of u* open up very interesting possibilities for studying contextual inputs and contextual dynamics. Contextual inputs are typically slowly varying inputs that are not considered input to the computation per se but rather shift the nature of the computation by changing the dynamics (Figure 4). The contextual input changes the dynamics by moving the average operating point in state space (A and B in Equation 6 are a function of u*). As a result, these contextual inputs can change both the recurrent dynamics and also the system response to other dimensions of the input (e.g., sensory input dimensions).

Figure 4.

Figure 4

Contextual inputs. A static or slowly varying input can be viewed as contextualizing a nonlinear system such that it produces different dynamics under different values of the contextual input. In this diagram, increasing the value of a static input (from purple for small values to cyan for large values) moves the operating point of the nonlinear system (circles) such that the dynamics smoothly change from a diamond-like oscillation (purple arrows) to a circular oscillation (cyan arrows).

2.3. Linear Subspaces and Manifolds

Here we define three mathematical concepts that are relevant for dynamics in high-dimensional data, including neuroscientific data. See Figure 5 for illustrative examples.

Figure 5.

Figure 5

Subspaces and manifolds. (a) Two-dimensional (2D) linear subspace (blue) embedded in a three-dimensional (3D) space. (b) One-dimensional (1D) ring manifold (green) embedded in 3D space. (c) A smooth 2D manifold (orange) embedded in 3D space. (d,e) Dynamics may live in subspaces or on manifolds. (d) A 2D oscillation embedded in a 2D linear subspace, all of which is embedded in 3D space. (e) A ring attractor, which is a 1D manifold of stable fixed points (blue ring) embedded in 3D space. Inputs move the state along the ring attractor (blue arrows; note arrows here indicate the effect of inputs, not flow fields).

2.3.1. Linear subspace.

A linear subspace is a hyperplane that passes through the origin of a higher-dimensional space. For example, in 3D space, a 2D plane passing through the origin is a linear subspace (Figure 5a). Linear subspaces play a key role in dimensionality reduction, whereby measured ND neural states that empirically lie near a P-dimensional linear subspace (where P < N) are summarized and often denoised by a P-dimensional representation within the linear subspace (Cunningham & Yu 2014).

2.3.2. Null space.

The null space of a linear system is the space of all inputs to that system that result in a zero output. For a linear system defined by matrix A, an input vector x is in the null space of A if Ax = 0. A geometrical illustration is provided in Figure 6. In systems neuroscience, the communication from an upstream brain region to a downstream region can be modeled as a linear system, y = Ax, and the null space of that system is the set of upstream-region neural states, x, that do not directly contribute to the downstream-region activity (with neural states y). This concept has provided a neural population mechanism by which signals can be flexibly communicated within and across brain areas (Semedo et al. 2019); an example of this is how the motor cortex can prepare a movement without its preparatory activity causing premature movement (Kaufman et al. 2014).

Figure 6.

Figure 6

Null and potent subspaces. A state space can be divided up into nonoverlapping subspaces, called the potent and null (sub)spaces, where the potent space may be read out by a downstream area, while the null space is not read out. (a) A dynamical trajectory (blue arrows) in state space follows a looping dynamic from the origin back to the origin. While doing so, the trajectory has projections on the null space (gray line) and potent space (purple line). We focus on four points in time, t1, t2, t3, and t4 (red arrows). (b) The state space trajectory in panel a can be decomposed into its projections onto the potent and null spaces (shown for time points t1, t2, t3, and t4). (c) The projection of the state space trajectory onto the potent dimension through time (yellow line), highlighting its value at time points t1, t2, t3, and t4. There is a corresponding time series for the projection onto the null space (not shown), but this is not read out by a downstream area. Examples of this in systems neuroscience can be found in these references (Hennig et al. 2018, Kaufman et al. 2014, Perich et al. 2018, Semedo et al. 2019, Stavisky et al. 2017a).

2.3.3. Manifold.

A manifold, like a linear subspace, is a low-dimensional topological space embedded within a higher-dimensional space. Unlike a linear subspace, the more-general manifold can bend about as long as it remains linear when zoomed in on closely enough. For example, if we zoomed in closely enough on the one-dimensional (1D) ring manifold shown in Figure 5b, the curvature of the ring would disappear, and it would appear as a line. The same is true of the 2D manifold in Figure 5c, which is a warped plane embedded in a 3D space. Dynamical systems can generate states that are confined to manifolds. For example, in Figure 5d, an unstable oscillation is confined to a 2D plane within a 3D space. Another example is the ring attractor, a manifold of fixed points that lie along a 1D ring, whereby neural states move around the ring as a result of the system integrating inputs (Figure 5e). A number of neuroscientific studies have derived insights into neural systems by leveraging manifold structures identified from neurophysiological data (e.g., Chaudhuri et al. 2019, Kim et al. 2017, Sadtler et al. 2014; for a review, see Gallego et al. 2017).

3. DYNAMICAL SYSTEMS IN MOTOR CONTROL

Much of our understanding of the neural control of movement stems from the pioneering work of Evarts (Evarts 1964, 1968; Tanji & Evarts 1976), who recorded the activity of individual neurons from awake behaving monkeys and related it to muscle activation. Investigators have since employed other electrophysiological (single- and multielectrode arrays) and optical (calcium imaging and optogenetics) techniques to investigate neural correlates underlying behavior (e.g., Kalaska 2019, O’Shea et al. 2017). Early investigations, following work in primary visual cortex (V1) (Hubel & Wiesel 1962), attempted to characterize response properties of motor cortical neurons. There was success in correlating responses of individual neurons to task parameters (Georgopoulos et al. 1982, 1986, 1989). However, unlike V1, there was little consensus regarding what parameters motor cortical activity represented—some favored representation of higher-level movement parameters (e.g., end-effector velocity), while others favored the view that activity drove muscle activity more directly. Here we ask a different but related question: What principles guide generation of the observed time-varying patterns of motor cortical activity (Churchland et al. 2012, Shenoy et al. 2013)?

3.1. Motor Preparation: First Cog in a Dynamical Machine

One common experimental paradigm used in laboratory settings is the instructed-delay reaching task (Ames et al. 2014). The position of the animal’s arm in physical space controls the position of a cursor on a screen. Animals are trained to make rapid and accurate reaching arm movements as targets appear on the screen. Since movements are prepared before they are executed (Riehle & Requin 1989, Rosenbaum 1980, Wise 1985), this preparatory process is probed by requiring that animals wait a randomly distributed and experimentally controlled amount of time before initiating a reach. During this delay period, animals are cued to the target that they must ultimately reach to, and thus the measured neural activity is typically referred to as delay-period or preparatory activity. Previous work has shown that animals use this delay period to perform behaviorally relevant computations: More time to prepare reduces reaction times, and intracortical microstimulation during the delay period causes a reaction time penalty (Churchland & Shenoy 2007a, Churchland et al. 2006b). Furthermore, neural activity during the delay period is tuned to distance, direction, and speed, among other parameters (Churchland et al. 2006b; Even-Chen et al. 2019; Messier & Kalaska 1997, 1999; Santhanam et al. 2009; Yu et al. 2009). During this task, neural measurements are typically made in contralateral-to-arm primary motor (M1) and dorsal premotor cortex (PMd), though comparisons between ipsilateral and contralateral regions have also been investigated (Ames & Churchland 2019, Cisek et al. 2003).

From the CTD perspective, preparatory activity is viewed as an initial condition. It brings the population neural state to an initial value from which movement period activity evolves, governed by both the local dynamics of motor cortex as well as inputs from other brain regions. This predicts that different movements have different preparatory states, though different initial conditions can result in qualitatively similar movements (e.g., with different reaction times) (Churchland et al. 2006c, 2010, 2012; Michaels et al. 2016; Sussillo et al. 2015). Nonetheless, initial conditions closer to the optimal preparatory state have been shown to confer behavioral benefits, e.g., faster reaction times (Afshar et al. 2011, Ames et al. 2014). This initial condition hypothesis is an alternative to a representational hypothesis, which suggests that preparatory activity is a subthreshold version of movement-period activity. By itself, preparatory activity is not sufficient to trigger movement but could shorten time needed to rise above threshold and trigger movement, thus resulting in shorter reaction times (Bastian et al. 1998, 2003). At present, evidence appears to favor the initial condition hypothesis (Churchland et al. 2006c, Shenoy et al. 2013).

Other competing hypotheses argue preparatory activity is suppressive rather than facilitatory, i.e., it functions to inhibit movement (Wong et al. 2015), or that preparatory activity, while facilitatory, is not essential (Ames et al. 2014) or exhibits large variability (Michaels et al. 2015). In particular, Ames et al. (2014) showed that the preparatory state could be bypassed in trials with no delay period, though many aspects of the subsequent population response were conserved across both delay and nondelay contexts. Perfiliev et al. (2010) recently showed that using moving targets, in lieu of fixed ones, exhibited low latency reaches. Lara et al. (2018b) adapted this task design to test what aspects of preparatory activity are present in the absence of the delay period. To disambiguate whether the delay period is suppressive or preparatory, the authors also trained animals in a self-initiated reach context where animals were free to initiate movement at any point after target onset. Lara et al. concluded that delay period activity does indeed reflect a preparatory process that occurs even in the absence of a delay. However, this process need not be time-consuming; when necessary, it can rapidly set up an appropriate preparatory state prior to movement onset. These results dovetail with RNN models trained to produce muscle activity using preparatory activity as initial conditions (Hennequin et al. 2014, Sussillo et al. 2015). The dynamics in these models show surprising similarity to those measured in motor cortex of behaving animals. Recent results suggest that this same preparatory process can also be used during reach correction, for example, when a new reach target appears at a new location while moving toward an initial target (Ames et al. 2019). Finally, Inagaki et al. (2019) found that delay period neural activity in mouse anterior lateral motor cortex moves toward stable fixed points that correspond to movement directions. Critically, these fixed points are robust to optogenetic perturbation; occasionally activity hops to the other fixed point following perturbation, producing incorrect actions. This serves as a compelling demonstration that discrete attractor dynamics underlie delay period activity.

Since neurons are preferentially active during both preparation and movement, how does this activity avoid causing movement? Classic theories suggested that preparatory activity was subthreshold (Tanji & Evarts 1976) or gated (Cisek 2006). However, Churchland et al. (2010) showed there was limited correlation in preferred directions between preparation and movement, and Kaufman et al. (2010, 2013) analyzed neuron classes during preparation and did not find a bias toward inhibitory cells. The CTD framework provides a computational scheme that can explain this phenomenon through the concept of null spaces (Figure 6). If neural activity during preparation evolved along neural dimensions that were orthogonal to dimensions explored during movement, then preparatory activity would avoid causing movement prematurely. Kaufman et al. (2014) together with Elsayed et al. (2016) found evidence for such a mechanism in motor cortex: Output-potent dimensions presumably relay information to muscles and are preferentially active during movement, and output-null dimensions are preferentially active during preparation. Recently, Stavisky et al. (2017a) provided a test of the more general theory that orthogonal dimensions can be used for flexible computation by using a brain-computer interface (BCI) where the output dimensions could be experimentally controlled. They showed that sensory feedback relating to error signals is initially isolated in decoder-null rather than decoder-potent dimensions. Li et al. (2016) showed that following optogenetic perturbation, preparatory activity in mouse anterior lateral motor cortex was restored along behaviorally relevant neural dimensions but not along other noninformative dimensions (Druckmann & Chklovskii 2012, Li et al. 2016). Despite these studies, there may be constraints to the role that these output-null dimensions play (Hennig et al. 2018).

3.2. Dynamical Motifs During Movement

During arm movements, the largest component of the population response is condition-invariant and reflects the timing of movement. This component, termed the condition-independent signal (CIS), appears following the onset of the go cue and causes a substantial change in dynamics (Kaufman et al. 2016). Such a CIS has also been observed during reach-to-grasp movements (Rouse & Schieber 2018). While the precise role of the CIS is still not known, presumably following a well-timed input into motor cortex, it transitions the preparatory state to another region in neural state space. Functionally, the translation in state space releases the preparatory state from an attractor-like region to exhibit rotation-like dynamics (Churchland et al. 2012, Michaels et al. 2016, Pandarinath et al. 2015). These rotations, while only an approximation to the true nonlinear dynamics, explain a large fraction of variance. Recently, Stavisky et al. (2019) found rotatory dynamics that followed a condition-invariant component in hand/knob motor cortex of humans during speech, suggesting common dynamical principles underlying control of arm and speech articulators. The significance of these rotational dynamics is that they appear to correspond to neither purely representational parameters nor purely movement parameters. Rather, they can be interpreted as a basis set for building complex time-varying patterns, like those required for generating muscle activity. Recently, Sauerbrei et al. (2020) demonstrated that cortical pattern generation during dexterous movement requires continuous input for the thalamus. Thus, future work will need to disentangle the relative roles and contribution of inputs and local dynamics in cortical pattern generation.

As a sufficiency test for the rotations motif, Sussillo et al. (2015) trained RNNs to transform simple preparatory activity–like inputs into complex, spatiotemporally varying patterns of muscle activity. They found that when networks were optimized to be simple, the underlying solution was a low-dimensional oscillator that generated the necessary multiphasic patterns. This highlights an additional, perhaps underappreciated, aspect of using RNNs in systems neuroscience: hypothesis generation. While not done here, one could have performed this study even before that of Churchland et al. (2012) and committed to a quantitatively rigorous hypothesis.

The CTD framework was born out of the hope that relatively simple population-level phenomena might explain seemingly complex single-neuron responses (Churchland & Shenoy 2007b, Churchland et al. 2012, Shenoy et al. 2013). Dynamical models often succeed in this goal. For example, RNNs trained by Sussillo et al. (2015) reproduced both single-neuron responses and population-level features (e.g., rotational dynamics) despite not being fit to neural data. While it is possible to obtain rotation-like features from non-dynamical models [e.g., sequence-like responses (Lebedev et al. 2019)], such rotations are dissimilar from those observed empirically (Michaels et al. 2016). Perhaps more critically, the hypothesis of sequence-like responses does not account for single-neuron responses, which are typically multiphasic and/or peak at different times for different conditions (e.g., Churchland et al. 2012, figure 2). Given such complex responses, it is reasonable to ask whether the presence of apparently dynamical patterns (e.g., rotations) is inevitable at the population level. In fact, rotational dynamics are neither observed upstream in the supplementary motor area (SMA) (Lara et al. 2018a) nor downstream at the level of muscles. This is true even though complex and multiphasic responses are present in both situations. This highlights an emerging need: determining when population-level observations provide new information beyond single-neuron properties. A method for doing so was recently introduced (Elsayed & Cunningham 2017) and confirms that dynamical structure is not a byproduct of simpler features (e.g., temporal, neural, and condition correlations) within the data. Thus, dynamical models are appealing because training them to perform a hypothesized computation is often sufficient to naturally account for the dominant features of single-neuron and population-level responses.

Neural population responses exhibit a wide variety of patterns. Do all patterns directly guide movements, or are there patterns that only support driving the dynamics themselves? A recent study by Russo et al. (2018) found evidence for the latter. They started by exploring a general phenomenon of dynamical systems: Two very similar neural states can result in very different future behaviors, which is defined as high tangling. Russo et al. found evidence that a large portion of the response during motor generation was used for “untangling.” Untangling neural dynamics is potentially quite important, as even a small amount of noise in a highly tangled state space could radically change the future evolution of the dynamical system. Russo et al. suggested that additional dimensions of neural activity function to reduce tangling and have no further relationship to inputs and outputs of motor cortex. This was further supported by Hall et al. (2014), who found a consistent rotational structure in dominant neural dimensions during wrist movements; precise movement direction was only decodable from the lower-variance neural dimensions.

Russo et al. (2019) recently characterized another geometric property of the neural population response: divergence, i.e., two neural trajectories never follow the same path before diverging. Population activity in SMA, which keeps track of context (e.g., time), must be structured so as to robustly differentiate between actions that are the same now but will soon diverge. Russo et al. (2019) found that population activity in SMA was structured in a fashion to achieve this by keeping divergence low, whereas M1, which does not keep track of context, showed high divergence. They found that RNNs trained to track context naturally had low divergence independent of the exact solution chosen, and they showed a similar geometry to SMA (helical). In contrast, RNNs that were context naïve did not have low divergence, and their state space geometry resembled that of M1 (elliptical).

Taken together, the studies described thus far constitute the core concepts of CTD applied to cortical control of arm movements. In summary, (a) preparatory activity sets an initial condition to a neural dynamical system, where preparatory dimensions are approximately orthogonal to movement dimensions; (b) presumably a well-timed input from other brain regions gives rise to a CIS, which transitions the preparatory state to another region in neural state space; (c) rotational dynamics are observed in motor cortex, where a large part of the response during motor generation appears to serve to untangle and support driving the dynamics themselves; (d) movement begins; (e) sensory feedback is isolated in null-dimensions before influencing output-potent feedback corrections; and (f) these dynamics are largely stable across time for a given behavior (Chestek et al. 2007, Flint et al. 2016, Gallego et al. 2020).

3.3. Population Dynamics Underlying Motor Adaptation

To evaluate if and how the core principles change when behaviors are made more complex, we shift our focus from CTD applied to motor pattern generation to CTD applied to motor adaptation. We start by discussing error-driven learning, which has a rich history in human psychophysics (for a review, see Krakauer et al. 2019). Here we focus on two tasks: visuomotor rotation (VMR) adaptation and curl force-field (CF) adaptation. During VMR adaptation, when a subject moves her arm in space, the cursor movement is offset by an angle that is experimentally controlled (e.g., 45° counterclockwise). In order to adapt, subjects must learn to move their arm in the anti-VMR direction; for example, for a 45° counterclockwise VMR, they must move their arm in a 45° clockwise direction to cause the cursor to move directly to the target.

The CTD framework predicts that adaptation should largely affect preparatory dimensions, as setting a new initial condition is sufficient to overcome the VMR. For example, if a 45° counterclockwise rotation is applied, generating preparatory activity that corresponds to making a reach to a 45° clockwise target would produce the correct behavior. Such changes in preparatory activity must match the timescale of behavioral changes; in humans and monkeys, adaptation happens gradually over the course of several tens of trials. Vyas et al. (2018) found that preparatory states corresponding to making a reach to a particular target systematically rotated across trials to align with the anti-VMR direction in neural state space. The preparatory state on single trials was highly correlated with behavioral changes on those trials. Similar changes have also been observed for other visuomotor manipulations [e.g., gain adaptation (Stavisky et al. 2017b)]. A hallmark of motor adaptation is washout, i.e., once the visuomotor perturbation is removed, subjects make errors in the opposite direction, compared to learning. Vyas et al. exploited this fact to show that VMR learning can transfer between overt reaching and covert movement (i.e., cursor movement guided by BCIs) contexts. Their analyses revealed that the preparatory subspace is shared between both contexts, and thus changes that happen in one context persist across context changes. Moreover, covert movements and overt reaching engage a similar neural dynamical system, despite covert context requiring no muscle activity production.

Given this potentially critical role of preparatory activity during learning, Vyas et al. (2020) investigated the causal role between motor preparation and VMR adaptation. Vyas et al. found that disrupting the preparatory state via intracortical microstimulation did not affect stimulated trials. Instead, error (and thus learning) on subsequent trials was disrupted. Through a series of causal experiments, the authors concluded that preparatory activity interacts with the update computation of the trial-by-trial learning process to produce effects that are dose, brain-region, and temporally specific. Moreover, their results, consistent with Bayesian estimation theory, argued that the learning process reduces its learning rate, by virtue of lowering its sensitivity to error, when faced with uncertainty introduced via stimulation (Vyas et al. 2020, Wei & Körding 2009).

Another common task that has recently been investigated is CF adaptation. Here subjects are required to learn new arm dynamics as they experience large velocity-dependent forces in an experimenter-defined direction (Cherian et al. 2013, Thoroughman & Shadmehr 1999). Sheahan et al. (2016) developed a two-sequence task where planning and execution could be separated. They found that planning, and not execution, allowed human participants to learn two opposing CF perturbations that would otherwise interfere. The implication is that neural activity during motor preparation can function, at least partly, as a contextual input, which the motor system leverages to separate different movements. Perich et al. (2018) studied the neural correlates of CF learning in PMd and M1 and found a surprising degree of stability in the correlation structure (i.e., functional connectivity) both within and across areas. While activity changed systematically with adaptation in the null space (of the map from PMd to M1), surprisingly, the output potent map was preserved. This perhaps functions to preserve the communication link between PMd and M1 during learning, with the advantageous feature that it does not require dramatic changes to the pattern-generation circuitry. As shown by Stavisky et al. (2017a) for error processing, this result serves as yet another example of where orthogonal subspaces are a general computational motif, which the brain can exploit for flexible and parallel computation. Taken together, these studies open the door to applying the computation through dynamics framework to study the neural control of error-driven motor learning.

3.4. Implications of Manifold Structure on Learning

Many studies have reported that neural population dynamics tend to lie within a low-dimensional subspace (or manifold) embedded within the high-dimensional neural population state space (Cunningham & Yu 2014). In this section, we review studies developed to understand the implications of these manifolds on the flexibility of neural population dynamics. Causally attributing changes in neural population activity to changes in behavior is a notoriously difficult problem in systems neuroscience (Jazayeri & Afraz 2017, Otchy et al. 2015). BCIs have emerged as a powerful experimental paradigm for addressing this challenge (Golub et al. 2016). A BCI maps the activity of a set of recorded neurons into control signals, which are used to drive movements of a computer cursor (Shenoy & Carmena 2014). This setup is powerful for studying learning because the experimenter has complete knowledge of and ability to manipulate the mapping between neural activity and behavior (Athlaye et al. 2017, Chase et al. 2012, Jarosiewicz et al. 2008, Orsborn et al. 2014, Sakellaridi et al. 2019, Zhou et al. 2019).

Sadtler et al. (2014) leveraged this feature of the BCI paradigm in an experiment in which monkeys learned to make BCI cursor movements under various perturbations of this mapping. Each perturbation was designed relative to the linear subspace, termed the intrinsic manifold, that encompassed the movement-period M1 dynamics that animals generated when making nonperturbed BCI cursor movements. The key result was that animals could readily learn (within a few hours) perturbations that could be solved using neural dynamics that lie within the intrinsic manifold, but they could not as readily learn perturbations that required outside-manifold dynamics. Thus, these observed neural manifolds likely reflect intrinsic constraints of cortical circuits, perhaps governed by inherent network connectivity, which limits the regions of neural state space through which neural population dynamics may explore.

When learning did occur (within-manifold perturbations, WMPs), what was the guiding neural population mechanism? Given the constraint that neural population dynamics remain within the intrinsic manifold, each WMP prescribes a behaviorally optimal set of neural states that would most swiftly drive the cursor toward each target, while remaining within the intrinsic manifold and respecting physiological constraints on individual neuron firing rates. Golub et al. (2018) identified these optimal states and compared them to neural states generated by animals after learning. Interestingly, animals did not learn these optimal states but rather learned systematically according to an even more stringent constraint. The learned population dynamics predominantly visited neural states that were within animals’ prelearning neural repertoire, defined as the set of neural states produced during an initial block of nonperturbed trials in the BCI task. Although this neural repertoire remained fixed, through learning animals changed associations between the neural states in that repertoire and corresponding intended movements. This mechanism of learning by neural reassociation thus explained animals’ postlearning neural states as well as levels of behavioral performance attained.

Could animals have better learned the outside-manifold perturbations if given multiple days to do so? Oby et al. (2019) demonstrated that this was indeed the case. Here, learned neural population dynamics were shown to drive behavioral improvements due to them visiting new neural population states of two kinds: (a) within the prelearning manifold but outside the prelearning repertoire, and (b) entirely outside the prelearning manifold (and repertoire). Thus, there appear to be multiple learning mechanisms operating at different timescales. A fast-timescale mechanism reassociates existing neural states with new intended behaviors (Golub et al. 2018), and slower-timescale mechanisms drive out-of-repertoire and outside-manifold changes to the neural population dynamics (Oby et al. 2019).

How are these different mechanisms implemented? Changes to the observed M1 population dynamics could result from changing initial conditions in M1 (Figure 2d), changing the inputs received by M1 (Figure 4), or changing the network connectivity within M1. Measuring these types of changes in vivo poses tremendous technical challenges. However, as in previous examples discussed, important insight can be derived from RNN modeling approaches that follow the CTD framework. Wärnberg & Kumar (2019) showed that, using three different classes of RNN models, substantial out-of-repertoire and outside-manifold RNN dynamics [as observed by Oby et al. (2019)] required substantial changes to the RNNs’ synaptic weights. Within-repertoire changes (Golub et al. 2018) were possible with either changes to inputs or comparatively smaller changes to the synaptic weights.

4. DYNAMICAL SYSTEMS IN COGNITION

Animals display the ability to volitionally modulate their behavior in response to contextual cues, sensory feedback, or environmental changes. Here we highlight studies that have applied the CTD framework to investigate neural computations underlying such flexible executive control.

4.1. Motor Timing

We start by reviewing studies that focus on uncovering neural mechanisms underlying an animal’s ability to flexibly control their movement initiation and/or duration time (Mauk & Buonomano 2004, Remington et al. 2018a). An early model of motor timing described a rise-to-threshold mechanism (Jazayeri & Shadlen 2015, Killeen & Fetterman 1988, Meck 1996, Treisman 1963). This model was derived from evidence that neurons exhibited ramping activity, triggering movement when a threshold was reached (e.g., Romo & Schultz 1992). This is also consistent with a central clock model where neural circuits trigger movement once a fixed number of ticks have been accumulated.

Ramping signals are not the only mechanism by which neural circuits can keep track of time. Recent modeling work has proposed state-dependent networks, which inherently keep track of time as they are composed of units that have time-dependent biological properties, and thus evolving states of these networks encode the passage of time (Buonomano 2000, Buonomano & Maass 2009, Karmarkar & Buonomano 2007, Maass et al. 2002). In such models, temporally patterned stimuli are encoded as a whole rather than as individual interval components. Furthermore, states are expanded to include not only population firing rates but also instantaneous membrane potentials and synaptic strengths. Interval sensitivity is thus established as a nonlinear function of connectivity, short-term plasticity, and synaptic strength. Intriguingly, such models degrade if initial conditions have noise or are otherwise perturbed. Karmarkar & Buonomano (2007) demonstrated that this is indeed the case: In both models and human psychophysics, performance is disrupted when distractors are presented at random times.

Recent experimental evidence suggests that another strategy for flexible motor timing involves controlling the speed of neural state evolution. Such control is achieved by manipulating inputs and/or initial conditions of the dynamical system. Wang et al. (2018) studied animals performing a cue-set-go task where animals were provided a context cue at trial-onset that indicated the timing at which they had to initiate movement. Temporal profiles of the neural population response in dorsomedial frontal cortex (DMFC) were scaled according to the production time. Neural trajectories in a low-dimensional state space were structured such that different production times were associated with an adjustment of neural trajectory speed. Wang et al. hypothesized that this structure might be a consequence of thalamocortical inputs into DMFC. They found that thalamic neurons provided approximately tonic input into DMFC with activity level that was correlated with the set-go interval. This suggests that thalamic neurons can manipulate their tonic activity level in order to influence dynamics of DMFC, thus contextualizing dynamics of DMFC in order to control neural trajectory speed. They tested this prediction in models by training RNNs to perform the same task and found a similarity to measured electrophysiology; intervals of different durations corresponded to similar trajectories in state space moving at different speeds. In a related study, Egger et al. (2019) studied a 1-2-3-go task where animals had to estimate the time interval between 1–2 and 2–3 epochs before initiating a movement at the anticipated 3–4 epoch, labeled as 3-go. The authors found that the speed command was updated based on the error between the timing measured during the 2–3 epoch and the estimated timing derived from the simulated motor plan based on the 1–2 epoch. These results suggest that speed modulation in DMFC can be understood in terms of similar computational principles such as closed-loop feedback control (Egger et al. 2019).

Remington et al. (2018b) studied a ready-set-go task where animals estimated a time interval during the ready-set epoch and then, depending on a context cue, scaled their estimate by a gain factor of 1 or 1.5 before making a movement at the appropriate time during the set-go epoch. The authors found that neural trajectories in DMFC separated along two dimensions: one associated with gain and one with the interval. Within a single context, speed modulation during the set-go epoch was governed by selecting appropriate initial conditions in the ready-set epoch. Across contexts, there was a gain-dependent tonic input into DMFC, which resulted in two groups of isomorphic neural trajectories that separated along a dimension in state space. The authors provided further evidence for these findings by training distinct RNNs, one where the gain came in as a transient input, thus only affecting the initial condition, and one where the gain was a tonic input. RNNs with gain as a tonic (or contextual) input more faithfully matched electrophysiology.

4.2. Decision-Making and Working Memory

One of the first studies to incorporate electrophysiology from behaving monkeys in a decision-making task with sophisticated neural network modeling was performed by Mante et al. (2013). The authors recorded from prefrontal cortex as animals performed a context-dependent variant of a classic decision-making task. Animals observed randomly moving dots, with each dot randomly colored red or green. During the color context, animals had to ignore motion and report the dominant color of the dots. During the motion context, animals had to ignore color and report the dominant motion direction of the dots. Surprisingly, context dependence was neither due to irrelevant inputs being blocked nor due to the decision axis differing between contexts. Using RNN modeling, the authors found that the decision axis in their networks was a line attractor: a linear arrangement of stable fixed points with end points corresponding to converged decisions (Figure 3c). Moreover, inputs corresponding to the irrelevant context decayed quickly, whereas context-dependent inputs were longer lasting and caused desirable perturbations to neural trajectory evolution. Context dependence was not a result of traditional gating but rather of dynamics quenching activity along certain directions in state space, achieved via a contextually dependent state space shift.

Chaisangmongkon et al. (2017) also employed RNNs to study neural mechanisms underlying delayed match-to-category (DMC) tasks. Animals were tasked to report if a particular test stimulus was a categorical match to a previously presented sample stimulus; animals were pre-trained to differentiate between categories. The authors trained RNNs to perform DMC tasks and confirmed that behavior and neural activity strongly resembled that of animals performing the task (Freedman & Assad 2006, Swaminathan & Freedman 2012). RNN dynamics resembled slow and stable states connected by robust trajectory tunnels. During working memory, these tunnels functioned to force neural states to flow along two possible routes (corresponding to the sample categories), thus maintaining memory in a dynamic manner. Misclassification can thus be interpreted as noise pushing the state outside the correct tunnel and into the tunnel corresponding to the other category. By analyzing the state space geometry, Chaisangmongkon et al. found that RNNs solved the task by strategically navigating around attracting fixed points and unstable saddle points through slow and reliable transitions. That is, RNNs started with rest states, moving to those corresponding to sample categories, then to category working memory, then sample-test category combination, and finally to decision states. This study put forth the idea that delayed association in biological circuits can be performed using such transient dynamics. The study also serves as yet another example of how different regions of state space can subserve different computational roles (Churchland et al. 2012; Hennequin et al. 2014; Mante et al. 2013; Murakami & Mainen 2015; Rabinovich et al. 2001, 2008; Raposo et al. 2014). Other studies spanning working memory and its representations are reviewed in depth by Chaudhuri & Fiete (2016) and Miller et al. (2018).

A critical aspect of flexible decision-making involves integrating sensory evidence with prior information. Sohn et al. (2019) used a time-interval reproduction task and found evidence for Bayesian integration in neural circuits. Priors resembled low-dimensional curved manifolds across the population and warped underlying neural representations in a manner such that behavior was Bayes-optimal, i.e., responses were biased consistently with priors. The authors performed a causal test of the curved manifold hypothesis in models by perturbing RNN activity in a manner that revealed that curvature induced regression toward the mean of the prior. Such causal tests might be an important stopgap before similar perturbations can be performed in vivo.

In all of the studies reviewed thus far, RNNs are typically trained and optimized to perform a single task, which is certainly a poor approximation to biological networks that can solve many tasks. Yang et al. (2019) trained RNNs to perform 20 tasks commonly studied in systems neuroscience. The key discovery in this study was that RNNs developed compositionality of task representation, i.e., one task could be performed by recombining instructions for other tasks, and groups of neurons became specialized for specific cognitive processes. At least at a high level, this is a critical feature of biological systems. We believe future work along these lines will be critical, as biological neural circuits are rarely optimized to only perform single or a small handful of tasks. To aid in such efforts, Song et al. (2016) recently released a flexible framework for training RNNs for cognitive tasks. These authors have also made strides in developing models that do not simply rely on graded error signals to train RNNs. Animals, for example, often learn by being rewarded on a trial-by-trail basis, where elements of confidence, past performance, and value judgement are critical to the underlying computations, even if not apparent in the motor action. The authors developed a set of tools for reward-based training of RNNs, which match behavioral and electrophysiological recordings from an array of experimental paradigms and thus might serve as a powerful framework for studying value-based neural computations (Song et al. 2017).

5. CHALLENGES AND OPPORTUNITIES

The CTD framework has been successfully applied to study the effect of inputs, initial conditions, and neural population geometry on behavior. Many challenging questions, however, still remain unanswered. For example, what does the input-output structure look like for a particular brain region? Where do the inputs come from? How do they shape the local dynamics of that brain region? How does this organization inform the overall function of the circuit? While commendable progress has been made toward characterizing inputs (e.g., Wang et al. 2018), how activity from one brain region affects another (e.g., Semedo et al. 2019), or how population covariance structure changes across tasks (e.g., Gallego et al. 2018), further headway will likely require recording neural activity from many more of the brain regions involved in a given task. Fortunately, such progress is already underway. Advances in two-photon calcium imaging in larval zebrafish (Ahrens et al. 2012, Andalman et al. 2019, Naumann et al. 2016, Portugues et al. 2014), Drosophila (Kim et al. 2017), rodents (Allen et al. 2019, Chaudhuri et al. 2019, Stringer et al. 2019), and Caenorhabditis elegans (Kato et al. 2015, Nichols et al. 2017), as well as high channel count electrophysiology via NeuroPixels (e.g., Jun et al. 2017, Stringer et al. 2019, Trautmann et al. 2019), have enabled measuring and manipulating neural activity at unprecedented scale. We believe these technologies will present new opportunities to test the CTD framework.

The CTD framework makes strong predictions about neural dynamics under novel conditions. While many of these predictions expose themselves to falsification, it has been difficult to falsify them empirically. The challenge is nontrivial; testing CTD models requires a high degree of control over neural activity. The experimenter must be able to manipulate neural states arbitrarily in state space and then measure behavioral and neural consequences of such perturbations, as described by Shenoy (Adamantidis et al. 2015). Nonetheless, a burgeoning body of recent research on using optogenetics to spatiotemporally pattern neural activity in visual cortex has demonstrated that experimental control over population activity is indeed possible (e.g., Bashivan et al. 2019, Carrillo-Reid et al. 2019, Chettih & Harvey 2019, Marshel et al. 2019, Yang et al. 2018). Perhaps the time is fast approaching when such technology can interface with predictions made by CTD. In our view, this could potentially serve as a powerful and causal test of the CTD framework.

In closing, application of the CTD framework has uncovered several invariant motifs concealed behind complex neural population activity. Studying these motifs may help establish rigorous computational theories linking brain and behavior. We believe that meaningful progress in this branch of systems neuroscience likely requires three steps: (a) continued development of the CTD framework and other kinds of quantitative models that make empirically testable predictions, (b) further research into methodological approaches toward causal experiments, and (c) providing concrete strategies for updating models once new measurements are made.

ACKNOWLEDGMENTS

We apologize in advance to all the investigators whose research could not be appropriately cited owing to space limitations. We thank Dr. Ryan J. McCarty (UC Irvine) for graphics assistance and feedback and Beverly Davis for administrative support. This work was supported by a National Science Foundation Graduate Fellowship, a Ric Weiland Stanford Graduate Fellowship, a National Institutes of Health (NIH) F31 fellowship (5F31NS103409 to S.V.); an NIH-NIMH K99/R00 award (K99MH121533 to M.D.G.); an Office of Naval Research Award (N000141812158 to M.D.G., D.S., and K.V.S.); a Simons Foundation Collaboration on the Global Brain award (543045 to D.S. and K.V.S.); and the following to K.V.S.: a DARPA BTO NeuroFAST award (W911NF-14-2-0013), an NIH-NINDS award (UO1 NS098968), an NIH-NIDCD R01 (DC014034), an NIH-NINDS UH2/3 (NS095548), and an NIH-NIDCD U01 (DC017844). K.V.S. is a Howard Hughes Medical Institute investigator.

DISCLOSURE STATEMENT

K.V.S. is on the scientific advisory boards of CTRL-Labs Inc., Heal Inc., MindX Inc., and Inscopix Inc. and is a consultant to Neuralink Inc. These entities in no way influenced or supported this work.

LITERATURE CITED

  1. Adamantidis A, Arber S, Bains JS, Bamberg E, Bonci A, et al. 2015. Optogenetics: 10 years after ChR2 in neurons—views from the community. Nat. Neurosci 18:1202–12 [DOI] [PubMed] [Google Scholar]
  2. Afshar A, Santhanam G, Yu BM, Ryu SI, Sahani M, Shenoy KV. 2011. Single-trial neural correlates of arm movement preparation. Neuron 71(3):555–64 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Ahrens MB, Li JM, Orger MB, Robson DN, Schier AF, et al. 2012. Brain-wide neuronal dynamics during motor adaptation in zebrafish. Nature 485(7399):471–77 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Allen WE, Chen MZ, Pichamoorthy N, Tien RH, Pachitariu M, et al. 2019. Thirst regulates motivated behavior through modulation of brainwide neural population dynamics. Science 364(6437):eaav3932. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Ames KC, Churchland MM. 2019. Motor cortex signals for each arm are mixed across hemispheres and neurons yet partitioned within the population response. eLife 8:e46159. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Ames KC, Ryu SI, Shenoy KV. 2014. Neural dynamics of reaching following incorrect or absent motor preparation. Neuron 81(2):438–51 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Ames KC, Ryu SI, Shenoy KV. 2019. Simultaneous motor preparation and execution in a last-moment reach correction task. Nat. Commun 10(1):2718. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Andalman AS, Burns VM, Lovett-Barron M, Broxton M, Poole B, et al. 2019. Neuronal dynamics regulating brain and behavioral state transitions. Cell 177(4):970–85.e20 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Athlaye VR, Ganguly K, Costa RM, Carmena JM. 2017. Emergence of coordinated neural dynamics underlies neuroprosthetic learning and skillful control. Neuron 93(4):955–970.e5 [DOI] [PubMed] [Google Scholar]
  10. Barak O 2017. Recurrent neural networks as versatile tools of neuroscience research. Curr. Opin. Neurobiol 46:1–6 [DOI] [PubMed] [Google Scholar]
  11. Bashivan P, Kar K, DiCarlo JJ. 2019. Neural population control via deep image synthesis. Science 364(6439):eaav9436. [DOI] [PubMed] [Google Scholar]
  12. Bastian A, Riehle A, Erlhagen W, Schöner G. 1998. Prior information preshapes the population representation of movement direction in motor cortex. Neuroreport 9(2):315–19 [DOI] [PubMed] [Google Scholar]
  13. Bastian A, Schöner G, Riehle A. 2003. Preshaping and continuous evolution of motor cortical representations during movement preparation. Eur. J. Neurosci 18(7):2047–58 [DOI] [PubMed] [Google Scholar]
  14. Briggman KL, Abarbanel HDI, Kristan WB. 2005. Optical imaging of neuronal populations during decision-making. Science 307(5711):896–901 [DOI] [PubMed] [Google Scholar]
  15. Briggman KL, Abarbanel HDI, Kristan WB. 2006. From crawling to cognition: analyzing the dynamical interactions among populations of neurons. Curr. Opin. Neurobiol 16(2):135–44 [DOI] [PubMed] [Google Scholar]
  16. Broome BM, Jayaraman V, Laurent G. 2006. Encoding and decoding of overlapping odor sequences. Neuron 51(4):467–82 [DOI] [PubMed] [Google Scholar]
  17. Brunton BW, Johnson LA, Ojemann JG, Kutz JN. 2016. Extracting spatial–temporal coherent patterns in large-scale neural recordings using dynamic mode decomposition. J. Neurosci. Methods 258:1–15 [DOI] [PubMed] [Google Scholar]
  18. Buonomano DV. 2000. Decoding temporal information: a model based on short-term synaptic plasticity. J. Neurosci 20(3):1129–41 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Buonomano DV, Maass W. 2009. State-dependent computations: spatiotemporal processing in cortical networks. Nat. Rev. Neurosci 10(2):113–25 [DOI] [PubMed] [Google Scholar]
  20. Carrillo-Reid L, Han S, Yang W, Akrouh A, Yuste R. 2019. Controlling visually guided behavior by holo-graphic recalling of cortical ensembles. Cell 178(2):447–57.e5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Chaisangmongkon W, Swaminathan SK, Freedman DJ, Wang XJ. 2017. Computing by robust transience: how the fronto-parietal network performs sequential, category-based decisions. Neuron 93(6):1504–17.e4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Chase SM, Kass RE, Schwartz AB. 2012. Behavioral and neural correlates of visuomotor adaptation observed through a brain-computer interface in primary motor cortex. J. Neurophysiol 108(2):624–44 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Chaudhuri R, Fiete I. 2016. Computational principles of memory. Nat. Neurosci 19(3):394–403 [DOI] [PubMed] [Google Scholar]
  24. Chaudhuri R, Gerçek B, Pandey B, Peyrache A, Fiete I. 2019. The intrinsic attractor manifold and population dynamics of a canonical cognitive circuit across waking and sleep. Nat. Neurosci 22(9):1512–20 [DOI] [PubMed] [Google Scholar]
  25. Cherian A, Ernandes HL, Miller LE. 2013. Primary motor cortical discharge during force field adaptation reflects muscle-like dynamics. J. Neurophysiol 110(3):768–83 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Chestek CA, Batista AP, Santhanam G, Yu BM, Afshar A, et al. 2007. Single-neuron stability during repeated reaching in macaque premotor cortex. J. Neurosci 27:10742–50 [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Chettih SN, Harvey CD. 2019. Single-neuron perturbations reveal feature-specific competition in V1. Nature 567(7748):334–40 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Churchland MM, Afshar A, Shenoy KV. 2006a. A central source of movement variability. Neuron 52(6):1085–96 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Churchland MM, Cunningham JP, Kaufman MT, Foster JD, Nuyujukian P, et al. 2012. Neural population dynamics during reaching. Nature 487(7405):51–56 [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Churchland MM, Cunningham JP, Kaufman MT, Ryu SI, Shenoy KV. 2010. Cortical preparatory activity: representation of movement or first cog in a dynamical machine? Neuron 68(3):387–400 [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Churchland MM, Santhanam G, Shenoy KV. 2006b. Preparatory activity in premotor and motor cortex reflects the speed of the upcoming reach. J. Neurophysiol 96(6):3130–46 [DOI] [PubMed] [Google Scholar]
  32. Churchland MM, Shenoy KV. 2007a. Delay of movement caused by disruption of cortical preparatory activity. J. Neurophysiol 97(1):348–59 [DOI] [PubMed] [Google Scholar]
  33. Churchland MM, Shenoy KV. 2007b. Temporal complexity and heterogeneity of single-neuron activity in premotor and motor cortex. J. Neurophysiol 97(6):4235–57 [DOI] [PubMed] [Google Scholar]
  34. Churchland MM, Yu BM, Ryu SI, Santhanam G, Shenoy KV. 2006c. Neural variability in premotor cortex provides a signature of motor preparation. J. Neurosci 26(14):3697–712 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Cisek P 2006. Integrated neural processes for defining potential actions and deciding between them: a computational model. J. Neurosci 26(38):9761–70 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Cisek P, Crammond DJ, Kalaska JF. 2003. Neural activity in primary motor and dorsal premotor cortex in reaching tasks with the contralateral versus ipsilateral arm. J. Neurophysiol 89(2):922–42 [DOI] [PubMed] [Google Scholar]
  37. Cohen MR, Maunsell JHR. 2009. Attention improves performance primarily by reducing interneuronal correlations. Nat. Neurosci 12(12):1594–600 [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Cunningham JP, Yu BM. 2014. Dimensionality reduction for large-scale neural recordings. Nat. Neurosci 17(11):1500–9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Druckmann S, Chklovskii DB. 2012. Neuronal circuits underlying persistent representations despite time varying activity. Curr. Biol 22(22):2095–103 [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Egger SW, Remington ED, Chang C-J, Jazayeri M. 2019. Internal models of sensorimotor integration regulate cortical dynamics. Nat. Neurosci 22(11):1871–82 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Elsayed G, Cunningham J. 2017. Structure in neural population recordings: an expected byproduct of simpler phenomena? Nat. Neurosci 20:1310–18 [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Elsayed GF, Lara AH, Kaufman MT, Churchland MM, Cunningham JP. 2016. Reorganization between preparatory and movement population responses in motor cortex. Nat. Commun 7:13239. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Evarts EV. 1964. Temporal patterns of discharge of pyramidal tract neurons during sleep and waking in the monkey. J. Neurophysiol 27:152–71 [DOI] [PubMed] [Google Scholar]
  44. Evarts EV. 1968. Relation of pyramidal tract activity to force exerted during voluntary movement. J. Neurophysiol 31(1):14–27 [DOI] [PubMed] [Google Scholar]
  45. Even-Chen N, Sheffer B, Vyas S, Ryu SI, Shenoy KV. 2019. Structure and variability of delay activity in premotor cortex. PLOS Comput. Biol 15(2):e1006808. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Flint RD, Scheid MR, Wright ZA, Solla SA, Slutzky MW. 2016. Long-term stability of motor cortical activity: implications for brain machine interfaces and optimal feedback control. J. Neurosci 36:3623–32 [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Freedman DJ, Assad JA. 2006. Experience-dependent representation of visual categories in parietal cortex. Nature 443:85–88 [DOI] [PubMed] [Google Scholar]
  48. Gallego JA, Perich MG, Chowdhury RH, Solla SA, Miller LE. 2020. Long-term stability of cortical population dynamics underlying consistent behavior. Nat. Neurosci 23:260–70 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Gallego JA, Perich MG, Miller LE, Solla SA. 2017. Neural manifolds for the control of movement. Neuron 94(5):978–84 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Gallego JA, Perich MG, Naufel SN, Ethier C, Solla SA, Miller LE. 2018. Cortical population activity within a preserved neural manifold underlies multiple motor behaviors. Nat. Commun 9:4233. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Georgopoulos A, Kalaska J, Caminiti R, Massey J. 1982. On the relations between the direction of two-dimensional arm movements and cell discharge in primate motor cortex. J. Neurosci 2(11):1527–37 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Georgopoulos AP, Crutcher MD, Schwartz AB. 1989. Cognitive spatial-motor processes. 3. Motor cortical prediction of movement direction during an instructed delay period. Exp. Brain Res 75(1):183–94 [DOI] [PubMed] [Google Scholar]
  53. Georgopoulos AP, Schwartz AB, Kettner RE. 1986. Neuronal population coding of movement direction. Science 233(4771):1416–19 [DOI] [PubMed] [Google Scholar]
  54. Golub M, Sussillo D. 2018. FixedPointFinder: a Tensorflow toolbox for identifying and characterizing fixed points in recurrent neural networks. J. Open Source Softw 3(31):1003 [Google Scholar]
  55. Golub MD, Chase SM, Batista AP, Yu BM. 2016. Brain-computer interfaces for dissecting cognitive processes underlying sensorimotor control. Curr. Opin. Neurobiol 37:53–58 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Golub MD, Sadtler PT, Oby ER, Quick KM, Ryu SI, et al. 2018. Learning by neural reassociation. Nat. Neurosci 21:607–16 [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Hall TM, deCarvalho F, Jackson A. 2014. A common structure underlies low-frequency cortical dynamics in movement, sleep, and sedation. Neuron 83(5):1185–99 [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Hennequin G, Vogels TP, Gerstner W. 2014. Optimal control of transient dynamics in balanced networks supports generation of complex movements. Neuron 82(6):1394–406 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Hennig JA, Golub MD, Lund PJ, Sadtler PT, Oby ER, et al. 2018. Constraints on neural redundancy. eLife 7:e36774. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Hodgkin AL, Huxley AF. 1952. A quantitative description of membrane current and its application to conduction and excitation in nerve. J. Physiol 117(4):500–44 [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Hubel DH, Wiesel TN. 1962. Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J. Physiol 160(1):106–54 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Inagaki HK, Fontolan L, Romani S, Svoboda K. 2019. Discrete attractor dynamics underlies persistent activity in the frontal cortex. Nature 566(7743):212–17 [DOI] [PubMed] [Google Scholar]
  63. Jarosiewicz B, Chase SM, Fraser GW, Velliste M, Kass RE, Schwartz AB. 2008. Functional network reorganization during learning in a brain-computer interface paradigm. PNAS 105(49):19486–91 [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Jazayeri M, Afraz A. 2017. Navigating the neural space in search of the neural code. Neuron 93(5):1003–14 [DOI] [PubMed] [Google Scholar]
  65. Jazayeri M, Shadlen MN. 2015. A neural mechanism for sensing and reproducing a time interval. Curr. Biol 25(20):2599–609 [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Jun JJ, Steinmetz NA, Siegle JH, Denman DJ, Bauza M, et al. 2017. Fully integrated silicon probes for high-density recording of neural activity. Nature 551:232–36 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Kalaska JF. 2019. Emerging ideas and tools to study the emergent properties of the cortical neural circuits for voluntary motor control in non-human primates [version 1; peer review: 4 approved]. F1000Research 8:749. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Karmarkar UR, Buonomano DV. 2007. Timing in the absence of clocks: encoding time in neural network states. Neuron 53(3):427–38 [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Kato S, Kaplan HS, Schrödel T, Skora S, Lindsay TH, et al. 2015. Global brain dynamics embed the motor command sequence of Caenorhabditis elegans. Cell 163(3):656–69 [DOI] [PubMed] [Google Scholar]
  70. Kaufman MT, Churchland MM, Ryu SI, Shenoy KV. 2014. Cortical activity in the null space: permitting preparation without movement. Nat. Neurosci 17:440–48 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Kaufman MT, Churchland MM, Santhanam G, Yu BM, Afshar A, et al. 2010. Roles of monkey premotor neuron classes in movement preparation and execution. J. Neurophysiol 104(2):799–810 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Kaufman MT, Churchland MM, Shenoy KV. 2013. The roles of monkey M1 neuron classes in movement preparation and execution. J. Neurophysiol 110(4):817–25 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Kaufman MT, Seely JS, Sussillo D, Ryu SI, Shenoy KV, Churchland MM. 2016. The largest response component in the motor cortex reflects movement timing but not movement type. eNeuro 3(4):ENEURO.0085–16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Killeen PR, Fetterman JG. 1988. A behavioral theory of timing. Psychol. Rev 95(2):274–95 [DOI] [PubMed] [Google Scholar]
  75. Kim SS, Rouault H, Druckmann S, Jayaraman V. 2017. Ring attractor dynamics in the Drosophila central brain. Science 356(6340):849–53 [DOI] [PubMed] [Google Scholar]
  76. Kobak D, Brendel W, Constantinidis C, Feierstein CE, Kepecs A, et al. 2016. Demixed principal component analysis of neural population data. eLife 5:e10989. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Krakauer JW, Hadjiosif AM, Xu J, Wong AL, Haith AM. 2019. Motor learning. Compr. Physiol 9(2):613–63 [DOI] [PubMed] [Google Scholar]
  78. Lara AH, Cunningham JP, Churchland MM. 2018a. Different population dynamics in the supplementary motor area and motor cortex during reaching. Nat. Commun 9:2754. [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Lara AH, Elsayed GF, Zimnik AJ, Cunningham JP, Churchland MM. 2018b. Conservation of preparatory neural events in monkey motor cortex regardless of how movement is initiated. eLife 7:e31826. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Lebedev MA, Ossadtchi A, Mill NA, Urpi NA, Cervera MR, Nicolelis MAL. 2019. Analysis of neuronal ensemble activity reveals the pitfalls and shortcomings of rotation dynamics. Sci. Rep 9:18978. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Li N, Daie K, Svoboda K, Druckmann S. 2016. Robust neuronal dynamics in premotor cortex during motor planning. Nature 532:459–64 [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Linderman SW, Johnson MJ, Miller AC, Adams RP, Blei DM, Paninski L. 2017. Bayesian learning and inference in recurrent switching linear dynamical systems. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, 54:914–922. Brookline, MA: Microtome Publ. [Google Scholar]
  83. Luenberger DG. 1979. Introduction to Dynamic Systems: Theory, Models, and Applications. Hoboken, NJ: John Wiley & Sons [Google Scholar]
  84. Maass W, Natschläger T, Markram H. 2002. Real-time computing without stable states: a new framework for neural computation based on perturbations. Neural Comput. 14(11):2531–60 [DOI] [PubMed] [Google Scholar]
  85. Mante V, Sussillo D, Shenoy KV, Newsome WT. 2013. Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature 503:78–84 [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Marshel JH, Kim YS, Machado TA, Quirin S, Benson B, et al. 2019. Cortical layer–specific critical dynamics triggering perception. Science 365(6453):eaaw5202. [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Mauk MD, Buonomano DV. 2004. The neural basis of temporal processing. Annu. Rev. Neurosci 27:307–40 [DOI] [PubMed] [Google Scholar]
  88. Mazor O, Laurent G. 2005. Transient dynamics versus fixed points in odor representations by locust antennal lobe projection neurons. Neuron 48(4):661–73 [DOI] [PubMed] [Google Scholar]
  89. Meck WH. 1996. Neuropharmacology of timing and time perception. Cogn. Brain Res 3(3–4):227–42 [DOI] [PubMed] [Google Scholar]
  90. Messier J, Kalaska JF. 1997. Differential effect of task conditions on errors of direction and extent of reaching movements. Exp. Brain Res 115:469–78 [DOI] [PubMed] [Google Scholar]
  91. Messier J, Kalaska JF. 1999. Comparison of variability of initial kinematics and endpoints of reaching movements. Exp. Brain Res 125(2):139–52 [DOI] [PubMed] [Google Scholar]
  92. Michaels JA, Dann B, Intveld RW, Scherberger H. 2015. Predicting reaction time from the neural state space of the premotor and parietal grasping network. J. Neurosci 35(32):11415–32 [DOI] [PMC free article] [PubMed] [Google Scholar]
  93. Michaels JA, Dann B, Scherberger H. 2016. Neural population dynamics during reaching are better explained by a dynamical system than representational tuning. PLOS Comput. Biol 12(11):e1005175. [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Miller EK, Lundqvist M, Bastos AM. 2018. Working memory 2.0. Neuron 100(2):463–75 [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. Mitchell JF, Sundberg KA, Reynolds JH. 2009. Spatial attention decorrelates intrinsic activity fluctuations in macaque area V4. Neuron 63(6):879–88 [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Murakami M, Mainen ZF. 2015. Preparing and selecting actions with neural populations: toward cortical circuit mechanisms. Curr. Opin. Neurobiol 33:40–46 [DOI] [PubMed] [Google Scholar]
  97. Naumann EA, Fitzgerald JE, Dunn TW, Rihel J, Sompolinsky H, Engert F. 2016. From whole-brain data to functional circuit models: the zebrafish optomotor response. Cell 167(4):947–60.e20 [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Nichols ALA, Eichler T, Latham R, Zimmer M. 2017. A global brain state underlies C. elegans sleep behavior. Science 356(6344):eaam6851. [DOI] [PubMed] [Google Scholar]
  99. Oby ER, Golub MD, Hennig JA, Degenhart AD, Tyler-Kabara EC, et al. 2019. New neural activity patterns emerge with long-term learning. PNAS 116(30):15210–15 [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Orsborn AL, Moorman HG, Overduin SA, Shanechi MM, Dimitrov DF, Carmena JM. 2014. Closed-loop decoder adaptation shapes neural plasticity for skillful neuroprosthetic control. Neuron 82(6):1380–93 [DOI] [PubMed] [Google Scholar]
  101. O’Shea DJ, Trautmann E, Chandrasekaran C, Stavisky S, Kao JC, et al. 2017. The need for calcium imaging in nonhuman primates: new motor neuroscience and brain-machine interfaces. Exp. Neurol 287(Pt. 4):437–51 [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Otchy TM, Wolff SBE, Rhee JY, Pehlevan C, Kawai R, et al. 2015. Acute off-target effects of neural circuit manipulations. Nature 528(7582):358–63 [DOI] [PubMed] [Google Scholar]
  103. Pandarinath C, Ames KC, Russo AA, Farshchian A, Miller LE, et al. 2018a. Latent factors and dynamics in motor cortex and their application to brain–machine interfaces. J. Neurosci 38(44):9390–401 [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Pandarinath C, Gilja V, Blabe CH, Nuyujukian P, Sarma AA, et al. 2015. Neural population dynamics in human motor cortex during movements in people with ALS. eLife 4:e07436. [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Pandarinath C, O’Shea DJ, Collins J, Jozefowicz R, Stavisky SD, et al. 2018b. Inferring single-trial neural population dynamics using sequential auto-encoders. Nat. Methods 15(10):805–15 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Paninski L, Cunningham JP. 2018. Neural data science: accelerating the experiment-analysis-theory cycle in large-scale neuroscience. Curr. Opin. Neurobiol 50:232–41 [DOI] [PubMed] [Google Scholar]
  107. Perfiliev S, Isa T, Johnels B, Steg G, Wessberg J. 2010. Reflexive limb selection and control of reach direction to moving targets in cats, monkeys, and humans. J. Neurophysiol 104(5):2423–32 [DOI] [PubMed] [Google Scholar]
  108. Perich MG, Gallego JA, Miller LE. 2018. A neural population mechanism for rapid learning. Neuron 100(4):964–76.e7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Petreska B, Yu BM, Cunningham JP, Santhanam G, Ryu SI, et al. 2011. Dynamical segmentation of single trials from population neural data In Advances in Neural Information Processing Systems 24, ed. Shawe-Taylor J, Zemel RS, Bartlett PL, Pereira F, Weinberger KQ. San Diego, CA: NeurIPS [Google Scholar]
  110. Portugues R, Feierstein CE, Engert F, Orger MB. 2014. Whole-brain activity maps reveal stereotyped, distributed networks for visuomotor behavior. Neuron 81(6):1328–43 [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Rabinovich M, Huerta R, Laurent G. 2008. Transient dynamics for neural processing. Science 321(5885):48–50 [DOI] [PubMed] [Google Scholar]
  112. Rabinovich M, Volkovskii A, Lecanda P, Huerta R, Abarbanel HDI, Laurent G. 2001. Dynamical encoding by networks of competing neuron groups: winnerless competition. Phys. Rev. Lett 87(6):068102. [DOI] [PubMed] [Google Scholar]
  113. Raposo D, Kaufman MT, Churchland AK. 2014. A category-free neural population supports evolving demands during decision-making. Nat. Neurosci 17(12):1784–92 [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Remington ED, Egger SW, Narain D, Wang J, Jazayeri M. 2018a. Time in the brain a dynamical systems perspective on flexible motor timing. Trends Cogn. Sci 22(10):938–52 [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Remington ED, Narain D, Hosseini EA, Jazayeri M. 2018b. Flexible sensorimotor computations through rapid reconfiguration of cortical dynamics. Neuron 98(5):1005–19.e5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  116. Riehle A, Requin J. 1989. Monkey primary motor and premotor cortex: single-cell activity related to prior information about direction and extent of an intended movement. J. Neurophysiol 61(3):534–49 [DOI] [PubMed] [Google Scholar]
  117. Romo R, Schultz W. 1992. Role of primate basal ganglia and frontal cortex in the internal generation of movements. III. Neuronal activity in the supplementary motor area. Exp. Brain Res 91(3):396–407 [DOI] [PubMed] [Google Scholar]
  118. Rosenbaum DA. 1980. Human movement initiation: specification of arm, direction, and extent. J. Exp. Psychol. Gen 109(4):444–74 [DOI] [PubMed] [Google Scholar]
  119. Rouse AG, Schieber MH. 2018. Condition-dependent neural dimensions progressively shift during reach to grasp. Cell Rep. 25(11):3158–68.e3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  120. Russo AA, Bittner SR, Perkins SM, Seely JS, London BM, et al. 2018. Motor cortex embeds muscle-like commands in an untangled population response. Neuron 97(4):953–66.e8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  121. Russo AA, Khajeh R, Bittner SR, Perkins SM, Cunningham JP, et al. 2019. Neural trajectories in the supplementary motor area and primary motor cortex exhibit distinct geometries, compatible with different classes of computation. bioRxiv 650002. 10.1101/650002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  122. Sadtler PT, Quick KM, Golub MD, Chase SM, Ryu SI, et al. 2014. Neural constraints on learning. Nature 512(7515):423–26 [DOI] [PMC free article] [PubMed] [Google Scholar]
  123. Sakellaridi S, Christopoulos VN, Aflalo T, Pejsa KW, Rosario ER, et al. 2019. Intrinsic variable learning for brain-machine interface control by human anterior intraparietal cortex. Neuron 102(3):694–705.e3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  124. Santhanam G, Yu BM, Gilja V, Ryu SI, Afshar A, et al. 2009. Factor-analysis methods for higher-performance neural prostheses. J. Neurophysiol 102(2):1315–30 [DOI] [PMC free article] [PubMed] [Google Scholar]
  125. Sauerbrei BA, Guo J, Cohen JD, Mischiati M, Guo W, et al. 2020. Cortical pattern generation during dexterous movement is input-driven. Nature 577:386–91 [DOI] [PMC free article] [PubMed] [Google Scholar]
  126. Saxena S, Cunningham JP. 2019. Towards the neural population doctrine. Curr. Opin. Neurobiol 55:103–11 [DOI] [PubMed] [Google Scholar]
  127. Seely JS, Kaufman MT, Ryu SI, Shenoy KV, Cunningham JP, Churchland MM. 2016. Tensor analysis reveals distinct population structure that parallels the different computational roles of areas M1 and V1. PLOS Comput. Biol 12(11):e1005164. [DOI] [PMC free article] [PubMed] [Google Scholar]
  128. Semedo JD, Zandvakili A, Machens CK, Yu BM, Kohn A. 2019. Cortical areas interact through a communication subspace. Neuron 102(1):249–59.e4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  129. Sheahan HR, Franklin DW, Wolpert DM. 2016. Motor planning, not execution, separates motor memories. Neuron 92(4):773–79 [DOI] [PMC free article] [PubMed] [Google Scholar]
  130. Shenoy KV, Carmena JM. 2014. Combining decoder design and neural adaptation in brain-machine interfaces. Neuron 84(4):665–80 [DOI] [PubMed] [Google Scholar]
  131. Shenoy KV, Sahani M, Churchland MM. 2013. Cortical control of arm movements: a dynamical systems perspective. Annu. Rev. Neurosci 36:337–59 [DOI] [PubMed] [Google Scholar]
  132. Sohn H, Narain D, Meirhaeghe N, Jazayeri M. 2019. Bayesian computation through cortical latent dynamics. Neuron 103(5):934–47.e5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  133. Song HF, Yang GR, Wang XJ. 2016. Training excitatory-inhibitory recurrent neural networks for cognitive tasks: a simple and flexible framework. PLOS Comput. Biol 12(2):e1004792. [DOI] [PMC free article] [PubMed] [Google Scholar]
  134. Song HF, Yang GR, Wang XJ. 2017. Reward-based training of recurrent neural networks for cognitive and value-based tasks. eLife 6:e21492. [DOI] [PMC free article] [PubMed] [Google Scholar]
  135. Stavisky SD, Kao JC, Ryu SI, Shenoy KV. 2017a. Motor cortical visuomotor feedback activity is initially isolated from downstream targets in output-null neural state space dimensions. Neuron 95(1):195–208.e9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  136. Stavisky SD, Kao JC, Ryu SI, Shenoy KV. 2017b. Trial-by-trial motor cortical correlates of a rapidly adapting visuomotor internal model. J. Neurosci 37(7):1721–32 [DOI] [PMC free article] [PubMed] [Google Scholar]
  137. Stavisky SD, Willett FR, Wilson GH, Murphy BA, Rezaii P, et al. 2019. Neural ensemble dynamics in dorsal motor cortex during speech in people with paralysis. eLife 8:e46015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  138. Stopfer M, Jayaraman V, Laurent G. 2003. Intensity versus identity coding in an olfactory system. Neuron 39(6):991–1004 [DOI] [PubMed] [Google Scholar]
  139. Stringer C, Pachitariu M, Steinmetz N, Reddy CB, Carandini M, Harris KD. 2019. Spontaneous behaviors drive multidimensional, brainwide activity. Science 364(6437):eaav7893. [DOI] [PMC free article] [PubMed] [Google Scholar]
  140. Sussillo D, Abbott LF. 2009. Generating coherent patterns of activity from chaotic neural networks. Neuron 63(4):544–57 [DOI] [PMC free article] [PubMed] [Google Scholar]
  141. Sussillo D, Barak O. 2013. Opening the black box: low-dimensional dynamics in high-dimensional recurrent neural networks. Neural Comput. 25(3):626–49 [DOI] [PubMed] [Google Scholar]
  142. Sussillo D, Churchland MM, Kaufman MT, Shenoy KV. 2015. A neural network that finds a naturalistic solution for the production of muscle activity. Nat. Neurosci 18:1025–33 [DOI] [PMC free article] [PubMed] [Google Scholar]
  143. Swaminathan SK, Freedman DJ. 2012. Preferential encoding of visual categories in parietal cortex compared with prefrontal cortex. Nat. Neurosci 15(2):315–20 [DOI] [PMC free article] [PubMed] [Google Scholar]
  144. Tanji J, Evarts EV. 1976. Anticipatory activity of motor cortex neurons in relation to direction of an intended movement. J. Neurophysiol 39(5):1062–68 [DOI] [PubMed] [Google Scholar]
  145. Thoroughman KA, Shadmehr R. 1999. Electromyographic correlates of learning an internal model of reaching movements. J. Neurosci 19(19):8573–88 [DOI] [PMC free article] [PubMed] [Google Scholar]
  146. Trautmann EM, Stavisky SD, Lahiri S, Ames KC, Kaufman MT, et al. 2019. Accurate estimation of neural population dynamics without spike sorting. Neuron 103(2):292–308.e4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  147. Treisman M 1963. Temporal discrimination and the indifference interval. Implications for a model of the “internal clock.” Psychol. Monogr 77(13):1–31 [DOI] [PubMed] [Google Scholar]
  148. Vyas S, Even-Chen N, Stavisky SD, Ryu SI, Nuyujukian P, Shenoy KV. 2018. Neural population dynamics underlying motor learning transfer. Neuron 97(5):1177–86.e3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  149. Vyas S, O’Shea DJ, Ryu SI, Shenoy KV. 2020. Causal role of motor preparation during error-driven learning. Neuron. 106(2):329–39 [DOI] [PMC free article] [PubMed] [Google Scholar]
  150. Wang J, Narain D, Hosseini EA, Jazayeri M. 2018. Flexible timing by temporal scaling of cortical responses. Nat. Neurosci 21(1):102–10 [DOI] [PMC free article] [PubMed] [Google Scholar]
  151. Wärnberg E, Kumar A. 2019. Perturbing low dimensional activity manifolds in spiking neuronal networks. PLOS Comput. Biol 15(5):e1007074. [DOI] [PMC free article] [PubMed] [Google Scholar]
  152. Wei K, Körding K. 2009. Relevance of error: What drives motor adaptation? J. Neurophysiol 101(2):655–64 [DOI] [PMC free article] [PubMed] [Google Scholar]
  153. Williams AH, Kim TH, Wang F, Vyas S, Ryu SI, et al. 2018. Unsupervised discovery of demixed, low-dimensional neural dynamics across multiple timescales through tensor component analysis. Neuron 98(6):1099–115.e8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  154. Williamson RC, Doiron B, Smith MA, Yu BM. 2019. Bridging large-scale neuronal recordings and large-scale network models using dimensionality reduction. Curr. Opin. Neurobiol 55:40–47 [DOI] [PMC free article] [PubMed] [Google Scholar]
  155. Wise S 1985. The primate premotor cortex: past, present, and preparatory. Annu. Rev. Neurosci 8:1–19 [DOI] [PubMed] [Google Scholar]
  156. Wong AL, Haith AM, Krakauer JW. 2015. Motor planning. Neuroscientist 21(4):385–98 [DOI] [PubMed] [Google Scholar]
  157. Yang GR, Joglekar MR, Song HF, Newsome WT, Wang XJ. 2019. Task representations in neural networks trained to perform many cognitive tasks. Nat. Neurosci 22(2):297–306 [DOI] [PMC free article] [PubMed] [Google Scholar]
  158. Yang W, Carrillo-Reid L, Bando Y, Peterka DS, Yuste R. 2018. Simultaneous two-photon imaging and two-photon optogenetics of cortical circuits in three dimensions. eLife 7:e32671. [DOI] [PMC free article] [PubMed] [Google Scholar]
  159. Yu BM, Afshar A, Santhanam G, Ryu SI, Shenoy KV, Sahani M. 2005. Extracting dynamical structure embedded in neural activity In Advances in Neural Information Processing Systems 18, ed. Weiss Y, Schölkopf B, Platt JC. San Diego, CA: NeurIPS [Google Scholar]
  160. Yu BM, Cunningham JP, Santhanam G, Ryu SI, Shenoy KV, Sahani M. 2009. Gaussian-process factor analysis for low-dimensional single-trial analysis of neural population activity. J. Neurophysiol 102(1):614–35 [DOI] [PMC free article] [PubMed] [Google Scholar]
  161. Zhou X, Tien RN, Ravikumar S, Chase SM. 2019. Distinct types of neural reorganization during long-term learning. J. Neurophysiol 121(4):1329–41 [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES