Skip to main content
The Journal of Neuroscience logoLink to The Journal of Neuroscience
. 2018 Oct 31;38(44):9390–9401. doi: 10.1523/JNEUROSCI.1669-18.2018

Latent Factors and Dynamics in Motor Cortex and Their Application to Brain–Machine Interfaces

Chethan Pandarinath 1,2,, K Cora Ames 3,4,5,6, Abigail A Russo 3,5,6, Ali Farshchian 7, Lee E Miller 7, Eva L Dyer 1,8, Jonathan C Kao 9,10
PMCID: PMC6209846  PMID: 30381431

Abstract

In the 1960s, Evarts first recorded the activity of single neurons in motor cortex of behaving monkeys (Evarts, 1968). In the 50 years since, great effort has been devoted to understanding how single neuron activity relates to movement. Yet these single neurons exist within a vast network, the nature of which has been largely inaccessible. With advances in recording technologies, algorithms, and computational power, the ability to study these networks is increasing exponentially. Recent experimental results suggest that the dynamical properties of these networks are critical to movement planning and execution. Here we discuss this dynamical systems perspective and how it is reshaping our understanding of the motor cortices. Following an overview of key studies in motor cortex, we discuss techniques to uncover the “latent factors” underlying observed neural population activity. Finally, we discuss efforts to use these factors to improve the performance of brain–machine interfaces, promising to make these findings broadly relevant to neuroengineering as well as systems neuroscience.

Keywords: motor control, motor cortex, dynamical systems, brain-machine interfaces, neural population dynamics, machine learning

Introduction

Our knowledge of the motor cortices (MC) is rapidly evolving. Traditional models of motor cortical activity held that the firing rates of individual neurons “represent” externally measurable movement covariates, such as hand or joint kinematics, forces, or muscle activity. Much effort in related studies was devoted to finding the “correct” coordinate system. However, the increased ability to record from many neurons simultaneously has revealed many features of population activity that are difficult to reconcile with a purely representational viewpoint. First, much of the observed, high-dimensional activity of neural populations in MC can be explained as a combination of a modest number of “latent factors”; abstract, time-varying patterns that cannot be observed directly, but represent the correlated activity of the neural population. Second, during movements, these factors appear to evolve in time by obeying consistent dynamic rules, much like the lawful dynamics that govern physical systems. Through this lens, the complex, often-puzzling responses of individual neurons are naturally explained as minor elements in a coordinated underlying dynamical system. These findings have provided a new framework for evaluating neural activity during many of the functions that are ascribed to MC, such as motor preparation and execution, motor learning, bimanual control, and the production of muscle activity.

Beyond its application to the motor cortices, the dynamical systems framework and related computational methods may have broad applicability throughout the brain. Over the past decade, the ability to record from large populations of neurons has increased exponentially (Stevenson and Kording, 2011; Sofroniew et al., 2016, Jun et al., 2017; Stringer et al., 2018). These data collection tools promise to further transform our understanding of the brain, but only if we can process and interpret the coming wave of massive datasets. Trying to interpret the “tuning” of 10,000 neurons is not only onerous but a missed opportunity; much of the brain's computation is inaccessible from the activity of individual neurons, but is instead instantiated via population-level dynamics. Fortunately, modeling neural populations as low-dimensional dynamical systems is providing new insights in many cortical areas, including areas that mediate cognitive processes such as decision-making (Mante et al., 2013; Raposo et al., 2014; Carnevale et al., 2015), interval timing (Remington et al., 2018), and navigation (Harvey et al., 2012; Morcos and Harvey, 2016). This has deep implications for systems neuroscience: moving forward, the central thrust in understanding how brain areas perform computations and mediate behaviors may be through uncovering their population structure and underlying dynamics. MC is a critical model for studying these phenomena, as its activity appears strongly governed by internal dynamics, yet is well related to observable behavior. These characteristics make MC an excellent “proving ground” for tools that may be useful in a wide variety of brain areas.

Furthermore, our increasing knowledge of latent factors and dynamics in MC creates new opportunities to harness cortical activity to build high-performance and robust brain–machine interfaces (BMIs) to restore mobility to people with paralysis. BMIs aim to restore function by directly interfacing with the brain and reading out neural activity related to a person's movement intent. To date, the vast majority of BMIs that use MC activity have been based on a representational viewpoint, with the assumption that individual neurons represent external movement covariates. Incorporating knowledge of the latent structure and dynamics of MC population activity potentially offers the means to develop BMIs whose performance and long-term stability are greatly improved.

Our review is divided to cover three broad areas: (1) an overview of the dynamical systems view of MC, including key studies that have tested its applicability and demonstrated new insight into the structure of population activity in MC; (2) current techniques to uncover latent structure and dynamics from the activity of neural populations; and (3) recent efforts to use latent factors and dynamics to improve BMI performance.

The dynamical systems view and evidence in motor cortex

Early work to understand the relationship between MC activity and movements drew inspiration from studies in sensory areas, such as the experiments of Hubel and Wiesel (1959) in visual cortex. In those experiments, the response of a neuron was modeled as a function of carefully controlled features of the presented stimuli (Hubel and Wiesel, 1959). Similarly, studies in MC revealed that the responses of individual neurons (e.g., spike counts over hundreds of milliseconds) could be reasonably well modeled as a function of kinetic or kinematic movement parameters (Evarts, 1968; Georgopoulos et al., 1982; Schwartz et al., 1988). A complication of the motor domain is that these movement covariates could only be studied by training animals to produce highly stereotypic movements, replete with many correlations across limb segments and measurement systems. Over the next decades, a long-simmering debate that had originated perhaps with Hughlings Jackson over which parameters of movement were represented (Jackson, 1873; Phillips, 1975) was given new fuel. Anatomical considerations argue for a strong, direct link between primary motor cortex and muscle activity (Landgren et al., 1962; Jankowska et al., 1975; Cheney and Fetz, 1985), supported by many studies which found that neural activity covaries with muscle activation and kinetics (Evarts, 1968; Hepp-Reymond et al., 1999; Gribble and Scott, 2002; Holdefer and Miller, 2002). Yet correlates of higher-level parameters such as endpoint position (Riehle and Requin, 1989), velocity (Georgopoulos et al., 1982), speed (Churchland et al., 2006a), and curvature (Hocherman and Wise, 1991) could all be found as well. As this list became longer, some began to notice that these representations could also break down quite badly (Fu et al., 1995; Churchland and Shenoy, 2007b), and that such correlations could be spurious (Mussa-Ivaldi, 1988). This led many to wonder whether viewing MC as a representational system is appropriate (Fetz, 1992; Scott, 2008; Churchland et al., 2010).

Rather than asking which parameters constitute the output of MC, one might instead view the system from a generative perspective: how does MC generate its output? From this perspective, MC is seen as a computational engine whose activity translates high-level movement intention into the complex patterns of muscle activity required to execute a movement (Todorov and Jordan, 2002; Scott, 2004; Shenoy et al., 2013). If so, how might this computation be performed?

For decades, theoreticians have posited that brain areas may perform computation through network-level phenomena in which information is distributed across the activity of many neurons, and processed via lawful dynamics that dictate how the activity of a neural population evolves over time (for review, see Yuste, 2015). We formalize this dynamical view in Fig. 1A. We assume that at a given time point t, the activity of a population of D neurons can be captured by a vector of spike counts n(t) = [n1(t), n2(t), … nD(t)]. The neural population acts as a coordinated unit, with a K-dimensional internal “state” x(t) = [x1(t), x2(t), … xK(t)]. In many brain areas, x(t) has been observed to be much lower-dimensional than the total number of observed neurons (i.e., KD; Fig. 1B; Cunningham and Yu, 2014). This dimensionality is likely somewhat constrained due to the recurrent connectivity of the network, which restricts the possible patterns of coactivation that may occur (for review, see Gallego et al., 2017a). A notable consideration, however, is that the observed dimensionality is often lower than might be expected due to network constraints alone; this particularly low dimensionality may be further induced by the simplicity of common behavioral paradigms (Gao and Ganguli, 2015; Gao et al., 2017).

Figure 1.

Figure 1.

Intuition for latent factors and dynamical systems. A, n(t) is a vector representing observed spiking activity. Each element of the vector captures the number of spikes a given neuron emits within a short time window around time t. n(t) can typically be captured by the neural state variable x(t), an abstract, lower-dimensional representation that captures the state of the network. Dynamics are the rules that govern how the state updates in time. For a completely autonomous dynamical system without noise, if the dynamics f(x) are known, then the upcoming states are completely predictable based on an initial state x(0). B, In a simple three-neuron example, the ensemble's activity at each point in time traces out a trajectory in a 3-D state space, where each axis represents the activity of a given neuron. Not all possible patterns of activity are observed, rather, activity is confined to a 2-D plane within the 3-D space. The axes of this plane represent the neural state dimensions. Adapted from Cunningham and Yu, 2014. C, Conceptual low-dimensional dynamical system: a 1-D pendulum. A pendulum released from point p1 or p2 traces out different positions and velocities over time, and the state of the system can be captured by two state variables (position and velocity). D, The evolution of the system over time follows a fixed set of dynamic rules, i.e., the pendulum's equations of motion. Knowing the pendulum's initial state [x(0), filled circles] and the dynamical rules that govern its evolution [f(x), gray vector flow-field] is sufficient to predict the system's state at all future time points.

The dynamical systems view posits an additional constraint: the evolution of the population's activity in time is largely determined by internal rules (dynamics). In the limit of an autonomous dynamical system (i.e., a system that operates independently of any external factors), and without noise, the system's evolution follows the equation = f(x); that is, its future state changes are completely dependent upon (and predicted by) the current state. A conceptual example of a low-dimensional system with simple rotational dynamics (a 1-D pendulum), and its related dynamical flow-field, is presented in Figure 1, C and D. We note that MC clearly cannot be autonomous. It must receive and process inputs, such as sensory information, to produce responsive behaviors. However, as discussed below, the model of an autonomous dynamical system is reasonable for MC activity during the execution of well prepared movements. During behaviors that are unprepared, or where unpredictable events necessitate corrections (such as responding to task perturbations), MC activity may be well modeled as an input-driven dynamical system, analogous to a pendulum started from particular initial conditions, and subject to external perturbations (Pandarinath et al., 2018).

The dynamical systems framework makes testable predictions about the nature of MC activity. First, it predicts that the initial conditions of the system, such as those observed during movement preparation, largely determine the subsequent evolution of activity. Second, the activity of neurons in MC should relate not only to the inputs and outputs of the system, but also to the computations being performed. Finally, distinct computations may be appropriated into different, non-overlapping neural dimensions. Here, we explore experimental evidence related to each of these predictions.

Early studies exploring the dynamical systems hypothesis in MC examined whether preparatory activity served as an “initial condition” for the subsequent dynamics. In tasks with delay periods, where a subject has knowledge of the movement condition before execution, neural activity in MC approaches distinct “preparatory states” for distinct movements (Tanji and Evarts, 1976). In a dynamical system, initial conditions determine subsequent activity patterns, so the same dynamical “rules” can give rise to different activity patterns and behaviors if the initial condition is different. Similarly, in MC, an altered preparatory state relates to altered movement execution. If neural preparatory activity is not in the right state at the time of the go cue, either due to natural fluctuations (Churchland et al., 2006b; Afshar et al., 2011; Michaels et al., 2015, 2018), subthreshold microstimulation during the delay period (Churchland and Shenoy, 2007a), or a change in the location of the target (Ames et al., 2014), the reaction time is delayed compared with well prepared trials. This suggests that, if motor preparation is incorrect, subjects do not move until their preparation has been corrected. Furthermore, motor adaptation to a visuomotor rotation (Vyas et al., 2018), visuomotor scaling (Stavisky et al., 2017a), or a curl field (Perich and Miller, 2017; Perich et al., 2017) has been shown to modify the motor preparatory state. These modifications correspond to altered execution trajectories, and the associated changes in preparation states and execution trajectories transfer from covert settings (BMI tasks without movement) to overt movements (normal reaching movements; Vyas et al., 2018).

Additional work has tested whether dynamical system models continue to predict observed neural activity during the transition from preparation to movement. A simple model in which preparatory activity seeds the initial condition for rotational dynamics during movement generation fits neural activity well in nonhuman primates (Fig. 2A; Churchland et al., 2012; Elsayed et al., 2016; Michaels et al., 2016; Pandarinath et al., 2018) and humans (Pandarinath et al., 2015, 2018). Furthermore, recurrent neural networks trained to generate muscle activity after receiving preparatory input display dynamics similar to those recorded in MC (Hennequin et al., 2014; Sussillo et al., 2015; Kaufman et al., 2016), suggesting that a dynamical system that uses preparatory activity as the initial condition for subsequent movement dynamics may be a natural strategy for generating muscle activity during reaching.

Figure 2.

Figure 2.

Overview of results supporting the dynamical systems view of motor cortex. A, The neural state achieved during the delay period (green-red dots) predicts the subsequent trajectory of movement activity (green-red lines). Each dot/line is a single reach condition, recorded from a 108-condition task (inset). Adapted from Churchland et al., 2012. B, In dynamical systems, places where neighboring points in state space have very different dynamics are indications of “tangling”. Such regions would be highly sensitive to noise; small perturbations yield very different trajectories. C, Conceptual example illustrating tangling. Imagine a system that needs to produce two sine waves, one of which has double the frequency and is phase-shifted 1/4 of a cycle relative to the other. If it contains these sine waves with no additional dimensions, activity would trace out a figure 8, with a point of “high tangling” in the center. By adding in a third dimension, the system can move from a high tangling to a “low tangling” configuration, using the third dimension to separate the tangled points. Adapted from Russo et al., 2018. D, Although EMG often displays highly-tangled points (x-axis), MC's neural activity maintains low tangling (y-axis). E, Illustration of muscle-potent/muscle-null concept. Imagine a muscle that is driven with a strength equal to the sum of the firing rates of two units. If the units change in such a way that one unit's firing rate decreases as the other increases, then the overall drive to the muscle will remain the same (muscle-null). If, on the other hand, the neurons increase or decrease together, then the drive to the muscle will change (muscle-potent). In this way, neural activity can change in the muscle-null space while avoiding causing a direct change in the command to the muscles. Adapted from Kaufman et al., 2014. F, Neural activity in MC occupies a different set of dimensions during motor preparation than during movement. Red, Neural activity across different reach conditions in “preparatory” dimensions; green, neural activity across different reach conditions in “movement” dimensions. Adapted from Elsayed et al., 2016.

Another important prediction of the dynamical systems model is that not all of the activity in MC must directly relate to task parameters or muscle activity, but may instead relate to internal processes that subserve the current computation. For example, the switch from movement preparation to movement generation is accompanied by a substantial change in dynamics (Kaufman et al., 2014; Elsayed et al., 2016). Recent work has posited that this change is accomplished by a large, condition-invariant translation in state-space, which triggers the activation of movement-generation dynamics. Indeed, this condition-invariant signal is not only present at the switch from preparation to generation, but is also the largest aspect of the motor cortical response (Kaufman et al., 2016). Furthermore, during movement generation itself, the dominant patterns of neural activity may also play a role in supporting neural dynamics, rather than directly encoding the output. One challenge for a dynamical system results when the flow-field is highly “tangled”: when there are points in the space in which very similar states lead to very different future behavior. If two nearby points lead to different paths, then small amounts of noise in the system can lead to dramatic differences in the evolution of the neural state (Fig. 2B). A robust dynamical system, therefore, must ensure that the tangling is low, potentially by adding in additional dimensions of activity whose job is to “pull apart” points of high tangling (Fig. 2C). In MC, although some components of neural activity resemble muscle-like signals during movement generation, the largest patterns of neural activity during movement generation appear to function to reduce tangling (Fig. 2D; Russo et al., 2018). Thus, within MC, evidence has been found that some signals primarily support and drive dynamics, rather than directly encoding input or output.

Finally, the dynamical systems framework predicts that to perform different computations, neural activity may use different dimensions (Mante et al., 2013). Although this need not be a property of every possible dynamical system, using different dimensions for different functions allows a system to better maintain independence between its different roles. It has long been observed that many neurons are active both during movement preparation and movement generation. How then does preparatory activity avoid causing movement? Traditional views held that preparatory activity lies below a movement-generation threshold (Tanji and Evarts, 1976; Erlhagen and Schöner, 2002) or is under the influence of a gate (Bullock and Grossberg, 1988; Cisek, 2006). However, subthreshold activation fails to explain why preferred directions are minimally correlated between preparation and movement (Churchland et al., 2010; Kaufman et al., 2010), and there is little evidence for gating within MC, as inhibitory neurons in MC are not preferentially activated during motor preparation (Kaufman et al., 2013). The dynamical systems model, by contrast, makes a different prediction: that unwanted movements can be avoided by avoiding specific neural dimensions. Some dimensions, termed “output-potent”, correspond to activation patterns that are output to the muscles, whereas others, termed “output-null”, do not (Fig. 2D; Kaufman et al., 2014). In MC, different dimensions are activated during movement preparation and generation (Fig. 2E; Elsayed et al., 2016). Furthermore, the dimensions that best correlate with muscle activity are preferentially active during movement generation, suggesting that output-potent dimensions are selectively avoided during preparation (Kaufman et al., 2014). Similarly, distinct dimensions may be explored during cortically-dependent movement versus non-cortically-dependent movement (Miri et al., 2017), and sensory feedback initially enters MC in different dimensions from those of the muscle-potent activity (Stavisky et al., 2017b).

This division of different functions into different neural dimensions is not limited to muscle-related activity. In BMIs, where output-potent dimensions can be specified explicitly, activity that is informative about experimentally induced perturbations is initially orthogonal to the corresponding corrective responses (Stavisky et al., 2017b). During long-term BMI use, activity in output-potent dimensions is more stable than output-null dimensions (Flint et al., 2016). Neural activity also tends to occupy different dimensions for different movement categories: for example, pedaling in a forward versus reverse direction (Russo et al., 2018), isometric force production versus limb movement (Gallego et al., 2017b), or moving with the contralateral versus ipsilateral arm (Ames and Churchland, 2018). Using different neural dimensions for different functions may give the MC flexibility to generate activity patterns that support a wide variety of functions without interfering with one another (Perich et al., 2017).

Methods for estimating and evaluating motor cortical dynamics

As detailed above, much of the activity of MC neurons is naturally explained as a reflection of a low-dimensional dynamical system. Studying such dynamic processes requires techniques that can infer latent structure and its dynamics from observed, high-dimensional data. Related techniques have been applied to a wide variety of model systems and brain areas over the last two decades (for review, see Cunningham and Yu, 2014). However, MC holds particular value for testing these techniques, as its activity is closely tied to observable behavior (e.g., movement conditions, arm or hand kinematics, reaction times), which provides a key reference for validating the inferred state estimates. In this section, we review common techniques for estimating latent state and dynamics that have been applied in MC. We first present a general framework for discussion. Next, we review techniques that are applied to time points independently (i.e., do not explicitly model neural dynamics). Finally, we review techniques that do explicitly model neural dynamics, thereby resulting in better latent state estimates.

Techniques for latent state estimation typically view spiking activity as being “generated” by an underlying state x(t) (Fig. 3A). A common assumption is that for any given trial, the observed high-dimensional spike counts n(t) reflect a noisy sample from each neuron's underlying firing rate distribution r(t), a distribution that is itself derived from the latent state x(t). For motor cortical data, the distinction between observations n(t) and underlying rates r(t) captures the empirical observation that the spiking of any given neuron across multiple repeats (trials) of the same movement is highly variable.

Figure 3.

Figure 3.

Applications of latent state and dynamics estimation methods to MC ensemble activity. A, Generative model of observed neural activity. Population spiking activity is assumed to reflect an underlying latent state x(t) whose temporal evolution follows consistent rules (dynamics). Firing rates for each neuron r(t) are derived from x(t), and observed spikes n(t) reflect a noisy sample from r(t). B, dPCA applied to trial-averaged MC activity during a delayed reaching task separates condition-invariant and condition-variant dimensions. Each bar shows the total variance captured by each dimension, with red portions denoting condition-invariant fraction, and blue portions denoting condition-variant fraction. Traces show projection onto first dimension found by dPCA. Each trace corresponds to a single condition (inset, kinematic trajectories with corresponding colors). Adapted from Kaufman et al., 2016. C, GPFA reveals single-trial state space trajectories during a delayed reaching task. Gray traces represent individual trials. Ellipses indicate across-trial variability of the neural state at reach target onset (red shading), go cue (green shading), and movement onset (blue shading). Adapted from Yu et al., 2009. D, SLDS enables segmentation of individual trials by their dynamics. Each horizontal trace represents a single trial for the first state dimension found by the SLDS. Trace coloring represents time periods with distinct (discrete) dynamics for each trial, recognized in an unsupervised fashion. Switching between dynamic states reliably follow target onset and precede movement onset, with time lags that are correlated with reaction time. Adapted from Petreska et al., 2011.

A standard approach to de-noising n(t) and approximating r(t) is trial-averaging. Trial-averaging assumes all trials of a given movement condition are identical, and reduces single-trial noise in the estimate of r(t) by averaging n(t) across repeated trials. r(t) is often further de-noised by convolving it with a smoothing kernel. A common approach to estimate the lower dimensional x(t) is to perform principal component analysis (PCA). Performing PCA on r(t) rather than n(t) is preferred; if performed on n(t), PCA often results in poor latent factor estimation, because it simply maximizes the variance captured by the low-dimensional space, without separating variance that is shared among neurons from variance that is independent across neurons (for review, see Yu et al., 2009). When performed on r(t), PCA is typically accompanied by firing rate normalization, so that neurons with high rates (and thus higher variability) do not dominate the dimensionality reduction. Further, PCA can be extended by integrating some supervision into the dimensionality reduction step, e.g., by integrating information about task conditions to identify dimensions that capture neural variability related to particular task variables, using de-mixed PCA (dPCA; Fig. 3B; Kobak et al., 2014; Kaufman et al., 2016; Gallego et al., 2017b). The strategy of trial-averaging followed by PCA has led to several insights into latent structure and dynamics in MC (Churchland et al., 2012; Ames et al., 2014; Kaufman et al., 2014; Pandarinath et al., 2015; Elsayed et al., 2016; Kaufman et al., 2016; Gallego et al., 2017b; Russo et al., 2018).

However, circumventing the need to average over trials is critical for elucidating inherently single-trial phenomena, such as the trial-to-trial variability of real movements (and their corresponding error corrections), non-repeated behaviors such as natural movements, random target tasks, and tasks involving learning. Likewise, studying “internal” processes that vary substantially across trials and have limited behavioral correlates, such as decision-making, vacillation, and internal state estimates (Golub et al., 2015; Kaufman et al., 2015), is also impossible with trial-averaged data. Factor analysis (FA; Everitt, 1984) is often favored for analyzing single-trial phenomena (Santhanam et al., 2009; Sadtler et al., 2014; Golub et al., 2015, 2018; Athalye et al., 2017). A key assumption of FA is that activity that is correlated across neurons represents “signal” [comprising the latent factors x(t)], and activity that is not correlated across neurons represents “noise”. This assumption matches the graphical model in Figure 3A. A recent, complementary approach to capturing trial-dependent variability in n(t) without corrupting the latent factors is to integrate information regarding trial ordering into the dimensionality reduction step, and introduce a set of “trial factors” that accommodate variability across trials, as in tensor components analysis (TCA; Williams et al., 2018).

A key limitation of the above techniques (PCA, FA) is that they treat neighboring time points as though they are independent. However, as discussed, a core assumption of the dynamical systems framework is that time points are intimately related, and in particular, previous states are predictive of future states. Therefore, methods that simultaneously infer latent states and dynamics should provide more accurate state estimation by leveraging the interdependencies of data points that are close in time. Two well developed families of models are Gaussian process-based approaches (Fig. 3C; Yu et al., 2009; Lakshmanan et al., 2015; Zhao and Park, 2017; Duncker and Sahani, 2018) and linear dynamical systems (LDS)-based approaches (Macke et al., 2011; Buesing et al., 2012; Gao et al., 2015, 2016; Kao et al., 2015, 2017; Aghagolzadeh and Truccolo, 2016). Gaussian Process approaches assume that the latent state x(t) is composed of factors that vary smoothly and independently in time, with each factor having its own characteristic time constant. In comparison, LDS-based approaches assume that the latent state at a given time point is a linear function of the previous state (i.e., = Ax), which incorporates linear interactions between latent dimensions. One issue with the LDS approach is that the matrix A is time-invariant, yet must capture the dynamics at all time points. In MC, this is potentially problematic, because activity during different behavioral phases (e.g., preparation and movement) is governed by very different dynamics (Kaufman et al., 2014, 2016; Elsayed et al., 2016). A promising approach to address this challenge is switching LDS (SLDS), which assumes that at any given time point, the system's evolution obeys one of a discrete set of possible dynamics, each of which must be learned (Fig. 3D; Petreska et al., 2011; Linderman et al., 2017; Wei et al., 2018).

An alternative approach to uncovering single-trial population dynamics uses recurrent neural networks (RNNs). Known as latent factor analysis via dynamical systems (LFADS; Sussillo et al., 2016a; Pandarinath et al., 2018), the approach trains an RNN as a generative model of the observed spiking activity. RNNs are powerful nonlinear function approximators, capable of modeling complex, highly nonlinear dynamics through adjustment of their recurrent connectivity (Fig. 4A). LFADS uses a sequential autoencoder (SAE) framework (Fig. 4B), allowing the potentially nonlinear dynamics to be learned from noisy, single-trial neural population activity using stochastic gradient descent. This allows LFADS to accurately infer dynamics on a single-trial, moment-to-moment basis (Fig. 4C). A critical confirmation that these dynamics are accurate and meaningful is that they lead to dramatic improvements in the ability to predict behavior. As shown (Fig. 4D), the LFADS-inferred latent representations were considerably more informative about subjects' reaching movements than was the population activity that was directly observed. These findings reinforce that population states, rather than the activity of individual neurons, may be a key factor in understanding how brain areas mediate behaviors, and further, that SAEs provide a powerful new avenue toward linking the activity of neural populations to the behaviors they mediate.

Figure 4.

Figure 4.

LFADS uses recurrent neural networks to infer precise estimates of single-trial population dynamics. A, A recurrent neural network (simplified) is a set of artificial neurons that implements a nonlinear dynamical system, with dynamics set by adjusting the weights of its recurrent connections. Conceptually, the RNN can be “unrolled” in time, where future states of the RNN are completely predicted based in an initial state g(0) and its learned recurrent connectivity (compare Fig. 3A). B, The SAE framework consists of an encoding network and decoding network. The encoder (RNN) compresses single-trial observed activity n(t) into a trial code g(0), which sets the initial state of the decoder RNN. The decoder attempts to re-create n(t) based only on g(0). To do so, the decoder must model the ensemble's dynamics using its recurrent connectivity. The output of the decoder is x(t), the latent factors, and r(t), the de-noised firing rates. C, The de-noised single-trial estimates produced by LFADS uncover known dynamic features (such as rotations; Fig. 2A) on single trials. D, Decoding the LFADS-de-noised rates using simple optimal linear estimation leads to vastly improved predictions of behavioral variables (hand velocities) over Gaussian smoothing, even with limited numbers of neurons. Adapted from Pandarinath et al., 2018.

Leveraging latent factors and dynamics for BMIs

BMIs aim to help compensate for lost motor function by directly decoding movement intent from neuron spiking activity to control external devices or restore movement (Taylor et al., 2002; Carmena et al., 2003; Hochberg et al., 2006, 2012; Ethier et al., 2012; Collinger et al., 2013; Sadtler et al., 2014; Gilja et al., 2015; Ajiboye et al., 2017; Pandarinath et al., 2017). BMIs have largely decoded neural activity into movement through a representational viewpoint: each neuron represents a reach direction and if the neuron fires, it votes for movement in that direction (Georgopoulos et al., 1982; Taylor et al., 2002; Gilja et al., 2012; Hochberg et al., 2012; Collinger et al., 2013). Efforts to decode EMG activity have been essentially similar, though in a higher-dimensional, more abstract space (Pohlmeyer et al., 2007; Ethier et al., 2012). However, as previously discussed, this representational model has important limitations in describing MC activity. In this section, we review recent studies that have asked whether representational assumptions also limit BMI performance, and if so, whether performance and robustness can be increased by incorporating MC latent factors and dynamics.

Using latent factors and dynamics to increase BMI performance

The dynamical systems view holds that movement-related variables (such as kinematics or EMG activity) are not the only factors that influence the activity of MC neurons. However, BMI decoders based on the standard representational model do not take other factors into account when relating observed activity to movement intention. Recent work introduced a decoding architecture (graphically represented in Fig. 5A) that incorporates latent factors and their dynamics (modeled as a simple linear dynamical system; Kao et al., 2015; Aghagolzadeh and Truccolo, 2016). One advantage of this architecture is that modeling latent factors can account for the multiple, diverse influences on observed neural activity to better uncover movement-related variables. A second advantage is that latent factors may be more easily de-noised than the observed high-dimensional activity, resulting in higher BMI performance. Briefly, the dynamical systems view assumes that the temporal evolution of MC states is largely predictable. If so, deviations from this prediction may primarily correspond to noise. To de-noise, MC dynamics can be used to adjust the latent factors so that they are more consistent with the dynamic predictions (Fig. 5B). In closed-loop BMI experiments, decoding the dynamically de-noised latent factors significantly increased performance over previous approaches (Fig. 5C), including previous representational decoders that de-noise activity by (1) smoothing using an experimenter-chosen filter (optimal linear estimator; Velliste et al., 2008; Collinger et al., 2013), (2) incorporating prior knowledge about kinematic smoothness (kinematic Kalman filter; Wu et al., 2003; Kim et al., 2008; Gilja et al., 2012, 2015; Hochberg et al., 2012), and (3) learning filtering parameters via least-squares regression (Wiener filter; Carmena et al., 2003; Hochberg et al., 2006).

Figure 5.

Figure 5.

Improving BMI performance and longevity by leveraging neural dynamics. A, Graphical model of decoder with dynamical smoothing. B, Illustration of smoothing latent state estimates using neural dynamics. The instantaneous estimate of the latent state (blue) is augmented by a dynamical prior (gray flow-field) to produce a smoother, de-noised estimate (orange). C, Smoothing using neural dynamics results in better closed-loop BMI performance than other approaches. Performance is achieved information bitrate. Adapted from Kao et al., 2015. D, Example of low-dimensional signals that can be used to augment intracortical BMIs. PCA applied to neural activity around the time of target selection identifies a putative “error signal”, allowing real-time detection and correction of user errors in a typing BMI. Adapted from Even-Chen et al., 2017. E, Remembering dynamics from earlier recording conditions can extend performance as neurons are lost. Performance measure is (off-line) mean velocity correlation. F, Comparison of closed-loop performance when 110 channels are “lost” shows a >3× improvement achieved by remembering dynamics. FIT-KF, state-of-the-art kinematic Kalman filter (Fan et al., 2014). Adapted from Kao et al., 2017. G, Dynamic neural stitching with LFADS. A single model was trained on 44 recording sessions. Each session used a 24-channel recording probe. Left, Recording locations in MC. Right, Single-trial reaches from an example session. Arc. Sp., arcuate spur; PCd, precentral dimple; CS, central sulcus. H, Neural state space trajectories inferred by LFADS. Each trace of a given color is from a separate recording session (44 traces per condition). Inferred trajectories are consistent across 5 months. jPC1 and jPC2 are the first two components identified by jPCA (Churchland et al., 2012). I, Using LFADS to align 5 months of data (“Stitched”) significantly improves decoding versus other tested methods. Adapted from Pandarinath et al., 2018. ***Significant improvement in median R2; P < 10−8, Wilcoxon signed-rank test.

BMI performance may also be increased through the use of non-movement signals that become apparent by examining latent factors. Recently, Even-Chen et al. (2017) exploited this idea to identify factors that reflect errors made during BMI control. The motivation for this work is that errors inevitably happen when controlling a BMI; however, instead of the user having to correct an error explicitly, it is possible to detect (or predict) its occurrence and automatically correct it (or prevent it). They applied PCA to identify an error-related signal in MC and found dimensions where projected neural data reflected task errors (example latent factors observed during errors are shown in Fig. 5D). In real-time experiments with monkeys, these latent factors were decoded to both prevent and autocorrect errors, improving BMI performance.

Using dynamics and latent factors to increase BMI longevity

Ideally, a BMI's performance would be maintained indefinitely. However, neural recording conditions frequently change across days, and even within-day in pilot clinical trials, e.g., due to neuron death or electrode movement and failure (Barrese et al., 2013; Perge et al., 2013; Sussillo et al., 2016b; Downey et al., 2018), which can lead to decoding instability. Current approaches to solve this problem include decoding more stable neural signals (e.g., threshold crossings and local field potentials; Flint et al., 2013; Nuyujukian et al., 2014; Gilja et al., 2015; Stavisky et al., 2015), gradually updating decoder parameters using a weighted sliding average (Orsborn et al., 2012; Dangi et al., 2013), automated decoder recalibration by updating “tuning” estimates daily (Bishop et al., 2014), continuous recalibration by retrospectively inferring the user's intention among a set of fixed targets (Jarosiewicz et al., 2015), and training robust neural network decoders on a diversity of conditions using large data volumes (Sussillo et al., 2016b).

A separate class of approaches aims to exploit the underlying neural latent space, which, as a property of the neural population, should have a stable relationship with the user's intention that is independent of the specific neurons observed at any moment (Gao and Ganguli, 2015; Dyer et al., 2017; Gallego et al., 2017b; Kao et al., 2017; Pandarinath et al., 2018). However, it is challenging to relate the observed neurons from a given recording condition to the underlying latent space. Recent studies using supervised alignment strategies have demonstrated the potential of latent dynamics to maintain BMI performance. Kao et al. (2017) exploited historical information about population dynamics (Fig. 5E,F), finding that even under severe neuron loss, aligning the remaining neurons to previously learned dynamics could partially rescue closed-loop performance, effectively extending BMI lifetime. Alternatively, Pandarinath et al. (2018) trained a single LFADS model using data from 44 independently recorded neural populations spanning many millimeters of MC and 5 months of recording sessions (Fig. 5G,H). They then used a single linear decoder to map these latent dynamics onto kinematics (Fig. 5I). This work demonstrated that, in the absence of learning, a single, consistent dynamical model describes neural population activity across long time periods and large cortical areas, and yields improved off-line decoding performance for any given recording session than was otherwise possible.

Some settings lack data for supervised alignment, i.e., directly linking neural activity from new recording conditions to motor intent may be challenging (settings without structured behaviors, or where intent is less clear on a moment-by-moment basis). In these settings, unsupervised techniques may be useful for aligning data. Recently, Dyer et al. (2017) introduced a semi-supervised approach called distribution alignment decoding (DAD; Fig. 6A–C). This approach aims to map neural data from a new recording condition (new data) onto a previously recorded low-dimensional movement distribution. To do this, DAD first reduces the dimensionality of the neural data (using PCA or a nonlinear manifold learning technique), and then searches for an affine transformation to match the low-dimensional neural data to movements, by minimizing the Kullback–Leibler (KL)-divergence between the two datasets (Fig. 6B). Their results demonstrate that DAD can achieve similar performance to that of a supervised decoder that has access to corresponding measurements of neural state and movement, if the underlying data distribution contains asymmetries that facilitate alignment (Fig. 6C). Although a powerful approach for aligning neural and movement data, DAD solves a non-convex optimization problem with many local minima (Fig. 6B) by using a brute force search. To improve alignment and avoid having to first perform dimensionality reduction, neural network architectures such as generative adversarial networks (GANs; Goodfellow et al., 2014; Molano-Mazon et al., 2018) provide a potential method to learn nonlinear mappings from one distribution to another (Fig. 6D). By leveraging the fact that low-dimensional representations of neural activity are consistent across days and even subjects, distribution alignment methods like DAD or GANs provide a strategy for decoding movements without labeled training data from new recording conditions.

Figure 6.

Figure 6.

Distribution alignment methods for stabilizing movement decoders across days and subjects. A, Data dimensionality is first reduced, and then low-dimensional projections are aligned onto a previously recorded movement distribution. B, KL-divergence provides a robust metric for alignment (displayed as a function of the angle used to rotate the data). Many local minima exist (points 1, 2, 3, 4), which makes alignment difficult. C, Prediction accuracy of 2-D kinematics for distribution alignment decoding and supervised methods. Left, Accuracy of DAD using movements from Subject M (DAD-M), from Subject C (DAD-C), and using movements from both Subjects M and C (DAD-MC). Right, Standard L2-regularized supervised decoder (Sup) and a combined decoder (Sup-DAD), which averages the results of the supervised and DAD decoders. All results are compared with an Oracle decoder (far right), which provides an upper bound for the best linear decoding performance for this task. Adapted from Dyer et al., 2017. D, A schematic of a generative adversarial network strategy for distribution alignment across multiple days: generator network (left) receives new data and learns a transformation of the data to match the prior (from a previous day).

Conclusions

The increasing ability to monitor large numbers of neurons simultaneously will present new opportunities to study neural activity at the population level. Mounting evidence shows that this provides a qualitatively different window into the nervous system from that of single-neuron recordings, and that population-level dynamics likely underlie neural population activity across a wide range of systems. Here we reviewed recent evidence that such dynamics shape activity and drive behavior in MC, outlined key methods for inferring latent factors and dynamics that have been applied to MC activity, and showed how uncovering latent factors and dynamics can yield higher-performing and more robust BMIs. Continuing advances in recording technologies, algorithms, and computational power will enable studies of dynamics that were not previously possible, and further, may open new avenues for neural prostheses to address a wide variety of disorders of the nervous system.

Footnotes

This work was supported by a Burroughs Wellcome Fund Collaborative Research Travel Grant (C.P.) and NIH NINDS R01NS053603 (L.E.M.).We thank Steven Chase, Chandramouli Chandrasekaran, Juan Gallego, Matthew Kaufman, Daniel O'Shea, David Sussillo, Sergey Stavisky, Xulu Sun, Eric Trautmann, Jessica Verhein, Saurabh Vyas, Megan Wang, and Byron Yu for their feedback on the paper.

The authors declare no competing financial interests.

References

  1. Afshar A, Santhanam G, Yu BM, Ryu SI, Sahani M, Shenoy KV (2011) Single-trial neural correlates of arm movement preparation. Neuron 71:555–564. 10.1016/j.neuron.2011.05.047 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Aghagolzadeh M, Truccolo W (2016) Inference and decoding of motor cortex low-dimensional dynamics via latent state-space models. IEEE Trans Neural Syst Rehabil Eng 24:272–282. 10.1109/TNSRE.2015.2470527 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Ajiboye AB, Willett FR, Young DR, Murphy BA, Miller JP, Walter BL, Sweet JA, Hoyen HA, Keith MW, Peckham PH, Simeral JD, Donoghue JP, Miller JP, Hochberg LR, Kirsch RF (2017) Restoration of reaching and grasping in a person with tetraplegia through brain-controlled muscle stimulation: a proof-of-concept demonstration. Lancet 389:1821–1830. 10.1016/S0140-6736(17)30601-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Ames KC, Churchland MM (2018) Population-level but not neuron-level similarity during movement of the contra- vs ispilateral hand. In: Computational and Systems Neuroscience (COSYNE). Denver, CO. [Google Scholar]
  5. Ames KC, Ryu SI, Shenoy KV (2014) Neural dynamics of reaching following incorrect or absent motor preparation. Neuron 81:438–451. 10.1016/j.neuron.2013.11.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Athalye VR, Ganguly K, Costa RM, Carmena JM (2017) Emergence of coordinated neural dynamics underlies neuroprosthetic learning and skillful control. Neuron 93:955–970.e5. 10.1016/j.neuron.2017.01.016 [DOI] [PubMed] [Google Scholar]
  7. Barrese JC, Rao N, Paroo K, Triebwasser C, Vargas-Irwin C, Franquemont L, Donoghue JP (2013) Failure mode analysis of silicon-based intracortical microelectrode arrays in non-human primates. J Neural Eng 10:66014. 10.1088/1741-2560/10/6/066014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bishop W, Chestek CC, Gilja V, Nuyujukian P, Foster JD, Ryu SI, Shenoy KV, Yu BM (2014) Self-recalibrating classifiers for intracortical brain–computer interfaces. J Neural Eng 11:026001. 10.1088/1741-2560/11/2/026001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Buesing L, Macke J, Sahani M (2012) Learning stable, regularised latent models of neural population dynamics. Network 23:24–47. [DOI] [PubMed] [Google Scholar]
  10. Bullock D, Grossberg S (1988) Neural dynamics of planned arm movements: emergent invariants and speed-accuracy properties during trajectory formation. Psychol Rev 95:49–90. 10.1037/0033-295X.95.1.49 [DOI] [PubMed] [Google Scholar]
  11. Carmena JM, Lebedev MA, Crist RE, O'Doherty JE, Santucci DM, Dimitrov DF, Patil PG, Henriquez CS, Nicolelis MA (2003) Learning to control a brain–machine interface for reaching and grasping by primates. PLoS Biol 1:e42. 10.1371/journal.pbio.0000042 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Carnevale F, de Lafuente V, Romo R, Barak O, Parga N (2015) Dynamic control of response criterion in premotor cortex during perceptual detection under temporal uncertainty. Neuron 86:1067–1077. 10.1016/j.neuron.2015.04.014 [DOI] [PubMed] [Google Scholar]
  13. Cheney PD, Fetz EE (1985) Comparable patterns of muscle facilitation evoked by individual corticomotoneuronal (CM) cells and by single intracortical microstimuli in primates: evidence for functional groups of CM cells. J Neurophysiol 53:786–804. 10.1152/jn.1985.53.3.786 [DOI] [PubMed] [Google Scholar]
  14. Churchland MM, Shenoy KV (2007a) Delay of movement caused by disruption of cortical preparatory activity. J Neurophysiol 97:348–359. 10.1152/jn.00808.2006 [DOI] [PubMed] [Google Scholar]
  15. Churchland MM, Shenoy KV (2007b) Temporal complexity and heterogeneity of single-neuron activity in premotor and motor cortex. J Neurophysiol 97:4235–4257. 10.1152/jn.00095.2007 [DOI] [PubMed] [Google Scholar]
  16. Churchland MM, Santhanam G, Shenoy KV (2006a) Preparatory activity in premotor and motor cortex reflects the speed of the upcoming reach. J Neurophysiol 96:3130–3146. 10.1152/jn.00307.2006 [DOI] [PubMed] [Google Scholar]
  17. Churchland MM, Afshar A, Shenoy KV (2006b) A central source of movement variability. Neuron 52:1085–1096. 10.1016/j.neuron.2006.10.034 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Churchland MM, Cunningham JP, Kaufman MT, Ryu SI, Shenoy KV (2010) Cortical preparatory activity: representation of movement or first cog in a dynamical machine? In: Neuron, pp 387–400: Elsevier. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Churchland MM, Cunningham JP, Kaufman MT, Foster JD, Nuyujukian P, Ryu SI, Shenoy KV (2012) Neural population dynamics during reaching. Nature 487:51–56. 10.1038/nature11129 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Cisek P. (2006) Integrated neural processes for defining potential actions and deciding between them: a computational model. J Neurosci 26:9761–9770. 10.1523/JNEUROSCI.5605-05.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Collinger JL, Wodlinger B, Downey JE, Wang W, Tyler-Kabara EC, Weber DJ, McMorland AJ, Velliste M, Boninger ML, Schwartz AB (2013) High-performance neuroprosthetic control by an individual with tetraplegia. Lancet 381:557–564. 10.1016/S0140-6736(12)61816-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Cunningham JP, Yu BM (2014) Dimensionality reduction for large-scale neural recordings. Nat Neurosci 17:1500–1509. 10.1038/nn.3776 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Dangi S, Orsborn AL, Moorman HG, Carmena JM (2013) Design and analysis of closed-loop decoder adaptation algorithms for brain–machine interfaces. Neural Comput 25:1693–1731. 10.1162/NECO_a_00460 [DOI] [PubMed] [Google Scholar]
  24. Downey JE, Schwed N, Chase SM, Schwartz AB, Collinger JL (2018) Intracortical recording stability in human brain–computer interface users. J Neural Eng 15:046016. 10.1088/1741-2552/aab7a0 [DOI] [PubMed] [Google Scholar]
  25. Duncker L, Sahani M (2018) Temporal alignment and latent Gaussian process factor inference in population spike trains. bioRxiv 331751 10.1101/331751 [DOI] [Google Scholar]
  26. Dyer EL, Azar MG, Perich MG, Fernandes HL, Naufel S, Miller LE, Körding KP (2017) A cryptography-based approach for movement decoding. Nat Biomed Eng 1:967 10.1038/s41551-017-0169-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Elsayed GF, Lara AH, Kaufman MT, Churchland MM, Cunningham JP (2016) Reorganization between preparatory and movement population responses in motor cortex. Nat Commun 7:13239. 10.1038/ncomms13239 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Erlhagen W, Schöner G (2002) Dynamic field theory of movement preparation. Psychol Rev 109:545–572. 10.1037/0033-295X.109.3.545 [DOI] [PubMed] [Google Scholar]
  29. Ethier C, Oby ER, Bauman MJ, Miller LE (2012) Restoration of grasp following paralysis through brain-controlled stimulation of muscles. Nature 485:368–371. 10.1038/nature10987 [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Evarts EV. (1968) Relation of pyramidal tract activity to force exerted during voluntary movement. J Neurophysiol 31:14–27. 10.1152/jn.1968.31.1.14 [DOI] [PubMed] [Google Scholar]
  31. Even-Chen N, Stavisky SD, Kao JC, Ryu SI, Shenoy KV (2017) Augmenting intracortical brain–machine interface with neurally driven error detectors. J Neural Eng 14:066007. 10.1088/1741-2552/aa8dc1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Everitt B. (1984) An introduction to latent variable models. London: Chapman and Hall. [Google Scholar]
  33. Fan JM, Nuyujukian P, Kao JC, Chestek CA, Ryu SI, Shenoy KV (2014) Intention estimation in brain–machine interfaces. J Neural Eng 11:016004. 10.1088/1741-2560/11/1/016004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Fetz EE. (1992) Are movement parameters recognizably coded in the activity of single neurons? Behav Brain Sci 15:679–690. 10.1017/S0140525X00072599 [DOI] [Google Scholar]
  35. Flint RD, Wright ZA, Scheid MR, Slutzky MW (2013) Long term, stable brain machine interface performance using local field potentials and multiunit spikes. J Neural Eng 10:056005. 10.1088/1741-2560/10/5/056005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Flint RD, Scheid MR, Wright ZA, Solla SA, Slutzky MW (2016) Long-term stability of motor cortical activity: implications for brain machine interfaces and optimal feedback control. J Neurosci 36:3623–3632. 10.1523/JNEUROSCI.2339-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Fu QG, Flament D, Coltz JD, Ebner TJ (1995) Temporal encoding of movement kinematics in the discharge of primate primary motor and premotor neurons. J Neurophysiol 73:836–854. 10.1152/jn.1995.73.2.836 [DOI] [PubMed] [Google Scholar]
  38. Gallego JA, Perich MG, Miller LE, Solla SA (2017a) Neural manifolds for the control of movement. Neuron 94:978–984. 10.1016/j.neuron.2017.05.025 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Gallego JA, Perich MG, Naufel SN, Ethier C, Solla SA, Miller LE (2017b) Multiple tasks viewed from the neural manifold: stable control of varied behavior. bioRxiv 176081 10.1101/176081 [DOI] [Google Scholar]
  40. Gao P, Ganguli S (2015) On simplicity and complexity in the brave new world of large-scale neuroscience. Curr Opin Neurobiol 32:148–155. 10.1016/j.conb.2015.04.003 [DOI] [PubMed] [Google Scholar]
  41. Gao Y, Buesing L, Shenoy KV, Cunningham JP (2015) High-dimensional neural spike train analysis with generalized count linear dynamical systems. Adv Neural Inform Process Syst 2:2044–2052. [Google Scholar]
  42. Gao Y, Archer E, Paninski L, Cunningham JP (2016) Linear dynamical neural population models through nonlinear embeddings. arXiv 1605.08454 [Google Scholar]
  43. Gao P, Trautmann E, Byron MY, Santhanam G, Ryu S, Shenoy K, Ganguli S (2017) A theory of multineuronal dimensionality, dynamics and measurement. bioRxiv 214262 10.1101/214262 [DOI] [Google Scholar]
  44. Georgopoulos AP, Kalaska JF, Caminiti R, Massey JT (1982) On the relations between the direction of two-dimensional arm movements and cell discharge in primate motor cortex. J Neurosci 2:1527–1537. 10.1523/JNEUROSCI.02-11-01527.1982 [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Gilja V, Nuyujukian P, Chestek CA, Cunningham JP, Yu BM, Fan JM, Churchland MM, Kaufman MT, Kao JC, Ryu SI, Shenoy KV (2012) A high-performance neural prosthesis enabled by control algorithm design. Nat Neurosci 15:1752–1757. 10.1038/nn.3265 [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Gilja V, Pandarinath C, Blabe CH, Nuyujukian P, Simeral JD, Sarma AA, Sorice BL, Perge JA, Jarosiewicz B, Hochberg LR, Shenoy KV, Henderson JM (2015) Clinical translation of a high-performance neural prosthesis. Nat Med 21:1142–1145. 10.1038/nm.3953 [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Golub MD, Yu BM, Chase SM (2015) Internal models for interpreting neural population activity during sensorimotor control. eLife 4:e10015. 10.7554/eLife.10015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Golub MD, Sadtler PT, Oby ER, Quick KM, Ryu SI, Tyler-Kabara EC, Batista AP, Chase SM, Yu BM (2018) Learning by neural reassociation. Nat Neurosci 21:607–616. 10.1038/s41593-018-0095-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In: Advances in Neural Information Processing Systems 27 (eds Ghahramani Z., et al.), pp 2672–2680 (Curran Associates, Red Hook, NY; ). [Google Scholar]
  50. Gribble PL, Scott SH (2002) Overlap of internal models in motor cortex for mechanical loads during reaching. Nature 417:938–941. 10.1038/nature00834 [DOI] [PubMed] [Google Scholar]
  51. Harvey CD, Coen P, Tank DW (2012) Choice-specific sequences in parietal cortex during a virtual-navigation decision task. Nature 484:62–68. 10.1038/nature10918 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Hennequin G, Vogels TP, Gerstner W (2014) Optimal control of transient dynamics in balanced networks supports generation of complex movements. Neuron 82:1394–1406. 10.1016/j.neuron.2014.04.045 [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Hepp-Reymond MC, Kirkpatrick-Tanner M, Gabernet L, Qi HX, Weber B (1999) Context-dependent force coding in motor and premotor cortical areas. Exp Brain Res 128:123–133. 10.1007/s002210050827 [DOI] [PubMed] [Google Scholar]
  54. Hochberg LR, Serruya MD, Friehs GM, Mukand JA, Saleh M, Caplan AH, Branner A, Chen D, Penn RD, Donoghue JP (2006) Neuronal ensemble control of prosthetic devices by a human with tetraplegia. Nature 442:164–171. 10.1038/nature04970 [DOI] [PubMed] [Google Scholar]
  55. Hochberg LR, Bacher D, Jarosiewicz B, Masse NY, Simeral JD, Vogel J, Haddadin S, Liu J, Cash SS, van der Smagt P, Donoghue JP (2012) Reach and grasp by people with tetraplegia using a neurally controlled robotic arm. Nature 485:372–375. 10.1038/nature11076 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Hocherman S, Wise SP (1991) Effects of hand movement path on motor cortical activity in awake, behaving rhesus monkeys. Exp Brain Res 83:285–302. [DOI] [PubMed] [Google Scholar]
  57. Holdefer R, Miller LE (2002) Primary motor cortical neurons encode functional muscle synergies. Exp Brain Res 146:233–243. 10.1007/s00221-002-1166-x [DOI] [PubMed] [Google Scholar]
  58. Hubel DH, Wiesel TN (1959) Receptive fields of single neurones in the cat's striate cortex. J Physiol 148:574–591. 10.1113/jphysiol.1959.sp006308 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Jackson JH. (1873) Observations on the localisation of movements in the cerebral hemispheres, as revealed by cases of convulsion, chorea, and “aphasia”. The West Riding Lunatic Asylum Medical Reports 3:175–195. [Google Scholar]
  60. Jankowska E, Padel Y, Tanaka R (1975) Projections of pyramidal tract cells to alpha-motoneurones innervating hind-limb muscles in the monkey. J Physiol 249:637–667. 10.1113/jphysiol.1975.sp011035 [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Jarosiewicz B, Sarma AA, Bacher D, Masse NY, Simeral JD, Sorice B, Oakley EM, Blabe C, Pandarinath C, Gilja V, Cash SS, Eskandar EN, Friehs G, Henderson JM, Shenoy KV, Donoghue JP, Hochberg LR (2015) Virtual typing by people with tetraplegia using a self-calibrating intracortical brain-computer interface. Sci Transl Med 7:313ra179. 10.1126/scitranslmed.aac7328 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Jun JJ, Steinmetz NA, Siegle JH, Denman DJ, Bauza M, Barbarits B, Lee AK, Anastassiou CA, Andrei A, Aydın ÇBarbic M, Blanche TJ, Bonin V, Couto J, Dutta B, Gratiy SL, Gutnisky DA, Häusser M, Karsh B, Ledochowitsch P, Lopez CM, et al. (2017) Fully integrated silicon probes for high-density recording of neural activity. Nature 551:232. 10.1038/nature24636 [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Kao JC, Nuyujukian P, Ryu SI, Churchland MM, Cunningham JP, Shenoy KV (2015) Single-trial dynamics of motor cortex and their applications to brain–machine interfaces. Nat Commun 6:7759. 10.1038/ncomms8759 [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Kao JC, Ryu SI, Shenoy KV (2017) Leveraging neural dynamics to extend functional lifetime of brain–machine interfaces. Sci Rep 7:7395. 10.1038/s41598-017-06029-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Kaufman MT, Churchland MM, Santhanam G, Yu BM, Afshar A, Ryu SI, Shenoy KV (2010) Roles of monkey premotor neuron classes in movement preparation and execution. J Neurophysiol 104:799–810. 10.1152/jn.00231.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Kaufman MT, Churchland MM, Shenoy KV (2013) The roles of monkey M1 neuron classes in movement preparation and execution. J Neurophysiol 110:817–825. 10.1152/jn.00892.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Kaufman MT, Churchland MM, Ryu SI, Shenoy KV (2014) Cortical activity in the null space: permitting preparation without movement. Nat Neurosci 17:440–448. 10.1038/nn.3643 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Kaufman MT, Churchland MM, Ryu SI, Shenoy KV (2015) Vacillation, indecision and hesitation in moment-by-moment decoding of monkey motor cortex. elife 4:e04677. 10.7554/eLife.04677 [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Kaufman MT, Seely JS, Sussillo D, Ryu SI, Shenoy KV, Churchland MM (2016) The largest response component in motor cortex reflects movement timing but not movement type. eNeuro 3:ENEURO.0085–0016.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Kim SP, Simeral JD, Hochberg LR, Donoghue JP, Black MJ (2008) Neural control of computer cursor velocity by decoding motor cortical spiking activity in humans with tetraplegia. J Neural Eng 5:455–476. 10.1088/1741-2560/5/4/010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Kobak D, Brendel W, Constantinidis C, Feierstein CE, Kepecs A, Mainen ZF, Rumo R, Qi XL, Uchida N, Machens CK (2014) Demixed principal component analysis of population activity in higher cortical areas reveals independent representation of task parameters. arXiv 1410.6031. [Google Scholar]
  72. Lakshmanan KC, Sadtler PT, Tyler-Kabara EC, Batista AP, Yu BM (2015) Extracting low-dimensional latent structure from time series in the presence of delays. Neural Comput 27:1825–1856. 10.1162/NECO_a_00759 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Landgren S, Phillips CG, Porter R (1962) Minimal synaptic actions of pyrimidal impulses on some alpha motoneurones of the baboon's hand and forearm. J Physiol 161:91–111. 10.1113/jphysiol.1962.sp006875 [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Linderman S, Johnson M, Miller A, Adams R, Blei D, Paninski L (2017) Bayesian learning and inference in recurrent switching linear dynamical systems. Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, in PMLR 54:914–922. [Google Scholar]
  75. Macke JH, Buesing L, Cunningham JP, Yu BM, Shenoy KV, Sahani M (2011) Empirical models of spiking in neuronal populations. In: Advances in Neural Information Processing Systems, 24 Red Hook, NY: Neural Information Processing Systems. [Google Scholar]
  76. Mante V, Sussillo D, Shenoy KV, Newsome WT (2013) Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature 503:78–84. 10.1038/nature12742 [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Michaels JA, Dann B, Intveld RW, Scherberger H (2015) Predicting reaction time from the neural state space of the premotor and parietal grasping network. J Neurosci 35:11415–11432. 10.1523/JNEUROSCI.1714-15.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Michaels JA, Dann B, Scherberger H (2016) Neural population dynamics during reaching are better explained by a dynamical system than representational tuning. PLoS Comput Biol 12:e1005175. 10.1371/journal.pcbi.1005175 [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Michaels JA, Dann B, Intveld RW, Scherberger H (2018) Neural dynamics of variable grasp movement preparation in the macaque frontoparietal network. J Neurosci 38:5759–5773. 10.1523/JNEUROSCI.2557-17.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Miri A, Warriner CL, Seely JS, Elsayed GF, Cunningham JP, Churchland MM, Jessell TM (2017) Behaviorally selective engagement of short-latency effector pathways by motor cortex. Neuron 95:683–696.e11. 10.1016/j.neuron.2017.06.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Molano-Mazon M, Onken A, Piasini E, Panzeri S (2018) Synthesizing realistic neural population activity patterns using generative adversarial networks. arXiv 180300338. [Google Scholar]
  82. Morcos AS, Harvey CD (2016) History-dependent variability in population dynamics during evidence accumulation in cortex. Nat Neurosci 19:1672–1681. 10.1038/nn.4403 [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Mussa-Ivaldi FA. (1988) Do neurons in the motor cortex encode movement direction? an alternative hypothesis. Neurosci Lett 91:106–111. 10.1016/0304-3940(88)90257-1 [DOI] [PubMed] [Google Scholar]
  84. Nuyujukian P, Kao JC, Fan JM, Stavisky SD, Ryu SI, Shenoy KV (2014) Performance sustaining intracortical neural prostheses. J Neural Eng 11:066003. 10.1088/1741-2560/11/6/066003 [DOI] [PubMed] [Google Scholar]
  85. Orsborn AL, Dangi S, Moorman HG, Carmena JM (2012) Closed-loop decoder adaptation on intermediate time-scales facilitates rapid BMI performance improvements independent of decoder initialization conditions. IEEE Trans Neural Syst Rehabil Eng 20:468–477. 10.1109/TNSRE.2012.2185066 [DOI] [PubMed] [Google Scholar]
  86. Pandarinath C, Gilja V, Blabe CH, Nuyujukian P, Sarma AA, Sorice BL, Eskandar EN, Hochberg LR, Henderson JM, Shenoy KV (2015) Neural population dynamics in human motor cortex during movements in people with ALS. eLife 4:e07436. 10.7554/eLife.07436 [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Pandarinath C, Nuyujukian P, Blabe CH, Sorice BL, Saab J, Willett FR, Hochberg LR, Shenoy KV, Henderson JM (2017) High performance communication by people with paralysis using an intracortical brain-computer interface. eLife 6:e18554. 10.7554/eLife.18554 [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Pandarinath C, O'Shea DJ, Collins J, Jozefowicz R, Stavisky SD, Kao JC, Trautmann EM, Kaufman MT, Ryu SI, Hochberg LR, Henderson JM, Shenoy KV, Abbott LF, Sussillo D (2018) Inferring single-trial neural population dynamics using sequential auto-encoders. Nat Methods 15:805–815. 10.1038/s41592-018-0109-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Perge JA, Homer ML, Malik WQ, Cash S, Eskandar E, Friehs G, Donoghue JP, Hochberg LR (2013) Intra-day signal instabilities affect decoding performance in an intracortical neural interface system. J Neural Eng 10:036004. 10.1088/1741-2560/10/3/036004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Perich MG, Miller LE (2017) Altered tuning in primary motor cortex does not account for behavioral adaptation during force field learning. Exp Brain Res 235:2689–2704. 10.1007/s00221-017-4997-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Perich MG, Gallego JA, Miller LE (2017) A neural population mechanism for rapid learning. bioRxiv 138743 10.1101/138743 [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Petreska B, Yu B, Cunningham J, Santhanam G, Ryu S, Shenoy K, Sahani M (2011) Dynamical segmentation of single trials from population neural data. In: Advances in Neural Information Processing Systems, 24 Red Hook, NY: Neural Information Processing Systems. [Google Scholar]
  93. Phillips CG. (1975) Laying the ghost of ‘muscles versus movements’. Can J Neurol Sci 2:209–218. 10.1017/S031716710002028X [DOI] [PubMed] [Google Scholar]
  94. Pohlmeyer EA, Solla SA, Perreault EJ, Miller LE (2007) Prediction of upper limb muscle activity from motor cortical discharge during reaching. J Neural Eng 4:369–379. 10.1088/1741-2560/4/4/003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. Raposo D, Kaufman MT, Churchland AK (2014) A category-free neural population supports evolving demands during decision-making. Nat Neurosci 17:1784–1792. 10.1038/nn.3865 [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Remington ED, Narain D, Hosseini EA, Jazayeri M (2018) Flexible sensorimotor computations through rapid reconfiguration of cortical dynamics. Neuron 98:1005–1019.e1005. 10.1016/j.neuron.2018.05.020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Riehle A, Requin J (1989) Monkey primary motor and premotor cortex: single-cell activity related to prior information about direction and extent of an intended movement. J Neurophysiol 61:534–549. 10.1152/jn.1989.61.3.534 [DOI] [PubMed] [Google Scholar]
  98. Russo AA, Bittner SR, Perkins SM, Seely JS, London BM, Lara AH, Miri A, Marshall NJ, Kohn A, Jessell TM, Abbott LF, Cunningham JP, Churchland MM (2018) Motor cortex embeds muscle-like commands in an untangled population response. Neuron 97:953–966.e8. 10.1016/j.neuron.2018.01.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Sadtler PT, Quick KM, Golub MD, Chase SM, Ryu SI, Tyler-Kabara EC, Yu BM, Batista AP (2014) Neural constraints on learning. Nature 512:423–426. 10.1038/nature13665 [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Santhanam G, Yu BM, Gilja V, Ryu SI, Afshar A, Sahani M, Shenoy KV (2009) Factor-analysis methods for higher-performance neural prostheses. J Neurophysiol 102:1315–1330. 10.1152/jn.00097.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  101. Schwartz AB, Kettner RE, Georgopoulos AP (1988) Primate motor cortex and free arm movements to visual targets in three-dimensional space: I. Relations between single cell discharge and direction of movement. J Neurosci 8:2913–2927. 10.1523/JNEUROSCI.08-08-02913.1988 [DOI] [PMC free article] [PubMed] [Google Scholar]
  102. Scott SH. (2004) Optimal feedback control and the neural basis of volitional motor control. Nat Rev Neurosci 5:532–546. 10.1038/nrn1427 [DOI] [PubMed] [Google Scholar]
  103. Scott SH. (2008) Inconvenient truths about neural processing in primary motor cortex. J Physiol 586:1217–1224. 10.1113/jphysiol.2007.146068 [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Shenoy KV, Sahani M, Churchland MM (2013) Cortical control of arm movements: a dynamical systems perspective. Annu Rev Neurosci 36:337–359. 10.1146/annurev-neuro-062111-150509 [DOI] [PubMed] [Google Scholar]
  105. Sofroniew NJ, Flickinger D, King J, Svoboda K (2016) A large field of view two-photon mesoscope with subcellular resolution for in vivo imaging. eLife 5:e14472. 10.7554/eLife.14472 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Stavisky SD, Kao JC, Nuyujukian P, Ryu SI, Shenoy KV (2015) A high performing brain–machine interface driven by low-frequency local field potentials alone and together with spikes. J Neural Eng 12:036009. 10.1088/1741-2560/12/3/036009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  107. Stavisky SD, Kao JC, Ryu SI, Shenoy KV (2017a) Trial-by-trial motor cortical correlates of a rapidly adapting visuomotor internal model. J Neurosci 37:1721–1732. 10.1523/JNEUROSCI.1091-16.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. Stavisky SD, Kao JC, Ryu SI, Shenoy KV, Stavisky SD, Kao JC, Ryu SI, Shenoy KV (2017b) Motor cortical visuomotor feedback activity is initially isolated from Downstream targets in output-null neural state space dimensions article motor cortical visuomotor feedback activity is initially isolated from Downstream targets in output-null neural state space dimensions. Neuron 95:195–208.e9. 10.1016/j.neuron.2017.05.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Stevenson IH, Kording KP (2011) How advances in neural recording affect data analysis. Nat Neurosci 14:139–142. 10.1038/nn.2731 [DOI] [PMC free article] [PubMed] [Google Scholar]
  110. Stringer C, Pachitariu M, Steinmetz N, Reddy CB, Carandini M, Harris KD (2018) Spontaneous behaviors drive multidimensional, brain-wide population activity. bioRxiv 306019 10.1101/306019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Sussillo D, Churchland MM, Kaufman MT, Shenoy KV (2015) A neural network that finds a naturalistic solution for the production of muscle activity. Nat Neurosci 18:1025–1033. 10.1038/nn.4042 [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Sussillo D, Jozefowicz R, Abbott LF, Pandarinath C (2016a) LFADS: latent factor analysis via dynamical systems. arXiv 1608.06315. [Google Scholar]
  113. Sussillo D, Stavisky SD, Kao JC, Ryu SI, Shenoy KV (2016b) Making brain–machine interfaces robust to future neural variability. Nat Commun 7:13749. 10.1038/ncomms13749 [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Tanji J, Evarts EV (1976) Anticipatory activity of motor cortex neurons in relation to direction of an intended movement. J Neurophysiol 39:1062–1068. 10.1152/jn.1976.39.5.1062 [DOI] [PubMed] [Google Scholar]
  115. Taylor DM, Tillery SI, Schwartz AB (2002) Direct cortical control of 3D neuroprosthetic devices. Science 296:1829–1832. 10.1126/science.1070291 [DOI] [PubMed] [Google Scholar]
  116. Todorov E, Jordan MI (2002) Optimal feedback control as a theory of motor coordination. Nat Neurosci 5:1226–1235. 10.1038/nn963 [DOI] [PubMed] [Google Scholar]
  117. Velliste M, Perel S, Spalding MC, Whitford AS, Schwartz AB (2008) Cortical control of a robotic arm for self-feeding. Nature 453:1098–1101. 10.1038/nature06996 [DOI] [PubMed] [Google Scholar]
  118. Vyas S, Even-Chen N, Stavisky SD, Ryu SI, Nuyujukian P, Shenoy KV (2018) Neural population dynamics underlying motor learning transfer. Neuron 97:1177–1186.e3. 10.1016/j.neuron.2018.01.040 [DOI] [PMC free article] [PubMed] [Google Scholar]
  119. Wei Z, Inagaki H, Li N, Svoboda K, Druckmann S (2018) An orderly single-trial organization of population dynamics in premotor cortex predicts behavioral variability. bioRxiv 376830 10.1101/376830 [DOI] [PMC free article] [PubMed] [Google Scholar]
  120. Williams AH, Kim TH, Wang F, Vyas S, Ryu SI, Shenoy KV, Schnitzer M, Kolda TG, Ganguli S (2018) Unsupervised discovery of demixed, low-dimensional neural dynamics across multiple timescales through tensor component analysis. Neuron 98:1099–1115.e8. 10.1016/j.neuron.2018.05.015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  121. Wu W, Black MJ, Gao Y, Serruya M, Shaikhouni A, Donoghue J, Bienenstock E (2003) Neural decoding of cursor motion using a Kalman filter. In: Advances in Neural Information Processing Systems, 15, pp 133–140. Cambridge, MA; London: MIT. [Google Scholar]
  122. Yu BM, Cunningham JP, Santhanam G, Ryu SI, Shenoy KV, Sahani M (2009) Gaussian-process factor analysis for low-dimensional single-trial analysis of neural population activity. J Neurophysiol 102:614–635. 10.1152/jn.90941.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  123. Yuste R. (2015) From the neuron doctrine to neural networks. Nat Rev Neurosci 16:487–497. 10.1038/nrn3962 [DOI] [PubMed] [Google Scholar]
  124. Zhao Y, Park IM (2017) Variational latent gaussian process for recovering single-trial dynamics from population spike trains. Neural Comput 29:1293–1316. 10.1162/NECO_a_00953 [DOI] [PubMed] [Google Scholar]

Articles from The Journal of Neuroscience are provided here courtesy of Society for Neuroscience

RESOURCES