Skip to main content
bioRxiv
  • Home
  • About
  • Submit
  • ALERTS / RSS
Advanced Search
New Results

Interplay between external inputs and recurrent dynamics during movement preparation and execution in a network model of motor cortex

Ludovica Bachschmid-Romano, Nicholas G. Hatsopoulos, Nicolas Brunel
doi: https://doi.org/10.1101/2022.02.19.481140
Ludovica Bachschmid-Romano
1Department of Neurobiology, Duke University
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Nicholas G. Hatsopoulos
2Department of Organismal Biology and Anatomy, The University of Chicago
3Committee on Computational Neuroscience, The University of Chicago
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Nicolas Brunel
1Department of Neurobiology, Duke University
4Department of Physics, Duke University
5Duke Institute for Brain Sciences, Duke University
6Center for Cognitive Neuroscience, Duke University
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • For correspondence: nicolas.brunel@duke.edu
  • Abstract
  • Full Text
  • Info/History
  • Metrics
  • Preview PDF
Loading

Abstract

The primary motor cortex has been shown to coordinate movement preparation and execution through computations in approximately orthogonal subspaces. The underlying network mechanisms, and in particular the roles played by external and recurrent connectivity, are central open questions that need to be answered to understand the neural substrates of motor control. We develop a recurrent neural network model that recapitulates the temporal evolution of single-unit activity recorded from M1 of a macaque monkey during an instructed delayed-reach task. We explore the hypothesis that the observed dynamics of neural covariation with the direction of motion emerges from a synaptic connectivity structure that depends on the preferred directions of neurons in both preparatory and movement-related epochs. We constrain the strength both of synaptic connectivity and of external input parameters by using the data as well as an external input minimization cost. Our analysis suggests that the observed patterns of covariance are shaped by external inputs that are tuned to neurons’ preferred directions during movement preparation, and they are dominated by strong direction-specific recurrent connectivity during movement execution, in agreement with recent experimental findings on the relationship between motor–cortical and motor–thalamic activity, both before and during movement execution. We also demonstrate that the manner in which single-neuron tuning properties rearrange over time can explain the level of orthogonality of preparatory and movement-related subspaces. We predict that the level of orthogonality is small enough to prevent premature movement initiation during movement preparation; however, it is not zero, which allows the network to encode a stable direction of motion at the population level without direction-specific external inputs during movement execution.

Introduction

The activity of the primary motor cortex (M1) during movement preparation and execution plays a key role in the control of voluntary limb movement [1, 2, 3, 4, 5, 6, 7]. Classic studies of motor preparation were performed in a delayed-reaching task setting, showing that firing rates correlate with task-relevant parameters during the delay period, despite no movement occurring [8, 9, 10, 11, 12, 13, 14, 14, 15, 16, 17]. More recent works have shown that preparatory activity is also displayed before non-delayed movements [18], that it is involved in reach correction [19], and that when multiple reaches are executed rapidly and continuously, each upcoming reach is prepared by the motor cortical activity while the current reach is in action [20]. Preparation and execution of different reaches are thought to be processed simultaneously without interference in the motor cortex through computation along orthogonal dimensions [20]. Indeed, the preparatory and movement-related subspaces identified by linear dimensionality reduction methods are almost orthogonal [21] so that simple linear readouts that transform motor cortical activity into movement commands will not produce premature movement during the planning stage [22]. However, response patterns in these two epochs of motion are nevertheless linked, as demonstrated by the fact that a linear transformation can explain the flow of activity from the preparatory subspace to the movement subspace [21]. How this population-level strategy is implemented at the circuit level is still under investigation [23]. A related open question [24] is whether inputs from areas upstream to the primary motor cortex (such as from the thalamus and other cortical regions, here referred to as external inputs) that have been shown to be necessary to sustain movement generation [25] are specific to the type of movement being generated throughout the whole course of the motor action, or if they serve to set the initial conditions for the dynamics of the motor cortical network to evolve as shaped by recurrent connections [26, 27, 28, 29].

Here, we use a network modeling approach to explain the relationship between network recurrent connectivity, external inputs, and computations in orthogonal dimensions. Our analysis is based on recordings from M1 of a macaque monkey performing a delayed centerout reach task. One easily measurable feature of motor cortical activity during straight limb reaches is its covariance with the direction of motion (see [30, 31, 32, 33] but also [34, 35]). Recorded neurons are tuned to the direction of motion both during movement preparation and execution, but their tuning properties change over time ([36, 37, 38]). Interestingly, major changes in single neuron tuning happen when the activity flows from the preparatory to the movement-related subspaces. Yet, the direction of motion that we can decode at the population level is stable throughout the course of the motor action. To model these observations, we considered a recurrent neural network of rate-based neurons and we reduced the network dynamics to a few latent variables, or order parameters, that recapitulate the temporal evolution of neurons tuning properties. We then fitted the model to the data, by imposing that the model reproduce the observed dynamics of the order parameters, and that the energy cost associated with large external inputs is minimized. Our analysis suggests that during the delay period, information encoding the direction of movement is inherited by the network through external inputs that are tuned to the preferred directions of the neurons. During movement execution, external inputs are untuned and information regarding the movement direction is maintained via strong direction-specific recurrent connections. We will discuss how this prediction is in line with recent findings [39] showing that the co-firing patterns of pairs of motor-thalamic and motor-cortical neurons significantly correlate with the difference in the neurons’ preferred directions before, but not during, movement execution. Finally, we show how the specific way in which neurons tuning properties rearrange over time produces the observed level of orthogonality between the preparatory- and movement-related subspaces.

Results

Subjects and Task

We analyzed multi-electrode recordings from the primary motor cortex (M1) of two macaque monkeys performing a previously reported [40] instructed-delay, center-out reaching task. The monkey’s arm was on a two-link exoskeletal robotic arm, so that the position of the monkey’s hand controlled the location of a cursor projected onto a horizontal screen. The task consisted of three periods (Fig. 1): a hold period, during which the monkey was trained to hold the cursor on a center target and wait 500 ms for the instruction cue; an instruction period, where the monkey was presented with one of eight evenly spaced peripheral targets and continued to hold at the center for an additional 1,000–1,500 ms; a movement period, signalled by a go cue, were the monkey initiated the reach to the peripheral target. Successful trials where the monkeys reached the target were rewarded with a juice or water reward. The peripheral target was present on the screen throughout the whole instruction and movement periods. In line with previous studies (e.g., [21]), the preparatory and movement-related epochs were defined as two 300ms time intervals beginning, respectively, 100ms after target onset and 50ms before the start of the movement.

Figure 1:
  • Download figure
  • Open in new tab
Figure 1:

(a) Schematic of the center-out delayed reach task and definition of the preparatory (blue) and of the movement-related (red) epochs. Black circles represent the time of: target onset; go cue; start of movement; end of movement, averaged across trials. (b) Pairwise correlation of trial averaged preparatory activity plotted against the correlation of trial averaged movement-related activity, for each pair of neurons. The correlation coefficient of the correlations is 0.2 (P < 10−5). (c) Percentage of variance of the preparatory (blue) and movement-related (red) recorded activity explained by the first 11 principal components calculated from preparatory (top) and movement-related (bottom) trial averaged activity.

Correlations and tuning properties of neural activity

As previously reported in [21], the correlation coefficient between the activity of pairs of neurons during movement preparation weakly correlates with the correlation coefficient of the same two neurons during movement execution (Fig. 1.b); moreover, subspaces of neural activity dedicated to movement-preparation and execution are close to being orthogonal (Fig. 1.c). We aim to understand how computations during movement preparation and execution are linked, while correlation patterns are fundamentally reorganized between epochs. To this end, we first analyzed how neurons tuning properties relate in the two epochs of motion. Studies of motor cortex have shown that tuning to movement direction is not a time-invariant property of motor cortical neurons, but rather varies in time throughout the course of the motor action; single-neuron encoding of entire movement trajectories has been also reported (e.g., [36, 37]). We measured temporal variations in neurons preferred direction by binning time into 290ms time bins and fitting the binned trial-averaged spike counts as a function of movement direction with a cosine function. The cumulative density function of circular variances of preferred directions (Fig. 2.e) shows that the variability in preferred direction during the delay period alone and during movement execution alone is smaller than the variability during the entire duration of the task. This justifies our choice to characterize neurons tuning properties only in terms of their preferred direction during movement preparation and their preferred direction during movement execution. In each epoch of motion, pairs of neurons have positively correlated activity if their preferred direction is similar, and negatively correlated activity if their preferred direction is opposite (Fig. 2.f). Conversely, correlations of preparatory activities depend less strongly on the preferred direction during movement execution, and vice versa (Fig. 2.f). We next built a network model whose architecture is shaped by neurons tuning properties.

Figure 2:
  • Download figure
  • Open in new tab
Figure 2:

(a) Example of the trial averaged firing rate of one neuron during movement preparation (blue dots) as a function of the target location. The dotted line represent the corresponding cosine fit. (b) Same as in (a), but for the activity of the same neuron during movement execution. (c) Scatter plot between preferred directions for all neurons, defined as the location of the peak of the cosine tuning function; circular correlation coefficient: rθ = 0.4. (d) Scatter plot between the normalized amplitude of the cosine tuning curve for all neurons; correlation coefficient: rη ≈ 0.3. (e) Cumulative density function of the circular variance of preferred directions. For each neuron, preferred directions are computed in 290ms time bins across the whole duration of the task (grey line, 5 time bins); during the delay period (blue line, 3 time bins) and in the time interval going from the start to the end of the movement (red line, 2 time bins). (f) Pairwise correlation of trial averaged activity during movement preparation (blue) and execution (red) of pairs of neurons as a function of the difference ΔPD in their preferred directions during preparation (top) and execution (bottom).

Two-cylinder model

We considered a recurrent network model where two distinct but correlated maps, denoted by A and B, encode the direction of movement during movement preparation and during movement execution, respectively. We will refer to the units in the network as neurons, even though each unit in the model rather represents a group of M1 neurons with similar functional properties, and the connection between two units in our model represents the effective connection between the two functionally similar groups of neurons in M1. The model is an extension of the ring model: while previously studied ring models are defined on one [41] or multiple [42] one-dimensional circular spaces, we added another dimension to the feature space, representing the degree of participation of each neuron to encoding the circular variable. Neurons are identified by four coordinates: θA ∈ [0, 2π) and θB ∈ [0, 2π), representing their preferred direction during movement preparation and during movement execution, and ηA ∈ [0, 1] and ηB ∈ [0, 1], describing how strongly the neuron participates into encoding the direction of motion in the two epochs of movement. Maps A and B are thus two cylindrical maps defined by one angular coordinate θ and one linear coordinate η. Neurons receive input currents Iext both from outside of the network - either homogeneous or anisotropic in maps A and B – and recurrent inputs Irec, mediated by direction-specific cortical interactions. The strength of synaptic connections from a pre-synaptic neuron with preferred directions (θA, θB) and participation strengths (ηA, ηB) to a post-synaptic neuron with preferred directions Embedded Image and participation strengths Embedded Image is denoted by Embedded Image. Since the two maps encode two non-concurrent features of the motor action, we assume that the couplings encode the maps additively [43]. In analogy with previously studied ring models [41, 44, 42], the interactions are given by: Embedded Image j0 represents a uniform inhibitory term; Embedded Image and Embedded Image measure the amplitude of the sym-metric connections storing map A and map B, respectively; ja measures the amplitude of asymmetric connections from map A to map B. Later in this section (also, Fig. 3), we will describe how the network can encode the direction of movement through a localized pattern of activity. The last term of (1) contributes to align location of the bump in map B with the location of the bump in map A. In the mean-field limit of a very large network, the dynamics of the activity rate is defined by: Embedded Image where [ ]+ is the threshold-linear (a.k.a. relu) function. We set the time constant to τ = 25ms, which is of the same order of magnitude of the membrane time constant, and we checked that for values of τ in the range 10ms − 100ms our results did not quantitatively change. The total input to a neuron is Embedded Image where x represents the coordinates vector x = {θA, θB, ηA, ηB} and ρ(x) is the joint probability density of the preferred directions and participation strengths, that we will specify later. The mean-field formalism allowed us to reduce the dimensionality of the system to only five order parameters: the total average activity, r0(t); the spatial modulation of the activity profile in either maps, rA(t) and rB(t); the location of the peak of the activity profile in either map: ψA and ψB. The equations governing the temporal evolution of the order parameters are derived in the Methods.

Figure 3:
  • Download figure
  • Open in new tab
Figure 3:

(a) Schematic of the recurrent neural network model. The model is defined on a 4−dimensional space, with coordinates: θA, θB, representing neurons preferred direction during movement preparation and execution, respectively; and ηA, ηB, representing neurons degree of participation to encoding the direction of motion during movement preparation and execution, respectively. In the schematic, the direction of the arrows represents the neurons preferred direction, while its thickness represents the degree of participation. Synaptic connectivity is defined by equation 1: it contains both a symmetric component, that depends on the distance between preferred directions of pre and post-synaptic neurons separately for the two epochs, and an asymmetric component that depends on the distance between the preferred preparatory direction of the pre-synaptic neuron and the preferred execution direction of the post-synaptic one. (b) Localized 4−dimensional stationary network activity shown in three different cross-sections. Left: activity plotted on map A (θA, ηA) at fixed θB, ηB. Center: activity plotted on map B (θB, ηB) at fixed θB, ηB. Right: activity plotted as a function of θA, θB, at fixed ηA, ηB;. (c) Localized 4−dimensional stationary network activity plotted as a function of θA, at fixed θB, ηB and for different values of ηA. The activity profiles represent tuning curves of different neurons during movement preparation. (d) Localized 4−dimensional stationary network activity plotted as a function of θB, at fixed θA, ηA and for different values of ηB. The activity profiles represent tuning curves of different neurons during movement execution. The couplings parameters used to generate the activity shown in (b-c-d) correspond to the solution (d) in Fig. 6

We first characterize the model by studying the fixed-points of the network dynamics when subject to a constant and homogeneous external input. If the external field is independent on θA, θB, the fixed points can be of two kinds, depending on the value of the couplings parameters. One solution is homogeneous, meaning that the activity rate is independent on θA, θB; in this case, the order parameters measuring the spatial modulation of the activity are zero: rA = 0 and rB = 0 (see Methods for details). If the coupling parameters Embedded Image modulating the cosine terms in the couplings (1) exceed a certain threshold, the activity rate is zero in a subset of the (θA, θB, ηA, ηB)−space. The only nonzero stationary states are localized patterns of activity (bump), which can be localized either more strongly in map A (rA > rB > 0), in map B (rB > rA > 0) or at the same level in both maps (rA = rB > 0). Since maps A and B are correlated, the location of the stationary bump of activity is constrained to be the same in the two maps: ψA = ψB ≡ ψ, with arbitrary ψ. The system can relax to a continuous manifold of fixed points parameterized by ψ that are marginally stable. In this continuous attractor regime, the system can store in memory any direction of motion ψ, as the activity is localized in absence of tuned inputs. In the space of parameters Embedded Image, we computed the bifurcation surface separating the homogeneous from the bump phase. The result is shown in Fig. 4.c where for different values of ja we plotted the bifurcation line in the space Embedded Image, that is implicitly defined by the following set of inequalities: Embedded Image

Figure 4:
  • Download figure
  • Open in new tab
Figure 4:

(a) Joint distribution of coordinates θA and θB inferred from the data shown in Fig. 2.b. (b) Distribution of coordinates ηA and ηB inferred from the data shown in Fig. 2.d. ηA and ηB are assumed to be independent. (c) Phase diagram of the model with homogeneous and constant external inputs, shown as a function of the parameters Embedded Image, modulating the two symmetric terms in the couplings (see equation 1). Different curves correspond to bifurcation lines for different values of the parameter ja, modulating the asymmetric term in the couplings. The area underneath each line (gray) corresponds to the homogeneous phase; the area beyond each line (white) corresponds to the marginal phase, where the network exhibit a narrowly localized activity pattern even in absence of external tuned inputs.

⟨…⟩ denoting average over the distribution of θA, θB, ηA, ηB. Moreover, j0 has to be smaller than 1 to avoid rate instability. Examples of the cross-section of the four-dimensional stationary activity profile r(θA, θB, ηA, ηB) in the marginal phase are shown in Fig. 3. The activity as a function of θA, θB at fixed values of ηA, ηB can be strictly positive, with a full cosine modulation profile (e.g., Fig. 3.c-d, for ηA = 0.4, ηB = 0.4); this is a major difference with respect to the double ring model [42], where the activity has a cosine threshold profile at least in one of the two rings.

Time-dependent activity profiles

The activity of motor cortical neurons shows no signature of being in a stationary state but instead displays complex transients. To model the data, we assumed that the neurons are responding both to recurrent inputs and to fluctuating external inputs that can be either homogeneous or tuned to θA, θB, with peak at constant location ΦA = ΦB ≡ Φ. In the limit of an infinitely large network, we derived a simplified description of the dynamics, where recurrent inputs are replaced by effective local inputs, that take into account the average effect of the many recurrent inputs the neuron is receiving. The total input in (3) is rewritten as the sum three local inputs: the first one is homogeneous, the second is tuned to map A and the third is tuned to map B: Embedded Image where Embedded Image and where we have introduced the external fields parameters C0, CA, CB, ϵA, ϵB representing, respectively, the magnitude of: an homogeneous input; an untuned preparatory input (proportional to ηA but homogeneous in θA); an untuned execution input (proportional to ηB but homogeneous in θB); a tuned preparatory input (proportional to ηA and tuned to θA); and a tuned execution input (proportional to ηB and tuned to θB). Equations (2) and (5) show that the pattern of activity is localized at a constant location Φ throughout the dynamics, while its shape changes in time. For example, the activity pattern can evolve from being strongly localized in map A to being strongly localized in map B; accordingly, the order parameters measuring the spatial modulation in either map, rA, rB, are also time dependent. In order to draw a correspondence between the model and the data, we note that the activity of each neuron in the data corresponds to the activity rate at a specific coordinate θA, θB, ηA, ηB in the model. From equations (2) and (5) we see that, if we assume that changes in the external inputs happen on a time scale larger than τ, the modulation of the activity profile in map A at each time t is centered at θA − Φ and has amplitude proportional to ηA, while the modulation of the activity profile in map B is centered at θB − Φ and has amplitude proportional to ηB. Accordingly, for each recorded neuron θA and θB represent the location of the peak of the neuron’s tuning curve computed during the preparatory and movement-related epoch, respectively, while ηA and ηB are proportional to the amplitude of the tuning curves. From the empirical distribution of the amplitude and location of the tuning curves that we measured from data, we inferred the probability density of the four coordinates ρ(θA, θB, ηA, ηB) that we used in our mean-field analysis (see Fig. 4 and Methods for details). It factorizes as follow: Embedded Image where the distribution of preferred directions is well fitted by: Embedded Image as shown in (Fig. S1), and where ρpA(ηA), ρpB(ηB) are estimated non-parametrically using kernel density estimation. Next, we notice that either of the two fields IA(t) and IB(t) modulating the amplitude of the activity are a sum of two terms (6): the strength of tuned external inputs and the strength of recurrent inputs, which is proportional to the couplings parameters Embedded Image. Hence, the localized pattern of activity can be sustained by either strongly tuned external inputs or by strong recurrent connections.

External inputs inferred from the dynamics of the order parameters

The dynamics of the order parameters rA, rB computed from data (Fig.5) shows that during the delay period neural activity is spatially modulated in map A but not in map B, while during movement execution the activity is strongly modulated in map B, with the degree of modulation in map A slowly decreasing to very small values. This behaviour can be reproduced by the model for different choices of recurrent connections and external inputs parameters, ranging in between the following two opposite scenarios (Fig.5):

  • The couplings are not direction specific Embedded Image and the localized activity is sustained by an external input that is tuned to θA during movement preparation and to θB during movement execution.

  • The bump is self-sustained via strong recurrent connections and fluctuating homogeneous external inputs allow the bump to be localized in map A during movement preparation and in map B during movement execution.

The value of the parameters that best fit the data are inferred through minimization of a cost function (Etot) composed of two terms: one is the reconstruction error of the temporal evolution of the order parameters (Erec) and the other represents an energetic cost penalizing large external inputs (Eext): Embedded Image where α is an hyperparameter of the fitting algorithm. To reduce the degeneracy in the results of the fit, we impose not only that the model reconstruct the value of the order parameters r0, rA, rB, but also the value of two additional parameters. We denote them by r0A and r0B and they represent the overlap between the variable ηA (or ηB) and the neural activity. The result of the fit depends on the hyperparameter α (Fig.5). If α is chosen so that the two terms αErec and Eext have equal magnitude – after the cost function is minimized – then the inferred values of the couplings parameters are above, but very close to, the bifurcation surface. In absence of tuned external inputs, the activity sustained by recurrent connections is localized in map B. The inferred value of the external fields is shown in Fig. 6.e. During the delay period, an external input tuned to map A sustains the bump of activity localized in map A and sets the direction of motion; during movement execution, the activity is driven by homogeneous external inputs and the activity localized in map B is sustained by the strong recurrent connections. The correlation between the two maps allows the information about the direction of motion to be encoded stably throughout the whole duration of the task. On the other hand, if we chose a larger value of the hyperparameter α, so to impose a better reconstruction of the recorded dynamics but penalize less for large external inputs, the solutions that we obtain are still close to the bifurcation surface but below it, and a small input tuned to θB is present during movement execution. Importantly, the same analysis applied to a second dataset recorded from a different macaque monkey performing the same task yielded qualitatively similar results (see Fig. S3).

Figure 5:
  • Download figure
  • Open in new tab
Figure 5:

(a) Dynamics of the order parameters rA (degree of spatial modulation of the activity in map A) and rB (degree of spatial modulation in map B) computed from data (solid line; shaded area: ± SEM across the population). Black dots on the x-axes represent the trial-averaged time of: target onset, go cue, start of movement and end of movement. (b) Our model can qualitatively explain the switching trajectories of the order parameters for different values of the coupling parameters Embedded Image. We illustrate here the schematic of two opposite scenarios: very small couplings (c), and very strong couplings, away from the bifurcation line (d). (c) Very small couplings scenario: the observed dynamics of the order parameters can result from external inputs that are tuned to θA during movement preparation and to θB during movement execution. (d) Very strong couplings scenario: the strong recurrent connections sustain the localized pattern of activity; a change in untuned external inputs makes the activity to be localized along the θA during preparation and along θB during preparation.

Figure 6:
  • Download figure
  • Open in new tab
Figure 6:

Inferred dynamics of the external fields required to sustain the observed dynamics of the order parameters. (a-b-c) Scenario where neurons are connected by uniform inhibitory connections, in absence of direction specific couplings. (a) Couplings parameters Embedded Image are set to zero (orange dot on the phase diagram). The black line represents the bifurcation surface in the space Embedded Image, at ja = 0. (b) Dynamics of the external inputs inferred from the data. Gray line: homogeneous input; light blue (\ light red) line: input that is direction unspecific, but proportional to ηA (\ηB); dark blue (\ dark red) line: input that is tuned to the preferred direction θA (\θB) and proportional to ηA (\ηB). Black dots on the x-axes represent the trial-averaged time of: target onset, go cue, start of movement and end of movement. (c) Dynamics of the order parameters r0, r0A, r0B, rA, rB computed from data (thin line; shaded area: ± standard error across the population) and analytical prediction from mean-field analysis (thick line). (d-e-f-g-h-i) Both couplings parameters and external inputs are inferred from data, through minimization of a cost function composed of two terms: Etot = αErec + Eext. Erec represents the reconstruction error of the order parameters, while Eext is the magnitude of the external inputs required to sustain the activity. α is an hyperparameter of the fitting algorithm. (d) Value of the couplings parameters (orange dot) inferred by choosing the hyperparameter α in such a way that the two terms of the cost function have equal magnitude. The obtained couplings parameters are above, but close to, the bifurcation line. (e) Dynamics of the external inputs inferred from the data. (f) Dynamics of the order parameters r0, rA, rB computed from data (thin line) analytical prediction from mean-field analysis (thick line). (g) Value of the couplings parameters (orange dot) inferred by choosing the hyperparameter α in such a way that the term Erec has a slightly larger weight than the term Eext in the cost function. The obtained couplings parameters are below, but close to, the bifurcation line. (h) Dynamics of the external inputs inferred from the data. (i) Dynamics of the order parameters r0, rA, rB computed from data (thin line) analytical prediction from mean-field analysis (thick line).

Simulations of the dynamics of the network

The fitting procedure and the results of Fig. 5 are based on the equations governing the dynamics of the order parameters that we derived in the mean-field limit of infinitely many neurons. We next checked that a network of finite size with the same parameters that we inferred in the last section reproduces the dynamics seen in the data. We built a network of 16000 neurons by assigning to each neuron i the coordinates Embedded Image so to match the empirical distribution of coordinates of Fig. 3. In particular, the values of Embedded Image and Embedded Image are chosen to be equally spaced along the lines Embedded Image (see Fig. S5). Neuronal firing rates obeyed standard rate equations (29), were a noise term modelled as an Ornstein-Uhlenbeck process is added to the total input. Simulations of a network whose coupling parameters correspond to the solution of Fig. 6d reproduce well the dynamics of the order parameters (Fig. 7) and the location of the bump fluctuates only weakly around a stable location (Fig. S5). We analyzed the results of the simulations similarly as how we analyzed the data. In particular, we computed tuning curves during the preparatory and execution epochs and estimated the values Embedded Image from the location and the amplitude of the tuning functions. The reconstructed values of neurons tuning parameters are consistent with the values initially assigned to them when we built the network (Fig. S5); moreover, simulating the dynamics with additive noise produces tuning curves whose shape resembles the data (Fig. S6). Next, we added noise to the variables Embedded Image, to break the rotational symmetry of the model. If the network has coupling parameters above the bifurcation surface and no tuned external input is present during movement execution, the location of the bump after movement initiation starts to drift towards a few discrete attractors. When the noise is weak, the drift happen on a time scale that is much larger that the time of movement execution; the larger the level of the noise, the faster the drift. If we instead consider a solution like the one of Fig. 6.g, where tuned inputs are weak but non-zero, the location of the bump remains stable throughout the dynamics. Hence, adding stability constraints to the cost function that we used for our minimization procedure will favor network parameters that are close to, but below the bifurcation surface.

Figure 7:
  • Download figure
  • Open in new tab
Figure 7:

(a) Dynamics of the order parameters r0, rA, rB computed from data (thin line; shaded area: ± SEM of the population) and numerical simulations (thick line) of the dynamics of a finite-size network model with additive noise. The couplings parameters of the network correspond to the scenario of Fig. 6, panel d: they are above, but close to, the bifurcation line. (b) Percentage of variance of the preparatory (blue) and movement-related (red) activity from simulations explained by the first 11 principal components calculated from preparatory (top) and movement-related (bottom) trial-averaged activity. (c) The alignment index quantifies the degree of orthogonality between two subspaces. Top: alignment index between the preparatory and movement-related activities computed from data, compared to the randomized test (random alignment index, distribution in light gray and average in dark grey). Bottom: alignment index computed from simulations, for 100 subsets of 140 neurons each, sampled uniformly at random from the larger network we simulated; the gray histogram shows the average random alignment index for all subsets. (d) Pairwise correlation of the trial-averaged activity from simulations during movement preparation (blue) and execution (red) of pairs of neurons as a function of the difference in their preferred directions ΔθA (top) and ΔθB (bottom).

PCA subspaces dedicated to movement preparation and execution

We performed principal component analysis (PCA) on the trial averaged activity to show that the two subspaces that capture most of the variance of neural activity during movement preparation and execution are close to being orthogonal. After identifying the preparatory and movement-related principal components (PCs) (see Methods for details), we quantified the level of orthogonality of the two supspaces by the aligment index as defined in [21], that measures the percentage of variance of each epoch’s activity explained by both sets of PCs. Fig 7 shows that the alignment index is much smaller than the one computed between two subspaces drawn at random (random alignment index, explained in the Methods), both for the simulations and for the data. The level of orthogonality of the preparatory and movement-related subspaces can be understood as follows. The activity of the ring model encoding the value of a singular angular variable is two-dimensional in the Euclidean space. Similarly, the activity of the double-ring model encoding two distinct circular maps is four-dimensional. Our model is an extension of the double-ring model, where both the connectivity matrix and the external fields () are a sum of several terms, each one composed of an η-dependent term multiplying a θ-dependent term. The connectivity matrix is still rank-four, but its eigenvectors are modulated by the η variables. Although the dynamics is four-dimensional, we have shown that during movement preparation the activity is localized only in map A, while during movement execution it is predominantly localized in map B: in either epoch, we expect only two eigenvalues to explain most of the activity variance, as we indeed see from simulations in absence of additive noise (Fig. S7). The degree of orthogonality between the preparatory and movement-related subspaces is determined by the level of orthogonality between map A and map B, which can be quantified in terms of the following correlation: Embedded Image where brackets ⟨…⟩ denote the average with respect to the distribution of (7-8), and where we used the distribution of ηA, ηB that we inferred from data (Fig. 4) to compute the last term on the right-hand side. The correlation between maps is smaller with respect to both the case where ηA = 1, ηB = 1 for all neurons (CAB = 0.33) and the case where ηA, ηB are uniformly distributed (CAB = 0.25). Finally, the noise term added to the dynamics introduces extra random dimensions; as the dynamics gets higher dimensional, both the alignment index and the random alignment index get smaller.

To conclude, we observed that from the PCA analysis alone it is not possible to discriminate between the scenario where neurons tuning is induced by tuned external inputs from the one where it emerges from recurrent connections. Simulations of the network dynamics in the two scenarios yielded very similar results in terms of the variance captured by the first principal components (Fig. S7).

Discussion

Studies on the dynamics of motor cortical activity during delayed reach tasks have shown that the primary motor cortex employs an ‘orthogonal but linked strategy’ [22, 21] to coordinate planning and execution of movements. In this work, we explored the hypothesis that this strategy emerges as result of a specific Hebbian-like recurrent functional architecture, in which synaptic connections store information about two distinct patterns of activity that underlie movement preparation and movement execution. In a simplifying modeling setting based on recordings in Macaque monkeys performing a straight reach task, we characterized response patterns in terms of their covariance with the direction of motion. Hence, the preparatory and movement related patterns stored in the couplings are formalized in terms of two maps (A and B) in the space defined by two features of neurons tuning properties: their preferred direction and their level of participation to the population encoding. We inferred the distribution of these tuning features from data and showed that the degree of correlation between the two maps is small enough to allow for almost orthogonal subspaces, which is thought to be important for the preparatory activity not to cause premature movement; at the same time, having a non-zero correlation between maps allows the activity to flow from the preparatory to the movement-related subspaces with minimal external inputs.

By using a mean field analysis, we derived a simple description of neural dynamics in terms of a few order parameters that can be easily computed from data. Different combinations of the strength of direction-specific recurrent connections and of tuned external inputs allow the model to accurately reproduce the dynamics of the order parameters, ranging from a scenario where neurons tuning properties emerge solely from recurrent inputs to one where the motor cortex is simply integrating tuned external inputs. The addition of an external input strength minimization constraint breaks the degeneracy of the space of solutions, leading to a solution where synaptic couplings depend on the tuning properties of the pre- and post-synaptic neurons, in such a way that in the absence of a tuned input, neural activity is localized in map B. During movement preparation, an external input tuned to map A sustains a localized activity in map A, and sets the direction of motion encoded at the population level. During movement execution, movement direction is stably encoded at the population level, thanks to recurrent inputs that sustain a localized activity in map B. The correlation between maps A and B allows the activity in map B during movement execution to be localized around the same location as it was in map A during movement preparation. These results, based on a mean-field analysis valid for infinitely large networks, were confirmed by simulating the dynamics of a finite-size network of 16000 neurons, with additive noise whose level was chosen so that the dimensionality of the activity in the neural state space matched the one from data. The presence of noise did not disrupt the temporal evolution of the order parameters; on single trials, the direction of motion encoded by the network slightly diffuses around the value predicted by the mean field analysis, but remains stable for trial-averaged activity. However, our solution requires an implausible fine tuning of the recurrent connections. Heterogeneity in the connectivity causes a systematic drift of the encoded direction of motion on a typical time scales of seconds - the larger the structural noise in the couplings, the faster the drift, as has been extensively studied in the literature of continuous attractor models [45, 46, 47, 48, 49]. It has been shown that homeostatic mechanisms could compensate for the heterogeneity in cellular excitability and synaptic inputs to reduce systematic drifts of the activity [47] and that short-term synaptic facilitation in recurrent connections could also significantly improve the robustness of the model [48] – even when combined with short-term depression [49]. While a full characterization of our model in the presence of structural heterogeneity is beyond the scope of this work, we considered a second version of the optimization procedure to infer the model parameters that takes into account the stability of the encoded direction of motion with respect to perturbations in the couplings. We showed that a solution that improves the stability is one where tuned inputs are also present during movement execution. Interestingly, the inferred tuned inputs are still much weaker than the untuned ones, during movement execution. Also, the inferred direction-specific couplings are strong and amplify the weak external inputs tuned to map B, therefore still playing a major role into shaping the observed dynamics during movement execution.

Our prediction that external inputs are direction-specific during movement preparation but non-specific during movement execution agrees with several other studies on the activity of the primary motor cortex during limb movement. In particular, [28] showed that the changes in neural activity that characterize the transition from movement preparation to execution reflect when movement is made but are invariant to movement direction and type; [24] argued that external input transients signalling the motor cortical network a new movement direction are more strongly time-locked to target appearance as compared to the start of movement; finally, [39] measured the correlations in firing between motor thalamic (m-Th) and motor cortical cells, in monkeys performing the same task considered in the present work. In order to investigate if the observed cofiring patterns resulted from neurons tuning properties, the authors looked at the effect of the difference in preferred directions on cofiring patterns between couples of M1–mTh cells. Interestingly, the study reported significant correlation between cofiring patterns and difference in preferred direction solely before movement, but not during movement execution (see Appendix, Fig. S8 of ref. [39]).

Comparison with the ring model

The idea that the tuning properties of motor cortical neurons could emerge from directionspecific synaptic connections goes back to the work of Lukashin and Georgopulos [50]. However, it was with the theoretical analysis of the so called ring model [51, 41, 52, 44] that localized patterns of activity were formalized as attractor states of the dynamics in networks with strongly specific recurrent connections. Related models were later used to describe maintenance of internal representations of continuous variables in various brain regions [53, 54, 55, 56, 57, 58, 59, 43, 60, 61] and were extended to allow for storage of multiple continuous manifolds [62, 42, 63] to model the firing patterns of place cells the hippocampus of rodents exploring multiple environments. While our formalism is built on the same theoretical framework of these works, we would like to stress two main differences between our model and the ones previously considered in the literature. First, we studied the dynamic interplay between fluctuating external inputs and recurrent currents, that causes the activity to flow from the preparatory map to the movement-related one and, consequently, neurons tuning curves and order parameters to change over time, while maintaining stable encoding of the direction of motion at the population level. Moreover, we introduced an extra dimension representing the degree of participation of single neurons to the population encoding of movement direction. We have discussed how the presence of this dimension is key to having tuning curves whose shape resembles the one computed from data, and decreases the level of orthogonality between the subspaces dedicated to the preparatory and movement-related activity.

Extension to modeling richer movements

By analysing neural patterns of covariation with the direction of motion in the simple context of delayed straight reaches, we were able to model the temporal evolution of neural trajectories in low-dimensional latent spaces. However, neurons directional tuning properties have been shown to be influenced by many contextual factors that we neglected in our analysis [64, 65], to depend on the acquisition of new motor skills [66, 67, 68] and other features of movement such as the shoulder abduction/adduction angle even for similar hand kinematic profiles [34, 35]. Moreover,a large body of work [69, 35, 70, 64, 71, 72, 73] has shown that the activity in the primary motor cortex covary with many parameters of movement other than the hand kinematics – for a review, see [74]. More recent studies have also provided evidence that encoding of movement-related variables is insufficient to explain the rich dynamics of neurons in M1 [75, 76, 73, 77]. These and other works [74] argue that the largest signals in motor cortex might not ‘represent’ task-relevant variables at all, but if they do, surely not in a simple way that allows for different features to be decoded by linear decoders [77]. Based on these observations, we will now speculate on two possible ways of extending our model to more realistic scenarios.

In the current model, maps A and B each live in the low-dimensional space defined by neurons preferred directions: θA for map A and θB for map B. In a more general model, maps A and B could be parameterized by additional latent variables other than movement direction. For different tasks, the external input would have a different structure and select the dimensions that are relevant for the required pattern of kinematics and muscle activation. Neurons tuning to one single feature of the motor action would then vary across tasks in a highly non-linear manner, depending on the other latent variables relevant to that type of movement.

Another simplifying assumption allowed by the study of a delayed straight limb task is that the motor action can be clearly separated into a preparatory and movement related phase, where the direction of motion encoded at the population level is constant. We hypothesize that our simple model could also explain encoding of the direction of motion during more complex trajectories, that can be seen as a concatenation of short segments [20], where the network keeps an internal representation of the ongoing movement direction in map B while the next movement in a different direction is being prepared in map A. As suggested in [20], this kind of analysis can also provide a framework to interpret the population encoding of dynamic hand trajectories [78, 79, 36], that could emerge from the superposition of preparatory and movement-related activities relative to consecutive segments.

Author contributions

LBR and NB conceptualized the project. LBR performed the analytical calculations, implemented the computer codes and supporting algorithms. NB supervised the project. NGH provided the data. LBR and NB wrote the paper, and all authors edited the manuscript.

Methods

Analysis of the model

We studied the rate model with couplings defined by (1) with two complementary approaches. First, an analytic approximation based on mean-field arguments and valid in the limit of large network size allowed us to derive a low-dimensional description of the network dynamics in terms of a few latent variables, and to fit the model parameters to the data. Next, we checked that simulations of the dynamics of a network of N = 104 neurons reproduce the results that we derived in the limit N → ∞.

The mean-field equations are derived following the methods introduced in [41, 42]. In the limit where the number of neurons is large, the average activity of a neuron with coordinates θA, θB, ηA, ηB is described by equations (2) and (3). In the following, we will denote the integration measure by the shorthand Embedded Image

In order to derive a lower dimensional description of the dynamics, we rewrite (2) in terms of the average activity rate Embedded Image and of the second Fourier components of the activity rate modulated by ηA/B : Embedded Image

The phase ψA(/B) is defined so that the parameter rA(/B) is a real nonnegative number. Together with (10), the order parameters of the dynamics are thus defined by: Embedded Image

rA(/B) is interpreted as a measure of the spatial modulation of the activity profile, while ψA(/B) represents the position of the peak of the activity profile in map A(/B). From (2), we see that the order parameters evolve in time according to the following set of equations, Embedded Image where the total input (3) is rewritten as: Embedded Image

Stationary states for homogeneous external inputs

We first study the properties of the fixed point solution focusing on the scenario where the joint distribution (7) of the θA, θB is of the form Embedded Image which fits the empirical distribution of the data well for x = 2/3 (Fig. S1). For now, we leave the distributions ρpA(ηA) and ρpB(ηB) unspecified. We first consider the case where the external input to the network is a constant that is independent of θA, θB: Embedded Image

The stationary solutions of (2,14) are of the form: Embedded Image where we have defined the fields Embedded Image

Here, {r0, rA, rB} are solutions of the system (13) with the left hand side set to zero. The second term on the r.h.s in the last equation of (18) is obtained from (14) by observing that in the stationary state ψA = ψB if either ja > 0 or if θA and θB are correlated (as we will assume in the following). As in [41, 44], we can distinguish broad from narrow activity profiles. The term broad activity profile refers to the scenario where the activity of all of the neurons is above threshold, the dynamics is linear and the stationary state reduces to: Embedded Image

By inserting the above equation in (12), we find that the only solution is homogeneous over the maps θA and θB: Embedded Image where the notation ⟨. ⟩ represents an average over the measure dµ, i.e. Embedded Image

First, we notice that a nonzero homogeneous state is present if Embedded Image

Then, the stability of this state with respect to a small perturbation {δr0, δrA, δrB} can be studied by linearizing (13) around the stationary solution, at fixed ψA = ψB = 0. The resulting Jacobian matrix is Embedded Image where Embedded Image

The homogeneous solution (20) is stable if j0 < 1 and if the couplings parameters Embedded Image satisfy the system of inequalities (4). If j0 ≥ 1, the system undergoes an amplitude instability. For values of Embedded Image that exceed the threshold implicitly defined by (4), the dynamics is no longer linear and the activity profile at the fixed point is narrowly localized and characterized by positive stationary values of the order parameters rA, rB.

Tuned and time-dependent external inputs

We next considered the case where the system is subject to a tuned time-dependent external input. While the external inputs have time-varying magnitude, their location Φ is constant and is the same in the two maps: ΦA = ΦB ≡ Φ. Analogously to the ring model [44], the external input pins the location of the bump of activity. We assume that at the initial time the location of the bump equals the location of the external input, so that ψA(t) = ψB(t) = Φ at all times t. Although the position of the bump is stable, the direction modulation of the activity rates change in time in response to time-varying external inputs. We parameterized the external input in analogy with the recurrent input: Embedded Image where C0, CA, CB, ϵA, ϵB represent, respectively, the magnitude of: an homogeneous input; an input that is proportional to ηA (\ηB) but homogeneous in θA (\θB); an input that is proportional to ηA (\ηB) and tuned to θA (\θB). The total inputs can be rewritten as in (5).

Fitting the model to the data

Neurons activity rates were computed by smoothing the spike trains with a Gaussian kernel with s.d. of 25ms and averaging them across all trials with the same condition; Embedded Image denotes the rate of neuron i at time t for condition k, each condition corresponding to one of the 8 angular locations of the target on the screen. Since trials had highly variable length, we normalized the responses along the temporal dimension before averaging them over trials, as follows. We divided the activity into three temporal intervals: from the target onset to the go cue; from the go cue to the start of the movement; from the start of the movement to the end of the movement. For each interval, we normalized the response times to the average length of the interval across trials. We then aggregated the three intervals together. We defined the preparatory and execution epochs – denoted by A and B – as two 300ms time intervals beginning, respectively, 100ms after target onset and 50ms before the start of the movement, in line with [21]. Fig. S2 shows that our results do not change qualitatively when the lengths of the preparatory and execution intervals are increased. For each neuron i, we fitted the activity rate averaged across time within each epoch as a function of the angular position Φ of the target with a cosine function: Embedded Image where the parameters Embedded Image and Embedded Image represent the neuron’s preferred direction during the preparatory and execution epochs. In our the model, the direction modulation of the rates, see (6), is proportional to ηA\B, that measures how strongly the neuron participates in the two epochs of movement; hence, we defined Embedded Image to be proportional to the amplitude of the tuning curve: Embedded Image

The scatter plot of ηA, ηB (Fig. 2.d) shows an outlier with ηB ∼ 1. We checked that our results hold true if we discard that point. Fig. S1 shows that ηA and ηB are not significantly correlated with either θA, θB, nor θA−θB, while Fig. 2 shows that ηA and ηB have significant but weak mutual correlation. Based on these observations, we assumed for simplicity that ηA and ηB are independent variables. The order parameters (10,12) are computed at each time t by approximating the integrals with the sums: Embedded Image where N is the number of neurons and nc = 8 is the number of conditions. The angular location of the localized activity Embedded Image can be computed from (12) as Embedded Image

However, this estimate is strongly affected by the heterogeneity in the distribution of θA, θB - deviating from the rotational symmetry of the model. That is why in Fig. 3 we approximated Embedded Image by the k − th angular location of the target on the screen. Fig. S2 shows that computing Embedded Image by using either method does not affect the dynamics of the order parameters Embedded Image.

We assumed that the observed dynamics of the order parameters r0(r), rA(t), rB(t) obeys equations (13,14) with time-dependent external inputs of the form (21). We inferred the value of the parameters {C0(t), CA(t), CB(t), ϵA(t), ϵB(t)}t of the external fields and Embedded Image of the coupling matrix that allow us to reconstruct the dynamics of the order parameters r0, rA, rB computed from data; since this inference problem is undetermined, we required as further constraint that the model reconstruct the dynamics of the following two additional order parameters: Embedded Image

In this way, for given coupling parameters Embedded Image, we can uniquely identify the external fields parameters that produced the observed dynamics. Still, an equally good reconstruction of r0, rA, rB, r0A, r0B can be obtained for different choices of coupling parameters (3). Hence, we inferred the model parameters by minimizing a cost function composed of two terms: one that is proportional to the reconstruction error of the temporal evolution of the order parameters and the other that represents an energetic cost penalizing large external inputs.

Fitting the model to the data: details

The fitting procedure was divided in the following steps:

  1. The time interval T going from the target onset till the end of the movement was binned into Δt = 5ms time bins: T = {Δt1, Δt2, … ΔtT}.

  2. The couplings parameters were initialized to zero: Embedded Image. At the first time bin, the external fields parameters were initialized to zero: Embedded Image and the reconstructed order parameters (r0, …) were initialized to the order parameters estimated from the data Embedded Image: Embedded Image

  3. For each time step Δti, i = 2, …, T :

    • We started from the reconstructed order parameters at the previous time step: Embedded Image and we let the dynamical system (13, 25) with external fields parameters C0, CA, CB, ϵA, ϵB evolve for Δt = 5ms to estimate the order parameters at the current time step: Embedded Image

    • We inferred the value of the external fields parameters Embedded Image by minimizing the reconstruction error: Embedded Image that quantifies the difference between the order parameters estimated from the data and the reconstructed ones; note that the dependence of the cost function E on C0, CA, CB, ϵA, ϵB is implicitly contained in the reconstructed order parameters. We minimized the cost function (26) by using an interior point method algorithm [Byrd et al, 1999] starting from the initial condition Embedded Image we imposed that ϵA > 0, ϵB > 0 and added a L1 regularization term to stabilize the solution.

      The external fields inferred with step 3 depend on our initial choice of of the couplings parameters Embedded Image.

  4. Using step 3, the value of the couplings parameters Embedded Image is inferred by mini-mizing the cost function Embedded Image composed of two terms: the reconstruction error and a term that favors small external fields: Embedded Image

The hyperparameter α was chosen so that the two terms have equal magnitude after minimization (see Fig. S4 for details) and a = 0.001 is a parameter rescaling the external fields. The minimization is done using a surrogate optimization algorithm [Gutmann et al., 2001; Wang et al., 2014].

The result does not depend on the choice of the time bin Δt. Also, the result weakly depends on the time constant τ in the mean field equations (e.g., 13), if τ varies on in the range: 10 − 100ms. We set τ = 25ms.

Simulations of the model

We simulated the dynamics of a finite network of N neurons. To each neuron i, we assigned the variables Embedded Image as follows.

  • In the θ A\B-space, we sampled Nθ points Embedded Image equally spaced along the lines Embedded Image in such a way that their joint distribution matches the distribution of the data (8), as shown in Fig.S5.

  • In the η A\B-space, we drew Nη points Embedded Image at random from the empirical distribution ρpA(ηA)ρpB(ηB).

  • For i = 1, 2, … Nθ, we assigned to a block of Nη neurons the same coordinates Embedded Image in the θA\B-space, and all possible coordinates Embedded Image in the ηA\B-space, so that the overall number of neurons is N = NθNη.

The network dynamics we simulated is defined by the following stochastic differential equation: Embedded Image Embedded Image where Embedded Image Embedded Image Embedded Image

. W in (29) is a Wiener process, and the parameters Embedded Image set the magnitude of the noise fluctuations. The level of noise is chosen so that the results of the PCA analysis (see next section) match the data. Note that by setting the noise to zero and taking the limit N → ∞ we recover the mean-field equations (2). The results of Fig. 4 are obtained from a network of N = 16000 neurons; we simulate the network dynamics for 8 location of the external input Φ and 20 trials for each of the 8 conditions, i.e. 20 instances of the noisy dynamics. The order parameters shown in Fig. 4 are computed from single trial activity, and then averaged over trials. The correlation-based analysis, instead, is obtained from trial-averaged activity.

Correlations and PCA analysis

The correlation-based analysis explained in this section was performed both on the smoothed and trial-averaged spike trains from recordings, and on the trial-averaged activity rates from simulations. To compute signal correlations, we preprocessed the data as follows - the same procedure holds both for the recordings and for the simulations. For each neuron, we normalized the activity by its standard deviation (computed across all times and all conditions); then, we mean-centered the activity across conditions. The T = 300ms long preparatory activity for all C = 8 conditions was concatenated into a N × TC matrix denoted by P, and the movement-related activity was grouped into an analogous matrix M. We obtained correlation matrices relative to preparatory and movement related activity by computing the correlations between the rows of the respective matrices. We then identified the prep-PCs and move-PCs by performing PCA separately on the matrices P and M. The degree of orthogonality between the prep- and move-subspaces was quantified by the Alignment Index A [21], measuring the amount of variance of the preparatory activity explained by the first K move-PCs: Embedded Image where Emov is the matrix defined by the top K move-PCs, Cprep is the covariance matrix of the preparatory activity and sprep(i) is the i-th eigenvalue of Cprep. K was set to the number of principal components needed to explain 88% of the execution activity variance. Hence, the Alignment Index ranges from 0 (orthogonal subspaces) to 1 (aligned subspaces). As random test, we computed the Random Alignment Index between two sets of K dimensions drawn at random within the space occupied by neural activity, using the Monte Carlo procedure described in [21]. We performed the same analysis on both the data (K = 12) and the model (K = 9) trial averaged activity. Since the number of recorded neurons was of the order of Ndata ∼ 102, while the simulated network was composed of Nsim ∼ 104 neurons, we computed signal correlations and performed the PCA analysis on a subset of Ndata neurons randomly sampled from the larger simulated network. We repeated this procedure 100 times and averaged the Alignment Index and the Random Alignment Index across the 100 random samples.

We also quantified the roational structure present in the data by applying the jPCA [26] dimensionality reduction technique to both the simulated activity and the recordings, and showed (Fig. S8) that the trajectories from simulations qualitatively resembled the ones from data when projected onto the dimensions that capture rotational dynamics.

Figure S1:
  • Download figure
  • Open in new tab
Figure S1:

(a) Distribution of θA − θB computed from recorded data (orange) and data generated from the distribution: 1+2 cos(θA −θB)/3 (gray). (b) Variables ηA, ηB plotted as a function of θA −θB. The circular correlation is r = 0.13 (P = 0.3), for both ηA and ηB. (c) Variables ηA, ηB plotted as a function of θA. The circular correlation is r = 0.08 (P = 0.6) for ηA and r = 0.06 (P = 0.7) for ηB. (d) Variables ηA, ηB plotted as a function of θB. The circular correlation is r = 0.01 (P = 0.9) for ηA and r = 0.1 (P = 0.2) for ηB. (e) Scatter plot of the amplitude of a cosine function fitted to the tuning curve of the preparatory activity vs the R2 coefficient of the fit. The neurons with larger tuning amplitude are the ones with higher R2. (f) Same as in (c) but for movement-related activity.

Figure S2:
  • Download figure
  • Open in new tab
Figure S2:

For two different definitions of the preparatory (blue segment) and movement-related (red segment) epochs: (a) vs (b), we show the corresponding distributions of θA, θB: (c) vs (e); the distributions of ηA, ηB: (d) vs (f); the dynamics of the order parameters: (g) vs (i); and the results of the PCA analysis: (h) vs (l).

Figure S3:
  • Download figure
  • Open in new tab
Figure S3:

Analysis based on data recorded from monkey Rj. (a) Couplings parameters and (b) dynamics of the external inputs inferred from data. (c) Dynamics of the order parameters (solid line: data, shaded area: ± standard error across the population; dotted line: analytical prediction).

Figure S4:
  • Download figure
  • Open in new tab
Figure S4:

The two terms composing the cost function Etot = αErec + Eext (see) as a function of α, obtained from fitting the model to data recorded from monkey Rk (a) and Rj (b). In both cases, the hyperparameter α is chosen so that the two terms composing the cost function have equal magnitude. (c) Cost function Etot computed at different values of Embedded Image and Embedded Image. (d) Same as in (c), but for monkey Rj.

Figure S5:
  • Download figure
  • Open in new tab
Figure S5:

(a) To build the network architecture, we assigned to 16000 neurons coordinates ηA, ηB (top) and θA, θB (bottom) drawn from the their respective empirical distribution. (b) Scatter plot of the coordinates initially assigned to the neurons (x-axes) vs the corresponding variables that we computed from simulations of the network activity during the preparatory and movement-related epochs (y-axes). Points clustered on the diagonal show self-consistency of the model. (c) Dynamics of the order parameters ψA (top) and ψB (bottom) for 8 different locations Φ of the external input; ψA and ψB represent the location of the bump of activity in map A and B, respectively.

Figure S6:
  • Download figure
  • Open in new tab
Figure S6:

Examples of tuning curves during the preparatory (blue) and movement-related (red) epochs computed from simulations (a) and from data (b).

Figure S7:
  • Download figure
  • Open in new tab
Figure S7:

(a) PCA analysis on the results of simulations for the solution of Fig.7.g (strong couplings) when no noise is added to the dynamics. (b) Alignment index quantifying the degree of orthogonality between the preparatory and movement-related subspaces from simulations, for the same solution as in (a): strong couplings, no noise added to the dynamics. Orange: alignment index for 100 subsets of 140 neurons each, sampled uniformly at random from the larger network we simulated; gray: the average random alignment index for each subset. In absence of noise, the dynamics is confined onto a four-dimensional space; the random alignment index quantifies the alignment between two two-dimensional subspaces drawn at random within the four-dimensional space occupied by neural activity. Note that the alignment index is significantly smaller than the random test. (c) PCA analysis on the results of simulations for the solution of Fig.7.g (strong couplings) and (d) of Fig.7.a (zero couplings).

Figure S8:
  • Download figure
  • Open in new tab
Figure S8:

(a) Trial-averaged activity rates during movement execution (from 50ms before the start of the movement, to 250ms after the start of the movement) projected onto the first two jPCA dimensions, defined as in [26]. In order to compare the data to the theory, we averaged the activity of all trials corresponding to the same condition – defining 8 groups (each denoted in a different color) corresponding to the 8 different conditions. However, hand kinematics corresponding to the activity within the same group can differ quite substantially, even if they end up at the same target location. This is a major difference with respect to the analysis of [26], where many more groups were defined, each group containing the activity corresponding to very similar hand trajectories. (b) Same as in (a), but for activity from simulations.

Acknowledgements

We thank Jason MacLean, Alex P. Vaz, Subhadra Mokashe, Alessandro Sanzeni for helpful discussions, and Stephen H. Scott for pointing the work of [39] to our attention. This work has been supported by NIH R01NS104898.

References

  1. [1].↵
    Edward V Evarts. Relation of pyramidal tract activity to force exerted during voluntary movement. Journal of neurophysiology, 31(1):14–27, 1968.
    OpenUrlCrossRefPubMedWeb of Science
  2. [2].↵
    Ian Q Whishaw, Sergio M Pellis, Boguslaw Gorny, Bryan Kolb, and Wolfram Tetzlaff. Proximal and distal impairments in rat forelimb use in reaching follow unilateral pyramidal tract lesions. Behavioural brain research, 56(1):59–76, 1993.
    OpenUrlCrossRefPubMedWeb of Science
  3. [3].↵
    Ian Q Whishaw. Loss of the innate cortical engram for action patterns used in skilled reaching and the development of behavioral compensation following motor cortex lesions in the rat. Neuropharmacology, 39(5):788–805, 2000.
    OpenUrlCrossRefPubMedWeb of Science
  4. [4].↵
    Michael SA Graziano, Charlotte SR Taylor, and Tirin Moore. Complex movements evoked by microstimulation of precentral cortex. Neuron, 34(5):841–851, 2002.
    OpenUrlCrossRefPubMedWeb of Science
  5. [5].↵
    Thomas C Harrison, Oliver GS Ayling, and Timothy H Murphy. Distinct cortical circuit mechanisms for complex forelimb movement and motor map topography. Neuron, 74(2):397–409, 2012.
    OpenUrlCrossRefPubMedWeb of Science
  6. [6].↵
    Stephen H Scott. The computational and neural basis of voluntary motor control and planning. Trends in cognitive sciences, 16(11):541–549, 2012.
    OpenUrlCrossRefPubMedWeb of Science
  7. [7].↵
    Andrew R Brown and G Campbell Teskey. Motor cortex is functionally organized as a set of spatially distinct representations for complex movements. Journal of Neuroscience, 34(41):13574–13585, 2014.
    OpenUrlAbstract/FREE Full Text
  8. [8].↵
    Doug P Hanes and Jeffrey D Schall. Neural control of voluntary movement initiation. Science, 274(5286):427–430, 1996.
    OpenUrlAbstract/FREE Full Text
  9. [9].↵
    JUN Tanji and Edward V Evarts. Anticipatory activity of motor cortex neurons in relation to direction of an intended movement. Journal of neurophysiology, 39(5):1062–1068, 1976.
    OpenUrlCrossRefPubMedWeb of Science
  10. [10].↵
    Mark M Churchland, Afsheen Afshar, and Krishna V Shenoy. A central source of movement variability. Neuron, 52(6):1085–1096, 2006.
    OpenUrlCrossRefPubMedWeb of Science
  11. [11].↵
    Mark M Churchland, Gopal Santhanam, and Krishna V Shenoy. Preparatory activity in premotor and motor cortex reflects the speed of the upcoming reach. Journal of neurophysiology, 96(6):3130–3146, 2006.
    OpenUrlCrossRefPubMedWeb of Science
  12. [12].↵
    Julie Messier and John F Kalaska. Covariation of primate dorsal premotor cell activity with direction and amplitude during a memorized-delay reaching task. Journal of neurophysiology, 84(1):152–165, 2000.
    OpenUrlCrossRefPubMedWeb of Science
  13. [13].↵
    Michael C Dorris, Martin Pare, and Douglas P Munoz. Neuronal activity in monkey superior colliculus related to the initiation of saccadic eye movements. Journal of Neuroscience, 17(21):8566–8579, 1997.
    OpenUrlAbstract/FREE Full Text
  14. [14].↵
    Paul W Glimcher and David L Sparks. Movement selection in advance of action in the superior colliculus. Nature, 355(6360):542–545, 1992.
    OpenUrlCrossRefPubMedWeb of Science
  15. [15].↵
    ROBERT H Wurtz and MICHAEL E Goldberg. Activity of superior colliculus in behaving monkey. 3. cells discharging before eye movements. Journal of Neurophysiology, 35(4):575–586, 1972.
    OpenUrlCrossRefPubMedWeb of Science
  16. [16].↵
    Timothy R Darlington, Jeffrey M Beck, and Stephen G Lisberger. Neural implementation of bayesian inference in a sensorimotor behavior. Nature neuroscience, 21(10):1442–1451, 2018.
    OpenUrl
  17. [17].↵
    Timothy R Darlington and Stephen G Lisberger. Mechanisms that allow cortical preparatory activity without inappropriate movement. Elife, 9:e50962, 2020.
    OpenUrl
  18. [18].↵
    Antonio H Lara, Gamaleldin F Elsayed, Andrew J Zimnik, John P Cunningham, and Mark M Churchland. Conservation of preparatory neural events in monkey motor cortex regardless of how movement is initiated. elife, 7:e31826, 2018.
    OpenUrlCrossRefPubMed
  19. [19].↵
    K Cora Ames, Stephen I Ryu, and Krishna V Shenoy. Simultaneous motor preparation and execution in a last-moment reach correction task. Nature communications, 10(1):1–13, 2019.
    OpenUrl
  20. [20].↵
    Andrew J Zimnik and Mark M Churchland. Independent generation of sequence elements by motor cortex. Nature neuroscience, 24(3):412–424, 2021.
    OpenUrl
  21. [21].↵
    Gamaleldin F Elsayed, Antonio H Lara, Matthew T Kaufman, Mark M Churchland, and John P Cunningham. Reorganization between preparatory and movement population responses in motor cortex. Nature communications, 7(1):1–15, 2016.
    OpenUrl
  22. [22].↵
    Matthew T Kaufman, Mark M Churchland, Stephen I Ryu, and Krishna V Shenoy. Cortical activity in the null space: permitting preparation without movement. Nature neuroscience, 17(3):440–448, 2014.
    OpenUrlCrossRefPubMed
  23. [23].↵
    Ta-Chu Kao, Mahdieh S Sadabadi, and Guillaume Hennequin. Optimal anticipatory control as a theory of motor preparation: a thalamo-cortical circuit model. Neuron, 109(9):1567–1581, 2021.
    OpenUrl
  24. [24].↵
    Peter J Malonis, Nicholas G Hatsopoulos, Jason N MacLean, and Matthew T Kaufman. M1 dynamics share similar inputs for initiating and correcting movement. bioRxiv, 2021.
  25. [25].↵
    Britton A Sauerbrei, Jian-Zhong Guo, Jeremy D Cohen, Matteo Mischiati, Wendy Guo, Mayank Kabra, Nakul Verma, Brett Mensh, Kristin Branson, and Adam W Hantman. Cortical pattern generation during dexterous movement is input-driven. Nature, 577(7790):386–391, 2020.
    OpenUrlCrossRefPubMed
  26. [26].↵
    Mark M Churchland, John P Cunningham, Matthew T Kaufman, Justin D Foster, Paul Nuyujukian, Stephen I Ryu, and Krishna V Shenoy. Neural population dynamics during reaching. Nature, 487(7405):51–56, 2012.
    OpenUrlCrossRefPubMedWeb of Science
  27. [27].↵
    Krishna V Shenoy, Maneesh Sahani, and Mark M Churchland. Cortical control of arm movements: a dynamical systems perspective. Annual review of neuroscience, 36:337–359, 2013.
    OpenUrlCrossRefPubMedWeb of Science
  28. [28].↵
    Matthew T Kaufman, Jeffrey S Seely, David Sussillo, Stephen I Ryu, Krishna V Shenoy, and Mark M Churchland. The largest response component in the motor cortex reflects movement timing but not movement type. Eneuro, 3(4), 2016.
  29. [29].↵
    Saurabh Vyas, Matthew D Golub, David Sussillo, and Krishna V Shenoy. Computation through neural population dynamics. Annual Review of Neuroscience, 43:249–275, 2020.
    OpenUrlCrossRefPubMed
  30. [30].↵
    Apostolos P Georgopoulos, Andrew B Schwartz, and Ronald E Kettner. Neuronal population coding of movement direction. Science, 233(4771):1416–1419, 1986.
    OpenUrlAbstract/FREE Full Text
  31. [31].↵
    Andrew B Schwartz, Ronald E Kettner, and Apostolos P Georgopoulos. Primate motor cortex and free arm movements to visual targets in three-dimensional space. i. relations between single cell discharge and direction of movement. Journal of Neuroscience, 8(8):2913–2927, 1988.
    OpenUrlAbstract/FREE Full Text
  32. [32].↵
    Apostolos P Georgopoulos, Joseph T Lurito, Michael Petrides, Andrew B Schwartz, and Joe T Massey. Mental rotation of the neuronal population vector. Science, 243(4888):234–236, 1989.
    OpenUrlAbstract/FREE Full Text
  33. [33].↵
    Apostolos P Georgopoulos, Masato Taira, and Alexander Lukashin. Cognitive neurophysiology of the motor cortex. Science, 260(5104):47–52, 1993.
    OpenUrlAbstract/FREE Full Text
  34. [34].↵
    Stephen H Scott and John F Kalaska. Reaching movements with similar hand paths but different arm orientations. i. activity of individual cells in motor cortex. Journal of neurophysiology, 77(2):826–852, 1997.
    OpenUrlCrossRefPubMedWeb of Science
  35. [35].↵
    Stephen H Scott, Paul L Gribble, Kirsten M Graham, and D William Cabel. Dissociation between hand motion and population vectors from neural activity in motor cortex. Nature, 413(6852):161–165, 2001.
    OpenUrlCrossRefPubMedWeb of Science
  36. [36].↵
    Nicholas G Hatsopoulos, Qingqing Xu, and Yali Amit. Encoding of movement fragments in the motor cortex. Journal of Neuroscience, 27(19):5105–5114, 2007.
    OpenUrlAbstract/FREE Full Text
  37. [37].↵
    Mark M Churchland and Krishna V Shenoy. Temporal complexity and heterogeneity of single-neuron activity in premotor and motor cortex. Journal of neurophysiology, 97(6):4235–4257, 2007.
    OpenUrlCrossRefPubMedWeb of Science
  38. [38].↵
    Jörn Rickert, Alexa Riehle, Ad Aertsen, Stefan Rotter, and Martin P Nawrot. Dynamic encoding of movement direction in motor cortical neurons. Journal of Neuroscience, 29(44):13870–13882, 2009.
    OpenUrlAbstract/FREE Full Text
  39. [39].↵
    Abdulraheem Nashef, Rea Mitelman, Ran Harel, Mati Joshua, and Yifat Prut. Areaspecific thalamocortical synchronization underlies the transition from motor planning to execution. Proceedings of the National Academy of Sciences, 118(6), 2021.
  40. [40].↵
    Doug Rubino, Kay A Robbins, and Nicholas G Hatsopoulos. Propagating waves mediate information transfer in the motor cortex. Nature neuroscience, 9(12):1549–1557, 2006.
    OpenUrlCrossRefPubMedWeb of Science
  41. [41].↵
    Rani Ben-Yishai, R Lev Bar-Or, and Haim Sompolinsky. Theory of orientation tuning in visual cortex. Proceedings of the National Academy of Sciences, 92(9):3844–3848, 1995.
    OpenUrlAbstract/FREE Full Text
  42. [42].↵
    Sandro Romani and Misha Tsodyks. Continuous attractors with morphed/correlated maps. PLoS Comput Biol, 6(8):e1000869, 2010.
    OpenUrl
  43. [43].↵
    Alexei Samsonovich and Bruce L McNaughton. Path integration and cognitive mapping in a continuous attractor neural network model. Journal of Neuroscience, 17(15):5900–5920, 1997.
    OpenUrlAbstract/FREE Full Text
  44. [44].↵
    David Hansel and Haim Sompolinsky. 13 modeling feature selectivity in local cortical circuits, 1998.
  45. [45].↵
    Misha Tsodyks and Terrence Sejnowski. Associative memory and hippocampal place cells. International journal of neural systems, 6:81–86, 1995.
    OpenUrlCrossRefPubMed
  46. [46].↵
    Kechen Zhang. Representation of spatial orientation by the intrinsic dynamics of the head-direction cell ensemble: a theory. Journal of Neuroscience, 16(6):2112–2126, 1996.
    OpenUrlAbstract/FREE Full Text
  47. [47].↵
    Alfonso Renart, Pengcheng Song, and Xiao-Jing Wang. Robust spatial working memory through homeostatic synaptic scaling in heterogeneous cortical networks. Neuron, 38(3):473–485, 2003.
    OpenUrlCrossRefPubMedWeb of Science
  48. [48].↵
    Vladimir Itskov, David Hansel, and Misha Tsodyks. Short-term facilitation may stabilize parametric working memory trace. Frontiers in computational neuroscience, 5:40, 2011.
    OpenUrl
  49. [49].↵
    Alexander Seeholzer, Moritz Deger, and Wulfram Gerstner. Stability of working memory in continuous attractor networks under the control of short-term plasticity. PLoS computational biology, 15(4):e1006928, 2019.
    OpenUrl
  50. [50].↵
    Alexander V Lukashin and Apostolos P Georgopoulos. A dynamical neural network model for motor cortical activity during movement: population coding of movement trajectories. Biological cybernetics, 69(5):517–524, 1993.
    OpenUrlPubMedWeb of Science
  51. [51].↵
    Shun-ichi Amari. Dynamics of pattern formation in lateral-inhibition type neural fields. Biological cybernetics, 27(2):77–87, 1977.
    OpenUrlCrossRefPubMedWeb of Science
  52. [52].↵
    David C Somers, Sacha B Nelson, and Mriganka Sur. An emergent model of orientation selectivity in cat visual cortical simple cells. Journal of Neuroscience, 15(8):5448–5465, 1995.
    OpenUrlAbstract/FREE Full Text
  53. [53].↵
    Kechen Zhang. Representation of spatial orientation by the intrinsic dynamics of the head-direction cell ensemble: a theory. Journal of Neuroscience, 16(6):2112–2126, 1996.
    OpenUrlAbstract/FREE Full Text
  54. [54].↵
    A David Redish, Adam N Elga, and David S Touretzky. A coupled attractor model of the rodent head direction system. Network: computation in neural systems, 7(4):671, 1996.
    OpenUrl
  55. [55].↵
    Bruce L McNaughton, Carol A Barnes, Jason L Gerrard, Katalin Gothard, Min W Jung, James J Knierim, H Kudrimoti, Y Qin, WE Skaggs, M Suster, et al. Deciphering the hippocampal polyglot: the hippocampus as a path integration system. The Journal of experimental biology, 199(1):173–185, 1996.
    OpenUrlAbstract/FREE Full Text
  56. [56].↵
    H Sebastian Seung, Daniel D Lee, Ben Y Reis, and David W Tank. Stability of the memory of eye position in a recurrent network of conductance-based model neurons. Neuron, 26(1):259–271, 2000.
    OpenUrlCrossRefPubMedWeb of Science
  57. [57].↵
    Misha Tsodyks. Attractor neural network models of spatial maps in hippocampus. Hippocampus, 9(4):481–489, 1999.
    OpenUrlCrossRefPubMedWeb of Science
  58. [58].↵
    Marcelo Camperi and Xiao-Jing Wang. A model of visuospatial working memory in prefrontal cortex: recurrent network and cellular bistability. Journal of computational neuroscience, 5(4):383–405, 1998.
    OpenUrlCrossRefPubMedWeb of Science
  59. [59].↵
    Albert Compte, Nicolas Brunel, Patricia S Goldman-Rakic, and Xiao-Jing Wang. Synaptic mechanisms and network dynamics underlying spatial working memory in a cortical network model. Cerebral cortex, 10(9):910–923, 2000.
    OpenUrlCrossRefPubMedWeb of Science
  60. [60].↵
    SM Stringer, TP Trappenberg, ET Rolls, and IETd Araujo. Self-organizing continuous attractor networks and path integration: one-dimensional models of head direction cells. Network: Computation in Neural Systems, 13(2):217–242, 2002.
    OpenUrl
  61. [61].↵
    Yoram Burak and Ila R Fiete. Accurate path integration in continuous attractor network models of grid cells. PLoS computational biology, 5(2):e1000291, 2009.
    OpenUrl
  62. [62].↵
    Francesco P Battaglia and Alessandro Treves. Attractor neural networks storing multiple space representations: a model for hippocampal place fields. Physical Review E, 58(6):7738, 1998.
    OpenUrl
  63. [63].↵
    Rémi Monasson and Sophie Rosay. Transitions between spatial attractors in place-cell models. Physical review letters, 115(9):098101, 2015.
    OpenUrl
  64. [64].↵
    M-C Hepp-Reymond, M Kirkpatrick-Tanner, L Gabernet, H-X Qi, and B Weber. Context-dependent force coding in motor and premotor cortical areas. Experimental brain research, 128(1-2):123–133, 1999.
    OpenUrlCrossRefPubMedWeb of Science
  65. [65].↵
    RB Muir and RN Lemon. Corticospinal neurons with a special role in precision grip. Brain research, 261(2):312–316, 1983.
    OpenUrlCrossRefPubMedWeb of Science
  66. [66].↵
    Rony Paz, Thomas Boraud, Chen Natan, Hagai Bergman, and Eilon Vaadia. Preparatory activity in motor cortex reflects learning of local visuomotor skills. Nature neuroscience, 6(8):882–890, 2003.
    OpenUrlCrossRefPubMedWeb of Science
  67. [67].↵
    Chiang-Shan Ray Li, Camillo Padoa-Schioppa, and Emilio Bizzi. Neuronal correlates of motor performance and motor learning in the primary motor cortex of monkeys adapting to an external force field. Neuron, 30(2):593–607, 2001.
    OpenUrlCrossRefPubMedWeb of Science
  68. [68].↵
    SP Wise, SL Moody, KJ Blomstrom, and AR Mitz. Changes in motor cortical activity during visuomotor adaptation. Experimental Brain Research, 121(3):285–299, 1998.
    OpenUrlCrossRefPubMedWeb of Science
  69. [69].↵
    Robert Ajemian, Daniel Bullock, and Stephen Grossberg. Kinematic coordinates in which motor cortical cells encode movement direction. Journal of Neurophysiology, 84(5):2191–2203, 2000.
    OpenUrlPubMedWeb of Science
  70. [70].↵
    Paul L Gribble and Stephen H Scott. Overlap of internal models in motor cortex for mechanical loads during reaching. Nature, 417(6892):938–941, 2002.
    OpenUrlCrossRefPubMedWeb of Science
  71. [71].↵
    Emanuel Todorov. Direct cortical control of muscle activation in voluntary arm movements: a model. Nature neuroscience, 3(4):391–398, 2000.
    OpenUrlCrossRefPubMedWeb of Science
  72. [72].↵
    RN Holdefer and LE Miller. Primary motor cortical neurons encode functional muscle synergies. Experimental Brain Research, 146(2):233–243, 2002.
    OpenUrlCrossRefPubMedWeb of Science
  73. [73].↵
    Lauren E Sergio, Catherine Hamel-Pâquet, and John F Kalaska. Motor cortex neural correlates of output kinematics and kinetics during isometric-force and arm-reaching tasks. Journal of neurophysiology, 94(4):2353–2378, 2005.
    OpenUrlCrossRefPubMedWeb of Science
  74. [74].↵
    Stephen H Scott. The role of primary motor cortex in goal-directed movements: insights from neurophysiological studies on non-human primates. Current opinion in neurobiology, 13(6):671–677, 2003.
    OpenUrlCrossRefPubMedWeb of Science
  75. [75].↵
    Jonathan A Michaels, Benjamin Dann, and Hansjörg Scherberger. Neural population dynamics during reaching are better explained by a dynamical system than representational tuning. PLoS computational biology, 12(11):e1005175, 2016.
    OpenUrl
  76. [76].↵
    Abigail A Russo, Sean R Bittner, Sean M Perkins, Jeffrey S Seely, Brian M London, Antonio H Lara, Andrew Miri, Najja J Marshall, Adam Kohn, Thomas M Jessell, et al. Motor cortex embeds muscle-like commands in an untangled population response. Neuron, 97(4):953–966, 2018.
    OpenUrl
  77. [77].↵
    Karen E Schroeder, Sean M Perkins, Qi Wang, and Mark M Churchland. Cortical control of virtual self-motion using task-specific subspaces. bioRxiv, pages 2019–12, 2021.
  78. [78].↵
    Yoram Ben-Shaul, Rotem Drori, Itay Asher, Eran Stark, Zoltan Nadasdy, and Moshe Abeles. Neuronal activity in motor cortical areas reflects the sequential context of movement. Journal of Neurophysiology, 91(4):1748–1762, 2004.
    OpenUrlCrossRefPubMedWeb of Science
  79. [79].↵
    Xiaofeng Lu and James Ashe. Anticipatory activity in primary motor cortex codes memorized movement sequences. Neuron, 45(6):967–973, 2005.
    OpenUrlCrossRefPubMedWeb of Science
Back to top
PreviousNext
Posted February 19, 2022.
Download PDF
Email

Thank you for your interest in spreading the word about bioRxiv.

NOTE: Your email address is requested solely to identify you as the sender of this article.

Enter multiple addresses on separate lines or separate them with commas.
Interplay between external inputs and recurrent dynamics during movement preparation and execution in a network model of motor cortex
(Your Name) has forwarded a page to you from bioRxiv
(Your Name) thought you would like to see this page from the bioRxiv website.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Share
Interplay between external inputs and recurrent dynamics during movement preparation and execution in a network model of motor cortex
Ludovica Bachschmid-Romano, Nicholas G. Hatsopoulos, Nicolas Brunel
bioRxiv 2022.02.19.481140; doi: https://doi.org/10.1101/2022.02.19.481140
Digg logo Reddit logo Twitter logo Facebook logo Google logo LinkedIn logo Mendeley logo
Citation Tools
Interplay between external inputs and recurrent dynamics during movement preparation and execution in a network model of motor cortex
Ludovica Bachschmid-Romano, Nicholas G. Hatsopoulos, Nicolas Brunel
bioRxiv 2022.02.19.481140; doi: https://doi.org/10.1101/2022.02.19.481140

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Subject Area

  • Neuroscience
Subject Areas
All Articles
  • Animal Behavior and Cognition (3707)
  • Biochemistry (7835)
  • Bioengineering (5709)
  • Bioinformatics (21372)
  • Biophysics (10616)
  • Cancer Biology (8218)
  • Cell Biology (11990)
  • Clinical Trials (138)
  • Developmental Biology (6794)
  • Ecology (10435)
  • Epidemiology (2065)
  • Evolutionary Biology (13920)
  • Genetics (9736)
  • Genomics (13119)
  • Immunology (8183)
  • Microbiology (20092)
  • Molecular Biology (7886)
  • Neuroscience (43219)
  • Paleontology (322)
  • Pathology (1285)
  • Pharmacology and Toxicology (2270)
  • Physiology (3367)
  • Plant Biology (7263)
  • Scientific Communication and Education (1317)
  • Synthetic Biology (2012)
  • Systems Biology (5554)
  • Zoology (1136)