Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2020 Feb 28;16(2):e1007414. doi: 10.1371/journal.pcbi.1007414

Stochastic optimal open-loop control as a theory of force and impedance planning via muscle co-contraction

Bastien Berret 1,2,3,*, Frédéric Jean 4
Editor: Adrian M Haith5
PMCID: PMC7065824  PMID: 32109941

Abstract

Understanding the underpinnings of biological motor control is an important issue in movement neuroscience. Optimal control theory is a leading framework to rationalize this problem in computational terms. Previously, optimal control models have been devised either in deterministic or in stochastic settings to account for different aspects of motor control (e.g. average behavior versus trial-to-trial variability). While these approaches have yielded valuable insights about motor control, they typically fail in explaining muscle co-contraction. Co-contraction of a group of muscles associated to a motor function (e.g. agonist and antagonist muscles spanning a joint) contributes to modulate the mechanical impedance of the neuromusculoskeletal system (e.g. joint viscoelasticity) and is thought to be mainly under the influence of descending signals from the brain. Here we present a theory suggesting that one primary goal of motor planning may be to issue feedforward (open-loop) motor commands that optimally specify both force and impedance, according to noisy neuromusculoskeletal dynamics and to optimality criteria based on effort and variance. We show that the proposed framework naturally accounts for several previous experimental findings regarding the regulation of force and impedance via muscle co-contraction in the upper-limb. Stochastic optimal (closed-loop) control, preprogramming feedback gains but requiring on-line state estimation processes through long-latency sensory feedback loops, may then complement this nominal feedforward motor command to fully determine the limb’s mechanical impedance. The proposed stochastic optimal open-loop control theory may provide new insights about the general articulation of feedforward/feedback control mechanisms and justify the occurrence of muscle co-contraction in the neural control of movement.

Author summary

This study presents a novel computational theory to explain the planning of force and impedance (e.g. viscoelasticity) in the neural control of movement. It assumes that one main goal of motor planning is to elaborate feedforward motor commands that determine both the force and the impedance required for the task at hand. These feedforward motor commands (i.e. that are defined prior to movement execution) are designed to minimize effort and variance costs considering the uncertainty arising from sensorimotor or environmental noise. A major outcome of this mathematical framework is the explanation of muscle co-contraction (i.e. the concurrent contraction of a group of muscles involved in a motor function). Muscle co-contraction has been shown to occur in many situations but previous modeling works struggled to account for it. Although effortful, co-contraction contributes to increase the robustness of motor behavior (e.g. small variance) upstream of sophisticated optimal closed-loop control processes that require state estimation from delayed sensory feedback to function. This work may have implications regarding our understanding of the neural control of movement in computational terms. It also provides a theoretical ground to explain how to optimally plan force and impedance within a general and versatile framework.

Introduction

Optimal control theory is a leading framework for understanding biological motor behavior in computational terms [14]. Historically, this research has been carried out along two lines. On the one hand, deterministic optimal control (DOC) theory focused on the planning stage and sought to explain average motor behaviors in humans or animals. The minimum jerk and minimum torque change models are well-known representatives of this line of research [5, 6], which provided researchers with simple models accounting for the formation of average trajectories (e.g. bell-shaped velocity profiles in reaching tasks). This laid the foundations for more advanced studies like inverse optimal control ones, where the goal is to recover relevant optimality criteria from (averaged) experimental motion data [7, 8]. On the other hand, stochastic optimal control (SOC) theory was used to account for the variability of biological movement observed across multiple repetitions of the same task [911]. The noise that affects the neuromusculoskeletal system, and the uncertainty it induces about movement performance, are taken into account in this approach [12, 13]. This class of model can also be used to explain motor planning (e.g. via the specification of feedback gains prior to movement onset) but the genuine motor commands are only revealed along the course of the movement, once the current state of the system has been optimally estimated (e.g. hand/joint positions, velocities etc.). The SOC theory led to a number of valuable predictions among which the minimal intervention principle, stating that errors are corrected on-line only when they affect the goal of the task [9].

However, both of these approaches fail at accounting for a fundamental motor control strategy used by the central nervous system (CNS) and often referred to as co-contraction or co-activation of muscles groups (see [14] for a recent review). This frequent phenomenon is known since more than a century and the work of Demenÿ [15], and has been investigated extensively since then. There is now a strong evidence that co-contraction is voluntarily used by the CNS in a number of tasks, especially those requiring high stability, robustness or end-point accuracy [1618]. Co-contraction indeed contributes to modulate the mechanical impedance of the neuromusculoskeletal system. For instance, co-contraction can drastically increase the apparent joint stiffness by a factor 4 to 7 [19]. This effect results both from the summation of intrinsic stiffness of muscles around a common joint [20, 21] and reflexes [2224]. The former short-range stiffness implements an instantaneous (feedback-free) mechanism. The latter implements both a short-latency (low-level) feedback mechanism via fast-conducting mono- or oligo-synaptic spinal pathways [response latency at muscle level ~20-40 ms after a mechanical perturbation] and a long-latency (high-level) feedback mechanism via transcortical pathways [response latency ~50-100 ms]. The two above-mentioned approaches (DOC and SOC) are not able to account for co-contraction in a principled way for fundamentally distinct reasons. First, co-contraction contributes to modulate the effective limb’s impedance (e.g. joint viscoelasticity), whose actual effect can only be seen when unexpected perturbations are applied to the limb [19, 25]. As there are no such random perturbations in deterministic models, they will usually not predict co-contraction. Indeed, there is no functional gain at co-contracting opposing muscles in those models. Co-contraction just appears as a waste of energy considering that such models typically aim at minimizing effort or energy-like costs [26, 27]. Therefore, whenever a deterministic model exhibits co-contraction, it is an artifact of muscle modeling (e.g. due to response times of muscle activation dynamics) that does not serve any task-oriented, functional purpose. In SOC models, the presence of sensorimotor and environmental noise is taken into account so that co-contraction could become a relevant strategy regarding disturbance rejection and task achievement. However, SOC controllers typically exhibit reciprocal muscle activation patterns on average because they also minimize (expected) effort costs (e.g. see Fig. 2 in [28] or Fig. 3a in [29]), and correct errors using sensory feedback and reciprocal activations that are less costly than co-contraction. A few studies have nevertheless attempted to predict co-contraction from the SOC framework. These studies had to rely on advanced noise models explicitly reducing signal-dependent variance during co-contraction or on advanced viscoelastic muscle models yielding co-contraction without clear task dependency or functional purpose [29, 30]. More fundamentally, a closed-loop optimal control scheme requires optimal state estimation combining delayed sensory signals with an efferent copy of the motor command –the latter being converted into state variables via forward internal models– [31]. The neural substrate underlying SOC is thought to involve the long-latency transcortical pathway passing through the primary motor cortex [3234]. This may seem to contrast with the feedforward nature of impedance that has been demonstrated in several studies [16, 18, 3537]. However, the planning of optimal feedback gains may be viewed as a form of feedforward control of impedance in SOC models [38]. The main difference with co-contraction is that control via feedback gains critically depends on the ability of the CNS to form accurate estimates of the current system state. As this ability may be limited in some cases (e.g. unpredictable interaction with the environment, unstable task or too fast motion), co-contraction may constitute an alternative strategy to regulate mechanical impedance without the involvement of high-level, long-latency feedback mechanisms. In this vein, several studies on deafferented monkeys (without feedback circuitry at all) suggested that an equilibrium point/trajectory resulting from the co-contraction of opposing muscles was preprogrammed by the CNS during point-to-point movements without sight of the arm [3942]. Similar conclusions were drawn with deafferented patients who were able to perform relatively accurate reaching movements without on-line vision –if allowed to see their arm transiently prior to movement execution– [43]. Furthermore, neurophysiological studies seem to agree that muscle co-contraction has a central origin with little contribution from spinal mechanisms [14, 19, 44]. Noticeably, during co-contraction of antagonistic muscles, disynaptic reciprocal inhibition has been shown to be reduced by central signals [45, 46]. This highlights the singularity of muscle co-contraction in impedance control and departs from the reciprocal activations predicted by standard models based on DOC or SOC theories. For these reasons, co-contraction may be a critical feature of descending motor commands (i.e. open-loop control in computational terms) which may serve to generate stable motor behaviors ahead of the optimal closed-loop control machinery.

In this paper, we thus propose a novel stochastic optimal control framework to determine force and impedance –via muscle co-contraction– at the stage of motor planning. Our approach lies in-between DOC and SOC theories from a mathematical standpoint and we refer to it as stochastic optimal open-loop control (SOOC) theory to stress that we consider stochastic dynamical systems controlled by open-loop, deterministic controls [47]. This work is in the vein of seminal motor control studies [5, 48, 49]. We generalize and extend these approaches to the planning of upper-limb movement within a versatile mathematical framework that can handle a variety of motor tasks, types of noise, nonlinear limb dynamics and cost functions. The proposed theory primarily accounts for co-contraction as a means to modulate the apparent mechanical impedance of the musculoskeletal system via feedforward, descending motor commands that do not require any advanced on-line estimation of the system state. Although we use the term open-loop –in the sense of control theory– we do not necessarily exclude the role of reflexes that contribute to the spring-like behavior of intact muscles beyond their short-range stiffness. However, we do exclude from this open-loop terminology all the optimal closed-loop control processes integrating sensory data during movement execution through transcortical feedback loops [32]. The critical role of SOC is rather attributed to those long-latency, sophisticated and task-dependent motor responses that are triggered by the CNS to correct large-enough external perturbations [33, 34].

Materials and methods

Our working hypothesis is that both force and mechanical impedance are planned by the brain via descending motor commands. To illustrate our purpose, we will focus on the control of arm posture and movement, and compare the predictions made by our framework to existing experimental data. In this work, the major premise is to assume open-loop control (which makes sense at the stage of the motor planning process) while acknowledging the stochastic nature of the neuromusculoskeletal system. We shall illustrate that this formulation of motor planning as a SOOC problem naturally accounts for optimal muscle co-contraction and impedance control without the need to estimate the state of the system during movement execution.

To introduce the SOOC theory, we first revisit the seminal work of [48]. Hogan considered the problem of maintaining the forearm in an unstable upright posture in presence of uncertainty and without feedback about the system state. The forearm was modeled as an inverted pendulum in the gravity field, actuated by a pair of antagonistic muscles as follows:

Iθ¨=T(u1-u2)-K(u1+u2)θ-bθ˙+mglcsin(θ)+Gη (1)

where θ is the joint angle (0° being the upright orientation of the forearm and a dot above a variable standing for its time-derivative), I is the moment of inertia, m, lc, and g are respectively the mass, length to the center-of-mass and gravity acceleration, b is a damping parameter, and η is some noise (typically Gaussian). Parameters T and K are constants –as well as the noise factor G for the moment– and ui are the “neural” non-negative inputs to the flexor (i = 1) and extensor (i = 2) muscles. With this simplified model, Hogan showed that the optimal open-loop controls (ui(t)i=1‥2) that should be used to maintain the forearm in the unstable upright position while minimizing an expected cost based on effort and variance led to some optimal amount of co-contraction (i.e. u1 = u2 > 0). The variable stiffness property of muscles, and the fact that stiffnesses of opposing muscles add, allowed to maintain this unstable posture even without on-line feedback about the actual system state. This minimal example captures a crucial feature for our subsequent theoretical developments: the controlled system to obtain this result involved interaction between control and state components (i.e. terms in ui θ). Without gravity (g = 0) or with linearization of gravitational torque (e.g. sin(θ) ≈ θ), this type of system is called “bilinear” in control theory and it will be the simplest class of systems for which the SOOC framework makes original and relevant predictions regarding force and impedance planning. For linear control systems, which are often assumed in the motor control literature for simplicity, no difference with a deterministic approach would be observed. In the following, we build upon these ideas to model movement planning (not only posture as in Hogan’s initial work) and extend the method to more general nonlinear dynamics (not only one degree-of-freedom or bilinear dynamics as in this example).

Stochastic optimal open-loop control for bilinear systems

Consider stochastic dynamics with bilinear drift of the form:

dxt=[(A+i=1pNiui(t))xt+Bu(t)]dt+G(u(t),t)dωt (2)

with ωt being a m-dimensional standard Brownian motion. The stochastic state is denoted by xtRn and the deterministic control is denoted by u(t)Rp. The matrix G(u(t),t)Rn×m can account for noise with both constant and signal-dependent variance.

In the simplest setting, our goal is to find the optimal open-loop control u(t) that minimizes a quadratic expected cost of the form:

C(u)=E[0tf(u(t)TRu(t)+xtTQxt)dt+xfTQfxf]. (3)

where R is a positive definite matrix and Q, Qf are positive semi-definite matrices, all of appropriate dimensions. Note that because u(⋅) is a deterministic function by hypothesis, the related integral value can be taken outside the expectation operator.

We assume that the system has an initial state distribution that is known, x0, at the initial time (hence state estimation from sensory feedback is at least required initially for motor planning). Time tf is the total movement duration, which can be fixed a priori or left free. For such a bilinear system, xt will be a Gaussian process because the associated stochastic differential equation is actually linear (since u(t) is deterministic in the drift and diffusion). Therefore, the process xt can be fully determined by propagation of its mean and covariance.

The propagation of the mean and covariance of the process xt (denoted respectively by m(t)=E[xt] and P(t)=E[etetT] with et = xtm(t)) are given by the following ordinary differential equations (see [50] for example):

{m˙=(A+iNiui)m+BuP˙=[A+iNiui]P+P[A+iNiui]T+GGT. (4)

The latter equation shows explicitly the dependence of the covariance propagation on the control u.

Next, a simple calculation shows that the expected cost C(u) in Eq 3 can be rewritten as follows:

C(u)=0tf(uTRu+mTQm+trace(QP))dt+mfTQmf+trace(QfPf) (5)

Therefore, we have just shown that the initial SOOC problem reduces to an exactly-equivalent DOC problem, the state of which is composed of the elements of the mean and covariance of the stochastic process xt.

It must be noted that this equivalent deterministic problem has nonlinear dynamics but a quadratic cost. This constraint of a quadratic cost is however not critical as any Lagrangian depending on the mean and covariance could be added to the original cost function. For example, the following more general type of costs could be considered as well:

C(u)=E[0tf(L(m(t),u(t))+xtTQxt)dt+xfTQfxf]. (6)

The deterministic term L(m, u) can be used for instance to implement a minimum hand jerk on the mean behaviour, in agreement with the deterministic control literature [5]. Note also that the term xtTQxt could be replaced by (xt-m(t))TQ¯(xt-m(t)) to introduce a penalty on the state covariance alone in the equivalent DOC problem, in agreement with the minimum variance model [49]. Terminal state constraints or path constraints could also be added on the mean and covariance of the state process xt without any difficulty but they are not described here. Typically, this could be useful to impose hard constraints on the mean state to reach and/or on the covariance state to set a desired final accuracy. Such modeling choices will be illustrated in the subsequent arm movement and posture simulations, and can easily be handled within the equivalent DOC framework. Remarkably, optimal solutions of such DOC problems can be obtained via efficient existing numerical methods (e.g. [51]).

Stochastic optimal open-loop control for general nonlinear systems

We now consider more general stochastic dynamics of the form

dxt=f(xt,u(t),t)dt+G(xt,u(t),t)dωt. (7)

An example of such nonlinear system would be the system of Eq 1 (due to the gravitational torque). Multijoint arms also exhibit complex nonlinear dynamics due to inertial, centripetal, Coriolis, and gravitational torques. Nonlinearities will also arise for more advanced musculoskeletal models of the upper-limb. Therefore, we need a method to solve SOOC problems for the general class of nonlinear stochastic systems described in Eq 7.

Here we thus seek for a deterministic control u(t) minimizing the expectation of the above quadratic cost function (see Eq 6) and acting on the stochastic dynamics of Eq 7. The random process xt is not necessarily Gaussian anymore. However, mean and covariance are still variables of major interest for movement control and their propagation would yield significant information about both mean behaviour and its variability. Actually we have shown in [47] that via statistical linearization techniques the control u(t) can be approximated by the solution of a DOC problem involving the propagation of the mean and covariance.

For instance, with Gaussian statistical linearization based on first order Taylor approximations, computations in [47] show that the dynamics of the mean and covariance in the corresponding DOC problem are:

{m˙(t)=f(m(t),u(t),t),P˙(t)=F(m(t),u(t),t)P(t)+P(t)F(m(t),u(t),t)T+E[G(xt,u(t),t)G(xt,u(t),t)T] (8)

where F(m(t),u(t),t)=fx(m(t),u(t),t).

If G = G(u(t), t) models constant and signal-dependent noise, then the latter expectation simply equates to G(u(t), t)G(u(t), t)T. For a more general term, such as G(xt, u(t), t), more terms may be used to approximate the covariance propagation and the reader is referred to [52] for more information.

In summary, approximate solutions of the original SOOC problem can also be obtained from an associated DOC problem based on the mean and covariance of a process approximating the two first moments of the original process xt. Then, state-of-the-art DOC solvers can be used to find numerical solutions and model other constraints if desired (e.g. set a desired final positional variance or final mean position as a hard constraint…). The accuracy of these approximate solutions can be tested by simulating the original stochastic equation (Eq 7) with the obtained optimal control, and by comparing the evolution of the mean and covariance with Monte Carlo sampling techniques.

Stochastic optimal open-loop control in the neural control of movement

The mathematical SOOC framework being formally introduced, we are now left with modeling choices to describe the effects of co-contraction. On the one hand, one may consider an end-effector or joint level description of the stiffness-like property of the neuromusculoskeletal system (e.g. [48] or the r- and c-commands in [14]). On the other hand, one may consider more advanced models representing the multiple muscles crossing each joint, the activation of which will modulate both the apparent stiffness of the musculoskeletal system and net joint torques (e.g. [53]). This choice is related to the hierarchical control hypothesis, as discussed in [4, 14] for instance. As this choice is still elusive, we will consider both joint and muscle levels of description. In particular, this will highlight the generality and consistency of the proposed theory beyond specific modeling choices.

Joint level modeling: Explicit description of force and impedance planning

In this paragraph, we extend Hogan’s model presented above to account for the control of movement. Consider again the forearm model given in Eq 1. To simplify the derivations, we note that the state of a joint can be modified only in two ways: it can either (1) be moved to another position via changes of net torques or (2) be stiffened with no apparent motion via co-contraction [14]. Accordingly, the forearm dynamics can be rewritten as

Iθ¨=τ(t)-Ksκ(t)(θ(t)-Θ(t))-Kdκ(t)(θ˙(t)-Θ˙(t))-bθ˙+mglcsin(θ)+Gη, (9)

where τ(t)R is the net joint torque and κ(t)R+ modulates the joint stiffness (two control variables). The function Θ(t) serves as a reference trajectory which is not present in Hogan’s original formulation but is critical to change the limb’s working position. A potential justification could be that intact muscles behave like “nonlinear springs with modifiable zero-length” [54]. Hence, we assume that the resultant joint-level effect of this characteristics allows to set an equilibrium joint position or trajectory. In addition to stiffness, damping also seems to be modified through co-contraction [55]. We thus added a term in factor of Kd=IKs such that the damping ratio, in terms of a second-order model, was constant (here equal to 1/2, e.g. see [56, 57]).

In order to determine the reference trajectory Θ(t), there are several choices. For instance, one could consider a third control variable to choose Θ(t), by adding an equation such as Θ˙=u(t). This reference trajectory might be very simple (e.g. steady state or linear). The drawback is to introduce a third control variable, the choice of which seems rather elusive (e.g. what cost function on it). Alternatively, a better choice may be to consider reference trajectories that are themselves solutions of the joint-level dynamics. In this case, we assume that Θ(t) satisfies the rigid body dynamics

IΘ¨=τ(t)-bΘ˙+mglcsin(Θ) (10)

with Θ(0) = θ(0). Hence, if we define Δ(t) = θ(t) − Θ(t), one can derive the following system from Eqs 9 and 10:

{IΘ¨=τ(t)-bΘ˙+mglcsin(Θ)IΔ¨=-Ksκ(t)Δ-Kdκ(t)Δ˙-bΔ˙+mglc(sin(Δ+Θ)-sin(Θ))+Gη (11)

with Δ(0) = 0, Θ(0) = θ(0) and Δ˙(0)=0, Θ˙(0)=0. The advantage of this formulation, relying on a reference angle Θ and deviations Δ from it, is that only two controls are needed, namely τ(t) that specifies the net joint torque and κ(t) that specifies the joint stiffness (and damping). As such, this modeling implements a separate control of force (via reciprocal commands) and impedance (via co-contraction commands), which is compatible with several experimental findings [18, 19, 58, 59].

Now assume that the goal is to minimize an expected cost of the form

C(τ,κ)=E[0tf(τ2+ακ2+qvar(Δ2+qvΔ˙2))dt+qvar(Δf2+qvΔ˙f2)]. (12)

where the cost elements in Δ and Δ˙ penalize deviations from the reference trajectory (variance), and the control costs penalize effort. Weight factors α, qv and qvar can be chosen to adjust the optimal behavior of the system. Typically, optimal solutions will yield minimal net joint torque and impedance to remain close to the reference trajectory to some extent determined by the weight qvar (and qv).

In the present form, the dynamics of Θ and Δ are coupled by the gravitational term. To derive an interesting result, let us focus on horizontal movements for a moment. The system then simplifies to:

{IΘ¨=τ(t)-bΘ˙IΔ¨=-Ksκ(t)Δ-Kdκ(t)Δ˙-bΔ˙+Gη. (13)

This system is actually a controlled SDE with a 4-D state vector (Θ,Θ˙,Δ,Δ˙). Interestingly, the first two states are deterministic (in Θ) and noise only affects deviations from these reference states (in Δ, which we now rewrite Δt to stress that it is a random variable). Remarkably, the original SOOC problem is completely decoupled in this case. On the one hand, we have a deterministic sub-problem with dynamics in Θ and cost in τ2. On the other hand, we have a stochastic sub-problem with dynamics in Δt and cost in κ2 and Δt,Δ˙t. This SOOC problem only involving the mean and covariance of the state (Δt,Δ˙t) can be solved by deriving an equivalent DOC problem as described before. Since the mean of (Δt,Δ˙t) is zero given the initial conditions, propagation of the mean can be neglected for this part. Regarding Θ(t), since it is a deterministic variable, the propagation of its covariance can be neglected. In summary, the components Θ(t),Θ˙(t) correspond to the mean of the actual state θt,θ˙t, and the covariance of Δt,Δ˙t corresponds to the covariance of the actual state θt,θ˙t. Therefore, the net torque τ(t) controls the mean of the stochastic process in θt,θ˙t whereas κ(t) controls its covariance independently. As the process in Eq 13 is Gaussian, it is thus fully characterized and controlled.

These derivations are useful to understand the functioning of the model but, in more realistic scenarios, the above decoupling will not hold anymore. This is in no way an issue or limitation because more general cases can be easily handled within the SOOC framework. For instance, if gravity is not neglected, Eq 11 can be used together with Eq 12 to formulate a nonlinear SOOC problem that can be resolved with the methods described in the previous subsections.

Finally, these considerations may be reminiscent of the equilibrium point theory [60]. However, the approach differs from equilibrium point theory in the sense that a feedforward torque controller τ(t) is assumed (hence our approach requires “internal models”). It nevertheless fits with some aspects of the equilibrium point theory in the sense that a “virtual” reference trajectory Θ is planned together with time-varying impedance parameters (which may be tuned in practice via co-contraction of opposing muscles). As such, τ(t) and κ(t) might resemble the c- and r-commands described in [14] even though movement is not generated only by the viscoelastic properties of the musculoskeletal systems and shifts in equilibrium points/trajectories in our case (see also Discussion).

Muscle level modeling: Implicit description of force and impedance planning

Here we use more advanced models of the musculoskeletal system. In this work, we used the muscle model proposed by Katayama and Kawato [53] and assume that a feedforward motor command can be sent to each muscle individually.

Single-joint arm

For a single-joint system like the forearm moving in the horizontal plane, Katayama and Kawato’s model writes as follows:

Iθ¨=τ1+τ2+Gη (14)

where τ1 and τ2 are the muscle torques that are respectively functions of muscle activations u1 and u2, defined as follows:

τi=-aiTi,i=1or2Ti=(k0+kiui)(riui+l0-lmi+aiθ)+(b0+biui)aiθ˙,i=1or2 (15)

In this case, the system state is x=(θ,θ˙). Muscle parameters were taken from [53]. Here we have I = 0.0588 kg.m2, ki = 1621.6 N/m, k0 = 810.8 N/m, bi = 108.1 N.s/m, b0 = 54.1 N.s/m, ai = 2.5 cm, ri = (−1)i × 2.182 cm for i = 1‥2, lm1-l0=5.67 cm and lm2-l0=0.436 cm. We also set m = 1.44 kg, lc = 0.21 m, and the forearm length was 0.35 m.

In this model, the muscle torque thus depends on the muscle activation as well as on the angular position and velocity. The muscle activations can therefore modulate both the net joint torque and the joint viscoelasticity, in particular via muscle co-contraction.

Two-joint arm

A two degrees-of-freedom (dof) version of the arm with 6 muscles was also considered to simulate planar arm reaching movements, corresponding to the full model of [53]. The state of the arm is then x=(q,q˙)R4 where q = (θ1, θ2) denotes the joint angle vector (1st component for shoulder and 2nd component for elbow) and q˙=(θ˙1,θ˙2) denote the corresponding joint velocity vector. The dynamics of the arm follows a rigid body equation of the form:

q¨=M-1(q)(τ(q,q˙,u)-C(q,q˙)q˙) (16)

where M(q) is the inertia matrix, Cq˙ is the Coriolis/centripetal term, τ is the net joint torque vector produced by muscles and uR6 is the muscle activation vector (restricted to be open-loop/deterministic in this work).

The net joint torque vector is a function τ(q,q˙,u) depending on the moment arms (assumed constant) and on the muscle lengths/velocities expressed as affine functions of the joint positions and velocities as in Eq 15. All the parameters of the complete model with 6 muscles can be found in the Tables 1, 2, and 3 in [53]. Finally, by introducing noise (ωt is a 2-dimensional standard Brownian motion), we obtain the following SDE modeling the noisy musculoskeletal dynamics of a multijoint arm:

dxt=f(xt,u(t))dt+Gdωt (17)

with

f(xt,u(t))=(q˙tM-1(qt)(τ(qt,q˙t,u(t))-C(qt,q˙t)q˙t)) (18)

and G is a matrix allowing to define the coupling of the system to the noise.

Results

In this section, we consider simulations accounting for results of several experimental findings about the planning of force and impedance as well as on the role of muscle co-contraction in posture and movement control. Different models, types of noise, and cost functions will be used to illustrate the flexibility of the framework in making consistent predictions about the role of co-contraction and impedance for the open-loop control of stochastic systems.

Co-contraction planning in 1-dof motor tasks

Unstable postural task with the forearm

In Hogan’s study described above [48], the maintenance of a human forearm in an upright posture was considered. Hogan described a system controlled by a pair of agonist/antagonist muscles and showed that co-contraction was a strategy used by participants to maintain such an unstable posture. Here, we reconsider this task to test our framework with this simple starting example. As already mentioned, Hogan modeled the forearm as an inverted pendulum in the gravity field driven by opposing muscles having the essential force-length dependence of real muscles (see Eq 1). Here we considered two scenarios tested in Hogan’s experiment depending on whether the forearm was loaded (mload = 2.268 kg attached at the hand) or unloaded. To efficiently resolve the problem, we showed in [47] that via statistical linearization techniques we can get a bilinear system as in Eq 2 (which amounts to linearize the gravitational torque, i.e. sin θθ with a small angle hypothesis), with matrices defined as follows:

A=(01mglcI-bI),B=(00TI-TI),andNi=(00-KI0),i=1or2. (19)

where the parameters are defined by Eq 1.

In our simulations, noise was taken of constant variance and acting at acceleration level, G = (0, .1). We set T = K = 1 and considered a fixed damping parameter b = 1 Nms/rad. The cost was defined as R = diag(1, 1), Q = Qf = diag(104, 103) (see Eq 3). These weights were chosen such as to make effort and variance costs of comparable magnitude (note that angles are in radians). We assumed that the system was at state x0 = 0 at initial time t = 0 with zero covariance. The goal of the task was to maintain the inverted pendulum around this position without on-line sensory feedback for tf = 5 s while minimizing a compromise between variance (cost depending on Q, Qf) and effort (cost depending on R). In the equivalent DOC problem, the final covariance Pf was left free whereas the final mean state was set to zero (mf = 0). Results of simulations are reported in Fig 1. In these graphs, one can see that an optimal level of stiffness is achieved to stabilize the forearm in this unstable posture. We checked that, if there was no co-contraction, the forearm would fall in about one second due to the combined effects of noise and gravity (remind that feedback control is prevented). Therefore, co-contraction creates a resultant stiffness that is just enough to compensate the task instability. Note that if noise magnitude is larger, a larger co-contraction becomes optimal (dashed-dotted line in Fig 1B). This change in co-contraction with noise magnitude agrees with a study of Hasson [61]. In the loaded case, the destabilizing gravitational torque increases and the optimal co-contraction level becomes larger to counteract it. Remarkably, it can be observed that, like in Hogan’s original work [48], the activation of the flexor muscle in the loaded case is larger than the corresponding activation of that muscle that would be necessary to maintain the forearm horizontal in the unloaded case (dotted line in the bottom-right panel). Note that the exact shape of the optimal solutions depends on the terminal state constraints (on mean and covariance), the weights in the cost function (i.e. the effort/variance trade-off) and the level of noise. The lack of steady-state behavior near the end of the simulation is in particular due to the finite time-horizon used in these simulations and the associated terminal cost/constraints.

Fig 1. Co-contraction during maintenance of an upright forearm posture.

Fig 1

A. Position (in degrees) and velocity (in degrees per second) for the unloaded case. Thick black lines depict the means and shaded areas depict standard deviations (from 500 samples). Three single trajectories are displayed to illustrate their stochastic nature. B. Corresponding optimal joint stiffness (solid black line) and flexor/extensor muscle torques (solid and dashed respectively) for the unloaded case. In dotted gray line, the “divergent” stiffness level that co-contraction must overcome for stability is depicted (i.e. mglc in our case). In dashed-dotted black line, we also report the optimal stiffness when noise magnitude is increased by a factor 5, i.e. G = (0, .5). C-D. Same information for the loaded case where I was increased by mload l2 and mglc was increased by mload gl. A significant increase of stiffness, originating from a larger co-contraction of flexor and extensor muscles, can be noticed. Parameters: m = 1.44 kg, mload = 2.268 kg, l = 0.35 m, lc = 0.21 m, I = 0.0588 m.kg2 and g = 9.81 m/s2.

Reaching task with the forearm

Joint-level modeling

Here we consider single-joint pointing movements performed with the forearm. We first use the joint-level description of force and impedance derived in Eqs 913. For these simulations, we focus more specifically on the controlled system described in Eq 13 and on the experimental data reported in [57, 62]. Bennett’s main findings were that the elbow joint stiffness varies within point-to-point movements (either cyclical or discrete). The forearm was found to be overall quite compliant (measured stiffness ranging between 2 Nm/rad and 20 Nm/rad). Yet, stiffness significantly increased when the hand approached the target and stiffness had minimal values near peak velocity. Additionally, mean joint stiffness was found to increase with peak velocity and to increase almost linearly with the magnitude of net joint torque. In Fig 2 we replicated these main observations within our framework. Fig 2A and 2B shows the optimal behavior for movements executed in presence of signal-dependent noise (proportional to net torque τ). Because the task in [62] involved cyclical forearm movements, we imposed equal initial and final covariances. We also chose b = 0 since damping was already modulated together with stiffness in this model such as to get a constant damping ratio as suggested in [57] (damping however seemed harder to identify in general but it tended to fluctuate around 0.5 Nms/rad in experimental data). We penalized the integral of the square of the controls (effort) plus a covariance term involving position and velocity in the cost function (see Eq 12). We considered multiplicative noise acting in torque space in these simulations (20% of net torques, i.e. G=G(τ)=1I(0,0,0,0.2τ(t))).

Fig 2. Simulations of a pointing movement with the forearm (elbow extension of 1 rad).

Fig 2

A. Optimal trajectories. Angular displacement and velocity profiles (mean in thick lines and standard deviation in shaded areas) and net torque responsible of the joint motion (third column). B. Optimal impedance. Inertia (constant for this single-joint system), stiffness and damping are depicted. Time-varying joint stiffness and damping, part of the optimal open-loop motor plan in our model, are responsible of robustness of motor behavior around the mean behavior (without needing on-line feedback for that purpose). Note an increase of stiffness at the end of the motion, to improve accuracy on target, in agreement with experimental data. Values can be quantitatively compared to [62]. Time-average values are represented by horizontal lines. C. Relationship between time-average net torques and time-average stiffnesses for movements of different speeds. Peak velocity is indicated in rad/s. An approximately linear relationship is observed as in [57]. D. Corresponding time-varying stiffness profiles for the different movement speeds. In panels A and B, parameters were as follows: qvar = 104, α = 1, qv = 0.01, tf = 0.75, and I = 0.072. In panels C and D, same parameters but with tf ranging between 0.25 s and 1.0 s to generate movements of different speeds.

In Fig 2C and 2D, we varied movement duration to test the effect of speed on joint stiffness. Results can be compared to [57, 62, 63]. We found that, indeed, stiffness tends to increase almost linearly with net torque (which also increases with speed). These values have been compared to phasic and tonic EMG data in experimental works. However, this joint-level description of force and impedance planning does not allow to see the origin of stiffness in terms of muscle co-contraction. Therefore, we next consider muscle models to further investigate the co-contraction phenomenon in reaching arm movements.

Muscle-level modeling

Here we consider the musculoskeletal arm model proposed in [53] (Eq 14) to simulate elbow flexion movements in the horizontal plane (hence with g = 0). This is a muscle-level description of force and impedance planning. We focus on the experimental results of [64] which showed that subjects can reach faster while preserving accuracy if asked to co-contract opposing muscles. A signal-dependent noise model was defined here as in [29] in order to model that co-contraction does not lead to increased variability [16] as it would be the case for a standard signal-dependent noise model. The noise model was as follows:

G(u(t))=(0d(|u1(t)-u2(t)|1.5+0.01|u1(t)+u2(t)|1.5)) (20)

where d is a factor to set the overall magnitude of this signal-dependent noise (here we fixed d = 4 in simulations because it yielded good quantitative predictions of the empirical variability found in such fast reaching movements).

The cost function of the associated deterministic problem (Eq 5) was as follows:

C=0tfu(t)u(t)dt+trace(QfPf) (21)

with Qf = qvardiag(1, 0.1). The term Qf simply penalizes the terminal state covariance and qvar is a free parameter. We further set m0 = (25°,0)T and mtarg = (65°,0)T as the initial and final mean positions of the reaching task as in [64]. The initial covariance P0 was zero and the final covariance Pf was left free but minimized because the goal of the task was to reach a target of width 5° (given that the amplitude of the movement was 40°, the index of difficulty was 4 bits for this task). As such, this cost function implements an effort-variance compromise as suggested in [64, 65].

Fig 3 shows the results of simulated pointing movements. In Fig 3A, we set tf = 475 ms as in Missenard’s experiment and qvar = 50. With these settings we reproduced quite well the spontaneous behavior of subjects in this experiment, which is representative of what occurs in a standard Fitts’ like paradigm. For instance, peak velocities (PV) were about 140 deg/s and the index of co-contraction (IC) was about 20% in experimental data (index of co-contraction was defined as in [59, 64]). In our simulations we obtained PV of 139.75 deg/s and IC of 22.35% (see Table 1). Yet, when asked to co-contract to an IC of ~80%, subjects in [64] were able in practice to perform the task with greater speeds without apparent loss of accuracy. We replicated this condition in Fig 3B by setting tf = 400 ms as in Missenard’s experiment and qvar = 500. Because of signal-dependent noise, going faster increases noise magnitude and endpoint variance unless co-contraction is used. With these parameters, we obtained a PV of 178.87 deg/s (compared to about 180 deg/s in [64]) and an IC of 80.15%. Even though movements were faster, the positional standard deviation of the endpoint (EPstd) was 1.89 deg –hence smaller than half the target’s width such that the task could be achieved successfully on successive trials–. Therefore, there should be no more overshoots or undershoots in this condition, as was observed in [64]. Table 1 shows that this improvement of speed at comparable accuracy is highly costly due to a clear co-contraction of agonist/antagonist muscles (see effort column), especially at the end of the movement (but co-contraction also appears at its beginning). For comparison, for strategies without co-contraction (e.g. obtained by setting a small weight, e.g. qvar = 1), positional standard deviations of the end-point would be respectively 2.82 deg and 3.35 deg for movements times of 475 ms and 400 ms (see Table 1). This justifies why a minimal level of co-contraction is indeed required to perform the task accurately enough (again, with our open-loop control assumption). The fact that a significant co-contraction appears at the beginning and at the end of the movement agrees well with the literature [17, 62]. This example confirms that a trade-off between effort, speed, and accuracy may be prevalent in Fitts-like tasks, i.e. when subjects are instructed to perform the task as fast and as accurately as possible.

Fig 3. Simulations of pointing movements following the experiment of [64].

Fig 3

A. Simulation with tf = 475 ms and qvar = 50. The mean position and velocity (thick black traces) are reported. Shaded areas correspond to standard deviations. Some individual trajectories are also depicted with thin gray traces. The corresponding optimal open-loop muscle activations are reported (black for flexor and gray for extensor). Co-contraction is necessary to achieve the requested accuracy as actually observed in Fitts-like reaching experiments. B. Simulation with tf = 400 ms and qvar = 500. Same information than in panel A. It is seen that with even more co-contraction, an acceptable accuracy can also be achieved but for faster movements. Note the asymmetry of velocity profiles, with a longer deceleration than acceleration, which is also typical of Fitts’ instructions [66].

Table 1. Parameters corresponding to the simulated movements of Fig 3.

In all cases, initial and final positions were 25° and 65° (amplitude of 40°). The target width was 5° such that the index of difficulty (ID) was 4 bits. The reference end-point standard deviation can thus be 2.5°. The effort is measured as the quadratic cost in u according to Eq 21.

tf (ms) EPstd (deg) PV (deg/s) qvar Effort (×10−2) IC (%)
475 2.51 139.75 50 16.99 22.35
400 2.41 161.56 50 27.36 27.36
400 1.89 178.87 500 94.16 80.15
475 2.82 134.79 1 4.03 <0.1
400 3.35 146.47 1 5.60 <0.1

Co-contraction planning in 2-dof motor tasks

Here we consider 2-dof arm reaching tasks and the musculoskeletal model described in [53] and Eqs 1618. This model contains 4 single-joint muscles acting around the shoulder and elbow joints and 2 double-joint muscles. It has been shown to capture the basic stiffness properties of the human arm and has been investigated to evaluate the equilibrium point hypothesis originally. Here we used this model to test our SOOC framework with a quite advanced musculoskeletal model and to see whether co-contraction may be an optimal strategy to regulate the limb’s mechanical impedance in open-loop in certain tasks.

Two-link arm reaching task in a divergent force field

Burdet and colleagues found that subjects succeeded in performing accurate pointing movements in an unstable environment by selective muscle co-contraction [16, 18]. In their experiment, participants had to point to a target placed in front of them with a force field applying perturbations proportional to their lateral deviation from a straight line. Because the hand would start with random lateral deviation due to motor noise, it was not possible for the subjects to predict whether the arm would be pushed to the left or to the right during movement execution. The strength of the perturbation force (proportional to the extent of lateral deviation) and delays in sensorimotor loops would prevent participants from using an optimal closed-loop control strategy that requires accurate estimation of the system state to function (e.g. [67]). Instead, experimental data clearly showed that the solution of the participants was to stiffen their limb, in particular via co-contraction mechanisms and in a feedforward manner (e.g. participants kept co-contracting when the divergent force field was unexpectedly removed) [18, 35, 68].

Here we used the Eqs 17 and 18 to model the arm dynamics but we added the external perturbation force applied to the endpoint and had to consider an appropriate cost function to model the task. More precisely, a term

q¨ext=M-1(q)J(q)(Fext0) (22)

was added to the right-hand side of Eq 16, with J being the Jacobian matrix of the arm and Fext = βx being the external force (x is the Cartesian position of the hand along the horizontal axis and β = 40 Nm−1 in our simulations).

The cost function was designed as

C(u)=E[0tfL(m,u)dt+qvarϕ(xf)] (23)

where L(m,u)=uu+12(x¨2+y¨2), and ÿ being the mean Cartesian accelerations of the endpoint along the x and y axes respectively (i.e. functions of m and u, which can be easily computed from the forward kinematic function), and ϕ(xf) is a function penalizing the covariance of the final state xf (qvar is a weighting factor). Hence this cost is a trade-off between minimum effort/variance and maximum smoothness in Cartesian space (e.g. see [5, 7, 69]). In these simulations, the smoothness term was needed because the minimum effort solution for this musculoskeletal model led to hand trajectories that were too curved compared to normal human behavior in the task (even without force field). For the variance term, we penalized the final positional variance in task space by defining

φ(xf)=[J(mq,f)(qf-mq,f)][J(mq,f)(qf-mq,f)] (24)

where mq,f is the mean of the final position of the process and qf is a 2-D random vector composed of final joint positions extracted from xf.

The expectation of ϕ(xf) can be rewritten as a function of the final mean mf and covariance Pf as Φ(mf, Pf) = J(mq,f)Pq,f J(mq,f) where Pq,f denotes the 2x2 covariance matrix of joint positions. Finally, the expected cost function can be rewritten as follows:

C(u)=0tfL(m,u)dt+qvarΦ(mf,Pf). (25)

The latter cost was used in the deterministic optimal control problem that approximates the solution to the original SOOC problem.

A simple noise model was considered in these simulations:

G(xt)=(diag(0,0)M-1(qt)diag(σ1,σ2)), (26)

where the parameters σ1 and σ2 were used to set the magnitude of constant additive noise (which was assumed to act in torque space, hence the inverse of the inertia matrix in the expression of G).

Results of simulations are reported in Fig 4 and Table 2. In these simulations, we set tf = 0.75 s according to the data of [16]. The initial arm’s position was located at (0,0.30) in Cartesian coordinates and the target was at (0,0.55). Noise magnitude was set to σ1 = σ2 = 0.025 in Eq 26.

Fig 4. Two-link arm reaching experiment in a divergent force field.

Fig 4

A. Endpoint trajectories (paths and velocity) and stiffness (Cartesian stiffness components Sxx, Syy and Sxy of the matrix S = J−⊤(q)KJ−1(q) where K=τq is the joint stiffness) when no penalization of the endpoint variance is considered in the cost (qvar = 0). Twenty sampled paths of the endpoint are depicted (light gray traces). Red traces represent the theoretical mean trajectory from the associated DOC problem and thick black traces represent the mean trajectory over 1000 samples. Vertical dotted lines are displayed to visualize deviations from the straight path (±3cm). The blue circle represents the target (radius of 2.5cm). The green ellipse represents the endpoint positional covariance. The temporal evolution of the mean endpoint stiffness is also depicted (components Sxx, Syy and Sxy). The six muscle activations (open-loop control variables) and the muscle tensions are also reported. Muscles with opposed biomechanical functions were paired (emphasized with the same color, solid lines for flexors and dashed lines for extensors). B. Same information with a weight on the variance cost equal to qvar = 104. A significant increase of co-contraction of agonist/antagonist muscles can be noticed and the improvement in final accuracy is also clear (green ellipse). Note that only an open-loop motor command was employed in these simulations (no on-line state feedback control to correct deviations due to noise and the divergent force field).

Table 2. Influence of model parameters on the simulated optimal movements.

The model parameters that were varied are σi, qvar and β. Effort is the integral cost in the control variable u(t). EP std is the final standard deviation of the endpoint along the x-axis, and S¯xx is the mean endpoint stiffness along the x-axis. Sensitivity of the results to increasing the magnitude of the force field, increasing noise or increasing qvar is tested. Note that the model predicts an increase of the lateral endpoint stiffness on average to perform the task accurately in open-loop (~2x factor in these simulations).

Noise σi qvar β (N/m) Effort (×10−2) EP std (cm) S¯xx (N/m)
0.025 0 0 3.73 0.79 28.02
0.025 0 40 3.98 4.46 28.29
0.025 1e4 0 18.18 0.52 32.20
0.025 1e4 40 120.77 0.89 50.69
0.025 1e5 40 306.28 0.42 60.66
0.05 1e4 40 213.13 1.13 57.03
0.025 1e4 80 356.54 1.21 81.69

Overall we found that it was possible to perform this unstable reaching task without on-line estimation of the actual system state by co-contracting pairs of opposing muscles (see Fig 4B). This finding agrees with [18, 68]. Muscle co-contraction increased when the endpoint variance was penalized more strongly (but at the cost of a greater effort) and when the divergent force field had a greater magnitude (see mean endpoint stiffness along the x-axis in Table 2). Co-contraction also increased when noise magnitude was increased, all other parameters being equal (Table 2). As a rule of thumb, endpoint stiffness was found to increase with (1) the magnitude of the divergent force field, (2) the weight of the variance cost and (3) the magnitude of noise, in agreement with experimental observations [17, 61, 70].

However, while endpoint stiffness increased along the direction of instability, we also found that it increased in the direction of the movement. As such, the stiffness geometry was not really shaped according to the direction of the destabilizing force (see Fig 5A). We found this is actually a limitation of the underlying musculoskeletal model used in these simulations, which does not allow arbitrary geometries of the endpoint stiffness for the current arm posture –the orientation of the depicted ellipses was actually the most horizontal that one can get from this model given the Jacobian matrix at the midpoint of the trajectory–. This was tested by considering all possible muscle activation vectors uR+6 and checking that the resultant stiffness ellipse was never oriented horizontally. To increase the stiffness along the x-axis, the algorithm thus had to increase the endpoint stiffness as a whole and not as selectively as in the data of [16, 23] (but see [71] in static tasks). Note that this observation does not preclude alternative musculoskeletal models from yielding stiffness ellipses elongated in the direction of instability (e.g. see [72, 73]). Instead of considering alternative muscle models, we investigated whether the selective tuning of stiffness geometry could be predicted by the SOOC framework by considering a simpler Cartesian model of the task (following the derivations of Eqs 913, but for a planar mass-point system). In this Cartesian model, the cost was only composed of effort and endpoint variance terms (no smoothness term was needed because optimal paths were straight when minimizing effort). Using this Cartesian model, it becomes clear that the optimal endpoint stiffness predicted by the SOOC framework is shaped according to the direction of the destabilizing force (see Fig 5B). The interest of this Cartesian model is to show that the change of stiffness geometry in the divergent force field can be explained within the SOOC framework. The interest of the muscle model was to show that muscle co-contraction may indeed underlie the increase of endpoint stiffness.

Fig 5. Endpoint stiffness geometry at the midpoint of movement path.

Fig 5

A. Case of the 6-muscles model with detailed results reported in Fig 4. B. Case of a 2-D Cartesian mass-point model. In green, the geometry of the optimal endpoint stiffness without divergent force field (NF) is represented. In red, the same data is reported when the divergent force field is on (DF). Solid lines correspond to β = 40 and dotted lines to β = 80 (note that qvar and σi were fixed in these simulations).

A no intervention principle

Finally, we revisit the minimum intervention principle [9] as illustrated in a pointing-to-a-line task in [9, 74, 75]. For this kind of tasks, DOC models will fail to explain the empirical structure of endpoint variability [75]. In contrast, SOC models will capture endpoint variability very well through the minimal intervention principle which states that deviations from the planned trajectory are corrected only when they interfere with the goal of the task [9]. Alternatively a terminal optimal feedback controller can also reproduce this variability but it requires on-line state estimation processes as well [74] (the model re-plans open-loop trajectories from each estimated initial state and is not stochastically optimal in the sense that it does not consider variability across repeated trajectories to determine the control action). Here we show that on-line state estimation through sensory feedback is even not necessary at all to reproduce that variability in task-irrelevant dimensions is larger than variability in task-relevant ones (e.g. uncontrolled manifold, [76]) as long as mechanical impedance is appropriately regulated via feedforward processes like co-contraction (Fig 6).

Fig 6. Top view of arm trajectories for a pointing-to-a-line experiment.

Fig 6

The targets are indicated by the solid lines (blue and red). The green ellipse represents the 90% confidence ellipse of the endpoint distribution. Noise was additive (σi = 0.1) in these simulations and movement time was tf = 0.75s for the forward motion (blue) and tf = 0.55s for the leftward motion (red). The variance weight in the cost, qvar, was set to 104 and endpoint variance was penalized in a the direction orthogonal to the target line (via the function n J(qf)Pq,f J(qf) n where n is the normal vector, J is the Jacobian matrix and Pq,f is the joint-space positional covariance). Note that hard terminal constraints were imposed on the mean state (mean endpoint position on the target line and zero final mean velocity). The main orientation of the endpoint confidence ellipses is compatible with experimental observations and shows that co-contraction may be used to increase accuracy in the task-relevant dimension.

We considered the same 6-muscle model than in the previous simulation. The simulations show that the endpoint variance is elongated along the target line (i.e. task-irrelevant dimension), showing that impedance regulation can lead to a phenomenon similar to a minimal intervention principle (except that here it should rather be called a no intervention principle as there is no state feedback at all during movement execution). Conceivably, a testable hypothesis to determine whether this consideration makes sense would be to check the presence of such task-dependent endpoint distributions in deafferented patients with no vision (but with initial vision of the arm prior to the movement as in [43] and of the redundant target). It has already been shown that, in healthy subjects, this principle still applies when on-line vision is removed (see [75]).

Discussion

In this study, we have presented a novel optimal control framework to account for the planning of force and impedance via muscle co-contraction. This framework models motor planning as the design of optimal open-loop controls for stochastic dynamics. One main implication is that such open-loop controls will seek to optimally exploit the limb’s impedance characteristics to perform the task accurately enough, taking into account the presence of uncertainty induced by sensorimotor noise. Optimality is considered with respect to a trade-off between effort and variance costs but other terms may be represented as well (e.g. smoothness) in agreement with the literature. Using several simulations, we have illustrated the relevance and versatility of the framework to explain well-known experimental observations involving co-contraction and impedance measurements. Below, we discuss the significance and implication of this framework with respect to existing motor control theories.

Planning of force and impedance via muscle co-contraction

At a computational level, the SOOC framework lies in-between deterministic optimal control and stochastic optimal control theories (see [1, 4, 11] for reviews). These previous frameworks have been useful to predict many aspects of sensorimotor control and learning. However, they usually do not account for the phenomenon of muscle co-contraction in a principled manner. Yet, co-contraction has been found in many motor tasks and is a general feature of motor control (e.g. [17, 18, 48, 77]). In SOOC, the crucial ingredients to obtain co-contraction in musculoskeletal systems are threefold: (1) the consideration of open-loop control, (2) the presence of noise in the dynamics and (3) a cost function including at least effort and variance terms. Each ingredient has found experimental support in the literature. The feedforward aspect of control for learned movements has been emphasized in [18, 38], the effects of sensorimotor noise have been described in [12, 13], and the relevance of effort and variance costs in motor control has been stressed in several studies [65, 78]. The class of models considered in this study is particularly in the spirit of the minimum variance model [49] but with a couple of notable differences. In our framework, effort and variance are separated cost elements such that an optimal motor strategy may involve a large effort without implying a large variance (i.e. co-contraction). In the classical minimum variance model, variance is indeed the same as effort because a signal-dependent noise is assumed to affect a linear system. In our approach, relevant predictions regarding co-contraction and impedance planning can be made only for nonlinear systems (e.g. bilinear systems) and irrespective of the precise type of noise that is modeled (signal-dependent, constant noise etc.). Concretely, the controller can reach different levels of endpoint variance by setting different levels of co-contraction whereas the standard minimum variance model would only yield one (optimal) level of variance (at fixed movement time). Besides variance, effort and energy-like criteria are often minimized in optimal control models which tend to prevent the relevance of co-contraction. In [79], it was demonstrated mathematically that co-contraction of opposing muscles is non-optimal with respect to the minimization of the absolute work of muscle torques in a deterministic model. In other optimal control models with muscle modeling, co-contraction does not occur neither in deterministic settings (Fig. 9 in [26]) nor in stochastic settings (Fig. 2 in [28] or Fig. 3a in [29]). Researchers have nevertheless attempted to explain co-contraction or its contribution to impedance in existing DOC or SOC frameworks, but this was often an ad-hoc modeling [80, 81]. One difficulty is that empirical works stressed a relatively complex task-dependency of muscle co-contraction –as assessed by EMG co-activation– [17, 48, 63, 64, 82]. For instance, co-contraction seems to depend on noise magnitude [25, 61] and to tune impedance according to the degree of instability of the task [16, 18, 23, 70, 71]. Finding general principles to automatically predict the adequate co-contraction or impedance required for the task at hand thus appears necessary. In [72, 73, 83], an algorithm focusing on the trial-by-trial learning of force and impedance was developed to acquire stability without superfluous effort. In contrast to this approach, our framework models learned behaviors with known dynamics (up to some uncertainty modeled by noise). Furthermore, while the previous learning algorithm requires a reference trajectory to be defined a priori to apply, the SOOC framework allows predicting the “reference” trajectory as the outcome of optimality. This aspect may be particularly important given that inertia is also a key component of mechanical impedance for a multi-joint system. Hence, besides viscoelasticity, the SOOC framework should be able to exploit kinematic redundancy to plan stable behaviors with low effort. Using the SOC framework, other authors have also attempted to predict a limb’s mechanical impedance via muscle co-contraction. In [29], a model based on an extended signal-dependent noise model (see Eq 20), which explicitly favors co-contraction by reducing the variance of noise during co-contraction, was proposed. One issue is that for simpler noise models (e.g. simple constant noise), this model would not command muscle co-contraction. In SOOC, co-contraction was planned for a variety of noise models. Co-contraction and stiffness regulation was also considered in another SOC model [84], but the simulated limb’s stiffness was mostly due to the intrinsic stiffness of the muscles in the model (that of [53]) without clear task dependency (signal-dependent noise was also a critical constraint). In these SOC models, the role of state feedback could actually duplicate the role of co-contraction –high-level feedback control accounts for a form of impedance but differently from a feedforward co-contraction, [47, 67]–. In [85], an optimal control model was introduced to account for the planning of both trajectory and stiffness. However, this model as well as others not related to optimal control (e.g. [59, 86, 87]), were derived along the lines of the equilibrium point theory. While co-contraction is often discussed within equilibrium point theory [14], it is worth stressing that our approach rather accounts for co-contraction of a group of muscles within optimal control theory. As such, an important point of departure from equilibrium point theory is the need for internal models of the arm dynamics (see [31]) in order to set net joint torques and derive an efficient feedforward strategy for the control force and impedance in SOOC. The present theory indeed proposes that force and impedance may be planned simultaneously within descending motor commands. This idea seems coherent with the several studies that emphasized that two separate control mechanisms may exist for the control of force and impedance, the latter being at least partly governed by muscle co-contraction [18, 19, 58, 59]. However, impedance can also be regulated via feedback gains in SOC as mentioned earlier and, therefore, the conceptual differences between SOOC and SOC should be discussed further.

Implications as a motor control theory

Our framework partly formulates motor planning as a stochastic optimal open-loop control problem. One primary outcome of the planning process would then be a feedforward motor command that optimally predetermines both the mean behaviour and the variability of the system via force and impedance control. The term “open-loop” may raise questions about the role of sensory feedback in this framework. Computationally, sensory feedback is only required to estimate the initial state of the motor apparatus during movement preparation in SOOC. This contrasts with optimal closed-loop control that critically requires an estimate the system state to create the motor command throughout movement execution [9, 11]. Indeed, if an optimal feedback gain can be elaborated at the motor planning stage in SOC, the actual motor command is only determined once the current state of the motor apparatus has been properly estimated at the execution stage (e.g. hand or joint position/velocities…). An optimal closed-loop control scheme is thought to involve the primary motor cortex and, therefore, to require on-line transcortical feedback loops [3234]. These neural pathways imply relatively long latencies with muscle responses occurring ~50-100 ms after a mechanical perturbation is applied to a limb. Because these responses are quite sophisticated and task-dependent, relatively complex cortical processes combining sensory information with predictions from internal models of the limb’s dynamics and the environment are likely necessary for their formation. Besides these long-latency responses, short-latency responses are also observed <40 ms after a mechanical perturbation. This stretch reflex only involves the spinal circuitry and has been shown to be relatively fast, simple and stereotypical. Nevertheless, background muscle activity is also known to modify the gain of the stretch reflex likely due to the size-recruitment principle [88]. Co-contraction is therefore a means to increase the apparent mechanical impedance of a joint by increasing the gains of stretch reflexes in opposing muscles (and not only by increasing the intrinsic stiffness). Nonlinear effects occurring during co-contraction have been shown to amplify this increase of the reflex gains beyond what would have been expected by considering each muscle alone [22]. As we do not exclude the contribution of the stretch reflex in SOOC, ambiguity may arise here. Indeed, the stretch reflex relies on sensory information from muscle spindles: hence it does implement a (low-level) feedback control loop. However, we argue that the functioning of the neuromusculoskeletal system with intact reflex circuitry may be well accounted for within the SOOC framework and the “open-loop” control assumption. Indeed, being mainly under the influence of descending motor commands via alpha/gamma motoneurons activity, the short-latency reflex arc plays a crucial role in the apparent spring-like properties of a muscle –which we model– beyond its intrinsic short-range stiffness [21]. We thus consider that these low-level feedback loops are part of a neuromuscular actuator with variable impedance which is under a (feedforward) control from higher-level centers. As mentioned earlier, mechanical impedance can also be modified via transcortical reflex loops relying on pre-determined feedback gains. However, the associated mechanisms are of a different nature in that they require evolved state estimation processes. It is noteworthy that the same low-level/high-level dichotomy applies to robotics (as does the SOOC framework actually). Moreover, experimental estimations of a limb’s impedance during posture or movement are normally unable to rule out the impact of reflexes on measurements (which can be as short as ~20 ms for biceps brachii in humans, [89]). In summary, the distinction between short-latency/low-level spinal loops and long-latency/high-level transcortical loops parallels the distinction between optimal open-loop and feedback control frameworks in computational terms. The crucial difference between SOC and SOOC theories thus regards the involvement or not of high-level state estimation processes in the on-line control mechanisms. One implication of the SOOC theory is that such sophisticated high-level feedback processes occurring during movement execution may not necessarily be critical to ensure reliable motor performance in well-learned motor behaviors (but this achievement may require muscle co-contraction to some extent).

Conclusion

A new theoretical framework to model human movement planning has been presented. It provides a specific emphasis on the elaboration of optimal feedforward motor commands for the control of noisy neuromusculoskeletal systems. Interestingly, optimal open-loop strategies spontaneously exhibit co-contraction to generate robust motor behaviors without relying on sophisticated feedback mechanisms that requires state estimation processes during movement execution. In this framework, the magnitude of co-contraction or joint/endpoint stiffness is kept as small as possible because effort is penalized. Yet, depending on the task constraints (e.g. instability, accuracy demand) and uncertainty (e.g. internal and/or external noise magnitudes), a significant feedforward co-contraction or stiffening of the joints/hand may become the optimal strategy. This prediction was very consistent as we found it for both joint-level and muscle-level descriptions of the musculoskeletal dynamics as well as for various noise models. The SOOC framework may thus complement SOC in the following sense: once a motor plan is elaborated, locally optimal feedback control strategies may be designed after linearization around the optimal mean trajectory and open-loop control. One general advantage of planning force and impedance via co-contraction could be to provide a nominal robustness to the system, thereby resisting small perturbations without the need for a continuous multi-sensory integration (e.g. merging of visual and somatosensory information at cortical levels) to optimally estimate the state of the system during movement execution. Adequately tuning muscle co-contraction (even to small levels) might allow the system to be less sensitive to delays, noise and task uncertainty (and might improve the reliability of state estimation as well). This theoretical framework should be tested more extensively in the future to see whether it constitutes a viable theory for the neural control of movement but it already provides an interesting conceptual trade-off between purely deterministic approaches and purely stochastic approaches. As far as motor planning is of concern and the elaboration of feedforward motor commands is thought to be a significant component of the neural control of movement (see also [90]), the SOOC theory may constitute a relevant theoretical approach. Finally, we note that the very same framework could prove useful in human-inspired robotics especially for robots with variable impedance actuators [47, 91].

Acknowledgments

The authors would like to thank Dr. Francesco Nori for having initiated this work during a visiting professorship. The authors also wish to thank Prof. Etienne Burdet for his insightful comments on a previous version of the article.

Data Availability

There is no new experimental data in this study. However, the Matlab codes used to produce the simulations are available from http://hebergement.universite-paris-saclay.fr/berret/software/SOOC.zip. Note that the codes depend on a third-party software (GPOPS).

Funding Statement

This work was partly supported by a public grant overseen by the French National Research Agency as part of the Investissement d’Avenir program, through the iCODE Institute project funded by the IDEX Paris-Saclay, ANR-11- IDEX-0003-02. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Engelbrecht S. Minimum Principles in Motor Control. J Math Psychol. 2001;45(3):497–542. 10.1006/jmps.2000.1295 [DOI] [PubMed] [Google Scholar]
  • 2. Todorov E. Optimality principles in sensorimotor control. Nat Neurosci. 2004;7(9):907–915. 10.1038/nn1309 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Friston K. What is optimal about motor control? Neuron. 2011;72:488–498. 10.1016/j.neuron.2011.10.018 [DOI] [PubMed] [Google Scholar]
  • 4. Berret B, Delis I, Gaveau J, Jean F. In: Venture G, Laumond JP, Watier B, editors. Optimality and Modularity in Human Movement: From Optimal Control to Muscle Synergies. Cham: Springer International Publishing; 2019. p. 105–133. [Google Scholar]
  • 5. Flash T, Hogan N. The coordination of arm movements: an experimentally confirmed mathematical model. J Neurosci. 1985;5(7):1688–1703. 10.1523/JNEUROSCI.05-07-01688.1985 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Uno Y, Kawato M, Suzuki R. Formation and control of optimal trajectory in human multijoint arm movement. Minimum torque-change model. Biol Cybern. 1989;61(2):89–101. 10.1007/bf00204593 [DOI] [PubMed] [Google Scholar]
  • 7. Berret B, Chiovetto E, Nori F, Pozzo T. Evidence for composite cost functions in arm movement planning: an inverse optimal control approach. PLoS Comput Biol. 2011;7(10):e1002183 10.1371/journal.pcbi.1002183 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Berret B, Jean F. Why Don’t We Move Slower? The Value of Time in the Neural Control of Action. J Neurosci. 2016;36(4):1056–1070. 10.1523/JNEUROSCI.1921-15.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Todorov E, Jordan MI. Optimal feedback control as a theory of motor coordination. Nat Neurosci. 2002;5(11):1226–1235. 10.1038/nn963 [DOI] [PubMed] [Google Scholar]
  • 10. Todorov E. Stochastic optimal control and estimation methods adapted to the noise characteristics of the sensorimotor system. Neural Comput. 2005;17(5):1084–1108. 10.1162/0899766053491887 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Diedrichsen J, Shadmehr R, Ivry RB. The coordination of movement: optimal feedback control and beyond. Trends Cogn Sci. 2009. 10.1016/j.tics.2009.11.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. van Beers RJ, Haggard P, Wolpert DM. The role of execution noise in movement variability. J Neurophysiol. 2004;91(2):1050–1063. 10.1152/jn.00652.2003 [DOI] [PubMed] [Google Scholar]
  • 13. Faisal AA, Selen LPJ, Wolpert DM. Noise in the nervous system. Nat Rev Neurosci. 2008;9(4):292–303. 10.1038/nrn2258 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Latash ML. Muscle coactivation: definitions, mechanisms, and functions. J Neurophysiol. 2018;120:88–104. 10.1152/jn.00084.2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Demenÿ G. Du rôle mécanique des muscles antagonistes dans les actes de locomotion. Archives de Physiologie. 1890;5(2):747. [Google Scholar]
  • 16. Burdet E, Osu R, Franklin DW, Milner TE, Kawato M. The central nervous system stabilizes unstable dynamics by learning optimal impedance. Nature. 2001;414(6862):446–449. 10.1038/35106566 [DOI] [PubMed] [Google Scholar]
  • 17. Gribble PL, Mullin LI, Cothros N, Mattar A. Role of cocontraction in arm movement accuracy. J Neurophysiol. 2003;89(5):2396–2405. 10.1152/jn.01020.2002 [DOI] [PubMed] [Google Scholar]
  • 18. Franklin DW, Osu R, Burdet E, Kawato M, Milner TE. Adaptation to stable and unstable dynamics achieved by combined impedance control and inverse dynamics model. J Neurophysiol. 2003;90:3270–3282. 10.1152/jn.01112.2002 [DOI] [PubMed] [Google Scholar]
  • 19. Humphrey DR, Reed DJ. Separate cortical systems for control of joint movement and joint stiffness: reciprocal activation and coactivation of antagonist muscles. Adv Neurol. 1983;39:347–372. [PubMed] [Google Scholar]
  • 20. Joyce G, Rack P, Westbury D. The mechanical properties of cat soleus muscle during controlled lengthening and shortening movements. The Journal of physiology. 1969;204(2):461–474. 10.1113/jphysiol.1969.sp008924 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Nichols T, Houk J. Improvement in linearity and regulation of stiffness that results from actions of stretch reflex. J Neurophysiol. 1976;39(1):119–142. 10.1152/jn.1976.39.1.119 [DOI] [PubMed] [Google Scholar]
  • 22. Carter RR, Crago PE, Gorman PH. Nonlinear stretch reflex interaction during cocontraction. J Neurophysiol. 1993;69(3):943–952. 10.1152/jn.1993.69.3.943 [DOI] [PubMed] [Google Scholar]
  • 23. Franklin DW, Liaw G, Milner TE, Osu R, Burdet E, Kawato M. Endpoint stiffness of the arm is directionally tuned to instability in the environment. J Neurosci. 2007;27(29):7705–7716. 10.1523/JNEUROSCI.0968-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Lewis GN, MacKinnon CD, Trumbower R, Perreault EJ. Co-contraction modifies the stretch reflex elicited in muscles shortened by a joint perturbation. Exp Brain Res. 2010;207(1-2):39–48. 10.1007/s00221-010-2426-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Soechting JF, Dufresne JR, Lacquaniti F. Time-varying properties of myotatic response in man during some simple motor tasks. J Neurophysiol. 1981;46:1226–1243. 10.1152/jn.1981.46.6.1226 [DOI] [PubMed] [Google Scholar]
  • 26. Guigon E, Baraduc P, Desmurget M. Computational motor control: redundancy and invariance. J Neurophysiol. 2007;97(1):331–347. 10.1152/jn.00290.2006 [DOI] [PubMed] [Google Scholar]
  • 27. Berret B, Darlot C, Jean F, Pozzo T, Papaxanthis C, Gauthier JP. The inactivation principle: mathematical solutions minimizing the absolute work and biological implications for the planning of arm movements. PLoS Comput Biol. 2008;4(10):e1000194 10.1371/journal.pcbi.1000194 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Li W, Todorov E. Iterative linearization methods for approximately optimal control and estimation of non-linear stochastic system. Int J Control. 2007;80(9):1439–1453. 10.1080/00207170701364913 [DOI] [Google Scholar]
  • 29. Mitrovic D, Klanke S, Osu R, Kawato M, Vijayakumar S. A computational model of limb impedance control based on principles of internal model uncertainty. PLoS One. 2010;5:e13601 10.1371/journal.pone.0013601 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Ueyama Y, Miyashita E. Signal-dependent noise induces muscle co-contraction to achieve required movement accuracy: a simulation study with an optimal control. Current Bioinformatics. 2013;8(1):16–24. 10.2174/157489313804871632 [DOI] [Google Scholar]
  • 31. Wolpert DM, Ghahramani Z. Computational principles of movement neuroscience. Nat Neurosci. 2000;3 Suppl:1212–1217. 10.1038/81497 [DOI] [PubMed] [Google Scholar]
  • 32. Scott SH. Optimal feedback control and the neural basis of volitional motor control. Nat Rev Neurosci. 2004;5(7):532–546. 10.1038/nrn1427 [DOI] [PubMed] [Google Scholar]
  • 33. Scott SH. The computational and neural basis of voluntary motor control and planning. Trends in cognitive sciences. 2012;16:541–549. 10.1016/j.tics.2012.09.008 [DOI] [PubMed] [Google Scholar]
  • 34. Pruszynski JA, Scott SH. Optimal feedback control and the long-latency stretch response. Exp Brain Res. 2012;218(3):341–359. 10.1007/s00221-012-3041-8 [DOI] [PubMed] [Google Scholar]
  • 35. Franklin DW, Burdet E, Osu R, Kawato M, Milner TE. Functional significance of stiffness in adaptation of multijoint arm movements to stable and unstable dynamics. Experimental brain research. 2003;151(2):145–157. 10.1007/s00221-003-1443-3 [DOI] [PubMed] [Google Scholar]
  • 36. Osu R, Burdet E, Franklin DW, Milner TE, Kawato M. Different mechanisms involved in adaptation to stable and unstable dynamics. J Neurophysiol. 2003;90(5):3255–3269. 10.1152/jn.00073.2003 [DOI] [PubMed] [Google Scholar]
  • 37. Osu R, Morishige Ki, Miyamoto H, Kawato M. Feedforward impedance control efficiently reduce motor variability. Neurosci Res. 2009;65:6–10. 10.1016/j.neures.2009.05.012 [DOI] [PubMed] [Google Scholar]
  • 38. Franklin DW, Wolpert DM. Computational mechanisms of sensorimotor control. Neuron. 2011;72(3):425–442. 10.1016/j.neuron.2011.10.006 [DOI] [PubMed] [Google Scholar]
  • 39. Polit A, Bizzi E. Processes controlling arm movements in monkeys. Science (New York, NY). 1978;201:1235–1237. 10.1126/science.99813 [DOI] [PubMed] [Google Scholar]
  • 40. Polit A, Bizzi E. Characteristics of motor programs underlying arm movements in monkeys. J Neurophysiol. 1979;42:183–194. 10.1152/jn.1979.42.1.183 [DOI] [PubMed] [Google Scholar]
  • 41. Bizzi E, Accornero N, Chapple W, Hogan N. Posture control and trajectory formation during arm movement. The Journal of neuroscience: the official journal of the Society for Neuroscience. 1984;4:2738–2744. 10.1523/JNEUROSCI.04-11-02738.1984 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Hogan N. Planning and execution of multijoint movements. Can J Physiol Pharmacol. 1988;66:508–517. 10.1139/y88-080 [DOI] [PubMed] [Google Scholar]
  • 43. Ghez C, Gordon J, Ghilardi MF. Impairments of reaching movements in patients without proprioception. II. Effects of visual information on accuracy. J Neurophysiol. 1995;73:361–372. 10.1152/jn.1995.73.1.361 [DOI] [PubMed] [Google Scholar]
  • 44. Nielsen JB. Human Spinal Motor Control. Annu Rev Neurosci. 2016;39:81–101. 10.1146/annurev-neuro-070815-013913 [DOI] [PubMed] [Google Scholar]
  • 45. Nielsen J, Kagamihara Y. The regulation of disynaptic reciprocal Ia inhibition during co-contraction of antagonistic muscles in man. The Journal of physiology. 1992;456:373–391. 10.1113/jphysiol.1992.sp019341 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Crone C, Nielsen J. Central control of disynaptic reciprocal inhibition in humans. Acta Physiol Scand. 1994;152:351–363. 10.1111/j.1748-1716.1994.tb09817.x [DOI] [PubMed] [Google Scholar]
  • 47.Berret B, Jean F. Efficient computation of optimal open-loop controls for stochastic systems; 2019; hal-02158875. 10.1016/j.automatica.2020.108874 [DOI]
  • 48. Hogan N. Adaptive control of mechanical impedance by coactivation of antagonist muscles. IEEE Trans Autom Control. 1984;29(8):681–690. 10.1109/TAC.1984.1103644 [DOI] [Google Scholar]
  • 49. Harris CM, Wolpert DM. Signal-dependent noise determines motor planning. Nature. 1998;394(6695):780–784. 10.1038/29528 [DOI] [PubMed] [Google Scholar]
  • 50. Stengel RF. Optimal Control and Estimation Dover books on advanced mathematics. Dover Publications; 1986. [Google Scholar]
  • 51. Rao AV, Benson DA, Darby CL, Patterson MA, Francolin C, Sanders I, et al. Algorithm 902: GPOPS, A MATLAB software for solving multiple-phase optimal control problems using the gauss pseudospectral method. ACM Trans Math Software. 2010;37(2):1–39. 10.1145/1731022.1731032 [DOI] [Google Scholar]
  • 52. Maybeck PS. Stochastic models, estimation, and control. vol. 2 Academic press; 1982. [Google Scholar]
  • 53. Katayama M, Kawato M. Virtual trajectory and stiffness ellipse during multijoint arm movement predicted by neural inverse models. Biol Cybern. 1993;69:353–362. 10.1007/BF01185407 [DOI] [PubMed] [Google Scholar]
  • 54. Feldman AG. Once more on the equilibrium-point hypothesis (lambda model) for motor control. J Mot Behav. 1986;18(1):17–54. 10.1080/00222895.1986.10735369 [DOI] [PubMed] [Google Scholar]
  • 55. Milner TE, Cloutier C. Compensation for mechanically unstable loading in voluntary wrist movement. Experimental Brain Research. 1993;94(3):522–532. 10.1007/bf00230210 [DOI] [PubMed] [Google Scholar]
  • 56. Weiss P, Hunter I, Kearney R. Human ankle joint stiffness over the full range of muscle activation levels. J Biomech. 1988;21(7):539–544. 10.1016/0021-9290(88)90217-5 [DOI] [PubMed] [Google Scholar]
  • 57. Bennett DJ. Torques generated at the human elbow joint in response to constant position errors imposed during voluntary movements. Exp Brain Res. 1993;95:488–498. 10.1007/bf00227142 [DOI] [PubMed] [Google Scholar]
  • 58. Yamazaki Y, Ohkuwa T, Itoh H, Suzuki M. Reciprocal activation and coactivation in antagonistic muscles during rapid goal-directed movements. Brain Res Bull. 1994;34:587–593. 10.1016/0361-9230(94)90144-9 [DOI] [PubMed] [Google Scholar]
  • 59. Scheidt RA, Ghez C. Separate adaptive mechanisms for controlling trajectory and final position in reaching. J Neurophysiol. 2007;98:3600–3613. 10.1152/jn.00121.2007 [DOI] [PubMed] [Google Scholar]
  • 60. Feldman AG. Functional tuning of the nervous system with control of movement or maintenace of a steady posture, II: Controllable parameters of the muscles. Biophysics. 1966;11:565–578. [Google Scholar]
  • 61. Hasson CJ, Gelina O, Woo G. Neural Control Adaptation to Motor Noise Manipulation. Front Hum Neurosci. 2016;10:59 10.3389/fnhum.2016.00059 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62. Bennett DJ, Hollerbach JM, Xu Y, Hunter IW. Time-varying stiffness of human elbow joint during cyclic voluntary movement. Exp Brain Res. 1992;88:433–442. 10.1007/bf02259118 [DOI] [PubMed] [Google Scholar]
  • 63. Suzuki M, Shiller DM, Gribble PL, Ostry DJ. Relationship between cocontraction, movement kinematics and phasic muscle activity in single-joint arm movement. Exp Brain Res. 2001;140(2):171–181. 10.1007/s002210100797 [DOI] [PubMed] [Google Scholar]
  • 64. Missenard O, Fernandez L. Moving faster while preserving accuracy. Neuroscience. 2011;197:233–241. 10.1016/j.neuroscience.2011.09.020 [DOI] [PubMed] [Google Scholar]
  • 65. Wang C, Xiao Y, Burdet E, Gordon J, Schweighofer N. The duration of reaching movement is longer than predicted by minimum variance. J Neurophysiol. 2016;116(5):2342–2345. 10.1152/jn.00148.2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66. Jean F, Berret B. On the duration of human movement: from self-paced to slow/fast reaches up to Fitts’s law In: Geometric and Numerical Foundations of Movements. Springer; 2017. p. 43–65. [Google Scholar]
  • 67.Berret B, Ivaldi S, Nori F, Sandini G. Stochastic optimal control with variable impedance manipulators in presence of uncertainties and delayed feedback. In: Proc. IEEE/RSJ Int Intelligent Robots and Systems (IROS) Conf; 2011. p. 4354–4359.
  • 68. Franklin DW, So U, Burdet E, Kawato M. Visual feedback is not necessary for the learning of novel dynamics. PLoS One. 2007;2(12):e1336 10.1371/journal.pone.0001336 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69. Wolpert DM, Ghahramani Z, Jordan MI. Are arm trajectories planned in kinematic or dynamic coordinates? An adaptation study. Exp Brain Res. 1995;103(3):460–470. 10.1007/bf00241505 [DOI] [PubMed] [Google Scholar]
  • 70. Franklin DW, So U, Kawato M, Milner TE. Impedance control balances stability with metabolically costly muscle activation. J Neurophysiol. 2004;92(5):3097–3105. 10.1152/jn.00364.2004 [DOI] [PubMed] [Google Scholar]
  • 71. Selen LP, Franklin DW, Wolpert DM. Impedance control reduces instability that arises from motor noise. J Neurosci. 2009;29(40):12606–12616. 10.1523/JNEUROSCI.2826-09.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72. Franklin DW, Burdet E, Tee KP, Osu R, Chew CM, Milner TE, et al. CNS learns stable, accurate, and efficient movements using a simple algorithm. Journal of neuroscience. 2008;28(44):11165–11173. 10.1523/JNEUROSCI.3099-08.2008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73. Tee KP, Franklin DW, Kawato M, Milner TE, Burdet E. Concurrent adaptation of force and impedance in the redundant muscle system. Biological cybernetics. 2010;102(1):31–44. 10.1007/s00422-009-0348-z [DOI] [PubMed] [Google Scholar]
  • 74. Guigon E, Baraduc P, Desmurget M. Optimality, stochasticity, and variability in motor behavior. J Comput Neurosci. 2008;24(1):57–68. 10.1007/s10827-007-0041-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75. Berret B, Chiovetto E, Nori F, Pozzo T. Manifold reaching paradigm: how do we handle target redundancy? J Neurophysiol. 2011;106(4):2086–2102. 10.1152/jn.01063.2010 [DOI] [PubMed] [Google Scholar]
  • 76. Scholz JP, Schöner G. The uncontrolled manifold concept: identifying control variables for a functional task. Exp Brain Res. 1999;126(3):289–306. 10.1007/s002210050738 [DOI] [PubMed] [Google Scholar]
  • 77. Milner TE. Adaptation to destabilizing dynamics by means of muscle cocontraction. Exp Brain Res. 2002;143(4):406–416. 10.1007/s00221-002-1001-4 [DOI] [PubMed] [Google Scholar]
  • 78. Liu D, Todorov E. Evidence for the flexible sensorimotor strategies predicted by optimal feedback control. J Neurosci. 2007;27(35):9354–9368. 10.1523/JNEUROSCI.1110-06.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79. Gauthier JP, Berret B, Jean F. A Biomechanical Inactivation Principle. Proceedings of the Steklov Institute of Mathematics. 2010;268:93–116. 10.1134/S0081543810010098 [DOI] [Google Scholar]
  • 80. Tee KP, Burdet E, Chew CM, Milner TE. A model of force and impedance in human arm movements. Biol Cybern. 2004;90(5):368–375. 10.1007/s00422-004-0484-4 [DOI] [PubMed] [Google Scholar]
  • 81. Yadav V, Sainburg RL. Motor lateralization is characterized by a serial hybrid control scheme. Neuroscience. 2011;196:153–167. 10.1016/j.neuroscience.2011.08.039 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82. Missenard O, Mottet D, Perrey S. The role of cocontraction in the impairment of movement accuracy with fatigue. Exp Brain Res. 2008;185(1):151–156. 10.1007/s00221-007-1264-x [DOI] [PubMed] [Google Scholar]
  • 83. Yang C, Ganesh G, Haddadin S, Parusel S, Albu-Schaeffer A, Burdet E. Human-like adaptation of force and impedance in stable and unstable interactions. IEEE transactions on robotics. 2011;27(5):918–930. 10.1109/TRO.2011.2158251 [DOI] [Google Scholar]
  • 84. Ueyama Y, Miyashita E. Optimal feedback control for predicting dynamic stiffness during arm movement. IEEE Trans Ind Electron. 2013;61(2):1044–1052. 10.1109/TIE.2013.2273473 [DOI] [Google Scholar]
  • 85. Hasan Z. Optimized movement trajectories and joint stiffness in unperturbed, inertially loaded movements. Biol Cybern. 1986;53(6):373–382. 10.1007/bf00318203 [DOI] [PubMed] [Google Scholar]
  • 86. Flash T. The control of hand equilibrium trajectories in multi-joint arm movements. Biol Cybern. 1987;57(4-5):257–274. 10.1007/bf00338819 [DOI] [PubMed] [Google Scholar]
  • 87. Gribble PL, Ostry DJ, Sanguineti V, Laboissière R. Are complex control signals required for human arm movement? J Neurophysiol. 1998;79(3):1409–1424. 10.1152/jn.1998.79.3.1409 [DOI] [PubMed] [Google Scholar]
  • 88. Pruszynski JA, Kurtzer I, Lillicrap TP, Scott SH. Temporal evolution of “automatic gain-scaling”. J Neurophysiol. 2009;102(2):992–1003. 10.1152/jn.00085.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89. Hammond PH. Involuntary activity in biceps following the sudden application of velocity to the abducted forearm. The Journal of physiology. 1955;127:23–5P. [PMC free article] [PubMed] [Google Scholar]
  • 90. Yeo SH, Franklin DW, Wolpert DM. When Optimal Feedback Control Is Not Enough: Feedforward Strategies Are Required for Optimal Control with Active Sensing. PLoS Comput Biol. 2016;12:e1005190 10.1371/journal.pcbi.1005190 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.Vanderborght B, Albu-Schaeffer A, Bicchi A, Burdet E, Caldwell D, Carloni R, et al. Variable impedance actuators: Moving the robots of tomorrow. In: Proc. IEEE/RSJ Int. Conf. Intelligent Robots and Systems; 2012. p. 5454–5455.
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007414.r001

Decision Letter 0

Adrian M Haith, Samuel J Gershman

29 Oct 2019

Dear Dr Berret,

Thank you very much for submitting your manuscript 'Stochastic optimal open-loop control as a theory of force and impedance planning via muscle co-contraction' for review by PLOS Computational Biology. Your manuscript has been fully evaluated by the PLOS Computational Biology editorial team and in this case also by independent peer reviewers. The reviewers overall agreed that the approach and results constitute an important advance in our understanding of co-contraction during movement. The reviewers did, however, raise a number of concerns, as well as many suggestions for how the manuscript could be further improved. While your manuscript cannot be accepted in its present form, we are willing to consider a revised version in which the issues raised by the reviewers have been adequately addressed. We cannot, of course, promise publication at that time.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

Your revisions should address the specific points made by each reviewer. Please return the revised version within the next 60 days. If you anticipate any delay in its return, we ask that you let us know the expected resubmission date by email at ploscompbiol@plos.org. Revised manuscripts received beyond 60 days may require evaluation and peer review similar to that applied to newly submitted manuscripts.

In addition, when you are ready to resubmit, please be prepared to provide the following:

(1) A detailed list of your responses to the review comments and the changes you have made in the manuscript. We require a file of this nature before your manuscript is passed back to the editors.

(2) A copy of your manuscript with the changes highlighted (encouraged). We encourage authors, if possible to show clearly where changes have been made to their manuscript e.g. by highlighting text.

(3) A striking still image to accompany your article (optional). If the image is judged to be suitable by the editors, it may be featured on our website and might be chosen as the issue image for that month. These square, high-quality images should be accompanied by a short caption. Please note as well that there should be no copyright restrictions on the use of the image, so that it can be published under the Open-Access license and be subject only to appropriate attribution.

Before you resubmit your manuscript, please consult our Submission Checklist to ensure your manuscript is formatted correctly for PLOS Computational Biology: http://www.ploscompbiol.org/static/checklist.action. Some key points to remember are:

- Figures uploaded separately as TIFF or EPS files (if you wish, your figures may remain in your main manuscript file in addition).

- Supporting Information uploaded as separate files, titled Dataset, Figure, Table, Text, Protocol, Audio, or Video.

- Funding information in the 'Financial Disclosure' box in the online system.

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see here

We are sorry that we cannot be more positive about your manuscript at this stage, but if you have any concerns or questions, please do not hesitate to contact us.

Sincerely,

Adrian M Haith

Associate Editor

PLOS Computational Biology

Samuel Gershman

Deputy Editor

PLOS Computational Biology

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: This paper addresses an important problem: the modelling of feedforward control necessary to deal with large sensory delays in human sensorimotor control i.e. motion planning. Specifically, the paper develops a method to compute the force and mechanical impedance (or muscles activation) for learned arm movements, considering intrinsic motor noise. The method is illustrated by simulating well selected experimental studies from the literature. The presentation is clear and accurate, and i have no major concern. In the following, i will first give general comments, then list minor suggestions.

General comments:

1. The paper seems to target co-contraction of antagonist muscle pairs, however i) mechanical impedance at a joint arises from the co-contraction of the group of all muscles actuating this joint, ii) i guess the technique developed in this paper is valid for such muscle groups, not only for antagonist muscle pairs. If ii) holds, the paper could be formulated in the more general case i).

2. Muscle viscoelasticity varies with their activation, and the human nervous system (NS) can control muscles to shape the interaction with the environment, i.e. impedance control. This environment may involve stable or unstable dynamics, and contain noise. In the paper's current formulation, only internal motor noise is considered, not environmental noise. This should be corrected.

3. Impedance control is shaped by muscle mechanics, stretch reflexes as well as long-delay reflexes, see [Franklin2007, Franklin2008]. Long-delay reflexes are missing in the corresponding description at lines 27-31. Note that they still come naturally under feedforward motor commands as defined by the authors as "defined prior to movement execution".

4. Mechanical impedance is not just stiffness (see e.g. line 33), but can be expressed as corresponding to viscoelasticity. In fact several studies such as [Milner1993] have demonstrated the ability of the NS to adapt wrist viscosity (thus not just stiffness can be controlled, although stiffness and viscosity will co-vary).

5. To facilitate the reading, would suggest adding a figure 1 to describe the different setups corresponding to equations 1, 14-16, close to these equations and with the related parameters.

6. The proposed model can predict muscles activation of movements learned in various stable and unstable environments. Similarly, the model of [Franklin2008,Tee2010] can predict at least the experiments of figures 1,2,4,5. It is currently just mentioned at line 523, but i think a comparison with the new model should be provided. In my understanding:

- The new model determines the trajectory, force and impedance corresponding to the learned behaviour of a limb with known kinematics and dynamics, in a known dynamical environment.

- On the other hand, the model of [Franklin2008,Tee2010] can learn the force and impedance along a reference trajectory, and does not need a-priori knowledge of the plant and environment dynamics or kinematics. Note that while model is formulated in an ad-hoc way in [Franklin2008,Tee2010], it in fact corresponds to the gradient descent minimisation of error and effort as analysed in [Yang2011].

- I guess that the simulations of reaching forward movements with lateral instability, the conditions are different from the experiment [Burdet2001] and the simulations of [Franklin2008,Tee2010], where the external force drops when the hand deviates laterally more than x centimeters from the straight line (the experiment would be dangerous and tiring to carry out without this). This may explain the different terminal muscle activation in the simulations in Fig.4,5.

7. As pointed out by the authors, the major difference of their model to SOC is that it is feedforward while SOC is "closed-loop". I would stress this difference even more by calling the SOC closed-loop control at every opportunity in the text. (also it would be possible to use SOC corresponding to the control community which invented it rather than SOFC used much later in the computational neuroscience community, and FSOC (i.e. feedforward SOC) could then be used instead of SOOC?)

Minor suggestions:

Abstract

While these approaches have yielded valuable insights about motor control, they typically fail explain a common phenomenon known as muscle co-contraction. Co-contraction of agonist and antagonist muscles contributes to modulate the mechanical impedance of the neuromusculoskeletal system (e.g. joint stiffness) and is thought to be mainly under the influence of descending signals from the brain.

->

While these approaches have yielded valuable insights about motor control, they typically fail in explaining muscle co-contraction. Co-contraction of a group of muscles associated to a motor function (e.g. agonist and antagonist muscles spanning a joint) contributes to modulate the mechanical impedance of the neuromusculoskeletal system (e.g. joint viscoelasticity) and is thought to be mainly under the influence of descending signals from the brain.

Optimal feedback (closed-loop) control, preprogramming feedback gains but requiring on-line state estimation processes through long-latency sensory feedback loops,

->

Optimal closed-loop control, ...

Author summary

to explain the planning of force and impedance (e.g. stiffness)

->

to explain the planning of force and impedance (e.g. viscoelasticity)

A major outcome of this mathematical framework is the explanation of a long-known phenomenon called muscle co-contraction (i.e. the concurrent contraction of opposing muscles).

->

A major outcome of this mathematical framework is the explanation of muscle co-contraction (i.e. the concurrent contraction of a group of muscles involved in a motor function).

line 9

On the other hand, stochastic optimal –feedback– control (SOC or SOFC) theory was developed to account for the

->

On the other hand, stochastic optimal control (SOC) was used to account for the ...

16

The SOFC theory led to a number of valuable predictions among which the minimal intervention principle, stating that errors are corrected on-line only when they affect the goal of the task, is a significant outcome [9].

->

The SOC theory led to a number of valuable predictions among which the minimal intervention principle, stating that errors are corrected on-line only when they affect the goal of the task [9].

19

However, these two prominent approaches have in common that they fail to simply account for a fundamental motor control strategy used by the central nervous system

->

However, both of these approaches fail at accounting for a fundamental motor control strategy ...

21

co-contraction or co-activation of antagonist muscles

->

co-contraction or co-activation of muscles groups

27

This effect does not only result from the summation of intrinsic stiffnesses of opposing muscles [20, 21] but also from nonlinear stretch reflex interaction [22, 23].

->

This effect results both from the summation of intrinsic stiffness of muscles around a common joint [20, 21] and reflexes [22,23,Franklin2007].

33

First, co-contraction contributes to modulate the effective limb’s impedance (e.g. joint stiffness),

->

... the effective limb’s impedance (e.g. joint viscoelasticity),

48

More fundamentally, an optimal feedback control scheme requires

->

More fundamentally, a closed-loop optimal control scheme requires ...

53

This may seem to contrast with the feedforward nature of impedance and co-contraction control that has been stressed in several studies [16, 18, 34–36].

->

This may seem to contrast with the feedforward nature of impedance and co-contraction control [16,18,34,Franklin2013B]

(xxx these papers present direct experimental evidence for the feedforward nature of impedance control)

57

As this ability may be limited in some cases (e.g. unstable task or too fast motion), co-contraction

->

As this ability is limited in some cases (e.g. unpredictable interaction with the environment, unstable task or too fast motion), co-contraction

82

Although we use the term open-loop –in the sense of control theory–we do not necessarily exclude the role of automatic short-latency reflexes that contribute to the spring-like behavior of intact muscles beyond their short-range stiffness.

->

... we do not necessarily exclude the role of reflexes that contribute to the spring-like behavior of intact muscles beyond their short-range stiffness.

90

Our working hypothesis is that both force and impedance are planned

->

Our working hypothesis is that both force and mechanical impedance are planned

93

open-loop controls

->

open-loop control

128

where R, Q and Qf are positive definite and positive semi-definite matrices with appropriate dimensions

respectively.

->

where R, Q are positive definite matrices and and Qf is a positive semi-definite matrix, all of appropriate dimensions.

130

it can be put out of the expectation

->

it can be taken outside the expectation value integral

145

has a nonlinear dynamics

->

has nonlinear dynamics

152

in agreement with the well-known minimum variance model

->

in agreement with the minimum variance model

173

1st order Taylor approximations

->

first order Taylor approximations

lines 200 to 220:

is it necessary to invoke Feldman (thus a physiological hypothesis) here, or would a linearisation do the same job

230

To illustrate an enlightening point, let us focus on horizontal movements now. The system then simplifies as follows:

->

Focusing on horizontal movements, the system then simplifies to:

268

A two degrees-of-freedom (dof) version of the arm with 6 muscles was also considered to simulate planar arm reaching movements, corresponding to the full model of [51].

->

A two degrees-of-freedom (dof) version of the arm with 6 muscles was also considered to simulate planar arm reaching movements. This is exactly the full model described in [51].

273

C is the Coriolis/centripetal term

->

C \\dot{q} is the Coriolis/centripetal term

276

The net joint torque vector was a function

->?

The net joint torque vector is a function

291

In the seminal study of Hogan described above [46],

->

In Hogan's study [46] described above,

302:

i) add: "where the parameters are defined by Eq.1."

ii) what are the units of the parameters in this simulation?

309

in such an unstable posture

->

in the unstable posture

311

(remind that we prevent feedback control)

->

(remind that feedback control is prevented)

314

In the loaded case, the task instability is increased

-> what does this mean? what would be the measure of stability?

Fig.1

the lines are currently hardly visible. To increase the visibility, one could e.g. reduce the position range to e.g. [-3,3] and the velocity range to e.g. [-10,10], and indicate the standard deviation using e.g. a fine dotted line?

348

in order to model that co-contraction does not lead to increased variability

->

in order to model that co-contraction does not lead to increased variability [Burdet2001]

Fig.2 For these simulations, why using different q_var values? could for example all simulations be done with q_var=5000 ?

362

behavior of subjects described in [61]

->

behavior of subjects in this experiment

Fig.3, similar to Fig.1, could the visibility be improved?

397

Therefore, it was impossible for the subjects to predict

->

Because the hand would start with random lateral deviation due to motor noise, it was not possible for the subjects to predict ...

404

(e.g. participants kept co-contracting when the divergent force field was unexpectedly removed) [18, 65–67].

->

[18, Franklin2003B,66]

418

there was a paper by Wolpert around 1996 showing how the visual feedback can lead to deforming the hand trajectory, which may back the use of a jerk term for this simulation

445

We noticed this is actually a limit of the 6-muscle model used in these simulations, which does not allow arbitrary geometries for the endpoint stiffness in a given posture

->

xxx note that the geometry of the 2 link model allows modifying the stiffness ellipse shape and orientation, see e.g. [Tee2010]

similarly the difference to [Tee2010] could be mentioned in lines 450-460.

462

Finally, we revisit the minimum intervention principle [9]. This well-known principle is most simply illustrated in a pointing-to-a-line task as in [9, 69, 70].

->

Finally, we revisit the minimum intervention principle [9] as illustrated in a pointing-to-a-line task in [9, 69, 70].

502

the consideration of open-loop controls

->

the consideration of open-loop control, ...

516

effort and energy-like criteria are often minimized in optimal control models which tends to prevent

->

... which tend to prevent

523

Researchers have nevertheless attempted to explain co-contraction or its contribution to impedance in existing DOC or SOFC frameworks, but this was often an ad-hoc modeling [75,76].

->

xxx i believe this requires more discussion, and a comparison with this paper's results. To note, [75, Franklin2008] are formulated in an ad-hoc way, but the underlying mathematical principle is described in [Yang2011].

[Milner1993] TE Milner and C Cloutier C (1993), Compensation for mechanically unstable loading in voluntary wrist movement. Experimental Brain Research 94(3): 522-32.

[Franklin2007]

DW Franklin, G Liaw, TE Milner, R Osu, E Burdet and Kawato (2007), Endpoint stiffness of the arm is directionally tuned to instability in the environment. Journal of Neuroscience 27(29): 7705-16.

[Franklin2008]

DW Franklin, E Burdet E, KP Tee, R Osu, CM Chew, TE Milner and M Kawato (2008), CNS learns stable, accurate, and efficient movements using a simple algorithm. Journal of Neuroscience 28(44): 11165-73.

[Franklin2003B]

DW Franklin, E Burdet, R Osu, M Kawato, TE Milner (2003), Functional significance of stiffness in adaptation of multijoint arm movements to stable and unstable dynamics. Experimental Brain Research 151(2): 145-57.

[Tee2010]

KP Tee, DW Franklin, M Kawato, TE Milner and E Burdet (2010), Concurrent adaptation of force and impedance in the redundant muscle system. Biological Cybernetics 102(1): 31-44.

[Yang2011]

C Yang, G Ganesh, S Haddadin, S Parusel, A Albu-Schaeffer and E Burdet (2011), Human-like adaptation of force and impedance in stable and unstable interactions. IEEE Transactions on Robotics 27(5): 918-30.

Reviewer #2: This paper proposes a stochastic optimal open-loop control theory which enable to plan the movement and the stiffness of a biological system moved by antagonistic muscles. The core of the work is to show how such a model is able to easily exploit the use of co-contraction to account for task uncertainties/disturbances.

The idea is very interesting, and the paper presents a novel contribution which is worth for consideration. However, I have several comments that I would ask the authors to consider as suggestions to improve their manuscript.

The major drawback that I see in this work is related to the significant variability in the definition of cost functions to implement the approach in the different experimental conditions. I am pretty convinced that there could be a unique definition of a general problem definition/cost function able to “work” in all the conditions. This is also motivated by a biological counterpart; indeed it is pretty unlikely that the human motor control employ different feedforward strategies for different tasks, but rather I would expect an unifying framework (which is one of the main point of strength for the equilibrium point hypothesis). I think that this aspect is at least worth a discussion in the manuscript, together with a clarification on the particular choices in defining the optimisation problems.

Additional comments are provided below, divided in major and minors.

Majors:

- In Fig. 1A, the plot of variance in position and velocity plots is not visible, maybe the authors could try with a different set of colors.

- The description of Fig.1 is a bit confused, I would suggest naming the 4 subplots of subfig A (and the same for B) and refer to those labels in the caption in an ordered way.

- I would expect that, as soon as an equilibrium is reached, the parameters are maintained constant for the whole execution. In the simulations shown in Fig.1, instead, it seems that the optimal solution shows some oscillations in the last 0.5 seconds. Is there a modeling reason for this, or it is related to the optimization itself? I think this is a relevant aspect to discuss.

- In section “Reaching task with the forearm” the authors refer to Fig. 2C to show the effect of trajectory-time on the resulting stiffness. However, this is not completely clear in the figure. I guess the higher trajectory-time the lower overall torque, but explicitly indicating the time dependency of the stiffness would be beneficial. Moreover, it could be really interesting observing and comparing the whole optimal stiffness profile at different trajectory-time values (with a suitable time-scaling to enable the comparison).

- Also, in the simulations of section “Reaching task with the forearm” I observe an oscillation in the optimal impedance at the beginning and at the end of the task, is there a “methodological” reason for this? I would expect instead a steady value (as shown in the reference [54] for humans)

- Fig 4, the plot of velocities is not clear. Please report them in dedicated subfigures. Controls and Muscles Tensions subfigures are not explained in the caption.

- I would include further discussions regarding the following points:

o what happens if the stiffness cartesian matrix is not diagonal?

o what happens if there is an unpredicted interaction with the environment (e.g. a contact with the environment, thus a force in a specific direction)?

o Does this approach scale well with the dimensionality of the problem? E.g. is it possible to generalize to full upper limb models?

o is it possible to model dual arm constraints (e.g. executing a task while holding an object)?

Minors and Typos:

- In abstract: “fail to explain”

- In abstract, the sentence “Optimal feedback (closed loop) control, preprogramming feedback gains but requiring on-line state estimation processes through long-latency sensory feedback loops, may then complement this nominal feedforward motor command to fully determine the limb’s mechanical impedance.” Is too long, I would suggest to rephrase by splitting in two.

- In caption of Fig. 1A, “corresponding individual muscle torques are depicted below (black for the flexor activation and gray for the extensor activation)” shouldn’t be filled and dashed line instead?

- Line 210 --- kd = sqrt(iks)? Is not 1/2

- Line 225 --- Weight factors α, β and qvar can be chosen to adjust the optimal behavior of the system. How do you select these parameters?

- Eq 26 should be followed by a comma and not a dot

Reviewer #3: In the article, the authors use an Optimal Control framework to develop multiple models in different state-spaces (muscle level and joint level) to show that the optimal control principles can be used to explain the co-contraction in human movements. This is an important attempt, as the modelling of co-contraction has not been done in this framework before, and the simulation results have replicated multiple experimental results.

Although this is a very interesting approach I have several concerns.

Major.

The key idea in the joint level model is that there is a reference trajectory, which is controlled by joint torque, and the deviation from this trajectory, which is controlled by the co-contraction. The two parts are separable. In other words, there is the part of the model that deal with the trajectory planning, which is not new when it comes to modelling. This planned trajectory is subjected to system noise, and therefore there is another, one-input controller, where the only control input is stiffness. Is this then a non-trivial result, that such model predicts stiffness control? Would the results still hold, if the control was dependent on the torque too?

Inferring from Fig. 1, numerically the state x and control u are of the similar order of magnitude. However, the selected costs for that model Q and R differ by 3 or 4 orders of magnitude (line 303). This would mean that the effect of activation cost is small compared to the state dependent costs, suggesting that the co-contraction is energetically cheap, which is not the case in humans. Is this correct? What is the relative weighting of these costs? How sensitive are the results to this cost? Normally, the cost parameters are selected so that the effects of the separate modalities are comparable, otherwise why have it as a cost in the first place?

Specific.

The joint-level model description considers the model as open-loop. However, the behaviour contains the corrections from the reference trajectory, which is clearly feedback control. The authors should clarify what they mean by open loop in this case?

Joint level modelling is clearly described and easy to follow. However, the muscle level modelling lacks clarity in definition. Authors provide the equations for the mechanical model behaviour, but the implementation of the controller (at least to me) is unclear and not nearly at the level of the joint level model.

I understand that it is probably beyond the scope of the paper, but I would like to see (at least the discussion) of how such model would extend to the case where the feedback control is available. Would co-contraction still be present?

The 6-muscle model is unable to learn to increase the endpoint stiffness only in the direction of the instability (Figure 5A). However previous muscle-based models have shown that this is possible when considering costs of stability, accuracy and metabolic cost (Franklin et al., J Neuroscience, 2008; Tee et al., Biological Cybernetics, 2010; Kadiallah et al., PLoS ONE, 2012). Is this because of the specific parameters of the muscle/joint model that was used, or specific to the newly developed SOOC model. Could this result from the extreme low cost of co-contraction such that learning the specific endpoint impedance is not necessary/optimal?

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: No: The code for the simulations is only available from the authors on request. According to my understanding this should be freely available online after acceptance of the manuscript.

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Etienne Burdet

Reviewer #2: Yes: GIUSEPPE AVERTA

Reviewer #3: No

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007414.r003

Decision Letter 1

Adrian M Haith, Samuel J Gershman

17 Dec 2019

Dear Dr Berret,

Thank you very much for submitting your manuscript, 'Stochastic optimal open-loop control as a theory of force and impedance planning via muscle co-contraction', to PLOS Computational Biology. As with all papers submitted to the journal, yours was fully evaluated by the PLOS Computational Biology editorial team, and in this case, by independent peer reviewers. The reviewers were satisfied with the revisions to the paper. However, Reviewer 3 still has some outstanding concerns. Although these comments do not require major revisions, it would helpful to briefly clarify the extent to which the SOOC approach might be applicable in more complex settings (e.g with gravity or for a 2-link arm - see their Q1), and whether the failure to predict endpoint stiffness in unstable environments reflects a limitation of the solution method or a limitation of the underlying model (see their Q3).

We would therefore like to ask you to modify the manuscript according to the review recommendations before we can consider your manuscript for acceptance. Your revisions should address the specific points made by each reviewer and we encourage you to respond to particular issues Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.raised.

In addition, when you are ready to resubmit, please be prepared to provide the following:

(1) A detailed list of your responses to the review comments and the changes you have made in the manuscript. We require a file of this nature before your manuscript is passed back to the editors.

(2) A copy of your manuscript with the changes highlighted (encouraged). We encourage authors, if possible to show clearly where changes have been made to their manuscript e.g. by highlighting text.

(3) A striking still image to accompany your article (optional). If the image is judged to be suitable by the editors, it may be featured on our website and might be chosen as the issue image for that month. These square, high-quality images should be accompanied by a short caption. Please note as well that there should be no copyright restrictions on the use of the image, so that it can be published under the Open-Access license and be subject only to appropriate attribution.

Before you resubmit your manuscript, please consult our Submission Checklist to ensure your manuscript is formatted correctly for PLOS Computational Biology: http://www.ploscompbiol.org/static/checklist.action. Some key points to remember are:

- Figures uploaded separately as TIFF or EPS files (if you wish, your figures may remain in your main manuscript file in addition).

- Supporting Information uploaded as separate files, titled 'Dataset', 'Figure', 'Table', 'Text', 'Protocol', 'Audio', or 'Video'.

- Funding information in the 'Financial Disclosure' box in the online system.

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com  PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

We hope to receive your revised manuscript within the next 30 days. If you anticipate any delay in its return, we ask that you let us know the expected resubmission date by email at ploscompbiol@plos.org.

If you have any questions or concerns while you make these revisions, please let us know.

Sincerely,

Adrian M Haith

Associate Editor

PLOS Computational Biology

Samuel Gershman

Deputy Editor

PLOS Computational Biology

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The paper is fine to me and the authors have addressed the previous comments in a suitable way.

Three minor points could be addressed:

1. The response had:

C \\dot{q} is the Coriolis/centripetal term

Answer: Changed.

however at line 275 i still read "C is the Coriolis/centripetal term"

i do not know whether this is an overlook.

2. Reference [47] seems to be incomplete:

47. Berret B, Jean F. Efficient computation of optimal open-loop controls for stochastic systems; 2019.

3. In the discussion at lines 570-605, if i understand correctly, mechanical impedance is still ascribed to intrinsic mechanical properties and spinal reflexes. However in the studies of Franklin et al. 2007, 2008 etc. mechanical impedance seems to stem also from long delay reflexes.

Reviewer #2: The authors addressed all the comments I had on the first version of the manuscript and, I believe, the final version of this work is suitable for publication.

Reviewer #3: The authors have revised the paper and answered several of my concerns, however I feel that three of these issues are not sufficiently addressed in the current revision.

Q1:

Original: The key idea in the joint level model is that there is a reference trajectory, which is controlled by joint torque, and the deviation from this trajectory, which is controlled by the co-contraction. The two parts are separable. In other words, there is the part of the model that deal with the trajectory planning, which is not new when it comes to modelling. This planned trajectory is subjected to system noise, and therefore there is another, one-input controller, where the only control input is stiffness. Is this then a non-trivial result, that such model predicts stiffness control? Would the results still hold, if the control was dependent on the torque too?

Answer: This joint-level model is used for the purpose of illustrating how stiffness control may arise in the proposed framework and for illustrating this theoretical uncoupling of net torque control and stiffness control. However, in more general cases (e.g. a planar 2-dof arm or a 1-dof arm with gravity torque), this result would not hold. However, the same theoretical derivation can still be applied. This was said Line 228 and Lines 243-246.

Response: If this result does not hold for a realistic case where there is gravity then how general is this solution? This feels like an important point that is glossed over in the current manuscript.

Q2:

Original: Inferring from Fig. 1, numerically the state x and control u are of the similar order of magnitude. However, the selected costs for that model Q and R differ by 3 or 4 orders of magnitude (line 303). This would mean that the effect of activation cost is small compared to the state dependent costs, suggesting that the cocontraction is energetically cheap, which is not the case in humans. Is this correct? What is the relative weighting of these costs? How sensitive are the results to this cost? Normally, the cost parameters are selected so that the effects of the separate modalities are comparable, otherwise why have it as a cost in the first place?

Answer: In Figure 1, we plotted angles in degrees but in the mathematical model they would be expressed in radians. Hence the order of magnitude is not really similar (x has values much smaller than u actually; typically, x is <0.05 radian while u>1 Nm). Given that these values are squared and integrated over time, this means that there are several orders of magnitude of difference between the variance magnitude and the effort magnitude. Hence, the weights for R and Q which are indeed adapted to make the two cost components (effort and variance) of comparable magnitude. Regarding the sensitivity of the results with respect to the weights of the cost, this is not very sensitive when the weights are similar orders of magnitude. A fine-tuning of the weights was not needed in these simulations.

Response: The authors response to this is clear, however it would be helpful to mention in the paper that these two cost components are of similar magnitude. As well, as the authors mention that the modelling is done in radians, Figure 1 could also use radians for easier comparison.

Q3

Original: The 6-muscle model is unable to learn to increase the endpoint stiffness only in the direction of the instability (Figure 5A). However previous muscle-based models have shown that this is possible when considering costs of stability, accuracy and metabolic cost (Franklin et al., J Neuroscience, 2008; Tee et al., Biological Cybernetics, 2010; Kadiallah et al., PLoS ONE, 2012). Is this because of the specific parameters of the muscle/joint model that was used, or specific to the newly developed SOOC model. Could this result from the extreme low cost of co-contraction such that learning the specific endpoint impedance is not necessary/optimal?

Answer: It is likely that other muscle models may have been used to change the shape of the stiffness ellipse (as mentioned in the referred works). Here we used the model of Katayama and Kawato (with exactly their parameters). When checking all the possible orientations of the stiffness ellipse in the considered posture, we could not find any muscle activation vector that would make the ellipse horizontal (but note that this was possible for other arm postures). We added a few more words on this issue (Lines 450-460). Finally, please note that co-contraction is always a costly strategy in our models, and it is not negligible compared to the variance cost (thanks to the weights in the cost function).

Response: This still does not address my concern. Is the SOOC model unable to find a solution that matches human subjects, or is it the feature of the kinematic model used. If the authors could implement this SOOC approach on a model that was already shown to produce this effect, we would know where this limitation arises. The fact that the kinematic model can’t produce this result does not show that the SOOC model can produce this. I realize that the Cartesian mass-point model is a partial approach but still does not fully address the issue.

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

Reviewer #2: None

Reviewer #3: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Etienne Burdet

Reviewer #2: Yes: Giuseppe Averta

Reviewer #3: No

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007414.r005

Decision Letter 2

Adrian M Haith, Samuel J Gershman

23 Dec 2019

Dear Dr Berret,

We are pleased to inform you that your manuscript 'Stochastic optimal open-loop control as a theory of force and impedance planning via muscle co-contraction' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. Please be aware that it may take several days for you to receive this email; during this time no action is required by you. Once you have received these formatting requests, please note that your manuscript will not be scheduled for publication until you have made the required changes.

In the meantime, please log into Editorial Manager at https://www.editorialmanager.com/pcompbiol/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production and billing process.

One of the goals of PLOS is to make science accessible to educators and the public. PLOS staff issue occasional press releases and make early versions of PLOS Computational Biology articles available to science writers and journalists. PLOS staff also collaborate with Communication and Public Information Offices and would be happy to work with the relevant people at your institution or funding agency. If your institution or funding agency is interested in promoting your findings, please ask them to coordinate their releases with PLOS (contact ploscompbiol@plos.org).

Thank you again for supporting Open Access publishing. We look forward to publishing your paper in PLOS Computational Biology.

Sincerely,

Adrian M Haith

Associate Editor

PLOS Computational Biology

Samuel Gershman

Deputy Editor

PLOS Computational Biology

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1007414.r006

Acceptance letter

Adrian M Haith, Samuel J Gershman

18 Feb 2020

PCOMPBIOL-D-19-01595R2

Stochastic optimal open-loop control as a theory of force and impedance planning via muscle co-contraction

Dear Dr Berret,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Laura Mallard

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: response_file3.pdf

    Attachment

    Submitted filename: response_file.docx

    Data Availability Statement

    There is no new experimental data in this study. However, the Matlab codes used to produce the simulations are available from http://hebergement.universite-paris-saclay.fr/berret/software/SOOC.zip. Note that the codes depend on a third-party software (GPOPS).


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES