Table 1.
Terminology
Model | |||
---|---|---|---|
Time | |||
Trial | All | ||
Sources of variability | |||
, | Motor noise | Inevitable variability that is always present. Also called unregulated variability (Dhawale et al. 2019) | Cashaback19 |
Dhawale19 | |||
Therrien16 | |||
Therrien18 | |||
, | Exploration | Variability that can be added and can be learnt from. Also called regulated variability (Dhawale et al. 2019) | Cashaback19 |
Dhawale19 | |||
Therrien16 | |||
Therrien18 | |||
Input exploration | |||
Exploration estimated following non-successful trials | |||
Exploration estimated following successful trials | |||
(A)TTC estimate of exploration | |||
Movement generation | |||
Aim point | Mean of the probability density of movement endpoints given a certain ideal motor command (van Beers 2009) | All | |
End point | Observable movement outcome | All | |
Reward-based motor learning | |||
Reward presence or absence | R = 0: no reward | All | |
R = 1: reward | |||
Reward prediction error | Difference between actual reward obtained and predicted reward | Dhawale19 | |
RPE > 0: Reward obtained | |||
RPE < 0: No reward obtained | |||
Low-pass filtered reward history | Low-pass filtered reward history of the τ previous trials | Dhawale19 | |
Reward-based learning parameter | Learning gain, adjustment fraction | Cashaback19 | |
Dhawale19 | |||
Therrien16 | |||
Therrien18 | |||
Reward rate update fraction | Gain of updating the reward rate estimate () with the most recent trial outcome | Dhawale19 | |
[tau] | Number of trials in reward history memory window | Inferred memory window for reinforcement on past trials, or the time-scale of the experimentally observed decay of the effect of single-trial outcomes on variability (Dhawale et al. 2019) | Dhawale19 |