|
Trial |
|
All |
Sources of variability |
|
|
|
,
|
Motor noise |
Inevitable variability that is always present. Also called unregulated variability (Dhawale et al. 2019) |
Cashaback19 |
Dhawale19 |
Therrien16 |
Therrien18 |
,
|
Exploration |
Variability that can be added and can be learnt from. Also called regulated variability (Dhawale et al. 2019) |
Cashaback19 |
Dhawale19 |
Therrien16 |
Therrien18 |
|
|
Input exploration |
|
|
|
Exploration estimated following non-successful trials |
|
|
|
Exploration estimated following successful trials |
|
|
|
(A)TTC estimate of exploration |
|
Movement generation |
|
|
|
|
Aim point |
Mean of the probability density of movement endpoints given a certain ideal motor command (van Beers 2009) |
All |
|
End point |
Observable movement outcome |
All |
Reward-based motor learning |
|
|
|
|
Reward presence or absence |
R = 0: no reward |
All |
R = 1: reward |
|
Reward prediction error |
Difference between actual reward obtained and predicted reward |
Dhawale19 |
|
|
RPE > 0: Reward obtained |
|
|
|
RPE < 0: No reward obtained |
|
|
Low-pass filtered reward history |
Low-pass filtered reward history of the τ previous trials |
Dhawale19 |
|
Reward-based learning parameter |
Learning gain, adjustment fraction |
Cashaback19 |
Dhawale19 |
Therrien16 |
Therrien18 |
|
Reward rate update fraction |
Gain of updating the reward rate estimate () with the most recent trial outcome |
Dhawale19 |
[tau] |
Number of trials in reward history memory window |
Inferred memory window for reinforcement on past trials, or the time-scale of the experimentally observed decay of the effect of single-trial outcomes on variability (Dhawale et al. 2019) |
Dhawale19 |