Skip to main content
Springer logoLink to Springer
. 2026 Feb 12;244(3):43. doi: 10.1007/s00221-026-07237-6

Identifying motor learning deficits in neurological conditions: a critical analysis of a perennial problem

Rajiv Ranganathan 1,2,, Sharo Costa 3,#, Angelyn May Mengote 3,#, Megan Thomas 3,#, Rohan Ved 3,#
PMCID: PMC12901082  PMID: 41677941

Abstract

Identifying deficits in motor learning has the potential to serve as an indicator of brain function in several neurological conditions. However, evidence for motor learning deficits can be confounded by factors unrelated to learning. In this review, we critically examine the evidence for these deficits in neurological conditions, focusing both on conceptual and methodological issues. Across a wide range of neurological conditions, we found that a majority of evidence for motor learning deficits is difficult to clearly interpret as learning-related and is potentially confounded by the presence of baseline differences. In addition, the use of low sample sizes, short time durations of practice, and the narrow focus on sequence learning paradigms also raise questions about the validity and generality of this evidence. Given that deficits in motor learning have implications for the early detection of neurological status and the design of rehabilitation strategies, we highlight the need for greater rigor when addressing this important, but perennially challenging, question.

Keywords: Baseline, Skill, Learning, Performance, Parkinson’s disease, Stroke

Introduction

There has been a long-standing interest in understanding motor learning differences associated with neurological conditions (Milner 1965; Corkin 1968; Heilman et al. 1975; Sanes et al. 1990). Although some motor learning studies have focused on conditioning of responses and reflexes (Glickstein and Yeo 1990; Thompson et al. 2009), here we focus on motor learning in the context of practice-related improvements in task performance. This process of motor learning has been associated with several different brain regions, including the basal ganglia, cerebellum, primary motor cortex, and premotor cortex, and depends on a number of factors such as the type of task being learned and the stage of learning (Willingham 1998; Doyon and Benali 2005; Seidler 2010; Penhune and Steele 2012; Spampinato and Celnik 2021). In light of these findings, identifying deficits in motor learning has been viewed as having the potential to reveal unique insights into neurological function, which have implications for early non-invasive identification of disease states, and for designing effective rehabilitation strategies.

Deficits in motor learning have been reported in a wide range of neurological conditions including Parkinson’s disease (Harrington et al. 1990; Krebs et al. 2001), stroke (Boyd and Winstein 2004), Alzheimer’s disease (Ferraro et al. 1993; Rouleau et al. 2002), autism (Mostofsky et al. 2000) and dyslexia (Stoodley and Stein 2006). However, the evidence for such learning differences has been somewhat mixed as several studies show that motor learning may not be affected in these populations (Grafman et al. 1990; Agostino et al. 1996; Winstein et al. 1999). There have been attempts to resolve these conflicting findings by distinguishing between the type of task, level of impairment, extremity tested, or the learning mechanism involved (Ashe et al. 2006; van Halteren-van Tilborg et al. 2007; Nieuwboer et al. 2009; Marinelli et al. 2017). However, given that motor learning encompasses several processes (Krakauer et al. 2019), these mixed findings raise the need for a greater understanding of both conceptual and methodological aspects of how such deficits are quantified across studies (Ranganathan et al. 2022c).

Addressing the question of whether there is a motor learning deficit has several challenges, primarily because of the “learning-performance distinction” (Salmoni et al. 1984; Schmidt and Bjork 1992; Kantak and Winstein 2012). Because motor learning is inferred based on observed behavior, several factors (such as fatigue, motivation, or attention) can influence performance without affecting learning. Specifically, in the context of neurological conditions, one factor that complicates this issue is that clinical populations often do not start at the same initial level of motor performance as age-matched unimpaired controls (i.e., they are at a different ‘baseline’ prior to learning). This means that the same results can be interpreted in different (sometimes even contradictory) ways depending on how learning is defined (Schmidt 1972; Kitago and Krakauer 2010). For example, when there are no group differences at baseline, different measures of learning using absolute measures (such as final performance level) or relative measures (such as change or relative change scores) will tend to result in the same conclusion (Fig. 1A). However, when the baseline levels are different, the question of which group “learned more” depends on how learning is measured. For example, consider two groups that have the same exponential learning rate, but plateau at different levels (Fig. 1B). In this case, using measures such as final performance level or relative change could indicate differences between groups, whereas using measures a learning rate or a change score could indicate no differences between the groups (depending on the exact values of baseline and plateau performance levels). Similarly, other such inconsistencies among learning metrics in the presence of baseline differences can be shown where the two groups have (i) same learning rates but different change scores (Fig. 1C), (ii) same learning rates but different relative change scores (Fig. 1D), and (iii) different learning rates but same change scores (Fig. 1E). These examples highlight the conceptual difficulty in measuring learning when baseline levels are different between groups. In addition to these conceptual factors, methodological factors associated with the evidence for learning deficits also need to be considered because they determine the strength of the evidence for a deficit (e.g., the sample size), as well as the implications of these findings for real-world applications such as rehabilitation (e.g., the type of task and the time duration of practice).

Fig. 1.

Fig. 1

Assessing learning deficits in the presence of baseline differences. Hypothetical learning curves are presented for two groups (indicated in blue and red), with practice on the x-axis and movement time/error on the y-axis in arbitrary units (such that decreasing values indicate “better” performance). Learning deficits are assessed using four different learning measures (final level, change, relative change, and rate). When there are no baseline differences, A all measures of learning tend to agree. However, in the presence of baseline differences (B)–(E), conclusions can be ambiguous because they are highly dependent on the choice of the learning measure

To characterize the prevalence of these issues in the literature, we reviewed studies that address the issue of motor learning deficits in neurological populations. Given the wide range of experimental paradigms used, we use the term ‘learning’ here to include any practice-dependent changes in task performance (without making the distinction between transient and persistent changes as highlighted in the learning-performance distinction (Salmoni et al. 1984; Schmidt and Bjork 1992; Kantak and Winstein 2012)). Moreover, because our focus was on the measurement of motor learning deficits in general, we sampled across a broad set of neurological conditions and critically evaluate the evidence provided for motor learning differences. We highlight critical conceptual and methodological issues involved in such studies and provide guidelines for future studies to address these questions with greater rigor.

Methods

The Web of Science database was used for the literature search, with the following keywords: “motor learning deficits”, “stroke and motor learning”, “motor learning and rehabilitation”, and “motor learning and neurological deficits”. Papers generated from the search (n = 268) were screened based on the following inclusion criteria: (i) the paper was an experimental paper, (ii) it was based on human-subject studies with a focus on neurological conditions, (iii) there was the presence of a control group for comparison, and (iv) the study tested a motor learning component (i.e., participants practiced a motor task over multiple trials with performance tracked over time/trials). It is important to note that because the focus of our review was to examine practices over a wide range of motor learning studies, we did not restrict our review based on any specific definitions (theoretical or practical) or measurements of motor learning.

All studies that did not meet any one of the inclusion criteria listed above, were excluded. The studies were screened twice, each time by a different author, to make sure they met the study criteria, and a third author acted as a tie-breaker when needed. For papers reporting multiple experiments or multiple tasks, each experiment was considered a separate study. Overall, a total of 58 articles (covering 72 studies) were included based on the criteria that were then subject to a full-text review (Fig. 2).

Fig. 2.

Fig. 2

PRISMA graph of study selection. A total of 58 papers (covering 72 studies) on motor learning deficits were included for this review

Data analysis

For each included study, the following data were extracted: (i) the neurological condition being investigated, (ii) the type of task studied, (iii) whether the study made a claim of a learning deficit, (iv) the type of deficit being examined, (v) the type of learning measure being examined, (vi) whether there were initial baseline differences in motor performance between experimental and control groups, and (vii) strategies to account for baseline differences, (viii) average group sample size, and (ix) the time duration of motor learning.

The neurological condition being investigated was based on the description of the participant group. The type of task was selected based on the classification from (Ranganathan et al. 2021), which divides the task into one of six categories - adaptation, applied, coordination, sequence learning, tracking, and variability. Briefly, adaptation tasks involved modification of well-learned movements (e.g., visuomotor rotation), applied tasks involved real-world tasks (e.g., a soccer kick), coordination tasks involved producing a spatiotemporal pattern using more than a single degree of freedom (e.g., bimanual coordination), sequence learning tasks involved producing a sequence of several movement responses (e.g., key pressing), tracking tasks involved producing a time-varying spatiotemporal pattern (e.g., tracking a moving target), and variability tasks involved producing a steady-state performance (e.g., throwing to a specific target). As with prior work, although some tasks did not fall exclusively into one category, we decided to code each task into the ‘best fitting’ category. The rationale behind coding the type of task was to examine the underlying learning mechanisms involved in learning.

The claim of a learning deficit was coded as a “yes”, “no”, or “unclear” based on the authors’ interpretation of the results. The type of deficit being examined was classified either as a “learning mechanism” or a “rate” deficit. Learning mechanism deficits involved comparison of learning in two different conditions (e.g., as characterized by a group × condition interaction). Rate deficits involved comparison of how ‘quickly’ participants learned the task relative to controls (e.g., characterized by a group × time interaction or a learning rate parameter comparison). The type of learning measure used to infer the deficit was classified based on one of four categories – (i) absolute level: the value of the dependent variable is directly compared between groups, (ii) change score: the difference in the dependent variable (between two or more conditions or time points) is compared between groups, (iii) normalized change: the ‘normalized’ or ‘relative’ difference (e.g., a % improvement) in the dependent variable (between two or more conditions or time points) is compared between groups and (iv) learning rate: the learning rate (e.g., as estimated by linear slopes or exponential parameters) is compared between groups.

Baseline difference was coded as a “yes” (baseline difference in performance between the experimental and control groups was reported in the manuscript and this difference was statistically significant), “no” (baseline difference in performance between the experimental and control groups was reported in the manuscript and this difference was not statistically significant) or “unclear” (no statistical comparisons were reported, or statistical comparisons were not made between the groups under identical task conditions).

Strategies to account for baseline differences were coded as: (i) using within-subject control - i.e., comparing the participant’s performance relative to their own performance in another condition, (ii) statistical adjustments (e.g., adding covariates or normalizing the dependent variable), (iii) task modification – i.e., modifying the task to make performance equivalent between groups (e.g., changing the target speed in a tracking task), or (iv) none.

The sample size of the study was determined as the average number of participants per group in the experiment (i.e., total sample size divided by number of groups; so this number could be a non-integer). Time duration was coded based on the length of time (measured in days) of practice, as a measure of its relevance toward capturing long-term learning and rehabilitation.

Results

The full table of articles with references is available at https://osf.io/v6af4/. The main results are summarized below.

Motor learning deficits were reported across several neurological conditions

The review included several neurological conditions (Fig. 3A), with the top categories being Parkinson’s disease (n = 13, 26.4%), stroke (n = 6, 8.3%), and developmental coordination disorder (n = 5, 6.9%). We found that the presence of motor learning deficits was reported in a majority of these studies (n = 42, 58.3%).

Fig. 3.

Fig. 3

A Neurological conditions where motor learning deficits were assessed. A majority of the studies (57%) reported deficits in motor learning. B Types of task paradigms used for assessing deficits and the nature of the deficit investigated. Sequence learning was the most popular paradigm, followed by adaptation and coordination tasks. Learning rate deficits were investigated more often, although sequence learning paradigms did tend to have a high number of studies that investigated learning mechanisms

Sequence learning was the dominant task paradigm, and the types of learning deficits examined were influenced by the task paradigm

The analysis of task category showed that sequence learning tasks were the dominant paradigm, used in over twice as many studies (n = 41, 56.9%) as the next highest task category (Fig. 3B). Tracking (n = 15, 20.8%) was the next most common category, followed by coordination (n = 7, 9.7%), adaptation (n = 7, 9.7%), and variability (n = 2, 2.8%) tasks.

The analysis of the type of motor learning deficits predominantly depended on the type of task paradigm (Fig. 3B). Studies examining deficits in learning rates were more common overall (n = 41, 56.9%). However, sequence learning (n = 21 out of 41, 51.2%) and adaptation paradigms (n = 4 out of 7, 57%) tended to have a fairly good representation of studies examining deficits in learning mechanisms.

Baseline differences in motor performance were reported frequently, and strategies to account for baseline differences depended on the type of task paradigm

Because baseline differences affect different learning measures differently, we quantified the types of learning measures used. The learning measure most used was a change score (n = 58, 80.5%) (Fig. 4A). This change score typically involved examining performance using a Group x Block interaction where Block could represent time (e.g., early vs. late in learning) or conditions (e.g. a practiced sequence vs. a random sequence). Other measures such as normalized change (n = 6, 8.3%), learning rates (n = 5, 6.9%) and absolute levels (n = 3, 4.1%) were also used, albeit much less frequently.

Fig. 4.

Fig. 4

A Type of learning measure used for assessing deficits – change scores were the most common, which can be problematic when there are baseline differences. B Presence of baseline differences, i.e., difference in performance at the onset of learning between the clinical and the control groups. Over half of the studies reported differences in baseline performance. C Accounting for baseline differences was dependent on the task paradigm. In sequence learning tasks, the majority of paradigms used a control condition with a random sequence acting as a ‘within-subject’ comparison to offset the presence of any baseline differences. However, in other paradigms, there were statistical corrections (such as adding covariates), task adjustments, or no corrections

About a little more than half of the studies (n = 39, 54.1%) reported significant differences in motor performance at baseline (Fig. 4B). The remainder reported no statistically different difference (n = 24, 33.3%) or this information was unclear (n = 9, 12.5%).

Strategies to account for baseline differences also tended to be dependent on the type of task being studied (Fig. 4C). Using a within-subject control was the most common way to account for baseline differences (n = 31, 43.1%) and was extensively used in sequence learning paradigms. Other experimental paradigms tended to adjust for baseline differences statistically (n = 10, 13.9%). Some tasks also adjusted task difficulty (n = 7, 9.7%) to ensure that baseline performances were equivalent between groups. Critically, a relatively large number of studies applied no correction for baseline differences (n = 24, 33.3%), even when significant baseline differences were identified (n = 12 out of 39, 30.8%).

Sample sizes tended to be small and time duration of practice was short

A majority of the studies (n = 51, 70.8%) reviewed used an average group sample size less than or equal to 20 (Fig. 5A). In terms of time duration of practice, most studies (n = 56, 77.7%) were conducted within a 1-day interval, with only a handful of studies (n = 5, 6.9%) going beyond 3 days of practice (Fig. 5B).

Fig. 5.

Fig. 5

A Sample size distribution across studies. Sample sizes were low, but typical of motor learning studies in the literature (~ 10–20/group). B Time duration of motor learning across studies. A majority of the studies focused only on a single day of practice, again typical of most motor learning studies

Discussion

The purpose of this review was to critically evaluate the evidence of motor learning deficits across a wide range of neurological conditions, focused on both conceptual and methodological issues. Across different motor learning paradigms, there was significant variability in the results reported as evidence for the claim of motor learning deficits including: (i) the type of deficits reported and the type of learning measures used to infer learning deficits, (ii) reporting of baseline differences and correction strategies, and (iii) the sample size and time duration of motor learning investigated. From a conceptual standpoint, we found that a large percentage of studies made claims of motor learning deficits even though about half of the studies indicated there were baseline differences in motor performance between experimental and control groups. From a methodological standpoint, we found that most studies used relatively small sample sizes with short time durations of practice. Overall, these conceptual and methodological issues highlight the significant ambiguity in the evidence for motor learning deficits in the literature.

Conceptual issues

Quantifying evidence for motor learning deficits is non-trivial

A central challenge in motor learning is that there is a large degree of fragmentation in defining what constitutes learning (Ranganathan et al. 2021). The use of different measures (e.g., final performance level, change scores, relative change scores, rates of learning) to assess a learning deficit is a non-trivial issue because each type of measure assumes a different model of learning. Furthermore, prior work has shown that even when a specific learning measure like ‘rate of learning’ is used, the same measure can be operationalized in very different ways (e.g., linear slopes, single or double exponentials, power laws) (Ranganathan et al. 2022a). This fragmentation makes it challenging to reliably establish evidence of motor learning deficits and compare them across studies.

This ambiguity in interpretations of learning deficits is further exacerbated when there are baseline differences between groups. Several prior papers on motor learning (Kitago and Krakauer 2010; Schmidt 1972), have argued that in the presence of baseline differences, the same data can be used to argue that a group learned ‘more’ or ‘less’ depending on how the measure of learning is computed (as shown in Fig. 1). For example, using a change score from pre- to post-test (or equivalently, a group × time interaction) relies on an additive model of learning, where the central assumption is that that the change score is independent of the baseline performance. However, this assumption may not be justifiable in many experimental paradigms, because learning curves tend to plateau over time (i.e., participants who have high performance levels have less room to improve and, therefore, are likely to show smaller change scores). Even in cases where the underlying measure accounts for such behavior - e.g., the learning rate parameter of an exponential curve (Krebs et al. 2001), direct comparison of rates is not always straightforward since the form of these functions is quite sensitive to averaging and noise that is often seen in real-world data (Newell et al. 2001; Olivier et al. 2021) and may require additional robust fitting procedures (Zhang et al. 2025). Moreover, learning curves can take a wide range of functional forms (e.g., S-shaped curves) (Liu et al. 2006), which makes estimation and comparison of learning rates across tasks and individuals challenging.

Fixing the issue of baseline differences

The issue of how to account for any baseline differences is also non-trivial, especially in the current context of non-randomized studies, where individuals cannot be randomly assigned to different groups. From a statistical standpoint, Lord (Lord 1967) stated, “there simply is no logical or statistical procedure that can be counted on to make proper allowances for uncontrolled pre-existing differences between groups” (p. 305). Similarly, Kitago and Krakauer (Kitago and Krakauer 2010) refer to the same issue as “a formidable challenge” (p.1250). This has been an ongoing discussion in the statistics literature with the suggestion that while using covariates is a somewhat better strategy compared to using a change score (Senn 2006), the standard experimental design with only two time point measures (pre-test/post-test) is insufficient for an accurate estimate of the effect since it is not possible to separate ‘exogenous change’ (i.e., the change caused by learning) from ‘random change’ (Tennant et al. 2022).

Attempts to ‘equate’ baseline differences between populations by adjusting task difficulty also have challenges. Although adjusting the task to the individual is necessary in some contexts so that the task is feasible (e.g., adjusting a force production task based on the maximum voluntary isometric contraction), in the context of motor learning, adjusting task difficulty can alter the task in ways that are not easy to predict (Willingham et al. 1997). For example, adjusting task difficulty in tracking tasks can alter the shape of the learning curve (Bahrick et al. 1957). Moreover, adjusting parameters like movement speed may also alter the underlying control strategy in terms of the use of feedback and feedforward processes. These highlight the difficulty of interpreting evidence for learning deficits across populations when baseline performance is different.

Perhaps the more convincing evidence for baseline corrections comes from ‘within-subject’ comparisons – where the participant’s own performance in another condition can be used as a control condition to rule out confounds. For example, in serial reaction time tasks, the same participant is tested in the learned sequence and a random sequence, and the difference between these two conditions is assumed to be a specific measure of skill, which minimizes contamination of performance effects (Robertson 2007). However, because these involve comparisons between conditions, these also raise similar issues (i.e., do we compare the absolute change, relative change, etc.). Moreover, these types of designs cannot be easily applied to all types of task paradigms because they require the ability to measure performance in a ‘control’ condition which is a variation of the motor task that can be performed without learning.

Methodological issues

Low sample size

Consistent with prior reviews on motor learning (Lohse et al. 2016; Ranganathan et al. 2021), our review indicated that sample sizes tend to be under 20. While the difficulty of recruiting clinical populations must be recognized, there are well-known issues with using small sample sizes like low power and the inflation of effect sizes (Button et al. 2013). In the current context, this lack of power is especially critical since the primary method of quantifying motor learning deficits involves detecting ordinal interactions (e.g., when both groups learn but one group learns more than the other). These interactions typically require higher power than main effects because the expected effect sizes are smaller (Lakens and Caldwell 2021). Moreover, interpreting non-significant differences in baseline performance as evidence of ‘no baseline differences’ can also be problematic in studies with low power. The latter point about inflation of effect sizes is due to a combination of suboptimal power and the use of selection thresholds (Ioannidis 2008), i.e., because small sample sizes can only detect large effects, only studies that overestimate an effect size will show statistically significant results (Button et al. 2013).

Short time duration of practice

Although the duration of the practice is often seen as a convenience variable (i.e., clinical assessments must ideally be short), there is a danger in using short durations of practice for assessing motor learning deficits. First, the use of short practice durations can often be problematic in clinical populations, who may require greater time to familiarize themselves with novel experimental settings or task instructions. Second, given that motor learning occurs over different time scales of practice, short-term studies primarily tap into the ‘fast time scales’ of learning which are not only transitory (Newell et al. 2009), but may also involve different learning processes related to cognition and explicit learning (McDougle et al. 2015). This overemphasis on the early phase of learning also blurs the distinction between whether participants show persistent deficits (i.e., participants do not reach the same level of performance regardless of the duration of practice) or whether they are simply slower learners (i.e., participants can reach the same level of performance but do so at a slower rate). Last, the use of short-duration experiments also reduces the relevance for tailoring rehabilitation strategies (Olson et al. 2019), which may require weeks or months of training.

Narrow emphasis on sequence learning paradigms

Finally, most of the work was based on the sequence learning paradigm. Although this is somewhat expected in the specific context given the theoretical backdrop of procedural/implicit learning (Penhune and Steele 2012) and the relative ease with which the paradigm can be studied even online (Bönstrup et al. 2020; Weinberger and Green 2022; Cubillos et al. 2023), the narrow focus on this one paradigm also raises the question of what aspects of motor learning are being captured. Sequence learning paradigms typically place relatively less emphasis on the actual movements themselves (instead typically focusing only on movement outcomes such as times and errors), which can obscure certain types of deficits such as the control of variability or multi-segment coordination. This lack of clinically relevant paradigms was also noted in a prior meta-analysis on implicit learning in stroke (Kal et al. 2016), which found only a single article (out of 20 included in the review) that examined learning using a whole-body coordination task. The lack of sufficient representation of other types of paradigms, coupled with the general observation that studies included in the review tended to be lab-based research (which is early in the translation pipeline), raises the risk of overgeneralization to rehabilitation based on tasks with limited ecological validity (Ranganathan et al. 2022b).

Recommendations

In light of these challenges in evaluating evidence of motor learning deficits, we provide the following recommendations. We use our prior work on general guidelines for motor learning studies(Ranganathan et al. 2022c) as a basis for these recommendations, but specifically tailor these recommendations to address the challenges raised in the current context.

Characterizing participant abilities at baseline

Because motor learning in many tasks involves the integration of sensory, motor, and cognitive abilities, and baseline differences in these abilities can affect conclusions, there is a need for better characterization of participant abilities at baseline. Therefore, in addition to condition-specific assessments (such as the Unified Parkinsons’ Disease Rating Scale for Parkinson’s Disease or the Fugl-Meyer for stroke), the use of standardized test batteries such as the NIH toolbox (Gershon et al. 2013) or physiological profile assessments (Lord et al. 2003; Ingram et al. 2019) that evaluate different sensory, motor, and cognitive components across multiple neurological conditions can help provide critical information to make inferences about any observed deficits in motor learning tasks. These assessment batteries also provide estimates of heterogeneity within these populations, which can be useful for (i) assessing if impairment levels in participant sample is representative of the population (especially because the inability to perform the motor task is used as an exclusion criterion for such experiments (Willingham et al. 1997) and (ii) individual-level analyses that go beyond characterizing average differences at the group level (highlighted later in this section).

Building a more robust evidence base with richer experimental designs

Given the learning-performance distinction, our view is that the standard experimental approach of comparing an experimental group against a neurologically unimpaired control group in a single experimental condition is, by itself, not fully compelling evidence of a learning deficit (Ranganathan et al. 2022c). Between-subject designs provide preliminary evidence of an impaired learning mechanism, but given the potential confound of baseline differences, the exact nature of learning deficits may become more apparent when combined with ‘within-subject’ experimental designs that compare learning within the same population using different conditions or training paradigms. For example, prior work has shown that when making reaching movements, cerebellar patients have deficits adapting to a force field when the force field was introduced abruptly, but not when the same force field was introduced gradually over several trials (Criscimagna-Hemminger et al. 2010). An alternative approach that could address the issue of baseline differences is using stratified sampling based on baseline performance (Kernan et al. 1999; Ziv et al. 2020; Krishnan et al. 2025), which can be used in contexts where there is some likelihood of finding individuals in different groups with similar baseline performance (e.g., comparing two different clinical populations). In some contexts, post-hoc matching algorithms could also be used to identify how much of the observed effect in an experiment is likely due to mismatched baselines (Mandava et al. 2010).

Similarly, quantifying learning using paradigms that test participants across longer time scales could also provide more insight. For example, by measuring performance across multiple days in a throwing task in individuals with Parkinson’s disease (instead of relying on a single retention test), it was possible to distinguish motor control deficits from retention deficits (Pendt et al. 2011). These experimental designs may also provide the ability to use more sophisticated statistical models for longitudinal data such as mixed regression models (Lohse et al. 2020) that incorporate time more explicitly into the model.

Developing model tasks and pre-registration

In light of difficulties in identifying appropriate models and variables across various paradigms, a proposed solution for cumulative science is to develop model task paradigms that can be compared across different studies (Ranganathan et al. 2021). In the context of motor learning deficits, there are three specific challenges that need to be addressed. First, the goal is to develop tasks where deficits are not ‘masked’ by motor impairments (Subramanian et al. 2018) or the use of alternative strategies to perform the task (Nicolson and Fawcett 1994). For example, in several task paradigms, there are both explicit and implicit processes involved in learning (Taylor and Ivry 2013), and require specific experimental manipulations to distinguish the relative contribution of these processes. Second, developing tasks that are sensitive enough to detect group differences and associated with specific underlying learning mechanisms must be balanced with relevance to real-world contexts such as rehabilitation. This means that task paradigms need to: (i) focus on movement quality and coordination in addition to task outcome variables such as times and errors, and (ii) be relatively complex enough so that longer timescales of learning beyond a single session can also be examined (i.e., so that performance does not reach an asymptote within a single session). Given the wide discrepancy between basic research (which typically deals with a single learning mechanism) and real-world contexts (which typically involve multiple learning mechanisms operating at different time scales), bridging this gap between basic research and clinical practice requires clear delineation of different stages of investigation, choosing tasks that are optimal for each stage, and triangulating the evidence across these stages. Third, pre-registration could help address the issue of fragmentation in how learning is quantified. Specifying and justifying in advance the type and calculation of the learning measure, sample size justification, outlier exclusion criteria, and accounting for baseline differences could provide more transparency into the choices of these methods. Finally, in cases where issues such as baseline differences exist, examining the sensitivity of the conclusions to the type of learning measures or the method of accounting for baseline differences would increase the robustness of the research.

Greater focus on individual-level data

In addition to theoretical insight, one goal of measuring motor learning deficits in neurological populations is that it can potentially serve as a non-invasive behavioral marker for neurological function and be used as an early predictor of disease states such as Alzheimer’s disease (Ruitenberg et al. 2023; Korte et al. 2024), Parkinsons’ disease (Gigi et al. 2024), Huntington’s disease (Ghilardi et al. 2008) and spinocerebellar ataxia (Bando et al. 2019). However, most of the studies reviewed tended to focus on group-level differences (e.g., by showing statistically significant differences using ANOVAs or t-tests) which can be misleading given the heterogeneity in many neurological conditions. In addition to more comprehensive baseline assessments, increasing transparency by including individual data in figures (Weissgerber et al. 2015) and incorporating appropriate statistical reporting for biomarkers (e.g., by using measures that are sensitive to base rates) (Ray et al. 2010) can provide information about whether a deficit can indeed be considered predictive of future disease states and be compared with existing alternatives. The emphasis on explaining individual data, coupled with the choice of specific model tasks, could also bolster computational models of learning (Scheidt et al. 2001; Izawa and Shadmehr 2011; Schweighofer et al. 2011; Schlerf et al. 2013; Therrien et al. 2016; Chu et al. 2016; Filoteo et al. 2017) that can provide more insight into ‘why’ performance of two groups are different, rather than simply stating ‘that’ they performed differently.

There are a couple of important limitations associated with the study. First, because we did not adopt an explicit definition of motor learning, the review included papers that do not have a common definition of motor learning. Although one advantage of casting a broad net is that it allowed us to capture and document the range of learning paradigms commonly used to assess deficits in clinical populations, a downside of including multiple learning paradigms is that it somewhat limited the specificity of the recommendations. As mentioned earlier, motor learning is a complex process spanning multiple phenomena (Krakauer et al. 2019; Tsay et al. 2024) over multiple timescales (Newell et al. 2001) and it is our view that no single definition of learning applies universally across all paradigms. For example, one definition of motor learning emphasizes the use of delayed retention tests (e.g., measuring performance after a 24 h delay) to assess relatively permanent changes in behavior (Schmidt and Lee 2005). However, it is important to note that while a single retention test can account for performance effects that dissipate over time (such as fatigue), it does not account for performance effects due to impairment or motor control deficits. Therefore, rather than using a ‘one-size-fits-all’ solution for all learning studies, future work that adapts our recommendations to address specific learning mechanisms and task contexts is critical. Second, from the viewpoint of translational impact, it is important to note that motor learning is only one component of neurorehabilitation, and that interventions have to be considered in the context of interaction between different levels of analysis of the ICF framework (body, function, activities, and participation) (Stucki et al. 2002). In other words, neurorehabilitation interventions can have different objectives such as reducing impairment, addressing activity limitations, or removing an environmental barrier (Lexell and Brogårdh 2015). Therefore, the presence of motor learning deficits in a clinical population should not be taken as directly indicative of the effectiveness of neurorehabilitation per se.

In summary, our review highlights that addressing motor learning deficits in the neurological population is a challenging research question. A central contribution of our paper is to document how variations in seemingly trivial choices in terms of researcher degrees of freedom (baseline correction, task selection, time duration of learning) across a range of motor learning paradigms result in ambiguous interpretations, which potentially explain why there is no consensus view on these issues despite several decades of work. We want to emphasize that there are no ‘quick fixes’ to the issues raised here; instead, the answer to this challenge requires the development of new research approaches with more representative tasks and new experimental designs. A more rigorous, convergent approach using multiple lines of evidence will eventually be necessary to create theoretical and clinical impact.

Acknowledgements

We thank Dr. Chandramouli Krishnan for his comments on a draft of this manuscript.

Author contributions

R.R. conceptualized the paper and wrote the first draft of the manuscript. S.C., A.M., M.T., R.V., and R.R. analyzed the papers and extracted data. R.V. and R.R. prepared the figures. S.C., A.M., M.T., and R.V. revised the manuscript. All authors reviewed the manuscript.

Funding

This material is based upon work supported by the National Science Foundation under Grant No. 1823889.

Data availability

No datasets were generated or analysed during the current study.

Declarations

Conflict of interest

The authors declare no competing interests.

Footnotes

Deceased: Sharo Costa.

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Sharo Costa, Angelyn May Mengote, Megan Thomas and Rohan Ved contributed equally to this work.

References

  1. Agostino R, Sanes JN, Hallett M (1996) Motor skill learning in Parkinson’s disease. J Neurol Sci 139:218–226. 10.1016/0022-510X(96)00060-3 [PubMed] [Google Scholar]
  2. Ashe J, Lungu OV, Basford AT, Lu X (2006) Cortical control of motor sequences. Curr Opin Neurobiol 16:213–221. 10.1016/j.conb.2006.03.008 [DOI] [PubMed] [Google Scholar]
  3. Bahrick HP, Fitts PM, Briggs GE (1957) Learning curves: facts or artifacts? Psychol Bull 54:256–268. 10.1037/h0040313 [Google Scholar]
  4. Bando K, Honda T, Ishikawa K et al (2019) Impaired adaptive motor learning is correlated with cerebellar hemispheric gray matter atrophy in spinocerebellar ataxia patients: a voxel-based morphometry study. Front Neurol. 10.3389/fneur.2019.01183 [Google Scholar]
  5. Bönstrup M, Iturrate I, Hebart MN et al (2020) Mechanisms of offline motor learning at a microscale of seconds in large-scale crowdsourced data. NPJ Sci Learn 5:7. 10.1038/s41539-020-0066-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Boyd LA, Winstein CJ (2004) Cerebellar stroke impairs Temporal but not Spatial accuracy during implicit motor learning. Neurorehabil Neural Repair 18:134–143. 10.1177/0888439004269072 [DOI] [PubMed] [Google Scholar]
  7. Button KS, Ioannidis JPA, Mokrysz C et al (2013) Power failure: why small sample size undermines the reliability of neuroscience. Nat Rev Neurosci 14:365–376. 10.1038/nrn3475 [DOI] [PubMed] [Google Scholar]
  8. Chu VW, Park S-W, Sanger TD, Sternad D (2016) Children with dystonia can learn a novel motor skill: strategies that are tolerant to high variability. IEEE Trans Neural Syst Rehabil Eng 24:847–858. 10.1109/TNSRE.2016.2521404 [DOI] [PubMed] [Google Scholar]
  9. Corkin S (1968) Acquisition of motor skill after bilateral medial temporal-lobe excision. Neuropsychologia 6:255–265. 10.1016/0028-3932(68)90024-9 [Google Scholar]
  10. Criscimagna-Hemminger SE, Bastian AJ, Shadmehr R (2010) Size of error affects cerebellar contributions to motor learning. J Neurophysiol 103:2275–2284. 10.1152/jn.00822.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Cubillos LH, Augenstein TE, Ranganathan R, Krishnan C (2023) Breaking the barriers to designing online experiments: a novel open-source platform for supporting procedural skill learning experiments. Comput Biol Med 154:106627. 10.1016/j.compbiomed.2023.106627 [DOI] [PubMed] [Google Scholar]
  12. Doyon J, Benali H (2005) Reorganization and plasticity in the adult brain during learning of motor skills. Curr Opin Neurobiol 15:161–167. 10.1016/j.conb.2005.03.004 [DOI] [PubMed] [Google Scholar]
  13. Ferraro FR, Balota DA, Connor LT (1993) Implicit memory and the formation of new associations in nondemented Parkinson′s disease individuals and individuals with senile dementia of the Alzheimer type: a serial reaction time (SRT) investigation. Brain Cogn 21:163–180. 10.1006/brcg.1993.1013 [DOI] [PubMed] [Google Scholar]
  14. Filoteo JV, Maddox WT, Ashby FG (2017) Quantitative modeling of category learning deficits in various patient populations. Neuropsychology 31:862–876. 10.1037/neu0000422 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Gershon RC, Wagster MV, Hendrie HC et al (2013) NIH toolbox for assessment of neurological and behavioral function. Neurology 80:S2–S6. 10.1212/WNL.0b013e3182872e5f [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Ghilardi MF, Silvestri G, Feigin A et al (2008) Implicit and explicit aspects of sequence learning in pre-symptomatic Huntington’s disease. Parkinsonism Relat Disord 14:457–464. 10.1016/j.parkreldis.2007.11.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Gigi I, Senatore R, Marcelli A (2024) The onset of motor learning impairments in Parkinson’s disease: a computational investigation. Brain Inf 11:4. 10.1186/s40708-023-00215-6 [Google Scholar]
  18. Glickstein M, Yeo C (1990) The cerebellum and motor learning. J Cogn Neurosci 2:69–80. 10.1162/jocn.1990.2.2.69 [DOI] [PubMed] [Google Scholar]
  19. Grafman J, Weingartner H, Newhouse PA et al (1990) Implicit learning in patients with Alzheimer’s disease. Pharmacopsychiatry 23:94–101. 10.1055/s-2007-1014490 [DOI] [PubMed] [Google Scholar]
  20. Harrington DL, Haaland KY, Yeo RA, Marder E (1990) Procedural memory in parkinson’s disease: impaired motor but not visuoperceptual learning. J Clin Exp Neuropsychol 12:323–339. 10.1080/01688639008400978 [DOI] [PubMed] [Google Scholar]
  21. Heilman KM, Schwartz HD, Geschwind N (1975) Defective motor learning in ideomotor apraxia. Neurology 25:1018–1020. 10.1212/wnl.25.11.1018 [DOI] [PubMed] [Google Scholar]
  22. Ingram LA, Butler AA, Walsh LD et al (2019) The upper limb physiological profile assessment: description, reliability, normative values and criterion validity. PLoS ONE 14:e0218553. 10.1371/journal.pone.0218553 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Ioannidis JPA (2008) Why most discovered true associations are inflated. Epidemiology 19:640–648. 10.1097/EDE.0b013e31818131e7 [Google Scholar]
  24. Izawa J, Shadmehr R (2011) Learning from sensory and reward prediction errors during motor adaptation. PLoS Comput Biol 7:e1002012. 10.1371/journal.pcbi.1002012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Kal E, Winters M, van der Kamp J et al (2016) Is implicit motor learning preserved after stroke? A systematic review with meta-analysis. PLoS ONE 11:e0166376. 10.1371/journal.pone.0166376 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Kantak SS, Winstein CJ (2012) Learning-performance distinction and memory processes for motor skills: a focused review and perspective. Behav Brain Res 228:219–231. 10.1016/j.bbr.2011.11.028 [DOI] [PubMed] [Google Scholar]
  27. Kernan WN, Viscoli CM, Makuch RW et al (1999) Stratified randomization for clinical trials. J Clin Epidemiol 52:19–26. 10.1016/s0895-4356(98)00138-3 [DOI] [PubMed] [Google Scholar]
  28. Kitago T, Krakauer JW (2010) Losing control: brain vs spinal cord. Neurology 74:1250–1251. 10.1212/WNL.0b013e3181dd4e0f [DOI] [PubMed] [Google Scholar]
  29. Korte JA, Weakley A, Donjuan Fernandez K et al (2024) Neural underpinnings of learning in dementia populations: a review of motor learning studies combined with neuroimaging. J Cogn Neurosci 36:734–755. 10.1162/jocn_a_02116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Krakauer JW, Hadjiosif AM, Xu J et al (2019) Motor learning. Compr Physiol 9:613–663. 10.1002/cphy.c170043 [DOI] [PubMed] [Google Scholar]
  31. Krebs H, Hogan N, Hening W et al (2001) Procedural motor learning in Parkinson’s disease. Exp Brain Res 141:425–437. 10.1007/s002210100871 [DOI] [PubMed] [Google Scholar]
  32. Krishnan C, Augenstein T, Ranganathan R (2025) Ten recommendations for improving research on stroke motor rehabilitation: a unified perspective. Restor Neurol Neurosci 10.1177/09226028251392269
  33. Lakens D, Caldwell AR (2021) Simulation-Based power analysis for factorial analysis of variance designs. Adv Methods Pract Psychol Sci 4:2515245920951503. 10.1177/2515245920951503 [Google Scholar]
  34. Lexell J, Brogårdh C (2015) The use of ICF in the neurorehabilitation process. NeuroRehabilitation 36:5–9. 10.3233/NRE-141184 [DOI] [PubMed] [Google Scholar]
  35. Liu Y-T, Mayer-Kress G, Newell KM (2006) Qualitative and quantitative change in the dynamics of motor learning. J Exp Psychol Hum Percept Perform 32:380–393. 10.1037/0096-1523.32.2.380 [DOI] [PubMed] [Google Scholar]
  36. Lohse KR, Buchanan T, Miller M (2016) Underpowered and overworked: problems with data analysis in motor learning studies. J Mot Learn Dev 4:37–58. 10.1123/jmld.2015-0010 [Google Scholar]
  37. Lohse KR, Shen J, Kozlowski AJ (2020) Modeling longitudinal outcomes: a contrast of two methods. J Mot Learn Dev 8:145–165. 10.1123/jmld.2019-0007 [Google Scholar]
  38. Lord FM (1967) A paradox in the interpretation of group comparisons. Psychol Bull 68:304–305. 10.1037/h0025105 [DOI] [PubMed] [Google Scholar]
  39. Lord SR, Menz HB, Tiedemann A (2003) A physiological profile approach to falls risk assessment and prevention. Phys Ther 83:237–252 [PubMed] [Google Scholar]
  40. Mandava P, Kalkonde YV, Rochat RH, Kent TA (2010) A matching algorithm to address imbalances in study populations: application to the National Institute of neurological diseases and stroke recombinant tissue plasminogen activator acute stroke trial. Stroke 41:765–770. 10.1161/STROKEAHA.109.574103 [DOI] [PubMed] [Google Scholar]
  41. Marinelli L, Trompetto C, Canneva S et al (2017) Learning how to learn: super declarative motor learning is impaired in Parkinson’s disease. Neural Plast 2017:1–8. 10.1155/2017/3162087 [Google Scholar]
  42. McDougle SD, Bond KM, Taylor JA (2015) Explicit and implicit processes constitute the fast and slow processes of sensorimotor learning. J Neurosci 35:9568–9579. 10.1523/JNEUROSCI.5061-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Milner B (1965) Visually-guided maze learning in man: effects of bilateral hippocampal, bilateral frontal, and unilateral cerebral lesions. Neuropsychologia 3:317–338. 10.1016/0028-3932(65)90005-9 [Google Scholar]
  44. Mostofsky SH, Goldberg MC, Landa RJ, Denckla MB (2000) Evidence for a deficit in procedural learning in children and adolescents with autism: implications for cerebellar contribution. J Int Neuropsychol Soc JINS 6:752–759. 10.1017/s1355617700677020 [DOI] [PubMed] [Google Scholar]
  45. Newell KM, Liu YT, Mayer-Kress G (2001) Time scales in motor learning and development. Psychol Rev 108:57–82. 10.1037/0033-295x.108.1.57 [DOI] [PubMed] [Google Scholar]
  46. Newell KM, Mayer-Kress G, Hong SL, Liu Y-T (2009) Adaptation and learning: characteristic time scales of performance dynamics. Hum Mov Sci 28:655–687. 10.1016/j.humov.2009.07.001 [DOI] [PubMed] [Google Scholar]
  47. Nicolson RI, Fawcett AJ (1994) Comparison of deficits in cognitive and motor skills among children with dyslexia. Ann Dyslexia 44:147–164. 10.1007/BF02648159 [DOI] [PubMed] [Google Scholar]
  48. Nieuwboer A, Rochester L, Müncks L, Swinnen SP (2009) Motor learning in Parkinson’s disease: limitations and potential for rehabilitation. Parkinsonism Relat Disord 15:S53–S58. 10.1016/S1353-8020(09)70781-3 [DOI] [PubMed] [Google Scholar]
  49. Olivier GN, Walter CS, Paul SS et al (2021) How common is the exponential decay pattern of motor skill acquisition? A brief investigation. Motor Control 25:451–461. 10.1123/mc.2020-0043 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Olson M, Lockhart TE, Lieberman A (2019) Motor learning deficits in Parkinson’s disease (PD) and their effect on training response in gait and balance: a narrative review. Front Neurol. 10.3389/fneur.2019.00062 [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Pendt LK, Reuter I, Müller H (2011) Motor skill learning, retention, and control deficits in parkinson’s disease. PLoS ONE 6:e21669. 10.1371/journal.pone.0021669 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Penhune VB, Steele CJ (2012) Parallel contributions of cerebellar, striatal and M1 mechanisms to motor sequence learning. Behav Brain Res 226:579–591. 10.1016/j.bbr.2011.09.044 [DOI] [PubMed] [Google Scholar]
  53. Ranganathan R, Tomlinson AD, Lokesh R et al (2021) A tale of too many tasks: task fragmentation in motor learning and a call for model task paradigms. Exp Brain Res 239:1–19. 10.1007/s00221-020-05908-6 [DOI] [PubMed] [Google Scholar]
  54. Ranganathan R, Cone S, Fox B (2022a) Predicting individual differences in motor learning: a critical review. Neurosci Biobehav Rev 141:104852. 10.1016/j.neubiorev.2022.104852 [DOI] [PubMed] [Google Scholar]
  55. Ranganathan R, Doherty C, Gussert M et al (2022b) Scientific basis and active ingredients of current therapeutic interventions for stroke rehabilitation. Restor Neurol Neurosci 40:97–107. 10.3233/RNN-211243 [DOI] [PubMed] [Google Scholar]
  56. Ranganathan R, Lee M-H, Krishnan C (2022c) Ten guidelines for designing motor learning studies. Braz J Mot Behav 16:112–133. 10.20338/bjmb.v16i2.283 [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Ray P, Manach YL, Riou B et al (2010) Statistical evaluation of a biomarker. Anesthesiology 112:1023–1040. 10.1097/ALN.0b013e3181d47604 [DOI] [PubMed] [Google Scholar]
  58. Robertson EM (2007) The serial reaction time task: implicit motor skill learning? J Neurosci 27:10073–10075. 10.1523/JNEUROSCI.2747-07.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Rouleau I, Salmon DP, Vrbancic M (2002) Learning, retention and generalization of a mirror tracing skill in Alzheimer’s disease. J Clin Exp Neuropsychol 24:239–250. 10.1076/jcen.24.2.239.997 [DOI] [PubMed] [Google Scholar]
  60. Ruitenberg MFL, Duff K, Koppelmans V (2023) Motor sequence learning in mild cognitive impairment and Alzheimer’s disease and its relationship to biomarkers. Alzheimers Dement 19:e075603. 10.1002/alz.075603 [Google Scholar]
  61. Salmoni AW, Schmidt RA, Walter CB (1984) Knowledge of results and motor learning: a review and critical reappraisal. Psychol Bull 95:355–386. 10.1037/0033-2909.95.3.355 [PubMed] [Google Scholar]
  62. Sanes JN, Dimitrov B, Hallett M (1990) Motor learning in patients with cerebellar dysfunction. Brain 113(Pt 1):103–120. 10.1093/brain/113.1.103 [DOI] [PubMed] [Google Scholar]
  63. Scheidt RA, Dingwell JB, Mussa-Ivaldi FA (2001) Learning to move amid uncertainty. J Neurophysiol 86:971–985. 10.1152/jn.2001.86.2.971 [DOI] [PubMed] [Google Scholar]
  64. Schlerf JE, Xu J, Klemfuss NM et al (2013) Individuals with cerebellar degeneration show similar adaptation deficits with large and small visuomotor errors. J Neurophysiol 109:1164–1173. 10.1152/jn.00654.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Schmidt RA (1972) The case against learning and forgetting scores. J Mot Behav 4:79–88. 10.1080/00222895.1972.10734922 [DOI] [PubMed] [Google Scholar]
  66. Schmidt RA, Bjork RA (1992) New conceptualizations of practice: common principles in three paradigms suggest new concepts for training. Psychol Sci 3:207–218. 10.1111/j.1467-9280.1992.tb00029.x [Google Scholar]
  67. Schmidt RA, Lee TD (2005) Motor control and learning: a behavioral emphasis, 4th edn. Human Kinetics, Champaign [Google Scholar]
  68. Schweighofer N, Lee J-Y, Goh H-T et al (2011) Mechanisms of the contextual interference effect in individuals poststroke. J Neurophysiol 106:2632–2641. 10.1152/jn.00399.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Seidler RD (2010) Neural correlates of motor learning, transfer of learning, and learning to learn. Exerc Sport Sci Rev 38:3–9. 10.1097/JES.0b013e3181c5cce7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Senn S (2006) Change from baseline and analysis of covariance revisited. Stat Med 25:4334–4344. 10.1002/sim.2682 [DOI] [PubMed] [Google Scholar]
  71. Spampinato D, Celnik P (2021) Multiple motor learning processes in humans: defining their neurophysiological bases. Neuroscientist 27:246–267. 10.1177/1073858420939552 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Stoodley CJ, Stein JF (2006) A processing speed deficit in dyslexic adults? Evidence from a peg-moving task. Neurosci Lett 399:264–267. 10.1016/j.neulet.2006.02.004 [DOI] [PubMed] [Google Scholar]
  73. Stucki G, Ewert T, Cieza A (2002) Value and application of the ICF in rehabilitation medicine. Disabil Rehabil 24:932–938. 10.1080/09638280210148594 [DOI] [PubMed] [Google Scholar]
  74. Subramanian SK, Feldman AG, Levin MF (2018) Spasticity may obscure motor learning ability after stroke. J Neurophysiol 119:5–20. 10.1152/jn.00362.2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Taylor JA, Ivry RB (2013) Implicit and explicit processes in motor learning. In: Prinz W, Beisert M, Herwig A (eds) Action science: foundations of an emerging discipline. The MIT Press, Cambridge, pp 63–87 [Google Scholar]
  76. Tennant PWG, Arnold KF, Ellison GTH, Gilthorpe MS (2022) Analyses of ‘change scores’ do not estimate causal effects in observational data. Int J Epidemiol 51:1604–1615. 10.1093/ije/dyab050 [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Therrien AS, Wolpert DM, Bastian AJ (2016) Effective reinforcement learning following cerebellar damage requires a balance between exploration and motor noise. Brain J Neurol 139:101–114. 10.1093/brain/awv329 [Google Scholar]
  78. Thompson AK, Chen XY, Wolpaw JR (2009) Acquisition of a simple motor skill: task-dependent adaptation plus long-term change in the human soleus H-reflex. J Neurosci 29:5784–5792. 10.1523/JNEUROSCI.4326-08.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Tsay JS, Kim HE, McDougle SD et al (2024) Fundamental processes in sensorimotor learning: reasoning, refinement, and retrieval. eLife 13:e91839. 10.7554/eLife.91839 [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. van Halteren-van Tilborg IADA, Scherder EJA, Hulstijn W (2007) Motor-skill learning in Alzheimer’s disease: a review with an eye to the clinical practice. Neuropsychol Rev 17:203–212. 10.1007/s11065-007-9030-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Weinberger AB, Green AE (2022) Dynamic development of intuitions and explicit knowledge during implicit learning. Cognition 222:105008. 10.1016/j.cognition.2021.105008 [DOI] [PubMed] [Google Scholar]
  82. Weissgerber TL, Milic NM, Winham SJ, Garovic VD (2015) Beyond bar and line graphs: time for a new data presentation paradigm. PLoS Biol 13:e1002128. 10.1371/journal.pbio.1002128 [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Willingham DB (1998) A neuropsychological theory of motor skill learning. Psychol Rev 105:558–584. 10.1037/0033-295X.105.3.558 [DOI] [PubMed] [Google Scholar]
  84. Willingham DB, Peterson EW, Manning C, Brashear HR (1997) Patients with Alzheimer’s disease who cannot perform some motor skills show normal learning of other motor skills. Neuropsychology 11:261–271. 10.1037/0894-4105.11.2.261 [DOI] [PubMed] [Google Scholar]
  85. Winstein CJ, Merians AS, Sullivan KJ (1999) Motor learning after unilateral brain damage. Neuropsychologia 37:975–987. 10.1016/s0028-3932(98)00145-6 [DOI] [PubMed] [Google Scholar]
  86. Zhang Y, Jayaswal S, Schweighofer N (2025) Two ways to learn in visuomotor adaptation. J Neurophysiol 134:1085–1096. 10.1152/jn.00046.2025 [DOI] [PubMed] [Google Scholar]
  87. Ziv G, Lidor R, Netz Y (2020) Dealing with possible baseline inequalities between experimental groups—The case of motor learning. J Mot Behav 52:502–513. 10.1080/00222895.2019.1649996 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

No datasets were generated or analysed during the current study.


Articles from Experimental Brain Research are provided here courtesy of Springer

RESOURCES