Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2016 Mar 1.
Published in final edited form as: Trends Cogn Sci. 2015 Jan 24;19(3):117–125. doi: 10.1016/j.tics.2014.12.010

Domain generality vs. modality specificity: The paradox of statistical learning

Ram Frost 1,2,3, Blair C Armstrong 3, Noam Siegelman 1, Morten H Christiansen 2,4,5
PMCID: PMC4348214  NIHMSID: NIHMS652678  PMID: 25631249

Abstract

Statistical learning is typically considered to be a domain-general mechanism by which cognitive systems discover the underlying distributional properties of the input. Recent studies examining whether there are commonalities in the learning of distributional information across different domains or modalities consistently reveal, however, modality and stimulus specificity. An important question is, therefore, how and why a hypothesized domain-general learning mechanism systematically produces such effects. We offer a theoretical framework according to which statistical learning is not a unitary mechanism, but a set of domain-general computational principles, that operate in different modalities and therefore are subject to the specific constraints characteristic of their respective brain regions. This framework offers testable predictions and we discuss its computational and neurobiological plausibility.

Keywords: Statistical learning, domain-general mechanisms, modality specificity, stimulus specificity, neurobiologically plausible models

The promise of statistical learning

Humans and other animals are constantly bombarded by streams of sensory information. Statistical learning (SL)—the extraction of distributional properties from sensory input across time and space—provides a mechanism by which cognitive systems discover the underlying structure of such stimulation. SL therefore plays a key role in the detection of regularities and quasi-regularities in the environment, results in discrimination, categorization and segmentation of continuous information, allows prediction of upcoming events, and thereby shapes the basic representations underlying a wide range of sensory, motor, and cognitive abilities.

In cognitive science, theories of SL have emerged as potential domain-general alternatives to the influential domain-specific Chomskyan account of language acquisition ([1], see also [2] for related claims). Rather than assuming an innate, modular, and neurobiologically hardwired human capacity for processing linguistic information, SL, as a theoretical construct, was offered as a general mechanism for learning and processing any type of sensory input that unfolds across time and space. To date, evidence for SL have been found across an array of cognitive functions, such as segmenting continuous auditory input [3], visual search [4], contextual cuing [5], visuomotor learning [6], conditioning (e.g., [7]), and in general, any predictive behavior (e.g., [8,9]).

In this paper, we propose a broad theoretical account of SL, starting with a discussion of how a domain-general ability may be subject to modality- (see glossary) and stimulus-specific constraints. We define ‘learning’ as the process responsible for updating internal representations given specific input and encoding potential relationships between them, thereby improving the processing of that input. Similarly, ‘processing’ is construed as determining how an input to a neural system interacts with the current knowledge stored in that system to generate internal representations. Knowledge in the system is thus continuously updated via learning. Specifically, we take SL to reflect updates based on the discovery of systematic regularities embedded in the input, and provide a mechanistic account of how distributional properties are picked up across domains, eventually shaping behavior. We further outline how this account is constrained by neuroanatomy and systems neuroscience, offering independent insights into the specific constraints on SL. Finally, we highlight individual differences in abilities for SL as a major, largely untapped source of evidence for which our account makes clear predictions.

Domain generality versus domain specificity

Originally, domain generality was invoked to argue against language modularity; its definition therefore implicitly implied “something that is not language specific”. Consequently, within this context, “domain” implies a range of stimuli that share physical and structural properties (e.g., spoken words, musical tones, tactile input), whereas “generality” is taken to be “something that does not operate along principles restricted to language learning”. Note, however, that this approach says what domain generality is not, rather than saying what it is (e.g., [10]). More recent accounts of SL ascribe domain generality to a unitary learning system (e.g., [11]), that executes similar computations across stimuli (e.g., [12]), and that can be observed across domains (e.g., [13]), and across species (e.g., [14,15]).

As a theoretical construct, SL promised to bring together a wide range of cognitive functions within a single mechanism. Extensive research over the last decade has therefore focused on mapping the commonalities involved in the learning of distributional information across different domains. From an operational perspective, these studies investigated whether overall performance in SL tasks is indeed similar across different types of stimuli [16], whether there is transfer of learning across domains (see Box 1), whether there is interference between simultaneously learning of multiple artificial grammars (e.g., [17]) or from multiple input streams within and across domains [18], or whether individual capacities in detecting distributional probabilities in a variety of SL tasks are correlated ([19]).

The pattern of results across these different studies is intriguingly consistent: contrary to the most intuitive predictions of domain-generality, the evidence persistently shows patterns of modality specificity and sometimes even stimulus specificity. For example, studies of artificial grammar learning (AGL, see Glossary) systematically demonstrate very limited transfer of learning across modalities, if at all (e.g., [20,21]). Similarly, the simultaneous learning of two artificial grammars can proceed without interference once they are implemented in separate modalities [17]. Modality specificity is also revealed by qualitative differences in patterns of SL in the auditory, visual, and tactile modalities [16], sometimes with opposite effects of presentation parameters across modalities [22]. To complicate matters even further, SL within modality reveals striking stimulus specificity, so that no transfer (and conversely, no interference) occurs within modality provided the stimuli have separable perceptual features (e.g., [17,23]). Finally, although performance in SL tasks displays substantial test-retest reliability within modality, there is no evidence of any correlation within individuals in their capacities to detect conditional probabilities across modalities and across stimuli (Siegelman & Frost, unpublished). This contrasts with what might be expected if SL was subserved by a unitary learning system: that individual variation in its basic function would manifest itself in at least some degree of correlation across different SL tasks. If not, its unitary aspect remains theoretically empty because it does not have an empirical reality in terms of specific testable predictions. Taken together, these studies suggest that there are independent modality constraints in learning distributional information [16], pointing to modality specificity, and further to stimulus specificity akin to perceptual learning [24].

Whereas this set of findings is not easy to reconcile with the notion of a unitary, domain-general system for SL, it does not necessarily invalidate the promise of SL to provide an overarching framework underlying learning across domains. Instead, what is needed is an account of SL that can explicate the manifestations of domain-generality in distributional learning with the evidence of its modality- and stimulus-specificity, restricted generalization, little transfer, and very low correlations of performance between tasks within individuals. More broadly, any general theory of learning that aims to describe a wide range of phenomena through a specific set of computational principles has to offer a theoretical account of how and why transfer, discrimination, and generalization take place, or not.

Towards a mechanistic model of SL

Our approach construes SL as involving a set of domain-general neurobiological mechanisms for learning, representation, and processing that detect and encode a wide range of distributional properties within different modalities or types of input (see [13], for a related approach). Crucially, though, in our account, these principles are not instantiated by a unitary learning system but, rather, by separate neural networks in different cortical areas (e.g., visual, auditory, and somatosensory cortex). Thus, the process of encoding an internal representation follows constraints that are determined by the specific properties of the input processed in the respective cortices. As a result, the outcomes of computations in these networks are necessarily modality specific, despite multiple cortical and subcortical regions invoking similar sets of computational principles and some shared brain regions (e.g., Hebbian learning, reinforcement learning; for discussion, see [25,26]).

For example, the auditory cortex displays lower sensitivity to spatial information but enhanced sensitivity to temporal information, whereas the visual cortex displays enhanced sensitivity to spatial information, but lower sensitivity to temporal information (e.g., [27,28]). Iconic memory is short-lived (scale of ms), whereas echoic memory lasts significantly longer (scale of seconds; e.g., [29]). Because auditory information unfolds in time, the auditory cortex must be sensitive to the temporal accumulation of information in order to make sense of the input. In contrast, visual information is instantaneous, and although temporal integration is necessary in some cases such as in deciphering motion, the visual cortex is relatively less sensitive to temporal accumulation of information over extended periods of time. These inherent differences are reflected in the way the sensory input eventually is encoded into internal representations for further computation. Moreover, within modality, encoding of events displays graded stimulus specificity given their complexity, similarity, saliency, and other factors related to the quality and nature of the input (see [30,31], for evidence in visual SL). For example, participants are able to learn two separate artificial grammars simultaneously in the visual domain when the stimuli are from separate perceptual dimensions—such as color and shape—but not when they are from within the same perceptual dimension [16]. Figure 1 represents a schematic account of our approach and shows how the same learning and representation principles result in modality and stimulus specificity because they are instantiated in different brain regions, each with their characteristic constraints.

Figure 1. Theoretical Model of Statistical Learning.

Figure 1

Schematic representation of the processing of distributional information in the visual, auditory, and somatosensory cortex, for unimodal and multimodal events. Different encoded representations of continuous input presented in time or space result in task-stimulus specificity, in spite of similar computations and contributions from partially shared neurocomputational networks.

Note that modality-specific constraints do not preclude the neurobiological ability to process multimodal events. Indeed, this has recently been shown within SL using the McGurk effect (see Glossary) in a cross-modal segmentation study [32]. More generally, perception of the world routinely involves multisensory integration (e.g., [33]), occurring at both low levels (i.e., the thalamus, [34]; the dorsal cochlear nucleus, [35]) and higher levels of cortical processing (e.g., anterior temporal poles; [36]). Critically, however, each of these multimodal areas would be subject to its own distinct set of constraints, which would not necessarily be the same as those from the unimodal regions that feed into it or to the constraints in other multimodal areas. For example, coherence in the timing at which an auditory and a visual stimulus unfold is important for specific types of integration [18] in audio-visual brain areas [37], but not as important for detecting regularities in the case of integrating two different visual representations in the visual system. Note that this view is distinct from alternative accounts suggesting that a unitary learning mechanism operates on “abstract” amodal representations (e.g., [38]; see Glossary). Instead, we suggest that multimodal regions are shaped by their own distinct sets of constraints.

This brings us to an operational definition of ‘domain generality’. Within our framework, domain generality primarily emerges because neural networks across modalities instantiate similar computational principles. Moreover, domain generality may also arise either through the possible engagement of partially-shared neural networks that modulate the encoding of the to-be-learned statistical structure [39], or if stimulus input representations encoded in a given modality (e.g., visual or auditory) are fed into a multi-modal region for further computation and learning. As we shall see next, the current neurobiological evidence is consistent with both of these latter possibilities.

The neurobiological bases of SL

Recent neuroimaging studies have shown that statistical regularities of visual shapes results in activation in higher-level visual networks (e.g., lateral occipital cortex, inferior temporal gyrus; [40,41]), whereas statistical regularities in auditory stimuli result in activation in analogous auditory networks (e.g., left temporal and inferior parietal cortices; frontotemporal networks including portions of the inferior frontal gyrus, motor areas involved in speech production, [42]; and the pars opercularis and pars triangularis regions of the left inferior frontal gyrus; [43]). Since these studies contrasted activation for structured vs. random blocks of stimuli, the stronger activation for structured stimuli in the above ROIs is consistent with the notion that some SL occurs already in brain regions that are largely dedicated to processing unimodal stimuli, thus allowing for modality-specific constraints to shape the outcome of computations.

In addition to identifying modality-specific learning mechanisms, imaging and ERP studies point to some brain regions that are active regardless of the modality in which the stimulus is presented. Often, this work has associated SL effects with the hippocampus, and more generally with the medial temporal lobe (MTL) memory system (see, e.g., [44]). This is consistent with considerable systems neuroscience work that has established the hippocampus as a locus for encoding and binding temporal and spatial contingencies presented in multiple different modalities [40,4448], as well as for consolidation of representations.

Hippocampal involvement in SL could consist of indirect modulation of the representations in sensory areas or direct computations on hippocampal representations that are driven by sensorimotor representations (see [48] for a discussion). Note, however, that even in the case of direct hippocampal computations, the computed representations are not necessarily amodal, as traces of their original specificity nevertheless remain (e.g., [49]). Sub-regions of the hippocampus have been shown to send and receive different types of information from different brain regions, while developing specialization for representing those different types of information [50]. In addition, representations within the hippocampus itself are typically sparse, and are wired to be maximally dissimilar even when stimuli evoke similar activation in a given sensorimotor region [5154]. Thus, even with a direct hippocampal involvement in SL, such computations would likely result in a high degree of stimulus specificity, as observed across many SL studies.

Additional imaging work has identified regions of the basal ganglia [55] and thalamus [42,56] as important collaborating brain regions that work with the MTL memory system to complete relevant sub-tasks involved in statistical learning. For instance, the thalamus may provide synchronizing oscillatory activity in the alpha-gamma and theta-gamma ranges that enables the rapid and accurate encoding of sequences of events [56]. Thus, as summarized in Figure 2, the current neurobiological evidence indeed suggests that detection of statistical regularities emerges from local computations carried out within a given modality, and through a multi-domain neurocognitive system that either modulates or operates on inputs from modality-specific representations. Whether unimodal computations are necessary or sufficient for SL, remains an open question. Whereas some studies report no learning following hippocampal damage [44], other report significant SL in spite of such damage (e.g., [57]). In this context we should note, that lack of SL cannot be unequivocally attributed to neurobiological impairment. Many normal participants do not show SL even with an intact MTL system (see, for example, performance of a subset of the control participants observed by [44], who do not fare better than the specific reported patient). This leads us to our next section on individual differences.

Figure 2. Key Neural Networks involved in Visual and Auditory Statistical Learning.

Figure 2

Key brain regions associated with domain-general (blue), and lower- and higher-level auditory (green) and visual (red) modality-specific processing and representation, plotted on a smoothed ICBM152 template brain. The depicted regions are not intended to constitute an exhaustive set of brain regions subserving each domain. C = Cuneus, FG = Fusiform Gyrus, STG = Superior Temporal Gyrus, IPL = Inferior Parietal Lobule, H = Hippocampus, T = Thalamus, CA = Caudate, IFG = Inferior Frontal Gyrus. Generated with the BrainNet Viewer [89].

Individual and group differences in SL

The proposed framework leads us to argue that individual differences provide key evidence for understanding the mechanism of SL. In past work, it has often been assumed that individual variance in implicit learning tasks is significantly smaller than that of explicit learning (e.g., [58]). Consequently, the source of variability in performance in SL has been largely overlooked, and had led researchers to focus on average success rate (but see [19,5961]).

In the context of SL, however, measures of central tendency can be particularly misleading, as often about one third of the sample or more is not performing the task above chance level (e.g., [12,60,61]). Moreover, tracking individual learning trajectories throughout the phases of a given SL task has recently suggested that there is a commensurate high level of variability in the learning curves of different individuals (e.g., [43,61]). In several areas of cognitive science, it is now well established that understanding the source of individual differences holds the promise of revealing critical insight regarding the cognitive operations underlying performance, leading to more refined theories of behaviour. Furthermore, a theory that addresses individual differences should aim to explain how learning mechanisms operate online to gradually extract statistical structure, as opposed to focusing strictly on the outcome of a learning phase in a subsequent test (e.g., [62]).

As a first approximation, our theoretical model splits the variance across individuals into two main sources. First, as indicated by Figure 1, there is the variance related to efficiency in encoding representations within modality in the visual, auditory, and somatosensory cortex. This variance could derive from individual differences in the efficacy of encoding fast sequential inputs or complex spatial stimuli, and thus potentially could be traced to the neuronal mechanisms that determine the effective resolution of one's sensory system. The second variance relates to the relative computational efficiency of processing multiple temporally and spatially encoded representations and detecting their distributional properties. This variance potentially could be traced to cellular- and systems-level differences in factors that include (but are not limited to) white matter density, which have been shown to affect AGL performance [63], and variation in the speed of changes in synaptic efficacy [64]. In modeling terms, these factors would relate to parameters such as connectivity, learning rates, and the quality and type of information to be encoded and transmitted by a given brain region (see Box 2).

The advantage of this approach is that it offers precise and testable predictions that can be empirically evaluated. Thus, individuals can display relatively increased sensitivity in encoding auditory information, but a relative disadvantage in encoding sequential visual information. Conversely, two individuals that have similar efficiency in terms of representational encoding in a given modality could differ in their relative efficiency in computing the distributional properties of visual or auditory events. In either case, low correlation in performance within individuals in two SL tasks, would be the outcome, as has been reported in recent studies (e.g., [19]). However, as exemplified in Box 3, accurate individual trajectories of SL can in principle be obtained by employing parametric designs that independently target the two sources of variance.

Individual differences are particularly intriguing given recent claims regarding developmental invariance in some types of SL (e.g., [65]). If SL capacities per se do not change, and brain maturation and experience are primarily driving improvements in processes “peripheral” to SL such as attention, then the bulk of variability in individual developmental trajectories in SL abilities should be explained by these peripheral factors only. We believe that the current empirical support for this claim is limited (see [66] for a discussion). Further progress, however, requires a better fundamental understanding of individual differences in SL, as elaborated in Box 3.

Summary and conclusions

The present paper offers a novel theoretical perspective on SL that considers computational and neurobiological constraints. Previous work on SL offered a possible cognitive mechanistic account of how distributional properties are computed, with explicit demonstrations being provided only within the domain of language [65,67]. The perspective we offer has the advantage of providing a unifying neurobiological account of SL across domains, modalities, neural and cognitive investigations, and cross-species studies, thus connecting with and explaining an extensive set of data. The core claim of our framework is that SL reflects contributions from domain-general learning principles that are constrained to operate in specific modalities, with potential contributions from partially shared brain regions common to learning in different modalities. Both of these notions are well grounded in neuroscience. Moreover, they provide our account with the flexibility needed to explain the apparently contradictory SL phenomena observed both within and between individuals, such as stimulus and modality specificity, while still being constrained by the capacities of the brain regions that subserve the processing of different types of stimuli. In addition to descriptive adequacy, our approach also provides targeted guidance for future investigations of SL via explicit neurobiological modeling and studies of the mechanics underlying individual differences. We therefore offer our framework as a novel platform for understanding and advancing the study of SL and related phenomena.

BOX 1: Generalization and transfer in statistical learning.

A key aspect of learning is to be able to apply knowledge gained from past experiences to novel input. In some studies of SL, for example, participants are first presented with a set of items generated by a pre-defined set of rules, and then in a subsequent test phase asked to distinguish unseen items generated by these rules (i.e. “grammatical items”) from another set of novel items that violate these rules (i.e. “ungrammatical items”). If they are able to correctly classify the unseen items as “grammatical” or “ungrammatical” at above chance levels, generalization from seen items to the novel exemplars is assumed.

Many scientists initially interpreted successful generalization as evidence that the participants had acquired the rules used to generate the stimuli and applied them to the novel stimuli. However, several studies have shown that participants’ performance at test can be readily explained by sensitivity to so-called “fragment” information, consisting of distributional properties of subparts of individual items [16]. Consider a hypothetical novel test item, ABCDE, which consists of various bigram (AB, BC, CD, DE) and trigram (ABC, BCD, CDE) fragments. The likelihood of a participant endorsing this test item as grammatical will depend on how frequently these bigram and trigram fragments have occurred in the training items. If a test item contains a fragment that has not been seen during training, then participants will tend to reject that item as ungrammatical (see [68]) . Thus, generalization in SL is often, if not always, driven by local stimulus properties and overall judgements of similarity, rather than the extraction of abstract rules.

Another possible way in which past learning could be extrapolated to new input is through the transfer of regularities learned in one domain to another (e.g., from visual input to auditory input). Although early studies appeared to support cross-modal transfer (e.g., [58,69]), more recent studies have shown that there is little, or no evidence for transfer effects, once learning during test based on repetition or simple fragment information is taken into account (e.g., [20,21,70]).

Generalization and transfer significantly differ in their contribution to theories of learning. Whereas generalization has been demonstrated in SL studies—which is important for the application of SL to language—there is little evidence of cross-modal transfer, likely because of the substantial differences in neurobiological characteristics of the visual, auditory and somatosensory cortices.

BOX 2: Advancing SL Theory via Computational Modeling.

Computational modeling serves an important dual role in both providing a quantitative account of observed empirical effects as well as the means to generate novel predictions that can guide empirical research (see, e.g., [67,71,72]). Within our framework, such modeling should reflect the relevant neural hardware of sensory cortices (e.g., different time-courses for iconic vs. echoic memory), as well as advances related to what and how the neural networks track distributional properties [40,56,73]. It should also make direct contact with neural measures as opposed to focusing strictly on behavioral measures that reflect the end-state of processing (for related discussion, see [72,74,75]).

The development of explicit models allows for the parametric variation of different aspects of the SL system, including the contributions of different learning mechanisms (e.g., Hebbian learning, reinforcement learning), different brain regions (e.g., the MTL system and modality specific areas), as well as of the quality and nature of the representations in different parts of the system (Figure I). This allows the probing of the model's ability to account not only for group-averaged effects, but also for individual differences (see Box 3; [76]), and establish how and why variation in parameters affecting different aspects of the system modulate overall performance.

Recent advances in “deep” neural networks have also enabled interesting insights into the effects of allowing intermediate representations to emerge as a function of learning [77,78], as opposed to being explicitly stipulated. This relates directly to the issues of modality and stimulus specificity that currently challenge SL theories. For instance, representations closer to the sensory cortices are learned earlier and are more strongly shaped by the specific characteristics of individual stimuli. This contrasts with higher-order (but possibly modality specific) areas that operate on these early sensory representations, and which can detect commonalities in higher-order statistics despite little similarity in the surface properties or lower-order statistical relationships amongst the stimuli (for related work using a Bayesian approach, see [79]). This predicts that SL tasks that involve stimuli whose relationships are only detectable in higher-order statistics should be more likely to show at least some generalization relative to early sensory regions, which are predicted to exhibit stronger stimulus-specificity (for a related proposal see [80]). For instance, the purpose of some brain regions is primarily to distinguish between highly-similar complex inputs (e.g., visual expertise areas such as the putative fusiform face area; [81]), or to transmit similar outputs to multiple brain regions regardless of the source of its input (e.g., the semantic memory system; [82]). Such a model is also able to account for stimulus specificity in some higher-order domains and to predict the possibility of generalization in others.

BOX 2: Advancing SL Theory via Computational Modeling

Box 2 - Figure I – Candidate computational architecture for explaining and predicting the neural and behavioural data pertaining to statistical learning

Depiction of candidate SL model architecture. In this model, visual and auditory sensory input are first encoded and processed in pools of units (neurons) that code for low-level sensory features (e.g., sound frequency, edge orientation). These pools then project to higher-level visual and auditory areas which are better suited for detecting higher-order statistics and developing more sophisticated representations (e.g., of objects or syllables). Bimodal representations may also be learned in an area that receives inputs from both modalities. All of these modality-specific and bimodal areas also project to and receive feedback from shared representation and memory modulation systems. Arrows denote connections that send representations from one pool to another; blue lines denote connections that can either send representations, modulate processing, or both. Note that this figure is not intended to be exhaustive: other representations (e.g., low-level audio-visual) are assumed to be part of a more complete model, as is the coding of more detailed sensory information inputs (e.g., color, shape, movement, taste, smell).

BOX 3: Mapping individual trajectories in statistical learning.

The present theoretical approach outlines a methodology for investigating individual performance in SL tasks by orthogonally manipulating the experimental parameters hypothesized to affect encoding efficacy on the one hand, and parameters related to efficiency in registering distributional properties, on the other. In general, manipulations that center on input encoding parameters (temporal rate of presentation, number of items in a spatial configuration, stimuli complexity, saliency of distinctive features, boundary information, etc.), will probe individual abilities in encoding stimuli in a given modality. In contrast, manipulations that center on transitional probabilities (i.e., the likelihood of Y following X, given the occurrence of X), level of predictability of events, type of statistical contingencies (e.g., adjacent or non-adjacent), or the similarity of foils to targets in the test phase, will probe the relative efficiency of a person's computational ability for registering distributional properties (see [6] for manipulation of transitional probabilities in an SRT (Serial Reaction Time) task). Such parametric experimental designs would reveal, for any given individual, a specific pattern of interactions of two main factors driving SL, outlining how their joint contribution determines his or her performance on a specific task.

Figure I presents hypothetical plots of the performance of two possible individuals in such parametric manipulations. The figure demonstrates differential trajectories of individual sensitivities to either type of manipulation, one that probes stimulus encoding efficacy (manipulating inter-stimulus intervals), and one that probes inter-stimulus statistic encoding efficiency (manipulating transitional probabilities). This experimental approach has the additional promise of revealing systematic commonalities or differences in sensitivity to various types of distributional properties across domains or modalities.

A possible extension of this line of research would incorporate the impact of prior knowledge on SL. The process of encoding representations of any continuous input is dependent on the characteristics of the representational space that exist at a given time point for a given individual. Thus, encoding an input of continuous syllabic elements (e.g., [12]) is different than encoding a sequence of non-linguistic novel sounds (e.g., [83]), affecting SL efficacy. This could generate significant individual differences in SL in domains such as language, where individuals differ significantly in their linguistic representations (e.g., vocabulary size, number of languages spoken).

Note that most current research on individual differences in SL focuses on predicting general cognitive or linguistic abilities from performance in SL tasks [19,5961,84,85] or showing similar neural correlates within subjects for SL and language [86,87]. Investigating the various facets of performance in SL, as outlined above, is a necessary further step to describe and explain the specific sources of potential correlations between SL test measures and the cognitive functions they are aimed to predict. Identifying these sources would, in turn, allow researchers to refine predictions and generate well-defined a priori hypotheses.

BOX 3: Mapping individual trajectories in statistical learning

Box 3 - Figure I - Predicted empirical results illustrating how stimulus encoding and transitional probability shape individual differences

The two graphs above present hypothetical data from two participants and illustrates how the ability to detect regularities and to encode inputs may be separated experimentally. Panel A demonstrates the manipulation of rate of presentation and shows that whereas Participant 1 performs well even in relatively fast rates, Participant 2 shows no learning when stimuli are presented at or above a rate of one per 600 ms. Panel B displays the manipulation of transitional probabilities. Here the rate of presentation is the same across all 5 tasks, but transitional probabilities vary from 0.6 to 1. The results show that Participant 2, who performs above chance in the test even when the transitional probabilities between elements are low, is more efficient in detecting probabilities than Participant 1.

BOX 4: Outstanding questions.

  • To what degree are high-level cognitive SL effects and low-level sensorimotor SL effects modulated by the partially shared SL systems (e.g., hippocampus, basal ganglia, inferior frontal gyrus) versus modality-specific systems?

  • Can a better understanding of low-level cellular and systems neurobiology guide theoretical advance by predicting the specific types of information that a brain region will be most suited to encode and, consequently, the types of statistical learning that may take place?

  • To what degree does variability in the quality and nature of an individual's modality-specific representations of individual stimuli, and variability in sensitivity to the dependencies between stimuli, explain individual differences in SL experiments?

  • To what degree are the modality-specific and partially-shared neural processing systems that underlie SL modulated by experience versus neuronal maturation throughout development?

Highlights.

  • Statistical learning (SL) theory is challenged by modality/stimulus-specific effects.

  • We argue SL is shaped by both modality-specific constraints and domain-general principles.

  • SL relies on modality-specific neural networks and partially-shared neural networks.

  • Studies of individual differences provide targeted insights into the mechanisms of SL.

Acknowledgments

This paper was supported by The Israel Science Foundation (Grant 217/14 awarded to Ram Frost), by the NICHD (RO1 HD 067364 awarded to Ken Pugh and Ram Frost, and PO1 HD 01994 awarded to Haskins Laboratories), and by a Marie Curie IIF award (PIIF-GA-2013-627784 awarded to Blair C. Armstrong).

GLOSSARY

Amodal representations

“Amodal” representations are typically taken to be “abstract” in the sense that they are not bound by specific sensory features (e.g., visual or auditory). Apart from the problem of defining a theoretical construct in terms of what it is not, the neurobiological evidence for such representations is scarce.

Artificial Grammar Learning (AGL)

In a typical AGL experiment, participants are exposed to sequences generated by a miniature grammar. Participants are only informed about the rule-based nature of the sequences after the exposure phase, when they are asked to classify a new set of sequences, some of which follow the grammar while others do not. AGL is also considered to be a kind of implicit learning task.

Generalization

Refers to extension of learned statistical structure to unseen stimuli, typically from within the same modality or stimulus domain.

Internal Representation

In neurobiological terms, an internal representation of a stimulus is the pattern of neural activity evoked by a stimulus in a brain region (or network of brain regions).

McGurk effect

The McGurk effect [88] illustrates the potentially complex interactions between two conflicting streams of information from the auditory and visual modalities. For instance, if a video of an individual pronouncing /ga/ is combined with the sound /ba/, a listener will tend to hear /da/ because the sound /da/ is most consistent with the visually-perceived positions of the lips and with the auditorily-perceived sound.

Modality

The sensorimotor mode in which the stimulus was presented (e.g., vision, audition, touch). One modality may contain several sub-modalities (e.g., visual motion, color), each of which is subserved by distinct neuroanatomy.

Multimodal representations

Representations that form when information from two or more modalities are integrated in a representational space and associated brain region (or network of regions). Importantly, these representations are, therefore, not “amodal”.

Transfer

A broader type of extension of learned knowledge than generalization, and refers to the application of learned regularities to novel domains and/or modalities.

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

References

  • 1.Chomsky N. A Review of B.F. Skinner's Verbal Behavior. Language (Baltim) 1959;35:26–58. [Google Scholar]
  • 2.Eimas PD, et al. Speech perception in infants. Science (80-. ) 1971;171:303–306. doi: 10.1126/science.171.3968.303. [DOI] [PubMed] [Google Scholar]
  • 3.Saffran JR, et al. Statistical Learning by 8-Month-Old Infants. Science (80-. ) 1996;274:1926–1928. doi: 10.1126/science.274.5294.1926. [DOI] [PubMed] [Google Scholar]
  • 4.Baker CI, et al. Role of attention and perceptual grouping in visual statistical learning. Psychol. Sci. 2004;15:460–466. doi: 10.1111/j.0956-7976.2004.00702.x. [DOI] [PubMed] [Google Scholar]
  • 5.Goujon A, Fagot J. Learning of spatial statistics in nonhuman primates: Contextual cueing in baboons (Papio papio). Behav. Brain Res. 2013;247:101–109. doi: 10.1016/j.bbr.2013.03.004. [DOI] [PubMed] [Google Scholar]
  • 6.Hunt RH, Aslin RN. Statistical learning in a serial reaction time task: access to separable statistical cues by individual learners. J. Exp. Psychol. Gen. 2001;130:658–680. doi: 10.1037//0096-3445.130.4.658. [DOI] [PubMed] [Google Scholar]
  • 7.Courville AC, et al. Bayesian theories of conditioning in a changing world. Trends Cogn. Sci. 2006;10:294–300. doi: 10.1016/j.tics.2006.05.004. [DOI] [PubMed] [Google Scholar]
  • 8.Friston K. The free-energy principle: a unified brain theory? Nat. Rev. Neurosci. 2010;11:127–138. doi: 10.1038/nrn2787. [DOI] [PubMed] [Google Scholar]
  • 9.Tishby N, Polani D. Perception-Action Cycle. Springer; 2011. Information theory of decisions and actions. pp. 601–638. [Google Scholar]
  • 10.Kirkham NZ, et al. Visual statistical learning in infancy: Evidence for a domain general learning mechanism. Cognition. 2002;83:B35–B42. doi: 10.1016/s0010-0277(02)00004-5. [DOI] [PubMed] [Google Scholar]
  • 11.Bulf H, et al. Visual statistical learning in the newborn infant. Cognition. 2011;121:127–132. doi: 10.1016/j.cognition.2011.06.010. [DOI] [PubMed] [Google Scholar]
  • 12.Endress AD, Mehler J. The surprising power of statistical learning: When fragment knowledge leads to false memories of unheard words. J. Mem. Lang. 2009;60:351–367. [Google Scholar]
  • 13.Saffran JR, Thiessen ED. Hoff E, Shatz M, editors. Domain-General Learning Capacities. Blackwell Handbook of Language Development. 2007. pp. 68–86.
  • 14.Hauser MD, et al. Segmentation of the speech stream in a non-human primate: Statistical learning in cotton-top tamarins. Cognition. 2001;78:B53–B64. doi: 10.1016/s0010-0277(00)00132-3. [DOI] [PubMed] [Google Scholar]
  • 15.Toro JM, Trobalón JB. Statistical computations over a speech stream in a rodent. Percept. Psychophys. 2005;67:867–875. doi: 10.3758/bf03193539. [DOI] [PubMed] [Google Scholar]
  • 16.Conway CM, Christiansen MH. Modality-constrained statistical learning of tactile, visual, and auditory sequences. J. Exp. Psychol. Learn. Mem. Cogn. 2005;31:24–39. doi: 10.1037/0278-7393.31.1.24. [DOI] [PubMed] [Google Scholar]
  • 17.Conway CM, Christiansen MH. Statistical learning within and between modalities: pitting abstract against stimulus-specific representations. Psychol. Sci. 2006;17:905–912. doi: 10.1111/j.1467-9280.2006.01801.x. [DOI] [PubMed] [Google Scholar]
  • 18.Mitchel AD, Weiss DJ. Learning across senses: Cross-modal effects in multisensory statistical learning. J. Exp. Psychol. Learn. Mem. Cogn. 2011;37:1081–1091. doi: 10.1037/a0023700. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Misyak JB, Christiansen MH. Statistical learning and language: An individual differences study. Lang. Learn. 2012;62:302–331. [Google Scholar]
  • 20.Redington M, Chater N. Transfer in artificial grammar learning: A reevaluation. J. Exp. Psychol. Gen. 1996;125:123–138. [Google Scholar]
  • 21.Tunney RJ, Altmann GTM. The transfer effect in artificial grammar learning: Reappraising the evidence on the transfer of sequential dependencies. J. Exp. Psychol. Learn. Mem. Cogn. 1999;25:1322–1333. [Google Scholar]
  • 22.Emberson LL, et al. Timing is everything: changes in presentation rate have opposite effects on auditory and visual implicit statistical learning. Q. J. Exp. Psychol. (Hove) 2011;64:1021–1040. doi: 10.1080/17470218.2010.538972. [DOI] [PubMed] [Google Scholar]
  • 23.Johansson T. Strengthening the case for stimulus-specificity in artificial grammar learning: no evidence for abstract representations with extended exposure. Exp. Psychol. 2009;56:188–197. doi: 10.1027/1618-3169.56.3.188. [DOI] [PubMed] [Google Scholar]
  • 24.Sigman M, Gilbert CD. Learning to find a shape. Nat. Neurosci. 2000;3:264–269. doi: 10.1038/72979. [DOI] [PubMed] [Google Scholar]
  • 25.Sjöström PJ, et al. Dendritic excitability and synaptic plasticity. Physiol. Rev. 2008;88:769–840. doi: 10.1152/physrev.00016.2007. [DOI] [PubMed] [Google Scholar]
  • 26.Samson RD, et al. Computational models of reinforcement learning: the role of dopamine as a reward signal. Cogn. Neurodyn. 2010;4:91–105. doi: 10.1007/s11571-010-9109-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Chen L, Vroomen J. Intersensory binding across space and time: a tutorial review. Atten. Percept. Psychophys. 2013;75:790–811. doi: 10.3758/s13414-013-0475-4. [DOI] [PubMed] [Google Scholar]
  • 28.Recanzone GH. Interactions of auditory and visual stimuli in space and time. Hear. Res. 2009;258:89–99. doi: 10.1016/j.heares.2009.04.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Sams M, et al. The Human Auditory Sensory Memory Trace Persists about 10 sec: Neuromagnetic Evidence. J. Cogn. Neurosci. 1993;5:363–70. doi: 10.1162/jocn.1993.5.3.363. [DOI] [PubMed] [Google Scholar]
  • 30.Fiser J, Aslin RN. Unsupervised statistical learning of higher-order spatial structures from visual scenes. Psychol. Sci. a J. Am. Psychol. Soc. / APS. 2001;12:499–504. doi: 10.1111/1467-9280.00392. [DOI] [PubMed] [Google Scholar]
  • 31.Fiser J, Aslin RN. Statistical learning of higher-order temporal structure from visual shape sequences. J. Exp. Psychol. Learn. Mem. Cogn. 2002;28:458–467. doi: 10.1037//0278-7393.28.3.458. [DOI] [PubMed] [Google Scholar]
  • 32.Mitchel AD, et al. Multimodal integration in statistical learning: Evidence from the McGurk illusion. Front. Psychol. 2014;5:407. doi: 10.3389/fpsyg.2014.00407. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Morein-Zamir S, et al. Auditory capture of vision: Examining temporal ventriloquism. Cogn. Brain Res. 2003;17:154–163. doi: 10.1016/s0926-6410(03)00089-2. [DOI] [PubMed] [Google Scholar]
  • 34.Tyll S, et al. Thalamic influences on multisensory integration. Commun. Integr. Biol. 2011;4:378–381. doi: 10.4161/cib.4.4.15222. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Basura GJ, et al. Multi-sensory integration in brainstem and auditory cortex. Brain Res. 2012;1485:95–107. doi: 10.1016/j.brainres.2012.08.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Patterson K, et al. Where do you know what you know? The representation of semantic knowledge in the human brain. Nat. Rev. Neurosci. 2007;8:976–987. doi: 10.1038/nrn2277. [DOI] [PubMed] [Google Scholar]
  • 37.Romanski LM, Hwang J. Timing of audiovisual inputs to the prefrontal cortex and multisensory integration. Neuroscience. 2012;214:36–48. doi: 10.1016/j.neuroscience.2012.03.025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Altmann GTM, et al. Modality independence of implicitly learned grammatical knowledge. J. Exp. Psychol. Learn. Mem. Cogn. 1995;21:899–912. [Google Scholar]
  • 39.Fedorenko E, Thompson-Schill SL. Reworking the language network. Trends Cogn. Sci. 2014;18:120–126. doi: 10.1016/j.tics.2013.12.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Turk-Browne NB, et al. Neural evidence of statistical learning: efficient detection of visual regularities without awareness. J. Cogn. Neurosci. 2009;21:1934–45. doi: 10.1162/jocn.2009.21131. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Bischoff-Grethe A, et al. Conscious and unconscious processing of nonverbal predictability in Wernicke's area. J. Neurosci. 2000;20:1975–1981. doi: 10.1523/JNEUROSCI.20-05-01975.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.McNealy K, et al. Cracking the language code: neural mechanisms underlying speech parsing. J. Neurosci. 2006;26:7629–7639. doi: 10.1523/JNEUROSCI.5501-05.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Karuza EA, et al. The neural correlates of statistical learning in a word segmentation task: An fMRI study. Brain Lang. 2013;127:46–54. doi: 10.1016/j.bandl.2012.11.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Schapiro AC, et al. The Necessity of the Medial-Temporal Lobe for Statistical Learning. J. Cogn. Neurosci. 2014;26:1736–1747. doi: 10.1162/jocn_a_00578. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Cohen NJ, Eichenbaum H. Memory, Amnesia, and the Hippocampal System. MIT Press; 1993. [Google Scholar]
  • 46.Eichenbaum H. Memory on time. Trends Cogn. Sci. 2013;17:81–88. doi: 10.1016/j.tics.2012.12.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Bornstein AM, Daw ND. Dissociating hippocampal and striatal contributions to sequential prediction learning. Eur. J. Neurosci. 2012;35:1011–23. doi: 10.1111/j.1460-9568.2011.07920.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Shohamy D, Turk-Browne NB. Mechanisms for widespread hippocampal involvement in cognition. J. Exp. Psychol. Gen. 2013;142:1159–70. doi: 10.1037/a0034461. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Papanicolaou AC, et al. The Hippocampus and Memory of Verbal and Pictorial Material. Learn. Mem. 2002;9:99–104. doi: 10.1101/lm.44302. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Poppenk J, et al. Long-axis specialization of the human hippocampus. Trends Cogn. Sci. 2013;17:230–40. doi: 10.1016/j.tics.2013.03.005. [DOI] [PubMed] [Google Scholar]
  • 51.McClelland JL, et al. Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. Psychol. Rev. 1995;102:419–457. doi: 10.1037/0033-295X.102.3.419. [DOI] [PubMed] [Google Scholar]
  • 52.Azab M, et al. Contributions of human hippocampal subfields to spatial and temporal pattern separation. Hippocampus. 2014;24:293–302. doi: 10.1002/hipo.22223. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.O'Reilly RC, et al. Complementary Learning Systems. Cogn. Sci. 2011;38:1229–1248. doi: 10.1111/j.1551-6709.2011.01214.x. [DOI] [PubMed] [Google Scholar]
  • 54.Rolls ET. The mechanisms for pattern completion and pattern separation in the hippocampus. Front. Syst. Neurosci. 2013;7:74. doi: 10.3389/fnsys.2013.00074. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Poldrack RA, et al. The neural correlates of motor skill automaticity. J. Neurosci. 2005;25:5356–5364. doi: 10.1523/JNEUROSCI.3880-04.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Roux F, Uhlhaas PJ. Working memory and neural oscillations: alpha-gamma versus theta-gamma codes for distinct WM information? Trends Cogn. Sci. 2014;18:16–25. doi: 10.1016/j.tics.2013.10.010. [DOI] [PubMed] [Google Scholar]
  • 57.Knowlton BJ, et al. Intact Artificial Grammar Learning in Amnesia: Dissociation of Classification Learning and Explicit Memory for Specific Instances. Psychol. Sci. 1992;3:172–179. [Google Scholar]
  • 58.Reber AS. Implicit Learning and Tacit Knowledge: An Essay on the Cognitive Unconscious. Oxford University Press; 1996. [Google Scholar]
  • 59.Arciuli J, Simpson IC. Statistical learning is related to reading ability in children and adults. Cogn. Sci. 2012;36:286–304. doi: 10.1111/j.1551-6709.2011.01200.x. [DOI] [PubMed] [Google Scholar]
  • 60.Frost R, et al. What predicts successful literacy acquisition in a second language? Psychol. Sci. 2013;24:1243–52. doi: 10.1177/0956797612472207. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Misyak JB, et al. On-line individual differences in statistical learning predict language processing. Front. Psychol. 2010;1:31. doi: 10.3389/fpsyg.2010.00031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Armstrong BC, Plaut DC. Simulating overall and trial-by-trial effects in response selection with a biologically-plausible connectionist network.. Proceedings of the 35th Annual Conference of the Cognitive Science Society.2013. pp. 139–144. [Google Scholar]
  • 63.Flöel A, et al. White matter integrity in the vicinity of Broca's area predicts grammar learning success. Neuroimage. 2009;47:1974–1981. doi: 10.1016/j.neuroimage.2009.05.046. [DOI] [PubMed] [Google Scholar]
  • 64.Matzel LD, et al. Synaptic efficacy is commonly regulated within a nervous system and predicts individual differences in learning. Neuroreport. 2000;11:1253–1258. doi: 10.1097/00001756-200004270-00022. [DOI] [PubMed] [Google Scholar]
  • 65.Thiessen ED, et al. The extraction and integration framework: a two-process account of statistical learning. Psychol. Bull. 2013;139:792–814. doi: 10.1037/a0030801. [DOI] [PubMed] [Google Scholar]
  • 66.Hagmann P, et al. White matter maturation reshapes structural connectivity in the late developing human brain. Proc. Natl. Acad. Sci. 2010;107:19067–19072. doi: 10.1073/pnas.1009073107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Thiessen ED, Pavlik PI. iMinerva: a mathematical model of distributional statistical learning. Cogn. Sci. 2013;37:310–43. doi: 10.1111/cogs.12011. [DOI] [PubMed] [Google Scholar]
  • 68.Reeder PA, et al. From shared contexts to syntactic categories: The role of distributional information in learning linguistic form-classes. Cogn. Psychol. 2013;66:30–54. doi: 10.1016/j.cogpsych.2012.09.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Marcus GF, et al. Infant rule learning facilitated by speech. Psychol. Sci. 2007;18:387–391. doi: 10.1111/j.1467-9280.2007.01910.x. [DOI] [PubMed] [Google Scholar]
  • 70.Gomez RL, et al. The basis of transfer in artificial grammar learning. Mem. Cognit. 2000;28:253–263. doi: 10.3758/bf03213804. [DOI] [PubMed] [Google Scholar]
  • 71.Elman JL. Finding structure in time. Cogn. Sci. 1990;14:179–211. [Google Scholar]
  • 72.Carreiras M, et al. The what, when, where, and how of visual word recognition. Trends Cogn. Sci. 2014;18:90–98. doi: 10.1016/j.tics.2013.11.005. [DOI] [PubMed] [Google Scholar]
  • 73.Sutskever I, et al. The Recurrent Temporal Restricted Boltzmann Machine. Neural Inf. Process. Syst. 2008;21:1601–1608. [Google Scholar]
  • 74.Laszlo S, Armstrong BC. PSPs and ERPs: Applying the dynamics of post-synaptic potentials to individual units in simulation of temporally extended Event-Related Potential reading data. Brain Lang. 2014;132C:22–27. doi: 10.1016/j.bandl.2014.03.002. [DOI] [PubMed] [Google Scholar]
  • 75.Laszlo S, Plaut DC. A neurally plausible Parallel Distributed Processing model of Event-Related Potential word reading data. Brain Lang. 2012;120:271–281. doi: 10.1016/j.bandl.2011.09.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Lambon Ralph MA, et al. Finite case series or infinite single-case studies? Comments on “Case series investigations in cognitive neuropsychology” by Schwartz and Dell (2010). Cogn. Neuropsychol. 2011;28:466–74. doi: 10.1080/02643294.2012.671765. [DOI] [PubMed] [Google Scholar]
  • 77.Di Bono MG, Zorzi M. Deep generative learning of location-invariant visual word recognition. Front. Psychol. 2013;4:635. doi: 10.3389/fpsyg.2013.00635. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Hinton GE, Salakhutdinov RR. Reducing the dimensionality of data with neural networks. Science (80-. ) 2006;313:504–507. doi: 10.1126/science.1127647. [DOI] [PubMed] [Google Scholar]
  • 79.Orbán G, et al. Bayesian learning of visual chunks by human observers. Proc. Natl. Acad. Sci. 2008;105:2745–2750. doi: 10.1073/pnas.0708424105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Aslin RN, Newport EL. Statistical learning: From acquiring specific items to forming general rules. Curr. Dir. Psychol. Sci. 2012;21:170–176. doi: 10.1177/0963721412436806. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Plaut D, Behrmann M. Complementary neural representations for faces and words: A computational exploration. Cogn. Neuropsychol. 2011;28:251–275. doi: 10.1080/02643294.2011.609812. [DOI] [PubMed] [Google Scholar]
  • 82.McClelland JL, Rogers TT. The parallel distributed processing approach to semantic cognition. Nat. Rev. Neurosci. 2003;4:310–22. doi: 10.1038/nrn1076. [DOI] [PubMed] [Google Scholar]
  • 83.Gebhart AL, et al. Statistical learning of adjacent and nonadjacent dependencies among nonlinguistic sounds. Psychon. Bull. Rev. 2009;16:486–490. doi: 10.3758/PBR.16.3.486. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Kidd E. Implicit statistical learning is directly associated with the acquisition of syntax. Dev. Psychol. 2012;48:171–184. doi: 10.1037/a0025405. [DOI] [PubMed] [Google Scholar]
  • 85.Kaufman SB, et al. Implicit learning as an ability. Cognition. 2010;116:321–340. doi: 10.1016/j.cognition.2010.05.011. [DOI] [PubMed] [Google Scholar]
  • 86.Christiansen MH, et al. Similar Neural Correlates for Language and Sequential Learning: Evidence from Event-Related Brain Potentials. Lang. Cogn. Process. 2012;27:231–256. doi: 10.1080/01690965.2011.606666. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Petersson KM, et al. What artificial grammar learning reveals about the neurobiology of syntax. Brain Lang. 2012;120:83–95. doi: 10.1016/j.bandl.2010.08.003. [DOI] [PubMed] [Google Scholar]
  • 88.McGurk H, MacDonald J. Hearing lips and seeing voices. Nature. 1976;264:746–748. doi: 10.1038/264746a0. [DOI] [PubMed] [Google Scholar]
  • 89.Xia M, et al. BrainNet Viewer: A Network Visualization Tool for Human Brain Connectomics. PLoS One. 2013;8:e68910. doi: 10.1371/journal.pone.0068910. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES