Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2014 Apr 2;1316(1):87–104. doi: 10.1111/nyas.12406

The Year in Cognitive Neuroscience

W Tecumseh Fitch 1, Mauricio D Martins 1,2
PMCID: PMC4285949  PMID: 24697242

Abstract

Sixty years ago, Karl Lashley suggested that complex action sequences, from simple motor acts to language and music, are a fundamental but neglected aspect of neural function. Lashley demonstrated the inadequacy of then-standard models of associative chaining, positing a more flexible and generalized “syntax of action” necessary to encompass key aspects of language and music. He suggested that hierarchy in language and music builds upon a more basic sequential action system, and provided several concrete hypotheses about the nature of this system. Here, we review a diverse set of modern data concerning musical, linguistic, and other action processing, finding them largely consistent with an updated neuroanatomical version of Lashley's hypotheses. In particular, the lateral premotor cortex, including Broca's area, plays important roles in hierarchical processing in language, music, and at least some action sequences. Although the precise computational function of the lateral prefrontal regions in action syntax remains debated, Lashley's notion—that this cortical region implements a working-memory buffer or stack scannable by posterior and subcortical brain regions—is consistent with considerable experimental data.

Keywords: hierarchy, language, music, computation, syntax

Introduction

In 1951, Karl Lashley1 suggested that the human capacity for serial ordering of action, whether of language or music or simply making coffee, presents central problems for our understanding of brain function. Lashley observed that then-prevailing models of action based on simple stimulus–response chaining had a fundamental flaw: although such serial models can capture the “one thing after another” characteristic of routine action, they fail to recognize the central importance of sustained goals and subgoals in more complex action. In complex actions, overall goals (e.g., making coffee) must persist over time while subgoals are initiated and completed (e.g., grinding beans, heating water, and adding cream) or interruptions occur (e.g., a phone call). Several aspects of such action sequences are inconsistent with a chaining model. Most obvious are errors of omission, where some step is accidentally left out (e.g., starting the coffee machine without first adding coffee). In chaining models, each completed action serves as the stimulus for the next one, and omission of any step would cause the sequence to grind to a halt. This and other phenomena, argued Lashley, demonstrate the need for hierarchical models of action planning—a requirement widely acknowledged today.2

Lashley further observed that core aspects of spoken language and music share this need for hierarchical structure, at multiple levels. For example, when pronouncing the words tire and right, the same basic vocal gestures appear but in different orders. How could a chaining system cope with such flexible rearrangement of basic actions into multiple larger composites? Phrasal syntax makes the point even more clearly. When reading the sentence, “The boy who patted the dog chased the girl,” an English speaker knows that the boy, and not the dog, chased the girl, despite the fact that this sentence contains the sequence “the dog chased the girl.” Linguistic syntax illustrates that serial order alone provides an inadequate basis for language processing, and that processes involving longer temporal scales must persist during the processing or production of shorter actions or simple chains. Such observations were foundational in early generative linguistics, which recognized the central role of phrase structure, rather than word order, in syntax.3,4

Lashley also suggested that these similarities in language, music, and other complex actions are not coincidental, instead hypothesizing that the hierarchical nature of music and language was inherited from more basic features of motor planning and action.

In this review, we reassess Lashley's observations and hypotheses from a modern perspective, concluding that his ideas were remarkably insightful and consistent with considerable new behavioral and neuroscientific data. We explore Lashley's idea that hierarchical structuring of temporal sequences is a key capability underlying human music and language, but one that has deeper phylogenetic roots in the domain of action. Unlike Lashley, who eschewed localizationist arguments, we make full use of brain imaging and lesion data as supporting evidence, with a particular focus on the role of Broca's area in processing hierarchical structures in the time domain.

Specifically, we will consider the following issues: (1) whether processing of sequential hierarchy is a well-defined specific capability, independent of other hierarchical processing (e.g., visuospatial) but applicable to both music and language; (2) to what extent the mechanisms subserving linguistic and musical syntax inherit properties from those typifying a syntax of action more generally; and (3) whether humans are unusual in the degree to which hierarchical abilities are developed, and if so, which neural mechanisms support this proclivity.

We evaluate these questions using behavioral, neuroanatomical, and brain-imaging data that go well beyond the data available in Lashley's day. We conclude that these data converge upon the notion of a distinct neural substrate for hierarchical sequence processing, unusually well developed in humans relative to other primates, and reflected in specific expansions of prefrontal brain regions including Broca's area, as well as a considerable remodeling of connections between prefrontal and posterior regions. Although the degree to which music and language inherit key properties from action syntax remains controversial, available data are consistent with overlap in these systems from both neural and computational viewpoints. More precise computational models and experimental designs will be required to further evaluate this hypothesis, which is quite different from currently popular hypotheses based upon mirror neurons and embodiment. Thus, 60 years on, Lashley's insights remain valid and intriguing and can inform current neuroscientific debates, and thus deserve renewed attention.

Defining temporal hierarchy: hierarchical sequences and hierarchical sets

Terms like syntax and hierarchy are notoriously ambiguous, so our first task in evaluating Lashley's hypotheses is to distinguish possible interpretations of the term hierarchy, which has many meanings in neuroscience. Especially in older literature, hierarchy connotes little more than the fact that cortical regions are above or below one another in processing terms, such that bottom–up processing proceeds from regions lower to regions higher in the visual hierarchy. A central characteristic of the visual system is that neurons in cortical regions closer to retinal input, such as V1, have smaller receptive fields than those in higher regions like V2. Our focus goes beyond this basic and widespread form of hierarchical layout of cortical regions themselves.

Here, we adopt a more specific and precise notion of hierarchy to denote a tree-like organization, where higher levels incorporate multiple lower levels in structural representations and/or processing. This notion is wholly independent of how the cortical regions that process such stimuli are arranged, entailing no specific commitments about neuroanatomical localization.

To be more exact we utilize the following mathematically-based definitions5,6 throughout this paper: a set is an unordered collection of distinct, unique objects; while a sequence is a collection of objects, perhaps including duplicates, ordered by some rule.

Although the set {a, b, c} is identical to the set {b, c, a}, the sequences [a b c] and [b c a] are distinct and different. Furthermore, because sequences but not sets can contain duplicate items, sequences are not, strictly speaking, a type of set. These distinctions are crucial in distinguishing Lashley's temporally ordered sequences, where order matters, from static hierarchy, in which order is typically irrelevant.

Hierarchy denotes a set or sequence of elements connected in the form of a rooted tree (a connected acyclic graph, in which one element is singled out as the root element). Hierarchies thus possess the following key properties: (1) all elements are combined into one structure (connectedness); (2) one element (the root) is superior to all others; and (3) no element is superior to itself (that is, there are no cycles, direct or indirect).

These definitions allow us to clarify a core distinction between temporally ordered and unordered hierarchies. Examples of hierarchical sets include visuospatial part–whole hierarchies, such as the schema of a face, in which certain components contain others without any necessary sequential ordering of components at each level. A face incorporates the set {eyes, nose, mouth}, but there is no need to consider one component to come first. Such hierarchies involve the minimal structure required to consider a representation hierarchical, namely that the superior/inferior relation between levels is specified. But at any given level, the elements form an unordered set.

In hierarchical sequences, in contrast, sequential order matters. The simple two-syllable utterances “callow” and “low-cal” contain the same set of syllables, but arranged in different orders, and they represent different concepts. Similarly, in music each arrangement of a set of notes constitutes a different melody. Finally, in action sequences, order is often important: if we grind the coffee beans at the end of the action sequence, after brewing is completed, poor coffee results. Thus temporal hierarchies incorporate an additional ordering component, where at least some elements at any given level represent a sequence rather than a set.

Lashley considered hierarchical structure to be an obvious component of language, music, and action. In each domain, some superordinate structure (his “determining tendency”) is required to lend coherence to subordinate components (speech sounds, notes, or simple actions). Although these higher level elements may differ considerably between domains (e.g., meaning for language versus an overall goal for action), Lashley's focus was what all three domains have in common: the need to impose the correct temporal ordering on the subelements. He cites, as a key example, language, where the elements of meaning are often cotemporal sets, but different languages have different ordering rules by which the same meaning must be expressed in a sequence: different “schemas of order.” Typing or speech errors (e.g., spoonerisms) or children's games like pig Latin provide evidence that the temporal ordering component is both independent of semantics, and a general characteristic of many “generalized schemata of action” that apply across the three domains. Thus, it is not the general problem of hierarchy but the specific issue of hierarchical sequencing that Lashley considered the essential problem of serial order, and that will be our focus in this review.

Lashley observed that we can often convert between spatial and temporal hierarchies. For instance, to reverse a piano melody, he stated, “I can only do it by visualizing the music spatially and then reading it backward” (p. 129). Similarly, when translating between languages with different word orders, such as English and German, a fluent bilingual can extract the meaning from one language, and convert it to a properly sequenced sentence in the other, freely reordering the individual words as needed. It is therefore possible that the distinction between sequential and nonordered hierarchy is of little importance at a fundamental neural or computational level. This is the first issue that we address using modern imaging data.

Sequential hierarchical processing is not colocalized with visual hierarchy systems

As noted above, there is often an increase in spatial receptive field size in cortical regions more distant from sensory input; this involves a widespread neural form of spatial hierarchy, where higher regions integrate information from multiple lower regions.7 In this functional sense, hierarchical processing is a widespread property of brain systems.816

In some cases, such patterns of neuroanatomical layout may reflect a functional form of cognitive hierarchy. For instance, in the motor domain, posterior frontal lobe regions encode basic movements (extending the arm), whereas anterior regions fire more selectively to particular contexts in which these movements are performed (e.g., making coffee and shaking hands).10 In the auditory cortex, core areas respond to simple acoustic characteristics, whereas anterior regions along the supratemporal plane display selective firing to more complex sound categories.12 In the visual ventral stream, anterior regions in the temporal cortex support categorical decisions of greater abstraction (e.g., animal versus nonanimal) than posterior regions (e.g., blue versus green).14 Given these pervasive patterns of brain organization, is there evidence for more abstract and modality independent hierarchical structuring in the brain?

Starting with hierarchical sets, many recent studies explore the brain regions involved in representing and processing hierarchical structures in both spatial and social domains.17,18 Representation of hierarchical structures in the social domain, which allows the evaluation of dominance relationships, appear to be encoded in the hippocampus.18 However, this structure is also active in encoding hierarchical ranks in nonsocial domains.18 In the visuospatial domain, for example, correct integration of landmarks within context frames recruits the parahippocampus,8 and also requires the integrity of the medial temporal lobe (MTL) and the retrosplenial cortex.17

Interestingly, this same MTL system may also encode high order hierarchical associations in the motor and linguistic domains.1922 It has been proposed that these medial structures are important for binding items within contexts,23,24 even for offline imagined scenes,25 as well as for episodic memory, which involves binding different perceptual and memory features into unified representations.26,27

In addition to these brain regions, a quite different set of activations is typically observed in temporal hierarchical processing, regardless of whether these involve musical, linguistic, or other tasks. A consistent finding within the neurolinguistic literature is that sequentially structured hierarchical processing activates the posterior prefrontal cortex, centering on Broca's area (a term we will use hereafter to denote both Brodmann's areas [BA] 44 and 45, typically in the left hemisphere, but not neighboring areas28). Increased Broca's activation typically correlates with increased demands on working memory, for example, when processing long-distance dependencies, where early information must remain active for proper interpretation of later elements. Ongoing controversy concerns the degree to which this active maintenance of past information is specific to hierarchical sequences per se22,29,30 or is found for nonhierarchical sequences as well,31,32 but it is clear that increasingly complex hierarchical structures lead to greater Broca's activation across multiple languages and laboratories.

Furthermore, in musical syntax, hierarchically dependent processing can stretch over long time spans (long-distance dependency).33 Neuroimaging studies on musical syntax (as probed with sequential harmonic or melodic structures) also consistently reveal activation in BA 44/45, typically in both the right and left hemispheres, but sometimes biased to the right side.3439 Evidence that these prefrontal activations are not restricted to meaningful language or music comes from artificial grammar research using nonsense syllables (spoken) or words (written) arranged according to particular rules. Again, irrespective of the input domain, both auditory and visual patterns, and even nonlinguistic visual symbols,40 elicit Broca's activation when they are sequentially presented. On the basis of these data, further explored below, we conclude that, despite MTL overlaps, the neural mechanisms for processing hierarchical sequences are not fully colocalized with those that process hierarchical sets. Despite formal similarities in these abstract hierarchical structures, these data argue against their identity. We now explore the functions of the prefrontal regions preferentially involved in hierarchical sequence processing.

Prefrontal cortex and Broca's area play central roles in processing hierarchical syntax

We first attempt to characterize the capacity for processing sequential hierarchies more precisely, to better understand its neural basis. Although Lashley was a confirmed agnostic regarding neural localization, there is a long tradition of associating the frontal lobes, specifically the prefrontal cortex (the portion of the frontal lobe anterior to the primary motor strip and the premotor cortex), with the organization of complex plans in general, and hierarchical sequences in particular, in multiple cognitive domains. The notion that the prefrontal cortex is specifically involved in action planning and executive control has been advanced notably by Luria,41 Shallice,42 and Baddeley,43 among many others. Clinical data leave little doubt that pathology in the frontal lobe can lead to disorganized action.44 We do not review this extensive literature here, but simply take it for granted that the frontal lobes play important roles in planning, working memory, and executive control. Our goal will be to specify this contribution more precisely, paying particular attention to the role of the inferior frontal gyrus (IFG) and Broca's area.

We note first that virtually any motor behavior can be conceptualized as hierarchical. A monkey eating peanuts involves a set of actions (grasp, ingest, chew) each of which can be further broken down (grasp: extend arm, open fingers, close fingers, retract arm). Thus, any action sequence that involves repeated subparts displays a simple form of hierarchy, pervasive to all motor action in all species.45 Such hierarchies have a fixed and limited depth, and were not Lashley's focus (because they can be accounted for by chaining). To clarify what he was getting at, we start with the clear case of language, and use this to guide our further inquiry through music, back to action.

Broca's area (BA 44 and 45 or, roughly, the pars opercularis and the pars triangularis of the IFG) is one of the most intensively studied brain regions, and its importance for speech production requires little emphasis.4648 Since the 1970s, this region has also been recognized in clinical studies to play a special role in syntax, both production and perception.49 Thus, a broad consensus exists that this region of the IFG plays an important role in language processing.30,5052

An elegant study by Pallier et al.53 explored the specificity of IFG involvement in hierarchical syntax by crossing two manipulations. In the first, they parametrically varied the chunk size of 12-word sentences, spanning from scrambled words (where there are no chunks, giving a constituent size of one) to full sentences (constituent size = 12). Intermediate strings had constituent sizes of two, three, four, and six words (an example with four-word chunks would be “mayor of the city/he hates this color/they read their names”). The prediction was that activation in a region specifically encoding constituent structure would increase with greater constituent size. Such a correlation was indeed found, including both IFG structures and temporal lobe areas, especially along the superior temporal sulcus.

To get at syntax specificity, a second manipulation involved substituting all content words with nonsense words to create “jabberwocky” constituents of varying sizes.53 For example, the four-word jabberwocky could read “tuyor of the roty/he futes this dator/they gead their wames.” This parametric manipulation removes the effect of semantic integration, but again elicited specific correlated activation in the IFG (centered on BA 45). The clever design controls for both semantics and word transition probability (because the nonsense words, by definition, have zero probability), and confirmed a specific role for Broca's area in processing hierarchical constituent structure in language.

Broca's region is involved in processing musical harmonic syntax

Regarding specificity of function, however, there has been considerable debate about the degree to which the function of Broca's area is specific to syntax, or whether it performs one or more computational functions that are shared by multiple aspects of language, including phonology or semantics,50 or even more broad cognition including cognitive control, action planning, and music.35,5456

The best characterized, and to us most convincing, nonlinguistic function of the IFG is in music processing and memory.5760 There is a long tradition, based on dissociations in lesion patients, of considering music and language to be instantiated in distinct brain regions.61 Initial evidence for a shared role for the IFG in musical processing, particularly processing harmonic syntax, came from event-related potential (ERP) and magnetoencephalography (MEG) studies.36,37,62 Furthermore, there are multiple studies indicating overlap of both cognitive and neural resources used in music-syntactic and language-syntactic processing based on interactions between the two domains.6365 These studies showed consistent similarities in early, anteriorly located reactions to anomalous items in linguistic and musical stimuli, and the MEG study by Maess et al.37 concluded simply that “musical syntax is processed in Broca's area.”

These results, from normal subjects, stand in contrast to the long clinical tradition identifying patients with damage to music and not language, or vice versa.61 Recent commentators57,58 have reconciled this apparent discrepancy by noting that although certain abstract operations may be shared by music and language, the units over which these functions operate are clearly quite different (phonemes and words for language, notes and chords for music). Furthermore, there are fundamental differences among different components of music. For example, a key component of musical rhythm is isochronicity—the tendency of key elements to be equally spaced in time.66 This is not an aspect of rhythm shared by normal speech or action, although other aspects like metrical structure may represent hierarchical sequencing.67 Given these basic differences, we do not expect complete overlap of musical and linguistic processing, either neurally or behaviorally. Rather, specific parallels in the manipulations applied to musical and linguistic stimuli (e.g., violations of hierarchical structural expectations33) are needed to elicit parallel neural effects.

Despite the broad consensus concerning a role for the IFG in these aspects of music processing, both ERP and MEG studies have difficulties with localization, and these early findings have not gone unchallenged by more recent functional magnetic resonance imaging (fMRI) studies.68,69 A particular issue concerns group analysis of different subjects. It is possible that each individual in a study could activate distinct but neighboring regions for linguistic versus musical syntax. Because of individual variability in precise localization, averaging all these individually distinct activations would give a false impression of thorough overlap in the two domains.

A recent study probed this issue by first identifying language regions in individual subjects, and then examining activations elicited in those regions by various other tasks involving working memory, cognitive control, and music.70 This study used a sentence-reading task, contrasted with reading nonsense words, to delineate the language regions of interest (ROIs). Although the study found considerable specificity to language in IFG activations, a musical task also preferentially activated their language-based ROIs, in particular overlapping with activations found in the left posterior IFG for both the sentence-reading task and a verbal working-memory task. Although the authors concluded from their results that specific language areas exist, not activated by most other tasks, this is an overly general conclusion applicable only to their specific musical and linguistic tasks. It is also important to recognize that the Patel or Koelsch shared-resource models57,58 do not predict exact overlap for these two different tasks (reading visually presented sentences versus listening to melodies) given the different input modalities. Thus, the significant language-area activations that were seen for the musical task support Lashley's contention that these two domains share important processing subsystems.

A second source of apparently contradictory evidence involves rhythm perception. A series of recent imaging studies concerning rhythm perception paint a consistent picture:7173 rhythmic stimuli tend to activate motor areas (such as the supplementary motor area, other dorsal prefrontal regions, or even the basal ganglia) involved in sequencing,74,75 and do not activate the IFG. But this work focuses on isochronous rhythms, and likely reflects the tendency of such rhythms to induce a desire to move to the regular beat—the so-called “groove” of the music.76,77 Precisely because this aspect of music is not shared with language, we do not consider such data to contradict Lashley's hypothesis or the well-established notion that the IFG plays a role in processing other components of musical syntax.

In summary, a considerable body of neuroimaging data, from many laboratories, is consistent with the hypothesis that the prefrontal cortex, and the IFG in particular, play important parallel roles in processing specific shared aspects of language and music. These data are consistent with both Lashley's early ideas and Patel and Koelsch's more specific hypotheses about the nature of the shared processing resources.

The role of frontocortical mirror neurons in understanding action

The third component of Lashley's triad is goal-directed action. There has been an explosion of interest recently in the role of the frontal lobes, and the IFG, in action understanding. This work typically makes no reference to Lashley's ideas, but was instead spurred by the discovery of mirror neurons in the macaque. Mirror neurons are neurons in motor and premotor cortex that fire, not only when a monkey executes an action itself, but also when it sees another agent perform the same action. The dual nature of the metaphoric mirror—in both production and perception—has led to revived interest in the mostly discredited motor theory of speech perception, which posits that speech perception is mediated by virtue of a resonance with the motor programs that generate the sounds (vocal gestures).78 By this hypothesis, the objects of speech perception are articulatory motor events, rather than acoustic or auditory events (for a critique, see Ref. 79).

Recent speculation concerning mirror neurons has extended the reach of the motor theory beyond its original domain of speech perception to language more generally. The region of macaque cortex where mirror neurons are found—F5—neighbors or overlaps the regions constituting the human Broca's area (BA 44/45). Many studies show activation of motor and premotor regions during language production and comprehension tasks,80,81 thought to support the notion that mirror neurons in this region play a central role in the parity of production and perception that typifies all aspects of language.8284 This idea has been used to support gestural hypotheses for language evolution, whereby initial stages of linguistic syntax and semantics evolved in the context of a gestural protolanguage, rather than a vocal system.83,8587

A detailed critique of mirror neurons or their possible importance in language evolution is not our purpose here (but see Refs. 8892). However, it is crucial to distinguish Lashley's hypothesis about action syntax from those linked to mirror neurons or embodiment. Lashley's model concerns a general and abstract notion of sequential hierarchy, and is only tied to motor actions insofar as these demand planning and control of a particular type. Lashley's model has nothing to do with grasping or recognizing graspable objects, and indeed involves structuring entities for which no such low-level action schemata are available. When applied to the role of the IFG, this leads to different empirical predictions from those of mirror-system hypotheses.

Action syntax: beyond mirror neurons

Because localization is central to mirror-based arguments, precision regarding specific regions of the frontal cortex is key to the following discussion. The cytoarchitecture and connectivity of neighboring frontal regions changes drastically over small distances. These properties range from agranular cortex in motor and premotor regions (BA 4 and 6), to dysgranular BA 44, to granular BA 45 and anterior frontal cortex. The motor cortex generates strong descending motor pathways (e.g., the corticospinal tract), but BA 44/45 are both truly prefrontal, and clearly functionally distinct from motor regions. With broad or imprecise spatial localization, or multisubject averaging of activations, it is easy to confuse (or conflate) activations in one of these regions with that of the others.

Starting with motor and premotor activations, brain-imaging studies reveal clear and consistent activations for certain classes of words or images connoting either direct actions (“throw” or “kick”) or tools providing affordances for action (“hammers,” “pliers”). Such motor-charged stimuli elicit (pre)motor activation; other more visually charged stimuli (e.g., animals) do not.80,91 Recent studies35 showing motor activation for particular components of language provide no special evidence for motor (or mirror neuron) involvement with language more broadly, because many linguistically structured concepts (color, clouds, beauty) have no motor component. Second, those few mirror neurons that have been directly documented in the human brain (during single-unit recording in 21 preoperative epilepsy patients) were found in the medial frontal (supplementary motor area) and the temporal (parahippocampal) cortex;93 it is therefore possible that mirror neurons are much more widespread in human brains than in macaques. Simple observation of IFG activation in the human brain therefore does not specifically implicate mirror neurons, and more direct and precise tests are needed to distinguish between general frontal involvement and mirror neuron involvement.

A key testable prediction of mirror neuron-based hypotheses follows from the fact that, by definition, mirror neurons only exist for motor actions the subject can actually perform itself. To the extent that mirror neurons are required to understand or process some observed action, this predicts that unfamiliar or unproduceable actions should not excite the mirror system, and should be uninterpretable (or at least difficult to process). This is rather difficult to test in the domain of language, because all normal humans are both producers and perceivers of language. Nonetheless, in brain-damaged patients with lesions in proposed sites of mirror neurons (the IFG and parietal regions), speech perception is not disturbed as predicted.94 This study only found impaired speech perception when lesions involved auditory regions in the temporal lobe, inconsistent with motor theories of speech perception and of the need for mirror neurons to support accurate speech perception.

Additional data do not support the prediction that motor familiarity is required to elicit appropriate processing of speech stimuli. An infant optical imaging study95 found frontal activation to speech-based patterns in newborns, despite their lack of experience in producing speech motor patterns themselves. Similarly, an artificial grammar-learning study by Bahlmann et al. employed abstract visual patterns, sequentially presented, which had no motor (or verbal) component whatsoever. Items were arranged into either context-free patterns favoring hierarchy or finite-state patterns favoring pure sequence. The authors found activation of BA 44 only for the first case, supporting a role for this region in sequential hierarchy in the absence of motor-action understanding.40

Perhaps the most direct test of the mirroring prediction is possible with music, where many humans are devoted music listeners but lack the motor capability to play this music themselves. Mirror neuron enthusiasts often cite studies indicating motor involvement and Broca's activation during music perception by expert musicians on their own instruments.35 But the crucial test is actually with nonmusicians, where the data are clear: Broca's activation is found during music perception, and musical syntax tasks in particular, even in nonmusicians, whose understanding of instrumental music presumably has little or no motor component.37,96 Furthermore, although many findings are consistent with the general notion that music perception relies partially on motor and action-related circuitry,60,72,75 most of the motor areas implicated are neither part of the putative mirror neuron system, nor specifically related to music production (e.g., singing or playing).

The dangers of rough localization and characterization are illustrated in an intriguing paper by Fazio et al.,54 which examined the abilities of patients with frontal aphasia to correctly understand event sequences. Previous work81 showed a particular deficit for prefrontally damaged patients for interpreting action. The Fazio study expanded this by first showing participants short video sequences of human actions or physical events and then presenting still images extracted from the videos on a touch screen. The patients’ task was to properly sequence these images by touching them (requiring no overt language). The six patients in this study showed a general deficit relative to healthy controls, but were more impaired for sequencing human actions (e.g., a man grasping a bottle) than physical events (e.g., a bicycle falling down). Superimposing the MRI-determined lesion locations of patients upon one another revealed that the only region consistently damaged in all patients was centered on BA 44. However, all but one of these patients also had much more extensive damage, often stretching back into the parietal lobe. Thus, although intriguing, these results do not support the strong inferences some draw about a specific role of Broca's area in this task,35 which involves action sequencing but not hierarchical sequencing per se.

Broca's area and hierarchical action planning

Much more convincing evidence for a specific role for Broca's area in hierarchical action, beyond music and language, comes from an elegant fMRI study by Koechlin and Jubault,97 which introduced a task designed to discriminate simple sequencing from hierarchical sequencing. This work built upon the computationally explicit model of hierarchical action planning of Dehaene and Changeux,98 which recognizes that at least three levels of nesting are needed to clearly discriminate simple gestures, motor sequences, and hierarchical plans. This model is based on the “Tower of London” task introduced by Shallice,42 which involves moving colored beads on three posts from some starting state to an end state designated with a target image. This task can be smoothly varied in complexity, from an easy variant where the beads are simply moved directly to the goal, to challenging end states that require five or more moves, some of which require moving beads away from their final goal state.99 Shallice found that patients with left prefrontal damage have difficulties specifically with these more complex problems,42 and brain-imaging results are broadly consistent with prefrontal involvement in this task.100 Dehaene's and Changeux's model specifies three levels of action: simple motor gestures (e.g., grasping a bead), operations (e.g., moving a bead from one post to another), and plans (nested sequences of operations to reach some specified goal state). Only the planning level provides clear evidence of hierarchical (nonchained) sequencing, particularly when it involves intermediate operations leading away from the goal. When the computational model was “lesioned” by removing planning units, these more difficult problems were specifically impaired, just as in Shallice's prefrontal patients.

Koechlin and Jubault96 designed an analogous task. Repeated button presses (motor condition) or simple, overlearned left—right sequences are produced in the scanner, under different stimulus conditions. In the simple chunk condition, overlearned sequences alternated with the motor control sequencing task. But in the superordinate condition, chunk sequences were arranged into higher level orders. All sequences were highly overlearned and automatized, so no learning occurred during scanning. Although simple motor acts elicited motor and premotor activity, higher order actions elicited more anterior activations within Broca's area, with transitions between chunks favoring BA 44, and termination of superordinate chunks activating BA 45. These data provide the clearest evidence to date of a specific role for Broca's area in planning hierarchically structured action.

We conclude that neuroimaging data support Lashley's hypothesis that music, language, and some types of action planning share a common substrate, and that Broca's area plays a key role in this distributed system. Although none of these data exclude a role for mirror neurons in action understanding in music or language, they argue against a necessity for motor knowledge to either understand sensory input or to elicit specific activity in the IFG. In contrast, these results remain consistent with Lashley's more abstract proposals, on the basis of computational similarities underlying planning of complex action in all three domains. The burning question thus remains concerning what, precisely, this supramodal system is doing at a computational level.

Broca's area as Lashley's “scannable buffer”

Lashley emphasized that the core requirement of sequential hierarchy is that higher level processes and goals can be placed on hold during the execution of lower level processes. To accomplish such persistence, he suggested, certain neural circuits must act as a buffer scannable by other circuits. In computational terms, sequential hierarchies require the equivalent of a register (holding one element), a queue (multiple freely accessible elements) or a stack (holding multiple elements, but with restricted last-in-first-out access), and require that these forms of persistent memory be available to other ongoing computational processes. Psychologically, these are all different forms of working memory that provide temporary storage of intermediate results (analogous to a scratchpad or blackboard). Although a reflex chain of actions, where each action triggers the next, does not require such persistent intermediate storage, both processing and production of hierarchical sequences do. Furthermore, the more complex and multileveled the hierarchy, the greater the capacity of the intermediate storage mechanism must be.

From a computational viewpoint, there is no doubt that effective, succinct computation requires such additional persistent storage. The most common means of achieving it in computer science is via a stack. For example, during the execution of complex computer programs, each call to a subroutine places a pointer to the current routine (and often intermediate results) onto the stack before branching to execute the subroutine. Once the subroutine completes its computation, typically returning some required result, execution of the original calling function can resume where it left off, by popping its pointer back off the stack. This can be embedded multiple times (with each level of nesting requiring additional stack space), and clearly the storage capacity of the stack limits the possible extent of embedding.

Both this storage capacity and restrictions on access influence the types of computations that are possible. Theoretical computer scientists have mapped out, in considerable detail, the computational limitations of registers versus stacks versus queues in abstract models of computation called automata. For instance, a finite-state automaton (FSA) is a system where discrete computational states are linked with no persistent memory. When such a system is augmented with additional working memory structures, new automata classes result. A push-down automaton (PDA), for example, is an FSA with an additional stack provided, whereas a Turing machine is an FSA with an endless queue (unrestricted order of access). Automata are models of processing, but important theorems link each of these types to specific classes of string sets (“languages”) that can be generated or recognized by the corresponding type of automaton. Unsurprisingly, the less restricted the additional memory store of the automaton, the broader and less restricted the type of string sets it can handle. These automaton/language pairs can thus be arranged into a classification system, termed the formal language hierarchy or extended Chomsky hierarchy:101 a fundamental theoretical concept in computer science.102104

One appealing version of such stack-based computing would involve domain-specific templates and filler information, retrieved and maintained by a multidomain processor. For example the sentence “John likes apples” would involve retrieving the template “X likes Y,” along with the specific elements “John” and “apples” (which might just as well be “Mary” and “chocolate”). The role of the scannable buffer here is simply to maintain the current template-element bindings during the processing of further perceptual information or motor planning. Such a mechanism could be used to deal with many types of templates, independent of format (e.g., verbal, melodic, and motor), and thus would be available for multiple domains.

Although the distinctions between registers, stacks, and queues are a major focus in computer science, from a psychological or neural point of view the relevance of such abstractions is less obvious. Lashley's critical point was that hierarchical sequence processing requires some form of scannable intermediate storage, and limitations on this capacity obviously restrict the capacity of a system to handle hierarchical structures. Formally, the crucial distinction is between FSA-only systems (which can generate the so-called regular languages and have only a very limited and inflexible capacity to process hierarchy) and the augmented automata above these (e.g., PDAs or Turing machines), which correspond to the supraregular languages.

A foundational result in computational linguistics is that natural languages require processing resources over and above those of an FSA, and are thus supraregular.105,106 Considerable debate has surrounded the question of how far beyond regular the natural languages go; current consensus is that languages go a bit beyond the capacities of a PDA and belong to the class termed mildly context-sensitive grammars,107,108 requiring multiple stacks, or even a stack of stacks to be computed. With music, the situation is less clear, but some theorists suggest that music requires computational supraregular resources including at least one stack109—the so-called context-free grammars.

Marrying these computational considerations with our prior conclusions on the basis of neural localization leads to the following hypothesis. Cortical resources in the IFG, including at least BA 44 and BA 45, implement a storage buffer scannable by other cortical and subcortical circuits subserving sequential behavior. This buffer is required to implement supraregular hierarchical sequence processing, and its processing load increases with the depth and complexity of the hierarchy being processed.

This hypothesis, which has been entertained in related forms by multiple researchers,52,97,110 extends Lashley's speculations by positing a discrete neural locus implementing his scannable buffer and by specifying its supraregular nature. Direct evidence that processing load increases with depth is provided by parametric studies,53,111,112 and the localization aspect to the IFG is consistent with a wealth of neuroimaging data. This hypothesis does not require that the IFG is the only site of a scannable neural buffer (similar buffers may be implemented, for example, in hippocampus or basal ganglia circuitry). Nor does this hypothesis suggest that the stack or buffer represents all of the relevant data or computations: the idea is that it can activate and link representations generated in other brain regions (akin in computer terms to storing a pointer, rather than all of the data). Nonetheless, the involvement of Broca's region in hierarchical sequencing in music and language, and in some complex action planning, seems abundantly clear. Koelsch has termed this idea the syntactic equivalence hypothesis.113

This hypothesis does not follow directly from the fact that cortical systems for both perception and action are laid out hierarchically. For if each level of a processing hierarchy mapped directly onto a neural hierarchy, implemented on the cortical surface, there would be strict limitations on the possible depth of hierarchical structures. For example, the data on hierarchical action planning above suggested a direct mapping from first-level motor gestures in posterior (motor and premotor) cortex and third-level superordinate planning to anterior (BA 45) regions. But prefrontal cortex cannot be extended forever, and a literal mapping would suggest a strict limit on the level of embedding possible in language, music, or action. Such a limit is not evident in action (even making coffee can be construed as involving four levels, and if interrupted by a phone call, five), nor in music or language. The well-known restriction of embedding to three levels in language, for example, applies only to a particular syntactic structure—center embedding—not to syntax or language as a whole.114,115

Having thus extended Lashley's hypothesis to a specific neural circuit, we turn to Lashley's final hypothesis: that the capacity to process musical and linguistic hierarchical sequences is unusually developed in humans, but builds upon and extends preexisting action sequencing mechanisms present in animals. Addressing this issue demands an examination of comparative data.

The phylogenetic roots of hierarchical sequencing: comparative data

Many comparative studies have addressed animals’ ability to produce and perceive hierarchical structure,116 mostly in the auditory domain but also exploring visual capabilities.45,117 Much of the recent animal work has adopted the computational principles outlined above, contrasting finite-state grammars with context-free grammars to see which (if any) animal species can go beyond the regular grammar level.

This comparative work built upon human experimental psychology research begun by George Miller in the late 1950s, in one of the first computerized testing laboratories at Harvard.118 Rule-learning problems were generated and tested at different computational levels. The results led Miller to put forth the following hypothesis, which we call the supraregular hypothesis: adult humans have a proclivity to induce rule systems at context-free (or higher) levels, even when the data do not require such systems. That is, even stimulus sets that could be captured by regular grammars tend to be captured using supraregular grammars.

Miller's hypothesis suggests not only that humans are able to go beyond regular grammars, but also that human observers are biased to view sequences as hierarchically structured, even when the generating algorithm is a serial, finite-state system. Because all stimulus sets generated by finite-state grammars can also be parsed by context-free grammar, this may be a cognitive example of the aphorism “to a man with a hammer everything looks like a nail.” We might think of this bias as a form of dendrophilia—an inordinate fondness for tree structures—in our species.119

Both hierarchy and supraregularity are clear and well defined mathematically. Unfortunately, it is not trivial to adjudicate between finite-state and context-free requirements on the basis of string processing alone.120 Structure is an invisible (mental) construct, and can only be empirically probed indirectly (e.g., by observing acceptance and rejection of well-chosen string sets,121 by adding prosodic cues at or between putative phrase boundaries,122 or by observing reactions to clicks played at or between putative boundaries123). Such indirectness renders empirical testing of these notions challenging, but by no means impossible.

Returning to animal research, existing studies paint a relatively consistent picture. Many species, from rats to songbirds to baboons, have well-developed sequencing capabilities that can be characterized as finite state systems. Such systems, for example, appear adequate to capture the syntactic structure seen in birdsong116 or the recognition of simple long-distance dependency in squirrel monkeys.124 In contrast, the capacity of animals to induce supraregular grammars remains contentious. Despite numerous examples of successful induction of regular grammars, only a few species have been claimed to go beyond this,116 and both the data and the methods in these studies have been sharply questioned.125127 Perhaps the most convincing attempt to show supraregular abilities in a nonhuman species came from Gentner et al. in their operant conditioning work128 with starlings (Sturnus vulgaris), songbirds with complex song and extended vocal learning in both females and males.129 Gentner's et al. used a supraregular grammar denoted AnBn, which means that the number of A units must be precisely matched by the number of B units (in this case the A and B elements were two types of starling song motifs, termed warbles and rattles, presented acoustically). This grammar is provably supraregular, and allows multiple possible structural interpretations ranging from count-and-compare, to center-embedded, to cross-serial structure.121 After very extensive training (between 9400 and 56,200 trials per bird, with a mean of 30,000 trials), three of four starlings tested with various probe strings showed evidence of having acquired the supraregular language AnBn. This suggested that, although quite difficult, it was at least possible for some birds to acquire a supraregular rule. However, a follow-up study in zebra finches125 showed that such levels of performance could be explained by simpler rules and averaging together of the performance of different birds.

On the basis of these and other results (reviewed in Ref. 116), we can tentatively formulate a comparative extension of Miller's human-oriented supraregular hypothesis—the supraregular exceptionality hypothesis, which posits that, in contrast to humans, nonhuman animals display a particular difficulty inducing hierarchical patterns from strings of data, especially any structures requiring computational resources above the finite-state level.

Evidence that an organism fails on supraregular grammars (e.g., AnBn, mirror grammars, or copy grammars101), although succeeding on regular grammars (e.g., (AB)n, AB*A, or sequential transition–probability grammars130,131) constitutes support for this hypothesis. Unlike Miller's human-specific supraregular hypothesis, which has abundant support from human studies,40,118,132134 our supraregular exceptionality hypothesis, which concerns nonhuman animals, remains tentative for several reasons. First, nearly every comparative study has used a different species, and only a handful of nonhuman species have been tested to date. This renders broad generalizations about animals impossible. Even statements about one species remain tentative: multiple empirical issues need to be considered in pattern-learning experiments that make it difficult for any single study to control for all interpretations. Finally, very few grammars have been tested (indeed, all animal-based claims for supraregularity are based on a single grammar AnBn), which renders generalizations to broad formal classes impossible. Although humans succeed on multiple supraregular grammars (including copy and mirror grammars), to our knowledge no one has tested such grammars in nonhumans. Firm conclusions will require more research, on different species and different grammars, and replications across laboratories.

Nonetheless, these comparative data provide initial support for Lashley's notion that a hierarchical syntax of action is particularly well developed in our species. It seems too early to draw conclusions concerning the degree to which this proclivity builds upon preexisting supraregular abilities in animals.135137 Certainly, we have no evidence of action syntax in animals even approaching the complexity of making a pot of coffee: the closest would be certain instances of tool use in some chimpanzee populations.138 Before returning to the possible neural basis for expanded hierarchical abilities in humans, we need to briefly forestall misunderstanding by noting an area of pervasive confusion in the recent comparative literature.

Recursion and supraregularity: an unfortunate confusion

For unclear reasons much of the comparative literature just discussed conflates several distinct issues, mixing up multiple computational abilities as if they were one. Our concern in this review is the processing of hierarchical sequences, defined as sequences with a tree-formed structure in which element order matters. Such structures are best processed by supraregular systems. Unfortunately, much of the last decade's research based on formal language theory has conflated hierarchy and supraregularity, in these precise senses, with various notions of recursion. Recursion refers to that broad class of computational processes in which a function calls itself (in computer science), or to those structures where the same hierarchical structure is repeated at multiple levels of the hierarchy (“self-embedding”). An example of a recursive structure is a fractal, where the same structure repeats itself ad infinitum at all hierarchical levels.139

Crucially, although all recursive trees are hierarchical, all hierarchies are not recursive. For many types of hierarchy discussed in this review, such as motor actions embedded within plans, it is unclear what self-embedding would even mean. Although a context-free grammar can be implemented recursively, for finite sets it need not be, and regular grammars can also be implemented recursively. The well-defined supraregular divide, between finite-state automata and those supraregular automata that have some additional scannable memory, is thus orthogonal to the issue of recursion. This issue has now been repeatedly discussed in the literature,121,140144 so we will not belabor it here. We can only hope that the unfortunate current tendency to substitute the ambiguous high-profile buzzword recursive for distinct (and more empirically tractable) concepts like supraregular or hierarchical will eventually run its course.

Evolutionary expansion of Broca's region in humans

We end with a brief review of recent comparative anatomical data pointing to a general expansion of prefrontal connectivity in our species, along with a more specific and pronounced increase in the size of Broca's area and changes in its pattern of connections.

The human brain is greatly expanded in size relative to other primates, including our nearest relatives the great apes (chimpanzees, gorillas, and orangutans).145 However, ours are not the largest brains in the animal kingdom: the brains of elephants and toothed whales like dolphins or orcas are larger.146 Nor do human brains occupy the greatest relative proportion of total body weight: that distinction goes to small mammals like shrews.147,148 It is only when brain size is considered relative to what would be predicted, given body size (the so-called encephalization quotient), that human brains reign supreme.149 In these terms, although great apes have relatively large brains for mammals, humans are still quite exceptional, with a roughly threefold increase in total brain size relative to chimpanzees (or early fossil hominids like Australopithecus150).

Recent comparative neuroanatomical work allows us to go beyond these long-known facts to inquire whether human brain expansion affected specific brain regions and circuits. Although human frontal regions, like the cortex overall, are much larger than in other primates, whether this increase is disproportionate to overall size increase remains contentious. One of the first studies to use MRI in living primates to address this issue151 compared the volume of the frontal lobe in four great ape species with humans, along with lesser ape (gibbon) and monkey (Macaca and Cebus) brains. They concluded that great apes in general have an expanded frontal cortex relative to gibbons or monkeys, but frontal lobe volume was not disproportionately different from humans relative to total brain volume. This is consistent with many findings that apes are cognitively superior to monkeys.152 However, this study looked at the entire frontal lobe, as demarcated by the central sulcus (thus including both motor and anterior areas).

To examine the relative size of the prefrontal cortex specifically, Schoeneman et al. used MRIs of 11 primate species to examine the brain volume anterior to the corpus callosum.153 A separate analysis of gray and white matter found a striking dissociation between these two components: human relative gray volumes differed significantly from only a few species, whereas white volumes were greater than all but two species. A subsequent analysis of gray:white size ratios indicates that any human expansion of the prefrontal cortex relative to total brain volume is almost entirely the result of an increase in white matter. This suggests that, in addition to their raw increase in absolute size, human prefrontal regions have become disproportionately connected to the rest of cortex. Whether human frontal lobes (gray or white) are significantly and disproportionately larger than predicted depends on the comparison group (apes or monkeys). But differences from nonape primates are established, whereas differences from apes remain controversial.154157 Furthermore, absolute frontal volume increase is undisputed, and relative white matter increases seem relatively clear.

To what extent do these increases in size and connectivity apply equally to all prefrontal regions? This question is challenging to answer, because it requires that multiple brains are analyzed using detailed cytoarchitectonic methods. In apes, such analysis has to date been performed only for prefrontal areas BA 10 (frontal pole), 44, and 45 (along with V1 and BA 13, part of the insula). Comparisons between the brains of 12 chimpanzees158 with previous work in humans159 revealed a striking finding: areas 44 and 45 are the most greatly expanded cortical areas yet identified in humans. Left areas 44 and 45 are six times larger in humans than in chimpanzees,158 disproportionate to the roughly threefold increase in total brain size, or the 4.5-fold increase in frontal cortex in total. This important study shows that Broca's area in particular has expanded disproportionately since our divergence from chimpanzees (roughly 6 million years ago).

Regarding connectivity, a recent study exploited the technique of diffusion tensor imaging (DTI), which uses MRI to estimate, in intact brains, axonal connectivity between brain regions.160 Applying DTI to human, macaque, and chimpanzee brains, Rilling et al.161 found significant changes in the pattern of connectivity between Broca's area and posterior brain regions. Humans, and to some degree chimpanzees, showed connectivity between prefrontal and temporal regions via a dorsal pathway through the parietal cortex, but only in humans did this pathway (the arcuate fasiculus) have a strong uninterrupted connection to the posterior temporal cortex. In contrast, rhesus macaques showed a very weak dorsal pathway, whereas ventral connections between the temporal cortex and frontal regions via the insula dominated. These results again converge on the conclusion that it is not simply the size of prefrontal regions that has increased in humans; the pattern of connectivity has changed as well.162

Conclusion: Broca meets Lashley

Combining the behaviorally anomalous status of human music and language relative to animal communication systems with the comparative anatomical data and imaging data reviewed above, we begin to see the outlines of a neurally and biologically grounded hypothesis concerning the undeniable differences between humans and other animals that does justice to the very deep cognitive and neural foundations that we share with other vertebrates (including other primates). For each of Lashley's three questions, laid out in the introduction, current data support a tentative positive answer: processing hierarchical sequences appears to be a well-defined and neutrally localizable function, shared by musical and linguistic syntax (Q1), humans are unusually well developed in this ability (Q3), and it is at least plausible to suggest that this ability may inherit key components from some type of action syntax that predated the evolution of human music and language (Q2).

From a modern perspective, Lashley's action syntax can be localized to a set of widespread brain circuits that have, as a key hub, prefrontal regions centered on Broca's region. The prefrontal regions play a primitive role in the hierarchical planning and sequencing of action, a function presumably shared with other primates, and particularly with chimpanzees, whose consistent and intelligent use of tools has no peer among other nonhuman primates.163165 However, the apparent remit of this type of hierarchical planning increased greatly during human evolution, to include both perception and production of all types of hierarchical sequences. The most prominent additions to this Broca-centered action sequencing capacity were, by hypothesis, those two great human achievements: music and language.

From an evolutionary viewpoint, this idea provides for a certain continuity between humans and other primates (particularly chimpanzees) while acknowledging the drastic increases in certain cognitive capacities that have occurred since our divergence with chimpanzees. They also make sense of the fact that although prefrontal cortex and Broca's area are not novel brain regions, they have expanded disproportionately and have changed their patterns of connectivity to other brain regions during recent human evolution. In proposing this, we are fully aware that Broca's region is by no means the sole seat of language or music. Both capacities rely on a far-flung network of both cortical and subcortical brain regions (all of them, again, shared with other primates). But Lashley's ideas bring into focus the precise cognitive and computational changes that might underlie our expanded capacities for planning, action, language, music, and thought itself, in the context of a wide array of shared capacities. This is, by hypothesis, our broadly developed ability and indeed proclivity to structure both action and perception hierarchically. The clear distinctions between currently popular mirror neuron-based hypotheses and the more abstract computational hypothesis of Lashley have the potential to drive more refined experimental procedures and to provide further empirical evidence relevant to understanding Broca's region and its function. We conclude that, although Lashley has been rarely cited in the large number of studies implicating Broca's region in action, music, and language, his ideas have withstood the passage of time well, and deserve renewed attention.

Acknowledgments

We thank Daniel Bowling, Bruno Gingras, Stefan Koelsch, Marisa Hoeschele, and two anonymous reviewers for comments on a previous version of this manuscript. We also acknowledge support from ERC Advanced Grant #230604 “SOMACCA” and University of Vienna Research Cluster Grant “Shared Neural Resources” (to W.T.F.), and FCT Grant SFRH/BD/64206/2009 (M.M.).

Conflicts of interest

The authors declare no conflicts of interest.

References

  • 1.Lashley K. The problem of serial order in behavior. In: Jeffress LA, editor. Cerebral Mechanisms in Behavior; the Hixon Symposium. New York: Wiley; 1951. pp. 112–146. [Google Scholar]
  • 2.Rosenbaum DA, et al. The problem of serial order in behavior: Lashley's legacy. Hum. Mov. Sci. 2007;26:525–554. doi: 10.1016/j.humov.2007.04.001. [DOI] [PubMed] [Google Scholar]
  • 3.Chomsky N. A note on phrase structure grammars. Inform. Control. 1959;2:393–395. [Google Scholar]
  • 4.Chomsky N. Language and Mind. New York: Harcourt, Brace & World; 1968. [Google Scholar]
  • 5.Upshall M. Hutchinson Dictionary of Mathematics. London: Brockhampton Press; 1993. [Google Scholar]
  • 6.Illingworth V. Dictinary of Computing. Oxford: Oxford University Press; 1983. [Google Scholar]
  • 7.Altmann CF, Bülthoff HH. Kourtzi Z. Perceptual organization of local elements into global shapes in the human visual cortex. Curr. Biol. 2003;13:342–349. doi: 10.1016/s0960-9822(03)00052-6. [DOI] [PubMed] [Google Scholar]
  • 8.Aminoff E, Gronau N. Bar M. The parahippocampal cortex mediates spatial and nonspatial associations. Cereb. Cortex. 2006;17:1493–1503. doi: 10.1093/cercor/bhl078. [DOI] [PubMed] [Google Scholar]
  • 9.Badre D. Cognitive control, hierarchy, and the rostro–caudal organization of the frontal lobes. Trends Cogn. Sci. 2008;12:193–200. doi: 10.1016/j.tics.2008.02.004. [DOI] [PubMed] [Google Scholar]
  • 10.Badre D. D'Esposito M. Is the rostro-caudal axis of the frontal lobe hierarchical? Nat. Rev. Neurosci. 2009;10:659–669. doi: 10.1038/nrn2667. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Badre D, et al. Hierarchical cognitive control deficits following damage to the human frontal lobe. Nat. Neurosci. 2009;12:515–522. doi: 10.1038/nn.2277. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Kikuchi Y, Horwitz B. Mishkin M. Hierarchical auditory processing directed rostrally along the monkey's supratemporal plane. J. Neurosci. 2010;30:13021–13030. doi: 10.1523/JNEUROSCI.2267-10.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Kourtzi Z, et al. Integration of local features into global shapes: monkey and human fMRI studies. Neuron. 2003;37:333–346. doi: 10.1016/s0896-6273(02)01174-1. [DOI] [PubMed] [Google Scholar]
  • 14.Kravitz DJ, et al. The ventral visual pathway: an expanded neural framework for the processing of object quality. Trends Cogn. Sci. 2013;17:26–49. doi: 10.1016/j.tics.2012.10.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Krumbholz K, et al. Hierarchical processing of sound location and motion in the human brainstem and planum temporale. Eur. J. Neurosci. 2005;21:230–238. doi: 10.1111/j.1460-9568.2004.03836.x. [DOI] [PubMed] [Google Scholar]
  • 16.Mormann F, et al. Latency and selectivity of single neurons indicate hierarchical processing in the human medial temporal lobe. J. Neurosci. 2008;28:8865–8872. doi: 10.1523/JNEUROSCI.1640-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Kravitz DJ, et al. A new neural framework for visuospatial processing. Nat. Rev. Neurosci. 2011;12:217–230. doi: 10.1038/nrn3008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Kumaran D, Melo HL. Duzel E. The emergence and representation of knowledge about social and nonsocial hierarchies. Neuron. 2012;76:653–666. doi: 10.1016/j.neuron.2012.09.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Schendan HE, et al. An fMRI study of the role of the medial temporal lobe in implicit and explicit sequence learning. Neuron. 2003;37:1013–1025. doi: 10.1016/s0896-6273(03)00123-5. [DOI] [PubMed] [Google Scholar]
  • 20.Meyer P, et al. Language processing within the human medial temporal lobe. Hippocampus. 2005;15:451–459. doi: 10.1002/hipo.20070. [DOI] [PubMed] [Google Scholar]
  • 21.Opitz B. Friederici AD. Interactions of the hippocampal system and the prefrontal cortex in learning language-like rules. Neuroimage. 2003;19:1730–1737. doi: 10.1016/s1053-8119(03)00170-8. [DOI] [PubMed] [Google Scholar]
  • 22.Opitz B. Friederici AD. Neural basis of processing sequential and hierarchical syntactic structures. Hum. Brain Mapping. 2007;28:585–592. doi: 10.1002/hbm.20287. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Eichenbaum H, et al. Towards a functional organization of episodic memory in the medial temporal lobe. Neurosci. Biobehav. Rev. 2012;36:1597–1608. doi: 10.1016/j.neubiorev.2011.07.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Ranganath C. A unified framework for the functional organization of the medial temporal lobes and the phenomenology of episodic memory. Hippocampus. 2010;20:1263–1290. doi: 10.1002/hipo.20852. [DOI] [PubMed] [Google Scholar]
  • 25.Hassabis D, et al. Patients with hippocampal amnesia cannot imagine new experiences. Proc. Natl. Acad. Sci. U S A. 2007;104:1726–1731. doi: 10.1073/pnas.0610561104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Baddeley A. The episodic buffer: a new component of working memory. Trends Cogn. Sci. 2000;4:417–423. doi: 10.1016/s1364-6613(00)01538-2. [DOI] [PubMed] [Google Scholar]
  • 27.Henke K. A model for memory systems based on processing modes rather than consciousness. Nat. Rev. Neurosci. 2010;11:523–532. doi: 10.1038/nrn2850. [DOI] [PubMed] [Google Scholar]
  • 28.Amunts K, et al. Broca's region: novel organizational principles and multiple receptor mapping. PLoS Biol. 2010;8:1000489. doi: 10.1371/journal.pbio.1000489. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Friederici AD. Towards a neural basis of auditory sentence processing. Trends Cogn. Sci. 2002;6:78–84. doi: 10.1016/s1364-6613(00)01839-8. [DOI] [PubMed] [Google Scholar]
  • 30.Friederici AD. The brain basis of language processing: from structure to function. Physiol. Rev. 2011;91:1357–1392. doi: 10.1152/physrev.00006.2011. [DOI] [PubMed] [Google Scholar]
  • 31.Petersson KM, Forkstam C. Ingvar M. Artificial syntactic violations activate Broca's region. Cogn. Sci. 2004;28:383–407. [Google Scholar]
  • 32.Forkstam C, et al. Neural correlates of artificial syntactic structure classification. Neuroimage. 2006;32:956–967. doi: 10.1016/j.neuroimage.2006.03.057. [DOI] [PubMed] [Google Scholar]
  • 33.Koelsch S, et al. Processing of hierarchical syntactic structure in music. Proc. Natl. Acad. Sci. U S A. 2013;110:15443–15448. doi: 10.1073/pnas.1300272110. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Brown S, Martinez MJ. Parsons LM. Music and language side by side in the brain: a PET study of the generation of melodies and sentences. Eur. J. Neurosci. 2006;23:2791–2803. doi: 10.1111/j.1460-9568.2006.04785.x. [DOI] [PubMed] [Google Scholar]
  • 35.Fadiga L, Craighero L. D'Ausilio A. Broca's area in language, action, and music. Ann. New York Acad. Sci. 2009;1169:448–458. doi: 10.1111/j.1749-6632.2009.04582.x. [DOI] [PubMed] [Google Scholar]
  • 36.Koelsch S, Maess B. Friederici AD. Musical syntax is processed in the area of Broca: an MEG study. Neuroimage. 2000;11:56. doi: 10.1038/87502. [DOI] [PubMed] [Google Scholar]
  • 37.Maess B, et al. Musical syntax is processed in Broca's area: an MEG study. Nat.Neurosci. 2001;4:540–545. doi: 10.1038/87502. [DOI] [PubMed] [Google Scholar]
  • 38.Patel AD, et al. Musical syntactic processing in agrammatic Broca's aphasia. Aphasiology. 2008;22:776–789. [Google Scholar]
  • 39.Sammler D, Koelsch S. Friederici AD. Are left fronto-temporal brain areas a prerequisite for normal music-syntactic processing? Cortex. 2011;47:659–673. doi: 10.1016/j.cortex.2010.04.007. [DOI] [PubMed] [Google Scholar]
  • 40.Bahlmann J, et al. Neural circuits of hierarchical visuo-spatial sequence processing. Brain Res. 2009;1298:161–170. doi: 10.1016/j.brainres.2009.08.017. [DOI] [PubMed] [Google Scholar]
  • 41.Luria AR. Higher Cortical Functions in Man. New York: Basic Books; 1966. [Google Scholar]
  • 42.Shallice T. Specific impairments of planning. Philos. Trans. R. Soc. B. 1982;298:199–209. doi: 10.1098/rstb.1982.0082. [DOI] [PubMed] [Google Scholar]
  • 43.Baddeley AD. Working Memory. Oxford: Clarendon Press; 1986. [Google Scholar]
  • 44.Passingham RE. The frontal lobes and voluntary action. Oxford, UK: Oxford University Press; 1993. [Google Scholar]
  • 45.Conway CM. Christiansen MH. Sequential learning in non-human primates. Trends Cogn. Sci. 2001;5:539–546. doi: 10.1016/s1364-6613(00)01800-3. [DOI] [PubMed] [Google Scholar]
  • 46.Passingham RE. Broca's area and the origins of human vocal skill. Philos. Trans. R. Soc. B. 1981;292:167–175. doi: 10.1098/rstb.1981.0025. [DOI] [PubMed] [Google Scholar]
  • 47.Bookheimer S. Functional MRI of language: new approaches to understanding the cortical organization of semantic processing. Annu. Rev. Neurosci. 2002;25:151–188. doi: 10.1146/annurev.neuro.25.112701.142946. [DOI] [PubMed] [Google Scholar]
  • 48.Friederici AD. Pathways to language: fiber tracts in the human brain. Trends Cogn. Sci. 2009;13:175–181. doi: 10.1016/j.tics.2009.01.001. [DOI] [PubMed] [Google Scholar]
  • 49.Zurif ER, Caramazza A. Myerson R. Grammatical judgments of agrammatic aphasics. Neuropsychologia. 1972;10(4):405–417. doi: 10.1016/0028-3932(72)90003-6. [DOI] [PubMed] [Google Scholar]
  • 50.Hagoort P. On Broca, brain, and binding: a new framework. Trends Cogn. Sci. 2005;9:416–423. doi: 10.1016/j.tics.2005.07.004. [DOI] [PubMed] [Google Scholar]
  • 51.Hickok G. Poeppel D. The cortical organization of speech processing. Nat. Rev. Neurosci. 2007;8:393–402. doi: 10.1038/nrn2113. [DOI] [PubMed] [Google Scholar]
  • 52.Pulvermüller F. Brain embodiment of syntax and grammar: discrete combinatorial mechanisms spelt out in neuronal circuits. Brain Lang. 2010;112:167–179. doi: 10.1016/j.bandl.2009.08.002. [DOI] [PubMed] [Google Scholar]
  • 53.Pallier C, Devauchelle A-D. Dehaene S. Cortical representation of the constituent structure of sentences. Proc. Natl. Acad. Sci. U S A. 2011;108:2522–2527. doi: 10.1073/pnas.1018711108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Fazio P, et al. Encoding of human action in Broca's area. Brain. 2009;132:1980–1988. doi: 10.1093/brain/awp118. [DOI] [PubMed] [Google Scholar]
  • 55.Thompson-Schill SL. Dissecting the language organ: a new look at the role of Broca's area in language processing. In: Cutler A, editor. Twenty-First Century Psycholinguistics: Four Cornerstones. London: Lawrence Erlbaum; 2005. pp. 173–190. [Google Scholar]
  • 56.Thompson-Schill SL, et al. Role of left inferior prefrontal cortex in retrieval of semantic knowledge: a reevaluation. Proc. Natl. Acad. Sci. U S A. 1997;94:14792–14797. doi: 10.1073/pnas.94.26.14792. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Patel AD. Sharing and nonsharing of brain resources for language and music. In: Arbib MA, editor. Language, Music, and the Brain: A Mysterious Relationship. Cambridge, Massachusetts: MIT Press; 2013. pp. 329–355. [Google Scholar]
  • 58.Koelsch S. Neural correlates of music perception. In: Arbib MA, editor. Language, Music, and the Brain: A Mysterious Relationship. Vol. 10. Cambridge, Massachusetts: MIT Press; 2013. pp. 141–172. [Google Scholar]
  • 59.Janata P. Parsons LM. Neural mechanisms of music, singing, and dancing. In: Arbib MA, editor; Language, Music, and the Brain: A Mysterious Relationship. Vol. 10. Cambridge, Massachusetts: MIT Press; 2013. pp. 307–328. [Google Scholar]
  • 60.Herholz SC, Halpern AR. Zatorre RJ. Neuronal correlates of perception, imagery, and memory for familiar tunes. J. Cogn. Neurosci. 2012;24:1382–1397. doi: 10.1162/jocn_a_00216. [DOI] [PubMed] [Google Scholar]
  • 61.Peretz I. Coltheart M. Modularity of music processing. Nat. Neurosci. 2003;6:688–691. doi: 10.1038/nn1083. [DOI] [PubMed] [Google Scholar]
  • 62.Patel AD. Syntactic processing in language and music: different cognitive operations, similar neural resources. Musical Percept. 1998;16:27–42. [Google Scholar]
  • 63.Koelsch S, et al. Interaction between syntax processing in language and in music: an ERP study. J. Cogn. Neurosci. 2005;17:1565–1577. doi: 10.1162/089892905774597290. [DOI] [PubMed] [Google Scholar]
  • 64.Slevc LR, Rosenberg JC. Patel AD. Making psycholinguistics musical: self-paced reading time evidence for shared processing of linguistic and musical syntax. Psychon. Bull. Rev. 2009;16:374–381. doi: 10.3758/16.2.374. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Steinbeis N. Koelsch S. Shared neural resources between music and language indicate semantic processing of musical tension-resolution patterns. Cereb. Cortex. 2008;18:1169–1178. doi: 10.1093/cercor/bhm149. [DOI] [PubMed] [Google Scholar]
  • 66.Fitch WT. The biology and evolution of music: a comparative perspective. Cognition. 2006;100:173–215. doi: 10.1016/j.cognition.2005.11.009. [DOI] [PubMed] [Google Scholar]
  • 67.Fitch WT. Rhythmic cognition in humans and animals: distinguishing meter and pulse perception. Front. Syst. Neurosci. 2013;7:1–16. doi: 10.3389/fnsys.2013.00068. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Abrams DA, et al. Decoding temporal structure in music and speech relies on shared brain resources but elicits different fine-scale spatial patterns. Cereb. Cortex. 2011;21:1507–1518. doi: 10.1093/cercor/bhq198. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Rogalsky C, et al. Functional anatomy of language and music perception: temporal and structural factors investigated using functional magnetic resonance imaging. J. Neurosci. 2011;31:3843–3852. doi: 10.1523/JNEUROSCI.4515-10.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Fedorenko E, Behr MK. Kanwisher N. Functional specificity for high-level linguistic processing in the human brain. Proc. Natl. Acad. Sci. U S A. 2011;108:16428–16433. doi: 10.1073/pnas.1112937108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Chen JL, Zatorre RJ. Penhune VB. Moving on time: brain network for auditory-motor synchronization is modulated by rhythm complexity and musical training. J. Cogn. Neurosci. 2008;20:226–239. doi: 10.1162/jocn.2008.20018. [DOI] [PubMed] [Google Scholar]
  • 72.Grahn JA. Neural mechanisms of rhythm perception: current findings and future perspectives. Top. Cogn. Sci. 2012;4:585–606. doi: 10.1111/j.1756-8765.2012.01213.x. [DOI] [PubMed] [Google Scholar]
  • 73.Grahn JA. Brett M. Rhythm and beat perception in motor areas of the brain. J. Cogn. Neurosci. 2007;19:893–906. doi: 10.1162/jocn.2007.19.5.893. [DOI] [PubMed] [Google Scholar]
  • 74.Wymbs NF. Grafton ST. Contributions from the left PMd and the SMA during sequence retrieval as determined by depth of training. Exp. Brain Res. 2013;224:49–58. doi: 10.1007/s00221-012-3287-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Janata P. Grafton ST. Swinging in the brain: shared neural substrates for behaviors related to sequencing and music. Nat. Neurosci. 2003;6:682–687. doi: 10.1038/nn1081. [DOI] [PubMed] [Google Scholar]
  • 76.Madison G. Experiencing groove induced by music: consistency and phenomenology. Music Percept. 2006;24:201–208. [Google Scholar]
  • 77.Janata P, Tomic ST. Haberman JM. Sensorimotor coupling in music and the psychology of the groove. J. Exp. Psychol.: General. 2012;141:54–75. doi: 10.1037/a0024208. [DOI] [PubMed] [Google Scholar]
  • 78.Liberman AM. Mattingly IG. The motor theory of speech perception revised. Cognition. 1985;21:1–36. doi: 10.1016/0010-0277(85)90021-6. [DOI] [PubMed] [Google Scholar]
  • 79.Diehl RL, Lotto AJ. Holt LL. Speech perception. Ann. Rev. Psychol. 2004;55:149–179. doi: 10.1146/annurev.psych.55.090902.142028. [DOI] [PubMed] [Google Scholar]
  • 80.Martin A, et al. Neural correlates of category-specific knowledge. Nature. 1996;379:649–652. doi: 10.1038/379649a0. [DOI] [PubMed] [Google Scholar]
  • 81.Tranel D, et al. Neural correlate of conceptual knowledge for actions. Cogn. Neuropsychol. 2003;20:409–432. doi: 10.1080/02643290244000248. [DOI] [PubMed] [Google Scholar]
  • 82.Arbib MA. The mirror system, imitation, and the evolution of language. In: Dautenhahn K, editor; Nehaniv C, editor. Imitation in Animals and Artifacts. Cambridge, MA: MIT Press; 2002. pp. 229–280. [Google Scholar]
  • 83.Arbib MA. From monkey-like action recognition to human language: an evolutionary framework for neurolinguistics. Behav. Brain Sci. 2005;28:105–167. doi: 10.1017/s0140525x05000038. [DOI] [PubMed] [Google Scholar]
  • 84.Rizzolatti G. Arbib MA. Language within our grasp. Trends Neurosci. 1998;21:188–194. doi: 10.1016/s0166-2236(98)01260-0. [DOI] [PubMed] [Google Scholar]
  • 85.Condillac ÉBD. Essai sur l'origine des Connaissances Humaines. Gainesville, FL: Scholar's Facsimiles and Reprints; 1971. (1747) [Google Scholar]
  • 86.Corballis MC. From Hand to Mouth: the Origins of Language. Princeton: Princeton University Press; 2002. [Google Scholar]
  • 87.Hewes GW. Primate communication and the gestural origin of language. Curr. Anthropol. 1973;14:5–24. [Google Scholar]
  • 88.Emmorey K. Sign languages are problematic for a gestural origins theory of language evolution. Behav. Brain Sci. 2005;28:130–131. [Google Scholar]
  • 89.Kendon A. Some considerations for a theory of language origins. Man. 1991;26:199–221. [Google Scholar]
  • 90.MacNeilage PF. Davis BL. The frame/content theory of evolution of speech: a comparison with a gestural-origins alternative. Interact. Stud. 2005;6:173–199. [Google Scholar]
  • 91.Toni I, et al. Language beyond action. J. Physiol. 2008;102:71–79. doi: 10.1016/j.jphysparis.2008.03.005. [DOI] [PubMed] [Google Scholar]
  • 92.Fitch WT. The Evolution of Language. Cambridge: Cambridge University Press; 2010. [Google Scholar]
  • 93.Mukamel R, et al. Single-neuron responses in humans during execution and observation of actions. Curr. Biol. 2010;20:750–756. doi: 10.1016/j.cub.2010.02.045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Rogalsky C, et al. Are mirror neurons the basis of speech perception? Evidence from five cases with damage to the purported human mirror system. Neurocase. 2011;17:178–187. doi: 10.1080/13554794.2010.509318. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95.Gervain J, et al. The neonate brain detects speech structure. Proc. Natl. Acad. Sci. U S A. 2008;105:14222–14227. doi: 10.1073/pnas.0806530105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 96.Koelsch S, et al. Bach speaks: a cortical “language-network” serves the processing of music. NeuroImage. 2002;17:956–966. [PubMed] [Google Scholar]
  • 97.Koechlin E. Jubault T. Broca's area and the hierarchical organization of human behavior. Neuron. 2006;50:963–974. doi: 10.1016/j.neuron.2006.05.017. [DOI] [PubMed] [Google Scholar]
  • 98.Dehaene S. Changeux J-P. A hierarchical neuronal network for planning behavior. Proc. Natl. Acad. Sci. U S A. 1997;94:13293–13298. doi: 10.1073/pnas.94.24.13293. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99.Berg WK, et al. Deconstructing the tower: parameters and predictors of problem difficulty on the Tower of London task. Brain Cogn. 2010;72:472–482. doi: 10.1016/j.bandc.2010.01.002. [DOI] [PubMed] [Google Scholar]
  • 100.Newman SD, Greco JA. Lee D. An fMRI study of the Tower of London: a look at problem structure differences. Brain Res. 2009;1286:123–132. doi: 10.1016/j.brainres.2009.06.031. [DOI] [PubMed] [Google Scholar]
  • 101.Jäger G. Rogers J. Formal language theory: refining the Chomsky Hierarchy. Philos. Trans. R. Soc. B. 2012;267:1956–1970. doi: 10.1098/rstb.2012.0077. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 102.Gersting JL. Mathematical Structures for Computer Science. New York: W H Freeman; 1999. [Google Scholar]
  • 103.Hopcroft JE, Motwani R. Ullman JD. Introduction to Automata Theory, Languages and Computation. Reading, Massachusetts: Addison-Wesley; 2000. [Google Scholar]
  • 104.Linz P. An Introduction to Formal Languages and Automata. Sudbury, Massachusetts: Jones & Bartlett; 2001. [Google Scholar]
  • 105.Chomsky N. Syntactic Structures. The Hague: Mouton; 1957. [Google Scholar]
  • 106.Levelt WJM. Formal Grammars in Linguistics and Psycholinguistics. Amsterdam: John Benjamins; 2008. [Google Scholar]
  • 107.Stabler EP. Varieties of crossing dependencies: structure dependence and mild context sensitivity. Cogn. Sci. 2004;28:699–720. [Google Scholar]
  • 108.Stabler EP. The epicenter of linguistic behavior. In: Tanenhaus MK, editor; Sanz M, Laka I, editors. Language Down the Garden Path: The Cognitive and Biological Basis of Linguistic Structures. New York: Oxford University Press; 2013. pp. 316–323. [Google Scholar]
  • 109.Rohrmeier M. Towards a generative syntax of tonal harmony. J. Math. Music. 2011;5:35–53. [Google Scholar]
  • 110.Uddén J. Bahlmann J. A rostro-caudal gradient of structured sequence processing in the left inferior frontal gyrus. Philos. Trans. R. Soc. Lond. B Biol. Sci. 2012;367:2023–2032. doi: 10.1098/rstb.2012.0009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 111.Braver TS, et al. A Parametric Study of Prefrontal Cortex Involvement in Human Working Memory. NeuroImage. 1997;5:49–52. doi: 10.1006/nimg.1996.0247. [DOI] [PubMed] [Google Scholar]
  • 112.Petersson KM, Folia V. Hagoort P. What artificial grammar learning reveals about the neurobiology of syntax. Brain Lang. 2010;120:83–95. doi: 10.1016/j.bandl.2010.08.003. [DOI] [PubMed] [Google Scholar]
  • 113.Koelsch S. Brain and Music. London, UK: John Wiley & Sons; 2012. [Google Scholar]
  • 114.Miller GA. Chomsky N. Finitary models of language users. In: Galanter E, editor; Luce RD, Bush RR, editors. Handbook of Mathematical Psychology. Vol. 35. New York: John Wiley & Sons; 1963. pp. 419–492. [Google Scholar]
  • 115.Bach E, Brown C. Marslen-Wilson W. Crossed and nested dependencies in German and Dutch: a psycholinguistic study. Lang. Cogn. Process. 1986;1:249–262. [Google Scholar]
  • 116.ten Cate C. Okanoya K. Revisiting the syntactic abilities of non-human animals: natural vocalizations and artificial grammar learning. Philos. Trans. R. Soc. B. 2012;367:1984–1994. doi: 10.1098/rstb.2012.0055. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 117.Stobbe N, et al. Visual artificial grammar learning: comparative research on humans, kea (Nestor notabilis) and pigeons (Columba livia) Philos. Trans. R. Soc. B. 2012;367:1995–2006. doi: 10.1098/rstb.2012.0096. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 118.Miller GA. Project Grammarama. In: Miller GA, editor. Psychology of Communication. New York: Basic Books; 1967. [Google Scholar]
  • 119.Fitch WT, Friederici AD. Hagoort P. Pattern perception and computational complexity. Philos. Trans. R. Soc. B. 2012;367:1925–1932. doi: 10.1098/rstb.2012.0099. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 120.Fitch WT. Toward a computational framework for cognitive biology: unifying approaches from cognitive neuroscience and comparative cognition. Phys. Life Rev. doi: 10.1016/j.plrev.2014.04.005. In press. [DOI] [PubMed] [Google Scholar]
  • 121.Fitch WT. Friederici AD. Artificial grammar learning meets formal language theory: an overview. Philos. Trans. R. Soc. B. 2012;367:1933–1955. doi: 10.1098/rstb.2012.0103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 122.Morgan JL, Meier RP. Newport EL. Structural packaging in the input to language learning: contributions of prosodic and morphological marking of phrases to the acquisition of language. Cogn. Psychol. 1987;19:498–550. doi: 10.1016/0010-0285(87)90017-x. [DOI] [PubMed] [Google Scholar]
  • 123.Fodor JA, Bever TG. Garrett MF. The Psychology of Language: An Introduction to Psycholinguistics and Generative Grammar. New York: McGraw-Hill; 1974. [Google Scholar]
  • 124.Ravignani A, et al. Action at a distance: dependency sensitivity in a New World primate. Biol. Lett. 2013;9:20130852. doi: 10.1098/rsbl.2013.0852. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 125.van Heijningen CAA, et al. Simple rules can explain discrimination of putative recursive syntactic structures by a songbird species. Proc. Natl. Acad. Sci. U S A. 2009;106:20538–20543. doi: 10.1073/pnas.0908113106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 126.Beckers GJL, et al. Birdsong neurolinguistics: songbird context-free grammar claim is premature. NeuroReport. 2012;23:139–145. doi: 10.1097/WNR.0b013e32834f1765. [DOI] [PubMed] [Google Scholar]
  • 127.Petkov CI. Jarvis ED. Birds, primates, and spoken language origins: behavioral phenotypes and neurobiological substrates. Front. Evol. Neurosci. 2012;4:e12. doi: 10.3389/fnevo.2012.00012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 128.Gentner TQ, et al. Recursive syntactic pattern learning by songbirds. Nature. 2006;440:1204–1207. doi: 10.1038/nature04675. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 129.Hausberger M, et al. Song sharing reflects the social organization in a captive group of European starlings (Sturnus vulgaris) J. Comp. Psychol. 1995;109:222–241. [Google Scholar]
  • 130.Saffran J, et al. Grammatical pattern learning by human infants and cotton-top tamarin monkeys. Cognition. 2008;107:479–500. doi: 10.1016/j.cognition.2007.10.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 131.Saffran JR, Aslin RN. Newport EL. Statistical learning by 8-month-old infants. Science. 1996;274:1926–1928. doi: 10.1126/science.274.5294.1926. [DOI] [PubMed] [Google Scholar]
  • 132.de Vries MH, et al. Syntactic structure and artificial grammar learning: the learnability of embedded hierarchical structures. Cognition. 2008;107:763–774. doi: 10.1016/j.cognition.2007.09.002. [DOI] [PubMed] [Google Scholar]
  • 133.Fitch WT. Hauser MD. Computational constraints on syntactic processing in a nonhuman primate. Science. 2004;303:377–380. doi: 10.1126/science.1089401. [DOI] [PubMed] [Google Scholar]
  • 134.Uddén J, et al. Implicit acquisition of grammars with crossed and nested non-adjacent dependencies: investigating the push-down stack model. Cogn. Sci. 2012;2012:1–24. doi: 10.1111/j.1551-6709.2012.01235.x. [DOI] [PubMed] [Google Scholar]
  • 135.Byrne RW. Clues to the origin of the human mind from primate observational field data. Jpn. J. Anim. Psychol. 2007;57:1–14. [Google Scholar]
  • 136.Hauser M, Chomsky N. Fitch WT. The language faculty: what is it, who has it, and how did it evolve. Science. 2002;298:1569–1579. doi: 10.1126/science.298.5598.1569. [DOI] [PubMed] [Google Scholar]
  • 137.Johnson-Pynn J, et al. Strategies used to combine seriated cups by chimpanzees (Pan troglodytes), bonobos (Pan paniscus), and capuchins (Cebus apella) J. Comp. Psychol. 1999;113:137–148. doi: 10.1037/0735-7036.113.2.137. [DOI] [PubMed] [Google Scholar]
  • 138.Sanz C, Call J. Morgan DB. Design complexity in termite-fishing tools of chimpanzees. Biol. Lett. 2009;5:293–296. doi: 10.1098/rsbl.2008.0786. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 139.Mandelbrot BB. The fractal geometry of nature. New York: Freeman; 1977. [Google Scholar]
  • 140.O'Donnell TJ, Hauser MD. Fitch WT. Using mathematical models of language experimentally. Trends Cogn. Sci. 2005;9:284–289. doi: 10.1016/j.tics.2005.04.011. [DOI] [PubMed] [Google Scholar]
  • 141.Fitch WT. Three meanings of “recursion”: key distinctions for biolinguistics. In: Yamakido H, editor; Larson R, Déprez V, editors. The Evolution of Human Language: Biolinguistic Perspectives. Cambridge, UK: Cambridge University Press; 2010. pp. 73–90. [Google Scholar]
  • 142.Fitch WT, Hauser MD. Chomsky N. The evolution of the language faculty: clarifications and Implications. Cognition. 2005;97:179–210. doi: 10.1016/j.cognition.2005.02.005. [DOI] [PubMed] [Google Scholar]
  • 143.Martins MD. Specific signatures of recursion. Philos. Trans. R. Soc. B. 2012;367:2055–2064. doi: 10.1098/rstb.2012.0097. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144.Van der Hulst H. Recursion and Human Language. Berlin: De Gruyter/Mouton; 2010. [Google Scholar]
  • 145.Huxley TH. Man's Place in Nature. Mineola, NY: Dover; 1863. [Google Scholar]
  • 146.Jerison HJ. Quantitative analysis of evolution of the brain in mammals. Science. 1961;133:1012–1014. doi: 10.1126/science.133.3457.1012. [DOI] [PubMed] [Google Scholar]
  • 147.Stephan H, Frahm H. Baron G. New and revised data on volumes of brain structures in insectivores and primates. Folia. Primatol. 1981;35:1–29. doi: 10.1159/000155963. [DOI] [PubMed] [Google Scholar]
  • 148.Deacon TW. Rethinking mammalian brain evolution. Am. Zool. 1990;30:629–705. [Google Scholar]
  • 149.Jerison HJ. Evolution of the Brain and Intelligence. New York: Academic Press; 1973. [Google Scholar]
  • 150.Holloway RL. Evolution of the human brain. In: Peters CR, editor; Lock A, editor. Handbook of Human Symbolic Evolution. Oxford: Clarendon Press; 1996. pp. 74–108. [Google Scholar]
  • 151.Semendeferi K, et al. Humans and great apes share a large frontal cortex. Nat. Neurosci. 2002;5:272–276. doi: 10.1038/nn814. [DOI] [PubMed] [Google Scholar]
  • 152.Deaner RO, van Schaik CP. Johnson V. Do some taxa have better domain-general cognition than others? A meta-analysis of nonhuman primate studies. Evol. Psychol. 2006;4:149–196. [Google Scholar]
  • 153.Schoenemann PT, Sheehan MJ. Glotzer LD. Prefrontal white matter volume is disproportionately larger in humans than in other primates. Nat. Neurosci. 2005;8:242–253. doi: 10.1038/nn1394. [DOI] [PubMed] [Google Scholar]
  • 154.Barton RA. Venditti C. Human frontal lobes are not relatively large. Proc. Natl. Acad. Sci. U S A. 2013;110:9000–9006. doi: 10.1073/pnas.1215723110. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 155.Sakai T, et al. Differential prefrontal white matter development in chimpanzees and humans. Curr. Biol. 2011;21:1397–1402. doi: 10.1016/j.cub.2011.07.019. [DOI] [PubMed] [Google Scholar]
  • 156.Sherwood CC, Holloway RL, Semendeferi K. Hof PR. Is prefrontal white matter enlargement a human evolutionary specialization? Nat. Neurosci. 2005;8:537–538. doi: 10.1038/nn0505-537. [DOI] [PubMed] [Google Scholar]
  • 157.Smaers JB, et al. Frontal white matter volume is associated with brain enlargement and higher structural connectivity in anthropoid primates. PLoS One. 2010;5:e9123. doi: 10.1371/journal.pone.0009123. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 158.Schenker NM, et al. Broca's area homologue in chimpanzees (Pan troglodytes): probabilistic mapping, asymmetry and comparison to humans. Cereb. Cortex. 2010;20:730–742. doi: 10.1093/cercor/bhp138. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 159.Amunts K, et al. Broca's region revisited: cytoarchitecture and intersubject variability. J. Comp. Neurol. 1999;412:319–341. doi: 10.1002/(sici)1096-9861(19990920)412:2<319::aid-cne10>3.0.co;2-7. [DOI] [PubMed] [Google Scholar]
  • 160.Koch MA, Norris DG. Hund-Georgiadis M. An investigation of functional and anatomical connectivity using magnetic resonance imaging. Neuroimage. 2002;16:241–250. doi: 10.1006/nimg.2001.1052. [DOI] [PubMed] [Google Scholar]
  • 161.Rilling JK, et al. The evolution of the arcuate fasciculus revealed with comparative DTI. Nat. Neurosci. 2008;11:426–428. doi: 10.1038/nn2072. [DOI] [PubMed] [Google Scholar]
  • 162.Catani M. Mesulam M. The arcuate fasciculus and the disconnection theme in language and aphasia: history and current state. Cortex. 2008;44:953–961. doi: 10.1016/j.cortex.2008.04.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 163.Goodall J. The Chimpanzees of Gombe: Patterns of Behavior. Cambridge, Massachusetts: Harvard University Press; 1986. [Google Scholar]
  • 164.McGrew WC. Marchant LF. Ethological studies of manual laterality in the chimpanzees of the Mahale Mountains, Tanzania. Behaviour. 2001;138:329–358. [Google Scholar]
  • 165.Sugiyama Y. Koman J. Tool-using and tool-making behavior in wild chimpanzees at Bossou, Guinea. Primates. 1979;20:513–524. [Google Scholar]

Articles from Annals of the New York Academy of Sciences are provided here courtesy of Wiley

RESOURCES