Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2015 Oct 1.
Published in final edited form as: Curr Opin Neurobiol. 2014 Jul 26;0:142–149. doi: 10.1016/j.conb.2014.07.005

The neuroanatomic and neurophysiological infrastructure for speech and language

David Poeppel 1,2
PMCID: PMC4177440  NIHMSID: NIHMS612235  PMID: 25064048

Abstract

New tools and new ideas have changed how we think about the neurobiological foundations of speech and language processing. This perspective focuses on two areas of progress. First, focusing on spatial organization in the human brain, the revised functional anatomy for speech and language is discussed. The complexity of the network organization undermines the well-regarded classical model and suggests looking for more granular computational primitives, motivated both by linguistic theory and neural circuitry. Second, focusing on recent work on temporal organization, a potential role of cortical oscillations for speech processing is outlined. Such an implementational-level mechanism suggests one way to deal with the computational challenge of segmenting natural speech.

Introduction

Experimental research on the neurobiological foundations of speech and language processing has taken considerable strides in the last decade, due in part to advances in the methods available to study the human brain (improved resolution of recording techniques) and in part to more theoretically motivated research that builds on critical distinctions provided by the results of linguistics, cognitive psychology and computer science (improved ‘conceptual resolution’). As the neurobiology of language matures, the units of analysis continue to change and become increasingly refined: from (i) broad (and somewhat pre-theoretical) categories such as ‘production’ versus ‘perception/comprehension’ to (ii) subroutines of language processing such as phonology, lexical processing, syntax, semantics, and so on, to (iii) ever more fine-grained representations and computational primitives argued to underpin the different subroutines of language, such as concatenation, linearization, etc.

In all areas of language processing, noteworthy new perspectives have been developed (reviewed, among many others, e.g., in [13], with special emphasis on speech, linguistic structure-building, and the sensorimotor basis of speech/language, respectively). Notwithstanding the novel approaches, many of the substantive challenges are only now becoming clear. The number and arrangement of the cortical and subcortical regions underpinning speech and language processing demonstrate that the system is considerably more complex and distributed; the age of Broca’s and Wernicke’s areas and the era of left-hemisphere imperialism are over. Here I focus on a two issues that are redefining the research agenda, pointing towards a computational neurobiology of language [4], a research direction that emphasizes the representational and computational primitives that form the basis of speech and language.

There are, of course, many ways to illustrate the progress that has been made, highlighting new ideas and directions. One approach would be to review the different aspects or levels of language processing that have been examined in new neuroscientific experimentation, i.e. phonetics, phonology [5,6], lexical access [710], lexical semantics [11], syntax [12,13], compositional semantics [14,15], discourse representation [16,17]; moreover, the interaction of the linguistic computational system with other domains has been investigated in interesting ways, including how language processing interfaces with attention [18], memory [19], emotion [20], cognitive control [21], predictive coding [2224], and even aesthetics [25]. A different approach is taken here, focusing first on the revised spatial map of brain and language; then, narrowing to one functional problem, a new ‘temporal view’ is discussed to illustrate a linking hypothesis between the computational requirements of speech perception and the neurobiological infrastructure that may provide a neural substrate.

The new functional anatomy: maps of regions, streams, and hemispheres

Our understanding of the anatomic foundations of language processing has changed dramatically in the last 10 years, ranging from the biggest to the most local levels. One might call this the maps problem [26], i.e. the challenge to define the best possible spatial map that describes the anatomic substrate [2729]. The older, ‘classical’ view and its limitations are discussed further in Hagoort, this volume, where a contrasting dynamic network view of local function is described.

a. Starting at the most coarse level, consider the role of hemispheric asymmetry. Historically, the lateralization of language processing to the ‘dominant hemisphere’ has been one of the principal defining features. It was uncontroversial that language processing is strongly lateralized. However, a more nuanced and theoretically informed view of language processing, breaking down the processes into constituent operations, has revealed that lateralization patterns are complex and subtle – and that not all language processing components are lateralized. For example, when examining the cortical regions mediating speech perception and lexical level comprehension, lesion [30,31], imaging [3234], and electrophysiological data [35,36] demonstrate convincingly that both left and right superior temporal cortical regions are implicated. Indeed, the operations mapping from input signals (e.g. sound) to lexical-level meaning, argued to be part of ventral stream processing (see b, below) appear to be robustly bilateral, as illustrated in Figure 1a (bottom panel).

Figure 1.

Figure 1

a. Dual stream model [1]. b. Note the bilateral arrangement of the ventral stream network (mapping from sound to meaning) and the lateralized dorsal stream projections (mapping to articulation). b. Bilateral processing of sensory-motor transformations for speech, from [37]. c. Spectrograms for three tasks: A’ Listen-Speak’ task: subjects heard a word and after a short delay had to repeat it; A ‘Listen-Mime’ task: subjects heard a word and after the same short delay had to move their articulators without vocalizing; ‘Listen’ task: subjects listened passively to a word. Sensory-motor (S-M) responses are seen in example electrodes in both the left (top row) and right (bottom row) hemispheres as demonstrated by a high gamma neural response (70–90+ Hz) present both when the subject listened to a word and when they repeated/mimed it. d. Population average brains with active electrodes demonstrate that S-M processes occur bilaterally (red). Electrodes that responded to the passive listen condition are also noted with green outlines (red with green outlines).

In contrast, it is typically argued that the structures and operations underlying production, for example, are lateralized. As illustrated in Figure 1a, one of the dorsal stream projections, suggested to underpin the sensory-motor mapping necessary for perception-production alignment, is depicted as fully lateralized. However, new data acquired in pre-surgical epilepsy patients using electrocorticography (ECog) seriously challenge even this generalization [37]. It is shown based on a range of tasks requiring sensory (sound)-to-motor (articulatory) transformation that the dorsal stream structures that provide the basis for this mapping are clearly bilateral as well (Figure 1b). Other, non-speech dorsal stream functions, for example operations that are part of grammatical relations, may be supported by other dorsal stream projections, and their lateralization pattern has not been fully established, although there appears to be a fair degree of lateralization to the dominant hemisphere (see [2] and Hagoort, this volume).

Various other imaging and physiology experiments on other aspects of language [38,23] also invite the interpretation that lateralization patterns are more complex than anticipated. Cumulatively, in other words, the language-ready brain (to use a phrase of Hagoort, this volume) appears to execute many of its subroutines bilaterally, unlike the 150-year-old dogma. And yet … There remain compelling reasons why an explanation for lateralization of function is required. Lesions to the left versus the right peri-Sylvian regions do not lead to the same linguistic deficits, so a new approach to this classical issue is necessary.

Which operations are functionally lateralized (and why) remains controversial. It has been debated for the speech perception case, for instance [3941], but the nature of the questions is at least fine-grained, that is to say at the level of circuits that execute specific (possibly network-dependent) computations [42]. One hypothesis that merits exploration is as follows: the operations that comprise the processing of input and output systems (the interfaces) are carried out bilaterally; the linguistic operations per se, beyond the initial mappings, and specifically those that require combinatorics and composition (COM) as well as linguistically (structurally-) based prediction (PRE) are lateralized. A COM-PRE view predicts that the neural circuits mediating those types of operations on linguistic data structures are asymmetrically distributed; but how such circuits might look remains pure speculation.

b. The organization of language regions within a hemisphere has also seen a major shift in perspective. The classical model – still the prevailing view in most textbooks – identifies a few critical areas (Broca’s and Wernicke’s regions, often the angular gyrus, connected by the arcuate fasciculus) and attributes entire chunks of linguistic cognition to these large brain regions. (See Hagoort, this volume, for critique.) There now exists consensus that a more likely organization involves processing streams organized along dorsal and ventral routes. This is argued to be the case for speech [1,43,44], lexical level processing [45], syntactic analysis [2], and semantics [46].

The existence of concurrent ‘what,’ ‘where,’ and ‘how’ streams in vision highlights how large-scale computational challenges (localization, identification, sensorimotor transformation/action– perception linking) can be implemented in parallel by streams consisting of hierarchically organized cortical regions. This key idea from visual neuroscience was adapted for speech and language in the past 10 years [1,2,43,44]. Figure 1a illustrates one such model, constructed to account for a range of phenomena in speech. The bilateral ventral streams are considered responsible for supporting the transformation from auditory input to lexical representations. The dorsal stream (one of at least two parallel dorsal streams, see [2]) is primarily critical for sensorimotor transformations.

c. Critical advances on local anatomy. Until about 2000, the interpretation of experimental data implicating Broca’s area were made at a level of analysis referring to left inferior frontal gyrus and, at best, about Brodmann areas 44 versus 45. There exist some interesting connectivity studies, as well [47], but by and large the functional anatomic characterization has been coarse. (Hagoort, this volume, provides a more extensive discussion of the functional role of Broca’s area.)

New anatomic techniques (imaging, immunocytochemistry) have now been applied. Data in an influential paper [48] show that the organization of this one cortical region is much more complex, incorporating (depending how one counts) 5 – 10 local fields that differ in their cellular properties. Figure 2a [48] illustrates the currently hypothesized organization of Broca’s region, highlighting anterior and posterior as well as dorsal and ventral subdivisions of known fields and pointing to additional, separate opercular and sulcal regions that are anatomically identifiable.

Figure 2.

Figure 2

A. Broca’s region, from [48]. Note the numerous subdivisions in the region. B. Major pathways associated with left MTG, from [50]. Tractography data from two subjects are show. Each row depicts the individual subject’s ROI, represented in their native space (left, yellow). Sagittal, axial, and coronal slices of the fiber bundles involved are shown. As per DTI mapping, left MTG is massively connected to critical language regions in a hub-like manner.

It is highly likely that each sub-region performs at least one different computation — after all, they are anatomically distinct. Suppose then that there are merely five anatomic subdomains in Broca’s region, and suppose that each anatomic subdomain supports, say, two computations. These are conservative numbers, but we must then still identify 10 computations (underestimating the actual complexity). Some of the computations will apply to any input data structure (structured linguistic representation, speech signal, action plan, rhythmic sequence, etc.), since it is established that Broca’s region is engaged by numerous cognitive tasks [49]. Other computations may be dedicated to data structures native to the linguistic cognitive system. We know little about precisely what kind of conditions need to be met to trigger selectively the many areas that constitute this part of the inferior frontal cortex.

d. A final point about anatomy: the vast majority of research on brain and language has focused on the traditional peri-Sylvian language areas; but both cortical and subcortical regions that have not been investigated before as much play a critical role. (This point is amplified, as well, in Hagoort, this volume.) Two regions, in particular, deserve emphasis with respect to linguistic computation: the left anterior temporal lobe (ATL), and the posterior middle temporal gyrus (pMTG). The left ATL has been implicated in a variety of experiments on elementary structure assembly, i.e. the primitive combinatorics (COM) that underlie syntactic and semantic structure building. fMRI data show quite consistently that combination into phrasal units, perhaps in the service of subsequent semantic interpretation, is mediated there [2]. Recent MEG data [14,15] provide evidence about the dynamics; the experiments demonstrate that between 200–300 ms after the onset of a critical word that can be combined with a previous item, this region reflects the execution of basic combinatoric operations.

The pMTG appears to be a crucial lexical node in the larger network of language areas – and heavily connected to lATL. This region is, in general, very richly connected to other language areas (see Figure 2b, [50] and is driven by many lexical tasks in imaging studies, including lexical access, lexical ambiguity resolution, and other processes. Lesions to the MTG lead to severe word comprehension difficulties [51] and profound aphasia, if paired with STG lesions. Recent MEG data on the dynamics in pMTG during lexical access show that, after initial activation of lexical targets, unambiguously lexical attributes such as surface frequency and neighborhood density elicit MTG modulation [52] no later than 350 ms after word onset.

In sum, the functional anatomy of speech and language looks quite different now than the classical model has taught us. Locally highly structured areas are connected into networks that are themselves organized into processing streams that support broader computational goals (e.g. sound to meaning mapping, sound to articulation mapping) and that are, in turn, bilaterally supported for many linguistic functions. The research program for the next years will ask about which basic computations are executed in the most local circuits and which computations then group together to generate the linguistic functions that constitute language, say, syntax, phonology, lexical access, or semantics. Structure building – be it for sound or meaning [14,53,54] – requires the assembly of elementary representations by elementary computations, and it is ultimately these basic operations we are seeking to identify in a new neurobiology of language.

The temporal view: a linking hypothesis for the neurobiology of speech perception

For nearly a century, speech perception has been studied primarily from the perspective of the acoustic signal. Spectral properties were thought to provide the principal cues for decoding phonemes, and from these building blocks words and phrases would be derived. In cognitive neuroscience studies, the basic intuitions of such a model are investigated regularly, testing which acoustic-phonetic primitives are essential and invariant, acoustic attributes [6], distinctive features [5,6,55,56], or phonemes [57]. Despite the successes of this research program, many issues have remained unanswered. One major issue concerns how the brain extracts the relevant units for analysis to begin with.

Speech and other dynamically changing auditory signals (as well as visual stimuli, including sign) contain critical information required for successful decoding that is carried at multiple temporal scales (e.g. intonation-level information at the scale of 500–2000ms, syllabic information that is closely correlated to the acoustic envelope of speech, ~150–300 ms, and rapidly changing featural information, ~20–80 ms). These different aspects of signals (slow and fast temporal modulation, frequency composition) must be analyzed for successful recognition. Recent research has aimed to identify the basis for the required multi-time resolution analysis [58,72]. A series of recent experiments suggests that intrinsic neuronal oscillations in cortex at ‘privileged’ frequencies (delta 1–3 Hz, theta 4–8 Hz, low gamma 30–50 Hz) may provide some of the relevant mechanisms to parse continuous speech into the necessary chunks for decoding [5963]. To achieve parsing of a naturalistic input signal (e.g. speech signal) into elementary pieces, one ‘mesoscopic-level’ mechanism is suggested to be the sliding and resetting of temporal windows, implemented as phase locking of low-frequency (delta, theta) activity to the envelope of speech and phase resetting of the intrinsic oscillations on privileged time scales [58,64]. The successful resetting of neuronal activity, triggered in part by stimulus-driven spikes, provides time constants (or temporal integration windows) for parsing and decoding speech signals. Recent studies link the infrastructure provided by neural oscillations (which reflect neuronal excitability cycles) to basic perceptual challenges in speech recognition, such as breaking the continuous input stream into chunks suitable for subsequent analysis [59,63,65].

Two recent studies serve to illustrate the logic of the research program. Figure 3a [65] shows the data from a magnetoencephalography (MEG) recording in which subjects listened to continuous regular or manipulated speech. The data show the close alignment between the stimulus envelope and the response from auditory cortex, as has also been shown in other recent studies [59,62,63]. How this alignment between speech acoustics and neural oscillations might underpin intelligibility has been unclear. This study tested the hypothesis that the ‘sharpness’ of temporal fluctuations in the critical band envelope was a temporal cue to syllabic rate, driving the intrinsic delta or theta rhythms to track the stimulus and thereby facilitating intelligibility. It was observed that ‘sharp’ events in the stimulus (i.e. auditory edges) cause cortical rhythms to re-align and parse the stimulus into syllable-sized chunks for further decoding. Using MEG recordings it was shown that by removing temporal fluctuations that occur at the syllabic rate, envelope-tracking activity is compromised. By artificially reinstating such temporal fluctuations, envelope-tracking activity is regained. Critically, changes in tracking correlate with stimulus intelligibility. These results suggest that the sharpness of stimulus edges, as reflected in the cochlear output, drive oscillatory activity to track and entrain to the stimulus, at its syllabic rate. This process facilitates parsing of the stimulus into meaningful chunks appropriate for subsequent decoding.

Figure 3.

Figure 3

A. Entrainment of cortical activity to speech envelope, from [65]. B. Spectrograms of stimulus and neural activity illustrate correspondence in frequency domain. C. Contour maps of two conditions show typical auditory distribution. D. Attention dependent processing of a complex speech signal with two simultaneous speakers [69]. Stimulus is same mixture of two voices. Upper trace: attend to same speaker/stream on different trials. Lower trace: attend to different speaker/stream on different trials.

If neuronal activity locks or entrains to stimulus features, is this process subject to the vagaries of naturalistic communication, with wandering attention, for example? Several recent studies have tested this [6669]; one result is shown in Figure 3b [69]. This ECog study investigated the manner in which speech streams are represented in brain activity and the way that selective attention governs the representation of speech using the ‘‘Cocktail Party’’ paradigm. It is shown that brain activity dynamically tracks speech streams using both low-frequency (delta, theta) phase and high-frequency (high gamma) amplitude fluctuations, and it is argued that optimal encoding likely combines the two – in the spirit of multi-time scale processing. Moreover, in and near low-level auditory cortices, attention modulates the representation by enhancing cortical tracking of attended speech streams (Figure 3b). Yet ignored speech remains represented. Jointly, these studies demonstrate the potential role that neuronal oscillations may play in the parsing, decoding, and attending to naturalistic speech [18].

Conclusion

There has, of course, been exciting progress in other more neuro-computational studies of language, including speech production [70], lexical representation and processing [(10)], and predictive coding [71]. This perspective is necessarily brief and selective. It is fair to say, however, that in the last 10–15 years, the model of how language is processed in the brain has changed dramatically from the classical perspective developed between 1861 (Broca), 1874 (Wernicke), 1885 (Lichtheim), and the 1970s (Geschwind). These changes are a consequence of a more mature linguistics, psychology, and neuroscience. It is important to acknowledge the immense contribution to basic and clinical neuroscience that the classical model has made – but to acknowledge as well that it cannot be carried forth as a realistic view on language and the brain.

Highlights.

  • The anatomic organization for language is less lateralized than assumed.

  • Speech and language processing are mediated by segregated concurrent processing streams.

  • The anatomic and functional complexity of local regions needs to be acknowledged.

  • Neural oscillations provide a possible mechanism to parse speech.

Acknowledgments

DP is supported by NIH 2R01DC05660, NSF BSC-1344285, and Army Research Office MURI ARO 54228-LS-MUR.

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

References and recommended reading

  • 1.Hickok G, Poeppel D. The cortical organization of speech processing. Nat Rev Neurosci. 2007;8:393–402. doi: 10.1038/nrn2113. [DOI] [PubMed] [Google Scholar]
  • 2.Friederici AD. The brain basis of language processing: From structure to function. Physiol Rev. 2011;91:1357–92. doi: 10.1152/physrev.00006.2011. [DOI] [PubMed] [Google Scholar]
  • 3.Pulvermüller F, Fadiga L. Active perception: Sensorimotor circuits as a cortical basis for language. Nat Rev Neurosci. 2010;11:351–60. doi: 10.1038/nrn2811. [DOI] [PubMed] [Google Scholar]
  • 4.Poeppel D, Embick D. Defining the relation between linguistics and neuroscience. In: Cutler A, editor. Twenty-first Century Psycholinguistics: Four Cornerstones. Lawrence Erlbaum; 2005. pp. 103–120. [Google Scholar]
  • 5.Scharinger M, Merickel J, Riley J, Idsardi WJ. Neuromagnetic evidence for a featural distinction of english consonants: Sensor- and source-space data. Brain Lang. 2011;116:71–82. doi: 10.1016/j.bandl.2010.11.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Mesgarani N, Cheung C, Johnson K, Chang EF. Phonetic feature encoding in human superior temporal gyrus. Science. 2014;343:1006–10. doi: 10.1126/science.1245994. ECog data show the sensitivity and specificity profiles of recordings over left STG. Both acoustic patterns and linguistic (distinctive featural) patterns are observed, consistent with the view that features (defining articulation) have acoustic correlates.
  • 7.Fiorentino R, Naito-Billen Y, Bost J, Fund-Reznicek E. Electrophysiological evidence for the morpheme-based combinatoric processing of English compounds. Cogn Neuropsychol. 2013 doi: 10.1080/02643294.2013.855633. PMID: 24279696. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Almeida D, Poeppel D. Word-specific repetition effects revealed by MEG and the implications for lexical access. Brain Lang. 2013;127:497–509. doi: 10.1016/j.bandl.2013.09.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Bozic M, Tyler LK, Su L, Wingfield C, Marslen-Wilson WD. Neurobiological systems for lexical representation and analysis in English. J Cogn Neurosci. 2013;25:1678–91. doi: 10.1162/jocn_a_00420. [DOI] [PubMed] [Google Scholar]
  • 10.Ettinger A, Linzen T, Marantz A. The role of morphology in phoneme prediction: Evidence from MEG. Brain Lang. 2014;129:14–23. doi: 10.1016/j.bandl.2013.11.004. [DOI] [PubMed] [Google Scholar]
  • 11.Crepaldi D, Berlingeri M, Cattinelli I, Borghese NA, Luzzatti C, Paulesu E. Clustering the lexicon in the brain: A meta-analysis of the neurofunctional evidence on noun and verb processing. Front Hum Neurosci. 2013;7:303. doi: 10.3389/fnhum.2013.00303. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Fedorenko E, Nieto-Castañón A, Kanwisher N. Syntactic processing in the human brain: What we know, what we don't know, and a suggestion for how to proceed. Brain Lang. 2012;120:187–207. doi: 10.1016/j.bandl.2011.01.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Moro A. On the similarity between syntax and actions. Trends Cogn Sci. 2014;18:109–10. doi: 10.1016/j.tics.2013.11.006. [DOI] [PubMed] [Google Scholar]
  • 14. Bemis DK, Pylkkänen L. Basic linguistic composition recruits the left anterior temporal lobe and left angular gyrus during both listening and reading. Cereb Cortex. 2013;23:1859–73. doi: 10.1093/cercor/bhs170. Building on a line of research investigating how the elementary structural and semantic composition of linguistic pieces is accomplished, this work highlights where and when basic linguistic composition is executed across language tasks.
  • 15.Westerlund M, Pylkkänen L. The role of the left anterior temporal lobe in semantic composition vs. Semantic memory. Neuropsychologia. 2014;57:59–70. doi: 10.1016/j.neuropsychologia.2014.03.001. [DOI] [PubMed] [Google Scholar]
  • 16.Hagoort P, van Berkum J. Beyond the sentence given. Philos Trans R Soc Lond B Biol Sci. 2007;362:801–11. doi: 10.1098/rstb.2007.2089. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Xu J, Kemeny S, Park G, Frattali C, Braun A. Language in context: Emergent features of word, sentence, and narrative comprehension. Neuroimage. 2005;25:1002–15. doi: 10.1016/j.neuroimage.2004.12.013. [DOI] [PubMed] [Google Scholar]
  • 18.Zion Golumbic EM, Poeppel D, Schroeder CE. Temporal context in speech processing and attentional stream selection: A behavioral and neural perspective. Brain Lang. 2012;122:151–61. doi: 10.1016/j.bandl.2011.12.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Li X, Yang Y. How long-term memory and accentuation interact during spoken language comprehension. Neuropsychologia. 2013;51:967–78. doi: 10.1016/j.neuropsychologia.2012.12.016. [DOI] [PubMed] [Google Scholar]
  • 20.Kotz SA, Paulmann S. When emotional prosody and semantics dance cheek to cheek: ERP evidence. Brain Res. 2007;1151:107–18. doi: 10.1016/j.brainres.2007.03.015. [DOI] [PubMed] [Google Scholar]
  • 21.January D, Trueswell JC, Thompson-Schill SL. Co-localization of stroop and syntactic ambiguity resolution in Broca's area: Implications for the neural basis of sentence processing. J Cogn Neurosci. 2009;21:2434–44. doi: 10.1162/jocn.2008.21179. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Poeppel D, Idsardi WJ, Van Wassenhove V. Speech perception at the interface of neurobiology and linguistics. Philosophical Transactions of the Royal Society B: Biological Sciences. 2008;363:1071. doi: 10.1098/rstb.2007.2160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Beeman MJ, Bowden EM, Gernsbacher MA. Right and left hemisphere cooperation for drawing predictive and coherence inferences during normal story comprehension. Brain Lang. 2000;71:310–36. doi: 10.1006/brln.1999.2268. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Federmeier KD. Thinking ahead: The role and roots of prediction in language comprehension. Psychophysiology. 2007;44:491–505. doi: 10.1111/j.1469-8986.2007.00531.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Obermeier C, Menninghaus W, von Koppenfels M, Raettig T, Schmidt-Kassow M, Otterbein S, Kotz SA. Aesthetic and emotional effects of meter and rhyme in poetry. Front Psychol. 2013;4:10. doi: 10.3389/fpsyg.2013.00010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Poeppel D. The maps problem and the mapping problem: Two challenges for a cognitive neuroscience of speech and language. Cogn Neuropsychol. 2012;29:34–55. doi: 10.1080/02643294.2012.710600. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Price CJ. A review and synthesis of the first 20 years of PET and fmri studies of heard speech, spoken language and reading. Neuroimage. 2012;62:816–47. doi: 10.1016/j.neuroimage.2012.04.062. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Vigneau M, Beaucousin V, Hervé PY, Duffau H, Crivello F, Houdé O, et al. Meta-analyzing left hemisphere language areas: Phonology, semantics, and sentence processing. Neuroimage. 2006;30:1414–32. doi: 10.1016/j.neuroimage.2005.11.002. [DOI] [PubMed] [Google Scholar]
  • 29.Vigneau M, Beaucousin V, Hervé PY, Jobard G, Petit L, Crivello F, et al. What is right-hemisphere contribution to phonological, lexico-semantic, and sentence processing? Insights from a meta-analysis. Neuroimage. 2011;54:577–93. doi: 10.1016/j.neuroimage.2010.07.036. [DOI] [PubMed] [Google Scholar]
  • 30.Poeppel D. Pure word deafness and the bilateral processing of the speech code. Cognitive Science. 2001;25:679–93. [Google Scholar]
  • 31.Hickok G, Poeppel D. Towards a functional neuroanatomy of speech perception. Trends Cogn Sci. 2000;4:131–8. doi: 10.1016/s1364-6613(00)01463-7. [DOI] [PubMed] [Google Scholar]
  • 32.Bozic M, Tyler LK, Ives DT, Randall B, Marslen-Wilson WD. Bihemispheric foundations for human speech comprehension. Proc Natl Acad Sci U S A. 2010;107:17439–44. doi: 10.1073/pnas.1000531107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Turkeltaub PE, Coslett HB. Localization of sublexical speech perception components. Brain Lang. 2010;114:1–15. doi: 10.1016/j.bandl.2010.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Binder JR, Frost JA, Hammeke TA, Bellgowan PS, Springer JA, Kaufman JN, Possing ET. Human temporal lobe activation by speech and nonspeech sounds. Cereb Cortex. 2000;10:512–28. doi: 10.1093/cercor/10.5.512. [DOI] [PubMed] [Google Scholar]
  • 35.Morillon B, Lehongre K, Frackowiak RS, Ducorps A, Kleinschmidt A, Poeppel D, Giraud AL. Neurophysiological origin of human brain asymmetry for speech and language. Proc Natl Acad Sci U S A. 2010;107:18688–93. doi: 10.1073/pnas.1007189107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Morillon B, Liégeois-Chauvel C, Arnal LH, Bénar CG, Giraud AL. Asymmetric function of theta and gamma activity in syllable processing: An intra-cortical study. Front Psychol. 2012;3:248. doi: 10.3389/fpsyg.2012.00248. This study capitalizes on rare bilateral intracortical recordings in a patient to test a specific computational hypothesis about asymmetry in speech perception.
  • 37. Cogan GB, Thesen T, Carlson C, Doyle W, Devinsky O, Pesaran B. Sensory-motor transformations for speech occur bilaterally. Nature. 2014;507:94–8. doi: 10.1038/nature12935. The ECog data presented demonstrate, among a range of findings, that the sensory-motor coordinate transformations that underlie the mapping from sounds to articulations are bilaterally mediated, against the prevailing dogma that dorsal stream sensory-motor functions are highly lateralized.
  • 38.Wlotko EW, Federmeier KD. Two sides of meaning: The scalp-recorded n400 reflects distinct contributions from the cerebral hemispheres. Front Psychol. 2013;4:181. doi: 10.3389/fpsyg.2013.00181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Poeppel D. The analysis of speech in different temporal integration windows: Cerebral lateralization as ‘asymmetric sampling in time’. Speech Communication. 2003;41:245–55. [Google Scholar]
  • 40.McGettigan C, Scott SK. Cortical asymmetries in speech perception: What's wrong, what's right and what's left? Trends Cogn Sci. 2012;16:269–76. doi: 10.1016/j.tics.2012.04.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Scott SK, McGettigan C. Do temporal processes underlie left hemisphere dominance in speech perception? Brain Lang. 2013;127:36–45. doi: 10.1016/j.bandl.2013.07.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Giraud AL, Kleinschmidt A, Poeppel D, Lund TE, Frackowiak RS, Laufs H. Endogenous cortical rhythms determine cerebral specialization for speech perception and production. Neuron. 2007;56:1127–34. doi: 10.1016/j.neuron.2007.09.038. [DOI] [PubMed] [Google Scholar]
  • 43.Hickok G, Poeppel D. Dorsal and ventral streams: A framework for understanding aspects of the functional anatomy of language. Cognition. 2004;92:67–99. doi: 10.1016/j.cognition.2003.10.011. [DOI] [PubMed] [Google Scholar]
  • 44.Rauschecker JP, Scott SK. Maps and streams in the auditory cortex: Nonhuman primates illuminate human speech processing. Nat Neurosci. 2009;12:718–24. doi: 10.1038/nn.2331. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Gow DW. The cortical organization of lexical knowledge: A dual lexicon model of spoken language processing. Brain Lang. 2012;121:273–88. doi: 10.1016/j.bandl.2012.03.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Lau EF, Phillips C, Poeppel D. A cortical network for semantics: (De)constructing the N400. Nat Rev Neurosci. 2008;9:920–33. doi: 10.1038/nrn2532. [DOI] [PubMed] [Google Scholar]
  • 47.Anwander A, Tittgemeyer M, von Cramon DY, Friederici AD, Knösche TR. Connectivity-Based parcellation of Broca's area. Cereb Cortex. 2007;17:816–25. doi: 10.1093/cercor/bhk034. [DOI] [PubMed] [Google Scholar]
  • 48. Amunts K, Lenzen M, Friederici AD, Schleicher A, Morosan P, Palomero-Gallagher N, Zilles K. Broca's region: Novel organizational principles and multiple receptor mapping. PLoS Biol. 2010;8(9) doi: 10.1371/journal.pbio.1000489. The anatomic analysis of Broca’s region summarized here demonstrates the complexity of the region that has not been sufficiently appreciated in functional studies. The state-of-the-art map of Broca’s region invites a much more granular view of the computations executed there.
  • 49.Fedorenko E, Duncan J, Kanwisher N. Language-selective and domain-general regions lie side by side within Broca's area. Curr Biol. 2012;22:2059–62. doi: 10.1016/j.cub.2012.09.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Turken AU, Dronkers NF. The neural architecture of the language comprehension network: Converging evidence from lesion and connectivity analyses. Front Syst Neurosci. 2011;5:1. doi: 10.3389/fnsys.2011.00001. An extensive connectivity study of several left-hemisphere temporal language regions. The left MTG shows a very extensive structural and functional connectivity pattern, consistent with the language impairments associated with posterior MTG lesions. These data reinforce an essential role for this region in language processing, probably at the lexical level.
  • 51.Dronkers NF, Wilkins DP, Van Valin RD, Redfern BB, Jaeger JJ. Lesion analysis of the brain areas involved in language comprehension. Cognition. 2004;92:145–77. doi: 10.1016/j.cognition.2003.11.002. [DOI] [PubMed] [Google Scholar]
  • 52.Lewis G, Poeppel D. The role of visual representations during the lexical access of spoken words. Brain Lang. 2014 Apr 30; doi: 10.1016/j.bandl.2014.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Pallier C, Devauchelle AD, Dehaene S. Cortical representation of the constituent structure of sentences. Proc Natl Acad Sci U S A. 2011;108:2522–7. doi: 10.1073/pnas.1018711108. The parametric fMRI experiment described tests how local syntactic structures of different sizes, i.e. constituent structures, are built. The results suggest a segregation into areas sensitive to structural features alone as wells complementary areas that exploit structure-meaning pairings.
  • 54.Tyler LK, Cheung TP, Devereux BJ, Clarke A. Syntactic computations in the language network: Characterizing dynamic network properties using representational similarity analysis. Front Psychol. 2013;4:271. doi: 10.3389/fpsyg.2013.00271. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Obleser J, Lahiri A, Eulitz C. Magnetic brain response mirrors extraction of phonological features from spoken vowels. J Cogn Neurosci. 2004;16:31–9. doi: 10.1162/089892904322755539. [DOI] [PubMed] [Google Scholar]
  • 56.Phillips C, Pellathy T, Marantz A, Yellin E, Wexler K, Poeppel D, et al. Auditory cortex accesses phonological categories: An MEG mismatch study. J Cogn Neurosci. 2000;12:1038–55. doi: 10.1162/08989290051137567. [DOI] [PubMed] [Google Scholar]
  • 57. DeWitt I, Rauschecker JP. Phoneme and word recognition in the auditory ventral stream. Proc Natl Acad Sci U S A. 2012;109:505–14. doi: 10.1073/pnas.1113427109. This meta-analysis of imaging data argues for a left lateralized, two-stage model of speech and auditory word-form recognition, with the analysis of phonemes occurring in mid-STG and word recognition occurring in anterior STG. The opens a new debate, identifying a region (anterior STG) typically not implicated in word recognition.
  • 58.Giraud AL, Poeppel D. Cortical oscillations and speech processing: Emerging computational principles and operations. Nat Neurosci. 2012;15:511–7. doi: 10.1038/nn.3063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Luo H, Poeppel D. Phase patterns of neuronal responses reliably discriminate speech in human auditory cortex. Neuron. 2007;54:1001–10. doi: 10.1016/j.neuron.2007.06.004. This is the first in a series of papers from different labs that argues that cortical oscillations entrain to continuous speech. In particular, the role of the phase pattern of theta band oscillations is implicated.
  • 60.Luo H, Liu Z, Poeppel D. Auditory cortex tracks both auditory and visual stimulus dynamics using low-frequency neuronal phase modulation. PLoS Biol. 2010;8(8):e1000445. doi: 10.1371/journal.pbio.1000445. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Luo H, Poeppel D. Cortical oscillations in auditory perception and speech: Evidence for two temporal windows in human auditory cortex. Front Psychol. 2012;3:170. doi: 10.3389/fpsyg.2012.00170. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Luo H, Tian X, Song K, Zhou K, Poeppel D. Neural response phase tracks how listeners learn new acoustic representations. Curr Biol. 2013;23:968–74. doi: 10.1016/j.cub.2013.04.031. [DOI] [PubMed] [Google Scholar]
  • 63.Peelle JE, Gross J, Davis MH. Phase-locked responses to speech in human auditory cortex are enhanced during comprehension. Cereb Cortex. 2013;23:1378–87. doi: 10.1093/cercor/bhs118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64. Ghitza O. Linking speech perception and neurophysiology: Speech decoding guided by cascaded oscillators locked to the input rhythm. Front Psychol. 2011;2:130. doi: 10.3389/fpsyg.2011.00130. A computational model is presented that outlines how low-frequency cortical oscillations, mainly in the theta band, can be exploited to parse an input signal into chunks that are subsequently decoded using higher frequency beta and gamma oscillations.
  • 65.Doelling KB, Arnal LH, Ghitza O, Poeppel D. Acoustic landmarks drive delta-theta oscillations to enable speech comprehension by facilitating perceptual parsing. Neuroimage. 2014;85:761–8. doi: 10.1016/j.neuroimage.2013.06.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66. Ding N, Simon JZ. Emergence of neural encoding of auditory objects while listening to competing speakers. Proc Natl Acad Sci U S A. 2012;109:11854–9. doi: 10.1073/pnas.1205381109. The data from this MEG study indicate that concurrently present auditory objects, even if spectrotemporally overlapping and not resolvable at the auditory periphery (like a cocktail party), are encoded individually in the auditory cortex. Moreover, each stream emerges as a basic representational unit subject to for top-down attentional modulation.
  • 67.Mesgarani N, Chang EF. Selective cortical representation of attended speaker in multi-talker speech perception. Nature. 2012;485:233–6. doi: 10.1038/nature11020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Zion Golumbic E, Cogan GB, Schroeder CE, Poeppel D. Visual input enhances selective speech envelope tracking in auditory cortex at a "cocktail party". J Neurosci. 2013;33:1417–26. doi: 10.1523/JNEUROSCI.3675-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Zion Golumbic EM, Ding N, Bickel S, Lakatos P, Schevon CA, McKhann GM, et al. Mechanisms underlying selective neuronal tracking of attended speech at a "cocktail party". Neuron. 2013;77:980–91. doi: 10.1016/j.neuron.2012.12.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Hickok G. Computational neuroanatomy of speech production. Nat Rev Neurosci. 2012;13:135–45. doi: 10.1038/nrn3158. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71. Gagnepain P, Henson RN, Davis MH. Temporal predictive codes for spoken words in auditory cortex. Curr Biol. 2012;22:615–621. doi: 10.1016/j.cub.2012.02.015. The rather complex design of this MEG study allows the authors to investigate alternative models, testing the role of lexical competition versus segment-based prediction in auditory word recognition. The latter view wins, providing strong support for predictive models in language processing.
  • 72.Santoro R, Moerel M, De Martino F, Goebel R, Ugurbil K, Yacoub E, Formisano E. Encoding of Natural Sounds at Multiple Spectral and Temporal Resolutions in the Human Auditory Cortex. PLoS Comput Biol. 2014;10(1):e1003412. doi: 10.1371/journal.pcbi.1003412. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES