Abstract
The influence of native-language experience on sensory-obligatory auditory-evoked potentials (AEPs) was investigated in native-English and native-Polish listeners. AEPs were recorded to the first word in nonsense word pairs, while participants performed a syllable identification task to the second word in the pairs. Nonsense words contained phoneme sequence onsets (i.e., /pt/, /pət/, /st/ and /sət/) that occur in the Polish and English languages, with the exception that /pt/ at syllable onset is an illegal phonotactic form in English. P1–N1–P2 waveforms from fronto-central electrode sites were comparable in English and Polish listeners, even though, these same English participants were unable to distinguish the nonsense words having /pt/ and /pət/ onsets. The P1–N1–P2 complex indexed the temporal characteristics of the word stimuli in the same manner for both language groups. Taken together, these findings suggest that the fronto-central P1–N1–P2 complex reflects acoustic feature processing of speech and is not significantly influenced by exposure to the phoneme sequences of the native-language. In contrast, the T-complex from bilateral posterior temporal sites was found to index phonological as well as acoustic feature processing to the nonsense word stimuli. An enhanced negativity for the /pt/ cluster relative to its contrast sequence (i.e., /pət/) occurred only for the Polish listeners, suggesting that neural networks within non-primary auditory cortex may be involved in early cortical phonological processing.
Keywords: Auditory-evoked potential (AEP), P1–N1–P2 complex, T-complex, Cross-linguistic, Phonetic, Speech perception
1. Introduction
The manner in which acoustic streams of speech are transformed by the cortical auditory pathways into integrated signals necessary for speech perception is not yet well understood. Listeners of a language such as Polish and English have differing auditory experiences with the phoneme sequences that constitute words. Comparison of neural processing to native and non-native phoneme sequences in native-language groups can determine whether experience with specific phoneme sequences influences stages of cortical speech processing. The current research explores the effects of native-language experience on early stages of cortical speech processing.
P1, N1 and P2 are obligatory components of auditory evoked potentials (AEPs) that index detection of the onset and offset of auditory stimuli (Näätänen and Picton, 1987). The P1, N1 and P2 deflections each have differing underlying neural sources and independent response patterns (for review Crowley and Colrain, 2004; Näätänen and Picton, 1987; Steinschneider et al., 2011a, 2011b) and are modulated by attention (Näätänen et al., 1978; Picton and Hillyard, 1974).
The sequential acoustic features within a word elicit P1– N1–P2 responses, which overlap and result in a signature waveform for the phoneme sequence within the word (Martin and Boothroyd, 1999, 2000; Ostroff et al., 1998). Investigation of the P1–N1–P2 complex to speech has important clinical implications because it can demonstrate cortical access to the acoustic features within the speech signal necessary for speech perception (Martin et al., 2007, 2008).
The T-complex is a series of negative and positive deflections obtained from lateral electrode sites that is also sensitive to the acoustic properties of a stimulus (Näätänen and Picton, 1987; Wolpaw and Perry, 1975). These deflections are best observed from lateral temporal electrode sites because their main sources have a radial orientation (Näätänen and Picton, 1987; Tonnquist-Uhlen et al., 2003; Wolpaw and Perry, 1975). Source analyses based on non-invasive recordings have been confirmed through intracranial recordings obtained directly from the lateral superior temporal gyrus (Howard et al., 2000; Steinschneider et al., 2011a, 2011b). Attenuated T-complex measures serve as markers of language impairment, suggesting that the underlying sources may play a role in language development and processing (Shafer et al., 2011; Tonnquist-Uhlen, 1996).
1.1. Language experience
Physiological investigations of native-language speech perception have revealed that experience with the phonological patterns of one’s native-language influences relatively late stages of neural processing (Dehaene-Lambertz et al., 2000; Näätänen et al., 1997; Sharma and Dorman, 2000; Wagner et al., 2012). However, it is unclear whether experience with the phonological patterns of a listener’s native-language influences earlier stages of cortical speech processing.
Research investigating this question has been equivocal. Several studies examining obligatory AEPs recorded from midline sites and elicited by syllables varying in their voice onset time (VOT) have shown no effect of listeners’ native-language experience (Elangovan and Stuart, 2011; Sharma and Dorman, 2000). In contrast, Zhang et al. (2005) using magnetoencephalography (MEG) found two N1m peaks to the syllable /la/ for Japanese, but not American listeners and argued that only the American listeners showed integration of the features of the phoneme /l/.
Brief exposure to sounds during psychoacoustic training (e.g., a VOT non-native sound contrast) modulates the frontocentral N1 and P2 waves (for review Kujala and Näätänen, 2010; Reinke et al., 2003; Tremblay et al., 2001; Tremblay and Kraus, 2002), suggesting a change related to speech perception. Also, P2 may reflect sensory mechanisms that lead to conscious perception (Ceponiene et al., 2005, 2008, 2009; Tremblay and Kraus, 2002). For example, Ceponiene et al. (2005) suggested that P2 reflects a transition between acoustic feature detection and integrated sensory processing through facilitating or inhibiting conscious perception of speech (Ceponiene et al., 2005, 2008, 2009). In the current study, we take advantage of the fact that adult listeners have a lifetime of exposure to native-language sound contrasts. If neural generators contributing to the frontocentral P2 wave facilitate or inhibit conscious perception of speech, then a language groups’ ability to perceive (or not perceive) a sound contrast should affect P2.
Further, the T-complex has been shown to reflect features of linguistic processing (Friedrich et al., 2009; Shafer et al., 2011, see also Chang et al., 2010). Therefore, we concurrently examined the effect of acoustic (physical) and phonological (meaningful in a language) contrasts within natural nonsense word stimuli on the P1–N1–P2 and T-complex responses.
1.2. Overview
The current research was part of a larger project that examined the effects of phoneme context on speech perception utilizing the consonant cluster /pt/ that occurs in English in word offset (e.g., “except”), but not in word onset. In contrast, native-Polish listeners experience the /pt/ cluster in their language in both word onset and offset.
Perception of /pt/ was explored using the word onset sequences /pt/ and /pət/ in nonsense word stimuli. These sequences are linguistically contrastive in word onset in Polish (e.g., Polish: “ptak” translated as bird, “Petronella”- a name), but not in English (e.g., English: “petunia”). The phoneme sequences /st/ and /sət/ were used as experimental controls because both languages contrast these sequences in word onset (e.g., Polish: “stal” translated as steel, “suterena” translated as basement; English: “stay”, “sateen”). Fig. 1 shows acoustic waveforms for specific productions of the naturally recorded stimuli.
Fig. 1.
Acoustic waveforms for specific productions of the naturally recorded stimuli. Total word durations in parentheses follow each word label. A vertical line within each waveform corresponds to initial segment durations from word onset through the burst for /t/ illustrating the longer durations for words containing fricative onsets relative to stop-consonant onsets.
Participants heard same and different nonsense word pairs (e.g., /pətima-pətima/, /pətima-ptima/ “peteema-pteema”) and determined whether the second word in the pairs had two or three syllables. Polish listeners were able to distinguish the 2 and 3-syllable nonsense words that began with /pt/ or /pət/, but English listeners were not (see Wagner et al., 2012).
Currently, we examine the P1–N1–P2 and T-complex responses to the first word in the word pairs to determine whether native-language experience is reflected within early sensory-obligatory components. We hypothesized that native-language patterns of speech perception would not be reflected at the early stages of cortical processing indexed by the P1–N1–P2 complex (Elangovan and Stuart, 2011; Sharma and Dorman, 2000), but might be observed within the T-complex whose main sources lie over non-primary auditory cortex.
2. Results
2.1. P1–N1–P2 complex
The P1–N1–P2 waveforms measured at fronto-central sites were highly similar between English and Polish listeners for both the 2 and 3-syllable st words and the 2 and 3-syllable pt words (Fig. 2). This similarity occurred even though these same English participants could not distinguish the /pt/versus /pət/ word onset contrasts (Wagner et al., 2012). Thus, acoustic features of the phoneme sequence onsets were registered by the P1–N1–P2, irrespective of native-language experience.
Fig. 2.
The grand mean P1–N1–P2 waveforms and standard error of the mean (SEM) for 2 and 3-syllable st words (top) and pt words (bottom) for English (left) and Polish participants (right) at electrode site 4. Notice the similarity in the signature patterns for both groups and the more negative response by the Polish participants for all word forms. Brackets illustrate that 2 and 3-syllable waveforms diverge for a longer duration of approximately 100 ms for the st relative to the pt stimuli (see text). Positive=up.
Dominance of acoustic feature processing can be further demonstrated by comparing waveforms to the words having fricative onsets relative to stop-consonant onsets. Brackets in Fig. 2 illustrate that waveforms for the /st/ and /sət/ onsets diverged for approximately 100 ms longer (232–472 ms) than waveforms for the /pt/ and /pət/ onsets (232–376 ms) for both the Polish and English listeners (Table 1). Thus, sensory processing for the /st/ and /sət/ onsets was longer in duration than for the /pt/ and /pət/ onsets, which corresponds to the acoustics of the fricative onsets being 100 ms longer in duration than the stop-consonant onsets (see Fig. 1; also see Wagner et al., 2012). Thus, temporal processing of the /pt/ and /pət/ phoneme sequences was reflected within the P1–N1–P2 complex for the English listeners, even though these individuals were unable to behaviorally distinguish the sequences.
Table 1.
Waveform amplitude values differ significantly for the 2 and 3-syllable word forms at similar time intervals at fronto-central and bilateral posterior temporal sites for the (a) st conditions and (b) pt conditions.
| P1–N1–P2 | T-complex LA | T-complex LP and RP |
|---|---|---|
| (a) | ||
| 232 ms* | 232 ms | 232 ms* |
| 256 ms* | 256 ms | 256 ms* |
| 280 ms* | 280 ms | 280 ms* |
| 304 ms | 304 ms | 304 ms |
| 328 ms* | 328 ms | 328 ms* |
| 352 ms* | 352 ms* | 352 ms* |
| 376 ms | 376 ms* | 376 ms |
| 400 ms* | 400 ms | 400 ms |
| 424 ms* | 424 ms | 424 ms* |
| 448 ms* | 448 ms | 448 ms* |
| (b) | ||
| 232 ms* | 232 ms* | 232 ms |
| 256 ms* | 256 ms* | 256 ms* |
| 280 ms | 280 ms | 280 ms |
| 304 ms* | 304 ms | 304 ms* |
| 328 ms* | 328 ms | 328 ms* |
| 352 ms* | 352 ms | 352 ms* |
Asterisk *=p<.05 Tukey HSD.
While the overall profile of the signature waveforms elicited by the 2 and 3-syllables were comparable between English and Polish listeners, later and slower components of the AEPs do diverge between the two language groups. AEPs recorded from Polish listeners were, in general, more negative for all conditions relative to the English listeners, suggesting processing differences between the two language groups. This finding is illustrated in Fig. 2, which shows that the waveforms elicited by /st/ and /sət/ became more negative for the Polish group beginning at 280 ms (280 through 496 ms: F (1, 22)=9.99; p=.005; partial eta squared=.312) and waveforms elicited by /pt/ and /pət/ became more negative for the Polish group beginning at 88ms (88 though 256 ms: F (1, 22)= 4.876; p=.038; partial eta squared=.181; 280 through 496 ms: F (1, 22)=10.726; p=.003; partial eta squared=.328).
2.2. T-complex
As for the P1–N1–P2 complex, the acoustic distinctions within the st and pt 2 and 3-syllable words were reflected within the T-complex. Table 1 illustrates that temporal distinctions in the acoustic contrasts (i.e., /st/sət/ and /pt/ pət/) reflected in the fronto-central P1–N1–P2 complex were similarly reflected in the T-complex from bilateral posterior temporal sites.
Also, T-complex waveforms for the /pt/ and /pət/ onsets diverged at the same time intervals for English and Polish listeners. Thus, neural networks underlying the T-complex were sensitive to the acoustic distinctions within the /pt/ and /pət/ onsets for both native and non-native listeners.
In contrast to the P1–N1–P2 complex, the T-complex from bilateral posterior temporal sites during the same time period showed a language effect specific to the 2-syllable /pt/ words, the legal word form in Polish, but illegal in English. Analysis of the T-complex revealed that the Polish listeners showed a more negative T-complex to the 2-syllable pt words relative to the 3-syllable words between 40 and 256 ms (F (1, 22)=8.516; p=.008; partial eta squared=.28) at bilateral posterior temporal sites (Fig. 3).
Fig. 3.
The T-complex and standard error of the mean (SEM) to the 2 and 3-syllable pt word forms from left and right-posterior temporal sites for the English (top) and Polish (bottom) participants. Notice the increased negativity to the 2-syllable pt word forms relative to the 3-syllable word forms only by the Polish listeners. Positive=up.
3. Discussion
3.1. P1–N1–P2 complex
The current research revealed that native-language phonotactic patterns were not reflected within the P1–N1–P2 complex, but rather responses were dominated by acoustic contrasts in a non-language specific manner. The 2 and 3- syllable pt words contain the word onset /pt-pət/ contrast that is a linguistic contrast in the Polish language, but not in the English language, and yet, the P1–N1–P2 signature waveforms differentiated the 2 and 3-syllable pt words for both the English and Polish listeners in the same manner. The current study supports research that has shown the P1–N1–P2 complex to primarily reflect acoustic, but not phonological aspects of the speech stimuli (Elangovan and Stuart, 2011; Sharma and Dorman, 2000). Contrasting results demonstrating different N1m responses by Japanese and English listeners may reflect particular feature processing for the /l/ phoneme or may reflect differences in experimental design that affect attention (Zhang et al., 2005).
Previous research suggests that the P2 obligatory AEP is sensitive to auditory experience (Tremblay et al., 2001; Tremblay and Kraus, 2002; Sheehan et al., 2005; Reinke et al., 2003). Exposure to one’s native-language provides the ultimate practice of a skill, speech perception of native-language contrasts. Why, then, does the Polish group not demonstrate a different P1–N1–P2 response to the 2-syllable /pt/ words?
Tremblay and Kraus (2002) suggested that the increased P2 response subsequent to training a non-native VOT distinction could have reflected the learning process because post-training modifications in behavioral perception and amplitude values were not correlated for individual subjects. This view is further supported by Atienza et al. (2004), who demonstrated an increased MMN subsequent to training a frequency deviation within a tone sequence; ERPs collected post training revealed a slow positive wave overlapping the P2 response attributed to new learning. The Polish listeners had highly established native-language sound patterns and P2 modifications attributed to new learning would not have been expected. Thus, the current results are consistent with the view that P2 modifications following non-native sound contrast training reflect processes involved in new learning, but not phonological perception.
The Polish listeners’ P1–N1–P2 complex showed an enhanced negativity for all four of the word types. This enhanced negativity may suggest that the Polish group engaged greater resources of attention (Näätänen, 1990; Näätänen et al., 2011) during the experiment relative to the English listeners. It is possible that these Polish participants, who learned English as young adults (i.e., late bilinguals), listened more carefully to all stimuli because they recognized the speaker as an early bilingual. It is equally possible that this group of Polish participants was more engaged in the syllable identification task because they recognized Polish language contrasts that are not present in English within the stimulus set. Future research will be necessary to determine the effects of attention on neural speech processing in bilingual and monolingual listeners.
The current results do not support the claim that P2 reflects modulation of access to speech perception or that the P2 reflects a pre-attentive alerting mechanism that contributes to improved phonological perception (Ceponiene et al., 2002, 2008, 2009; Tremblay and Kraus, 2002). Polish listeners’ P2 responses to words having the /pt/ onset did not differ from the English listeners’ responses despite years of experience with this onset cluster and its contrast sequence (i.e., /pət/). Ceponiene and colleagues’ argue that the P2 peak indexes a transition between acoustic feature detection and integrated speech processing and implies a hierarchical serial network with neural generators underlying the P2 response to lead to neural processes contributing to speech perception. In contrast, our results support additional parallel acoustic and linguistic processing networks (Näätänen et al., 2011; Näätänen and Winkler, 1999; Davis and Zerlin, 1966) because early language-group differences in response to the 2 and 3-syllable pt words were found from lateral posterior temporal sites, but not fronto-central sites.
3.2. T-complex
An increased negativity recorded from bilateral posterior temporal sites was restricted to the Polish listeners and specific for the /pt/ onset cluster relative to its contrast sequence (i.e., /pət/). This negativity specific for the /pt/ onset cluster suggests increased sound analysis by the Polish listeners to a phoneme sequence that is present only in the Polish language (Katarzyna Dziubalska-Kolaczyk, personal communication, January 9, 2010; also see Wagner et al., 2012) and thus, suggests an effect of language experience. Interestingly, the T-complex has its main sources underlying bilateral posterior temporal cortex (Howard et al., 2000) and phonological distinctions have been mapped to this region through electrocorticographic recordings from the exposed lateral temporal cortex (Chang et al., 2010).
Language group differences would be expected at higher levels of speech processing. Previous research had shown that minimal phonological differences within words were reflected from lateral temporal electrode sites at later stages of cortical processing around 350ms (Friedrich et al., 2009). The current research suggests an even earlier effect, given the enhanced negativity seen in Polish listeners for the /pt/ onset cluster.
3.3. Auditory signatures waveforms
The P1–N1–P2 waveforms and the T-complex showed longer duration sensory processing for the 2 and 3-syllable st words relative to the 2 and 3-syllable pt words, consistent with the acoustic characteristics of the stimuli. However, the P1–N1–P2 waveforms for both the pt and st stimuli did not begin to diverge until 232 ms, subsequent to the N1 response, even though the contrasting vowel within the /pt-pət/ phoneme sequences began at approximately 26 ms and the contrasting vowel within the /st-sət/ sequences began at approximately 122 ms. Interestingly, within the same time frame from bilateral temporal sites, the /pt-pət/ contrast was differentiated by the Polish listeners, suggesting feature integration within underlying auditory-cortical networks.
3.4. Conclusion
Current findings suggest that the P1–N1–P2 and T-complex signature waveforms are sensitive to the acoustic characteristics of speech. In addition, the T-complex reflected language experience, suggesting that neural networks within nonprimary auditory cortex from bilateral posterior temporal brain regions may be involved in early stage linguistic processing. The timing of this activity supports the view that linguistic and acoustic processing of speech occurs in parallel, as well as hierarchical, serial networks.
4. Experimental procedures
4.1. Participants
Twelve native-English (8 female) adults between the ages of 24 and 35 (mean age=29) and 12 native-Polish (8 female) adults between the ages of 23 and 34 (mean=30) participated in the study. Native-Polish listeners were bilingual speakers who had learned English as young adults, after the age of 15 years. The English listeners reported having no experience with Polish or other Slavic languages. All participants demonstrated normal hearing and were without a history of speech, language or cognitive impairment.
4.2. Stimuli
Stimuli consisted of 800 same and different nonsense word pairs. Nonsense words began with /pt/, /pət/, /st/ and /sət/ and were potential real words in the Polish and English languages with the exception of nonsense words that began with /pt/, an illegal phonotactic form in English. Phoneme sequences that followed the /pt/, /pət/, /st/ and /sət/ onsets in the words varied and words in the pt and st conditions were matched for rhyme (e.g., /ptɛsa/, /pətɛsa/, /stɛsa/, /sətɛsa/). Each word type (e.g., /st/words) consisted of 2 natural productions of 50 nonsense words (e.g., /stIna/). These 100 natural productions were presented twice within the experiment, once within a same pair and once within a different pair for a total of 200 trials per word type. Stimuli were produced by a male native-Polish speaker who emigrated from Poland to the United States with his family at age 6 years. The speaker learned English in school, but continued to converse only in Polish at home. The Polish participants described the speaker as having good Polish pronunciation. He spoke English without a Polish accent.
Stimuli were presented through speakers in a sound-treated electrically shielded room using E-prime (Version 1.1) software. Word pairs were presented with a 2 s inter-trial interval (ITI) and 250ms inter-stimulus interval (ISI). Procedures were identical to those reported in Wagner et al. (2012) with the exception that, for the current study, ERP epochs time-locked to the first word in the word pairs were analyzed.
4.3. EEG acquisition and analysis
A 65 channel net was used for data collection (Electrical Geodesic Inc.). Prior to testing, we confirmed that impedance levels were at or below 40 kOhms, an acceptable level for the high-impedance amplifiers used (Ferree et al., 2001). The EEG was recorded at a sampling rate of 250 Hz, bandpass filtered between .1 and 30 Hz and referenced to Cz. Net station software (Version 4.1.2) was used for data processing (i.e., segmentation, artifact rejection, averaging, baseline correction). The continuous EEG wave was segmented into an 1800 ms segment, which consisted of a 200ms baseline and a 1600ms post word onset segment. Data was baseline corrected between −100 ms and 0ms and re-referenced to an average reference.
The P1–N1–P2 complex was obtained from fronto-central site 4, left fronto-central site 5, site 65 (Cz) and right fronto-central site 55. The T-complex was obtained from left-anterior temporal site 20, right-anterior temporal site 56, left-posterior temporal site 24 and right-posterior temporal site 52 (see Wagner et al., 2012).
The mean number of epochs averaged (and range) for each participant group and word type follows. For the Polish participants: a mean of 161 epochs (range 68–197) was averaged for the /pt/ word type, 162 epochs (range 69–194) for the /pət/ word type, 158 (range 67–200) for the /st/ word type and 158 epochs (range 59–195) for the /sət/ word type. For the English participants: a mean of 166 epochs (range 82–193) were averaged for the /pt/ word type, 166 epochs (range 101–197) for the /pət/ word type, 166 (range 88–192) for the /st/ word type and 167 epochs (range 83– 200) for the /sət/ word type. Raw voltage amplitudes were averaged across 24 ms time intervals for statistical analysis. A mixed ANOVA with group (12 native-English and 12 native-Polish participants) as the between subject factor and condition (2 or 3-syllable word forms), time and site as the within subject factors was used. Greenhouse-Geisser correction was applied when appropriate and Tukey’s HSDwas used for post hoc testing (p<.05). Main effects of condition and language group and significant interactions involving condition were reported.
Acknowledgments
This publication was made possible by grant number HD-46193 from the National Institute of Child Health and Human Development (NICHD) at the National Institutes of Health (NIH) to Valerie Shafer. The contents of this work are solely the responsibility of the authors and do not necessarily represent the official views of NIH.
We acknowledge linguists Katarzyna Dziubalska-Kolaczyk, Malgosia Pyrzanowski, and Zosia Stankiewicz for providing information about the phonetic and phonotactic rules of the Polish language. We thank Dr. Erich Schröger and the reviewer who provided excellent suggestions for an improved manuscript.
REFERENCES
- Atienza M, Cantero J, Stickgold R. Post-training sleep enhances automaticity in perceptual discriminationJ. Cogn. Neurosci. 2004;16:53–64. doi: 10.1162/089892904322755557. [DOI] [PubMed] [Google Scholar]
- Ceponiene R, Alku M, Westerfield M, Torki M, Townsend J. ERPs differentiate syllable and nonphonetic sound processing in children and adults. Psychophysiology. 2005;42:391–406. doi: 10.1111/j.1469-8986.2005.00305.x. [DOI] [PubMed] [Google Scholar]
- Ceponiene R, Cummings A, Wulfeck B, Ballantyne A, Townsend J. Spectral vs temporal auditory processing in specific language impairment: a developmental ERP study. Brain Lang. 2009;110:107–120. doi: 10.1016/j.bandl.2009.04.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ceponiene R, Torki M, Alku M, Koyama A, Townsend J. Event-related potentials reflect spectral differences in speech and nonspeech stimuli in children and adults. Clin. Neurophysiol. 2008;119:1560–1577. doi: 10.1016/j.clinph.2008.03.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ceponiene R, Rinne T, Näätänen R. Maturation of cortical sound processing as indexed by event-related potentials. Clin. Neurophysiol. 2002;113:870–882. doi: 10.1016/s1388-2457(02)00078-0. [DOI] [PubMed] [Google Scholar]
- Chang E, Rieger J, Johnson K, Berger M, Barbaro N, Knight R. Categorical speech representation in human superior temporal gyrus. Nat. Neurosci. 2010;13:1428–1432. doi: 10.1038/nn.2641. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Crowley K, Colrain I. A review of the evidence for P2 being an independent component process: age, sleep and modality. Clin. Neurophysiol. 2004;115:732–744. doi: 10.1016/j.clinph.2003.11.021. [DOI] [PubMed] [Google Scholar]
- Davis H, Zerlin S. Acoustic relations of the human vertex potential. J. Acoust. Soc. Am. 1966;39:109–116. doi: 10.1121/1.1909858. [DOI] [PubMed] [Google Scholar]
- Dehaene-Lambertz G, Dupoux E, Gout A. Electrophysiological correlates of phonological processing: a cross-linguistic study. J. Cogn. Neurosci. 2000;12:635–647. doi: 10.1162/089892900562390. [DOI] [PubMed] [Google Scholar]
- Elangovan S, Stuart A. A cross-linguistic examination of cortical auditory evoked potentials for a categorical voicing contrast. Neurosci. Lett. 2011;490:140–144. doi: 10.1016/j.neulet.2010.12.044. [DOI] [PubMed] [Google Scholar]
- Ferree TC, Luu P, Russell GS, Tucker DM. Scalp electrode impedance, infection risk, and EEG data quality. Clin. Neurophysiol. 2001;112:536–544. doi: 10.1016/s1388-2457(00)00533-2. [DOI] [PubMed] [Google Scholar]
- Friedrich CK, Schild U, Roder B. Electrophysiological indices of word fragment priming allow characterizing neural stages of speech recognition. Biol. Psychol. 2009;80:105–113. doi: 10.1016/j.biopsycho.2008.04.012. [DOI] [PubMed] [Google Scholar]
- Howard M, Volkov I, Mirsky R, Garell P, Noh M, Granner M, Damasio H, Steinschneider M, Reale R, Hind J, Brugge J. Auditory cortex on the human posterior superior temporal gyrus. J. Comp. Neurol. 2000;416:79–92. doi: 10.1002/(sici)1096-9861(20000103)416:1<79::aid-cne6>3.0.co;2-2. [DOI] [PubMed] [Google Scholar]
- Kujala T, Näätänen R. The adaptive brain: a neurophysiological perspective. Prog. Neurobiol. 2010;91:55–67. doi: 10.1016/j.pneurobio.2010.01.006. [DOI] [PubMed] [Google Scholar]
- Martin B, Boothroyd A. Cortical, auditory, event-related potentials in response to periodic and aperiodic stimuli with the same spectral envelope. Ear Hear. 1999;20:33–44. doi: 10.1097/00003446-199902000-00004. [DOI] [PubMed] [Google Scholar]
- Martin B, Boothroyd A. Cortical, auditory, event-related potentials in response to changes of spectrum and amplitude. J. Acoust. Soc. Am. 2000;107:2155–2161. doi: 10.1121/1.428556. [DOI] [PubMed] [Google Scholar]
- Martin B, Tremblay K, Korczak P. Speech evoked potentials: from the laboratory to the clinic. Ear Hear. 2008;29:285–313. doi: 10.1097/AUD.0b013e3181662c0e. [DOI] [PubMed] [Google Scholar]
- Martin B, Tremblay K, Stapells D. Principles and application of cortical auditory evoked potentials. In: Burkard R, Don M, Eggermont J, editors. Auditory Evoked Potentials: Basic Principles and Clinical Application. New York: Lippincott, Williams & Wilkins; 2007. [Google Scholar]
- Näätänen R. The role of attention in auditory information processing as revealed by event-related potentials and other brain measures of cognitive function. Behav. Brain Sci. 1990;13:201–288. [Google Scholar]
- Näätänen R, Gaillard A, Mantysalo S. Early selectiveattention effect on evoked potential reinterpreted. Acta Psychol. 1978;42:313–329. doi: 10.1016/0001-6918(78)90006-9. [DOI] [PubMed] [Google Scholar]
- Näätänen R, Kujala T, Winkler I. Auditory processing that leads to conscious perception: a unique window to central auditory processing opened by the mismatch negativity and related responses. Psychophysiology. 2011;48:4–22. doi: 10.1111/j.1469-8986.2010.01114.x. [DOI] [PubMed] [Google Scholar]
- Näätänen R, Lehtokoski A, Lennes M, Cheour M, Huotllainen M, Livonen A, Vainio M, Alku P, IImoniemi R, Luuk A, Allik J, Sinkkonen J, Alho K. Language-specific phoneme representations revealed by electric and magnetic brain responses. Nature. 1997;385:432–434. doi: 10.1038/385432a0. [DOI] [PubMed] [Google Scholar]
- Näätänen R, Picton T. The N1 wave of the human electric and magnetic response to sound: a review and an analysis of the component structure. Psychophysiology. 1987;24:375–425. doi: 10.1111/j.1469-8986.1987.tb00311.x. [DOI] [PubMed] [Google Scholar]
- Näätänen R, Winkler I. The concept of auditory stimulus representation in cognitive neuroscience. Psychol. Bull. 1999;126:826–859. doi: 10.1037/0033-2909.125.6.826. [DOI] [PubMed] [Google Scholar]
- Ostroff J, Martin B, Boothroyd A. Cortical evoked response to acoustic change within a syllable. Ear Hear. 1998;19:290–297. doi: 10.1097/00003446-199808000-00004. [DOI] [PubMed] [Google Scholar]
- Picton T, Hillyard S. Human auditory evoked potentials. II Effects of attention. Electroencephalogr. Clin. Neurophys. 1974;36:191–199. doi: 10.1016/0013-4694(74)90156-4. [DOI] [PubMed] [Google Scholar]
- Reinke K, He Y, Wang C, Alain C. Perceptual learning modulates sensory evoked response during vowel segregation. Cogn. Brain Res. 2003;17:781–791. doi: 10.1016/s0926-6410(03)00202-7. [DOI] [PubMed] [Google Scholar]
- Shafer VL, Schwartz RG, Martin B. Evidence of deficient central speech processing in children with specific language impairment: the T-complex. Clin. Neurophysiol. 2011;122:1137–1155. doi: 10.1016/j.clinph.2010.10.046. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sharma A, Dorman M. Neurophysiologic correlates of cross-language phonetic perceptionJ. Acoust. Soc. Am. 2000;107:2697–2703. doi: 10.1121/1.428655. [DOI] [PubMed] [Google Scholar]
- Sheehan K, McArthur G, Bishop D. Is discrimination training necessary to cause changes in the P2 auditory eventrelated brain potential to speech sounds. Cogn. Brain Res. 2005;25:547–553. doi: 10.1016/j.cogbrainres.2005.08.007. [DOI] [PubMed] [Google Scholar]
- Steinschneider M, Liegeois-Chauvel C, Brugge JF. Auditory evoked potentials and their utility in the assessment of complex sound processing. In: Winer JA, Schreiner CE, editors. The Auditory Cortex. New York: Springer-Verlag; 2011a. pp. 535–560. [Google Scholar]
- Steinschneider M, Nourski K, Kawasaki H, Oya H, Brugge J, Howard M., III Intracranial study of speech-elicited activity on human posterolateral superior temporal gyrus. Cerebral Cortex. 2011b;21:2332–2347. doi: 10.1093/cercor/bhr014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tonnquist-Uhlen I. Topography of auditory evoked longlatency potentials in children with severe language impairment: the P2 and N2 components. Ear Hear. 1996;17:314–326. doi: 10.1097/00003446-199608000-00003. [DOI] [PubMed] [Google Scholar]
- Tonnquist-Uhlen I, Ponton C, Eggermont J, Kwong B, Don M. Maturation of human central auditory system activity: the T-complex. Clin. Neurophysiol. 2003;114:685–701. doi: 10.1016/s1388-2457(03)00005-1. [DOI] [PubMed] [Google Scholar]
- Tremblay K, Kraus N. Auditory training induces asymmetrical changes in cortical neural activity. J. Speech Lang. Hear. Res. 2002;45:564–572. doi: 10.1044/1092-4388(2002/045). [DOI] [PubMed] [Google Scholar]
- Tremblay K, Kraus N, McGee T, Ponton C, Otis B. Central auditory plasticity: changes in the N1-P2 complex after speech-sound training. Ear Hear. 2001;22:70–90. doi: 10.1097/00003446-200104000-00001. [DOI] [PubMed] [Google Scholar]
- Wagner M, Shafer VL, Martin B, Steinschneider M. The phonotactic influence on the perception of a consonant cluster/pt/ by native English and native Polish listeners: a behavioral and event related potential (ERP) study. Brain Lang. 2012;123:30–41. doi: 10.1016/j.bandl.2012.06.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wolpaw JR, Perry JK. A temporal component of the auditory evoked response. Electroencephalogr. Clin. Neurophysiol. 1975;39:609–620. doi: 10.1016/0013-4694(75)90073-5. [DOI] [PubMed] [Google Scholar]
- Zhang Y, Kuhl P, Imada T, Kotani M, Tohkura Y. Effects of language experience: neural commitment to language-specific auditory patterns. NeuroImage. 2005;26:703–720. doi: 10.1016/j.neuroimage.2005.02.040. [DOI] [PubMed] [Google Scholar]



