Abstract
Categorical perception, an increased sensitivity to between- compared to within-category contrasts, is a stable property of native speech perception that emerges as language matures. While recent research suggests that categorical responses to speech sounds can be found in left prefrontal as well as temporo-parietal areas, it is unclear how the neural system develops heightened sensitivity to between-category contrasts. In the current study, two groups of adult participants were trained to categorize speech sounds taken from a dental-retroflex-velar continuum according to two different boundary locations. Behavioral results suggest that for successful learners, categorization training led to increased behavioral sensitivity for between-category contrasts with no concomitant increase for within-category contrasts. Neural responses to the learned category schemes were measured using a short-interval habituation design during fMRI scanning. While both inferior frontal and temporal regions showed sensitivity to phonetic contrasts sampled from the continuum, only the bilateral middle frontal gyri exhibited a pattern consistent with encoding of the learned category scheme. Taken together, these results support a view in which top-down information about category membership may reshape perceptual sensitivities via attention or executive mechanisms in the frontal lobes.
INTRODUCTION
Research on speech perception has proposed an intriguing duality: the existence of a perceptual system for processing speech that is both flexible and stable. On the one hand, there is evidence of apparent plasticity in accommodating acoustic variations due to lexical context (Ganong, 1980) speech rate (Miller, 1981) and speaker identity (Nygaard & Pisoni, 1998), while at the same time, perceptual sensitivities to native-language speech contrasts acquired in childhood appear to be permanently ingrained, and in fact may interfere with the subsequent acquisition of non-native contrasts in adulthood (Best & McRoberts, 2003). Such a dichotomy is particularly observable in the categorical perception of native-language speech. Adults have a robust tendency to perceive speech stimuli that fall into two different categories (e.g. ‘d’ and ‘t’) as more distinctive than stimuli that fall into the same category (e.g. ‘d1’ and ‘d2’) even when they are equally spaced on a physical continuum (Liberman, Delattre, & Cooper, 1958; Liberman, Harris, Hoffman, & Griffith, 1957; Pisoni & Tash, 1974), a phenomenon known as categorical perception. While native language categories learned over the course of a lifetime likely have ingrained, stable representations, categorical perception, defined behaviorally as selective improvements in discrimination of between-category contrasts, can be induced in adults in the short term via category training (Swan & Myers, under review; Golestani & Zatorre, 2009). In the current study, participants were trained to categorize tokens along a non-native speech continuum in order to examine the neural systems underlying the contribution of top-down category-level information to emerging categorical perception.
Perception of non-native speech contrasts is notoriously difficult for adults, although with sufficient training, participants show gains in their ability to categorize sounds from non-native speech continua (Bradlow, Pisoni, Akahane-Yamada, & Tohkura, 1997; Flege, MacKay, & Meador, 1999; Golestani & Zatorre, 2009). The emergence of discontinuous perception may be impacted by the manner in which categories are learned (Guenther, Husain, Cohen, & Shinn-Cunningham, 1999), how the sounds map onto native language categories or minimal pairs in the learning environment (Best, McRoberts, & Goodell, 2001; Yeung & Werker, 2009) or even the way attention is distributed among category-relevant features (Francis & Nusbaum, 2002; Francis, Baldwin, & Nusbaum, 2000). For example, second language learners often receive explicit instruction as to the nature of phonetic categories in the new language, which may be reinforced by a mapping between the new contrast and the orthography of the language. Previous behavioral work from our lab (Swan & Myers, under review) and others (Golestani & Zatorre, 2009; McCandliss, Fiez, Protopapas, Conway, & McClelland, 2002) suggests that participants who successfully learn to categorize items from a non-native speech continuum using labels also show increased ability to discriminate tokens that cross the category boundary, with no change in discrimination accuracy for tokens within a learned category. This asymmetry in the ability to discriminate between-category and within-category contrasts provides a behavioral measure of categorical perception.
Ultimately, category learning may impact between-category discrimination accuracy via at least two, mutually compatible routes. First, category training may result in a re-warping of perceptual sensitivities to speech sounds (Guenther & Gjaja, 1996; Guenther et al., 1999; Guenther, Nieto-Castanon, Ghosh, & Tourville, 2004; Kuhl, 1991), effectively retuning neural sensitivity to tokens that cross a category boundary. An alternative hypothesis is that improvements in discrimination for between-category contrasts instead reflect the application of a decision threshold once the learned category boundary has been established. According to this view, the underlying perceptual space is not altered, but rather category learning is mediated by cognitive processes; for example, the distribution of attention toward meaningful differences in the speech signal and away from irrelevant distinctions (Francis & Nusbaum, 2002; Francis et al., 2000). Note that these two accounts are not entirely incompatible: namely, it may be the case that the redistribution of attention to relevant parts of the acoustic signal in the short term leads to retuning of perceptual sensitivities to phonetic contrasts over time. It is assumed that behavioral sensitivity to within- and between-category contrasts has its root in neural sensitivity of category distinctions. However, it is unclear where in the neural processing stream category-level information is encoded, and in particular, whether such coding suggests a retuning of perceptual space, or rather the top-down imposition of a learned category scheme. Neuroimaging studies, in particular fMRI, offer a tool for exploring these mechanisms.
Neural systems underlying phonetic category learning
Perception of native language phonetic category continua has revealed a network of areas involving frontal (IFG, precentral) and posterior (STG, SMG) areas (Binder, Liebenthal, Possing, Medler, & Ward, 2004; Blumstein, Myers, & Rissman, 2005; Guenther et al., 2004; Liebenthal et al., 2010; Myers, Blumstein, Walsh, & Eliassen, 2009). Broadly speaking, these studies suggest a division of labor between temporal and frontal areas. The left superior temporal gyrus (STG) shows sensitivity to the mapping between the fine-grained acoustic details of the signal and phonetic category space, and responds to acoustic variation both within and between phonetic categories (Myers, 2007; Myers et al., 2009). In contrast, frontal areas show patterns of activation which reflect a greater reliance on the details of the category structure: the left inferior frontal gyrus (IFG) responds preferentially to tokens which are ambiguous, and as such may be involved in resolving competition at the level of the phonetic category (Binder et al., 2004; Myers, 2007). An adjacent area which spans BA 44 and precentral gyrus has been shown to exhibit a pattern of phonetic category invariance: namely, it responds to phonetic contrasts that cross the category boundary, but not to acoustically-equivalent contrasts within a phonetic category (Myers, et al., 2009). Taken together, these results are consistent with a model in which the acoustic-phonetic details of the speech signal are processed in the left (or perhaps bilateral) STG, and phonetic categorization processes occur in the IFG and precentral gyrus.
Investigations of the neural systems activated in categorizing non-native speech sounds before and after training have examined native English speakers learning a non-native dental/retroflex contrast (Golestani & Zatorre, 2004), a lexical tone contrast (Wang, Sereno, Jongman, & Hirsch, 2003), and Japanese speakers learning a non-native r/l contrast (Callan et al., 2003). In all three studies, participants showed post-training increases in activation for phonetic categorization tasks in a network of areas associated with performance on native language phonetic categorization tasks (Blumstein et al., 2005; Myers, 2007), including the bilateral STG, left supramarginal gyrus (SMG), and sometimes left or bilateral IFG. However, these studies were not designed to consider the question of how categorization training in turn affects sensitivity to within and between-category contrasts in the learned acoustic-phonetic space. Only a handful of studies have examined the relationship between a category structure imposed on a novel acoustic space and the brain’s sensitivity to exemplars from this space, and these studies generally support the notion that regions in the STG show sensitivity to the learned category structure (Guenther et al., 2004; Leech, Holt, Devlin, & Dick, 2009). Given that the left STG has been implicated in the perception of fine-grained details of native-language phonetic categories (Chang et al., 2010; Formisano, De Martino, Bonte, & Goebel, 2008; Liebenthal, Binder, Spitzer, Possing, & Medler, 2005; Liebenthal et al., 2010; Myers, 2007), a finding that the STG are also modulated by a learned phonetic category scheme would suggest that category training induces changes in the perceptual representation of speech sounds.
While the left or bilateral superior temporal areas may play an important role in detecting differences within and between acquired sound categories as well as during native language acquisition, frontal regions, especially the left IFG, are likely to play a role in the perception of learned speech categories. With regard to mature native-language categories, the left IFG has been implicated in segmenting phonemes from the surrounding phonetic context (Burton, Small, & Blumstein, 2000) and resolving competition between acoustically similar phonetic categories (Blumstein et al., 2005). Moreover, studies of non-native category learning suggest that inferior frontal regions are among the set of regions recruited in successful non-native sound categorization (Callan, Jones, Callan, & Akahane-Yamada, 2004; Callan et al., 2003; Golestani & Zatorre, 2004; Golestani, Molko, Dehaene, LeBihan, & Pallier, 2007). These results are all consistent with a role for inferior frontal areas in categorizing the sounds of speech in order to map them to an established language category. Modulation of activity in the left IFG and/or left precentral gyrus associated with the learned category structure, in the absence of modulation of temporal activity, would suggest that discontinuous behavioral performance along an acoustic-phonetic continuum is due to cognitive, rather than perceptual factors.
The current study was designed to investigate the nature of neural sensitivity to non-native phonetic contrasts after categorization training, and in particular to determine whether newly acquired sensitivities to non-native contrasts are encoded in temporal or frontal regions. Native English-speakers were trained to categorize items along a synthetic acoustic continuum which spanned three voiced stop categories, from a dental /d̪a/, to a retroflex /ɖa/, to a velar /ga/ sound (see Methods & Materials for details). While the endpoints approximate native-English phonemes (/da/ and /ga/), the retroflex category is a non-native category. Participants were trained to categorize tokens from this continuum into categories ‘A’ and ‘B’ using one of two different boundaries, and the behavioral consequences of categorization training were measured with an AX discrimination task. Neural sensitivity to the between and within-category contrasts was assessed using a short-interval habituation (SIH) paradigm in fMRI (Zevin & McCandliss, 2005). Participants listened to trains of five syllables, with the final syllable either the same or different from the first four. Regions showing increased activation for ‘different’ trials compared to ‘same’ trials are thought to either signify a release from adaptation (Grill-Spector & Malach, 2001; Grill-Spector, Henson, & Martin, 2006) or an active change detection response to the contrast (Zevin, Yang, Skipper, & McCandliss, 2010). Importantly, while in the scanner, participants were not required to explicitly categorize stimuli or discriminate phonetic contrasts, but instead monitored auditory stimuli for rare, high-pitched target stimuli. This implicit paradigm has the advantage of avoiding some of the additional, extra-linguistic task demands that accompany explicit categorization and discrimination tasks.
To examine possible changes in behavioral and neural sensitivity due to category learning, the current study compared stimulus pairs that form a between-category contrast for one training group, but a within-category contrast for the other group. Consistent with previous work (Joanisse, Zevin, & McCandliss, 2007; Myers et al., 2009), it was predicted that between-category contrasts would show greater activation than within-category contrasts, as dictated by the learned category scheme. Given the role of the left, and perhaps also right, STG in processing fine-grained details of the speech stream, between-group differences in activation for the same speech stimuli in temporal areas would be taken as evidence that categorization training results in a reshaping of the auditory sensitivities to tokens from the learned categories. Alternatively, between-group differences in activation in left inferior frontal regions would support the hypothesis that reshaped perceptual sensitivities reflect the application of a decision mechanism as tokens are mapped to the new category representation. Finally, sensitivity of both frontal and temporal regions to the learned category boundaries would suggest that changes in perceptual sensitivity are achieved via both reshaping of auditory sensitivities to learned speech contrasts as well as diversion of attention to the relevant parts of acoustic space.
METHODS AND MATERIALS
Participants
Twenty-nine adults (21 females, 8 males) between the ages of 18 and 45 years old were recruited from the Brown University community. Having completed a behavioral study of category training (Swan & Myers, under review), participants were invited to take part in a subsequent study using functional magnetic resonance imaging (fMRI). All participants were native speakers of American English, were right-handed, as confirmed by the Edinburgh Handedness Inventory, (Oldfield, 1971), and reported no hearing or neurological deficits. Informed consent was obtained and all participants were screened for ferromagnetic materials according to guidelines approved by the Human Subjects Committees of Brown University. Participants were compensated $30 for the two-hour session. Movement of more than 3mm (the width of one voxel) was cause for exclusion of 1 participant from further analyses; thus the total number of participants in the study was 28, with 14 (10 females, 4 males) in each group.
Stimuli
Speech syllables used in both the behavioral and fMRI testing sessions were taken from a synthetic continuum that varied in the onset frequencies of the second and third formants and in the frequency of the burst, constituting a nine-point acoustic continuum ranging from a dental /d̪a/ to a retroflex /ɖa/ to a velar /ga/ places of articulation (Stevens & Blumstein, 1975). Although Stevens and Blumstein (1975) report individual differences in placement of dental/retroflex and retroflex/velar boundaries among native speakers of languages utilizing the dental/retroflex/velar contrast, they found that in general, the first three points can be classified as Dental sounds, the middle three points as Retroflex sounds, and the remaining three points as Velar sounds. For native speakers of English, the endpoints of the continuum comprise a near-native phonetic contrast, while the dental and retroflex points are more difficult for these speakers to discriminate (Polka, 1991; Werker & Tees, 1984). A high-pitched version of the nine-point continuum was created by shifting the entire pitch contour of a stimulus upwards by 100 Hz using Praat (Boersma, 2001). These stimuli were used in the TARGET trials, replacing one of the five tokens in ten trials per run.
Procedure
In the weeks prior to this experiment, participants had completed a single 45-minute behavioral testing session which included: 1) an AX discrimination pre-test, 2) six blocks of categorization training with feedback, 3) an identification task without feedback, and 4) a final AX discrimination post-test. The mean number of days between testing sessions was 60.11 (SE = 7.31, median = 39, range = 9 to 147) and did not differ between training groups, t(26) = 1.24, p=0.228. Because it was unknown how much the effects of training would decay between sessions, participants underwent a replication of the behavioral training upon arriving at the scanner. Participants were assigned to two training groups, Dental/Retroflex and Retroflex/Velar, just as in their first session; thus their training was consistent with the category boundary with which they were already familiar. The replication included 1) the AX discrimination pre-test, 2) categorization training with feedback, and 3) the identification task without feedback. Participants listened to speech sounds over headphones and indicated their responses by pushing the appropriate buttons on the button box as quickly and accurately as possible.
Discrimination
The pre- and post-tests evaluated sensitivity to the dental/retroflex/velar contrasts before and after learning to categorize the tokens. Participants heard 60 pairs of syllables (separated by a 250ms ISI) from the three category centers, Dental (point 2), Retroflex (point 5), and Velar (point 8). Thirty pairs were repeated (e.g. 2 vs. 2), and 30 pairs were contrasts (e.g. 5 vs. 8). They judged whether the stimuli sounded the same or different from one another by pushing the corresponding button.
Categorization Training
Participants heard six blocks of 40 single-token trials, beginning with the endpoints of the continuum and stepping inward on each subsequent block to present progressively narrower phonetic contrasts. The Dental/Retroflex training group categorized phonetic contrasts converging on the boundary between points 3 and 4, whereas the Retroflex/Velar training group categorized contrasts converging on the boundary between points 6 and 7, as illustrated in Figure 1. Critically, neither group was exposed within a given block to the stimulus contrasts that were tested in the discrimination pre- and post-tests. Participants were familiarized with the labels “Category A” and “Category B”, and after each token was played, they pressed a button to indicate in which category the sound belonged. Auditory feedback (i.e. unique sounds for a correct or incorrect response) was given immediately after each response.
Identification
Following training, participants heard ten tokens of each sound on the nine-point continuum, presented in a random order for a total of 90 trials. They identified each token as either “Category A” or “Category B” by pushing a corresponding button and no feedback was given.
Immediately after replicating these behavioral tasks, participants completed a one-hour functional magnetic resonance imaging (fMRI) session. During the scanning session, participants alternated between two tasks: a high-pitch detection task during the functional runs and a shortened category training practice during the anatomical scan and between functional runs. Participants listened to speech sounds over headphones adjusted to ensure comfortable listening over the noise of the scanner, and they responded by pushing buttons on an MRI-compatible button box. They were instructed to keep their eyes closed and refrain from movement, especially of the head and upper body.
A short-interval habituation paradigm was used during the functional scans, with each trial containing five 266-ms tokens separated by 50-ms inter-stimulus intervals (see Figure 2). In SAME trials, a token from one of the three category centers, Dental (point 2), Retroflex (point 5), and Velar (point 8), was repeated five times, while in DIFFERENT trials, one token was repeated four times and followed by a contrasting fifth token. There were 3 functional runs, each containing 55 trials: 30 DIFFERENT trials (10 of each contrast), 15 SAME trials (5 of each) and 10 high-pitched TARGET trials.
High-pitch detection
Though participants heard SAME and DIFFERENT stimulus trains, their task was not to discriminate between the phonetic tokens. Rather, they were instructed to listen for rare TARGET trials, in which one of the five syllables was replaced by a high-pitched syllable, and to push a button whenever they detected one. TARGET trials, as well as trials in which participants erroneously detected a high-pitched target, were included in a separate stimulus regressor and were not analyzed further. Participants were highly accurate at detecting these stimuli, with only five errors across all participants.
Category training practice
To be sure that effects of training were maintained in the scanner, participants practiced categorizing sounds with feedback during acquisition of the anatomical scan and between functional runs. Participants completed a single 60-trial block with just 10 trials for each progressively narrower step of training. No functional data was acquired during these blocks. Finally, immediately after scanning participants completed one more 60-trial block of practice category training followed by the AX discrimination post-test.
Image Acquisition and Analysis
A 3T Siemens Trio scanner was used to perform anatomical and functional scans. High-resolution 3D anatomical images were acquired with a T1-weighted magnetization prepared rapid acquisition gradient echo (MPRAGE) sequence (TR = 1900ms, TE=4.15ms, TI = 1100ms, 1 mm3 isotropic voxel size, 256 × 256 matrix). Functional images were acquired in ascending, interleaved order with an echo-planar imaging (EPI) sequence (15 slices, 5mm thick, 3 mm2 axial in-plane resolution, 64 × 64 matrix, 192 mm3 FOV, flip angle = 90°). The functional slab was aligned parallel to the sylvian fissure, with coverage from the middle temporal gyrus caudally to the middle frontal gyrus dorsally, including coverage of the STG, MTG, SMG, AG, IFG, and portions of the MFG adjacent to the IFG. Each functional volume was acquired with a 1s acquisition time, followed by 1.830s of silence during which auditory stimuli were presented, for a total repetition time of 2.830s. Trials occurred every fifth volume acquisition (see Figure 2), and two additional volumes were added to the beginning of each run to accommodate saturation effects, resulting in a total of 277 volumes collected per run.
Images were analyzed using AFNI (Cox, 1996). Pre-processing of images included motion correction using a six-parameter rigid body transform co-registration with the anatomical dataset (Cox & Jesmanowicz, 1999), normalization to Talairach space (Talairach & Tournoux, 1988), and spatial smoothing with a 6mm Gaussian kernel. Masks were created using each participant’s anatomical data to eliminate activated voxels located outside of the brain. Individual masks were used to generate a group mask, which included only those voxels imaged in at least 25 out of 28 participants’ functional datasets.
Time series vectors were generated for each stimulus type: DIFFERENT trials (Dental vs. Retroflex, Retroflex vs. Velar and Dental vs. Velar), SAME trials (Dental vs. Dental, Retroflex vs. Retroflex, and Velar vs. Velar) and TARGET trials. These vectors contained the start time of each stimulus train and were convolved with a gamma-variate function. Each participant’s preprocessed functional data was submitted to a regression analysis with thirteen regressors, namely the seven condition vectors and the six parameters of the motion correction process, which were included as nuisance variables. This 3dDeconvolve analysis returned by-voxel fit coefficients for each condition for each participant.
Fit coefficients for each condition were entered into a mixed-factor ANOVA with condition as a fixed factor and subject as a random factor. The contrasts of interest (Dental vs. Retroflex and Retroflex vs. Velar) were compared both between and within groups, and main effects and interactions were also investigated. In addition, the effect of adaptation (DIFFERENT – SAME) was considered overall and for each contrast type. The ANOVA results were masked with a small volume corrected group mask, which included anatomically defined regions typically implicated in language processing: bilateral IFG, MFG, STG, and MTG, left SMG, TTG and AG. To correct for multiple comparisons, Monte Carlo simulations were run on the small-volume corrected group mask, producing a corrected threshold of p<0.05, and statistical maps were thresholded to include only clusters of 63 contiguous voxels at a voxel-level threshold of p<0.05. Mean fit coefficients were then extracted from each participant’s data and submitted to paired and two-sample t-tests.
RESULTS
Behavioral Results
Categorization training
Since categorization training became more difficult over the six blocks as the phonetic contrasts narrowed, all participants were required to reach a criterion for inclusion of at least four blocks completed before falling below 60% accuracy. The mean number of blocks completed above criterion was 5.29 (SE =0.12) with no significant difference between groups (t(26) = 0.57, p<0.576).
Post-training identification task responses to each of the nine stimulus tokens are plotted alongside a baseline function provided by a group of 14 untrained native English speakers, shown in Figure 3. Responses were then converted to z-scores, and the point at which each response function crossed the 0.50 mark was calculated as the participant’s category boundary. The two groups demonstrated very different functions following training, particularly on the middle tokens 4–6, with members of the Retroflex/Velar training group exhibiting a mean category boundary of 5.396 (SE = 0.119) and members of the Dental Retroflex training group exhibiting a mean category boundary of 4.027 (SE = 0.132). A univariate ANOVA comparing the baseline and training groups’ boundaries showed a significant main effect of group, F(2,41)=20.11, p<0.001. Post-hoc tests (Bonferroni-corrected at p<0.05, yielding a functional threshold of p<0.017) revealed that while the Dental/Retroflex trainees’ category boundary was significantly different from both the baseline (t(26)=3.97, p<0.001) and the Retroflex/Velar trainees’ boundaries, the Retroflex/Velar trainees’ boundary did not differ from baseline (t(26)=1.67, p=0.110). This result suggests that categorization training had a strong impact on how participants identified tokens from the continuum.
Discrimination
D’ scores were computed for all three contrasts in the AX discrimination task (displayed in Table 1) for Pre-test 1 and Post-test 1, completed in the first behavioral testing session, and for Pre-test 2 and Post-test 2, completed before and after the one-hour scanning session. Of particular interest is the mean difference in d’ sensitivity from pre-test to post-test, shown in Figures 4 and 5, as these reflect the effects of categorization training between the two categorization training groups.
Table 1.
Prior Behavioral Testing Session | Change: Post 1 – Pre 1 | fMRI Scanning Session | Overall Change: Post 2 – Pre 1 | ||||
---|---|---|---|---|---|---|---|
Pre-Test 1 | Post-Test 1 | Pre-Test 2 | Post-Test 2 | ||||
Dental/Retroflex Trainers | Dental vs. Retroflex | 1.471 | 2.930 | 1.459 | 2.781 | 3.202 | 1.731 |
SE | 0.292 | 0.298 | 0.331 | 0.374 | 0.372 | 0.480 | |
Retroflex vs. Velar | 1.001 | 0.160 | −0.841 | 0.756 | 1.035 | 0.034 | |
SE | 0.275 | 0.267 | 0.325 | 0.289 | 0.190 | 0.375 | |
Dental vs. Velar | 3.281 | 4.004 | 0.723 | 4.499 | 4.196 | 0.916 | |
SE | 0.362 | 0.187 | 0.259 | 0.104 | 0.172 | 0.441 | |
Retroflex/Velar Trainers | Dental vs. Retroflex | 1.978 | 1.918 | −0.060 | 2.005 | 2.533 | 0.554 |
SE | 0.437 | 0.433 | 0.267 | 0.372 | 0.302 | 0.420 | |
Retroflex vs. Velar | 1.408 | 1.074 | −0.334 | 1.471 | 1.667 | 0.260 | |
SE | 0.446 | 0.452 | 0.477 | 0.428 | 0.338 | 0.306 | |
Dental vs. Velar | 3.750 | 4.419 | 0.669 | 4.497 | 4.574 | 0.824 | |
SE | 0.314 | 0.106 | 0.358 | 0.082 | 0.079 | 0.341 |
These d’ scores were submitted to two separate three-way repeated-measures analyses of variance (ANOVA). The first ANOVA was conducted using d’ scores from the first behavioral testing session to examine the effects of the original categorization training manipulation, and it included factors of Group (Dental/Retroflex and Retroflex/Velar), Contrast (Dental vs. Retroflex and Retroflex vs. Velar only), and Effect of Training (Pre-test 1 and Post-test 1). This ANOVA revealed a main effect of Contrast, F(1,26) = 8.26, p<0.008, and no other significant main effects. The ANOVA also showed a significant interaction of Contrast x Training, F(1,26) = 10.77, p<0.003, an interaction of Group x Training approaching significance, F(1,26) = 2.47, p<0.128, and a significant three-way interaction of Group x Contrast x Training, F(1,26) = 6.68, p<0.016.
The results of the first ANOVA were examined for simple effects with a two-way repeated measures ANOVA for the Dental vs. Retroflex contrast, which revealed a significant main effect of Training, F(1,26) = 10.81, p<0.003, as well as an interaction of Group x Training, F(1,26) = 12.56, p<0.001. The two-way repeated measures ANOVA for the Retroflex vs. Velar contrast showed a significant main effect of Training, F(1,26)=4.14, p<0.052, but no interaction with Group. These simple effects indicated that while the Dental/Retroflex training group improved considerably in discriminating the Dental vs. Retroflex contrast, the Retroflex/Velar group did not increase or decrease their performance on either contrast.
The second three-way repeated measures ANOVA was conducted using d’ scores from the first pre-test and the final post-fMRI post-test to consider the overall effects of training across the two sessions. It included factors of Group (Dental/Retroflex and Retroflex/Velar), Contrast (Dental vs. Retroflex and Retroflex vs. Velar only), and Effect of Training (Pre-test 1 and Post-test 2). This ANOVA revealed a significant main effect of Training, F(1,26) = 12.18, p<0.002,a significant main effect of Contrast, F(1,26) = 8.37, p<0.008, and an interaction of Contrast x Training, F(1,26)=5.39, p<0.028. A three-way interaction of Group x Contrast x Training approached significance, F(1,26) = 2.67, p<0.114.
The second ANOVA was also examined for simple effects with a two-way repeated measures ANOVA for the Dental vs. Retroflex contrast, which revealed a significant main effect of Training, F(1,26) = 12.85, p<0.001, and a near-significant interaction of Group x Training, F(1,26)=3.407, p<0.076. A two-way repeated-measures ANOVA for the Retroflex vs. Velar contrast did not show any main effects or interactions. These simple effects again indicated that the Dental/Retroflex training group’s substantial improvement in discriminating the Dental vs. Retroflex contrast after training was driving the three-way interaction.
Imaging Results
Comparison of DIFFERENT trials to SAME trials across groups showed clusters in left and right IFG, pictured in Figure 6, that responded differentially to contrasting syllables. This effect was significantly stronger among members of the Dental/Retroflex training group (in the left IFG, t(1,79) = 3.009, p< 0.003 and in the right IFG, t(1,82) = 2.051, p< 0.043) than for the Retroflex/Velar training group (p>0.342 for both regions).
The mixed-factor ANOVA was conducted to look for main effects, interactions, and between group and within group comparisons, all summarized in Table 2. The Group x Contrast type interaction revealed significant clusters of activation in primarily in the left and right MFG, extending into the pars triangularis of the IFG on each side: on the left, F(1,26) = 11.16, p<0.003, and on the right, F(1,26) = 9.20, p<0.005. The results suggest that these regions respond differently depending on the nature of the categorization training experienced (see Figure 7). The clusters from the Group x Contrast interaction are paralleled by significant clusters in the same regions for the Within Group comparisons, indicating that the middle frontal gyri respond less actively for the trained contrast than for the untrained contrast. Paired t-tests showed a significant difference between contrasts in the right MFG for members of the Retroflex/Velar training group, t(13) = 3.086, p<0.009, and in the left MFG for members of the Dental/Retroflex training group, t(13) = −3.167, p<0.007.
Table 2.
Maximum Intensity Coordinates (T-T) | |||||
---|---|---|---|---|---|
Area | Number of Activated Voxels | x | y | z | Maximum t value |
Main Effect of Adaptation [All DIFFERENT trials > All SAME trials] | |||||
Left inferior frontal gyrus and left Brodmann area 9 | 213 | −46 | 14 | 26 | 3.581 |
Right middle frontal gyrus | 63 | 47 | 17 | 29 | 2.921 |
Left middle temporal gyrus and left Brodmann area 21* | 51 | −61 | −37 | 2 | 2.153 |
Left middle temporal gyrus and left Brodmann area 19* | 46 | −37 | −79 | 23 | −3.255 |
Main Effect of Adaptation: [Dent vs. Vel DIFFERENT > Dent SAME + Vel SAME] | |||||
Left inferior frontal gyrus and left Brodmann area 47 | 94 | −46 | 17 | 2 | 2.442 |
Right middle temporal gyrus | 81 | 41 | −61 | 23 | 2.450 |
Left inferior frontal gyrus and left Brodmann area 9* | 59 | −43 | 11 | 26 | 3.037 |
Right middle frontal gyrus* | 57 | 38 | 8 | 32 | 3.331 |
Left middle temporal gyrus and left Brodmann area 19* | 43 | −37 | −79 | 23 | −2.829 |
Main Effect of Group [Ret/Vel Trainees > Dent/Ret Trainees] | |||||
Right inferior frontal gyrus | 120 | 38 | 20 | −7 | 2.914 |
Interaction Group x Contrast type | |||||
Left middle frontal gyrus | 116 | −40 | 35 | 26 | 8.500 |
Right middle frontal gyrus and right Brodmann area 10 | 78 | 38 | 41 | 17 | 9.060 |
Between Group Effects: Dent vs. Ret [Ret/Vel Trainees > Dent/Ret Trainees] | |||||
Right inferior frontal gyrus | 146 | 44 | 41 | 2 | 3.810 |
Left middle frontal gyrus | 80 | −37 | 38 | 26 | 2.964 |
Right inferior frontal gyrus* | 52 | 50 | 5 | 20 | 3.908 |
Between Group Effects: Ret vs. Vel [Dent/Ret Trainees > Ret/Vel Trainees] | |||||
- none - | |||||
Within Group Effects: Ret/Vel Trainers [Dent vs. Ret > Ret vs. Vel] | |||||
Right middle frontal gyrus | 67 | 38 | 38 | 17 | 2.980 |
Within Group Effects: Dent/Ret Trainers [Dent vs. Ret < Ret vs. Vel] | |||||
Left middle frontal gyrus and left Brodmann area 10* | 53 | −40 | 41 | 20 | −2.715 |
Right middle temporal gyrus* | 40 | 53 | −40 | −1 | −3.833 |
= cluster is sub-threshold
Between-groups comparisons revealed two clusters in frontal regions for which the Retroflex/Velar training group showed greater signal change in response to the Dental vs. Retroflex contrast than the Dental/Retroflex training group: in the right IFG, t(26) = 4.174, p<0.001, and in the left MFG, t(26) = 3.954, p<0.001. However, no significant differences between groups were apparent in response to the Retroflex vs. Velar contrast.
Results of the regression analysis, reported in Table 3 and pictured in Figure 8, indicated that a change in discrimination sensitivity (Post-test 2 – Pre-test 1 d’ score) was negatively correlated with signal change in two regions, right superior/middle temporal gyrus, r(84) = −0.309, and left IFG, r(84) = −0.334. This latter cluster was immediately adjacent to the MFG cluster revealed in the interaction analysis, and overlapped this cluster by five voxels. The negative correlation suggests that participants who improved at discriminating the contrasts showed less activation than those who did not improve.
Table 3.
Maximum Intensity Coordinates (T-T) | |||||
---|---|---|---|---|---|
Area | Number of Activated Voxels | x | y | z | Maximum t value |
Regression: Change in d’ perceptual sensitivity with % signal change | |||||
Right superior/middle temporal gyrus | 237 | 56 | −40 | 11 | −3.194 |
Left inferior frontal gyrus | 77 | −37 | 29 | 5 | −3.586 |
= cluster is sub-threshold
DISCUSSION
A central question in the study of speech perception is how categorical perception of speech emerges. The results of the present study suggest that explicit categorization training can produce increased discrimination sensitivity to between-category contrasts, as well as differential neural sensitivity to the learned category scheme. Specifically, one group (Dental/Retroflex training group) showed improvements in behavioral discrimination performance for tokens that crossed the learned category boundary (Dental vs. Retroflex) but no improvement for those within a single learned category (Retroflex vs. Velar). This same group also showed a significant shift in categorization function compared to individuals who had not undergone training, whereas the Retroflex/Velar group showed no such shift. Of interest, previous work from our lab using the same stimuli and presentation order (Swan & Myers, under review) showed that mere exposure to the tokens was not sufficient to produce category-specific changes in discrimination. In the present study, discrimination asymmetries that resulted from categorization training were accompanied by differences in activation almost exclusively in left and right frontal areas, specifically in the left IFG and left and right MFG. Surprisingly, no interaction between Group and Contrast was found in temporal areas, which have often been associated with the perception of fine-grained acoustic details of speech. The implications of these results for our understanding of the neural systems underlying speech perception are discussed below.
The effect of categorization training on behavioral discrimination differed significantly between groups. Behavioral data showed that, despite similar success during categorization training, only the Dental/Retroflex training group demonstrated a persisting shift in the phonetic category boundary and subsequently transferred their category learning to a discrimination paradigm (c.f. also Swan & Myers, under review)1. These behavioral results lead to the suggestion that only the Dental/Retroflex group showed a reorganization of perceptual space around the learned categories. Nonetheless, fMRI data revealed an interaction between group and contrast in left and right MFG, with significantly greater activation for the learned within-category contrast compared to the learned between-category contrast in both groups. This finding is consistent with the view that sensitivity to category-level information may precede the reshaping of sensitivities along the acoustic-phonetic continuum, as measured by discrimination (Liebenthal et al., 2010). Evidence from ERP paradigms also suggests that neural correlates of phonetic category learning may be found prior to behavioral sensitivity to the same contrast (Tremblay, Kraus, & McGee, 1998).
Both groups showed greater activation for ‘different’ trials than ‘same’ trials in the left STG as well as left IFG, replicating previous studies using this paradigm (Myers, et al, 2009). However, the relationship between activation patterns on between- and within-category trials was unexpected. Previous research from our group (Myers, et al 2009) and others (Joanisse et al., 2007; Zevin & McCandliss, 2005; Zevin et al., 2010) has shown greater activation in adaptation/habituation designs for more perceptible stimulus contrasts. In the current study, participants showed the expected behavioral pattern: between-category contrasts were easier to perceive than within-category contrasts after category training, yet greater activation was shown for the difficult within-category contrast in both groups. A possible explanation accounting for this difference is that the results reflect a cognitive set that participants enter when they are tasked with performing an active categorization task immediately before scanning. That is, while participants were not required to attend to category-level information during scanning, they may nonetheless be implicitly categorizing stimuli, consistent with the demands of the behavioral task performed outside the scanner. If participants are implicitly assigning category labels to stimuli, the results may be expected to reflect active detection of category distinctions akin to a discrimination task (Desai, Liebenthal, Waldron, & Binder, 2008; Hutchison, Blumstein, & Myers, 2008; Liebenthal et al., 2005). Of interest, Desai and colleagues (2008) report activation of the left and right MFG in a study which used an ABX discrimination task to measure participants’ sensitivity to acoustic distinctions in non-speech sine-wave sounds that can be perceived either as non-speech sounds or with phonetic content. Changes in activation in both the left and right MFG were larger for a continuum that participants ultimately perceived as speech than for a set of sounds that could not be mapped to speech categories, suggesting that the MFG bilaterally are sensitive to category-level information about speech tokens. Of interest, a separate region in the left STG was correlated with increases in the sharpness of participants’ categorization function. Given that these authors did not investigate correlations between brain and behavior in the MFG, it is unknown whether frontal activation shows a similar relationship to changes in behavior.
The left MFG has also been shown to discriminate between naturally produced and synthetic speech (Benson et al., 2001). Similarly, in a meta-analysis of 23 studies of speech perception, Turkeltaub and Coslett (2010) report a small focus in the posterior left MFG which was related to activation for speech stimuli compared to low-level control stimuli. While greater activation in the MFG has been observed while attending rather than ignoring auditory, and particularly speech, stimuli (Sabri et al., 2008), the MFG are also modulated by nonlinguistic input. The bilateral MFG have been implicated in access to category-level information in abstract visual patterns (Vogels, Sary, Dupont, & Orban, 2002), and sensitivity to learned visual object categories has been shown in a homologous region of lateral prefrontal cortex in non-human primates (Freedman, Riesenhuber, Poggio, & Miller, 2001). This evidence suggests that the recruitment of the bilateral MFG may reflect a domain-general resource for categorization processes.
Contrary to our hypotheses, while the IFG responded more to ‘Different’ than ‘Same’ trials, there was no interaction centered in the IFG between contrast and group, suggesting that this region responded similarly regardless of how participants had been trained. Most previous studies showing IFG sensitivity to phonetic category-level information have used active categorization tasks (Golestani, et al, 2004, Callan et al, 2004), and activation in the IFG for categorization tasks has been linked to difficulty in phonetic tasks (Binder et al, 2004, Blumstein et al, 2005). Moreover, some have suggested that the contribution of Broca’s area to speech perception is limited to situations in which participants are required to actively inspect the acoustic signal (Hickok & Poeppel, 2004). It is possible that the implicit nature of the current design did not require participants to inspect the signal, and thus did not differently engage the IFG as a function of the trained category membership
Most remarkably, no evidence of reorganization of cortical sensitivity as a function of training was seen in temporal areas. While the left STG did show sensitivity to phonetic differences, with a sub-threshold (50-voxel) cluster exhibiting greater activation for phonetic change trials than repeated trials, the two different training groups showed no differences in sensitivity in temporal areas to the learned within- and between-category contrasts. This may seem surprising given that some previous studies (Guenther, et al, 2004, Liebenthal et al, 2010, Desai, et al, 2008) have shown engagement of temporal areas related to categorization training. The relatively rapid learning that our participants displayed may result from an ability to deploy attention appropriately to the distinguishing aspects of the acoustic waveform rather than a fundamental re-organization of perceptual space (c.f. also Alain, Campeanu, & Tremblay, 2010). Accordingly, the presence of category labels may modulate attention and momentarily highlight stimulus features containing meaningful distinctions (cf. Landau, Dessalegn & Goldberg, 2009), or may reinforce a progressively more consistent decision threshold, but does not alter the underlying perceptual space or its representation in the temporal lobes in the short term. The suggested role of attention in learning to perceive new speech contrasts is not unique to this study. ‘Attention to dimension’ models of speech perception (Francis & Nusbaum, 2000, Francis et al, 2002) propose that shifts in attention to relevant parts of acoustic phonetic space may be crucial not only for learning new speech sounds, but for attending to distinctions between sounds in one’s native language repertoire.
It is possible that the activation observed in the middle frontal gyri does not relate to learned sensitivities to category information, but rather results from attentional biases that arise during training. For instance, it may be the case that as participants learn to assign different category labels to tokens along the acoustic phonetic continuum, attention is directed to between-category contrasts in a way that is unrelated to or independent of the demonstrated changes in behavioral sensitivity to the same contrasts. This account might explain the finding that while only one group showed significant changes in their behavioral performance, there was a full cross-over interaction in the bilateral MFG, with significant differences in activation for both groups.
We argue that the engagement of regions involved in attention and categorization need not be seen as epiphenomenal. Categorization training in this study resulted in shifts in discrimination accuracy that resemble those seen for native-language phonetic category structure. The engagement of frontal regions associated with a learned phonetic category structure may be the first step in encoding more long-term sensitivities to speech sounds. The fact that sensitivity to category membership was observed in the MFG rather than the temporal lobes suggests that this change in apparent perceptual sensitivity results from the influence of an enhanced distinctiveness from mapping two tokens to two distinct abstract categories, rather than due to the changes in the perceptual encoding of these stimuli per se. One hypothesis is that over the course of learning, top-down feedback signals mediate the tuning of acoustic-phonetic processers in the temporal lobes, which over time may result in long-term changes in tuning properties of temporal areas. Results of the regression analysis support this view: less activation in inferior frontal regions was seen for individuals who showed the greatest improvements from pretest to posttest. Such a model would account for data showing that incidental auditory category learning over many sessions, or even over a lifetime of second-language learning (Leech et al., 2009; Liebenthal, et al, 2010, Raizada, Tsao, Liu, & Kuhl, 2010) produce changes in encoding of between-category contrasts in temporal areas.
Results of the present study highlight the similarities between category learning for speech and category learning in other modalities. Across a variety of domains, the use of category labels seems to influence how items from physical continua are perceived. Increased discrimination of between-category distinctions for which there are established labels has been shown in perception of color, size and brightness, facial identity and emotional valence (Kay & Kempton, 1984; Kikutani, Roberson, & Hanley, 2008; Özgen & Davies, 2002). In the current study, exposure to category-level information produced increases in behavioral sensitivity to between-category contrasts. Parallels also exist at the level of neural function: primate studies of learned animal categories (Freedman et al., 2001; Seger & Miller, 2010) and face identity categories (Rotshtein, Henson, Treves, Driver, & Dolan, 2005), implicate inferior frontal regions in the processing of category membership. As such, it seems plausible that changes in behavioral sensitivity arising from category learning in non-native speech acquisition may be driven by domain-general executive processes, operating in the frontal lobes.
Conclusion
Speech scientists are forced to reconcile evidence of both plasticity and durability in the speech perceptual system. This contrast is readily apparent in the discontinuous perception of speech categories, which although robust, nonetheless adapts flexibly to changes in the language environment. We suggest that such tensions are resolved in the neural system through parallel processes; in this case combining stable long-term sensitivities encoded in the temporal lobes with flexible, context-dependent decision processes in the frontal lobes.
Acknowledgments
This work was funded by R03 DC009495 to Emily Myers. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIDCD or the NIH. The authors thank Sheila Blumstein for providing stimuli and Laura Mesite for assistance with data analysis.
Footnotes
The failure of the Retroflex/Velar group to show persisting effects of categorization training on subsequent categorization and discrimination tasks remains somewhat mysterious. The most likely explanation is that the retroflex and velar tokens are closer to one another in acoustic space, which has been shown to affect the relative discriminability of non-native contrasts (e.g. Polka, 1991). This was reflected by dramatically lower d’ scores at pre-test for both groups on the Retroflex/Velar contrast compared to the Dental/Retroflex contrast (Retroflex/Velar contrast: mean d’ =0.9325, Dental/Retroflex contrast: mean d’ =3.225).
Contributor Information
Emily B. Myers, University of Connecticut, Department of Communication Sciences, Department of Psychology.
Kristen Swan, Centre for Brain & Cognitive Development, Birkbeck, University of London.
Works Cited
- Alain C, Campeanu S, Tremblay K. Changes in sensory evoked responses coincide with rapid improvement in speech identification performance. Journal of Cognitive Neuroscience. 2010;22(2):392–403. doi: 10.1162/jocn.2009.21279. [DOI] [PubMed] [Google Scholar]
- Benson RR, Whalen DH, Richardson M, Swainson B, Clark VP, Lai S, Liberman AM. Parametrically dissociating speech and nonspeech perception in the brain using fMRI. Brain Lang. 2001;78(3):364–96. doi: 10.1006/brln.2001.2484. [DOI] [PubMed] [Google Scholar]
- Best CC, McRoberts GW. Infant perception of non-native consonant contrasts that adults assimilate in different ways. Lang Speech. 2003;46(Pt 2–3):183–216. doi: 10.1177/00238309030460020701. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Best CT, McRoberts GW, Goodell E. Discrimination of non-native consonant contrasts varying in perceptual assimilation to the listener’s native phonological system. J Acoust Soc Am. 2001;109(2):775–94. doi: 10.1121/1.1332378. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Binder JR, Liebenthal E, Possing ET, Medler DA, Ward BD. Neural correlates of sensory and decision processes in auditory object identification. Nature Neuroscience. 2004;7(3):295–301. doi: 10.1038/nn1198. [DOI] [PubMed] [Google Scholar]
- Blumstein SE, Myers EB, Rissman J. The perception of voice onset time: an fMRI investigation of phonetic category structure. Journal of Cognitive Neuroscience. 2005;17(9):1353–1366. doi: 10.1162/0898929054985473. [DOI] [PubMed] [Google Scholar]
- Boersma P. Praat, a system for doing phonetics by computer. Glot International. 2001;5(9/10):341–345. [Google Scholar]
- Bradlow AR, Pisoni DB, Akahane-Yamada R, Tohkura Y. Training Japanese listeners to identify English /r/ and /l/: IV. Some effects of perceptual learning on speech production. J Acoust Soc Am. 1997;101(4):2299–310. doi: 10.1121/1.418276. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Burton MW, Small SL, Blumstein SE. The role of segmentation in phonological processing: An fMRI investigation. J Cognitive Neurosci. 2000;12(4):679–690. doi: 10.1162/089892900562309. [DOI] [PubMed] [Google Scholar]
- Callan DE, Jones JA, Callan AM, Akahane-Yamada R. Phonetic perceptual identification by native- and second-language speakers differentially activates brain regions involved with acoustic phonetic processing and those involved with articulatory-auditory/orosensory internal models. Neuroimage. 2004;22(3):1182–94. doi: 10.1016/j.neuroimage.2004.03.006. [DOI] [PubMed] [Google Scholar]
- Callan DE, Tajima K, Callan AM, Kubo R, Masaki S, Akahane-Yamada R. Learning-induced neural plasticity associated with improved identification performance after training of a difficult second-language phonetic contrast. Neuroimage. 2003;19(1):113–24. doi: 10.1016/s1053-8119(03)00020-x. [DOI] [PubMed] [Google Scholar]
- Chang EF, Rieger JW, Johnson K, Berger MS, Barbaro NM, Knight RT. Categorical speech representation in human superior temporal gyrus. Nature Neuroscience. 2010;13(11):1428–1432. doi: 10.1038/nn.2641. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cox RW. AFNI: Software for analysis and visualization of functional magnetic resonance neuroimages. Computers and Biomedical Research. 1996;29:162–173. doi: 10.1006/cbmr.1996.0014. [DOI] [PubMed] [Google Scholar]
- Cox RW, Jesmanowicz A. Real-time 3D image registration for functional MRI. Magn Reson Med. 1999;42(6):1014–8. doi: 10.1002/(sici)1522-2594(199912)42:6<1014::aid-mrm4>3.0.co;2-f. [DOI] [PubMed] [Google Scholar]
- Desai R, Liebenthal E, Waldron E, Binder JR. Left posterior temporal regions are sensitive to auditory categorization. Journal of Cognitive Neuroscience. 2008;20(7):1174–1188. doi: 10.1162/jocn.2008.20081. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Flege JE, MacKay IR, Meador D. Native Italian speakers’ perception and production of English vowels. The Journal of the Acoustical Society of America. 1999;106:2973. doi: 10.1121/1.428116. [DOI] [PubMed] [Google Scholar]
- Formisano E, De Martino F, Bonte M, Goebel R. “Who” is saying “what”? Brain-based decoding of human voice and speech. Science (New York, NY) 2008;322(5903):970–973. doi: 10.1126/science.1164318. [DOI] [PubMed] [Google Scholar]
- Francis AL, Nusbaum HC. Selective attention and the acquisition of new phonetic categories. J Exp Psychol Hum Percept Perform. 2002;28(2):349–66. doi: 10.1037//0096-1523.28.2.349. [DOI] [PubMed] [Google Scholar]
- Francis AL, Baldwin K, Nusbaum HC. Effects of training on attention to acoustic cues. Percept Psychophys. 2000;62(8):1668–80. doi: 10.3758/bf03212164. [DOI] [PubMed] [Google Scholar]
- Freedman DJ, Riesenhuber M, Poggio T, Miller EK. Categorical representation of visual stimuli in the primate prefrontal cortex. Science. 2001;291(5502):312–6. doi: 10.1126/science.291.5502.312. [DOI] [PubMed] [Google Scholar]
- Ganong WF. Phonetic categorization in auditory word perception. Journal of Experimental Psychology: Human Perception and Performance. 1980;6(1):110–125. doi: 10.1037//0096-1523.6.1.110. [DOI] [PubMed] [Google Scholar]
- Golestani N, Zatorre RJ. Learning new sounds of speech: reallocation of neural substrates. Neuroimage. 2004;21(2):494–506. doi: 10.1016/j.neuroimage.2003.09.071. [DOI] [PubMed] [Google Scholar]
- Golestani N, Zatorre RJ. Individual differences in the acquisition of second language phonology. Brain Lang. 2009;109(2–3):55–67. doi: 10.1016/j.bandl.2008.01.005. [DOI] [PubMed] [Google Scholar]
- Golestani N, Molko N, Dehaene S, LeBihan D, Pallier C. Brain structure predicts the learning of foreign speech sounds. Cereb Cortex. 2007;17(3):575–82. doi: 10.1093/cercor/bhk001. [DOI] [PubMed] [Google Scholar]
- Grill-Spector K, Malach R. fMR-adaptation: a tool for studying the functional properties of human cortical neurons. Acta Psychologica (Amsterdam) 2001;107(1–3):293–321. doi: 10.1016/s0001-6918(01)00019-1. [DOI] [PubMed] [Google Scholar]
- Grill-Spector K, Henson R, Martin A. Repetition and the brain: neural models of stimulus-specific effects. Trends in Cognitive Sciences. 2006;10(1):14–23. doi: 10.1016/j.tics.2005.11.006. [DOI] [PubMed] [Google Scholar]
- Guenther FH, Gjaja MN. The perceptual magnet effect as an emergent property of neural map formation. Journal of the Acoustical Society of America. 1996;100(2 Pt 1):1111–21. doi: 10.1121/1.416296. [DOI] [PubMed] [Google Scholar]
- Guenther FH, Husain FT, Cohen MA, Shinn-Cunningham BG. Effects of categorization and discrimination training on auditory perceptual space. Journal of the Acoustical Society of America. 1999;106(5):2900–12. doi: 10.1121/1.428112. [DOI] [PubMed] [Google Scholar]
- Guenther FH, Nieto-Castanon A, Ghosh SS, Tourville JA. Representation of sound categories in auditory cortical maps. Journal of Speech Language and Hearing Research. 2004;47(1):46–57. doi: 10.1044/1092-4388(2004/005). [DOI] [PubMed] [Google Scholar]
- Hickok G, Poeppel D. Dorsal and ventral streams: a framework for understanding aspects of the functional anatomy of language. Cognition. 2004;92(1–2):67–99. doi: 10.1016/j.cognition.2003.10.011. [DOI] [PubMed] [Google Scholar]
- Hutchison ER, Blumstein SE, Myers EB. An event-related fMRI investigation of voice-onset time discrimination. NeuroImage. 2008;40(1):342–352. doi: 10.1016/j.neuroimage.2007.10.064. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Joanisse MF, Zevin JD, McCandliss BD. Brain mechanisms implicated in the preattentive categorization of speech sounds revealed using FMRI and a short-interval habituation trial paradigm. Cerebral Cortex (New York, NY: 1991) 2007;17(9):2084–2093. doi: 10.1093/cercor/bhl124. [DOI] [PubMed] [Google Scholar]
- Kay P, Kempton W. What Is the Sapir-Whorf Hypothesis? American Anthropologist. 1984;86(1):65–79. [Google Scholar]
- Kikutani M, Roberson D, Hanley J. What’s in the name? Categorical perception for unfamiliar faces can occur through labeling. Psychonomic bulletin & review. 2008;15(4):787–794. doi: 10.3758/pbr.15.4.787. [DOI] [PubMed] [Google Scholar]
- Kuhl PK. Human adults and human infants show a “perceptual magnet effect” for the prototypes of speech categories, monkeys do not. Perception and Psychophysics. 1991;50(2):93–107. doi: 10.3758/bf03212211. [DOI] [PubMed] [Google Scholar]
- Landau B, Dessalegn B, Goldberg A. Language and Space: Momentary Interactions. In: Chilton P, Evans V, Evans V, Bergen B, Zinken J, editors. Language, cognition and space: The state of the art and new directions Advances in Cognitive Linguistics Series. London: Equinox Publishing; 2010. [Google Scholar]
- Leech R, Holt LL, Devlin JT, Dick F. Expertise with artificial nonspeech sounds recruits speech-sensitive cortical regions. J Neurosci. 2009;29(16):5234–9. doi: 10.1523/JNEUROSCI.5758-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Liberman AM, Delattre PC, Cooper FS. Some cues for the distinction between voiceless and voiced stops in initial position. Lang Speech. 1958;1:153–157. [Google Scholar]
- Liberman AM, Harris KS, Hoffman HS, Griffith BC. The discrimination of speech sounds within and across phoneme boundaries. Journal of Experimental Psychology. 1957;54(5):358–68. doi: 10.1037/h0044417. [DOI] [PubMed] [Google Scholar]
- Liebenthal E, Binder JR, Spitzer SM, Possing ET, Medler DA. Neural substrates of phonemic perception. Cerebral Cortex. 2005;15(10):1621–31. doi: 10.1093/cercor/bhi040. [DOI] [PubMed] [Google Scholar]
- Liebenthal E, Desai R, Ellingson MM, Ramachandran B, Desai A, Binder JR. Specialization along the left superior temporal sulcus for auditory categorization. Cerebral Cortex (New York, NY: 1991) 2010;20(12):2958–2970. doi: 10.1093/cercor/bhq045. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McCandliss BD, Fiez JA, Protopapas A, Conway M, McClelland JL. Success and failure in teaching the [r]-[l] contrast to Japanese adults: Tests of a Hebbian model of plasticity and stabilization in spoken language perception. Cognitive, Affective, & Behavioral Neuroscience. 2002;2(2):89. doi: 10.3758/cabn.2.2.89. [DOI] [PubMed] [Google Scholar]
- Miller JL. Some effects of speaking rate on phonetic perception. Phonetica. 1981;38(1–3):159–80. doi: 10.1159/000260021. [DOI] [PubMed] [Google Scholar]
- Myers EB. Dissociable effects of phonetic competition and category typicality in a phonetic categorization task: An fMRI investigation. Neuropsychologia. 2007;45(7):1463–73. doi: 10.1016/j.neuropsychologia.2006.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Myers EB, Blumstein SE, Walsh E, Eliassen J. Inferior frontal regions underlie the perception of phonetic category invariance. Psychol Sci. 2009;20(7):895–903. doi: 10.1111/j.1467-9280.2009.02380.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nygaard LC, Pisoni DB. Talker-specific learning in speech perception. Perception and Psychophysics. 1998;60(3):355–376. doi: 10.3758/bf03206860. [DOI] [PubMed] [Google Scholar]
- Oldfield RC. The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia. 1971;9(1):97–113. doi: 10.1016/0028-3932(71)90067-4. [DOI] [PubMed] [Google Scholar]
- Özgen E, Davies IRL. Acquisition of categorical color perception: A perceptual learning approach to the linguistic relativity hypothesis. Journal of Experimental Psychology: General. 2002;131:477–493. [PubMed] [Google Scholar]
- Pisoni DB, Tash J. Reaction times to comparisons within and across phonetic categories. Perception and Psychophysics. 1974;15:289–290. doi: 10.3758/bf03213946. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Polka L. Cross-language speech perception in adults: phonemic, phonetic, and acoustic contributions. J Acoust Soc Am. 1991;89(6):2961–77. doi: 10.1121/1.400734. [DOI] [PubMed] [Google Scholar]
- Raizada RDS, Tsao F-M, Liu H-M, Kuhl PK. Quantifying the adequacy of neural representations for a cross-language phonetic discrimination task: prediction of individual differences. Cerebral Cortex (New York, NY: 1991) 2010;20(1):1–12. doi: 10.1093/cercor/bhp076. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rotshtein P, Henson RNA, Treves A, Driver J, Dolan RJ. Morphing Marilyn into Maggie dissociates physical and identity face representations in the brain. Nat Neurosci. 2005;8(1):107–113. doi: 10.1038/nn1370. [DOI] [PubMed] [Google Scholar]
- Sabri M, Binder JR, Desai R, Medler DA, Leitl MD, Liebenthal E. Attentional and linguistic interactions in speech perception. Neuroimage. 2008;39(3):1444–56. doi: 10.1016/j.neuroimage.2007.09.052. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Seger CA, Miller EK. Category Learning in the Brain. Annual Review of Neuroscience. 2010;33:203–219. doi: 10.1146/annurev.neuro.051508.135546. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stevens K, Blumstein S. Quantal aspects of consonant production and perception: A study of retroflex stop consonants. Journal of Phonetics. 1975;3(4):215–233. [Google Scholar]
- Swan K, Myers E. Category labels induce boundary-dependent perceptual warping in learned speech categories. (under review) [DOI] [PMC free article] [PubMed] [Google Scholar]
- Talairach J, Tournoux P. A co-planar stereotaxic atlas of a human brain. Stuttgart: Thieme; 1988. [Google Scholar]
- Tremblay K, Kraus N, McGee T. The time course of auditory perceptual learning: neurophysiological changes during speech-sound training. Neuroreport. 1998;9(16):3557–60. doi: 10.1097/00001756-199811160-00003. [DOI] [PubMed] [Google Scholar]
- Turkeltaub PE, Branch Coslett H. Localization of sublexical speech perception components. Brain and Language. 2010;114(1):1–15. doi: 10.1016/j.bandl.2010.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Vogels R, Sary G, Dupont P, Orban GA. Human Brain Regions Involved in Visual Categorization. NeuroImage. 2002;16(2):401–414. doi: 10.1006/nimg.2002.1109. [DOI] [PubMed] [Google Scholar]
- Wang Y, Sereno JA, Jongman A, Hirsch J. fMRI evidence for cortical modification during learning of Mandarin lexical tone. J Cogn Neurosci. 2003;15(7):1019–27. doi: 10.1162/089892903770007407. [DOI] [PubMed] [Google Scholar]
- Werker JF, Tees RC. Phonemic and phonetic factors in adult cross-language speech perception. J Acoust Soc Am. 1984;75(6):1866–78. doi: 10.1121/1.390988. [DOI] [PubMed] [Google Scholar]
- Yeung HH, Werker JF. Learning Words’ Sounds before Learning How Words Sound: 9-Month-Olds Use Distinct Objects as Cues to Categorize Speech Information. Cognition. 2009;113(2):234–243. doi: 10.1016/j.cognition.2009.08.010. [DOI] [PubMed] [Google Scholar]
- Zevin JD, McCandliss BD. Dishabituation of the BOLD response to speech sounds. Behav Brain Funct. 2005;1(1):4. doi: 10.1186/1744-9081-1-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zevin JD, Yang J, Skipper JI, McCandliss BD. Domain General Change Detection Accounts for“ Dishabituation” Effects in Temporal-Parietal Regions in Functional Magnetic Resonance Imaging Studies of Speech Perception. Journal of Neuroscience. 2010;30(3):1110. doi: 10.1523/JNEUROSCI.4599-09.2010. [DOI] [PMC free article] [PubMed] [Google Scholar]