Skip to main content
The Journal of Neuroscience logoLink to The Journal of Neuroscience
. 2016 Oct 19;36(42):10782–10790. doi: 10.1523/JNEUROSCI.1595-16.2016

Involvement of the Serotonin Transporter Gene in Accurate Subcortical Speech Encoding

Lenka Selinger 1,2, Katarzyna Zarnowiec 1,2, Marc Via 1,2,3, Immaculada C Clemente 1,2,3, Carles Escera 1,2,3,
PMCID: PMC6601888  PMID: 27798133

Abstract

A flourishing line of evidence has highlighted the encoding of speech sounds in the subcortical auditory system as being shaped by acoustic, linguistic, and musical experience and training. And while the heritability of auditory speech as well as nonspeech processing has been suggested, the genetic determinants of subcortical speech processing have not yet been uncovered. Here, we postulated that the serotonin transporter-linked polymorphic region (5-HTTLPR), a common functional polymorphism located in the promoter region of the serotonin transporter gene (SLC6A4), is implicated in speech encoding in the human subcortical auditory pathway. Serotonin has been shown as essential for modulating the brain response to sound both cortically and subcortically, yet the genetic factors regulating this modulation regarding speech sounds have not been disclosed. We recorded the frequency following response, a biomarker of the neural tracking of speech sounds in the subcortical auditory pathway, and cortical evoked potentials in 58 participants elicited to the syllable /ba/, which was presented >2000 times. Participants with low serotonin transporter expression had higher signal-to-noise ratios as well as a higher pitch strength representation of the periodic part of the syllable than participants with medium to high expression, possibly by tuning synaptic activity to the stimulus features and hence a more efficient suppression of noise. These results imply the 5-HTTLPR in subcortical auditory speech encoding and add an important, genetically determined layer to the factors shaping the human subcortical response to speech sounds.

SIGNIFICANCE STATEMENT The accurate encoding of speech sounds in the subcortical auditory nervous system is of paramount relevance for human communication, and it has been shown to be altered in different disorders of speech and auditory processing. Importantly, this encoding is plastic and can therefore be enhanced by language and music experience. Whether genetic factors play a role in speech encoding at the subcortical level remains unresolved. Here we show that a common polymorphism in the serotonin transporter gene relates to an accurate and robust neural tracking of speech stimuli in the subcortical auditory pathway. This indicates that serotonin transporter expression, eventually in combination with other polymorphisms, delimits the extent to which lifetime experience shapes the subcortical encoding of speech.

Keywords: 5-HTTLPR, frequency-following response, genetics, individual differences, SLC6A4, speech perception

Introduction

Why do some people have a better ear for music and some learn to pronounce a new language so much better than others? Clearly, individuals differ in their musical ability, and in their capacity to learn a new language and to correctly pronounce it (Díaz et al., 2008). More generally, it could be said that people differ in their capacity for sound and, importantly, speech perception (Pallier et al., 1997). An accurate and fine-grained neural representation of the sound signal is the basis for speech perception.

The subcortical auditory pathway has been shown to encode acoustic features of speech and nonspeech stimuli with great accuracy (Skoe and Kraus, 2010a). Even in the typically developing population, however, the strength with which features important for language perception are represented varies systematically with language and reading competence (Banai et al., 2009; Chandrasekaran et al., 2009; Hornickel et al., 2009, 2011). Moreover, the auditory brainstem response, comprising both the transient and the periodic frequency following response (FFR) to sound, has been shown to be altered in children with specific language impairment (Basu et al., 2010), attention deficit disorder with hyperactivity (Jafari et al., 2015), and auditory processing disorder (Rocha-Muniz et al., 2012). In preschooler children, it has been proposed as a biomarker to predict literacy (White-Schwoch et al., 2015).

The FFR is a neural response to periodic sounds that can be recorded from the human scalp (Chandrasekaran and Kraus, 2010), and whose generators have been attributed to auditory brainstem nuclei, particularly the inferior colliculus (IC; Smith et al., 1975; Møller et al. 1988), yet a cortical contribution cannot be disregarded for certain stimulus characteristics (Coffey et al., 2016). It faithfully reflects the envelope and frequency contents of sounds. The FFR grants a noninvasive, objective means for examining the neural encoding of sound at subcortical stages of the auditory pathway. These nuclei play an important role as a computational hub, receiving bottom-up auditory input from brainstem structures, as well as direct cortigofugal feedback projections from the auditory cortex. Remarkably, the fidelity with which the ascending auditory pathway represents the physical characteristics of sounds in the FFR is experience dependent, as musical training (Parbery-Clark et al., 2013), bilingualism (Krizman et al., 2012), the socioeconomic status of the person (Skoe et al., 2013a), and the context in which a stimulus is presented (Skoe and Kraus, 2010b) shape the subcortical response to sound. It has been hypothesized that increased variability in the FFR leads to a compromised interaction between sensory and cognitive processing, resulting in disparities in cognitive abilities between individuals (Skoe et al., 2013a). Whether genetic factors contribute to individual differences regarding sound processing at a subcortical level, however, has not been established so far.

It has been shown that the IC receives major serotonergic innervation originating in the raphe nuclei (Klepper and Herbert, 1991; Hurley and Pollak, 1999). This suggests that serotonin [5-hydroxytryptamine (5-HT)] is crucial for the modulation of information processing in the ascending auditory pathway (Obara et al., 2014). In particular, animal studies have demonstrated that serotonin alters multiple aspects of the neural response to sound. Not only does it change the number of spikes evoked by auditory stimuli (Hurley and Pollak, 1999, 2001), but also the latency and precision of initial spikes and the timing of spike trains and therefore all response properties contributing to the encoding of sensory stimuli (Hurley and Pollak, 2005). As the serotonin transporter [5-HT transporter (5-HTT)] is considered to be crucial for the regulation of serotonin availability and a common functional polymorphism has been identified in the serotonin transporter gene (SLC6A4; Heils et al., 1996), we hypothesized that it might be implicated in subcortical auditory processing.

The aim of this study was to determine whether the 5-HTT linked polymorphic region (5-HTTLPR) is involved in speech encoding at subcortical stages. Two measures of encoding accuracy and robustness of neural phase locking to the stimulus periodicity, namely, the signal-to-noise ratio (SNR) and pitch strength, revealed a more finely tuned auditory processing at subcortical stages for individuals with low serotonin transporter expression compared with those with medium to high expression.

Materials and Methods

Participants

Seventy-nine young adults ranging in age from 18 to 31 years [mean (±SEM) age, 22.32 ± 3.28 years; 59 females] participated in the study. They were mostly psychology students, recruited from a larger sample of volunteers who were interviewed according to an adapted version of the Clinical Interview of the Diagnostic and Statistical Manual of Mental Disorders, fourth edition, for the exclusion of participants with neurological or psychiatric illness or drug consumption. Participants were compensated economically for their time (∼7€/h). All participants gave their written informed consent at each phase of the study (interview, buccal cell extraction, and EEG recordings), according to the Declaration of Helsinki and approved by the Bioethics Committee of the University of Barcelona. Before the recording session, participants completed the State and Trait Anxiety inventory (Spielberger et al., 1988) and a questionnaire regarding musical experience. All participants had normal or corrected-to-normal vision and normal audition as assessed at the beginning of the recording session with a standard pure-tone audiometry using DT48-A headphones (Beyerdynamic). After exclusion by diagnostic criteria and after obtaining the 5-HTTLPR polymorphism, participants were invited for an EEG recording session. A total of 21 participants were excluded, as follows: 5-HTTLPR testing was inconclusive for 14 participants; 5 participants had very low SNRs (≤2); and 2 more participants had other technical issues regarding their EEG recordings.

Genotyping

Samples were collected using buccal cell cheek swabs (BuccalAmp DNA Extraction Kit, Epicentre), and DNA was extracted following manufacturer specifications. DNA samples were sent to an external genomics core facility (Progenika Biopharma) to determine genetic variants.

The 5-HTTLPR is a variable repeat sequence in the promoter region of the SLC6A4 gene. There are two allelic variants of this region: a short (S) variant comprising 14 copies of a 20–23 bp irregular repeat unit, and a long (L) variant comprising 16 copies. The 5-HTT removes 5-HT released into the synaptic cleft. Being a carrier of either the S or the L allele results in differential 5-HTT expression (Heils et al., 1996). Presence of the S allele is associated with reduced transcriptional activity and 5-HTT expression, relative to the L allele (Lesch et al., 1996), and therefore an increased availability of 5-HT in the synaptic cleft. Additionally, the L allele contains an A/G single nucleotide polymorphism (SNP; rs25531) which may render the LG allele functionally similar to the S allele (Hu et al., 2005).

Both the 5-HTTLPR (rs63749047) and the A/G SNP (rs25531) were determined through PCR amplification, digestion with the restriction enzyme Mspl, and migration in an agarose gel. Digestion was performed via incubation of 10 μl of the PCR product with 5 U of Mspl (New England Biolabs) for 1.5 h at 37°C. The PCR products and digestion were run in 2.5% agarose gels and separated through electrophoresis. Visualization took place in a UV transilluminator through SYBR Safe staining.

Thus, genotypes resulting from either the L or S allele and the A/G SNP contained in the L allele could be one of the following six: LA/LA, LA/LG, LG/LG, LA/S, LG/S, and S/S (Hu et al., 2005). The participants were initially assigned to one of three groups according to their level of 5-HTT expression, as follows: low (S/S, LG/S, LG/LG; 23 participants); medium (LA/LG, LA/S; 21 participants); and high (LA/LA; 14 participants). As pairwise comparisons did not yield any statistically significant differences regarding SNR (p = 0.93) and autocorrelation (p = 0.86) for the medium-expressing and the high-expressing groups, these participants were pooled together in one group (medium–high-expressing group, 35 participants).

Stimuli

Brainstem responses were recorded in response to the consonant-vowel syllable /ba/, which was generated with the Klatt speech synthesizer (Klatt, 1980) and designed as in the study by Slabu et al. (2012; Fig. 1A). The stimulus had a duration of 170 ms and a fundamental frequency (F0) of 100 Hz. The third (F3), fourth (F4), and fifth (F5) formants were set to 2900, 3500, and 4900 Hz, respectively. To elicit a large onset response, the first 5 ms included a rapid glide in the first (F1; from 400 to 1700 Hz) and second (F2; from 1700 to 1240 Hz) formant. As sounds never occur in isolation, for a more natural listening situation, and as differences between participants in signal extraction accuracy have been shown especially during challenging listening conditions (Song et al., 2011), a looped semantically anomalous speech babble was played in the background, 10 dB SPL softer than the syllable. The babble was recorded in a sound-attenuated booth, at a 44 kHz sampling rate and 16 bit accuracy using Audacity Software. It consisted in a short, semantically anomalous text, which was read in a natural and conversational manner by six native Spanish speakers (two male, four female), and then circularly looped with each speaker starting 10 s after the previous one and normalized off-line by root mean square (rms) amplitude normalization in Matlab version 7.4 (MathWorks; RRID: SCR_001622). To avoid any kind of interaction with the response to the syllable, the babble was looped with no silent intervals and the stimulation with the syllable started at a random phase of the babble.

Figure 1.

Figure 1.

Experimental design and FFR results. A, Design. Participants were presented with the consonant-vowel syllable /ba/ with a duration of 170 ms. The fundamental frequency was 100 Hz, while the third, fourth, and fifth formants were set to 2900, 3500, and 4900 Hz respectively. The syllable was presented at 80 dB SPL over a 10 dB SPL lower looped semantically anomalous speech babble in either a constant (SOA, 300 ms) or a jittered (SOA, 256–344 ms; mean SOA, 300 ms) condition. B, FFR grand average. Left, FFR responses for both conditions, measured at the CPz electrode. The steady-state part on which we focused in our analyses is highlighted in gray. Right, Power spectrum of the FFR for both timing conditions with slightly, but not significantly higher power in the constant condition.

The syllable was presented binaurally, with alternating polarities via ER-3A ABR insert earphones (Etymotic Research) at an intensity of 80 dB SPL and in the following two different timing conditions: constant and jittered stimulus onset asynchrony (SOA). In the constant condition, the SOA was set to 300 ms, while it ranged from 256 to 344 ms in the jittered condition (mean SOA, 300 ms). On the whole, 1008 trails/condition were presented in two separate blocks (the order of blocks was randomized across participants). Each block lasted ∼5 min, and participants were allowed to rest between blocks. During the whole recording session, participants sat in a comfortable chair, in an acoustically and electrically shielded room. As brainstem responses to sound are of a preattentive nature, participants were instructed to relax and focus on a silent movie with subtitles, while ignoring the auditory stimulation.

EEG recordings

The FFRs and cortical responses were extracted from continuous EEG recordings, acquired with a SynAmpsRT amplifier (Compumedics Neuroscan) and Scan version 4.4 software (Compumedics Neuroscan). Data were recorded from 36 scalp Ag/AgCl electrodes mounted in a nylon cap (Quick-Cap, Compumedics Neuroscan) at the standard 10–20 system locations. Two additional electrodes were positioned at the left and the right mastoids (M1 and M2, respectively). The electrooculogram (EOG) was measured with two bipolar electrodes placed above and below the left eye, and two horizontal electrodes placed on the outer canthi of the eyes. The ground electrode was located between Fz and FPz, and the right earlobe (A2) served as the on-line reference. All impedances were kept to <10 kΩ during the whole recording session, and data were on-line bandpass filtered from 0.05 to 3000 Hz and digitized with a sampling rate of 20 kHz.

Data processing and analyses

FFRs.

Data analysis was performed off-line using EEGlab version 7 toolbox (Delorme and Makeig, 2004; RRID: SCR_007292) running under Matlab version 2012a. The continuous recordings extracted from the CPz electrode were filtered off-line with a bandpass Kaiser window finite impulse response filter from 70 to 1500 Hz and epoched from 40 ms before the stimulus onset to 180 ms after stimulus onset.

An independent component analysis second-order blind identification (SOBI; Delorme and Makeig, 2004; Delorme et al., 2007) procedure was implemented to clean the recorded EEG from artifacts and to retain the signal activity corresponding to the FFR only. This was applied for each participant separately, on demeaned and detrended epochs of 10–170 ms from sound onset. All of the EEG and EOG channels were used in this step, and a total of 1008 epochs per condition were created. All epochs possibly containing artifacts were removed from the dataset via the automatic rejection algorithm implemented in EEGLAB (i.e., epochs with amplitudes exceeding ±95 μV), and epochs where the peak-to-peak values were exceeding 150 μV were removed from each participant's data (up to a maximum of 30% of all epochs). The cleaned data were input into the SOBI algorithm, and 40 independent components were found for each subject. Component activations were averaged, windowed with a Hanning taper, and zero padded to obtain a resolution of 1 Hz/bin, then fast Fourier transform (FFT) was computed on each of the components. First, the signal was estimated by reading out mean power values corresponding to a 10 Hz window around the fundamental frequency (100 Hz). Then an estimation of the baseline (noise) was performed, by taking the mean power value of a 10 Hz window on each side of the selected peak (with a separation of 20 Hz from the peak). Next, the SNR was calculated to select only components in which the amplitude of the signal at the frequency of interest was larger than the amplitude of the noise (according to an F test; significance level p < 0.05). All remaining components were treated as irrelevant for FFR analysis and pruned. On average, 36 components were retained for each subject. Finally, weights corresponding to FFR components were applied to the initially created epochs. The epochs were baseline corrected to a 40 ms interval preceding the sound onset (Russo et al., 2008).

Responses to stimuli of alternating polarities were averaged together to minimize stimulus artifacts and cochlear microphonic, preserving the FFR to the stimulus envelope (Aiken and Picton, 2008). In the FFR to the syllable, two parts can be distinguished: the consonant–vowel transition (10–65 ms), which poses a particular perceptual challenge to the listener (Assmann and Summerfield, 2004); and the steady-state part corresponding to the processing of the vowel (65–179 ms), which is characterized by constant temporal features (Song et al., 2011).

Epochs for the constant and the random blocks were averaged separately, and the grand average was calculated. To examine the FFR in the frequency domain, FFT (Cooley and Tukey, 1964) was applied to demeaned, zero-padded (1 Hz resolution) averages, windowed with a Hanning taper. The mean response amplitude was computed using a 10 Hz window surrounding F0 (90–110 Hz) and the subsequent three harmonics: H2, H3, and H4. Higher harmonic components were not reliably present in all subjects; therefore, only the response to F0 was statistically analyzed. As an index for response robustness, the SNR was determined by calculating the mean value of noise in a window of 10 Hz on either side of the peak, with a separation of 20 Hz between the window surrounding the peak and the window taken for noise, and subsequently dividing the mean signal value (power; e.g., squared amplitude) by the mean noise value. Overall condition and genetic effects were assessed by means of a two-way repeated-measures ANOVA with the within-subjects factor timing (constant vs jittered) and the between-subjects factor 5-HTT expression (low vs medium–high). Years of musical training were introduced to the ANOVA as a covariate. As the experimental manipulation of presentation timing of the syllable did not yield any significant results (SNR: F(1,65) = 0.04, p = 0.83; pitch strength: F(1,65) = 1.3, p = 0.26), we merged these two conditions. To exclude the possibility that differences in the prestimulus activity between groups could account for any measured effect, we calculated the rms values for both groups (mean ± SEM: low-expression group, 0.232 ± 0.027; medium–high-expression group, 0.267 ± 0.022) and found no significant difference (F(1,65) = 0.714; p = 0.493).

Pitch is an important psychoacoustic perceptual attribute for language as well as music processing. The perceived pitch is closely related to the periodicity of the acoustic stimulus waveform (Shofner, 2002). Therefore, pitch strength is a measure of response periodicity, indicating the robustness of neural phase locking to the stimulus F0 contour. To determine the pitch strength of the response for the different genetic groups, autocorrelation analysis was performed. Periodicity is most often quantified by the fundamental frequency, which is the number of times the period repeats in 1 s. For quantifying the neural pitch strength, the average response of each subject was copied, the copy delayed, and the cross-correlation for the original and the delayed copy was computed by applying a short-term correlation analysis with a 40 ms sliding window and a step size of 1 ms to find the highest peak in the correlation at the time shift equal to the period (Boersma, 1993; Krishnan et al., 2005, 2010; Jeng et al., 2011a, 2011b). Correlation is maximal at time lag 0, as the signal and its copy are identical. Furthermore, in the case of a periodic signal, the correlation is high for a time shift equal to the period. Hence, periodic sounds have a peak in the correlation at time shifts corresponding to the period (and multiples of the period). In our case, a peak was found at a time lag of 10 ms, corresponding to the F0 of the sound (frequency = 1/periodicity, so 100 Hz = 1/10 ms). To account for the transmission delay of the headphones, as well as neural delay, the analysis of the responses started at 15 ms. Pitch strength values were Fisher transformed, and averaged and analyzed with a two-way repeated-measures ANOVA with the within-subject factor condition (constant, jittered) the between-subjects factor 5-HTT expression (low and medium to high). The Greenhouse–Geisser correction was applied whenever the assumption of sphericity was violated.

Cortical responses.

To obtain cortical responses, we downsampled the data to 500 Hz, reaveraged to the average reference, applied a digital bandpass filter from 0.5 to 35 Hz, and averaged the resulting long-latency auditory-evoked potentials for both timing conditions off-line in epochs ranging from −40 to 300 ms. All epochs in which the signal exceeded ±100 Hz were rejected as containing artifacts. Seven subjects were excluded from analyses as >25% of all trials were rejected. Grand averages for both groups were computed. The resulting cortical potentials showed no prominent N1 response, probably due to the fast presentation rate (Tremblay et al., 2001), which is not optimal for the extraction of cortical potentials due to the refractory period of neurons in the auditory cortex. We performed a two-way nonparametric permutation ANOVA with the within-subjects factor timing (constant vs jittered) and the between-subjects factor 5-HTT expression (low vs medium to high) to objectively assess whether differences between conditions and groups existed at any time point or electrode. This method prevents from incurring circular analysis of the data (double dipping) by selecting electrodes and time windows (e.g., from difference waveforms) maximizing the observed effect (Kriegeskorte et al., 2009). Also, this approach ensures that no significant differences are observed in the baseline, and it effectively corrects for multiple comparisons. We set significance to p = 0.05, then lowered it to p = 0.01 and performed 4000 permutations, applying the false discovery rate correction for multiple comparisons. There was no significant difference between groups. Further permutation analysis with a less stringent criterion (2000 and 1000 permutations instead of the initial 4000) also failed to reveal any significant effects of group on cortical potentials. As the auditory response has been shown to adapt with stimulus repetition (Haenschel et al., 2005), we performed the analyses twice, once for all trials and once for the 100 first trials. Again, as there were no main effects of the timing condition or any interaction effect, we merged the two timing conditions.

Results

We recorded the FFR and cortical responses of 79 healthy young adults to a consonant-vowel syllable (/ba/; F0 = 100 Hz), presented in an either constant or jittered timing condition (Fig. 1) and in a challenging listening context, over a multispeaker background speech babble. Beforehand, we had determined the allelic variation of the 5-HTTLPR from all participants and formed two groups according to serotonin transporter expression: one group with low and one group with medium–high serotonin transporter expression. As a measure of response robustness, we analyzed the SNR of the FFR to the F0 of the stimulus. A higher SNR implies that the neural response to the stimulus, in this case the syllable /ba/, is more precise, either due to a gain in the response to the stimulus or due to a more successful suppression of surrounding noise. Moreover, we copied part of the periodic steady-state response signal and then compared it to a time-shifted part of itself to determine periodicity, an indicator for pitch strength. To investigate whether the 5-HTTLPR polymorphism would also influence sound processing at the cortical level, we extracted and analyzed long-latency auditory-evoked potentials. A nonparametric two-way ANOVA with the factors timing (constant vs jittered) and 5-HTTLPR expression (low vs medium to high) was performed on all trials and just on the first 100 trials of each condition. Participants watched a silent movie during the experiment.

Grand average FFRs for both the constant and the jittered conditions are shown in Figure 2. We found a main effect for the between-subjects factor 5-HTT expression (F(2,65) = 5.71; p = 0.005; η2 = 0.15) with higher SNRs for the low-expression group (mean ± SEM: low-expression group, 48.55 ± 8.02; medium–high-expression group, 16.63 ± 6.5) for the steady-state vowel part of the response. There was no significant difference for the consonant–vowel transition (F(2,65) = 0.64; p = 0.53), nor did years of musical training influence the SNR (F(1,54) = 0.53; p = 0.47). Pitch strength differed significantly between the low- and medium–high-expressing groups (F(2,65) = 4.45; p = 0.016; η2 = 0.12), with higher autocorrelation values for the low-expressing group than for the medium–high-expressing group (0.88 ± 0.06 and 0.68 ± 0.05, respectively).

Figure 2.

Figure 2.

SNR and pitch strength results. A, FFR grand average for the low and medium–high 5-HTT-expressing groups. B, SNR and pitch strength barplots. The SNR for the low 5-HTT-expressing group was significantly higher than for the medium–high-expressing group (mean ± SEM: low-expressing group, 48.55 ± 8.02; medium–high-expressing group, 16.63 ± 6.5; p = 0.005), and pitch strength was significantly higher for that group (0.88 ± 0.06 and 0.68 ± 0.05, respectively; p = 0.016). C, Pitch strength. Autorcorrelogram functions for the genetic groups. The grand average of the response for each group has been copied, and this copy was delayed to obtain autocorrelograms indicating pitch strength. The x-axis values refer to the center of each window (e.g., window 1 at 40 ms, window 2 at 41 ms), and the y-axis values refer to the time lag of the autocorrelation function. The pitch strength at the time lag of 10 ms, corresponding to the fundamental frequency of the stimulus, is significantly higher for the low-serotonin transporter expression group than for the medium–high-expression group.

Grand average cortical potentials and scalp distribution maps for the two prominent peaks in the cortical response for both groups are shown in Figure 3. There were no significant differences between groups, with neither taking into account all trials nor considering only the first 100 trials. These results suggest that genetically determined 5-HTT expression influences the faithful neural representation at a subcortical level of the processing hierarchy, with independent cortical processing. However, since recording parameters, particularly the presentation rate of 3 Hz, were suboptimal for acquiring cortical potentials, caution is required in interpreting these results.

Figure 3.

Figure 3.

Long-latency cortical auditory-evoked potentials to the syllable /ba/ and scalp topographies for the two prominent peaks. On the left, results for all trials are shown, while the right part of the figure shows results for the first 100 trials only. Permutation analyses revealed no significant differences between groups at any time point or electrode.

Discussion

The present study revealed that individuals with lower serotonin transporter expression, compared with individuals with medium to high expression had higher signal-to-noise ratios in the frequency following a response to the vowel of the syllable (/ba/), as well as higher pitch strength for their subcortical neuronal responses, both pointing toward a sharper speech signal extraction in the subcortical auditory pathway. The lack of significant results at cortical level indicates the specificity of the influence of 5-HTT expression to the subcortical neuronal representation of a sound, rather than its more abstract cortical processing. Although previous studies have shown the importance of serotonin for auditory perception, the 5-HTTLPR had not been previously implicated in speech signal extraction at subcortical stages as an important mechanism for language and music perception. The influence of genetics on both speech (Morell et al., 2007) and nonspeech auditory processing (Brewer et al., 2016) has been suggested by twin studies, which have shown that auditory processing is highly determined by the individual genetic makeup.

Former studies have linked serotonin to auditory disorders, such as tinnitus (Norena et al., 1999; Simpson and Davies, 2000; Caperton and Thompson, 2011) and hyperacusis (Marriage and Barnes, 1995; Attri and Nagarkar, 2010), and the auditory symptoms of psychiatric disorders, such as an altered loudness dependence of auditory-evoked potentials in schizophrenia (Juckel et al., 2003, 2008; Park et al., 2010, 2015; Juckel, 2015), and to differential auditory brain responses in depression (Gopal et al., 2000, 2005; Chen et al., 2002; Kampf-Sherf et al., 2004; brainstem response waveforms, loudness levels, and long-latency auditory-evoked potentials N1 and P2). Variations in the serotonin level have been linked to auditory processing at the cortex, where changes in endogenous serotonin availability have been shown to influence the amplitude of the N1/P2 auditory-evoked component (Manjarrez et al., 2005), as well as the loudness dependence of this component (Gallinat et al., 2003; Juckel et al., 2008; Pogarell et al., 2008). Drugs related to the release of serotonin and dopamine significantly reduced P300 amplitude in a further study (Lee et al., 2016).

While we identified no previous study implying the 5-HTTLPR in speech encoding, serotonin has been linked to auditory processing downstream of the cortex, in the auditory brainstem and, particularly, in the IC. Different methodological approaches ranging from histological to immunohistochemical, to chromatographic techniques and electrical as well as chemical stimulation have corroborated the existence of serotonergic fibers in the IC of different mammalian species (Moore et al., 1978; Klepper and Herbert, 1991; Keesom and Hurley, 2016). The neuromodulator in the IC that has received the most attention is serotonin. In general, the presence of serotonin in the IC is considered essential to modulate the brain response to auditory stimuli (Hurley and Pollak, 1999, 2001, 2005; Hurley and Hall, 2011; Obara et al., 2014; Lavezzi et al., 2015). Postmortem studies in humans have shown high densities of serotonergic fibers, especially in the central nucleus of the IC (Lavezzi et al., 2015). Serotonin receptors fall into seven main families, and members of at least the 5-HT1 (Chalmers and Watson, 1991; Pompeiano et al., 1992; Thompson et al., 1994; Wright et al., 1995), 5-HT2 (Wright et al., 1995), 5-HT3 (Bohorquez and Hurley, 2009), 5-HT4 (Waeber et al., 1994), and 5-HT7 (To et al., 1995) families have been found in auditory neurons of the IC. There is suggestive evidence that serotonin influences acoustic processing in the IC by modulating the tone of activity of GABAergic inhibitory interneurons (Peruzzi and Dut, 2004; Obara et al., 2014). When serotonin is applied endogenously or exogenously, usually a suppression or decrease in the amplitude of evoked responses and spontaneous activity in the IC is observed (Hurley and Pollak, 1999). Functionally, this suppression might result in an increased selectivity of responses to auditory stimuli (Hurley and Sullivan, 2012), possibly leading to a more robust neural response by tuning synaptic activity to the stimulus properties and decreasing the representation of noise, as observed in the present study.

Coffey et al. (2016) suggested in a recent article that although the FFR is originated mainly in subcortical nuclei, there might exist cortical contributions to the signal as well. The F0 of 100 Hz of the syllable used in this study might well be within the phase-locking capabilities of single neurons of the cortex. Nevertheless, although cortical activity might contribute to the FFR, we found no difference between groups in the cortical response to the stimulus, indicating that genetically mediated 5-HTT expression possibly only affects the auditory processing at an early, subcortical level. This should be confirmed, however, by future studies with a slower stimulus presentation rate, which is more suitable for the extraction of cortical potentials.

It has been stated that individual differences in auditory function can be observed even in typically developing individuals (Chandrasekaran et al., 2012; Hairston et al., 2013; Skoe et al., 2013b), and the ultimate determinants for subcortical differences may arise from a combination of environmental and genetic factors (Chandrasekaran et al., 2012). Different models of how acoustic experience and auditory training mold the subcortical processing of sound have been proposed (Krishnan and Gandour, 2009; Kraus and Chandrasekaran, 2010; Patel, 2011). These converge in the so-called “layering hypothesis” (Skoe and Chandrasekaran, 2014), a recent proposal for how auditory function is shaped by different types of experience at different time scales (from seconds to years) and taking into account the metaplasticity of the system. The layering hypothesis states that not only cortical, but also subcortical auditory processing is influenced by the specific auditory experience of each individual, resulting in individual differences in auditory function (Chandrasekaran et al., 2012; Hairston et al., 2013; Skoe et al., 2013b). Our results show that common variants in the 5-HTTLPR influence the robustness and accuracy of speech encoding at subcortical stages, and could therefore constrain or expand the extent to which experience can shape the structure and function of the subcortical auditory pathway. The scope of this influence yet has to be determined, and eventually may entail a complex interplay both with other genes and the environmental factors that have been disclosed so far (Skoe and Chandrasekaran, 2014). Because a large range of studies has shown that encoding accuracy at subcortical stages is experience dependent and can be trained (for review, see Skoe and Chandrasekaran, 2014), special attention might be given at an early developmental stage to individuals with medium to high 5-HTT expression, a possible genetically determined difficulty for auditory processing that may be counterbalanced by musical or other specific auditory training. This may be of interest, for example, regarding the auditory processing disorder, in which the major symptom is a difficulty in efficiently extracting speech from noise despite normal hearing sensitivity (Rosen et al., 2010). The prevalence of this disorder is especially high in the pediatric population (in combination with co-occurring learning disorders, it is estimated to be as high as 10%; Brewer et al., 2016), and the identification of the genetic and neurophysiologic underpinnings might significantly contribute to the understanding of the disorder and yield to early intervention. Our study identifies the 5-HTTLPR as being implicated in auditory encoding. Whether it might be involved in the genetic predisposition to the disorder, eventually in combination with other genes and influenced by acoustic experience, cannot be answered by the present study, and remains an interesting question for future studies.

On the other hand, it is of note that a number of studies have revealed cognitive advantages for individuals carrying the short allele of the 5-HTTLPR (Roiser et al., 2007; Borg et al., 2009; Enge et al., 2011), which in the present study were shown to have a more accurate speech sound encoding at subcortical stages. Since a relation between the auditory brainstem responses, including the FFR, and higher-order cognitive functions, such as rapid learning (Skoe et al., 2013b), reading abilities (Hornickel and Kraus, 2013), and executive function (Krizman et al., 2012), has been established, we are tempted to suggest the FFR to complex speech sounds as an intermediate phenotype not only for auditory processing, but eventually for a range of specific cognitive assets.

We believe our results pave the way for a series of questions for future research. As it has been shown that the FFR is plastic and experience dependent, it would be interesting to study the potential interactions between carrying specific allelic variants of the 5-HTTLPR and musical experience or bilingualism, for instance. In addition, it may turn out that the 5-HTTLPR is implicated in specific clinical groups involving several types of speech impairments, such as in stuttering, which has been linked not only to production, but also to auditory perception (Corbera et al., 2005) and dyslexia. While awaiting future confirmation, our results point to a higher ratio of individuals with medium to high serotonin transporter expression in these groups than in the general population. Likewise, low 5-HTT expression might result in an advantage for musical ability or a better achievement in the correct pronunciation of a newly learned foreign language.

To conclude, we found a significantly higher SNR and pitch strength of the subcortical auditory response to speech sounds in individuals with a low serotonin transporter expression, compared with those with medium to high expression, with both measures pinpointing toward more accurate speech encoding at subcortical stages of the auditory pathway in the former group. While experience-dependent plasticity of subcortical sound and speech encoding has been shown in an established line of research, the involvement of the 5-HTTLPR polymorphism adds a new layer to the understanding of how the interplay of different hardwired and experiential factors shapes a function as complex as speech encoding. Future studies should aim at establishing the extent to which auditory processing at subcortical stages is influenced by the genetic makeup, and how genetic expression can be modulated by experience-dependent plasticity.

Footnotes

This work was supported by grants from the Spanish Ministry of Economy and Knowledge (Grant PSI2015-63664-P), the Catalan Government (Grant SGR2014-177), and the ICREA Acadèmia Distinguished Professorship awarded to C.E.

The authors declare no competing financial interests.

References

  1. Aiken SJ, Picton TW. Envelope and spectral frequency-following responses to vowel sounds. Hear Res. 2008;245:35–47. doi: 10.1016/j.heares.2008.08.004. [DOI] [PubMed] [Google Scholar]
  2. Assmann PF, Summerfield AQ. The perception of speech under adverse conditions. In: Greenberg S, Ainsworth WA, Popper AN, Fay RR, editors. Speech processing in the auditory system. New York: Springer; 2004. pp. 231–308. [Google Scholar]
  3. Attri D, Nagarkar AN. Resolution of hyperacusis associated with depression, following lithium administration and directive counselling. J Laryngol Otol. 2010;124:919–921. doi: 10.1017/S0022215109992258. [DOI] [PubMed] [Google Scholar]
  4. Banai K, Hornickel J, Skoe E, Nicol T, Zecker S, Kraus N. Reading and subcortical auditory function. Cereb Cortex. 2009;19:2699–2707. doi: 10.1093/cercor/bhp024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Basu M, Krishnan A, Weber-Fox C. Brainstem correlates of temporal auditory processing in children with specific language impairment. Dev Sci. 2010;13:77–91. doi: 10.1111/j.1467-7687.2009.00849.x. [DOI] [PubMed] [Google Scholar]
  6. Boersma J. Accurate short-term analysis of the fundamental frequency and the harmonics-to-noise ratio of a sampled sound. Proc Inst Phonetic Sci. 1993;17:97–110. [Google Scholar]
  7. Bohorquez A, Hurley LM. Activation of serotonin 3 receptors changes in vivo auditory responses in the mouse inferior colliculus. Hear Res. 2009;251:29–38. doi: 10.1016/j.heares.2009.02.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Borg J, Henningsson S, Saijo T, Inoue M, Bah J, Westberg L, Lundberg J, Jovanovic H, Andrée B, Nordstrom AL, Halldin C, Eriksson E, Farde L. Serotonin transporter genotype is associated with cognitive performance but not regional 5-HT 1 A receptor binding in humans. Int J Neuropsychopharmacol. 2009;12:783–792. doi: 10.1017/S1461145708009759. [DOI] [PubMed] [Google Scholar]
  9. Brewer CC, Zalewski CK, King KA, Zobay O, Riley A, Ferguson MA, Bird JE, McCabe MM, Hood LJ, Drayna D, Griffith AJ, Morell RJ, Friedman TB, Moore DR. Heritability of non-speech auditory processing skills. Eur J Hum Genet. 2016;24:1137–1144. doi: 10.1038/ejhg.2015.277. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Caperton KK, Thompson AM. Activation of serotonergic neurons during salicylate-induced tinnitus. Otol Neurotol. 2011;32:301–307. doi: 10.1097/MAO.0b013e3182009d46. [DOI] [PubMed] [Google Scholar]
  11. Chalmers DT, Watson SJ. Comparative anatomical distribution of 5-HT1A receptor mRNA and 5-HT1A binding in rat brain—a combined in situ hybridisation/in vitro receptor autoradiographic study. Brain Res. 1991;561:51–60. doi: 10.1016/0006-8993(91)90748-K. [DOI] [PubMed] [Google Scholar]
  12. Chandrasekaran B, Kraus N. The scalp-recorded brainstem response to speech: neural origins and plasticity. Psychophysiology. 2010;47:236–246. doi: 10.1111/j.1469-8986.2009.00928.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Chandrasekaran B, Hornickel J, Skoe E, Nicol T, Kraus N. Context-dependent encoding in the human auditory brainstem relates to hearing speech in noise: implications for developmental dyslexia. Neuron. 2009;64:311–319. doi: 10.1016/j.neuron.2009.10.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Chandrasekaran B, Kraus N, Wong PC. Human inferior colliculus activity relates to individual differences in spoken language learning. J Neurophysiol. 2012;107:1325–1336. doi: 10.1152/jn.00923.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Chen TJ, Yu YW, Chen MC, Tsai SJ, Hong CJ. Association analysis for serotonin transporter promoter polymorphism and auditory evoked potentials for major depression. Neuropsychobiology. 2002;46:57–60. doi: 10.1159/000065412. [DOI] [PubMed] [Google Scholar]
  16. Coffey EBJ, Herholz SC, Chepesiuk AM, Baillet S, Zatorre RJ. Cortical contributions to the auditory frequency-following response revealed by MEG. Nat Commun. 2016;7:11070. doi: 10.1038/ncomms11070. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Cooley JW, Tukey JW. An algorithm for the machine calculation complex Fourier series. Int J Comput Math. 1964;19:297–301. [Google Scholar]
  18. Corbera S, Corral MJ, Escera C, Idiazábal MA. Abnormal speech sound representation in persistent developmental stuttering. Neurology. 2005;65:1246–1252. doi: 10.1212/01.wnl.0000180969.03719.81. [DOI] [PubMed] [Google Scholar]
  19. Delorme A, Makeig S. EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J Neurosci Methods. 2004;134:9–21. doi: 10.1016/j.jneumeth.2003.10.009. [DOI] [PubMed] [Google Scholar]
  20. Delorme A, Sejnowski T, Makeig S. Enhanced detection of artifacts in EEG data using higher-order statistics and independent component analysis. Neuroimage. 2007;34:1443–1449. doi: 10.1016/j.neuroimage.2006.11.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Díaz B, Baus C, Escera C, Costa A, Sebastián-Gallés N. Brain potentials to native phoneme discrimination reveal the origin of individual differences in learning the sounds of a second language. Proc Natl Acad Sci U S A. 2008;105:16083–16088. doi: 10.1073/pnas.0805022105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Enge S, Fleischhauer M, Lesch KP, Strobel A. On the role of serotonin and effort in voluntary attention: evidence of genetic variation in N1 modulation. Behav Brain Res. 2011;216:122–128. doi: 10.1016/j.bbr.2010.07.021. [DOI] [PubMed] [Google Scholar]
  23. Gallinat J, Senkowski D, Wernicke C, Juckel G, Becker I, Sander T, Smolka M, Hegerl U, Rommelspacher H, Winterer G, Herrmann WM. Allelic variants of the functional promoter polymorphism of the human serotonin transporter gene is associated with auditory cortical stimulus processing. Neuropsychopharmacology. 2003;28:530–532. doi: 10.1038/sj.npp.1300042. [DOI] [PubMed] [Google Scholar]
  24. Gopal KV, Daly DM, Daniloff RG, Pennartz L. Effects of selective serotonin reuptake inhibitors on auditory processing: case study. J Am Acad Audiol. 2000;11:454–463. [PubMed] [Google Scholar]
  25. Gopal KV, Briley KA, Goodale ES, Hendea OM. Selective serotonin reuptake inhibitors treatment effects on auditory measures in depressed female subjects. Eur J Pharmacol. 2005;520:59–69. doi: 10.1016/j.ejphar.2005.07.019. [DOI] [PubMed] [Google Scholar]
  26. Haenschel C, Vernon DJ, Dwivedi P, Gruzelier JH, Baldeweg T. Event-related brain potential correlates of human auditory sensory memory-trace formation. J Neurosci. 2005;25:10494–10501. doi: 10.1523/JNEUROSCI.1227-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Hairston WD, Letowski TR, McDowell K. Task-related suppression of the brainstem frequency following response. PLoS One. 2013;8:e55215. doi: 10.1371/journal.pone.0055215. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Heils A, Teufel A, Petri S, Stöber G, Riederer P, Bengel D, Lesch KP. Allelic variation of human serotonin transporter gene expression. J Neurochem. 1996;66:2621–2624. doi: 10.1046/j.1471-4159.1996.66062621.x. [DOI] [PubMed] [Google Scholar]
  29. Hornickel J, Kraus N. Unstable representation of sound: a biological marker of dyslexia. J Neurosci. 2013;33:3500–3504. doi: 10.1523/JNEUROSCI.4205-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Hornickel J, Skoe E, Nicol T, Zecker S, Kraus N. Subcortical differentiation of stop consonants relates to reading and speech-in-noise perception. Proc Natl Acad Sci U S A. 2009;106:13022–13027. doi: 10.1073/pnas.0901123106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Hornickel J, Chandrasekaran B, Zecker S, Kraus N. Auditory brainstem measures predict reading and speech-in-noise perception in school-aged children. Behav Brain Res. 2011;216:597–605. doi: 10.1016/j.bbr.2010.08.051. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Hu X, Oroszi G, Chun J, Smith TL, Goldman D, Schuckit MA. An expanded evaluation of the relationship of four alleles to the level of response to alcohol and the alcoholism risk. Alcohol Clin Exp Res. 2005;29:8–16. doi: 10.1097/01.ALC.0000150008.68473.62. [DOI] [PubMed] [Google Scholar]
  33. Hurley LM, Hall IC. Context-dependent modulation of auditory processing by serotonin. Hear Res. 2011;279:74–84. doi: 10.1016/j.heares.2010.12.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Hurley LM, Pollak GD. Serotonin differentially modulates responses to tones and frequency-modulated sweeps in the inferior colliculus. J Neurosci. 1999;19:8071–8082. doi: 10.1523/JNEUROSCI.19-18-08071.1999. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Hurley LM, Pollak GD. Serotonin effects on frequency tuning of inferior colliculus neurons. J Neurophysiol. 2001;85:828–842. doi: 10.1152/jn.2001.85.2.828. [DOI] [PubMed] [Google Scholar]
  36. Hurley LM, Pollak GD. Serotonin shifts first-spike latencies of inferior colliculus neurons. J Neurosci. 2005;25:7876–7886. doi: 10.1523/JNEUROSCI.1178-05.2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Hurley LM, Sullivan MR. From behavioral context to receptors: serotonergic modulatory pathways in the IC. Front Neural Circuits. 2012;6:58. doi: 10.3389/fncir.2012.00058. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Jafari Z, Malayeri S, Rostami R. Subcortical encoding of speech cues in children with attention deficit hyperactivity disorder. Clin Neurophysiol. 2015;126:325–332. doi: 10.1016/j.clinph.2014.06.007. [DOI] [PubMed] [Google Scholar]
  39. Jeng FC, Chung HK, Lin CD, Dickman B, Hu J. Exponential modeling of human frequency-following responses to voice pitch. Int J Audiol. 2011a;50:582–593. doi: 10.3109/14992027.2011.582164. [DOI] [PubMed] [Google Scholar]
  40. Jeng FC, Hu J, Dickman B, Lin CY, Lin CD, Wang CY, Chung HK, Li X. Evaluation of two algorithms for detecting human frequency-following responses to voice pitch. Int J Audiol. 2011b;50:14–26. doi: 10.3109/14992027.2010.515620. [DOI] [PubMed] [Google Scholar]
  41. Juckel G. Serotonin: from sensory processing to schizophrenia using an electrophysiological method. Behav Brain Res. 2015;277:121–124. doi: 10.1016/j.bbr.2014.05.042. [DOI] [PubMed] [Google Scholar]
  42. Juckel G, Gallinat J, Riedel M, Sokullu S, Schulz C, Möller HJ, Müller N, Hegerl U. Serotonergic dysfunction in schizophrenia assessed by the loudness dependence measure of primary auditory cortex evoked activity. Schizophr Res. 2003;64:115–124. doi: 10.1016/S0920-9964(03)00016-1. [DOI] [PubMed] [Google Scholar]
  43. Juckel G, Gudlowski Y, Müller D, Ozgürdal S, Brüne M, Gallinat J, Frodl T, Witthaus H, Uhl I, Wutzler A, Pogarell O, Mulert C, Hegerl U, Meisenzahl EM. Loudness dependence of the auditory evoked N1/P2 component as an indicator of serotonergic dysfunction in patients with schizophrenia—a replication study. Psychiatry Res. 2008;158:79–82. doi: 10.1016/j.psychres.2007.08.013. [DOI] [PubMed] [Google Scholar]
  44. Kampf-Sherf O, Zlotogorski Z, Gilboa A, Speedie L, Lereya J, Rosca P, Shavit Y. Neuropsychological functioning in major depression and responsiveness to selective serotonin reuptake inhibitors antidepressants. J Affect Disord. 2004;82:453–459. doi: 10.1016/j.jad.2004.02.006. [DOI] [PubMed] [Google Scholar]
  45. Keesom SM, Hurley LM. Socially induced serotonergic fluctuations in the male auditory midbrain correlate with female behavior during courtship. J Neurophysiol. 2016;115:1786–1796. doi: 10.1152/jn.00742.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Klatt DH. Software for cascade/parallel formant synthesizer. J Acoust Soc Am. 1980;67:971–975. doi: 10.1121/1.383940. [DOI] [Google Scholar]
  47. Klepper A, Herbert H. Distribution and origin of noradrenergic and serotonergic fibers in the cochlear nucleus and inferior colliculus of the rat. Brain Res. 1991;557:190–201. doi: 10.1016/0006-8993(91)90134-H. [DOI] [PubMed] [Google Scholar]
  48. Kraus N, Chandrasekaran B. Music training for the development of auditory skills. Nat Rev Neurosci. 2010;11:599–605. doi: 10.1038/nrn2882. [DOI] [PubMed] [Google Scholar]
  49. Kriegeskorte N, Simmons WK, Bellgowan PS, Baker CI. Circular analysis in systems neuroscience: the dangers of double dipping. Nat Neurosci. 2009;12:535–540. doi: 10.1038/nn.2303. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Krishnan A, Gandour JT. The role of the auditory brainstem in processing linguistically-relevant pitch patterns. Brain Lang. 2009;110:135–148. doi: 10.1016/j.bandl.2009.03.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Krishnan A, Xu Y, Gandour J, Cariani P. Encoding of pitch in the human brainstem is sensitive to language experience. Brain Res Cogn Brain Res. 2005;25:161–168. doi: 10.1016/j.cogbrainres.2005.05.004. [DOI] [PubMed] [Google Scholar]
  52. Krishnan A, Bidelman GM, Gandour JT. Neural representation of pitch salience in the human brainstem revealed by psychophysical and electrophysiological indices. Hear Res. 2010;268:60–66. doi: 10.1016/j.heares.2010.04.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Krizman J, Marian V, Shook A, Skoe E, Kraus N. Subcortical encoding of sound is enhanced in bilinguals and relates to executive function advantages. Proc Natl Acad Sci U S A. 2012;109:7877–7881. doi: 10.1073/pnas.1201575109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Lavezzi AM, Pusiol T, Matturri L. Cytoarchitectural and functional abnormalities of the inferior colliculus in sudden unexplained perinatal death. Medicine (Baltimore) 2015;94:e487. doi: 10.1097/MD.0000000000000487. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Lee H, Wang GY, Curley LE, Sollers JJ, Kydd RR, Kirk IJ, Russell BR. Acute effects of BZP, TFMPP and the combination of BZP and TFMPP in comparison to dexamphetamine on an auditory oddball task using electroencephalography: a single-dose study. Psychopharmacology (Berl) 2016;233:863–871. doi: 10.1007/s00213-015-4165-x. [DOI] [PubMed] [Google Scholar]
  56. Lesch KP, Bengel D, Heils A, Sabol SZ, Greenberg BD, Petri S, Benjamin J, Müller CR, Hamer DH, Murphy DL. Association of anxiety-related traits with a polymorphism in the serotonin transporter gene regulatory region. Science. 1996;274:1527–1531. doi: 10.1126/science.274.5292.1527. [DOI] [PubMed] [Google Scholar]
  57. Manjarrez G, Hernandez E, Robles A, Hernandez J. N1/P2 component of auditory evoked potential reflect changes of the brain serotonin biosynthesis in rats. Nutr Neurosci. 2005;8:213–218. doi: 10.1080/10284150500170971. [DOI] [PubMed] [Google Scholar]
  58. Marriage J, Barnes NM. Is central hyperacusis a symptom of 5-hydroxytryptamine (5-HT) dysfunction? J Laryngol Otol. 1995;109:915–921. doi: 10.1017/s0022215100131676. [DOI] [PubMed] [Google Scholar]
  59. Møller AR, Jannetta PJ, Sekhar LN. Contributions from the auditory nerve to the brain-stem auditory evoked potentials (BAEPs): results of intracranial recording in man. Electroencephalogr Clin Neurophysiol. 1988;71:198–211. doi: 10.1016/0168-5597(88)90005-6. [DOI] [PubMed] [Google Scholar]
  60. Moore RY, Halaris AE, Jones BE. Serotonin neurons of the midbrain raphe: ascending projections. J Comp Neurol. 1978;180:417–438. doi: 10.1002/cne.901800302. [DOI] [PubMed] [Google Scholar]
  61. Morell RJ, Brewer CC, Ge D, Snieder H, Zalewski CK, King KA, Drayna D, Friedman TB. A twin study of auditory processing indicates that dichotic listening ability is a strongly heritable trait. Hum Genet. 2007;122:103–111. doi: 10.1007/s00439-007-0384-5. [DOI] [PubMed] [Google Scholar]
  62. Norena A, Cransac H, Chéry-Croze S. Towards an objectification by classification of tinnitus. Clin Neurophysiol. 1999;110:666–675. doi: 10.1016/S1388-2457(98)00034-0. [DOI] [PubMed] [Google Scholar]
  63. Obara N, Kamiya H, Fukuda S. Serotonergic modulation of inhibitory synaptic transmission in mouse inferior colliculus. Biomed Res. 2014;35:81–84. doi: 10.2220/biomedres.35.81. [DOI] [PubMed] [Google Scholar]
  64. Pallier C, Bosch L, Sebastián-Gallés N. A limit on behavioral plasticity in speech perception. Cognition. 1997;64:B9–B17. doi: 10.1016/S0010-0277(97)00030-9. [DOI] [PubMed] [Google Scholar]
  65. Parbery-Clark A, Strait DL, Hittner E, Kraus N. Musical training enhances neural processing of binaural sounds. J Neurosci. 2013;33:16741–16747. doi: 10.1523/JNEUROSCI.5700-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Park YM, Lee SH, Kim S, Bae SM. The loudness dependence of the auditory evoked potential (LDAEP) in schizophrenia, bipolar disorder, major depressive disorder, anxiety disorder, and healthy controls. Prog Neuropsychopharmacol Biol Psychiatry. 2010;34:313–316. doi: 10.1016/j.pnpbp.2009.12.004. [DOI] [PubMed] [Google Scholar]
  67. Park YM, Jung E, Kim HS, Hahn SW, Lee SH. Differences in central serotoninergic transmission among patients with recent onset, subchronic, and chronic schizophrenia as assessed by the loudness dependence of auditory evoked potentials. Schizophr Res. 2015;168:180–184. doi: 10.1016/j.schres.2015.07.036. [DOI] [PubMed] [Google Scholar]
  68. Patel AD. Why would musical training benefit the neural encoding of speech? The OPERA hypothesis. Front Psychol. 2011;2:142. doi: 10.3389/fpsyg.2011.00142. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Peruzzi D, Dut A. GABA, serotonin and serotonin receptors in the rat inferior colliculus. Brain Res. 2004;998:247–250. doi: 10.1016/j.brainres.2003.10.059. [DOI] [PubMed] [Google Scholar]
  70. Pogarell O, Koch W, Schaaff N, Pöpperl G, Mulert C, Juckel G, Möller HJ, Hegerl U, Tatsch K. [123I] ADAM brainstem binding correlates with the loudness dependence of auditory evoked potentials. Eur Arch Psychiatry Clin Neurosci. 2008;258:40–47. doi: 10.1007/s00406-008-5011-5. [DOI] [PubMed] [Google Scholar]
  71. Pompeiano M, Palacios JM, Mengod G. Distribution and cellular localization of mRNA coding for 5-HT1A receptor in the rat brain: correlation with receptor binding. J Neurosci. 1992;12:440–453. doi: 10.1523/JNEUROSCI.12-02-00440.1992. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Rocha-Muniz CN, Befi-Lopes DM, Schochat E. Investigation of auditory processing disorder and language impairment using the speech-evoked auditory brainstem response. Hear Res. 2012;294:143–152. doi: 10.1016/j.heares.2012.08.008. [DOI] [PubMed] [Google Scholar]
  73. Roiser JP, Müller U, Clark L, Sahakian BJ. The effects of acute tryptophan depletion and serotonin transporter polymorphism on emotional processing in memory and attention. Int J Neuropsychopharmacol. 2007;10:449–461. doi: 10.1017/S146114570600705X. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Rosen S, Cohen M, Vanniasegaram I. Auditory and cognitive abilities of children suspected of auditory processing disorder (APD) Int J Pediatr Otorhinolaryngol. 2010;74:594–600. doi: 10.1016/j.ijporl.2010.02.021. [DOI] [PubMed] [Google Scholar]
  75. Russo NM, Skoe E, Trommer B, Nicol T, Zecker S, Bradlow A, Kraus N. Deficient brainstem encoding of pitch in children with Autism Spectrum Disorders. Clin Neurophysiol. 2008;119:1720–1731. doi: 10.1016/j.clinph.2008.01.108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Shofner WP. Perception of the periodicity strength of complex sounds by the chinchilla. Hear Res. 2002;173:69–81. doi: 10.1016/S0378-5955(02)00612-3. [DOI] [PubMed] [Google Scholar]
  77. Simpson JJ, Davies WE. A review of evidence in support of a role for 5-HT in the perception of tinnitus. Hear Res. 2000;145:1–7. doi: 10.1016/S0378-5955(00)00093-9. [DOI] [PubMed] [Google Scholar]
  78. Skoe E, Chandrasekaran B. The layering of auditory experiences in driving experience-dependent subcortical plasticity. Hear Res. 2014;311:36–48. doi: 10.1016/j.heares.2014.01.002. [DOI] [PubMed] [Google Scholar]
  79. Skoe E, Kraus N. Auditory brain stem response to complex sounds: a tutorial. Ear Hear. 2010a;31:302–324. doi: 10.1097/AUD.0b013e3181cdb272. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Skoe E, Kraus N. Hearing it again and again: on-line subcortical plasticity in humans. PLoS One. 2010b;5:e13645. doi: 10.1371/journal.pone.0013645. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Skoe E, Krizman J, Kraus N. The impoverished brain: disparities in maternal education affect the neural response to sound. J Neurosci. 2013a;33:17221–17231. doi: 10.1523/JNEUROSCI.2102-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Skoe E, Krizman J, Spitzer E, Kraus N. The auditory brainstem is a barometer of rapid auditory learning. Neuroscience. 2013b;243:104–114. doi: 10.1016/j.neuroscience.2013.03.009. [DOI] [PubMed] [Google Scholar]
  83. Slabu L, Grimm S, Escera C. Novelty detection in the human auditory brainstem. J Neurosci. 2012;32:1447–1452. doi: 10.1523/jneurosci.2557-11.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Smith JC, Marsh JT, Brown WS. Far-field recorded frequency-following responses: evidence for the locus of brainstem sources. Electroencephalogr Clin Neurophysiol. 1975;39:465–472. doi: 10.1016/0013-4694(75)90047-4. [DOI] [PubMed] [Google Scholar]
  85. Song JH, Skoe E, Banai K, Kraus N. Perception of speech in noise: neural correlates. J Cogn Neurosci. 2011;23:2268–2279. doi: 10.1162/jocn.2010.21556. [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Spielberger CD, Gorsuch RL, Lushene RE. Cuestionario de ansiedad estado-rasgo—manual. Ed 3. Madrid, Spain: TEA Ediciones (by permission of Consulting Psychologists Press, Sunnyvale, CA); 1988. [Google Scholar]
  87. Thompson GC, Thompson AM, Garrett KM, Britton BH. Serotonin and serotonin receptors in the central auditory system. Otolaryngol Head Neck Surg. 1994;110:93–102. doi: 10.1177/019459989411000111. [DOI] [PubMed] [Google Scholar]
  88. To ZP, Bonhaus DW, Eglen RM, Jakeman LB. Characterization and distribution of putative 5-ht7 receptors in guinea-pig brain. Br J Pharmacol. 1995;115:107–116. doi: 10.1111/j.1476-5381.1995.tb16327.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Tremblay K, Kraus N, McGee T, Ponton C, Otis B. Central auditory plasticity: changes in the N1–P2 complex after speech-sound training. Ear Hear. 2001;22:79–90. doi: 10.1097/00003446-200104000-00001. [DOI] [PubMed] [Google Scholar]
  90. Waeber C, Sebben M, Nieoullon A, Bockaert J, Dumuis A. Regional distribution and ontogeny of 5-HT4 binding sites in rodent brain. Neuropharmacology. 1994;33:527–541. doi: 10.1016/0028-3908(94)90084-1. [DOI] [PubMed] [Google Scholar]
  91. White-Schwoch T, Woodruff Carr K, Thompson EC, Anderson S, Nicol T, Bradlow AR, Zecker SG, Kraus N. Auditory processing in noise: a preschool biomarker for literacy. PLoS Biol. 2015;13:e1002196. doi: 10.1371/journal.pbio.1002196. [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Wright DE, Seroogy KB, Lundgren KH, Davis BM, Jennes L. Comparative localization of serotonin1A, 1C, and 2 receptor subtype mRNAs in rat brain. J Comp Neurol. 1995;351:357–373. doi: 10.1002/cne.903510304. [DOI] [PubMed] [Google Scholar]

Articles from The Journal of Neuroscience are provided here courtesy of Society for Neuroscience

RESOURCES