Skip to main content
Journal of Neurophysiology logoLink to Journal of Neurophysiology
. 2015 Nov 4;115(2):868–886. doi: 10.1152/jn.00953.2015

Two distinct representations of social vocalizations in the basolateral amygdala

Marie A Gadziola 1,2, Sharad J Shanbhag 1, Jeffrey J Wenstrup 1,2,
PMCID: PMC4839489  PMID: 26538612

Abstract

Acoustic communication signals carry information related to the types of social interactions by means of their “acoustic context,” the sequencing and temporal emission pattern of vocalizations. Here we describe responses to natural vocal sequences in adult big brown bats (Eptesicus fuscus). We first assessed how vocal sequences modify the internal affective state of a listener (via heart rate). The heart rate of listening bats was differentially modulated by vocal sequences, showing significantly greater elevation in response to moderately aggressive sequences than appeasement or neutral sequences. Next, we characterized single-neuron responses in the basolateral amygdala (BLA) of awake, restrained bats to isolated syllables and vocal sequences. Two populations of neurons distinguished by background firing rates also differed in acoustic stimulus selectivity. Low-background neurons (<1 spike/s) were highly selective, responding on average to one tested stimulus. These may participate in a sparse code of vocal stimuli, in which each neuron responds to one or a few stimuli and the population responds to the range of vocalizations across behavioral contexts. Neurons with higher background rates (≥1 spike/s) responded broadly to tested stimuli and better represented the timing of syllables within sequences. We found that spike timing information improved the ability of these neurons to discriminate among vocal sequences and among the behavioral contexts associated with sequences compared with a rate code alone. These findings demonstrate that the BLA contains multiple robust representations of vocal stimuli that can provide the basis for emotional/physiological responses to these stimuli.

Keywords: acoustic communication, bat, electrocardiogram, Eptesicus fuscus, vocalizations


social communication by sound depends on acoustic signals composed of basic spectro-temporal elements—syllables—that are combined in temporal patterns. Both the types of vocal syllables emitted and the temporal emission patterns, or sequencing, carry significant information related to the behavioral state of the sender and can influence the behavior of conspecifics: bats (Bastian and Schmidt 2008; Gadziola et al. 2012a; Gould and Weinstein 1971; Janssen and Schmidt 2009; Russ et al. 2005), primates (Fichtel and Hammerschmidt 2002; Fischer et al. 2002; Rendall 2003), tree shrews (Schehka et al. 2007), ground squirrels (Leger and Owings 1977; Mateo 2010), dogs (Pongracz et al. 2006; Yin and McCowan 2004), African elephants (Soltis et al. 2009), and guinea pigs (Arvola 1974). An accurate interpretation of these social signals thus depends on appropriate analysis of the “acoustic context,” defined here as the sequencing and temporal emission pattern of vocalizations within which they are presented. This study examines how a brain region critical for the response to social signals, the amygdala, responds to the natural sequencing of syllables used in vocal communication.

Given its central roles in establishing associations between sensory cues and biologically important events and in orchestrating emotional responses (LeDoux 2000, 2007), the amygdala likely serves an important function in acoustic communication (Gadziola et al. 2012b; Grimsley et al. 2013; Johnstone et al. 2006; Morris et al. 1999; Naumann and Kanwal 2011; Parsana et al. 2012; Peterson and Wenstrup 2012; Sander and Scheich 2001). Auditory information from the medial geniculate body and auditory association cortex enters the amygdala via the lateral nucleus (LeDoux et al. 1991; Mascagni et al. 1993). With widespread projections to hypothalamic, midbrain, and medullary nuclei, the central nucleus of the amygdala is well positioned to coordinate behavioral, autonomic, and neuroendocrine responses to biologically significant stimuli (Cardinal et al. 2002; Price 2003; Sah et al. 2003). Amygdalar responses also modulate subsequent auditory processing through direct projections to auditory cortex (Amaral and Price 1984), nucleus basalis (Sah et al. 2003), and, at least in bats, the inferior colliculus (Marsh et al. 2002). This circuitry establishes the basis for emotional responses to social vocalizations, but how amygdalar neurons respond to these signals and encode the emotional content within them remain open questions.

Several studies utilizing different species suggest different answers to these questions. Although a common finding is that neurons generally display stronger excitatory discharge in response to vocalizations associated with negative affect (Gadziola et al. 2012b; Naumann and Kanwal 2011; Parsana et al. 2012; Peterson and Wenstrup 2012), some studies report excitatory responses to stimuli associated with both positive and negative affect (Gadziola et al. 2012b; Peterson and Wenstrup 2012). These and another study (Grimsley et al. 2013) describe a key role for the temporal patterning of spike discharge to convey information about acoustic stimuli and the associated behavioral context. A limitation of most of these studies, however, is that they have relied on the presentation of social syllables in isolation (Gadziola et al. 2012b; Naumann and Kanwal 2011; Peterson and Wenstrup 2012) and show considerable variation in choice of stimulus repetition rates and randomization procedures. Because syllable ordering and temporal emission patterns provide the necessary acoustic cues for distinguishing the affective meaning of vocalizations (Bastian and Schmidt 2008; Bohn et al. 2008; Gadziola et al. 2012a), such methodological differences in stimulus presentation may alter the meaning of a stimulus to a listener. Amygdalar neurons, which appear to encode the meaning of vocal stimuli (Grimsley et al. 2013; Parsana et al. 2012; Wang et al. 2008a) should be affected by these methodological features.

The present study examines responses to natural, species-specific vocal sequences in adult big brown bats. This species (Eptesicus fuscus) is a highly social and long-lived (Brunet-Rossinni and Austad 2004; Kurta and Baker 1990; Paradiso and Greenhall 1967) mammal with conspecific relationships that last over multiple years (Willis et al. 2003; Willis and Brigham 2004). In its social interactions, it relies heavily on acoustic communication and possesses a complex vocal repertoire that mediates mother-infant interactions (Gould et al. 1973; Gould and Weinstein 1971; Monroy et al. 2011), social foraging behavior (Wright et al. 2014), a range of affiliative and aggressive behaviors during roosting (Gadziola et al. 2012a), and, presumably, mating (no reports). Across these studies, >20 distinct syllables have been identified. Bats are also one of the few mammalian taxa to display vocal learning (Boughman 1998; Esser 1994; Knörnschild et al. 2010), and this may extend to the family that includes the big brown bats (Knörnschild 2014). For all of these reasons, we believe this species forms a valuable mammalian model of social vocal interactions that may be mediated by the amygdala in humans.

We first assess how vocal sequences modify the internal affective state of a listener [via heart rate (HR)] and then characterize single-neuron responses in the basolateral division of the amygdala (BLA) to both isolated syllables and vocal sequences. The neurophysiological studies addressed several questions: 1) Do neuronal responses distinguish among individual vocal sequences? 2) Do neuronal responses distinguish among vocalizations associated with different behavioral contexts? 3) By what response features do neurons distinguish among vocalizations or contexts? Since previous studies suggest that both stimulus timing and spike timing are relevant to the BLA encoding of vocal stimuli, we compared responses to syllables and sequences, examined responses as a function of syllable repetition rate, and assessed the role of spike timing in coding. The results show that vocal sequences differentially modify the internal affective state of a listening bat, with moderately aggressive vocalizations evoking a greater change in HR compared with appeasement or neutral sequences. Furthermore, amygdalar neurons form distinct populations that participate in very different representations of social vocalizations. Finally, the results indicate that the timing of action potentials plays an important role in one of these forms of representation.

MATERIALS AND METHODS

Data were obtained from big brown bats (E. fuscus) maintained in a captive research colony at Northeast Ohio Medical University. Electrocardiogram (ECG) recordings were obtained from four adults (3 males, 1 female), and neurophysiological recordings were obtained from the BLA in five additional adults (1 male, 4 females). All animal husbandry and experimental procedures were approved by the Institutional Animal Care and Use Committee at Northeast Ohio Medical University (protocol no. 12-033).

Surgical Procedures

Animals received an injection of Torbugesic (5 mg/kg ip; Fort Dodge Animal Health, Fort Dodge, IA) before surgery and then were anesthetized to effect with isoflurane (2–4%; Abbott Laboratories, North Chicago, IL). Hair overlying the dorsal surface of the skull was removed with depilatory lotion, and the skin was disinfected with iodine solution. A midline incision was made, and the underlying muscle was reflected laterally to expose the skull. A metal pin was cemented onto the skull to secure the head to a stereotaxic apparatus. For bats used in neurophysiological studies, a sharpened tungsten ground electrode was inserted through a small hole in the skull and cemented in place. Based on stereotaxic coordinates, a craniotomy (<1-mm diameter) was made to access the amygdala. The underlying dura was removed and the craniotomy covered with sterile bone wax. For bats used in ECG monitoring studies, hair overlying the dorsal flanks on both sides of the back was removed to allow electrodes to contact the skin. Immediately after surgery, local anesthetic (lidocaine) and antibiotic cream were applied to the surgical areas, an injection of carprofen (3–5 mg/kg sc; Butler Schein Animal Health, Pittsburgh, PA) was administered, and the animal was returned to the holding cage for a 2-day recovery period before physiological experiments.

Acoustic Stimuli

Digitized acoustic stimuli, including both synthetic stimuli and prerecorded vocalizations, were converted to analog signals at 250–500 kHz and 16-bit depth with SciWorks (DataWave Technologies, Loveland, CO). Signals were antialias filtered (cutoff frequency of 125 kHz) and attenuated [Tucker-Davis Technologies (TDT)] and then amplified and sent to an electrostatic speaker (TDT model ES1). The system response had a gradual roll-off of ∼3 dB per 10 kHz. Harmonic distortion components were not detectable ∼45 dB below the signal level. Speaker placement differed in the two experiments and is described separately below.

Big brown bat social vocalizations were previously recorded from captive animals in defined social interactions and characterized in detail (Gadziola et al. 2012a). The present study used as stimuli both isolated syllables as well as vocal sequences. A “syllable” is the smallest acoustic unit of a vocalization, defined as one continuous emission surrounded by background noise. A “vocal sequence” is a longer emission pattern that can be defined by unique syllable sequencing and inter- and intracall boundaries. Syllable types were named according to their acoustic structure.

These vocalizations were assigned to a behavioral context based on nonvocal, overt behavioral displays (Gadziola et al. 2012a). Lower-aggression vocalizations were emitted by roosting bats in conjunction with aggressive behavioral displays (baring teeth, biting) after they had been jostled or disturbed by a conspecific. Bats transitioned to high-aggression vocalizations after prolonged or repeated disturbances, producing distinct behavioral displays (snapping, darting, and defensive posturing before physical contact). Appeasement vocalizations were never associated with aggressive displays but were instead related to increased social contact among conspecifics. Furthermore, the distinction between lower- and high-aggression calls was supported by differences in the change in HR in response to tactile irritation: high-aggression calls occurred after longer periods of tactile irritation and were associated with larger HR changes (duration and magnitude). Both males and females emitted social vocalizations in each behavioral context.

Although our previous study identified 18 distinct social syllable types, we reduced our stimulus set to include 12 syllable types (Fig. 1), choosing the most probable syllables and those with diverse spectro-temporal characteristics. An echolocation syllable type was also included to represent a neutral context. Because there are many more syllables used in aggressive interactions than in appeasement interactions, our syllable set includes more aggression than appeasement syllable types (n = 9 aggression, 3 appeasement, 1 neutral). In choosing natural vocal sequences, we balanced our stimulus design by choosing two or three characteristic exemplars to represent each of the four behavioral contexts: lower aggression, high aggression, appeasement, and neutral (Fig. 2). Vocalizations recorded under these behavioral contexts were unique in terms of the syllable type and sequencing of syllables (Fig. 2) (Gadziola et al. 2012a). For example, lower-aggression vocal sequences were typically composed of the same, rapidly repeating DFMs (downward frequency modulation short) syllable (Fig. 2C), whereas high-aggression vocal sequences included a more diverse range of syllable types emitted at slower repetition rates (Fig. 2D). Appeasement vocal sequences also contained unique syllable types and sequencing structure (Fig. 2B). All vocal stimuli selected were recorded in a natural behavioral context and not provoked by the experimenter. Appeasement vocalizations are typically emitted at a level ∼30 dB less intense than aggression or biosonar vocalizations.

Fig. 1.

Fig. 1.

All syllable types used as stimuli: waveforms (top) and spectrograms (bottom) for 13 representative syllable types previously identified as neutral (E), aggressive (DFMs, UFM, rBNBs, rBNBl, QCF, torQCF, sHFM, sAFM, and sinFM), or appeasing (DFMl, DFMl-QCFl, and DFMl-QCFl-UFM) based on the behavioral context in which they were predominantly observed (Gadziola et al. 2012a). Peak amplitude of all syllables is normalized. The time interval between syllables is for illustrative purposes and does not reflect natural emission rates. Syllable types were named according to acoustic structure (Gadziola et al. 2012a). E, echolocation; DFMs, downward frequency modulation short; UFM, upward frequency modulation; rBNBs, rectangular broadband noise burst short; rBNBl, rectangular broadband noise burst long; QCF, quasi-constant frequency; torQCF, torus quasi-constant frequency; sHFM, single-humped frequency modulation; sAFM, single-arched frequency modulation; sinFM, sinusoidal frequency modulation; DFMl, downward frequency modulation long; DFMl-QCFl, downward frequency modulation long to quasi-constant frequency long; DFMl-QCFl-UFM, downward frequency modulation long to quasi-constant frequency long to upward frequency modulation.

Fig. 2.

Fig. 2.

Acoustic sequences used as neurophysiological stimuli: waveforms (top) and spectrograms (bottom) for 8 acoustic sequences. For each context (A–D), 2 exemplars were used. Except for the tone sequence, natural vocal sequences were selected that are representative of such sequences (Gadziola et al. 2012a). Sequence composition: Neutral 1 (biosonar), natural sequence of echolocation calls; Neutral 2 (pure tones), sequence of five 25 kHz tone bursts; Appeasement 1, repeated syllable pairs of DFMl-QCFl-UFM and DFMl; Appeasement 2, repeated quartet of DFMl-QCFl-UFM, DFMl-QCFl-UFM, DFMl-QCFl-UFM, and DFMl-QFCl; Lower Aggression 1, sequence of DFMs syllables; Lower Aggression 2, sequence of DFMs syllables; High Aggression 1, sequence of rBNBs, rBNBs, rBNBl, rBNBs; High Aggression 2, sequence of sAFM, sAFM, sHFM, QCF-DFM, sHFM, QCF-DFM, UFM.

Electrocardiogram Monitoring

ECG monitoring procedures were identical to those described in Gadziola et al. (2012a). Briefly, animals were lightly anesthetized by isoflurane inhalation prior to attachment of the HR monitor. The ECG was recorded differentially with two surface electrodes temporarily fixed to the dorsal flanks, allowing full range of movement. Physiological signals were transmitted via a wireless headstage (4.7 g) to a receiver (Triangle Biosystems), permitting bats to behave freely within the experimental chamber. On an initial training day, we first tested that the bats tolerated the weight of the headstage and placement of the surface electrodes by allowing them to move freely within the test chamber for 1 h. Although the headstage was 15–25% of body weight, the animals moved freely in their cage and would also interact with other bats if given the opportunity. The loudspeaker was placed in the center of the chamber floor to maintain the most consistent sound field for bats hanging from the ceiling or walls of the experimental chamber.

Analog output from the receiver was amplified and band-pass filtered (100–500 Hz, model 3600; A-M Systems, Sequim, WA), digitized (10 kHz/channel, 12-bit depth; DataWave Technologies), and stored on a computer. DataWave software controlled physiological and video data acquisition. A threshold was used to detect the positive deflection of each R wave within the QRS complex, and the corresponding time stamps were saved.

Bats were allowed 30 min to recover from anesthesia, acclimate to the experimental chamber, and achieve a stable baseline HR. The ECG was recorded over 5-min trials, along with video and vocalizations. The first 60 s of a trial was used to monitor the baseline HR. After 60 s, sounds were presented from a loudspeaker for 30 s and then instantaneous HR was measured for the remainder of the trial. Within a recording session, bats were presented with acoustic sequences assigned to one of five behavioral contexts: lower aggression, high aggression, appeasement, and a neutral control sequence of pure tones at two different intensity levels. The order of sequence presentation was randomized within each session. Vocalizations were presented at sound levels consistent with naturally occurring intensities; appeasement vocalizations were attenuated by 20 dB relative to aggression vocalizations. To directly test level-dependent changes, pure tone sequences were presented at 0-dB and 20-dB attenuation. As a reference, a 25-kHz tone (the average peak frequency across vocal sequences) at 0-dB attenuation corresponds to 97 dB SPL. For calls associated with a behavioral context, stimuli were composed of three exemplars that were randomized with equal weighting to minimize habituation to repeated sounds. Stimuli were presented at a rate of 1/s for 30 trials. Bats were not tested with isolated syllables.

For further analysis of ECG data, DataWave files were imported into MATLAB (MathWorks, Natick, MA). Instantaneous HR was computed as the reciprocal of R-R intervals in beats per minute (bpm). Movement artifact was reduced by removing outliers and applying a sliding window with a shape-preserving piecewise cubic interpolation to fill missing values. Baseline HR was calculated as the instantaneous HR averaged over the first 60 s of each trial. For each trial, change in HR was calculated by subtracting the baseline from the maximum HR value.

Neurophysiological Recordings

General methods used in auditory neurophysiological experiments were described previously (Gadziola et al. 2012b). These experiments were conducted in a single-walled acoustic chamber. On experiment days, animals were placed in a custom-built stereotaxic device used to guide electrode penetrations and aid histological reconstruction of recording sites. Animals were secured with a nonadhesive wrap and placed inside a small tube, where they rested comfortably while head-fixed within the stereotaxic apparatus. Bats were not sedated during experiments. If an animal showed signs of discomfort or distress, the experiment was terminated for the day. Recording sessions did not exceed 6 h and were limited to one per day. Physiological recordings were obtained from both sides of the amygdala. On average, animals were used in seven recording sessions (range: 4–9 sessions) on different days.

Single-neuron activity was recorded with micropipette electrodes filled with 1 M NaCl (tip diameters of 2–5 μm) and advanced by a hydraulic micropositioner (Kopf 560). Extracellular potentials were amplified, band-pass filtered (300-5,000 Hz; A-M Systems 3600), and sampled at 40 kHz with 16-bit resolution (DataWave Technologies). DataWave software recorded spike occurrence and displayed peristimulus time histograms (PSTHs) and raster plots in real time. Spikes of constant waveform shape and amplitude determined single-neuron activity. Peak voltage of spikes typically exceeded background noise by a factor of 5. Neuronal spiking was identified either through background discharge or in response to acoustic search stimuli (noise bursts and social vocalizations). An explicit feature of the experimental design was an attempt to characterize auditory responses at any site with clear spiking activity (i.e., at least 2 overlapping spike waveforms per 5-min recording period). As a result, our sample includes many more auditory-unresponsive neurons, highly selective neurons, and neurons with low background discharge (<1 spike/s) than our previous study of the big brown bat amygdala (Gadziola et al. 2012b).

The loudspeaker was placed 10 cm from the animal and 25° into the sound field contralateral to the amygdala under study. Synthetic acoustic stimuli (noise and tone bursts, 2-ms rise-fall time, 30-ms duration; tonal sequences) and previously recorded social vocalizations from unfamiliar bats (13 syllables, 7 vocal sequences) were presented in blocks of 30 trials at a rate of 1/s. Duration of syllables ranged from 4 to 59 ms, and vocal sequences ranged from 148 to 766 ms. Peak amplitude was normalized across all stimuli. Once a neuron was encountered and isolated, initial data collection involved a standard series of rate/level functions in response to social vocalizations (syllables and sequences), broadband noise, and tones. Stimuli were typically presented at three attenuation levels: 0, 20, and 40 dB. Simple stimuli were used to assess onset latency and responsiveness to pure tones or broadband noise.

Since the repetition rate of individual syllables within a vocal sequence may distinguish different vocalizations or behavioral contexts, we examined the responsiveness of neurons to repetition rate. Artificial sequences consisting of repeated presentations of five different syllables were constructed. The syllables used were two aggression syllables [DFMs, rBNBs (rectangular broadband noise burst short)], one appeasement syllable [DFMl-QCFl-UFM (downward frequency modulation long to quasi-constant frequency long to upward frequency modulation)], and two neutral stimuli (echolocation pulse, 25-kHz tone burst). Repetition rates were varied (1, 2, 4, 10, 30, 50, 80, 110 Hz) while the number of syllable repeats remained the same (4 repetitions). The same syllable was used for all four repetitions within the sequence. Syllables with short durations (DFMs, echolocation pulse) were presented at the maximum repetition rates, while longer-duration syllables (e.g., DFMl-QCFl-UFM, rBNBs) were limited to repetition rates with periods greater than the syllable duration. Each artificial sequence stimulus was presented 30 times.

Verification of Recording Location

Stereotaxic coordinates developed in our laboratory for this species guided electrode penetrations through the amygdala (Gadziola et al. 2012b). Responses were recorded from an amygdala in several successive penetrations, the last of which was marked by a deposit of a neural tracer [Fluoro-Gold (Fluorochrome, Englewood, CO) or Fluoro-Ruby (Molecular Probes, Eugene, OR)]. Thus nearly all electrode penetrations were placed close to electrode tracks marked by deposits. Tracers were iontophoretically deposited with a constant current source; the electrode was kept in position for 10 additional minutes and then removed from the brain. Details of tracer-specific techniques have been described previously (Marsh et al. 2002; Yavuzoglu et al. 2010). Tracer deposit sites were photographed with a SPOT RT3 camera and SPOT Advanced Plus imaging software (version 4.7) mounted on a Zeiss Axio Imager M2 fluorescence microscope.

Spike Data Analysis

Only neurons that were presented with a complete syllable and sequence stimulus set were included in the analysis. Post hoc spike sorting using spike features and principal component analysis confirmed well-isolated single-neuron recordings and allowed removal of artifact when necessary. In some cases (n = 5 sites), two well-isolated neurons were present at the same recording site. Spike times and stimulus information were then exported to MATLAB for further analysis. A baseline background or “spontaneous” discharge rate was calculated from an initial 30-s nonstimulus control period. As background discharge can vary with stimulus history, a prestimulus background rate was also calculated from the 200 ms prior to stimulus onset for each block of trials featuring the identical stimulus type and level. Mean firing rates across trials were calculated in 10-ms bins, along with the 95% confidence interval. The criterion for a significant stimulus-evoked response was at least one 10-ms bin between stimulus onset and 100 ms after stimulus offset that did not have overlapping confidence intervals with the prestimulus background. For neurons with little or no prestimulus background discharge, the firing rate in a 10-ms bin was required to exceed 10 spikes/s. For neurons with high background rates, the response of each neuron was captured in a population response plot by determining whether the activity within each 10-ms bin was excitatory, inhibitory, or not significantly different by comparing each bin's confidence interval with the average confidence interval computed across the 200-ms prestimulus background.

The ability of amygdalar neurons to synchronize to the individual syllables of repeated stimuli was assessed by computing the vector strength (VS) for each neuron's response to each stimulus repetition. The Rayleigh statistic was used to assess significance of the VS measure, with P = 0.001 indicating significant synchronization (Mardia and Jupp 2000).

We examined the correspondence between each neuron's significant responses to the 13 individual syllables in our stimulus set and to the natural vocal sequences in which those syllables were present. For each neuron, we determined the number of sequences for which a significant response would be expected based upon the syllables that elicited a significant response. This expectation assumed that if a neuron responded to a syllable, it would respond to every sequence containing that syllable. The number of sequences that elicited significant responses was then divided by the expected number and expressed as a percentage. A complementary analysis used the sequences eliciting a significant response to determine a count of syllables for which a significant response was expected. Again, this expectation assumed that if a neuron responded to a sequence, it would respond to every syllable, presented in isolation, within that sequence. While this is less likely to be the case, our sequences were composed of relatively few syllable types. The composition of each test sequence is indicated in the legend of Fig. 2.

The spike distance metric (SDM), developed by Victor and Purpura (1996, 1997) and implemented in MATLAB by the Spike Train Analysis Toolkit (http://neuroanalysis.org/toolkit/), was used to assess the relative importance of spike timing to the underlying neural code and to determine whether spike discharge patterns can be used to discriminate among individual stimuli or the behavioral contexts associated with these (Victor and Purpura 1996, 1997). According to the SDM, Dspike[q], the distance (or dissimilarity) between two spike trains is the minimum total “cost” to transform one spike train into the other by a series of insertions, deletions, and time shifts of spikes. Both the insertion and deletion of a spike are set to a cost of 1, whereas shifting spikes by an amount of time t is set to a cost of qt. The cost parameter q (in units of 1/s) can be varied to characterize the relative importance of spike times and spike counts. When q = 0, the cost to move a spike in time is “free,” and distances are computed solely on the number of spikes (i.e., a rate code). As q increases, timing of individual spikes begins to exert a stronger influence on the distance metric, since a change in spike time of 1/q will have the same cost as simply deleting the spike.

A confusion matrix was constructed to assess how well responses to the same stimulus cluster together compared with pairs of responses from different stimuli. Each row represents instances of stimulus conditions, while each column represents instances of predicted stimulus conditions; correct assignments fall along the diagonal. For each trial of a given stimulus, Dspike was calculated between a given spike train and spike trains from other stimuli. The confusion matrix bin that obtained the minimal value was incremented by 1 to indicate the predicted stimulus or, in the case of a tie, was divided equally among the stimuli that shared the minimum Dspike value. The information, H (in units of bits), was then computed to evaluate the ability of the neuron to classify stimuli based on temporal patterns of auditory responses. H will be 0 when classification is random, whereas perfect discrimination of sound stimuli corresponds to H = log2(13) = 3.7 bits for syllables or H = log2(8) = 3 bits for vocal sequences.

The above process was repeated for a range of cost values (q = 0, 4, 8, 10, 12, 13, 15, 20, 25, 30, 35, 40, 45, 50, 60, 70, 80, 90, 100, 200, 300, 400, 500, 1,000, 2,000) to characterize the associated temporal integration window at which maximal information is found (Hmax). The contribution of spike timing to the stimulus classification was assessed by subtracting the information when the cost, q, was 0 (Hrate) from Hmax.

To further control for any bias associated with using a finite data set, the chance level of information, Hshuffle, was estimated by randomly reassigning spike trains to stimulus conditions and computing the information held in the confusion matrix of shuffled values (Victor and Purpura 1996). This process was repeated 100 times for each neuron to obtain a robust estimate of Hshuffle. All reported values of H calculated from the observed responses were corrected by subtracting Hshuffle from the original value. Any negative H values after correction were set to 0.

To determine whether spike discharge patterns transmit information about the behavioral context of vocal stimuli, a second SDM analysis was conducted as described above but combined all spike trains belonging to the same associated behavioral context instead of by stimulus identity. Pure tone sequences were excluded from this analysis, leaving four behavioral context categories: biosonar, appeasement, lower aggression, and high aggression. Thus perfect discrimination of behavioral context in this analysis corresponds to H = log2(4) = 2 bits.

RESULTS

HR Response of Listeners to Acoustic Stimuli

To understand how social vocalizations modify the behavioral and affective state of a listener, we first monitored ECG signals from animals during presentation of acoustic signals in five different categories: conspecific vocalizations associated with lower aggression, high aggression, and appeasement and neutral controls of pure tone sequences that were presented at each of two intensity levels.

Bats showed an average baseline HR of 460 ± 58 bpm during the 60 s prior to sound stimulus presentation. The HR of listeners showed significant elevations in response to sound on the majority of trials (81%, or 57 trials). However, conspecific vocalizations evoked a greater change in HR compared with the neutral pure tones (142 ± 35 vs. 76 ± 26 bpm, 2-sample t-test, P < 0.01).

There were two major findings from analysis of changes in HR. First, aggression vocalizations evoked the largest change in HR compared with appeasement vocalizations (171 ± 123 vs. 85 ± 72 bpm; 1-way ANOVA, P < 0.05) or neutral tones (171 ± 123 vs. 77 ± 70; 1-way ANOVA, P < 0.01). Second, and more surprising, lower-aggression vocalizations evoked a greater change in HR compared with all other behavioral contexts, including high aggression (220 ± 121 vs. 122 ± 107 bpm; 1-way ANOVA, P < 0.05; Fig. 3). An additional finding was that HR responses to neutral tones showed sound level-dependent changes, with larger elevation of HR for higher-intensity tones than for lower-intensity tones (106 ± 34 vs. 47 ± 36 bpm, independent t-test, P < 0.05; Fig. 3A).

Fig. 3.

Fig. 3.

Differential modulation of heart rate (HR) in response to sound. A: mean change in HR calculated by subtracting the baseline rate from the maximum HR value. Values were averaged across all trials for each of the 5 behavioral contexts. Error bars are 95% confidence intervals. Post hoc tests with Bonferroni correction revealed that lower-aggression calls evoked a greater change in HR compared with all other behavioral contexts. Higher-intensity tones evoked a greater change in HR compared with lower-intensity tones. *P < 0.05, **P < 0.01, ***P < 0.001. B: mean instantaneous HR and 95% confidence intervals averaged across lower-aggression and high-aggression trials. HR was first normalized to the average baseline value across all trials. Black bar represents timing of sound presentation. bpm, Beats per minute.

Amygdalar Responsiveness Is Related to Background Firing Rate

Neurophysiological approaches were used to examine how neurons in the BLA respond to and discriminate among vocal sequences and to compare differences in responses to isolated syllables and vocal sequences. Sixty-four neurons were histologically localized to the BLA. From these, 43 neurons (67%) were considered auditory responsive based on our criteria (see materials and methods).

Although significant periods of inhibition were observed in 19% of neurons, inhibition was generally difficult to detect unless there was a sufficiently high background firing rate. Testing responsiveness with 50-ms time bins, rather than the 10-ms bins normally used, resulted in an additional eight neurons being classified as auditory responsive and was more likely to detect significant suppression. However, the longer time bins were not useful for the temporal analysis of response patterns and were therefore not employed in subsequent analyses.

Background firing rates.

We first examined the distribution of mean firing rates of the background activity recorded from an initial 30-s period before any test stimulus presentation. This ensured that background firing rates were not influenced by stimulus history. The average background firing rate was 1.9 spikes/s for all neurons and 2.8 spikes/s for auditory-responsive neurons. Among the auditory-responsive neurons, background rate departed from a normal distribution (Kolmogorov-Smirnov test, P < 0.001) and was bimodal (coefficient of bimodality, b = 0.73; Fig. 4). For subsequent analyses, the population of auditory-responsive neurons was split into two groups based on a break in the distribution at a firing rate of 1 spike/s. Roughly half of the auditory-responsive neurons (22 of 43 neurons) showed background firing rates < 1 spike/s and are referred to as the “low-background” group; the remaining neurons showed background rates that were ≥1 spike/s and are referred to as the “high-background” group. Among auditory-unresponsive neurons (n = 21), the mean background rate was 0.1 spikes/s, with all such neurons showing background rates < 1 spike/s. No differences were found in action potential half-width or recording depth between the low- and high-background groups.

Fig. 4.

Fig. 4.

Distribution of background firing rates: histogram of background firing rates from all recorded neurons. Among auditory-responsive neurons, there is a nonnormal (Kolmogorov-Smirnov test, P < 0.001) and bimodal (coefficient of bimodality b > 0.55) distribution with a break at 1 spike/s. Histogram bins are 0.5 spikes/s.

To test whether the background activity of BLA neurons was modulated during the stimulus testing period, mean firing rates calculated during the initial 30-s silent period were compared to the 200-ms window prior to each stimulus onset. The majority of neurons (72%, or 46 of 64 neurons) showed significant modulation of the background activity to at least one stimulus. The background rates of most auditory-responsive neurons were modulated by acoustic stimulation (81%, or 35 of 43 neurons): roughly half of the responses showed an increase in background firing, and the other half showed a decrease. A smaller proportion of auditory-unresponsive neurons had significant modulation of background firing during the testing period (52%, or 11 of 21 neurons). For each of these “auditory-unresponsive” neurons, background firing increased in response to at least one stimulus. To control for this stimulus-induced modulation of background activity, all measures of stimulus-evoked spiking were computed relative to the background activity recorded during the 200 ms prior to stimulus onset.

Neuronal responsiveness to acoustic stimuli.

We next examined the relationship between background firing and the number of syllables and sequences to which a neuron responded (Fig. 5A). Neurons with low background rates were highly selective, responding on average to less than one tested syllable and one tested sequence (Fig. 5A). Across the entire suite of stimuli, most of these neurons responded to only one tested stimulus. In contrast, neurons with high background rates responded to most tested stimuli (Fig. 5A). On average, high-background neurons responded to 10 of 13 tested syllables and 7 of 8 tested sequences. The low- and high-background groups had minimal overlap in the number of stimuli to which the neurons responded (Fig. 5A). The distributions of responsiveness to syllables and sequences were both nonnormal (Kolmogorov-Smirnov test: syllables, P ≤ 0.001; sequences, P < 0.01) and bimodal (coefficients of bimodality: syllables, b = 0.74; sequences, b = 0.77).

Fig. 5.

Fig. 5.

Background discharge is related to auditory responsiveness. A: background firing rate for high- and low-background neurons plotted against the number of syllables (left) and sequences (right) that evoked a response from each neuron. Horizontal dashed line indicates 1 spike/s cutoff for low-background neurons. Although some high-background neurons were highly selective, most responded to multiple stimuli (mean = 11/13 syllables and 7/8 sequences). Neurons with low background rates respond to very few stimuli. B, left: high-background neurons (n = 21) responded to most acoustic stimuli. Social vocalizations were named according to acoustic structure (Gadziola et al. 2012a). Right: low-background neurons (n = 22) responded selectively but represented the range of acoustic stimuli tested.

Neurons in the BLA responded to a broad range of acoustic stimuli, including isolated syllables, vocal sequences, and synthetic acoustic stimuli. In general, tested stimuli were similarly effective at evoking a response among the sample of neurons. However, there was a dramatic difference in the proportion of neurons responding to stimuli between the low- and high-background groups (Fig. 5B). On average, 83 ± 11% of high-background neurons responded to each stimulus (Fig. 5B, left), while only 7 ± 5% of low-background neurons responded to each stimulus (Fig. 5B, right). Among low-background neurons, vocal sequences were more likely than syllables to evoke responses (2.6 ± 1.1 neurons per sequence vs. 1.2 ± 1.4 neurons per syllable; independent t-test, P < 0.05). Note that several low-background neurons were not responsive to either individual syllables or the natural sequences; such neurons responded only to broadband noise, tones, or syllables presented at certain repetition rates.

These differences in responsiveness between high- and low-background neurons are illustrated in Figs. 6 and 7. In Fig. 6, a neuron with low background firing (0.1 spikes/s) displayed a significant response to only two of the tested syllables and one of the tested vocal sequences. For this neuron, mean firing rates across a 200-ms window ranged from 0 to 11 spikes/s, but there was a robust response (peak firing rate of 77 spikes/s) only to the appeasement syllable, DFMl-QCFl. The high selectivity of this neuron is illustrated by the fact that closely related syllables (DFMl or DFMl-QCFl-UFM) evoked very little or no significant spiking. To sequences, there was little or no response; in particular, the neuron responded poorly to repetitions of the DFMl-QCFl syllable within the Appeasement 2 sequence. This result suggests that a neuron's firing depended substantially on the acoustic context within which the excitatory syllable was presented.

Fig. 6.

Fig. 6.

A low-background neuron (0.1 spikes/s) was highly selective, with a clear preference for the DFMl-QCFl syllable. For the 13 syllables (A) and 8 sequences (B) tested, mean firing rate (black line) and 95% confidence intervals (light gray outline) are shown, binned at 10 ms. The Appeasement 2 sequence includes the DFMl-QCFl syllable, but note the clear suppression in response relative to when the syllable is presented alone. Black bar represents stimulus timing and duration for syllables, whereas the waveforms of vocal sequences are plotted above. Inset: overlaid spike waveforms demonstrate a well-isolated neuron.

Fig. 7.

Fig. 7.

A high-background neuron (4.5 spikes/s) responded to all of the tested stimuli but showed clear differences in discharge pattern. For syllables (A) and sequences (B), mean firing rate (black line) and 95% confidence intervals (light gray outline) are shown, binned at 10 ms. Black bar represents stimulus timing and duration for syllables, whereas the waveforms of vocal sequences are plotted above. Note that responses to individual syllables do not always predict responses to sequences. Inset: overlaid spike waveforms demonstrate a well-isolated neuron.

In stark contrast, the high-background neuron (4.5 spikes/s) in Fig. 7 was responsive to all tested syllables and vocal sequences presented at the 20 dB attenuation level. Average firing rates calculated over a 200-ms window from stimulus onset ranged from 3 to 41 spikes/s across all stimuli, with peak firing rates within the same window ranging from 17 to 113 spikes/s. To syllables (Fig. 7A), the neuron's response magnitude and duration varied but were unrelated to syllable duration or behavioral context. The only consistent pattern was that most of the strongest responses were to syllables with an initial downward frequency-modulated component [e.g., E (echolocation), DFMs, DFMl, DFMI-QCFl, DFMl-QCFl-UFM]. The pattern of response to vocal sequences was more varied than the response to syllables. For several sequences (e.g., Biosonar, Appeasement 1, and Appeasement 2), the response was well locked to the temporal patterning of the constituent syllables. For other sequence responses (e.g., Lower Aggression 1 and Lower Aggression 2), the relationship to the syllable response was less clear, in part because of the high repetition rate of syllables in those sequences. It is noteworthy that both these sequences (Lower Aggression 1 and Lower Aggression 2) are exclusively composed of the DFMs syllable, yet the temporal response patterns are quite different. This suggests that the acoustic context within which syllables were presented affected the response of this neuron.

Since high-background neurons responded to most stimuli yet show considerable variability in their response properties, we explored how social vocalizations were represented across the sampled population of high-background neurons within the BLA. The proportion of neurons showing significant excitation was calculated for each 10-ms time bin in response to syllables and sequences in order to estimate the synchronized activity across the sample population (Figs. 8 and 9).

Fig. 8.

Fig. 8.

Population responses to isolated syllables. Eight representative syllables are illustrated, organized by the associated behavioral context: neutral syllable (black label), appeasement syllables (blue labels), and aggression syllables (red labels). Colored plots of population activity represent each high-background neuron in a single row. Each 10-ms response bin is colored based on excitation, inhibition, or no significant difference compared with the 200-ms prestimulus background window. Order of neurons is identical across all plots. Population responses are summarized below by calculating the proportion of neurons showing excitatory discharge at each 10-ms bin. Peak responsiveness is similar across the majority of syllables, regardless of associated behavioral context. Black bars represent stimulus timing and duration.

Fig. 9.

Fig. 9.

Population responses to 8 vocal sequences at 2 different sound levels (0-dB and 40-dB attenuation). Responses show unique temporal activation patterns that could not be predicted from responses to isolated syllables. Aggressive vocalizations (red labels) show a greater proportion of neurons with excitatory discharge than appeasement vocalizations (blue labels). At the lowest sound level (40-dB attenuation), population responses to 3 aggression sequences remained strong, while responses to appeasement and neutral sequences (black labels) dramatically decreased.

In response to all tested syllables, BLA neurons showed similar patterning of excitation regardless of syllable structure or behavioral context (Fig. 8). Isolated syllables typically evoked a single response peak across the sampled BLA neurons. Specifically, Fig. 8 shows that nearly all tested syllables, including both aggression and appeasement syllables, evoked a response in nearly 50% of high-background neurons during a 10-ms poststimulus bin. The noteworthy exception was the DFMs syllable, evoking a response among many fewer neurons. There was no clear preference for syllable type or calls associated with a particular behavioral context at any of the three sound levels presented (data not shown).

In response to vocal sequences, the population displayed temporal activation patterns that usually featured multiple response peaks (Fig. 9). Across the population, aggressive vocal sequences tended to evoke a more robust and potentially synchronized population response than appeasement sequences, and did so across attenuation levels. Furthermore, the population response patterns to the aggressive vocalizations differed substantially. These results suggest that the temporal patterning of discharge among the population may discriminate among vocal sequences. As with single neurons (previous section), the temporal patterns of the population did not always correspond to the response patterns to individual syllables. For example, the population response to single appeasement syllables (Fig. 8) was much stronger than that to the appeasement sequences (Fig. 9). Conversely, the population response to the Lower Aggression 1 sequence was much stronger than the response to isolated presentation of its constituent DFMs syllable.

The single neuron and population responses described in this section suggest that low- and high-background neurons participate in different amygdalar representations of vocal stimuli. Responses of low-background neurons may form a sparse code of vocal stimuli, based on their highly selective, relatively high firing rate response to a particular stimulus. In contrast, the high-background neurons appear to participate in a neural code based on broad responsiveness, in which discrimination among vocal stimuli may be expressed by differences in response timing and magnitude. For both groups of neurons, the data suggest that acoustic context influences the manner in which the neuron responds to constituent syllables. We explore this issue in the next section.

Responses to Syllables and Sequences Suggest Role for Acoustic Context

Since most studies of amygdalar coding of social vocalizations have examined the response to syllables, it is important to ask how the response to syllables relates to the response to sequences containing that syllable. If a neuron responds to a syllable in isolation, does it also respond to the sequences that contain that syllable? Conversely, if a neuron responds to a particular sequence, does it respond to the isolated syllable or syllables contained in that sequence? How does the response to syllables change as the repetition rate increase? The responses of the neurons illustrated in Figs. 69 suggest that the context within which syllables occurs may be important.

We assessed the correspondence between a high-background neuron's response to syllables and its response to the sequences that contain those syllables by dividing the number of observed sequence responses by the number expected on the basis of syllable responses. This value, expressed as a percentage, quantifies the correspondence between syllable responses and sequence responses, with 100% indicating maximum correspondence. Figure 10A displays this correspondence across all tested neurons that showed at least one syllable response. Across neurons, the median value of the correspondence was 73.3%, indicating an observed response to 73.3% of the sequences to which a neuron was expected to respond. A neuron's response to a syllable was not a highly reliable predictor of its response to a sequence containing that syllable.

Fig. 10.

Fig. 10.

Correspondence between syllable and sequence responses. A: distribution of % correspondence values based on syllable responses: the number of significant sequence responses that are expected based upon observed significant responses to syllables was counted for each neuron and expressed as %. Median % was 73.3, indicated by gray line on histogram. B: distribution of % correspondence values based on sequence responses: % of significant syllable responses that are expected based upon significant observed responses to sequences was computed for each neuron. Median % was 22.5 (gray line).

Do sequence responses predict the syllables to which a neuron will respond? We used each high-background neuron's response to sequences to develop an expectation of the number of syllables to which it would respond and then divided the number of observed syllable responses by the number expected on the basis of sequence responses. Figure 10B displays this correspondence across all tested neurons that showed at least one sequence response. Across neurons, the median value of the correspondence was 22.5%, indicating an observed response to 22.5% of the syllables expected on the basis of sequence responses. A neuron's response to sequences was a poor predictor of its response to syllables contained within the sequence. These observations suggest that the acoustic context within which syllables are presented may alter the neuron's response.

The rate at which syllables are repeated is a specific feature of acoustic context that carries behavioral contextual information in big brown bats (Gadziola et al. 2012a) and other species (Bastian and Schmidt 2008; Fischer et al. 2002; Schehka et al. 2007). Figures 69 suggest that low- and high-background amygdalar neurons may differ in their ability to follow the rate of syllable repetition in natural sequences. To explore this further, we examined responses of 25 neurons to tones and syllables repeated at rates from 1/s to 110/s, rates that occur naturally among vocal sequences used by big brown bats (Gadziola et al. 2012a). Figure 11 displays the results of this analysis. For low-background neurons (n = 13), very few responded to any of these stimuli, an expected result of the high selectivity of these neurons for acoustic stimuli. Furthermore, there was virtually no evidence that these neurons could follow natural rates of syllable repetition: in only one neuron's response was there significant locking to a repeated sequence at rates above 2/s. For high-background neurons (n = 12), most neurons responded to most presented rates of syllable presentation. However, synchronization to repeated syllables occurred in a minority of neurons and rarely occurred at tested rates higher than 10/s. These results indicate differences between the two groups of neurons in the ability to respond to and follow syllable repetitions but in general showed that BLA neurons do not follow the high rates of stimulus repetition present in many vocal sequences, such as the low-aggression sequences shown in Fig. 9.

Fig. 11.

Fig. 11.

Responses of amygdalar neurons as a function of syllable or tone repetition rate. Left: low-background neurons (n = 13). Right: high-background neurons (n = 12). Each graph displays % of sample neurons responding to (open bars) or synchronized to (filled bars) stimulus repetitions. n.s., No significant auditory response (open bars) or no significant synchronization (filled bars) to any rate of repetition of the test syllable or tone. Most low-background neurons did not synchronize to or respond to repetitive syllables or tones. In contrast, most high-background neurons responded across syllable repetition rates. However, very few of these neurons synchronized to repetition rates above 10/s.

Neural Coding Strategies

To investigate coding strategies by which BLA neurons discriminate among social vocalizations, we analyzed spike discharge with Victor and Purpura's (1996, 1997) spike distance metric (SDM). The SDM characterizes the dissimilarity between spike trains. For our analysis, the SDM was used to examine the differences in spike trains elicited by vocal stimuli and then quantify the relative stimulus discrimination contributions of spike rate and spike timing across a range of temporal precision values. Analyses were performed with an 800-ms-poststimulus onset spike train time window to encompass all stimulus durations. Preliminary analyses showed that this longer analysis window was more informative than shorter analysis windows for both sequences and isolated syllables (data not shown).

The SDM analysis revealed that the amount of information carried in a pure rate code was similar for vocal sequences and isolated syllables (Fig. 12A, top). However, additional information was gained in encoding vocal sequences when spike timing was considered, with best performance in this neuronal sample achieved with an integration window of 40 ms (Fig. 12A, top). In contrast, discrimination performance was not improved by considering spike timing in response to isolated syllables. To evaluate the information added by temporal discharge patterns, the difference between the maximum amount of information for each neuron across the different temporal integration windows, Hmax, was compared to that using a rate code alone, Hrate (Fig. 12A, bottom). The median information difference across all neurons was greater than zero for both syllables and sequences, indicating a role for spike timing in coding strategies in the BLA. However, spike timing information had a greater contribution to discrimination among sequences than among syllables (Wilcoxon rank sum test, P < 0.001). These analyses suggest greater variation in the temporal patterns of responses to vocal sequences compared with isolated syllables. This analysis is consistent with the spiking patterns shown by high-background neurons in Figs. 79.

Fig. 12.

Fig. 12.

Comparison of neural coding strategies based on spike distance metric. A: discrimination among vocal stimuli. Top: information for discriminating among syllables and sequences across a range of temporal integration values. Error bars show 95% confidence intervals. Bottom: difference between information at best temporal integration value (Hmax) and information in rate code (Hrate), for both syllables and vocal sequences. Vocal sequences carried more information than syllables in the rate code and gained a significant amount of information when the timing of spikes was considered. The amount of information gained from timing information was significantly different for syllables vs. sequences (Wilcoxon rank sum test, P < 0.001). Error bars show 95% confidence intervals of the median. B: discrimination among behavioral contexts associated with vocal stimuli. Top: information for discriminating among the contexts associated with vocal stimuli (biosonar, appeasement, lower aggression, and high aggression) for syllables and sequences across different temporal integration values. Bottom: difference between context discrimination information at best temporal integration window for syllables and sequences. Context discrimination information was higher in response to vocal sequences when spike timing was considered, and spike timing information contributed more to discrimination among behavioral contexts in response to sequences than syllables (Wilcoxon rank sum test, P < 0.001).

Can neurons in the BLA discriminate among vocal stimuli based on their associated behavioral context? To address this question we ran a second SDM analysis that collapsed vocal stimuli into four categories: biosonar, appeasement, lower aggression, and high aggression (Fig. 12B). Responses to the pure tone sequence were not included. This analysis revealed that information was available to discriminate the behavioral context associated with vocal stimuli. Information was higher in response to vocal sequences when spike timing was considered, with best performance achieved with an integration window of 40 ms (Fig. 12B, top). The median information difference between Hmax and Hrate across all neurons was greater than zero for both syllables and sequences, and spike timing information had a greater contribution to discrimination among behavioral contexts in response to sequences than to syllables (Wilcoxon rank sum test, P < 0.001) (Fig. 12B, bottom). Consistent with the spiking patterns observed in high-background neurons in Figs. 79, these analyses suggest that the temporal patterning of responses to vocal sequences can be used to discriminate the behavioral context of vocal stimuli.

DISCUSSION

Since a fundamental role of the amygdala is to orchestrate emotional responses to sensory stimuli, we (Gadziola et al. 2012b; Grimsley et al. 2013; Peterson and Wenstrup 2012) and others (Naumann and Kanwal 2011; Parsana et al. 2012) have hypothesized that amygdalar neurons should respond to and encode vocal stimuli that are associated with different behavioral states or social contexts. To test this hypothesis, we examined responses of BLA neurons to natural vocal sequences and their constituent syllables. The vocal sequences used here evoke increases in HR that are scaled to the associated behavioral context, with greater increases in HR associated with aggressive calls than with appeasement calls or neutral sequences. Among BLA neurons, we report a new finding that auditory responsiveness was bimodally distributed and that the degree of selectivity to vocal stimuli is closely associated with background discharge rate. These data indicate that distinct populations of BLA neurons participate in different representations of vocal stimuli. Low-background neurons form a sparse code in which each neuron responds to one or very few vocal stimuli, and different neurons respond to different vocalizations. High-background neurons respond to most stimuli. A new finding among BLA neurons is that the temporal patterning of action potentials carries more information than a rate code alone, but this applies only to encoding of vocal sequences, not syllables. The temporal patterning also conveys information about the behavioral context of vocal sequences. These findings demonstrate that the BLA contains multiple robust representations of vocal stimuli that can provide the basis for emotional/physiological responses to these stimuli by an animal.

Several analyses of syllable and sequence responses support the conclusion that the acoustic context within which syllables are presented, i.e., their sequencing and rate, has a major impact on the response of BLA neurons. From a methodological view, this raises questions about the utility of studies, ours and others, that examined BLA responses to isolated syllables. From a functional view of the amygdala, this reinforces the concept that the sequencing and rate of acoustic stimuli can have a major impact on the amygdala's response to a sound and, by extension, on the animal's emotional response to these stimuli.

Emotional Content in Vocalizations

Our previous work shows that the HR of vocalizing bats provides a measure of internal state that is well matched to investigator classifications of social context based on behavior and vocalizations (Gadziola et al. 2012a). In the present study, we find that the HR of a listener is differentially modulated by vocalizations from different behavioral contexts, with aggressive vocalizations evoking greater elevations than appeasement or neutral sequences. Each of these findings is consistent with the general understanding that vocalizations represent the internal state of a vocalizing animal and that they may alter the internal state of the listening animal. While not novel, our present result is important in demonstrating a link between vocalizations and emotional state in our animal model.

The novel feature of these results is that we observed indication of a mismatch between the states of the sender and the receiver. Thus we found that lower-aggression sequences evoke the greatest increase in HR, even compared with high-aggression sequences. Multiple lines of evidence (Gadziola et al. 2012a) support our original classification of the relationship between vocalizations and the sender's internal state. What explains the mismatch?

One explanation is partly methodological. We employed three randomly presented sequences for each social context, and it is possible that the chosen high-aggression sequences do not all have the same impact on the emotional state of a listener. In fact, our neurophysiological results reveal differences in amygdalar responses between the two tested high-aggression sequences. Future experiments will test whether these sequences elicit distinct HR/behavioral responses. This may reveal subtler categories of highly aggressive vocal interactions than our initial studies revealed.

A second explanation may relate to a real mismatch between a sender's “intent” and a listener's interpretation. In the present case, a potential explanation for the mismatch relates to the proximity of the listener. When a vocalizing bat produces high-aggression calls, the listening bat is not typically in close proximity, resulting in a decreased probability of biting behavior relative to lower-aggression contexts (Gadziola et al. 2012a). Thus one might predict that high-aggression vocalizations may actually be less emotionally arousing to a listening bat that is out of harm's way.

Subpopulations of Amygdalar Neurons

Our experimental approach included an attempt to record from neurons of all background rates and the use of an extended prestimulus recording period to assess this rate. As a result, we recorded many more neurons with low background discharge compared with our previous study (Gadziola et al. 2012b). With few exceptions, these low-background neurons were either unresponsive to our subset of auditory stimuli or highly selective for acoustic stimuli. These response types are much more common in the present study than in our previous work. This association between background discharge and stimulus selectivity has been described in studies of other brain regions (Ison et al. 2011; Kerlin et al. 2010; Mruczek and Sheinberg 2012). In these studies, there was a clear linkage between these features and cell type: projection neurons have low background discharge and high stimulus selectivity, while interneurons have high background discharge and low stimulus selectivity. Does this hold for auditory responsive neurons in the BLA?

The BLA contains two main cell types. Glutamatergic, pyramidal-like projection neurons make up the large majority (>90%) of BLA neurons, with the remaining identified as local-circuit inhibitory neurons (Rainnie et al. 1993; Sah et al. 2003; Washburn and Moises 1992). Background discharge rate is a useful but not definitive feature in characterizing these two cell types. Neurons with rates below 1 spike/s are generally considered to be projection neurons (Likhtik et al. 2006; Paré and Gaudreau 1996; Rosenkranz and Grace 1999, 2001). However, other projection neurons have background rates as high as 7 spikes/s, with a substantial number in the 2–5 spikes/s range (Likhtik et al. 2006). Likhtik and colleagues proposed that neurons above 7 spikes/s are almost certainly interneurons, but neurons with rates in the 2–7 spikes/s range could not be authoritatively classified. They further showed that spike width is not a reliable identifier of the two cell types in the BLA, unlike other brain regions (Domich et al. 1986). The differences in background discharge may relate to underlying differences in resting membrane potential, synaptic inhibition, and voltage-gated conductances that have been characterized in BLA neurons (Paré et al. 1995; Rainnie et al. 1993; Sah et al. 2003; Washburn and Moises 1992).

Except for the inclusion in our study of four neurons with rates above 7 spikes/s, our distribution of background rates from awake big brown bats corresponds remarkably well to the rates of projection neurons identified in isoflurane-anesthetized cats (Likhtik et al. 2006). This correspondence, and the established literature described above, strongly suggests that the neurons we categorized as “low background” are projection neurons. Also based on this correspondence, we conclude that the “high-background” group in our study likely includes a substantial number of projection neurons as well as interneurons. This conclusion is supported by the predominance of projection neurons in the BLA (Rainnie et al. 1993; Sah et al. 2003; Washburn and Moises 1992). In our study as in that of Likhtik and colleagues, there was no clear relationship between background discharge, spike width, and, by inference, cell type.

Our results clearly show a close relationship between background discharge rate and auditory stimulus responsiveness/selectivity. Neurons with background discharge in the 0–1 spike/s range generally responded to 0–1 syllables and 0–1 vocal sequences. Neurons with background discharge only slightly higher, in the 2–5 spikes/s range, generally responded to at least 75% of syllables and vocal sequences. The new finding for the BLA is that there are two forms of representation of vocal stimuli that likely characterize BLA outputs. Some previous studies of BLA responses to vocal stimuli appear to record from subsets of these neural populations. Our previous study in big brown bats (Gadziola et al. 2012b) recorded mostly from high-background neurons with average background firing of ∼5 spikes/s and found that nearly all of these neurons were highly responsive to vocal stimuli. Parsana et al. (2012) reported similar background firing rates of ∼5 spikes/s in their freely moving rats. Such neurons may be a high-background subset of all rat BLA neurons as recorded by Bordi and coworkers under similar conditions (Bordi et al. 1993). It is not clear how selective these neurons were to a full range of social vocalizations. In the mustached bat, two studies report similar background discharge rates (∼3.7 spikes/s) that appear to include both low- and high-background neurons as defined in our study (Naumann and Kanwal 2011; Peterson and Wenstrup 2012). Both mustached bat studies show population selectivity for syllables that is overall higher than that demonstrated in the big brown bat. While species may vary in the use of representation strategies involving high-selectivity vs. high-responsiveness types, our data suggest that a consideration of the background discharge rates among the sample of neurons may be crucial for understanding the nature of the vocal representation.

Neural Coding Strategies

There are likely multiple neural circuits within the BLA, responsible for coding different aspects of behaviorally relevant stimuli. Other groups have found strong evidence for valence-coding neurons (Belova et al. 2007, 2008; Paton et al. 2006), intensity-coding neurons (Anderson and Sobel 2003; Small et al. 2003), and even fear extinction-resistant neurons (An et al. 2012). The present study suggests that neurons within the BLA employ a variety of neural strategies for encoding social vocalizations.

Representation of vocal stimuli by low-background neurons.

In the big brown bat BLA, presumptive projection neurons with low background discharge (≤1 spike/s) provide a distinct representation of vocal stimuli. These neurons display a robust response to one or a very few stimuli. The response is usually phasic and does not show synchronous firing to the individual syllables within a vocal sequence or to a train of acoustic stimuli. The sampled population responded to each of the syllables and sequences in roughly equal proportion. These observations suggest a sparse code of vocalizations, in which the firing of an individual neuron indicates the occurrence of a particular vocalization and the population as a whole may efficiently encode a range of vocalizations (Olshausen and Field 2004). There is no clear indication among this sample of a preference for vocalizations with negative or positive affect, but the sample size may be too small to determine this.

Similar to the low-background neurons in this study, the majority of BLA neurons in the mustached bat are highly selective, preferentially responding to only a few of the tested syllables (Naumann and Kanwal 2011; Peterson and Wenstrup 2012). However, the mustached bat sample included both high- and low-background neurons (data from Peterson and Wenstrup 2012). Furthermore, although both antagonistic and affiliative calls evoked excitation, there was a predominance of excitatory responses to antagonistic signals, especially to a broadband noisy signal, similar to the rBNBl (rectangular broadband noise burst long), used in aggressive interactions. Thus the predominant form of representation, at least to syllables, was a sparse code in which vocal stimuli are represented by a population of selective neurons that use a rate code to discriminate among vocal stimuli.

Representation of vocal stimuli by high-background neurons.

The present study finds that both isolated syllables and vocal sequences evoke stimulus-dependent responses across a population of high-background neurons. In big brown bats, this population forms a second, distinct representation of vocal stimuli that is based on broadly responsive neurons that discriminate among stimuli on the basis of response magnitude and timing. Gadziola et al. (2012b) demonstrated the importance of response duration in discriminating among syllables in the vocal repertoire, and the present study further supports the importance of spike timing to the neural code. Spike timing becomes particularly informative when neurons respond to natural vocal sequences.

We find that neurons may employ different coding strategies depending on the complexity of acoustic signals. We relied on a spike distance metric to compare the relative performance of rate and temporal codes within the population of high-background neurons when responding to syllables or vocal sequences. We focused our analysis on the high-background neurons, as these broadly responsive neurons had the potential to discriminate among the stimulus set. These metrics are not suitable for neurons with a high degree of selectivity for stimuli or for neurons that discharge no, or very few, spikes in response to many of the tested stimuli. Among the high-background population, we find that neural responses are able to transmit more information about vocal sequences than syllables and that vocal sequences show significant gains in information when temporal information is considered. From work in other sensory systems, it has been suggested that pure rate codes are sufficient at signaling gross sensory differences but spike timing is important for signaling finer differences (Samonds and Bonds 2004; Victor 2005). Thus spike timing may be particularly important for coding complex acoustic sequences (Huetz et al. 2009; Narayan et al. 2006; Schnupp et al. 2006) and for signaling more subtle differences between the intensity of a behavioral state (i.e., lower and high levels of aggression). Furthermore, as natural vocal sequences often contain several rapidly repeated elements, different neural strategies may be adopted depending on the repetition rate (Wang et al. 2008b). The observed enhancement in discriminating social context provided by vocal sequences is carried, at least in part, by the temporal patterning of spikes. These findings contribute significantly to our understanding of valence coding in the amygdala. By highlighting the importance of spike timing to the neural code, our results provide a potential explanation for why some human imaging studies have failed to detect valence coding within the amygdala (Anderson and Sobel 2003).

Although our spike timing analyses in the present and previous study (Gadziola et al. 2012b) demonstrate the presence of valence coding by BLA neurons, the recorded population of high-background BLA neurons reveals a striking inconsistency in responses to the two examples of vocal sequences that are frequently observed during high-aggression interactions (Gadziola et al. 2012a). These two sequences differ acoustically from one another in that the “High Aggression 1” sequence included many low-frequency, broadband noise syllable types (e.g., rBNBs, rBNBl) and evoked robust responses in the amygdala, whereas the “High Aggression 2” sequence contained frequency-modulated syllables [e.g., sAFM (single-arched frequency modulation), sHFM (single-humped frequency modulation), UFM] and was relatively ineffective at eliciting a strong population response. Although these frequency-modulated syllable types evoke strong population responses when presented alone, responses are suppressed when they are combined within natural sequences. What underlies this apparent discrepancy?

Although both example sequences are commonly observed during behavioral interactions classified as high aggression, there may have been a more subtle distinction that we were unable to differentiate in an earlier behavioral study (Gadziola et al. 2012a) based on overt behavioral displays. While classification labels of “aggression” help specify the type of negative affect associated with the vocalization, there are several distinct forms of aggression (e.g., conspecific, defensive, predatory). Thus there is not likely to be a unitary state with a single neural representation, but rather multiple forms depending on the underlying survival circuit activated (LeDoux 2012; Moyer 1976). Moreover, within an aggressive social interaction, it is possible that the same vocalizing bat may oscillate between different forms of aggression, or even between different emotional states. This differential responsiveness may explain why high-aggression vocalizations evoked less change in HR compared with lower-aggression signals: if increased HR reflects activity within the amygdala, then the High Aggression 2 sequence would not be expected to evoke as large a change in HR as the High Aggression 1 sequence.

Challenges in the Study of Auditory Responses of BLA Neurons

In our evaluation of the role of the amygdala in responses to natural acoustic stimuli, it has become clear how much of the conclusion depends on a broad set of experimental variables. Thus our observation of two forms of representation of acoustic stimuli depended on the recording of neurons with a broad range of background firing rates, including those neurons that are difficult to detect with standard approaches.

A second consideration includes the role of species differences. There are differences in the representation of vocal stimuli in different species of bats and further differences in representation in rats and mice. For instance, rat BLA neurons are primarily excited by antagonistic vocalizations, while many big brown bat neurons respond across the affective range. Species effects may include different forms of representation of vocal stimuli but may also include species-typical responses to restraint, a contextual feature that alters BLA neuronal function (Kavushansky and Richter-Levin 2006; Roozendaal et al. 2006; Shekhar et al. 2005), including responsiveness to sounds (Grimsley et al. 2015).

While there has been clear evidence that other sensory context and internal state affects amygdalar responses to sensory stimuli (Grimsley et al. 2013; Kuraoka and Nakamura 2007; LeDoux 2000, 2007; Yan and Scott 1996; Yasoshima et al. 1995), previous work and the present study indicate additional effects related to acoustic stimuli. As syllable ordering and temporal emission patterns provide the necessary acoustic cues for distinguishing the affective context of vocalizations (Arnold and Zuberbuhler 2006; Banse and Scherer 1996; Bastian and Schmidt 2008; Bohn et al. 2008; Clucas et al. 2004; Coutinho and Dibben 2013; Fichtel and Hammerschmidt 2002; Fischer et al. 2002; Gadziola et al. 2012a; Rendall 2003), we hypothesized that amygdalar neurons would show better discrimination of vocalizations, and of the social context, when presented with natural vocal sequences rather than isolated syllables. The results support this hypothesis, showing that the important role of spike timing in neuronal discrimination of vocal stimuli and of valence is only revealed in testing based on vocal sequences. Our study further shows that BLA neurons do not generally display synchrony to the elements of acoustic trains presented at high rates. Vocal sequences that include high syllable rates elicit either onset responses or temporal summation effects, e.g., for lower-aggression sequences. Finally, the response to syllables is not highly informative of the response to sequences that contain those syllables, suggesting that combinatorial interactions may play important roles. Even here, however, there may be species effects. The study by Parsana et al. (2012) suggests that there is relatively little difference in responsiveness or valence coding by BLA neurons when responding to negative-valence 22-kHz vocal sequences vs. 22-kHz continual tones. We have no doubt that the high-background BLA neurons in the big brown bat would distinguish between these stimuli.

Other stimulus presentation parameters, such as stimulus repetition rate or randomization, will likely alter the meaning of the stimulus to a listener and modulate amygdalar responsiveness. Thus comparisons across studies may prove difficult if stimulus presentation parameters are not the same. For example, in the mustached bat amygdala, Peterson and Wenstrup (2012) find that amygdalar neurons are less responsive to artificial stimuli and more selective for vocalizations and display a greater degree of persistent activity compared with the results of Naumann and Kanwal (2011). These differences are likely attributable to methodological differences in stimulus repetition rates, as well as stimulus ordering (e.g., repeated presentation vs. pseudorandomized). Future studies should focus on the specific effects of repetition rate and stimulus ordering procedures to better understand the influence of the acoustic context on amygdalar responses.

GRANTS

This study was supported by National Institute on Deafness and Other Communication Disorders Grant R01 DC-00937-23 (J. J. Wenstrup).

DISCLOSURES

No conflicts of interest, financial or otherwise, are declared by the author(s).

AUTHOR CONTRIBUTIONS

Author contributions: M.A.G., S.J.S., and J.J.W. conception and design of research; M.A.G. performed experiments; M.A.G., S.J.S., and J.J.W. analyzed data; M.A.G., S.J.S., and J.J.W. interpreted results of experiments; M.A.G., S.J.S., and J.J.W. prepared figures; M.A.G., S.J.S., and J.J.W. drafted manuscript; M.A.G., S.J.S., and J.J.W. edited and revised manuscript; M.A.G., S.J.S., and J.J.W. approved final version of manuscript.

ACKNOWLEDGMENTS

We thank Kelsey Gerbig for assisting with electrocardiogram recordings, Carol Grose for technical assistance, and Jasmine Grimsley and the Auditory Neuroscience Group at Northeast Ohio Medical University (NEOMED) for discussion of data. We are grateful to the Ohio Department of Natural Resources for permission to collect bats.

Present address of M. A. Gadziola: Dept. of Neurosciences, School of Medicine, 2109 Adelbert Ave., Case Western Reserve University, Cleveland, OH 44106-4975.

REFERENCES

  1. Amaral DG, Price JL. Amygdalo-cortical projections in the monkey (Macaca fascicularis). J Comp Neurol 230: 465–496, 1984. [DOI] [PubMed] [Google Scholar]
  2. An B, Hong I, Choi S. Long-term neural correlates of reversible fear learning in the lateral amygdala. J Neurosci 32: 16845–16856, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Anderson AK, Sobel N. Dissociating intensity from valence as sensory inputs to emotion. Neuron 39: 581–583, 2003. [DOI] [PubMed] [Google Scholar]
  4. Arnold K, Zuberbuhler K. Language evolution: semantic combinations in primate calls. Nature 441: 303, 2006. [DOI] [PubMed] [Google Scholar]
  5. Arvola A. Vocalization in the guinea-pig, Cavia porcellus l. Ann Zool Fennici 11: 1–96, 1974. [Google Scholar]
  6. Banse R, Scherer KR. Acoustic profiles in vocal emotion expression. J Pers Soc Psychol 70: 614–636, 1996. [DOI] [PubMed] [Google Scholar]
  7. Bastian A, Schmidt S. Affect cues in vocalizations of the bat, Megaderma lyra, during agonistic interactions. J Acoust Soc Am 124: 598–608, 2008. [DOI] [PubMed] [Google Scholar]
  8. Belova MA, Paton JJ, Morrison SE, Salzman CD. Expectation modulates neural responses to pleasant and aversive stimuli in primate amygdala. Neuron 55: 970–984, 2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Belova MA, Paton JJ, Salzman CD. Moment-to-moment tracking of state value in the amygdala. J Neurosci 28: 10023–10030, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Bohn KM, Schmidt-French B, Ma ST, Pollak GD. Syllable acoustics, temporal patterns, and call composition vary with behavioral context in Mexican free-tailed bats. J Acoust Soc Am 124: 1838–1848, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Bordi F, LeDoux J, Clugnet MC, Pavlides C. Single-unit activity in the lateral nucleus of the amygdala and overlying areas of the striatum in freely behaving rats: rates, discharge patterns, and responses to acoustic stimuli. Behav Neurosci 107: 757–759, 1993. [DOI] [PubMed] [Google Scholar]
  12. Boughman JW. Vocal learning by greater spear-nosed bats. Proc Biol Sci 265: 227–233, 1998. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Brunet-Rossinni AK, Austad SN. Ageing studies on bats: a review. Biogerontology 5: 211–222, 2004. [DOI] [PubMed] [Google Scholar]
  14. Cardinal RN, Parkinson JA, Hall J, Everitt BJ. Emotion and motivation: the role of the amygdala, ventral striatum, and prefrontal cortex. Neurosci Biobehav Rev 26: 321–352, 2002. [DOI] [PubMed] [Google Scholar]
  15. Clucas BA, Freeberg TM, Lucas JR. Chick-a-dee call syntax, social context, and season affect vocal responses of Carolina chickadees (Poecile carolinensis). Behav Ecol Sociobiol 57: 187–196, 2004. [Google Scholar]
  16. Coutinho E, Dibben N. Psychoacoustic cues to emotion in speech prosody and music. Cogn Emot 27: 658–684, 2013. [DOI] [PubMed] [Google Scholar]
  17. Domich L, Oakson G, Steriade M. Thalamic burst patterns in the naturally sleeping cat: a comparison between cortically projecting and reticularis neurones. J Physiol 379: 429–449, 1986. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Esser KH. Audio-vocal learning in a non-human mammal: the lesser spear-nosed bat Phyllostomus discolor. Neuroreport 5: 1718–1720, 1994. [DOI] [PubMed] [Google Scholar]
  19. Fichtel C, Hammerschmidt K. Responses of redfronted lemurs to experimentally modified alarm calls: evidence for urgency-based changes in call structure. Ethology 108: 763–777, 2002. [Google Scholar]
  20. Fischer J, Hammerschmidt K, Cheney DL, Seyfarth RM. Acoustic features of male baboon loud calls: influences of context, age, and individuality. J Acoust Soc Am 111: 1465–1474, 2002. [DOI] [PubMed] [Google Scholar]
  21. Gadziola MA, Grimsley JM, Faure PA, Wenstrup JJ. Social vocalizations of big brown bats vary with behavioral context. PLoS One 7: e44550, 2012a. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Gadziola MA, Grimsley JM, Shanbhag SJ, Wenstrup JJ. A novel coding mechanism for social vocalizations in the lateral amygdala. J Neurophysiol 107: 1047–1057, 2012b. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Gould E, Weinstein SA. Studies of maternal-infant communication and development of vocalizations in the bats Myotis and Eptesicus. In: Communications in Behavioral Biology. New York: Academic, 1971, p. 263–313. [Google Scholar]
  24. Gould E, Woolf NK, Turner DC. Double-note communication calls in bats: occurrence in three families. J Mammal 54: 998–1001, 1973. [PubMed] [Google Scholar]
  25. Grimsley JM, Hazlett EG, Wenstrup JJ. Coding the meaning of sounds: contextual modulation of auditory responses in the basolateral amygdala. J Neurosci 33: 17538–17548, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Grimsley JM, Hazlett E, Vallabh N, Sheth S, Latsko M, Jasnow AM, Wenstrup JJ. Brief restraint in mice alters stress levels, vocal behavior and the neural representation of sound in the basolateral amygdala (Abstract). Neuroscience Meeting Planner 2015: 508, 2015. [Google Scholar]
  27. Huetz C, Philibert B, Edeline JM. A spike-timing code for discriminating conspecific vocalizations in the thalamocortical system of anesthetized and awake guinea pigs. J Neurosci 29: 334–350, 2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Ison MJ, Mormann F, Cerf M, Koch C, Fried I, Quiroga RQ. Selectivity of pyramidal cells and interneurons in the human medial temporal lobe. J Neurophysiol 106: 1713–1721, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Janssen S, Schmidt S. Evidence for a perception of prosodic cues in bat communication: contact call classification by Megaderma lyra. J Comp Physiol A Neuroethol Sens Neural Behav Physiol 195: 663–672, 2009. [DOI] [PubMed] [Google Scholar]
  30. Johnstone T, van Reekum CM, Oakes TR, Davidson RJ. The voice of emotion: an FMRI study of neural responses to angry and happy vocal expressions. Soc Cogn Affect Neurosci 1: 242–249, 2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Kavushansky A, Richter-Levin G. Effects of stress and corticosterone on activity and plasticity in the amygdala. J Neurosci Res 84: 1580–1587, 2006. [DOI] [PubMed] [Google Scholar]
  32. Kerlin AM, Andermann ML, Berezovskii VK, Reid RC. Broadly tuned response properties of diverse inhibitory neuron subtypes in mouse visual cortex. Neuron 67: 858–871, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Knörnschild M. Vocal production learning in bats. Curr Opin Neurobiol 28: 80–85, 2014. [DOI] [PubMed] [Google Scholar]
  34. Knörnschild M, Nagy M, Metz M, Mayer F, von Helversen O. Complex vocal imitation during ontogeny in a bat. Biol Lett 6: 156–159, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Kuraoka K, Nakamura K. Responses of single neurons in monkey amygdala to facial and vocal emotions. J Neurophysiol 97: 1379–1387, 2007. [DOI] [PubMed] [Google Scholar]
  36. Kurta A, Baker RH. Eptesicus fuscus. Mamm Species 356: 1–10, 1990. [Google Scholar]
  37. LeDoux J. The amygdala. Curr Biol 17: R868–R874, 2007. [DOI] [PubMed] [Google Scholar]
  38. LeDoux J. Rethinking the emotional brain. Neuron 73: 653–676, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. LeDoux JE. Emotion circuits in the brain. Annu Rev Neurosci 23: 155–184, 2000. [DOI] [PubMed] [Google Scholar]
  40. LeDoux JE, Farb CR, Romanski LM. Overlapping projections to the amygdala and striatum from auditory processing areas of the thalamus and cortex. Neurosci Lett 134: 139–144, 1991. [DOI] [PubMed] [Google Scholar]
  41. Leger DW, Owings DH. Responses to alarm calls by California ground squirrels: effects of call structure and maternal status. Behav Ecol Sociobiol 3: 177–186, 1977. [Google Scholar]
  42. Likhtik E, Pelletier JG, Popescu AT, Paré D. Identification of basolateral amygdala projection cells and interneurons using extracellular recordings. J Neurophysiol 96: 3257–3265, 2006. [DOI] [PubMed] [Google Scholar]
  43. Mardia K, Jupp P. Directional Statistics. New York: Wiley, 2000. [Google Scholar]
  44. Marsh RA, Fuzessery ZM, Grose CD, Wenstrup JJ. Projection to the inferior colliculus from the basal nucleus of the amygdala. J Neurosci 22: 10449–10460, 2002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Mascagni F, McDonald AJ, Coleman JR. Corticoamygdaloid and corticocortical projections of the rat temporal cortex: a Phaseolus vulgaris leucoagglutinin study. Neuroscience 57: 697–715, 1993. [DOI] [PubMed] [Google Scholar]
  46. Mateo JM. Alarm calls elicit predator-specific physiological responses. Biol Lett 6: 623–625, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Monroy JA, Carter ME, Miller KE, Covey E. Development of echolocation and communication vocalizations in the big brown bat, Eptesicus fuscus. J Comp Physiol A Neuroethol Sens Neural Behav Physiol 197: 459–467, 2011. [DOI] [PubMed] [Google Scholar]
  48. Morris JS, Scott SK, Dolan RJ. Saying it with feeling: neural responses to emotional vocalizations. Neuropsychologia 37: 1155–1163, 1999. [DOI] [PubMed] [Google Scholar]
  49. Moyer K. The Psychobiology of Aggression. New York: Harper & Row, 1976. [Google Scholar]
  50. Mruczek RE, Sheinberg DL. Stimulus selectivity and response latency in putative inhibitory and excitatory neurons of the primate inferior temporal cortex. J Neurophysiol 108: 2725–2736, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Narayan R, Graña G, Sen K. Distinct time scales in cortical discrimination of natural sounds in songbirds. J Neurophysiol 96: 252–258, 2006. [DOI] [PubMed] [Google Scholar]
  52. Naumann RT, Kanwal JS. Basolateral amygdala responds robustly to social calls: spiking characteristics of single unit activity. J Neurophysiol 105: 2389–2404, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Olshausen BA, Field DJ. Sparse coding of sensory inputs. Curr Opin Neurobiol 14: 481–487, 2004. [DOI] [PubMed] [Google Scholar]
  54. Paradiso JL, Greenhall AM. Longevity records for American bats. Am Midl Nat 78: 251–252, 1967. [Google Scholar]
  55. Paré D, Gaudreau H. Projection cells and interneurons of the lateral and basolateral amygdala: distinct firing patterns and differential relation to theta and delta rhythms in conscious cats. J Neurosci 16: 3334–3350, 1996. [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Paré D, Pape HC, Dong J. Bursting and oscillating neurons of the cat basolateral amygdaloid complex in vivo: electrophysiological properties and morphological features. J Neurophysiol 74: 1179–1191, 1995. [DOI] [PubMed] [Google Scholar]
  57. Parsana AJ, Li N, Brown TH. Positive and negative ultrasonic social signals elicit opposing firing patterns in rat amygdala. Behav Brain Res 226: 77–86, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Paton JJ, Belova MA, Morrison SE, Salzman CD. The primate amygdala represents the positive and negative value of visual stimuli during learning. Nature 439: 865–870, 2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Peterson DC, Wenstrup JJ. Selectivity and persistent firing responses to social vocalizations in the basolateral amygdala. Neuroscience 217: 154–171, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Pongracz P, Molnar C, Miklosi A. Acoustic parameters of dog barks carry emotional information for humans. Appl Anim Behav Sci 100: 228–240, 2006. [Google Scholar]
  61. Price JL. Comparative aspects of amygdala connectivity. Ann NY Acad Sci 985: 50–58, 2003. [DOI] [PubMed] [Google Scholar]
  62. Rainnie DG, Asprodini EK, Shinnick-Gallagher P. Intracellular recordings from morphologically identified neurons of the basolateral amygdala. J Neurophysiol 69: 1350–1362, 1993. [DOI] [PubMed] [Google Scholar]
  63. Rendall D. Acoustic correlates of caller identity and affect intensity in the vowel-like grunt vocalizations of baboons. J Acoust Soc Am 113: 3390–3402, 2003. [DOI] [PubMed] [Google Scholar]
  64. Roozendaal B, Hui GK, Hui IR, Berlau DJ, McGaugh JL, Weinberger NM. Basolateral amygdala noradrenergic activity mediates corticosterone-induced enhancement of auditory fear conditioning. Neurobiol Learn Mem 86: 249–255, 2006. [DOI] [PubMed] [Google Scholar]
  65. Rosenkranz JA, Grace AA. Modulation of basolateral amygdala neuronal firing and afferent drive by dopamine receptor activation in vivo. J Neurosci 19: 11027–11039, 1999. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Rosenkranz JA, Grace AA. Dopamine attenuates prefrontal cortical suppression of sensory inputs to the basolateral amygdala of rats. J Neurosci 21: 4090–4103, 2001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Russ JM, Jones G, Racey PA. Responses of soprano pipistrelles, Pipistrellus pygmaeus, to their experimentally modified distress calls. Anim Behav 70: 397–404, 2005. [Google Scholar]
  68. Sah P, Faber ES, Lopez De Armentia M, Power J. The amygdaloid complex: anatomy and physiology. Physiol Rev 83: 803–834, 2003. [DOI] [PubMed] [Google Scholar]
  69. Samonds JM, Bonds AB. From another angle: differences in cortical coding between fine and coarse discrimination of orientation. J Neurophysiol 91: 1193–1202, 2004. [DOI] [PubMed] [Google Scholar]
  70. Sander K, Scheich H. Auditory perception of laughing and crying activates human amygdala regardless of attentional state. Cogn Brain Res 12: 181–198, 2001. [DOI] [PubMed] [Google Scholar]
  71. Schehka S, Esser KH, Zimmermann E. Acoustical expression of arousal in conflict situations in tree shrews (Tupaia belangeri). J Comp Physiol A Neuroethol Sens Neural Behav Physiol 193: 845–852, 2007. [DOI] [PubMed] [Google Scholar]
  72. Schnupp JW, Hall TM, Kokelaar RF, Ahmed B. Plasticity of temporal pattern codes for vocalization stimuli in primary auditory cortex. J Neurosci 26: 4785–4795, 2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Shekhar A, Truitt W, Rainnie D, Sajdyk T. Role of stress, corticotrophin releasing factor (CRF) and amygdala plasticity in chronic anxiety. Stress 8: 209–219, 2005. [DOI] [PubMed] [Google Scholar]
  74. Small DM, Gregory MD, Mak YE, Gitelman D, Mesulam MM, Parrish T. Dissociation of neural representation of intensity and affective valuation in human gustation. Neuron 39: 701–711, 2003. [DOI] [PubMed] [Google Scholar]
  75. Soltis J, Leighty KA, Wesolek CM, Savage A. The expression of affect in African elephant (Loxodonta africana) rumble vocalizations. J Comp Psychol 123: 222–225, 2009. [DOI] [PubMed] [Google Scholar]
  76. Victor JD. Spike train metrics. Curr Opin Neurobiol 15: 585–592, 2005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Victor JD, Purpura KP. Nature and precision of temporal coding in visual cortex: a metric-space analysis. J Neurophysiol 76: 1310–1326, 1996. [DOI] [PubMed] [Google Scholar]
  78. Victor JD, Purpura KP. Sensory coding in cortical neurons. Ann NY Acad Sci 835: 330–352, 1997. [DOI] [PubMed] [Google Scholar]
  79. Wang H, Liang S, Burgdorf J, Wess J, Yeomans J. Ultrasonic vocalizations induced by sex and amphetamine in M2, M4, M5 muscarinic and D2 dopamine receptor knockout mice. PLoS One 3: e1893, 2008a. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Wang X, Lu T, Bendor D, Bartlett E. Neural coding of temporal information in auditory thalamus and cortex. Neuroscience 157: 484–494, 2008b. [DOI] [PubMed] [Google Scholar]
  81. Washburn MS, Moises HC. Electrophysiological and morphological properties of rat basolateral amygdaloid neurons in vitro. J Neurosci 12: 4066–4079, 1992. [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Willis CK, Brigham RM. Roost switching, roost sharing and social cohesion: forest-dwelling big brown bats, Eptesicus fuscus, conform to the fission-fusion model. Anim Behav 68: 495–505, 2004. [Google Scholar]
  83. Willis CK, Kolar KA, Karst AL, Kalcounis-Rueppell MC, Brigham RM. Medium- and long-term reuse of trembling aspen cavities as roosts by big brown bats (Eptesicus fuscus). Acta Chiropt 5: 85–90, 2003. [Google Scholar]
  84. Wright GS, Chiu C, Xian W, Wilkinson GS, Moss CF. Social calls predict foraging success in big brown bats. Curr Biol 24: 885–889, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Yan J, Scott TR. The effect of satiety on responses of gustatory neurons in the amygdala of alert cynomolgus macaques. Brain Res 740: 193–200, 1996. [DOI] [PubMed] [Google Scholar]
  86. Yasoshima Y, Shimura T, Yamamoto T. Single unit responses of the amygdala after conditioned taste aversion in conscious rats. Neuroreport 6: 2424–2428, 1995. [DOI] [PubMed] [Google Scholar]
  87. Yavuzoglu A, Schofield BR, Wenstrup JJ. Substrates of auditory frequency integration in a nucleus of the lateral lemniscus. Neuroscience 169: 906–919, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Yin S, McCowan B. Barking in domestic dogs: context specificity and individual identification. Anim Behav 68: 343–355, 2004. [Google Scholar]

Articles from Journal of Neurophysiology are provided here courtesy of American Physiological Society

RESOURCES