Skip to main content
eNeuro logoLink to eNeuro
. 2015 Apr 30;2(2):ENEURO.0077-14.2015. doi: 10.1523/ENEURO.0077-14.2015

Temporal Integration of Auditory Information Is Invariant to Temporal Grouping Cues1,2,3

Andrew SK Liu 1, Joji Tsunada 2, Joshua I Gold 3,*, Yale E Cohen 2,3,*,
PMCID: PMC4596088  PMID: 26464975

Auditory perception depends on perceptual grouping cues, which relate to how the brain parses the auditory scene into distinct perceptual units, and auditory decisions, which relate to how the brain identifies a sound. These two processes are not independent; both rely on the temporal structure of the acoustic stimulus.

Keywords: auditory cortex, decision-making, scene analysis, speed-accuracy tradeoff, streaming

Abstract

Auditory perception depends on the temporal structure of incoming acoustic stimuli. Here, we examined whether a temporal manipulation that affects the perceptual grouping also affects the time dependence of decisions regarding those stimuli. We designed a novel discrimination task that required human listeners to decide whether a sequence of tone bursts was increasing or decreasing in frequency. We manipulated temporal perceptual-grouping cues by changing the time interval between the tone bursts, which led to listeners hearing the sequences as a single sound for short intervals or discrete sounds for longer intervals. Despite these strong perceptual differences, this manipulation did not affect the efficiency of how auditory information was integrated over time to form a decision. Instead, the grouping manipulation affected subjects’ speed−accuracy trade-offs. These results indicate that the temporal dynamics of evidence accumulation for auditory perceptual decisions can be invariant to manipulations that affect the perceptual grouping of the evidence.

Significance Statement

Auditory perception depends on perceptual grouping cues, which relate to how the brain parses the auditory scene into distinct perceptual units, and auditory decisions, which relate to how the brain identifies a sound. These two processes are not independent; both rely on the temporal structure of the acoustic stimulus. However, the effects of this temporal structure on perceptual grouping and decision-making are not known. Here, we combined psychophysical testing with computational modeling to test the interaction of temporal perceptual grouping cues with the temporal processes that underlie perceptual decision-making. We found that temporal grouping cues do not affect the efficiency by which sensory evidence is accumulated to form a decision. Instead, the grouping cues modulate a subject’s speed−accuracy trade-off.

Introduction

Auditory perception depends on both perceptual grouping and decision-making. Perceptual grouping is a form of feature-based stimulus segmentation that determines whether acoustic events are grouped into a single sound or segregated into distinct sounds (Bregman, 1990). Auditory decision-making involves the brain’s interpretation of information within and across discrete stimuli to detect, discriminate, or identify their source or content.

Auditory perceptual grouping and decision-making each depend critically on the temporal structure of incoming acoustic events. For instance, when a person is walking, each step is a unique acoustic event, but our auditory system groups these events together to form a stream of footsteps. However, if the time between events is long, the auditory system segregates these events into unique, discrete sounds. Decision-making can also depend on the temporal structure of a sound because decision-making is a deliberative process in which listeners often accumulate and interpret auditory information over time to form categorical judgments (Green et al., 2010; Brunton et al., 2013; Mulder et al., 2013).

Although we know that perceptual grouping can affect some forms of decision-making (Bey and McAdams, 2002; Roberts et al., 2002; Micheyl and Oxenham, 2010; Borchert et al., 2011; Thompson et al., 2011), the interplay between the temporal properties of an auditory stimulus, perceptual grouping, and decision-making is not known. We cannot infer this interplay from visual studies because analogous manipulations in the visual domain (Kiani et al., 2013) do not relate directly to auditory perceptual grouping. Further, because temporal processing is fundamentally different for audition than for vision (Bregman, 1990; Griffiths and Warren, 2004; Shinn-Cunningham, 2008; Shamma et al., 2011; Bizley and Cohen, 2013), it is reasonable to hypothesize that this interplay may be different in these two sensory systems. Thus, it remains an open and fundamental question whether and how grouping temporal information interacts with auditory grouping and decision-making.

To examine this question, we measured the performance of human subjects who participated in a series of auditory tasks in which they reported whether a sequence of tone bursts was increasing or decreasing in frequency. Temporal information was manipulated by changing the interval between the onsets of consecutive tone bursts. This manipulation affected the subjects’ perceptual grouping of the tone-burst sequence: they heard one sound when the interval was short but a series of discrete sounds when it was long. The quality of the sensory evidence was manipulated by changing the proportion of tone bursts that linearly increased or decreased in frequency. We found that subjects accumulated sensory evidence over time to form their decisions. However, the time interval between consecutive tone bursts did not affect how this incoming stimulus was accumulated to form the decision about the change in frequency. Instead, the time between the tone bursts affected how the subjects balanced the speed and accuracy of their decisions, which fundamentally trade-off for certain decisions like this one that require incoming, noisy information to be accumulated over time (Gold and Shadlen, 2007; Bogacz et al., 2010). Specifically, for our task, longer time intervals between tone bursts (i.e., slower rates of incoming sensory information) led to a higher premium on speed at the expense of accuracy. Overall, these findings indicate that temporal manipulations that affect the perceptual grouping of sounds do not necessarily affect how information from those sounds are accumulated over time to form a decision, even when the temporal manipulations have clear effects on the trade-off between the speed and accuracy of the decision.

Materials and Methods

Prior to their participation, subjects provided informed consent. Human subjects were recruited at the University of Pennsylvania. All subjects (age range: 25-48 years) reported normal hearing; three of the subjects were authors on the study.

Experimental setup

All experimental sessions took place in a single-walled acoustic chamber (Industrial Acoustics) that was lined with echo-absorbing foam. Each subject was seated with his or her chin in a chin rest that was approximately 2 ft from a calibrated Yamaha (model MSP7) speaker. Auditory stimuli were generated using the RX6 digital-signal-processing platform (TDT Inc.). The task structure was controlled through the Snow Dots toolbox (http://code.google.com/p/snow-dots) that ran in the Matlab (Mathworks) programming environment. Subjects indicated their responses by pressing a button on a gamepad (Microsoft Sidewinder). Tasks instructions and feedback were presented on a LCD flat-panel monitor (Dell E171FP) that was placed above the speaker.

Auditory stimuli

The auditory stimulus was a sequence of tone bursts (duration: 30 ms with a 5 ms cos2 gate; level: 65 dB SPL). The interburst interval (IBI) was the time between the offset of one tone burst and the onset of the next tone burst (range: 10-150 ms).

At the beginning of each trial, the frequency of a sequence’s first tone burst was randomly sampled from a uniform distribution (500-3500 Hz). Next, we generated a monotonically increasing or decreasing sequence by adding or subtracting a 7.5 Hz to the previous tone-burst frequency. Finally, on a trial-by-trial basis, we perturbed the temporal order of the tone bursts. This perturbation changed the quality of the sensory evidence. In particular, for each trial, we defined a sequence’s coherence, which was the proportion of tone bursts in a sequence whose frequency value changed by a fixed increment relative to the previous tone burst. If the coherence was 100%, all of the tone bursts monotonically increased or decreased. If the coherence was <100%, the temporal order of a subset of tone bursts was randomly shuffled. For example, if the coherence was 50%, half of the tone bursts were shuffled. If the coherence was 0%, all of the tone bursts in the sequence were shuffled.

This stimulus design ensured that the frequency range of the sequences overlapped and that, on average, each sequence contained the same frequency values (but in a different temporal order). This procedure also minimized the possibility that subjects based their decisions on the specific frequency content of a sequence rather than its sequence direction.

Overall, each tone-burst sequence could be characterized by three parameters (Fig. 1): (1) direction, which indicated whether a sequence of tone bursts increased or decreased in frequency; (2) IBI; and (3) coherence. The values of each of these three parameters were determined on a trial-by-trial basis depending on the constraints of each auditory task; see Behavioral tasks, below.

Figure 1.

Figure 1

Example auditory sequences characterized in terms of coherence and different interburst intervals. A, Frequency direction (i.e., increasing) is easiest to discriminate when the stimulus is 100% coherent. As coherence gets smaller (60% in B and 0% in C), frequency direction becomes more ambiguous and, thus, can lead to more errors. In AC, we show increasing auditory sequences. Decreasing sequences are analogous but with negative coherence values. D shows a 100% coherence auditory sequence at three different IBIs: 10 ms (red), 60 ms (green), and 150 ms (blue).

Behavioral tasks

Subjects participated in three versions of the discrimination task. For all three versions, subjects reported the direction (increasing or decreasing) of the tone-burst sequence and received visual feedback, via the LCD monitor, regarding their report on every trial. The time between trials was ∼2 s and was independent of the subjects’ report.

Response-time task

For this task, subjects reported their perceptual decision at any time following sequence onset (Fig. 2A ). They were instructed to respond as quickly as possible but not to sacrifice accuracy. We tested six subjects (5 male and 1 female) in four weekly 1.5 h sessions. Each session contained four blocks of trials; a short break was provided between blocks. In each block, we varied sequence direction (increasing or decreasing), IBI (10, 60, or 150 ms), and coherence (0, 10, 25, 50, or 100%) on a trial-by-trial basis. Each combination of these sequence parameters was presented five times within a block for a total of 150 trials/block. The maximum response-time (RT) (and, hence, maximum sequence duration) was 5000 ms; a trial was aborted if a subject did not respond by the end of the sequence. The stimulus was terminated as soon as the subject reported the decision.

Figure 2.

Figure 2

Task design. A, For the RT task, subjects indicated their choice (i.e., sequence direction) any time after onset of the sequence. B, For the variable-duration task, subjects indicated their choice after sequence offset. C, For the hybrid task, subjects indicated sequence direction and whether they heard the sequence as one sound or discrete sounds after sequence offset in separate response periods. For all three tasks, subjects were provided feedback regarding frequency direction at the end of each trial (not shown).

Four subjects participated in additional sessions (four blocks of trials in each of four sessions per subject) of the RT task to test for effects of changes in the speed−accuracy trade-off. For these sessions, the IBI was held constant at 60 ms and subjects were instructed to “emphasize speed” or “emphasize accuracy” in alternating blocks. Stimulus tone-sequence direction (increasing or decreasing) and coherence (0, 10, 25, 50, or 100%) were varied on a trial-by-trial basis.

Variable-duration task

For this task, the experimenter controlled listening duration: subjects reported their direction decision following offset of the auditory sequence (Fig. 2B ). Five of the six subjects tested in the RT task (4 male and 1 female) also participated in six weekly 1.5 h sessions of the variable-duration task. On each trial, we chose the duration of the sequence by sampling from a truncated exponential distribution (rate parameter = 2000 ms for all IBIs); choosing the sequence duration from this distribution minimized the possibility that subjects could anticipate the end of the sequence (Gold and Shadlen, 2003). The upper- and lower-stimulus durations were a function of IBI: for IBI = 10 ms, min = 160, max = 1400 ms; for IBI = 60 ms, min = 360, max = 3150 ms; and for IBI = 150 ms, min = 720, max = 6300 ms. These limits were chosen so that each sequence, independent of IBI, contained 4-35 tone bursts. At the end of each sequence, a response cue was flashed on the LCD monitor and subjects had 800 ms to respond. Each session contained four blocks of trials; a short break was provided between blocks. In each block, we varied sequence direction (increasing or decreasing), IBI (10, 60, or 150 ms), and coherence (0, 10, 25, 50, or 100%) on a trial-by-trial basis. Each combination of these sequence parameters was presented five times within a block, for a total of 150 trials per block.

Hybrid task

For this task, subjects participated in a version of the variable-duration task that also required them to report, on each trial, whether they heard the sequence as one sound or as a series of discrete sounds (Fig. 2C ). Four of the six subjects tested in the RT task plus one new subject (3 male and 2 female) participated in four weekly 1.25 h sessions of the hybrid task. Sequence duration was sampled from a truncated exponential distribution and the limits of the stimulus durations were set to ensure that each sequence had 4-35 tone bursts. Subjects reported their two decisions during two separate 800 ms response periods. Prior to each sequence’s onset, a colored cue, which was presented on the LCD monitor, indicated the temporal order in which subjects were to report their decisions; this order alternated on a block-by-block basis.

For this task, we set both IBI and coherence to values that were centered on each subject’s psychophysical threshold. IBI threshold was the IBI value rated as one sound 50% of the time. Because the subjects’ 50%-IBI threshold varied on a day-by-day basis, we measured this threshold daily, prior to their participation in the hybrid task. IBI threshold was measured using a one-up/one-down adaptive procedure (Treutwein, 1995; García-Pérez, 1998). The sequence always had 16 tone bursts and used a 50% coherence stimulus.

Coherence threshold was defined to be 70.7% correct performance, which corresponds to a d′ of 0.77. Because preliminary experiments indicated that coherence threshold was constant across experimental sessions, it was measured once for each subject. Coherence threshold was calculated using a two-up/one-down adaptive procedure (Treutwein, 1995; García-Pérez, 1998). During this procedure, the sequence’s IBI was set to the IBI threshold.

For the hybrid task, each session contained four blocks of trials; a short break was provided between blocks. We varied IBI, on a trial-by-trial basis, between three different values: (1) IBI threshold minus 15 ms (20 trials per block), (2) IBI threshold (80 trials per block), and (3) IBI threshold plus 15 ms (20 trials per block). Coherence was set to each subject’s coherence threshold.

Fitting of behavioral data to sequential-sampling models

Behavioral data were fit to variants of sequential-sampling models related to the drift diffusion model (DDM) (Ratcliff et al., 2004; Smith and Ratcliff, 2004; Gold and Shadlen, 2007; Green et al., 2010; Brunton et al., 2013; Mulder et al., 2013) to quantify the effects of sequence coherence and IBI on the decision-making process. These models describe the process of converting incoming sensory evidence, which is represented in the brain as the noisy spiking activity of populations of relevant sensory neurons, into a decision variable that can guide behavior.

RT task

For the RT task, a key benefit of these sequential-sampling models is that they make quantitative predictions about both choice and RT as a function of the coherence of the auditory sequence. In other words, these models simultaneously fit: (1) the psychometric function, which describes accuracy versus sequence coherence; and (2) the chronometric function, which describes RT versus sequence coherence. We used several model variants.

Model variant #1

The first variant was a standard, symmetric DDM in which a perceptual decision is based on an accumulation over time of noisy evidence to a fixed bound, a process that is mathematically equivalent to the one-dimensional movement of a particle undergoing Brownian motion to a boundary (Ratcliff et al., 1999; Gold and Shadlen, 2002; Shadlen et al., 2006; Eckhoff et al., 2008; Green et al., 2010; Ding and Gold, 2012; Brunton et al., 2013; Mulder et al., 2013). In brief, this version had seven free parameters: one drift rate (k) per IBI; one symmetric bound for either increasing or decreasing (+A or −A; i.e., the height of the bound was the same for both choices but with an opposite sign) choices per IBI; and a single non-decision time (TND) that accounts for sensory-processing and motor-preparation time. Drift rate governs sensitivity and is implemented in terms of the moment-by-moment sensory evidence, which has a Gaussian distribution N(µ,1) with a mean µ that scales with sequence coherence (C): µ = kC. The decision variable is the temporal accumulation of this momentary sensory evidence. A decision (i.e., the subject reports that the sequence is increasing or decreasing) occurs when this decision variable reaches a decision bound (+A or −A, respectively). The decision time is operationally defined as the time between auditory-sequence onset and the cross of either bound. RT is the sum of the decision time and the associated non-decision time. The probability that the decision variable first crosses the +A bound is e2μA1e2μAe2μA. The mean decision time is 2Aμcoth(2μA)Aμcoth(μA) for increasing choices and 2Aμcoth(2μA)Aμcoth(μA) for decreasing choices.

Model variant #2

The second variant was a leaky accumulator, in the form of a stable Ornstein−Uhlenbeck process (Busemeyer and Townsend, 1993; Bogacz et al., 2006), in which positive values of the leak term imply that a given decision is influenced most strongly by the most recent samples of sensory evidence. Because these processes do not have simple, exact analytic solutions for both the psychometric and chronometric functions, we conducted simulations to fit the data. For each fit, we simulated 40,000 trials per iteration, with the decision variable computed in 1 ms steps.

Model variant #3

The final variant was an accumulate-to-bound model with non-leaky drift and bounds that can collapse (decrease towards zero) as a function of time within a trial (Ditterich, 2006b; Drugowitsch et al., 2012; Thura et al., 2012; Hawkins et al., 2015). The dynamics of the collapsing process were governed by a two-parameter Weibull function (Hawkins et al., 2015). These fits were also obtained using simulations like the leaky-accumulator model, above.

All model fits were computed by first using a pattern-search algorithm to find suitable initial conditions (“patternsearch” in the Matlab programming environment) and then a gradient descent to find the best-fitting parameters (“fmincon” in Matlab) that minimized the negative log-likelihood of the data, given the model fits (Palmer et al., 2005). The likelihood function for subjects’ choices was modeled as binomial errors and the mean response times were modeled as Gaussian errors.

Variable-duration and hybrid tasks

For these tasks, the experimenter, not the subject, controlled listening duration. Therefore, we used models that had a different stopping rule than those used for the RT tasks. Specifically, these models assumed that the decision was made based on the sign of the accumulated evidence at the end of the stimulus presentation. These models included two basic parameters: (1) a coherence-scaling term (k), which governed the relationship between sequence coherence and the strength of the evidence to accumulate; and (2) accumulation leak (λ), which governed the efficiency of the accumulation process. We computed the probability (p) of a correct response as a function of sequence coherence (C) and listening duration (D) as: p(C,D)=0.5(1(1L)φ(kC2(eλD1)λ(eλD+1))), where φ is the normal standard cumulative distribution function. The lapse rate (L) was set to a small value (0.01) to provide better fits (Klein, 2001; Wichmann and Hill, 2001). The models were fit using Matlab’s “fmincon” function to minimize the log-likelihood of the data, given the parameters, and assuming binomial errors.

Fitting of RT task data to the LATER model

RT distributions from the RT task were also analyzed using the LATER (Linear Approach to Threshold in Ergodic Rate) model. This model assumes that RT distributions are distributed as an inverse Gaussian because they result from a process with a linear rate of rise, which are distributed across trials as a Gaussian, that trigger a movement when reaching a fixed threshold (Carpenter and Williams, 1995; Reddi et al., 2003). We used maximum-likelihood methods to fit RT distributions to a model with two free parameters that represented the mean rate-of-rise and the threshold. We tested if and how each parameter varied with IBI for each subject, direction, and stimulus coherence.

Calculating psychophysical kernels from the RT data

Finally, we calculated the subjects’ psychophysical kernels from the RT task to further support the idea that the subjects were using a process akin to bounded accumulation. The kernels were computed directly from the data by taking, from each 0% coherence trial, the mean-subtracted stimulus sequence and then computing the mean (and standard error) value of these time-dependent sequences separately for trials leading to increasing or decreasing choices. We fit these kernels to a model that sorted the 0%-coherence trials into two categories based on the sign of the slope of a linear regression of frequency versus burst number. We used a grid search to find the values of two parameters—one governing the number of tone bursts from stimulus onset that is used to compute the linear regression and a second that scales the stimulus-frequency values used to compute the kernel—that maximized the likelihood of obtaining the IBI-specific kernels measured from the data given the model.

Results

We used human psychophysics to test if and how the time course of evidence accumulation for an auditory-discrimination task was affected by a temporal manipulation that modified the perceptual grouping of the sensory evidence. The task required subjects to report whether the frequency direction of a tone-burst sequence was increasing or decreasing (Fig. 1A ). Task difficulty was manipulated by controlling the coherence of the sequence, which corresponded to the fraction of tone bursts whose frequencies increased or decreased systematically (Fig. 1A−C ).

The temporal manipulation was a change in the IBI of the stimulus sequence, which affected perceptual grouping. For short IBIs (<∼30 ms), subjects tended to report that the sequence was one sound. For medium IBIs (∼30-100 ms), subjects alternated trial-by-trial between reports that the sequence was one sound or a series of discrete sounds. For long IBIs (>∼100 ms), subjects reliably reported that the sequence was a series of discrete sounds (Fig. 3). To test how this grouping manipulation affected the temporal dynamics of the perceptual decision (i.e., the frequency direction of the tone-burst sequence), we used three versions of a discrimination task: an RT task in which subjects controlled listening duration; a variable-duration task in which the experimenter controlled listening duration; and a hybrid task that combined the variable-duration task with an explicit grouping judgment about whether the sound was one sound or a series of discrete sounds. Results from each task are presented below.

Figure 3.

Figure 3

Influence of IBI on reports of perceived grouping. Subjects reported whether they perceived the stimulus as one sound or discrete sounds. The graph shows proportion of trials in which each subject chose one sound as a function of IBI. The points indicate each subject’s performance. Each curve represents a logistic function that was fit to each subject’s reports across four sessions. The gray line indicates 50% of trials reported as “one sound”, which was IBI threshold. Colors represent different subjects.

Response time task

Both accuracy and RT depended systematically on coherence and IBI. Figure 4 summarizes the behavioral performance of all six subjects. Choices tended to be more accurate and faster when stimulus coherence was high than when it was low.

Figure 4.

Figure 4

Performance on the RT task. A, The fraction of trials in which subjects reported a sequence of tone bursts was increasing in frequency as a function of signed coherence and IBI. Positive coherence values indicate that the sequence was increasing in frequency; negative indicates decreasing. B, The mean RT (i.e., the time between sequence onset and button press) as a function of signed coherence and IBI on correct trials only (and all 0% coherence trials). C, As in B, but using signal RT (i.e., RT not including cumulative IBI). The solid curves are simultaneous fits of psychometric and chronometric data to a DDM (see Results). The psychometric data (A) only show the fit to the signal-RT data (C). In each panel, the points indicate performance data that was pooled across subjects. Colors represent different IBIs, as indicated in A.

Subjects’ choices were clearly based on the direction of frequency change and not simply the absolute frequency value of the sequence. Specifically, when we analyzed behavior relative to the beginning, mean, and ending frequency value of the sequence, performance was near chance: the median (full range) percent correct of all six subjects’ choices, when collapsed across coherences and IBIs, was 56.9% (55.3-58.0), 50.6% (50.2-51.9), and 55.4% (54.6-56.6), respectively.

We also found that the relationships between choice, RT, and coherence were modulated by IBI. Longer IBIs led to shallower psychometric functions (i.e., lower sensitivity; Fig. 4A ) and longer RTs (Fig. 4B ). When we considered the portion of the RT that only included presentation of the tone bursts and not the silent periods by subtracting out the cumulative IBIs (which we refer to as “signal RT” and which treats the cumulative IBIs, like sensory and motor processing, as part of the non-decision time on a given trial), the effects of IBI on RT were reversed: the longest signal RTs corresponded to the shortest IBIs (Fig. 4C ).

We quantified the effects of coherence and IBI on the time-dependent decision process by fitting the choice and RT data for each subject and from pooled data across subjects to several variants of models that are related to the DDM. All of the models had the same basic form. They assumed that the decision was based on the temporal accumulation of noisy evidence until reaching one of two prespecified boundaries. The signal-to-noise ratio of this decision variable was governed by a drift rate, which was proportional to stimulus coherence and, in some cases, was also subject to leaky accumulation (Busemeyer and Townsend, 1993; Usher and McClelland, 2001; Tsetsos et al., 2012). Choice was governed by the identity of the reached boundary, which, in some cases, could change as a function of time within a trial to reflect an increasing urgency to respond (Ditterich, 2006b; Drugowitsch et al., 2012; Thura et al., 2012; Hawkins et al., 2015). RT was governed by the time to reach the boundary plus extra non-decision time. The height of the boundary governed the trade-off between speed and accuracy: higher boundaries provided longer decision times and higher accuracy, whereas lower boundaries increased speed at the expense of accuracy (Gold and Shadlen, 2007). Because performance tended to be symmetric for the two choices [across all subjects and IBIs and using balanced stimulus presentations, the median (interquartile range) absolute value of the difference in the fraction of increasing vs decreasing choices was only 0.001 (0.000–0.004)], we used an unbiased model with seven parameters: one non-decision time, three parameters representing drift rate per each of the three IBI conditions, and three parameters representing the bound height per IBI.

Fits of this model to choice data and either raw or signal RTs were consistent with a decision variable that was based on the signal portion of the stimulus sequence but not the time between bursts (the IBI). Drift rate has units of change of standardized evidence per unit time. Therefore, converting from raw to signal RT affects the (linear) scaling of this value, which is governed by the duty cycle associated with the given IBI: b/(IBI + b), where b is burst duration of 30 ms. Accordingly, best-fitting values of the DDM parameters that were fit to choices and raw RT were scaled versions of those fit to choices and signal RTs, with the scale factors approximately equal to the IBI-specific duty cycles [slopes of linear regressions of subject-specific signal vs raw drift rates = 0.80 (duty cycle = 0.75), 0.43 (0.33), and 0.23 (0.17) for IBI = 10, 60, and 150 ms, respectively]. Furthermore, best-fitting signal drift rates that were rescaled and expressed in units of the change in evidence per unit of raw time (i.e., multiplied by the duty cycle) were strongly correlated with the associated, best-fitting raw drift rates across IBIs and subjects (r = 0.98, p < 0.01a; Table 1). These results imply that the IBI manipulation affected only the duty-cycle-dependent scaling of drift rates. Therefore, we used signal RTs for the model fits, which corresponded to drift rates that had the same temporal scaling and thus could be compared directly across IBI conditions.

Table 1.

Statistical table

Data structure Statistical test Power
a Normal distribution Pearson Correlation p < 0.01
b Normal distribution Likelihood-ratio test; Bonferroni-corrected for three parameters p < 0.001
c Normal distribution Likelihood-ratio test; Bonferroni-corrected for three parameters p < 0.01
d Normal distribution Likelihood-ratio test p > 0.24
e Normal distribution Likelihood-ratio test p > 0.1
f Normality not assumed Mann−Whitney test p < 0.01
g Normality not assumed Kruskal−Wallis test p < 0.001
h Normal distribution Likelihood-ratio test, Bonferroni-corrected for two parameters p < 0.01
i Normality not assumed Kruskal−Wallis test p > 0.05
j Normal distribution Likelihood-ratio test, Bonferroni-corrected for two parameters p < 0.01
k Normal distribution Likelihood-ratio test p = 0.2838
l Normality not assumed Kruskal−Wallis test p = 0.011
m Normality not assumed Kruskal−Wallis test p = 0.125
n Normal distribution Likelihood-ratio test p > 0.05
o Normality not assumed Kruskal−Wallis test p > 0.05
p Normality not assumed Mann−Whitney test p < 0.05
q Normal distribution Likelihood-ratio test, Bonferroni-corrected for two parameters p < 0.01
r Normal distribution Likelihood-ratio test, Bonferroni-corrected for two parameters p < 0.01

We found that the effects of IBI on choice and signal RT primarily reflected changes in the decision boundary but not the drift rate (Fig. 5). The height of a symmetric, fixed decision boundary (i.e., the same height for both increasing and decreasing choices) declined systematically with increasing IBI for all six subjects and for data combined across subjects (likelihood-ratio test comparing a seven-parameter model with separate values of drift rate and bound height per IBI plus a non-decision time to a five-parameter model with a single value of drift rate shared across IBIs, p < 0.001b in all cases; Bonferroni-corrected for three parameters; Fig. 5A ). In contrast, drift rate depended on IBI for only one of the six subjects (p < 0.01c; Bonferroni-corrected for three parameters) and not for the other subjects or combined data (Fig. 5B ). These model fits were not improved by adding to the model either leaky accumulation (likelihood-ratio test, p > 0.24d across subjects and for all data combined) or collapsing bounds (likelihood-ratio test, p > 0.1e for five of the six subjects and for all data combined). There also was little evidence for slow errors that can be expected in models with collapsing bounds, with only eight of 216 conditions separated by subject/coherence/IBI showing such an effect (Mann−Whitney test comparing median correct vs error RTs, p < 0.01; Ditterich, 2006a)f. Likewise, there was little evidence for fast errors that can be expected in models with variable bounds, with only three conditions showing such an effect (Ratcliff and Rouder, 1998). Thus, changes in IBI, which affected perceptual grouping and the rate of arrival of decision-relevant signals, caused systematic, robust changes in the speed−accuracy trade-off governed by a fixed, time-independent bound. In contrast, the changes in IBI did not cause systematic changes in the efficiency with which sensory evidence was accumulated over time to form the decision.

Figure 5.

Figure 5

Parameter values from fits of the basic DDM to the RT-task data. Each panel shows best-fitting values of bound height (A) or drift rate (B) plotted as a function of IBI for fits to data from individual subjects (black) or combined across all subjects (red). Dark lines/symbols indicate that the model fits were improved significantly by fitting the given parameter separately for each IBI condition (likelihood-ratio test, p < 0.01, Bonferroni-corrected for three parameters).

These results were supported by independent analyses of the signal-RT distributions. A useful way to assess possible changes in drift rate and/or bound height in a simple accumulate-to-bound framework is to use the LATER model (Carpenter and Williams, 1995; Reddi et al., 2003). According to this model, a decision variable rises linearly to a threshold (bound) in order to trigger a motor response. Assuming a fixed bound, but a noisy decision variable with a rate of rise that is normally distributed across trials, RT is distributed as an inverse Gaussian. This distribution can be plotted as a straight line on “reciprobit” axes (i.e., percent cumulative frequency on a probit scale vs the reciprocal of RT from 0%-coherence trials; Fig. 6A ). Horizontal shifts of these lines imply changes in the mean rate-of-rise of the decision variable, whereas swivels about a fixed point at infinite RT imply changes in the bound height (Reddi et al., 2003). When we fit the LATER model to signal-RT data separately for each subject, coherence, and IBI (correct trials only), we found that increasing IBI caused systematic decreases in the bound (Kruskal−Wallis test for H0: equal median values per IBI, across subjects and coherences, p < 0.001g; 34 of 36 individual subject−coherence pairs had a significant dependence of bound height on IBI, all of which had a lower bound for the longest vs the shortest IBI, p < 0.01h, likelihood-ratio test, Bonferroni-corrected for two parameters; Fig. 6B ). In contrast, increasing IBI did not cause a systematic change in the best-fitting mean rate-of-rise (Kruskal−Wallis test, p > 0.05i; 13 of 36 individual subject−coherence pairs had a significant dependence of rate-of-rise on IBI, of which seven showed an increasing rate-of-rise and six showed a decreasing rate-of-rise; p < 0.01j, likelihood-ratio test, Bonferroni-corrected for two parameters; Fig. 6C ).

Figure 6.

Figure 6

LATER model fits to signal RT data. A, Distributions of signal RT from 0%-coherence trials for one subject are plotted on a reciprobit plot: reciprocal RT versus percentage of cumulative frequency on a probit scale (Reddi et al., 2003), separately per IBI. Best-fitting values of the bound height (B) and mean rate-of-rise (C) of the LATER model (see Results and Materials and Methods for details) are plotted as a function of IBI for each subject and coherence (black/gray lines and data points). The data in black indicate that the model fits were improved significantly by fitting the given parameter separately for each IBI condition (likelihood-ratio test, p < 0.01, Bonferroni-corrected for two parameters). Shaded lines/symbols indicate that the model fits were not improved significantly. Red data points/lines represent the median values across all conditions.

These results were also supported by correlation analyses that related choices to the noisy auditory stimulus (Knoblauch and Maloney, 2008; Murray, 2012). We computed two kernels per IBI condition, one for increasing choices and the other for decreasing choices, from the 0%-coherence trials from all six subjects. Each kernel represented the mean, within-trial time course of the mean-subtracted, stochastic auditory sequence that led to the given choice (Fig. 7). Subjects made increasing choices when the frequency tended to increase throughout most of the trial, with average kernels that started below the within-trial mean, then increased steadily to a peak value above the within-trial mean around the time of the median RT, then reverted back towards the mean. Likewise, subjects made decreasing choices on trials in which the frequency progression of the stochastic stimulus moved in the opposite direction, starting relatively high and then decreasing for much of the trial.

Figure 7.

Figure 7

Psychophysical kernels. Kernels represent the average of the stimulus sequence (mean subtracted on each trial) presented on all 0%-coherence trials across subjects for increasing (red) and decreasing (blue) choices. A−C, Kernels computed per IBI condition, as indicated, and smoothed using a 21-sample moving mean. D, E, Kernels for the 60 ms IBI condition when subjects were told to emphasize accuracy or speed, respectively. Thick and broken lines are mean and standard error, respectively. Data are aligned relative to onset of the sequence. Asterisks indicate that the increasing and decreasing kernels were significantly different from one another for a given time bin (Mann−Whitney test, p < 0.05p, using the raw, unsmoothed kernels). Black lines indicate best-fitting simulated kernels (see text for details). Solid vertical lines indicate median RT. Dashed vertical lines indicate the end of the integration time from the best-fitting simulated kernels.

These kernels were consistent with a DDM-like decision process that had lower bounds for longer IBIs, corresponding to an increasing emphasis on speed at the expense of accuracy. Specifically, the choice selectivity of these kernels (i.e., the time bins in which increasing and decreasing kernels differed from each other), measured in terms of tone bursts within a stimulus sequence and thus consistent with the signal RT analyses described above, was longest for the shortest IBI and shortest for the longest IBI (compare asterisks in Fig. 7A−C ). We found a similar effect when using just the 60 ms IBI but providing explicit instructions to the subjects about the speed−accuracy trade-off, with relatively longer choice selectivity under an “emphasize accuracy” condition (Fig. 7D ) and relatively shorter choice selectivity under an “emphasize speed” condition (Fig. 7E ). These kernels were also qualitatively consistent with an evidence-accumulation process with little or no leak because choice selectivity was strongest at the beginning of a trial; if leak was present, it would tend to show up as less choice selectivity at the beginning of a trial.

To more quantitatively relate these kernels to the underlying decision process, we fit them to a simple, two-parameter model that assumed that choices were based on particular frequency progressions within the given stimulus. One parameter governed the time course of the relevant progression, which could range from just the first two bursts to the full sequence; i.e., simulated increasing or decreasing choices occurred when the slope of a linear regression of frequency versus burst number for the first n bursts in a sequence was >0 or <0, respectively. The second parameter scaled the contribution of each stimulus sequence to the final kernel, akin to the signal-to-noise ratio (SNR) of the internal stimulus representation. We found that the best-fitting value of the integration time decreased systematically with increasing IBI (and for the “emphasize speed” relative to the “emphasize accuracy” instruction), in each case closely matching the IBI-specific median RTs (compare solid and dashed vertical lines in Fig. 7). In contrast, the best-fitting scale factor did not differ as a function of IBI (likelihood-ratio test, p = 0.28k), implying a consistent SNR across conditions (like the IBI-independent drift rate in Fig. 5). Thus, like the DDM- and LATER-based analyses described above, these kernel analyses implied that the IBI manipulation affected the speed−accuracy trade-off on the RT task but not how information was accumulated over time to form the decision.

Variable-duration task

Because the psychometric and chronometric data from the RT task depended on each subject’s individual speed−accuracy tradeoff, it is possible that the DDM-model fits reflected complex interactions between the rate of sensory-evidence accumulation and bound heights across IBI conditions (Ratcliff and Tuerlinckx, 2002). To directly test the relationship between IBI and the rate of evidence accumulation, subjects participated in the variable-duration task. During this task, we experimentally controlled the duration of the auditory sequence and, hence, the amount of sensory evidence (Fig. 2B ). Analogous to the RT task, we analyzed performance as a function of signal time to standardize the sequence duration with respect to the rate of tone-burst presentation.

Mean performance accuracy for all five subjects improved systematically as a function of both coherence and signal time, in a manner that was qualitatively similar for all three IBI conditions (Fig. 8A−C ). For each condition, accuracy tended to reach an upper asymptote of >99% correct in <1000 ms of signal time for the highest coherences; accuracy rose steadily at longer listening times for lower coherences.

Figure 8.

Figure 8

Performance on the variable-duration task. A−C, Psychometric data are plotted as a function of listening duration for different coherences and IBIs, as indicated. Each data point reflects mean performance for all five subjects as a function of coherence and signal time (plotted in 0.2 s bins, up to 1.0 s, but fit using unbinned data). The solid curves are fits from the best-fitting model with two parameters: drift rate and accumulation leak. D, E, Best-fitting values of drift rate (D) and accumulation leak (E) plotted as a function of IBI for fits to data from individual subjects (black) or combined across all subjects (red). Dark lines and symbols indicate that the model fits were improved significantly by fitting the given parameter separately for each IBI condition (likelihood-ratio test, p < 0.01q, Bonferroni-corrected for two parameters). Shaded lines and symbols indicate that the model fits were not improved significantly.

We again quantified these effects by fitting the choice data to DDM-like models (Fig. 8D,E ). Like the DDM described in the Response time task section (above), all of these models assumed that the decision was based on the value of a decision variable, which represented the accumulation of noisy sensory evidence over time. Like the RT fits, we assumed a drift rate that scaled linearly with coherence and an accumulation process that might include a leak. However, unlike bounded diffusion in the DDM when it was applied to RT data, these models assumed that the accumulation process continued until the stimulus was turned off, at which point the decision was based on the current sign of the decision variable (Gold and Shadlen, 2003; Bogacz et al., 2006). Therefore, these models did not have parameters representing bounds or non-decision times and were fit to psychometric data only (percent correct as a function of both coherence and listening duration).

The model-fitting results suggested only modest, if any, dependence of drift rate or leak on IBI. Drift rate was independent of IBI for all five subjects and depended non-monotonically on IBI for data combined across subjects (Fig. 8D ). There was also a slight trend for the best-fitting values of drift rate to depend systematically on IBI across subjects (Kruskal−Wallis test for H0: equal median values per IBI, across subjects, p = 0.011l). Accumulation leak was not significantly affected by IBI for any of the individual subjects or the data combined across subjects (Kruskal−Wallis test for H0: equal median values per IBI, across subjects p = 0.125m; Fig. 8E ). Thus, the subjects’ decisions improved systematically as a function of the number of tone bursts but were largely independent of the time between bursts.

Hybrid task

To test directly the relationship between perceptual grouping and the decision process, subjects participated in the hybrid task, which is a variant of the variable-duration task (Fig. 2C ). In this task, on each trial, tone-burst sequence was set to each subject’s coherence threshold and IBI threshold (see Materials and Methods, above). At the end of each trial, the subject gave two sequential responses to indicate: (1) perceptual grouping (i.e., one sound or a series of discrete sounds) and (2) sequence direction (i.e., increasing or decreasing). Figure 9 shows psychometric data pooled across all of the subjects, separated into trials in which the subject reported perceiving the sequence as one sound (Fig. 9A ) or a series of discrete sounds (Fig. 9B ). Accuracy tended to increase steadily as a function of listening duration in a similar manner for the RT and variable-duration tasks.

Figure 9.

Figure 9

Performance on the hybrid task. A, B, Psychometric data are plotted as a function of listening duration for trials in which the subject reported that the sequence was one sound (A) or a series of discrete sounds (B). The solid curves are fits from the best-fitting model with two parameters: drift rate and accumulation leak. C, D, Best-fitting values of drift rate (C) and accumulation leak (D) plotted as a function of perceptual grouping for fits to data from individual subjects (black) or combined across all subjects (red). The model fits were not improved by fitting the given parameter separately for each IBI condition (likelihood-ratio test, p > 0.1r in all cases, Bonferroni-corrected for two parameters).

To quantify these effects, we fit the data to the same models as those described for the variable-duration task but applied to one sound versus discrete conditions instead of different IBIs (Fig. 9C,D ). These fits indicted that the grouping report did not have any effect on the perceptual decision. For both the drift-rate and accumulation-leak parameters, the data from individual subjects and across subjects were better fit by models that used a single parameter for all trials, as opposed to separate parameters for one sound and discrete reports (likelihood-ratio test, p > 0.05n). Moreover, when the data were fit separately for the two grouping reports, the resulting best-fitting values did not differ from each other across subjects (Kruskal−Wallis test for H0: equal median values per grouping judgment, across subjects, p > 0.05o for both drift and leak). Thus, the perceptual-grouping judgment did not appear to have a substantial effect on the accumulation efficiency of the sensory evidence.

Discussion

We examined the relationship between auditory perceptual grouping and decision-making. Our focus was on the role of time in both processes. Specifically, does the temporal manipulation of a stream of acoustic events, which affects its perceptual grouping (Bregman, 1990), affect decisions about its identity? We found that the time interval between sequentially presented tone bursts had strong effects on whether the tone bursts were perceptually grouped as a single sound or heard as discrete sounds (Fig. 3). In contrast, this manipulation did not systematically affect how subjects accumulated information to form a decision about whether frequencies of the tone bursts were increasing or decreasing (Figs. 4-9). Thus, for this task and stimulus, the temporal accumulation of sensory evidence is invariant to the temporal intervals (gaps) between pieces of evidence that affects perceptual grouping.

The effect of time gaps on the time-course of evidence accumulation has also been studied in the visual system (Kiani et al., 2013). In that study, the accumulation of visual evidence was invariant to the temporal gap between pulses of motion evidence for a visual motion direction-discrimination task. We extended those findings by demonstrating that this accumulation invariance is accompanied by a change in the speed−accuracy trade-off that accounts for the different time intervals between the tone bursts. Our results also show that similar principles may govern the temporal dynamics of auditory and visual decisions, despite differences in how the underlying sensory mechanisms process temporal information (Carr and Friedman, 1999; Schnupp and Carr, 2009; Raposo et al., 2012). Below, we discuss the relationships between auditory perceptual grouping and decision-making and then discuss potential neural bases for our findings.

Temporal dynamics of auditory perceptual grouping and decision-making

Perceptual-grouping cues can affect auditory judgments (Bey and McAdams, 2002; Roberts et al., 2002; Micheyl and Oxenham, 2010; Borchert et al., 2011; Thompson et al., 2011). For example, judgments about the timing differences between auditory stimuli are more accurate when stimuli are grouped into the same auditory stream versus when they are segregated into different auditory streams (Roberts et al., 2002). Similarly, a listener’s ability to detect a deviant tone burst improves when the tone burst is segregated (e.g., by frequency) into a separate auditory stream (Rahne and Sussman, 2009; Sussman and Steinschneider, 2009). In other situations, stream segregation enhances a listener’s ability to identify a tone sequence (Bey and McAdams, 2002).

However, despite evidence for the roles of temporal cues in both auditory grouping and decision-making, little is known about how those roles interact. We addressed this issue by building upon on the rich history of auditory psychophysics and quantitative modeling (Green et al., 1957; Green, 1960; Greenwood, 1961; Luce and Green, 1972; Green and Luce, 1973). Specifically, we applied sequential-sampling models, in particular the DDM, to assess how auditory information presented sequentially over time was used to form a decision about the direction of change of the stimulus frequency.

We used a series of complementary approaches to demonstrate that subjects’ decisions were consistent with a DDM-like process that accumulates sensory evidence over time. First, we fit choice and mean RT data from the RT task directly to several variants of the DDM, all of which effectively described the relationships between stimulus coherence (strength), IBI, and the subjects’ speed−accuracy trade-offs (Figs. 4, 5) (Green and Luce, 1973; Wickelgren, 1977; Palmer et al., 2005; Ratcliff and McKoon, 2008). Second, the subjects’ full RT distributions were also consistent with a rise-to-bound process (Fig. 6), which we fit using a simplified version of DDM models (i.e., the LATER model) that assumes that the rising process is stochastic across trials (as opposed to within trials, for the DDM) and is effective at describing RT distributions across a range of conditions (Carpenter and Williams, 1995; Reddi et al., 2003). Third, analyses of our noisy auditory stimulus indicated that, at least on average, subjects were using information that extended from the beginning of the trial until around the time of the response (Fig. 7). That is, choice selectivity was strongest at the beginning of a trial. Fourth, performance on the variable-duration task increased systematically as a function of listening duration, in a manner consistent with the evidence-accumulation process described by the DDM (Figs. 8, 9).

Our primary result from these analyses was that the accumulation process was invariant to the IBI manipulation. In particular, we found that there was no systematic leak associated with the accumulation process in any of the tested task conditions. Further, the psychophysical kernels were consistent with an evidence-accumulation process with little leak: choice selectivity was strongest at the beginning of a trial and not at the end (which would be expected of a leaky process; Fig. 7). Thus, decisions were consistent with a lossless form of information accumulation (Brunton et al., 2013; Kiani et al., 2013). Moreover, the rate of information accumulation—which was measured in our models as a drift rate and represents the average rate of change of the underlying decision variable—depended only on the coherence of the tone bursts and not the temporal gaps between them. The temporal gaps, instead, affected the speed−accuracy trade-off of the decision (Figs. 57). Thus, the evidence-accumulation process was able to use the signals as they arrived without losing information in the intervening gaps, regardless of their duration.

This result was particularly striking in light of the fact that the temporal gaps had a strong effect on the subjects’ percept of the tone-burst stimulus: short gaps gave rise to the percept of a single, grouped sound, whereas longer gaps gave rise to the percept of discrete sounds (Fig. 3). Our findings are, therefore, somewhat surprising, given that previous work has noted an interaction between perceptual grouping and auditory judgments (Bey and McAdams, 2002; Roberts et al., 2002; Micheyl and Oxenham, 2010; Borchert et al., 2011; Thompson et al., 2011). Further work is needed to fully explore these interactions and their relationship to temporal processing. For example, it might be interesting to add nonuniform temporal manipulations to each stimulus sequence (e.g., variable IBIs or tone-burst durations) to get a better sense of how specific timing cues presented at specific times in the stimulus sequence affect both grouping and decision-making.

Neural basis

Much of auditory perceptual grouping is pre-attentive and has substantive neural signatures in the auditory midbrain and auditory cortex (Carlyon et al., 2001; Fishman et al., 2004; Micheyl et al., 2007; Sussman et al., 2007; Pressnitzer et al., 2008; Shinn-Cunningham, 2008; Winkler et al., 2009). In contrast, auditory decision-making is generally associated with neural mechanisms that are found in the ventral auditory pathway of cortex (Romanski et al., 1999; Romanski and Averbeck, 2009; Rauschecker, 2012; Bizley and Cohen, 2013). In primates, this pathway includes the core and belt fields of the auditory cortex, which project directly and indirectly to regions of the frontal lobe. Because neural activity in these regions, particularly in the belt fields, is not modulated by subjects’ choices, it is thought that this activity represents the sensory evidence used to form an auditory decision but not the decision itself (Tsunada et al., 2011; Tsunada et al., 2012; Bizley and Cohen, 2013; but see Niwa et al., 2012; Bizley et al., 2013). In contrast, frontal lobe activity is modulated by subjects’ choices, consistent with the notion that neural activity in this part of the brain reflects a transition from a representation of sensory evidence to a representation of choice (Binder et al., 2004; Kaiser et al., 2007; Russ et al., 2008; Lee et al., 2009). This hierarchy of information processing is qualitatively similar to that seen in the visual and somatosensory systems (Shadlen and Newsome, 1996; Parker and Newsome, 1998; Gold and Shadlen, 2001; Romo and Salinas, 2001; Romo et al., 2002; Gold and Shadlen, 2007; Hernández et al., 2010).

Thus, our results imply that frontal-mediated decision-making can temporally accumulate evidence from the auditory cortex, independent of how that evidence has been parsed into temporally continuous or distinct groups earlier in the auditory pathway. One possible explanation for our invariance to grouping is that the decision computations use information that is processed separately from the grouping percept. Unfortunately, whereas several studies have reported signatures of grouping in the core auditory cortex (Fishman et al., 2004; Micheyl et al., 2005; Brosch et al., 2006; Selezneva et al., 2006; Elhilali et al., 2009; Shamma and Micheyl, 2010; Fishman et al., 2013; Noda et al., 2013) and representations of the grouping percept in non-core regions of the human auditory cortex (Gutschalk et al., 2005; Gutschalk et al., 2008), neurophysiological studies elucidating where and how perceptual-grouping cues interact with decisions have yet to be conducted. Taken together, however, these aforementioned studies predict that the grouping percept may be mediated in the auditory cortex and frontal activity represents the decision process. A second, alternative explanation is that the decision process may be very flexible and able to efficiently accumulate different forms of noisy evidence under different conditions (Brunton et al., 2013; Kaufman and Churchland, 2013). The degree of this flexibility might depend on the type and quality of the sensory information, memory load, the nature of the environment in which the subject is making the decisions, and other task demands.

Acknowledgments

We thank Heather Hersh for helpful comments on the manuscript and Adam Croom for early help in collecting the psychophysical data.

Synthesis

The decision was a result of the Reviewing Editor Tatyana Sharpee and the peer reviewers coming together and discussing their recommendations until a consensus was reached. A fact-based synthesis statement explaining their decision and outlining what is needed to prepare a revision is listed below. The following reviewer agreed to reveal his identity: Mitchell Steinschneider

Review #1:

The finding that perception of one attribute of sound sequences (i.e., melody, in this case) is independent of whether the component sounds are integrated into one sound stream or heard as independent events. Overall, though, feel this paper is better suited to a more specialized journal.

Summary: For reasons of transparency, this reviewer acknowledges that perceptual modeling is not their expertise. Thus, a detailed critique of the modeling methods will not be provided. Instead, this review will focus on broader issues. In essence, using tone burst sequences as stimuli, the authors clarify the relationship between stream segregation/integration and the determination of whether the melody of a sound sequence rises or falls in frequency. Results indicate that the decision of whether a melody is rising or falling in frequency (one attribute of the sound stream) is independent of whether the individual tones were grouped into one stream, or heard as discrete acoustic events. This reviewer likens this to the perception of a sound melody sequentially played by different instruments. Results are interesting and of potentially great relevance in understanding the neural organization of sound and its discrete attributes. While this reviewer appreciates the significance of the results, the paper is very dense with detailed modeling results, and may not be of major interest to a broad neuroscience audience. In contrast, the paper would be of high interest in a more specialized journal that focuses on hearing and perception. This issue is best left as an editorial decision.

Major Issues:

1) Despite the intensive modeling, ultimately, the results of the experiments are not a major surprise. A) shorter inter-stimulus-intervals (ISIs) lead to a greater degree of sound sequence integration re longer ISIs, B) Reaction time task: accuracy and speed of response (rising or falling melody) were degraded when the rising and falling sequences were degraded and at longer ISIs, C) subjects decisions with regard to whether the sound sequence was rising (falling) were made based on if the tone frequencies tended to increase (decrease) through much of the trial, D) accuracy of behavioral responses increased with increases in coherence and the number of tone bursts presented, E) accuracy increased with signal duration.

2) Three of the subjects (total # < 10) were authors of the paper and thus not naive listeners. Further, it is quite possible that their responses could be biased toward supporting underlying hypotheses of the work. The data of these subjects should be specifically identified in at least the graphs in order to allow the readership to determine whether these expert listeners significantly biased the results of the experiments.

3) Feedback was always provided to the subjects. Did feedback bias results? Was there a progressive effect of "learning" such that results obtained at the end of a trial block were more accurate and faster than at the beginning? Similarly, was there improved performance in later blocks re earlier blocks?

4) This reviewer is confused as to the validity of the "signal RT" where the silent intervals are subtracted out from the computation of RT. It is stated (line 268) that these periods do not correspond to decision-making time. How can the authors empirically determine that within these intervals the subjects are not involved in the decision-making process? Surely, there will be neural activity in these time periods related to the task. Please explain.

Minor Issues:

5) Line 91: the phrase "adding or subtracting a fixed‐frequency increment (i.e., 7.5 Hz) to the previous tone‐burst frequency" reads as if the increment was always 7.5 Hz. '(e.g., 7.5 Hz)' would be more accurate.

6) Line 234: The purpose of the Table is not clear to this reviewer.

7) Lines 503-505: The authors cite the midbrain as a major auditory center involved in pre-attentive mechanisms subserving stream segregation/integration, but then go on to cite literature pertaining to the auditory cortex.

8) Figure 4 caption: It is unclear why only the "correct" response data are depicted.

Review #2:

The paper is relatively well written and appears sound, although I am not well acquainted with the modeling methodology, and some details were unclear. My main concern is that the connection with auditory "grouping" is extremely tenuous. It does not go beyond the observation that tone presentation rate, known to affect decisions as to whether a tone sequence is heard as one event or a sequence of events, also affects the dynamics of decision as to whether the sequence is increasing or decreasing in frequency. This outcome does not seem inconsistent with the hypothesis that both vary with the same stimulus manipulation, but are otherwise unrelated, contrary to what the Abstract suggests (2nd to last sentence).

Two aspects of the design and analysis are puzzling. (a) The models assume a fixed criterion, whereas the stimulus design and order of presentation involve a high degree of variability, so the subject's criterion is likely to vary widely before stabilizing. (b) Performance on each trial is related to the parameters of the statistics of the stimulus, rather than to the statistics of the actual stimuli (an exception is the estimation of the psychophsical kernels). This is a concern particularly as it is not clear whether the conditions are fully balanced (for example in the variable-duration task it is not clear that there are equal numbers of short vs long sequences in each condition). There are many sources of variability, some possibly unnecessary (see details).

Details:

l81: "tone bursts" --> "tones" (burst suggest a cluster of tones)

l85, 86: not intelligible until next paragraph is read

Why do you use "coherence" to manipulate sensory evidence quality? The quality of each sequence will depend on the sequence drawn, introducing a source of variability particularly for response time. Would it not be better to manipulate frequency step size?

l91: With fixed absolute frequency steps the perceptual salience will vary greatly over the range of base frequencies (500-3500 Hz). This is another source of unnecessary variabiliity. Why such a wide range? Why fixed step size? At 3500 Hz the step size (∼0.2%) is similar to the difference limen of selected and highly trained subjects. Most people have much higher thresholds (∼3-10%) so the individual steps (and possibly entire sequences) may be imperceptible for some subjects.

l106: In order to conform to instructions ("be fast but not sacrify accuracy") the subjects must evaluate the difficulty of the task. Given that IBI, coherence, direction, and baseline frequency vary randomly from trial to trial, I imagine that the subjects must take a while to stabilize their criterion, an additional source of variability. The choice is up to the subjects arbitrary decision, and I don't see how one can exclude that this decision depends on IBI in a way unrelated to sensory accumulation.

l110: why is maximum RT equal to maximum sequence duration? What is the intersequence interval, does it depend on the subject's response to the previous sequence?

l130: Were stimulus durations balanced across conditions?

l147: How do you know that coherence threshold was constant across sessions if you do not measure it?

l173: At this point "symmetric bound" is hard to understand. Perhaps precede by a short paragraph explaining the model.

l175: It is not clear to me what is "sensory evidence". Is it pitch, or pitch change, or some estimate of coherence or "signed coherence"? Coherence is a parameter of the stimulus statistics, it is not clear what an internal correlate might look like. Not clear also whether the noise in the model is purely stimulus related or whether it includes a neural component. Why is variance equal to one? Is this a model of a perceptual process, or of the stimulus process? If the latter, why do we need a model since we have perfect knowledge of the stimulus?

l259: In what way is it clear from Fig 4 that choices are not dependent on absolute frequency?

l294: DDM --> DDM parameters

l290: not clear what you mean by "signal". The word is introduced l58 as synonymous with "stimulus", and l267 in the definition of "signal RT". Provide a clear definition.

l299: Not clear what quantities are correlated. Is the high correlation not the trivial consequence of how raw and re-scaled rates are related?

l340: This analysis differs from the usual reverse correlation that averages the stimulus preceding each event in units of time relative to that event. Here the average is over time relative to stimulus onset, regardless of decision time. A decision can't depend on the stimulus that follows it, so some of the data going into this graph is irrelevant (which explains the relaxation to the mean in the later parts). Perhaps there is a better way to quantify what you're describing? Btw it might make more sense to look at dependency on tone-to-tone frequency change rather than frequency (mean subtracted).

l464: As far as I can understand the DDM does not explain grouping.

l476: Obviously they cant use information after their response...

l477: Would one not expect performance to increase with duration for almost any model?

l483: I don't think plots in Fig 7 reflect "choice selectivity". Values reflect a combination of stronger dependency on the initial slope (because post-decision values are irrelevant) and subtaction of the mean over the entire sequence.

l487: Isn't "drift rate" a parameter of the stimulus, rather than of an internal decision variable?

l503 to end: This section is vague and speculative. Any decision based on auditory stimulation must involve all stages of processing up to primary auditory cortex at least, so it does not make sense to assign some to midbrain, others to cortex, etc.

References

  1. Bey C, McAdams S (2002) Schema-based processing in auditory scene analysis. Percept Psychophys 64:844-854. [DOI] [PubMed] [Google Scholar]
  2. Binder JR, Liebenthal E, Possing ET, Medler DA, Ward BD (2004) Neural correlates of sensory and decision processes in auditory object identification. Nat Neurosci 7:295-301. 10.1038/nn1198 [DOI] [PubMed] [Google Scholar]
  3. Bizley JK, Cohen YE (2013) The what, where, and how of auditory-object perception. Nat Rev Neurosci 14:693-707. 10.1038/nrn3565 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bizley JK, Walker KM, Nodal FR, King AJ, Schnupp JW (2013) Auditory cortex represents both pitch judgments and the corresponding acoustic cues. Curr Biol 23:620-625. 10.1016/j.cub.2013.03.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bogacz R, Brown E, Moehlis J, Holmes P, Cohen JD (2006) The physics of optimal decision making: a formal analysis of models of performance in two-alternative forced-choice tasks. Psychol Rev 113:700-765. 10.1037/0033-295X.113.4.700 [DOI] [PubMed] [Google Scholar]
  6. Bogacz R, Wagenmakers EJ, Forstmann BU, Nieuwenhuis S (2010) The neural basis of the speed-accuracy tradeoff. Trends Neurosci 33:10-16. 10.1016/j.tins.2009.09.002 [DOI] [PubMed] [Google Scholar]
  7. Borchert EM, Micheyl C, Oxenham AJ (2011) Perceptual grouping affects pitch judgments across time and frequency. J Exp Psychol Hum Percept Perform 37:257-269. 10.1037/a0020670 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bregman AS (1990) Auditory scene analysis. Boston, MA: MIT. [Google Scholar]
  9. Brosch M, Oshurkova E, Bucks C, Scheich H (2006) Influence of tone duration and intertone interval on the discrimination of frequency contours in a macaque monkey. Neurosci Lett 406:97-101. 10.1016/j.neulet.2006.07.021 [DOI] [PubMed] [Google Scholar]
  10. Brunton BW, Botvinick MM, Brody CD (2013) Rats and humans can optimally accumulate evidence for decision-making. Science 340:95-98. 10.1126/science.1233912 [DOI] [PubMed] [Google Scholar]
  11. Busemeyer JR, Townsend JT (1993) Decision field theory: a dynamic-cognitive approach to decision making in an uncertain environment. Psych Rev 100:432-459. 10.1037/0033-295X.100.3.432 [DOI] [PubMed] [Google Scholar]
  12. Carlyon RP, Cusack R, Foxton JM, Robertson IH (2001) Effects of attention and unilateral neglect on auditory stream segregation. J Exp Psychol 27:115-127. [DOI] [PubMed] [Google Scholar]
  13. Carpenter RH, Williams ML (1995) Neural computation of log likelihood in control of saccadic eye movements. Nature 377:59-62. 10.1038/377059a0 [DOI] [PubMed] [Google Scholar]
  14. Carr CE, Friedman MA (1999) Evolution of time coding systems. Neural Comput 11:1-20. [DOI] [PubMed] [Google Scholar]
  15. Ding L, Gold JI (2012) Separate, causal roles of the caudate in saccadic choice and execution in a perceptual decision task. Neuron 865-874. 10.1016/j.neuron.2012.07.021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Ditterich J (2006a) Stochastic models of decisions about motion direction: behavior and physiology. Neural Netw 8:981-1012. [DOI] [PubMed] [Google Scholar]
  17. Ditterich J (2006b) Evidence for time-variant decision making. Eur J Neurosci 24:3628-3641. 10.1111/j.1460-9568.2006.05221.x [DOI] [PubMed] [Google Scholar]
  18. Drugowitsch J, Moreno-Bote R, Churchland AK, Shadlen MN, Pouget A (2012) The cost of accumulating evidence in perceptual decision making. J Neurosci 32:3612-3628. 10.1523/JNEUROSCI.4010-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Eckhoff P, Holmes P, Law C, Connolly PM, Gold JI (2008) On diffusion processes with variable drift rates as models for decision making during learning. New J Phys 10:nihpa49499. 10.1088/1367-2630/10/1/015006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Elhilali M, Ma L, Micheyl C, Oxenham AJ, Shamma SA (2009) Temporal coherence in the perceptual organization and cortical representation of auditory scenes. Neuron 61:317-329. 10.1016/j.neuron.2008.12.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Fishman YI, Arezzo JC, Steinschneider M (2004) Auditory stream segregation in monkey auditory cortex: effects of frequency separation, presentation rate, and tone duration. J Acoust Soc Am 116:1656-1670. [DOI] [PubMed] [Google Scholar]
  22. Fishman YI, Micheyl C, Steinschneider M (2013) Neural representation of harmonic complex tones in primary auditory cortex of the awake monkey. J Neurosci 33:10312-10323. 10.1523/JNEUROSCI.0020-13.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. García-Pérez MA (1998) Forced-choice staircases with fixed step sizes: asymptotic and small-sample properties. Vision Res 38:1861-1881. 10.1016/S0042-6989(97)00340-4 [DOI] [PubMed] [Google Scholar]
  24. Gold JI, Shadlen MN (2001) Neural computations that underlie decisions about sensory stimuli. Trends Cogn Sci 5:10-16. [DOI] [PubMed] [Google Scholar]
  25. Gold JI, Shadlen MN (2002) Banburismus and the brain: decoding the relationship between sensory stimuli, decisions, and reward. Neuron 36:299-308. [DOI] [PubMed] [Google Scholar]
  26. Gold JI, Shadlen MN (2003) The influence of behavioral context on the representation of a perceptual decision in developing oculomotor commands. J Neurosci 23:632-651. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Gold JI, Shadlen MN (2007) The neural basis of decision making. Annu Rev Neurosci 30:535-574. 10.1146/annurev.neuro.29.051605.113038 [DOI] [PubMed] [Google Scholar]
  28. Green CS, Pouget A, Bavelier D (2010) Improved probabilistic inference as a general learning mechanism with action video games. Curr Biol 20:1573-1579. 10.1016/j.cub.2010.07.040 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Green DM (1960) Auditory detection of a noise signal. J Acoust Soc Am 32:121-131. 10.1121/1.1907862 [DOI] [Google Scholar]
  30. Green DM, Luce RD (1973) Speed-accuracy trade off in auditory detection In: Attention and performance IV (Kornblum S, ed), pp 547–570. New York: Academic. [Google Scholar]
  31. Green DM, Birdsall TG, Tanner WP Jr (1957) Signal detection as a function of signal intensity and duration. J Acoust Soc Am 29:523-531. 10.1121/1.1908951 [DOI] [Google Scholar]
  32. Greenwood DD (1961) Auditory masking and the critical band. J Acoust Soc Am 33:484-502. 10.1121/1.1908699 [DOI] [PubMed] [Google Scholar]
  33. Griffiths TD, Warren JD (2004) What is an auditory object? Nat Rev Neurosci 5:887-892. 10.1038/nrn1538 [DOI] [PubMed] [Google Scholar]
  34. Gutschalk A, Micheyl C, Melcher JR, Rupp A, Scherg M, Oxenham AJ (2005) Neuromagnetic correlates of streaming in human auditory cortex. J Neurosci 25:5382-5388. 10.1523/JNEUROSCI.0347-05.2005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Gutschalk A, Micheyl C, Oxenham AJ (2008) Neural correlates of auditory perceptual awareness under informational masking. PLoS Biol 6:e138. 10.1371/journal.pbio.0060138 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Hawkins GE, Forstmann BU, Wagenmakers EJ, Ratcliff R, Brown SD (2015) Revisiting the evidence for collapsing boundaries and urgency signals in perceptual decision-making. J Neurosci 35: 2476–2484. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Hernández A, Nácher V, Luna R, Zainos A, Lemus L, Alvarez M, Vázquez Y, Camarillo L, Romo R (2010) Decoding a perceptual decision process across cortex. Neuron 29:300-314. 10.1016/j.neuron.2010.03.031 [DOI] [PubMed] [Google Scholar]
  38. Kaiser J, Lennert T, Lutzenberger W (2007) Dynamics of oscillatory activity during auditory decision making. Cereb Cortex 17:2258-2267. 10.1093/cercor/bhl134 [DOI] [PubMed] [Google Scholar]
  39. Kaufman MT, Churchland AK (2013) Cognitive neuroscience: sensory noise drives bad decisions. Nature 496:172-173. 10.1038/496172a [DOI] [PubMed] [Google Scholar]
  40. Kiani R, Churchland AK, Shadlen MN (2013) Integration of direction cues is invariant to the temporal gap between them. J Neurosci 33:16483-16489. 10.1523/JNEUROSCI.2094-13.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Klein SA (2001) Measuring, estimating, and understanding the psychometric function: a commentary. Percept Psychophys 63:1421-1455. [DOI] [PubMed] [Google Scholar]
  42. Knoblauch K, Maloney LT (2008) Estimating classification images with generalized linear and additive models. J Vis 8(16):10.1-10.19. 10.1167/8.16.10 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Lee JH, Russ BE, Orr LE, Cohen YE (2009) Prefrontal activity predicts monkeys' decisions during an auditory category task. Fron Integr Neurosci 3:16 10.3389/neuro.07.016.2009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Luce RD, Green DM (1972) A neural timing theory for response times and the psychophysics of intensity. Psych Rev 79:14-57. [DOI] [PubMed] [Google Scholar]
  45. Micheyl C, Oxenham AJ (2010) Objective and subjective psychophysical measures of auditory stream integration and segregation. J Assoc Res Otolaryngol 11:709-724. 10.1007/s10162-010-0227-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Micheyl C, Tian B, Carlyon RP, Rauschecker JP (2005) Perceptual organization of tone sequences in the auditory cortex of awake macaques. Neuron 48:139-148. 10.1016/j.neuron.2005.08.039 [DOI] [PubMed] [Google Scholar]
  47. Micheyl C, Carlyon RP, Gutschalk A, Melcher JR, Oxenham AJ, Rauschecker JP, Tian B, Courtenay Wilson E (2007) The role of auditory cortex in the formation of auditory streams. Hear Res 229:116-131. 10.1016/j.heares.2007.01.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Mulder MJ, Keuken MC, van Maanen L, Boekel W, Forstmann BU, Wagenmakers EJ (2013) The speed and accuracy of perceptual decisions in a random-tone pitch task. Atten Percept Psychophys 75:1048-1058. 10.3758/s13414-013-0447-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Murray RF (2012) Classification images and bubbles images in the generalized linear model. J Vis 12:2. 10.1167/12.7.2 [DOI] [PubMed] [Google Scholar]
  50. Niwa M, Johnson JS, O'Connor KN, Sutter ML (2012) Activity related to perceptual judgment and action in primary auditory cortex. J Neurosci 32:3193-3210. 10.1523/JNEUROSCI.0767-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Noda T, Kanzaki R, Takahashi H (2013) Stimulus phase locking of cortical oscillation for auditory stream segregation in rats. PLoS One 8:e83544. 10.1371/journal.pone.0083544 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Palmer J, Huk AC, Shadlen MN (2005) The effect of stimulus strength on the speed and accuracy of a perceptual decision. J Vis 5:376-404. 10.1167/5.5.1 [DOI] [PubMed] [Google Scholar]
  53. Parker AJ, Newsome WT (1998) Sense and the single neuron: probing the physiology of perception. Annu Rev Neurosci 21:227-277. 10.1146/annurev.neuro.21.1.227 [DOI] [PubMed] [Google Scholar]
  54. Pressnitzer D, Sayles M, Micheyl C, Winter IM (2008) Perceptual organization of sound begins in the auditory periphery. Curr Biol 18:1124-1128. 10.1016/j.cub.2008.06.053 [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Rahne T, Sussman E (2009) Neural representations of auditory input accommodate to the context in a dynamically changing acoustic environment. Eur J Neurosci 29:205-211. 10.1111/j.1460-9568.2008.06561.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Raposo D, Sheppard JP, Schrater PR, Churchland AK (2012) Multisensory decision-making in rats and humans. J Neurosci 32:3726-3735. 10.1523/JNEUROSCI.4998-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Ratcliff R, McKoon G (2008) The diffusion decision model: theory and data for two-choice decision tasks. Neural Comput 20:873-922. 10.1162/neco.2008.12-06-420 [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Ratcliff R, Rouder JN (1998) Modeling response times for two-choice decisions. Psychol Sci 9:347-356. 10.1111/1467-9280.00067 [DOI] [Google Scholar]
  59. Ratcliff R, Tuerlinckx F (2002) Estimating parameters of the diffusion model: approaches to dealing with contaminant reaction times and parameter variability. Psychon Bull Rev 9:438-481. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Ratcliff R, Van Zandt T, McKoon G (1999) Connectionist and diffusion models of reaction time. Psychol Rev 106:261-300. [DOI] [PubMed] [Google Scholar]
  61. Ratcliff R, Gomez P, McKoon G (2004) A diffusion model account of the lexical decision task. Psychol Rev 111:159-182. 10.1037/0033-295X.111.1.159 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Rauschecker JP (2012) Ventral and dorsal streams in the evolution of speech and language. Front Evol Neurosci 4:7 10.3389/fnevo.2012.00007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Reddi BA, Asrress KN, Carpenter RH (2003) Accuracy, information, and response time in a saccadic decision task. J Neurophysiol 90:3538-3546. 10.1152/jn.00689.2002 [DOI] [PubMed] [Google Scholar]
  64. Roberts B, Glasberg BR, Moore BC (2002) Primitive stream segregation of tone sequences without differences in fundamental frequency or passband. J Acoust Soc Am 112:2074-2085. 10.1121/1.1508784 [DOI] [PubMed] [Google Scholar]
  65. Romanski LM, Averbeck BB (2009) The primate cortical auditory system and neural representation of conspecific vocalizations. Annu Rev Neurosci 32:315-346. 10.1146/annurev.neuro.051508.135431 [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Romanski LM, Tian B, Fritz J, Mishkin M, Goldman-Rakic PS, Rauschecker JP (1999) Dual streams of auditory afferents target multiple domains in the primate prefrontal cortex. Nat Neurosci 2:1131-1136. 10.1038/16056 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Romo R, Salinas E (2001) Touch and go: decision-making mechanisms in somatosensation. Annu Rev Neurosci 24:107-137. 10.1146/annurev.neuro.24.1.107 [DOI] [PubMed] [Google Scholar]
  68. Romo R, Hernández A, Salinas E, Brody CD, Zainos A, Lemus L, de Lafuente V, Luna R (2002) From sensation to action. Behav Brain Res 135:105-118. [DOI] [PubMed] [Google Scholar]
  69. Russ BE, Orr LE, Cohen YE (2008) Prefrontal neurons predict choices during an auditory same-different task. Curr Biol 18:1483-1488. 10.1016/j.cub.2008.08.054 [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Schnupp J, Carr CE (2009) On hearing with more than one ear: lesions from evolution. Nat Neurosci 12:692-697. 10.1038/nn.2325 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Selezneva E, Scheich H, Brosch M (2006) Dual time scales for categorical decision making in auditory cortex. Curr Biol 16:2428-2433. 10.1016/j.cub.2006.10.027 [DOI] [PubMed] [Google Scholar]
  72. Shadlen MN, Newsome WT (1996) Motion perception: seeing and deciding. Proc Natl Acad Sci U S A 93:628-633. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Shadlen MN, Hanks TD, Churchland A, Kiani R, Yang T (2006) The speed and accuracy of a simple perceptual decision: a mathematical primer In: Bayesian brain (Rao R, Pouget A, eds). Cambridge, MA: MIT. [Google Scholar]
  74. Shamma S, Micheyl C (2010) Behind the scenes of auditory perception. Curr Opin Neurobiol 20:361-366. 10.1016/j.conb.2010.03.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Shamma S, Elhilali M, Micheyl C (2011) Temporal coherence and attention in auditory scene analysis. Trends Neurosci 34:114-123. 10.1016/j.tins.2010.11.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Shinn-Cunningham BG (2008) Object-based auditory and visual attention. Trends Cogn Sci 12:182-186. 10.1016/j.tics.2008.02.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Smith PL, Ratcliff R (2004) Psychology and neurobiology of simple decisions. Trends Neurosci 27:161-168. 10.1016/j.tins.2004.01.006 [DOI] [PubMed] [Google Scholar]
  78. Sussman E, Steinschneider M (2009) Attention effects on auditory scene analysis in children. Neuropsychologia 47:771-785. 10.1016/j.neuropsychologia.2008.12.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  79. Sussman ES, Horváth J, Winkler I, Orr M (2007) The role of attention in the formation of auditory streams. Percept Psychophys 69:136-152. [DOI] [PubMed] [Google Scholar]
  80. Thompson SK, Carlyon RP, Cusack R (2011) An objective measurement of the build-up of auditory streaming and of its modulation by attention. J Exp Psychol Hum Percept Perform 37:1253-1262. 10.1037/a0021925 [DOI] [PubMed] [Google Scholar]
  81. Thura D, Beauregard-Racine J, Fradet CW, Cisek P (2012) Decision making by urgency gating: theory and experimental support. J Neurophysiol 108:2912-2930. 10.1152/jn.01071.2011 [DOI] [PubMed] [Google Scholar]
  82. Treutwein B (1995) Adaptive psychophysical procedures. Vision Res 35:2503-2522. [PubMed] [Google Scholar]
  83. Tsetsos K, Chater N, Usher M (2012) Salience driven value integration explains decision biases and preference reversal. Proc Natl Acad Sci U S A 109:9659-9664. 10.1073/pnas.1119569109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Tsunada J, Lee JH, Cohen YE (2011) Representation of speech categories in the primate auditory cortex. J Neurophysiol 105:2634-2646. 10.1152/jn.00037.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  85. Tsunada J, Lee JH, Cohen YE (2012) Differential representation of auditory categories between cell classes in primate auditory cortex. J Physiol 590:3129-3139. 10.1113/jphysiol.2012.232892 [DOI] [PMC free article] [PubMed] [Google Scholar]
  86. Usher M, McClelland JL (2001) The time course of perceptual choice: the leaky, competing accumulator model. Psychol Rev 108:550-592. [DOI] [PubMed] [Google Scholar]
  87. Wichmann FA, Hill NJ (2001) The psychometric function: I. Fitting, sampling, and goodness of fit. Percept Psychophys 63:1293-1313. [DOI] [PubMed] [Google Scholar]
  88. Wickelgren WA (1977) Speed-accuracy tradeoff and information processing dynamics. Acta Psychologica 41:67-85. 10.1016/0001-6918(77)90012-9 [DOI] [Google Scholar]
  89. Winkler I, Denham SL, Nelken I (2009) Modeling the auditory scene: predictive regularity representations and perceptual objects. Trends Cogn Sci 13:532-540. 10.1016/j.tics.2009.09.003 [DOI] [PubMed] [Google Scholar]

Articles from eNeuro are provided here courtesy of Society for Neuroscience

RESOURCES