Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2021 Aug 2;17(8):e1009251. doi: 10.1371/journal.pcbi.1009251

Mathematical framework for place coding in the auditory system

Alex D Reyes 1,*
Editor: Lyle J Graham2
PMCID: PMC8360601  PMID: 34339409

Abstract

In the auditory system, tonotopy is postulated to be the substrate for a place code, where sound frequency is encoded by the location of the neurons that fire during the stimulus. Though conceptually simple, the computations that allow for the representation of intensity and complex sounds are poorly understood. Here, a mathematical framework is developed in order to define clearly the conditions that support a place code. To accommodate both frequency and intensity information, the neural network is described as a space with elements that represent individual neurons and clusters of neurons. A mapping is then constructed from acoustic space to neural space so that frequency and intensity are encoded, respectively, by the location and size of the clusters. Algebraic operations -addition and multiplication- are derived to elucidate the rules for representing, assembling, and modulating multi-frequency sound in networks. The resulting outcomes of these operations are consistent with network simulations as well as with electrophysiological and psychophysical data. The analyses show how both frequency and intensity can be encoded with a purely place code, without the need for rate or temporal coding schemes. The algebraic operations are used to describe loudness summation and suggest a mechanism for the critical band. The mathematical approach complements experimental and computational approaches and provides a foundation for interpreting data and constructing models.

Author summary

One way of encoding sensory information in the brain is with a so-called place code. In the auditory system, tones of increasing frequencies activate sets of neurons at progressively different locations along an axis. The goal of this study is to elucidate the mathematical principles for representing tone frequency and intensity in neural networks. The rigorous, formal process ensures that the conditions for a place code and the associated computations are defined precisely. This mathematical approach offers new insights into experimental data and a framework for constructing network models.

Introduction

Many sensory systems are organized topographically so that adjacent neurons have small differences in the receptive fields. The result is that minute changes in the sensory features causes an incremental shift in the spatial distribution of active neurons. This is has led to the notion of a place code where the location of the active neurons provides information about sensory attributes. In the auditory system, the substrate for a place code is tonotopy, where the preferred frequency of each neuron varies systematically along one axis [1]. Tonotopy originates in the cochlea [2, 3] and is inherited by progressively higher order structures along the auditory pathway [4]. The importance of a place code [5] is underscored by the fact that cochlear implants, arguably the most successful brain-machine interface, enable deaf patients to discriminate tone pitch simply by delivering brief electrical pulses at points of the cochlea corresponding to specific frequencies [6, 7].

Although frequency and intensity may be encoded in several ways [8], there are regimes where place-coding seems advantageous. Humans are able to discriminate small differences in frequencies and intensities even for stimuli as brief as 5–10 ms [913]. Therefore, the major computations have already taken place within a few milliseconds. This is of some significance because in this short time interval, neurons can fire only bursts of 1–2 action potentials [14, 15], indicating that neurons essentially function as binary units. Therefore, it seems likely that neither frequency nor intensity can be encoded via the firing rate of individual cells since the dynamic range would be severely limited. Similarly, coding schemes based on temporal or ‘volley’ schemes are difficult to implement at the level of cortex because neurons can phase-lock only to low frequency sounds [1618]. However, a purely place code cannot be used for dynamically complex sound; indeed, coding and perception are enhanced significantly when temporal and rate cues are factored in [8, 12, 1922] and when longer duration stimuli are used [912].

There are several challenges with implementing a purely place coding scheme. First, the optimal architecture for representing frequency is not well-defined. Possible functional units include individual neurons, cortical columns [23, 24], or overlapping neuron clusters [25]. The dimension of each unit ultimately determines the range and resolution at which frequencies and intensities that can be represented and discriminated. Second, how both frequency and intensity can be encoded with a place code is unclear, particularly for brief stimuli when cells function mostly as binary units. Third, the rules for combining multiple stimuli is lacking. Physiological sounds are composed of pure tones with differing frequencies and intensities, resulting in potentially complex spatial activity patterns in networks. Finally, the role of inhibition in a place coding scheme has not been established.

Here, a mathematical model is developed in order to gain insights into: 1) the functional organization of the auditory system that supports a place coding scheme for frequency and intensity: and 2) the computations that can be performed in networks. To simplify analyses and to reveal the inherent advantages and limitations, the model focuses on how simple tones are represented and combined with a pure place code and excludes the dynamic variables that mediate temporally complex sounds. The approach is to use mathematical principles to construct the acoustic and neural spaces, find a mapping between the spaces, and then develop the algebraic operations. The predictions of the math model are then tested with simulations. With this formal approach, the variables that are important for a place coding scheme are defined precisely.

Results

The mathematical model is subject to the following biological constraints. First, the neural network inherits the tonotopic organization of the cochlea [2, 3] so that the preferred frequency of neurons changes systematically with location along one axis. Second, a pure tone activates a population of neurons within a confined area [26], with the location of the area varying with the tone frequency. Third, the area grows with sound intensity [26], paralleling the increase in the response area of the basilar membrane [27, 28]. The model is broadly related to the “spread-of-excitation” class of models [8, 19]

The basic computations that can be performed using a place code are shown schematically in a 2-dimensional network of neurons (Fig 1). In response to a pure tone stimulus, a synaptic field from a presynaptic population of neurons is generated within an enclosed area of the network (a, left panel, cyan disk), causing a population of cells to fire (filled circles). A tone with a higher frequency and intensity activates a larger area at a different location (right panel). A sound composed of the two pure tones activates both regions simultaneously; the regions may overlap if the difference in frequencies is small (Fig 1B). Finally, excitatory synaptic fields and the activated neuron clusters are modulated by inhibition (Fig 1C). The mathematical basis for these computations is developed below. For clarity, only the main results are shown, with details in S1 Appendix.

Fig 1. Computations with a place code.

Fig 1

a, left, hypothetical neural representation of a low frequency (fα), small amplitude (pα) pure tone stimulus in a two-dimensional neural network. A stimulus-evoked synaptic field covers a circular area (cyan disk) and causes a subset of cells to fire (filled circles). Projection of the synaptic field onto the tonotopic axis (cyan bar) gives the location and size of the activated area. right, synaptic field generated by a tone with higher frequency (fβ) and sound pressure (pβ). b, synaptic field generated by a sound composed of the two tones. c, modulation by inhibition (red).

Neural space

Although the brain has three spatial dimensions, only one dimension -that corresponding to the tonotopic axis- is relevant for a place code. Thus, the circular synaptic field in Fig 1 is projected onto the tonotopic axis (bars). In the presence of sound, afferents from an external source generates a synaptic field that covers a contiguous subset of neurons. In the following, a mathematical description of the neural space will be developed that accommodates the neural elements and synaptic fields.

The neural space is defined as an interval, bounded by minimum and maximum values xmin and xmax (Fig 2B, magenta). The neural space is partitioned into Nh sections that represent the projections of neurons to the tonotopic axis (Fig 1). For explanation purposes, the neural space is constructed from single row of neurons (Fig 2A) (see S1 Appendix for a more general definition with multiple layers of staggered neurons). The neural space and the partitions are ‘half-open’ intervals, which are closed (“[”) on one end and open on the other (“)”). This is convenient mathematically because there is no space between intervals, allowing for a formal definition of a partition (S1 Appendix). Each interval may be expressed as non-overlapping subintervals of the form [x, x + Δx). Therefore, Δx may be viewed as the width of an individual neuron and Nh as the number of cells (Fig 2B). Each interval can be uniquely identified by the point at the closed end, which also gives its location along the tonotopic axis. The set containing these points is:

X={x0,x1,,x(Nh1)} (1)

Fig 2. Mathematical representation of neural and acoustic spaces.

Fig 2

a, neurons on the tonotopic axis are positioned next to each other with no space in between. Pure tone stimuli activate afferents onto a subset of neurons (cyan, blue). b, mathematical representation of neural space. The tonotopic axis is a half-open interval (magenta) partitioned into smaller intervals that represent the space (Δx) taken up by neurons. The synaptic fields (hxα,λα,hxβ,λβ) are also half-open intervals. c, The acoustic space has two dimensions, with frequency as one axis and pressure as the other. The frequency and pressure spaces are partitioned into half-open intervals of length Δf, Δp, respectively. d, Mapping tones in the acoustic space to intervals in neural space via a function ψ.

A synaptic field spans an integral number (nλ) of neurons and is also defined as a half-open interval with length λ = nλΔx. The length λ ranges from Δx (1 interval) to a maximum λmax = nλ,maxΔx. Each synaptic interval, designated as hx = [x, x + λ), is uniquely identified by the location of the cell at the closed end (xα, xβ in Fig 2B) and by its length (λα, λβ). The set of starting points Xλ and the set of achievable lengths Λ are given by:

Xλ={xiiN0,i(Nhnλ,max)}XΛ={λ1,λ2,,λmax} (2)

The set XλX takes into account the maximum interval length to ensure that the synaptic intervals are within neural space. A more formal and general definition of neural space and its topology is in S1 Appendix. As will be shown below, Xλ will contain information about the frequency while Λ will contain information about the sound pressure.

Representing sound in neural space

The elements of acoustic space A are pure tones, each of which is characterized by a sine wave of a given frequency (f) and amplitude corresponding to sound pressure (p). Theoretically, frequencies and pressures are unbounded and can take an uncountable number of values but under physiological conditions, the audible range is likely bounded by minimum and maximum values and consists of a finite number of discriminable frequencies and pressure levels.

The acoustic space has two dimensions with frequency as one axis and sound pressure as the other. As was done for neural space, the frequency and pressure axes are defined as a half-open intervals and divided into non-overlapping subintervals expressed as [f, f + Δf) and [p, p + Δp), respectively (Fig 2C). For example, two tones with frequencies fα, fβ that are in the interval [f1, f1 + Δf) will both be ‘assigned’ to f1 (S1 Appendix). Physiologically, Δf and Δp limit the resolutions of frequency and sound pressure perception and set the lower bounds of difference limens (see Discussion). The sets of Nf audible frequencies and Np pressures are:

F={f0,f1,,f(Nf1)}P={p0,p1,,p(Np1)} (3)

where the elements are the first points of each interval.

The number of audible frequencies and pressures are limited by the number of synaptic intervals that fit into the tonotopic axis (|Xλ|) and the number of cells that fit into a single synaptic interval (|Λ|), respectively. For simplicity, F and P are on a linear, rather than logarithmic scales.

Single tones in acoustic space are represented in neural space via a mapping ψ (Fig 2D; see S1 Appendix for formal treatment). A pure tone af,pA is mapped to an interval hx by first mapping the components f to x and p to λ via ψf(f) = x and ψp(p) = λ.

ψ(af,p)=hψf(f),ψp(p)=hx,λ (4)

By adjusting |F| and |P| to match |Xλ| and |Λ|, respectively, the mapping of single tones onto intervals can be made bijective (one-to-one, onto).

A mapping from acoustic space to intervals that are inhibitory can be similarly defined. The mapping is complicated by the fact that there fewer inhibitory (I) cells than excitatory (E) [29], may have different tuning properties [30, 31], and can be in the co-tuned or lateral inhibitory configuration depending on the stimulus [32]. The mapping is under ongoing investigation but for purposes of the present analyses, the mapping is taken to be identical to that for E.

Algebraic operations with synaptic intervals

Having formally described the mathematical structure of neural space, it is now possible to define the algebraic operations -addition and multiplication- for combining and modulating synaptic intervals (Fig 1B and 1C). To simplify notation, the intervals will henceforth be identified with a single subscript or superscript.

‘Addition’ of synaptic intervals is defined as their union. Let hα, hβ be two synaptic intervals. Then,

hα+hβ=defhαhβ (5)

The addition operation yields two possible results (S1 Appendix). If the two synaptic intervals do not overlap (hβ, hγ in Fig 3Ai), the union yields a set with the same two intervals (Fig 3Aii, bottom). If the two intervals overlap (hα and hβ, hα and hγ), the result is a single interval whose length depends on the amount of overlap (Fig 3Aii, top two traces). Note that summation is sublinear: |hα + hγ| < |hα| + |hγ|. Moreover, if the starting points are the same (hα and hβ), the length is equal to that of longer summand (|hα + hβ| = |hα|, |hα| > |hβ|).

Fig 3. Addition and multiplication.

Fig 3

a, schematic of network receiving three excitatory afferent inputs. i, half-open intervals representing the synaptic fields. ii, addition (union) of different combinations of intervals. b, schematic of network receiving two excitatory inputs (orange, cyan) and an inhibitory input (red). i, half-open intervals associated with the activated afferents. ii, multiplication (set minus) of each excitatory intervals by the inhibitory interval.

It is noteworthy that there is some ambiguity from a decoding perspective if there are multiple tones because the addition operation will fuse overlapping intervals into a larger, single interval (e.g. hα + hγ in Fig 3A). It would not be possible to determine whether a synaptic interval is a result of a single high intensity pure tone, multiple low intensity pure tones with small differences in frequencies, or band limited noise. There is some evidence of this ambiguity in psychophysical experiments (see Discussion).

One example of addition that occurs under biological conditions is when a pure tone arrives simultaneously to the two ears. The signal propagates separately through the auditory pathway but eventually converges at some brain region. Because each input is due to the same tone, the resultant synaptic intervals will be at same location (i.e. have the same starting points) on the tonotopic axis, though their lengths may differ because of interaural intensity differences (orange and blue intervals in Fig 3A). Fig 4A (top panel) shows the predicted total length when two intervals with different lengths are added (the length of the cyan interval is fixed while that of the blue is increased). The total length is equal to the length of the longer interval: hence, it is initially constant and equal to that of the cyan interval but then increases linearly when the length of the blue interval becomes longer.

Fig 4. Predicted effects of addition and multiplication.

Fig 4

a, top, addition of two intervals (cyan, blue) when the length of one interval (blue) is increased. bottom, one interval (blue) is shifted to the right of the other (cyan). b, top, multiplication of an excitatory interval (blue) by an inhibitory interval of increasing length (red). bottom, multiplication when the inhibitory interval is shifted to the right. c, top, effects of inhibition (red) on two excitatory intervals (blue, cyan). With the inhibitory interval at a fixed location, the distance between the excitatory intervals is increased systematically. bottom, Effects of two inputs configured as center-surround where the excitatory intervals (blue, cyan) are each flanked by two inhibitory intervals (red, magenta). One of the inputs is shifted systematically to the right of the other. Predicted product length when the inputs occur simultaneously (orange) and when calculated with inputs delivered sequentially.

Addition also takes place when the sound is composed of two pure tones with different frequencies. This would generate two synaptic intervals with different starting points and possibly different lengths (e.g. orange and cyan intervals in Fig 3A). The total length depends on the degree of overlap between the intervals. In Fig 4A (bottom panel), the location of one interval (cyan) is fixed while the other (blue) is shifted rightward. When the two intervals completely overlap, the total length is equal to the length of one interval. As the blue interval is shifted, the total length increases linearly and plateaus when the two intervals become disjoint.

Inhibition decreases the excitability of the network and would be expected to reduce the size of the synaptic interval. This is not possible with the addition operation because the union operation has no inverse (i.e. ‘subtraction’ is not defined; S1 Appendix). Therefore, to incorporate the effects of inhibition, a ‘multiplication’ operation is introduced (Fig 3B). Multiplication (‘⋅’) of an excitatory synaptic interval hE by an inhibitory interval hI is defined as:

hI·hE=defhE\hI (6)

The set minus operation “\” eliminates the points from the multiplicand hE that it has in common with the multiplier hI (Fig 3Bi and 3Bii) thereby decreasing the multiplicand’s length. Multiplication yields several results, depending on the relative locations and size of the multiplicand and the multiplier (S1 Appendix). If the excitatory (E) and inhibitory (I) intervals do not overlap, then the E interval is unaffected. If the intervals overlap (hαE and hI), then the E interval is shortened (hI·hαE in ii). If the E interval (hβE) is completely within the I interval, the product is the empty set, indicating complete cancellation (hI·hβE in ii). Multiplication can also change the starting points of the intervals and split an interval into two separate intervals. The algebraic properties of the multiplication operation are discussed in S1 Appendix.

If the excitatory and inhibitory inputs are co-activated (as in feedforward circuits), then the E and I intervals will be at the same location (same starting points) on the tonotopic axis but may have different lengths. Fig 4B (top panel) plots the predicted length of the product when the E interval is multiplied by I intervals of increasing lengths. The product length decreases and becomes zero when length of the I interval exceeds that of the E interval.

If the E and I inputs are independent of each other, the synaptic intervals could be at different locations on the tonotopic axis. Fig 4B (bottom panel) plots the length of the product when the starting point of the I interval (red) is shifted systematically to the right of the E interval (blue). The product length is zero when the E and I intervals overlap completely (separation = 0) and increases linearly as the overlap decreases, eventually plateauing to a constant value when the intervals become disjoint.

With addition and multiplication defined, the rules for combining the two operations can now be determined. A simple case is when a network receives two excitatory inputs that results in synaptic intervals (hαE, hβE) and a single inhibitory input that result in an inhibitory interval (hI). This scenario would occur if binaural excitatory inputs that converge in a network are then acted on by local inhibitory neurons. When all three inputs are activated simultaneously, the intervals combine in neural space as hI·(hαE+hβE). It can be shown that multiplication is left distributive (S1 Appendix) so that:

hI·(hαE+hβE)=(hI·hαE)+(hI·hβE) (7)

Intuitively, this means that the effect of a single inhibitory input on two separate excitatory inputs can be calculated by computing the inhibitory effects on each separately and then adding the results. Multiplication, however, is not right distributive. Thus, given two inhibitory intervals (hαI and hβI) acting on a single excitatory interval (hE):

(hαI+hβI)·hE(hαI·hE)+(hβI·hE) (8)

Fig 4C (top panel) plots the predicted length when two E intervals are multiplied by an I interval. The two E intervals (blue, cyan) are shifted, respectively, left- and rightward relative to the I interval (red). The product length is zero as long as the two E intervals are within the I interval. When the two E intervals reach and exceed the borders of the I interval, the product length increases and reaches a plateau when the E and I intervals become disjoint.

A common physiological scenario is when sound is composed of two pure tones and each tone results in an excitatory synaptic field surrounded by an inhibitory field (center surround inhibition, Fig 4C, bottom panel). The corresponding composite interval contains an excitatory interval that is flanked by two inhibitory intervals (inset). Letting the I -E -I interval triplet generated by each tone be described by (hαI+hβI)·hϵE and (hγI+hδI)·hζE, the expression when both occur simultaneously is:

(hαI+hβI+hγI+hδI)·(hϵE+hζE) (9)

In Fig 4C (bottom panel, orange curve), the location of one composite interval is shifted rightward. When the composite intervals coincide (separation = 0), the product length is equal to that of a single excitatory interval. With increasing separation, the product length decreases towards zero but then increases, reaching a plateau when the excitatory and inhibitory components of each composite interval no longer overlap.

Because of the distributive properties, the effect of introducing two tones simultaneously cannot be predicted by introducing each separately and then combining the results. That is,

(hαI+hβI+hγI+hδI)·(hϵE+hζE)(hαI+hβI)·hϵE+(hγI+hδI)·hζE (10)

The green curve in Fig 4C (bottom panel) is the predicted product length when the I -E -I triplet pairs are delivered separately and their product lengths subsequently summed. Intuitively, the curves differ because the effects of inhibition on the adjacent excitatory interval is absent; indeed, the result resembles that of adding two excitatory intervals (Fig 4A, bottom panel). A practical implication is that the intervals due to complex sound cannot be predicted by presenting individual tones separately (see Discussion).

Simulations with spiking neurons

Key features of the mathematical model were examined with simulations performed on a 2 dimensional network model of spiking excitatory and inhibitory neurons in auditory cortex [32] (code available at https://github.com/AlexDReyes/ReyesPlosComp.git). This model was chosen because the firing properties of and connection schemes between E and I, which determine the size of the synaptic fields, have been fully characterized experimentally [33] and can be modified readily. Extensive simulations also showed how the firing behavior is affected by the interaction of E and I cells [32]. Both E and I neuron population receive a Gaussian distributed excitatory drive from an external source (Fig 5A); the E cells in addition receive feedforward inhibitory inputs from the I cells. Stimulation evokes Gaussian distributed excitatory inward currents in both populations and also inhibitory currents in the E cells (profiles of currents shown in insets). With brief stimuli, the recurrent connections between neurons [33] do not contribute significantly to activity in auditory cortical circuits [32] and were omitted. The region encompassing neurons that fire is henceforth referred to as the activated area (Fig 5B, top panel). The underlying synaptic field (bottom panel) is described by the area of the network where the net synaptic inputs to cells exceeded (were more negative than) rheobase, the minimum current needed to evoke an action potential (IRh, inset). As defined, the synaptic field is a composite of all the inputs, both excitatory and inhibitory, that are evoked during a stimulus. Both the activated area and the synaptic field are quantified either by the diameters of circles fitted to the boundary points (magenta) or by the length of their projections to one axis (orange bar). Note that the spatial dimensions have units of cell number (see Methods to convert to microns).

Fig 5. Simulations with spiking neurons.

Fig 5

a, Network consists of excitatory and inhibitory neuron populations. An external drive evokes excitatory inputs in both populations (blue disks) and inhibitory inputs to the E cells (red disk). insets, profiles of excitatory (blue) and inhibitory (red) currents evoked in the E and I populations. b, bottom, synaptic field evoked in the network during a stimulus. The spatial extent of the synaptic field is quantified either by the diameter of a circle fitted to its outermost points (magenta) or by the length of its projection to the tonotopic axis (orange bar). inset, profile of net synaptic current generated in the E cell population. The perimeter of the synaptic field encompasses cells whose net synaptic current input exceeded rheobase (IRh). top, activated area contains cells that fired action potentials (dots).

To test the addition operation, two external excitatory drives were delivered to the center of the network simultaneously (without inhibition). Increasing the width (by increasing the standard deviation σα of the external drive) of one stimulus, while keeping that (σβ) of the other fixed, increased the diameters of the synaptic field and activated areas (Fig 6A, top panel, i-iii). As predicted, the diameters of the synaptic field (bottom panel, orange) and activated area (black) initially did not change but then increased as σα continued to widen. However, because the synaptic currents were Gaussian distributed (Fig 5A, bottom panel), the curve started to increase before σα became equal to σβ (σασβ=1). When delivered simultaneously, the magnitude of the composite current increased, causing the region that exceeded rheobase to widen (Fig 6A, top panel, compare synaptic field evoked with a single stimulus (i) to that evoked with 2 stimuli (ii)). The diameter can be calculated from the standard deviations of the two inputs (diameterσα2+σβ2).

Fig 6. Test of addition.

Fig 6

a, top, Activated area (top; 1 sweep) and underlying synaptic field (bottom; average of 25 sweeps). i, one stimulus. ii-iii, two stimuli delivered to the center of the network. The width of one input was systematically increased (σα: 2- 45 cells) while that of the other (σβ = 20 cells) was kept constant. bottom, plot of synaptic field (orange) and activated area (black) diameters vs σασβ. Dashed curve is predicted relation. b, addition of two spatially separated excitatory inputs (σα = σβ = 10). top, i–iii, activated areas and synaptic fields with increasing stimulus separation. Inset in ii shows example of excitatory synaptic current profiles. bottom, projection length vs. separation distance for synaptic field (orange) and activated area (black). Dashed curves are predicted changes.

To examine the addition of spatially disparate synaptic fields, two excitatory inputs were delivered at different distances from each other (Fig 6B, top panel). Consistent with the prediction, the projection lengths of the synaptic field (bottom panel, orange) and activated areas (black) increased with stimulus separation and reached a plateau when the two inputs became disjoint (iii). The projection lengths were greater than predicted (dashed lines) when the separation was small (< 10 cells) and when the intervals were just becoming disjoint (at separation ∼40 cells, ii) due to the summation properties of the Gaussian distributed inputs discussed above.

To test the multiplication operation, the E and I neurons were stimulated simultaneously, resulting in excitatory and inhibitory synaptic currents in the E cells (inset in top panel of Fig 7Aii). The width of the excitatory input (σexc) was kept constant while that of the inhibitory input (σinh) was increased systematically. As predicted, the diameter of the synaptic field (bottom panel, orange) and activated area (black) decreased with increasing σinh. However, the diameter asymptoted towards a non-zero value. Because the network was feedforward, the inhibitory input was delayed relative to excitation by about 10–15 ms; as a result, there was always a time window where excitation dominated [32]. The excitatory synaptic input was not canceled even when the inhibition was twice as wide (top panel, iii).

Fig 7. Test of multiplication.

Fig 7

a, top, i-iii activated areas and synaptic fields evoked with excitatory (σexc = 20) and inhibitory (σinh = 2 − 45) inputs. The spatial extent of the inhibition is demarcated by the red circles (inner circle: 1 σinh; outer: 2 σinh). Inset in ii shows an example of excitatory (blue) and inhibitory (red) synaptic current profiles. bottom, plot of activated area (black) and synaptic field (orange) vs σinhσexc. b, Same as in a except that the excitatory input (σexc = 10) was shifted systematically to the right of inhibition (σinh = 10). bottom, Diameters of activated area (black) and synaptic field (orange) plotted against separation between excitatory and inhibitory synaptic fields.

To examine multiplication of spatially disparate E and I inputs, the excitatory input was shifted systematically to the right of the inhibitory input (Fig 7B). As predicted, the diameters of the synaptic field (bottom panel, orange) and activated area (black) increased with the E -I separation and plateaued when the E and I inputs became disjoint (iii).

To examine how multiplication distributes over addition, two excitatory inputs and one inhibitory input were delivered simultaneously to the network (Fig 8A). This is the analog of the left hand side of (Eq 7). All three inputs were initially at the center and then with the inhibition stationary, the two excitatory inputs were shifted left and right (Fig 8A, top panel, i-iii). As predicted, the projection length of the synaptic field increased towards an asymptotic value (bottom panel, orange). To reproduce the right hand side of Eq 7, simulations were performed with inhibition, first with one of the excitatory inputs and then with the other; the resultant projection lengths of each were then summed (green). As was observed with simultaneous stimulation, the projection length increased with separation. The match was poor at small separations <10 cells (i) and at separation of ∼ 30 cells (ii) because the interaction between the Gaussian excitatory currents (see above) did not factor in when each input was delivered separately. The two curves were nearly identical at separations of 40–80 cells. In this range the E inputs were disjoint (as indicated by the plateauing of the excitation-only curve (dashed orange)) but still overlapped with the inhibitory synaptic field (the orange and green curves were below the excitation-only dotted curve).

Fig 8. Test of distributive properties.

Fig 8

a, top, i-iii Representative activated areas and synaptic fields generated by two excitatory inputs and a single inhibitory input (σexc = σinh = 10). With the location of inhibition fixed, the two excitatory inputs were separated systematically. The spatial extent of the inhibition is demarcated by the red circles (inner circle: 1 σinh; outer: 2 σinh). Inset in iii shows the excitatory (blue) and inhibitory (red) synaptic current profiles. bottom, plot of synaptic field projection lengths (orange) vs separation of the excitatory inputs. Green symbols are projection lengths obtained with the sequential stimulation protocol (see text). Dotted curve is with no inhibition. b, Simulations with two excitatory-inhibitory pairs, each with center-surround configuration (see inset in iii). bottom, legend as in a except that the green traces plot the projected lengths obtained when each input (σexc = 10, σinh = 17) was delivered sequentially (see text).

Finally, the interaction of inputs with center-surround inhibition (Eq 9) was examined by delivering two excitatory inputs, each with associated inhibitory components (inset in Fig 8B, top panel, iii), to the network. The distance between the inputs was then increased systematically and the projection lengths measured (bottom panel, orange curve). At separations > 20 cells, the projection length of the synaptic field decreased to a minimum (ii) and then increased towards a plateau (iii), consistent with the prediction (Fig 4C, bottom panel). However, at separations < 20 cells, the length increased instead of decreasing; this is likely due to the interaction of the Gaussian distributed excitatory fields discussed above. To confirm that the same result cannot be obtained by presenting each stimulus separately (right hand side of Eq 10), each E -I pair was delivered sequentially and the individual projection lengths summed. Unlike with simultaneous stimulation and consistent with the prediction (green curve in Fig 4C, bottom panel), the projection length increased monotonically to a plateau without a dip (green curve).

Application to loudness summation

In the auditory system, the perceived loudness of band limited noise or simultaneously presented tones depends on whether the frequency components are within the so-called critical band (CB) of frequencies [3436]. An important property is that increasing the bandwidth of the noise does not increase the perceived loudness until the bandwidth exceeds CB, after which it increases linearly [37]. Moreover, this property is maintained at different sound intensities, indicating that CB does not change. The origin of the CB is unclear and there is debate as to whether it is peripheral involving mainly excitatory processes [38, 39], or central, which may also recruit inhibition [4042]. The tonotopic axis is often divided into 24 CBs, each uniquely identified by the center frequency [35]. In the following, algebraic operations are used to describe features of loudness summation and to suggest network mechanisms.

A band-limited noise stimulus, or more generally a complex stimulus with multiple tones, may be expressed, after discretization, as a set of increasing frequencies, say: Fm = {f1, f2, …, fn}. The ‘bandwidth’ is defined as the difference between the highest and lowest frequency components (BW = fnf1). In neural space, the stimulus results in an interval that is the union of individual excitatory intervals generated by each tonal component: Hm=i=1n(hiE=[xi,xi+λ)), where λ is the length of each interval and is the same for all intervals.

The model assumes that for multi-tone stimulus, one of the tones is dominant and generates inhibitory intervals (hLI and hRI) that abut an excitatory interval hdEHm with no overlap (hdEhLIhRI=), as in a so-called lateral inhibitory configuration (see S1 Appendix for formal definitions). Physiologically, the dominant tone may correspond to the tone at the center of a CB [35] or to the tone with the lowest frequency, which has been shown to mask higher frequency components [43]. The union of these 3 intervals is defined to be the critical interval: HCI=hdEhLIhRI. The boxed inset in Fig 9 shows the relationship between Hm (gray), hdE (blue), and the two inhibitory intervals (red). The length of the interval hl that results from the interaction of these intervals is given by |hl|=|(hLI+hRI)·Hm| and is taken to be a proxy for loudness perception. As shown in S1 Appendix, |hl| is equal to |hdE| as long as HmHCI.

Fig 9. Algebra of loudness summation.

Fig 9

Predicted interval lengths resulting from the interaction of multi-tone stimulus delivered simultaneously. Boxed inset, overlapping synaptic intervals (Hm, gray) generated by stimulus with 3 frequency components. Tic marks show location of interval centers (xα, xd, xβ) along tonotopic axis. The dominant tone (blue) also generates two inhibitory side bands (red). Plot shows resultant length (|hl|=|(hLI+hRI)·(Hm)|) after the operations (see text) as the number of intervals in Hm is increased (abscissa). Green bars in insets show portion of Hm that was not cancelled by inhibition. Dotted vertical line marks deviation of curves from a constant value. Compare with Figs 9 of [37].

Fig 9 shows the result graphically when Hm is lengthened by adding more tones to the stimulus. |hl| is constant (=|hdE|) until the number of components is such that Hm exceeds the boundaries of the critical interval. In this example, the deviation occurs when the number of intervals, and hence the number of frequency components, exceed 9 (dotted vertical line). The CB is then f9f1.

Increasing the intensity of each component of Fm causes an increase in the length of the interval components of Hm. As shown in S1 Appendix, the CB will not change provided that the lateral inhibitory configuration is maintained and the lengths of the inhibitory intervals are constant. Under this condition, |Hm| and HCI| increase equally (compare lower and upper curves in Fig 9). Because hdE increases, there is an increase in baseline (upward shift of curves) without a change in CB.

An all-excitatory version without inhibition will not reproduce the data: the critical interval would then be hdE and since hdEHm, |hl| will be greater than |hdE| if Hm has more than one component and will grow with increasing number of tonal components. Unlike the data, the curves would have no flat region.

The operations also describe a related experiment where instead of noise, the stimuli consisted of 4 tones whose frequency separations were varied systematically [37] (Fig B of S1 Appendix).

The above analysis elucidates the general requirements for loudness summation. While there is some evidence for a dominant tone [43] and inhibitory processes [42], the extent of the inhibitory intervals is less clear and is likely to reflect the combined effect of the individual excitatory and inhibitory intervals generated by other tones in the stimulus. The precise mechanisms needs to be systematically explored with more detailed analyses, simulations, and experiments.

Discussion

The aim of this study was to develop a mathematical framework for a place-code and derive the underlying principles for how tones of varying frequencies and intensities are represented, assembled, and modulated in networks of excitatory and inhibitory neurons. The analyses are not intended to replicate the detailed aspects of biological networks and dynamic behavior but rather to clarify the minimal conditions that must be met for a viable place coding scheme, to aid in the interpretation of experimental data, and to provide a blueprint for developing computational models. The advantage of this formal approach is that it ensures that the terms and advantages/limitations of a purely place-coding model are defined precisely, providing a foundation for examining the role of other auditory cues that enhance coding and perception (see below). In addition, the mathematical rules effectively constrain the computations that may be performed with a purely place code.

Place code framework in auditory processing: Evidence and implications

The model has several implications with regards to auditory processing. In this section, the advantages of the place coding framework are discussed and experimental data are interpreted within the context of the mathematical framework.

Representation of frequency and sound pressure

A key feature of the model is that the ‘functional unit’ of neural space is a set of contiguous neurons that have flexible borders. The associated mathematical architecture is a collection of half-open intervals of varying lengths. The model provides a framework for encoding both frequency and intensity (or sound pressure) with a purely place-coding scheme. This is advantageous for brief stimuli where firing rate and spike timing [8, 12, 19] may not be available (see Introduction). Some information may be carried by single spike latency [20]; however, spike latencies depend on other variables besides frequency and does not appear to have the dynamic range to represent the full range of audible sound pressure levels [44]. Frequency and intensity discrimination does improve with stimulus duration, suggesting that the other variables play complementary roles in improving coding and perception [912, 22].

A network with flexible functional units is also advantageous for maintaining both high resolution frequency and pressure representations. This can be appreciated by comparing the resolutions attainable with the classical columnar organization [23, 25] (the stimulus is assumed brief so that firing rate information is unavailable; see Introduction). In this scheme, the neural space is divided into non-overlapping columns with fixed dimensions and distinct borders. The frequency of a stimulus is encoded by the location of the active column and sound pressure by the number of active neurons within the column (i.e. population rate code). The relation between the maximum number of achievable frequency and sound pressure levels is given by |P|=Nh|F| (see S1 Appendix). Intuitively, to maximize the number of frequency levels, the columns should be as small as possible so that more can fit along the tonotopic axis; however, this reduces the number of pressure levels that can be encoded because there are fewer neurons within a column. In contrast, for a network with flexible borders, the relation is: |F|=Nh|P|+1. Fewer neurons (Nh=|F|+|P|1) are needed to represent the full range of frequency and pressure levels as compared to columns (Nh=|F|×|P|).

The advantage of a columnar organization is that the components of a multi-frequency stimulus remain separated in neural space. With flexible units, two intervals generated by two tones with small frequency differences and/or high intensities can fuse into a single interval and hence be perceived as a single tone. As discussed below, ambiguities in perception of complex stimulus are more consistent with a flexible unit organization.

Relation between Δf and frequency difference limen

In the model, the acoustic space is discretized to reflect the resolution limits on frequency and intensity perception imposed by neural space composed of neurons. The number of frequency levels and Δf is determined by the number of intervals that can be contained within the neural space (Eq 2). Though the model was introduced with Δx equivalent to the diameter of a cell in a single layer (Fig 2), Δx (and hence Δf) can be much smaller if several layers of neurons are considered (Fig A of S1 Appendix).

The frequency difference limen (ΔfDL), gives the smallest difference in frequency of two tones that can be discriminated by subjects. The measured ΔfDL does not have a fixed value but depends on a number of stimulus parameters including duration, intensity, and test frequencies [10, 45]. Moreover, ΔfDL, which is related to the psychophysical measure of sensitivity (‘d-prime’, [46, 47]), is affected by unspecified sources of internal noise within subjects such as trial-to-trial variability in pitch perception [48]. For these reasons, ΔfDL is likely to be larger than Δf. Thus, Δf may be viewed as the lower bound for ΔfDL for a purely place-coding scheme that would be realized under optimal, noiseless conditions.

Addition operation

The addition operation applied to synaptic intervals is defined as their union: hα+hβ=defhαhβ. An important consequence is that if the intervals overlap, they will fuse into a single, longer interval. Under physiological conditions, this would occur if tones of a multifrequency stimulus have small differences in frequencies. This is in line with psychophysical experiments, which show that subjects perceive tones with small differences in frequencies as a single tone [43, 49] and have difficulties distinguishing the individual components of a multi-frequency stimulus [50, 51].

Another consequence is that addition of two overlapping non-empty intervals is sublinear: |hα| + |hβ| > |hα + hβ|. If one interval is also a subset of the other (hαhβ), then the sum is equal to the larger of the two intervals: |hα| + |hβ| = |hβ|. This scenario would occur when binaural inputs converge onto a common site. Consistent with the prediction, electrophysiological recordings from neurons in inferior colliculus show that the frequency response areas (FRAs, assumed to be representative of activity spread, see below) evoked binaurally is equal to the larger of two responses evoked monaurally [52]. Similarly, assuming that loudness perception is linked to the length of the interval, a possible psychophysical analog is that a tone presented binaurally to a subject is perceived to be less than twice as loud as monaural stimulation [53]. The apparent sublinear effects can be explained by the properties of addition operation, though inhibitory processes may also contribute.

Multiplication operation and distributive properties

Multiplication of two synaptic intervals is defined as the set minus operation: hαhβ = hβ \ hα. The operation removes from the multiplicand (hβ) elements that it has in common with the multiplier (hα), thereby shortening it. The effect of inhibition can be inferred from the FRAs of neurons. Applying GABA blockers causes the FRAs to widen [54, 55]. If the FRA can be used as a proxy for the spatial extent of activated neurons (see below), then the result is consistent with inhibition shortening the synaptic intervals.

The manner in which multiplication distributes over addition has important implications for combining information from multiple sources. In auditory cortex, excitatory pyramidal neurons receive convergent afferent inputs from the thalamus and other pyramidal cells [56, 57]. The two afferents also appear to innervate a common set of local inhibitory neurons [33, 57]. The fact that multiplication is left distributive (Eq 7) means that the effect can be estimated by measuring the effects of inhibition (hI) on each excitatory inputs (hctxE, hthalE) separately and then summing the results: |hI·(hctxE+hthalE)|=|hI·hctxE|+|hI·hthalE|. However, because multiplication is not right distributive (Eq 8) a similar approach cannot be used to examine two sources of inhibition acting on a single excitatory interval. The analyses, for example, suggest that the combined effects of two types of inhibitory neurons on excitatory cells [31] should be examined by activating both interneurons simultaneously rather than separately.

More generally, the representation of complex sound with a place coding scheme cannot be predicted by combining the representations of individual components if the inhibition generated by each component interact. As shown in Eq 10 and Fig 4C (bottom), the response of two tones presented simultaneously is not a simple combination of the responses to each tone separately. It should be emphasized that this conclusion was derived mathematically from the distributive properties; it is not trivially related to non-linearities contributed by e.g. inhibitory conductances or voltage gated channels since the model has no biophysical variables.

Assumptions and limitations

As evidenced by cochlear implants, at least rudimentary pitch perception can be achieved with a purely place code [6, 7]. However, extracting the auditory features completely requires additional cues. Firing rate and spike timing information has been shown to enhance coding and perception [8, 12, 1922]. Indeed, some neurons are specialized for extracting precise temporal information [16, 58]. Moreover, frequency and intensity discrimination improves with stimulus duration [912], indicating the contribution of dynamic processes at the synaptic [59] and network [32] levels. Sound localization [60] and beat generation [5], both of which use phase information, cannot be implemented with a purely place code. Perception of a fundamental frequency absent from a harmonic (missing fundamental [61]) also cannot be explained with a place code as the model predicts that only intervals generated by sound can be perceived. Finally, variables that affect the intervals and operations on intervals such as non-linearities due to biophysical properties of cells (Figs 6 and 7, see below) and cochlea [62] are absent from the model. The formal approach used here can in principle be used to incorporate these variables, with the place-coding framework as a starting point.

The mathematical model is based on two salient features of the auditory system. One is that the neural space is organized tonotopically. Tonotopy has been described in most neural structures in the auditory pathway, from the cochlea and auditory nerve [2, 3, 63, 64] to brainstem areas [4, 65, 66] to at least layer 4 of primary auditory cortex. Whether tonotopy is maintained throughout cortical layers is controversial, with some studies (all in mice) showing clear tonotopy [6770] and others showing a more ‘salt-and-pepper’ organization [7072]. A salt-and-pepper organization suggests that the incoming afferents are distributed widely in the neural space rather than confined to a small area. The model needs a relatively prominent tonotopy to satisfy the requirement that synaptic intervals encompass a contiguous set of cells.

A second requirement is that the size of the synaptic interval and activated area increase with the intensity of the sound. Intensity-related expansion of response areas occurs in the cochlea [27, 28, 73] and can also be inferred from the excitatory frequency-response areas (FRAs) of individual neurons. The excitatory FRAs, which document the firing of cells to tones of varying frequencies and intensities, are typically “V-shaped”. At low intensities, neurons fire only when the tone frequencies are near its preferred frequency (tip of the V). At higher intensities, the range of frequencies that evoke firing increases substantially [68, 74]. If adjacent neurons have comparably-shaped FRAs but have slightly different preferred frequencies, an increase in intensity would translate to an increase in the spatial extent of activated neurons.

For mathematical convenience, the location of the synaptic intervals was identified by the leftmost point (closed end) of the interval, with increases in intensity signaled by a lengthening of the interval in the rightward (high frequency) direction. Similar behavior has been observed in the cochlea albeit in the opposite direction: an increase in the intensity causes response area to increase towards low frequency region of the basilar membrane while the high frequency cutoff remains fixed [3, 28, 73]. The choice of the leftmost point to tag the interval is arbitrary and any point in the interval will suffice provided an analogous point can be identified uniquely in each interval in the set. Experimentally, using the center of mass of active neurons as the identifier might be more practical.

For simplicity, both Δf and Δp are kept constant along the tonotopic axis, which is inaccurate because the range of frequencies and sound pressure changes with frequency and sound pressure level. To represent the full ranges, the frequency and pressure can be transformed into an octave and decibel scales prior to mapping to neural space.

The algebraic operations were derived from set theoretic operations and the magnitude of the underlying synaptic inputs were irrelevant. Under biological conditions, the input magnitude determines the degree to which biophysical, synaptic, and network processes become engaged, which will affect the length of the synaptic intervals and activated areas. Not surprisingly, the results of the network simulations deviated quantitatively from the mathematical predictions in some regimes (compare Fig 4 to Figs 6 and 7). Most of the discrepancies in the simulations were because the magnitudes of the synaptic inputs were Gaussian distributed along the tonotopic axis. In biological networks, the discrepancies may be exacerbated by the presence of threshold processes such as regenerative events [75, 76]. The underlying algebraic operations may be obscured in regimes such as these.

The model incorrectly assumes that the strength of inhibition is sufficiently strong to fully cancel excitation. This facilitated analysis because the effect of multiplication depends solely on the overlap between the multiplicand and multiplier. As the simulations with the feedforward network showed, the excitation cannot be fully canceled by inhibition owing to synaptic delay. Moreover, the balance may be spatially non-homogeneous: in center-surround suppression, excitation dominates at the preferred frequency with inhibition becoming more prominent at non-preferred frequencies [54, 55, 74]. To apply multiplication to biological systems, it may be necessary to define empirically an “effective” inhibitory field that takes into account for E -I imbalances.

For convenience, the simulations that were used to test the analyses predictions used a network model based on cortical circuits where the properties of the cells and patterns of connections betwen E and I cells have been fully characterized [32, 33]. However, the results should generalize to other network types provided the stimuli are brief (50 ms) so that cells fire only a single action potential. The mathematical model treats neurons as binary units and so only the first action potential is important. Hence, if the stimulus is brief and suprathreshold, the results obtained with a network consisting of e.g. repetitively firing cortical neurons [15, 33] or transiently firing bushy cells [58] will be qualitatively similar. The results are likely to differ with longer duration stimuli, which would allow various time- and voltage-dependent channels to become active and engage recurrent connections. It would also be important to confirm the operations for combining tones using cochlear/auditory nerve models that implements tonotopy derived directly from the basilar membrane [77, 78].

Methods

Simulations were performed with a modified version of a network model used previously [32]. Briefly, the model is a 200 x 200 cell network composed of 75% excitatory (E) and 25% inhibitory (I) neurons. The connection architecture, synaptic amplitudes/dynamics, and intrinsic properties of neurons were based on experimental data obtained from paired whole-cell recordings of excitatory pyramidal neurons and inhibitory fast-spiking and low threshold spiking interneurons [33]. For this study, the low-threshold spiking interneurons and the recurrent connections between the different cell types were removed, leaving only the inhibitory connections from fast spiking interneurons to pyramidal neurons. The connection probability between the inhibitory fast-spiking cells and the excitatory pyramidal cells was Gaussian distributed with a standard deviation of 75 μm and peak of 0.4 [33].

Both E and I cells received excitatory synaptic barrages from an external source. The synaptic barrages to each cell (50 ms duration) represented the activity of a specified number of presynaptic neurons. The average number (nin(x, y)) of inputs that each neuron at location x, y received followed a Gaussian curve so that cells at the center of the network received more inputs (Fig 5A, bottom). For each run, the number was randomized by drawing a number from a Gaussian distribution with mean nin(x, y) and a standard deviation 0.25 * nin(x, y) so that the synaptic fields and activated areas varied from trial to trial. Excitatory synaptic currents were evoked in the E and I cell populations and inhibitory synaptic currents in the E cell population after the I cells fired (insets in Fig 5A). The spatial extents of the synaptic inputs were varied by changing the standard deviations of the external drive. In some simulations, the E and I cell populations were uncoupled and received separate inputs that could be varied independently of each other. The neurons are adaptive exponential integrate-and-fire units with parameters adjusted to replicate pyramidal and fast spiking inhibitory neuron firing (see [32] for the parameter values).

The synaptic field was defined as the area of the network where the net synaptic currents to the cells exceeded rheobase, the minimum current needed to evoke an action potential in the E cells (IRh, inset in Fig 5B, bottom panel). IRh was estimated by calculating the net synaptic current near firing threshold (Vθ): Inet = gexc * (VθEexc)+ ginh * (VθEinh) where gexc, ginh are the excitatory and inhibitory conductances, respectively, and Eexc = 0 mV, Einh = −80 mV are the reversal potentials. For the E cells, rheobase is approximately -0.27 nA.

The spatial extent of the synaptic field or activated area was quantified as the diameter of a circle fitted to the outermost points (maroon circles in Fig 5B). In simulations with multiple components, the spatial extents were quantified as the total length of the projection onto the tonotopic axis (orange bar in Fig 5B, bottom panel). The diameters and lengths have units of cell number but can be converted to microns by multiplying by 7.5 μm, the distance between E cells in the network. For all plots, the data points are plotted as mean +/- standard deviation compiled from 20–100 sweeps.

Supporting information

S1 Appendix. Detailed description of mathematical analyses and proofs.

Fig. A: Projections of multiple layers of staggered neurons on tonotopic axis decreases Δx. Fig. B: Algebra of loudness summation applied to stimuli consisting of 4 tones with equally spaced frequencies.

(PDF)

Acknowledgments

I thank L-S Young for her insightful critiques and A. Bose for commenting on an early version of manuscript.

Data Availability

The code for the simulations can be downloaded at https://github.com/AlexDReyes/ReyesPlosComp.git.

Funding Statement

The author received no specific funding for this work.

References

  • 1.Siebert WM. Frequency discrimination in the auditory system: Place or periodicity mechanisms? Proceedings of the IEEE. 1970;58(5):723–730. doi: 10.1109/PROC.1970.7727 [DOI] [Google Scholar]
  • 2.Von Bekesy G, Wever EG. Experiments in hearing. vol. 8. McGraw-Hill; New York; 1960. [Google Scholar]
  • 3.Zwislocki JJ, Nguyen M. Place code for pitch: A necessary revision. Acta oto-laryngologica. 1999;119(2):140–145. doi: 10.1080/00016489950181530 [DOI] [PubMed] [Google Scholar]
  • 4.Hackett TA, Barkat TR, O’Brien BM, Hensch TK, Polley DB. Linking topography to tonotopy in the mouse auditory thalamocortical circuit. Journal of Neuroscience. 2011;31(8):2983–2995. doi: 10.1523/JNEUROSCI.5333-10.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Moore BC, Ernst SM. Frequency difference limens at high frequencies: evidence for a transition from a temporal to a place code. The Journal of the Acoustical Society of America. 2012;132(3):1542–1547. doi: 10.1121/1.4739444 [DOI] [PubMed] [Google Scholar]
  • 6.Zeng FG. Temporal pitch in electric hearing. Hearing research. 2002;174(1-2):101–106. doi: 10.1016/S0378-5955(02)00644-5 [DOI] [PubMed] [Google Scholar]
  • 7.Carlyon RP, Macherey O, Frijns JH, Axon PR, Kalkman RK, Boyle P, et al. Pitch comparisons between electrical stimulation of a cochlear implant and acoustic stimuli presented to a normal-hearing contralateral ear. Journal of the Association for Research in Otolaryngology. 2010;11(4):625–640. doi: 10.1007/s10162-010-0222-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Delgutte B. Physiological models for basic auditory percepts. In: Auditory computation. Springer; 1996. p. 157–220. [Google Scholar]
  • 9.Doughty J, Garner W. Pitch characteristics of short tones. II. Pitch as a function of tonal duration. Journal of Experimental Psychology. 1948;38(4):478. [DOI] [PubMed] [Google Scholar]
  • 10.Moore BC. Frequency difference limens for short-duration tones. The Journal of the Acoustical Society of America. 1973;54(3):610–619. doi: 10.1121/1.1913640 [DOI] [PubMed] [Google Scholar]
  • 11.Micheyl C, Xiao L, Oxenham AJ. Characterizing the dependence of pure-tone frequency difference limens on frequency, duration, and level. Hearing Research. 2012;292(1-2):1–13. doi: 10.1016/j.heares.2012.07.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Micheyl C, Schrater PR, Oxenham AJ. Auditory frequency and intensity discrimination explained using a cortical population rate code. PLoS computational biology. 2013;9(11):e1003336. doi: 10.1371/journal.pcbi.1003336 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Florentine M. Level discrimination of tones as a function of duration. The Journal of the Acoustical Society of America. 1986;79(3):792–798. doi: 10.1121/1.393469 [DOI] [PubMed] [Google Scholar]
  • 14.Hefti BJ, Smith PH. Anatomy, Physiology, and Synaptic Responses of Rat Layer V Auditory Cortical Cells and Effects of Intracellular GABAABlockade. Journal of Neurophysiology. 2000;83(5):2626–2638. doi: 10.1152/jn.2000.83.5.2626 [DOI] [PubMed] [Google Scholar]
  • 15.Oswald AMM, Reyes AD. Maturation of intrinsic and synaptic properties of layer 2/3 pyramidal neurons in mouse auditory cortex. Journal of neurophysiology. 2008;99(6):2998–3008. doi: 10.1152/jn.01160.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Reyes A, Rubel E, Spain W. Membrane properties underlying the firing of neurons in the avian cochlear nucleus. Journal of Neuroscience. 1994;14(9):5352–5364. doi: 10.1523/JNEUROSCI.14-09-05352.1994 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Wang X, Lu T, Bendor D, Bartlett E. Neural coding of temporal information in auditory thalamus and cortex. Neuroscience. 2008;154(1):294–303. doi: 10.1016/j.neuroscience.2008.03.065 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Elhilali M, Fritz JB, Klein DJ, Simon JZ, Shamma SA. Dynamics of precise spike timing in primary auditory cortex. Journal of Neuroscience. 2004;24(5):1159–1172. doi: 10.1523/JNEUROSCI.3825-03.2004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Siebert WM, et al. Stimulus transformations in the peripheral auditory system. Recognizing patterns. 1968;104(133):602–615. [Google Scholar]
  • 20.Bizley JK, Walker KM, King AJ, Schnupp JW. Neural ensemble codes for stimulus periodicity in auditory cortex. Journal of Neuroscience. 2010;30(14):5078–5091. doi: 10.1523/JNEUROSCI.5475-09.2010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Chase SM, Young ED. First-spike latency information in single neurons increases when referenced to population onset. Proceedings of the National Academy of Sciences. 2007;104(12):5175–5180. doi: 10.1073/pnas.0610368104 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Oxenham AJ. How we hear: The perception and neural coding of sound. Annual review of psychology. 2018;69. doi: 10.1146/annurev-psych-122216-011635 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Mountcastle VB. Modality and topographic properties of single neurons of cat’s somatic sensory cortex. Journal of neurophysiology. 1957;20(4):408–434. [DOI] [PubMed] [Google Scholar]
  • 24.Hubel DH, Wiesel TN. Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. The Journal of physiology. 1962;160(1):106–154. doi: 10.1113/jphysiol.1962.sp006837 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Towe AL. Notes on the Hypothesis of Columnar Organization in Somatosensory Cerebral Cortex; pp. 32–47. Brain, Behavior and Evolution. 1975;11(1):32–47. doi: 10.1159/000123621 [DOI] [PubMed] [Google Scholar]
  • 26.Schreiner CE. Spatial distribution of responses to simple and complex sounds in the primary auditory cortex. Audiology and Neurotology. 1998;3(2-3):104–122. doi: 10.1159/000013785 [DOI] [PubMed] [Google Scholar]
  • 27.Ren T, He W, Kemp D. Reticular lamina and basilar membrane vibrations in living mouse cochleae. Proceedings of the National Academy of Sciences. 2016;113(35):9910–9915. doi: 10.1073/pnas.1607428113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Chatterjee M, Zwislocki JJ. Cochlear mechanisms of frequency and intensity coding. I. The place code for pitch. Hearing research. 1997;111(1-2):65–75. doi: 10.1016/S0378-5955(97)00089-0 [DOI] [PubMed] [Google Scholar]
  • 29.Lefort S, Tomm C, Sarria JCF, Petersen CC. The excitatory neuronal network of the C2 barrel column in mouse primary somatosensory cortex. Neuron. 2009;61(2):301–316. doi: 10.1016/j.neuron.2008.12.020 [DOI] [PubMed] [Google Scholar]
  • 30.Li Ly, Ji Xy, Liang F, Li Yt, Xiao Z, Tao HW, et al. A feedforward inhibitory circuit mediates lateral refinement of sensory representation in upper layer 2/3 of mouse primary auditory cortex. Journal of Neuroscience. 2014;34(41):13670–13683. doi: 10.1523/JNEUROSCI.1516-14.2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Lakunina AA, Nardoci MB, Ahmadian Y, Jaramillo S. Somatostatin-expressing interneurons in the auditory cortex mediate sustained suppression by spectral surround. Journal of Neuroscience. 2020;40(18):3564–3575. doi: 10.1523/JNEUROSCI.1735-19.2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Levy RB, Reyes AD. Coexistence of lateral and co-tuned inhibitory configurations in cortical networks. PLoS computational biology. 2011;7(10):e1002161. doi: 10.1371/journal.pcbi.1002161 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Levy RB, Reyes AD. Spatial profile of excitatory and inhibitory synaptic connectivity in mouse primary auditory cortex. Journal of Neuroscience. 2012;32(16):5609–5619. doi: 10.1523/JNEUROSCI.5158-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Moore BCJ. Relation between the critical bandwidth and the frequency-difference limen. The Journal of the Acoustical Society of America. 1974;55(2):359–359. doi: 10.1121/1.1914512 [DOI] [PubMed] [Google Scholar]
  • 35.Fastl H, Zwicker E. Psychoacoustics: facts and models. vol. 22. Springer Science & Business Media; 2006. [Google Scholar]
  • 36.Scharf B. Complex sounds and critical bands. Psychological Bulletin. 1961;58(3):205. [DOI] [PubMed] [Google Scholar]
  • 37.Zwicker E, Flottorp G, Stevens SS. Critical band width in loudness summation. The Journal of the Acoustical Society of America. 1957;29(5):548–557. doi: 10.1121/1.1908963 [DOI] [Google Scholar]
  • 38.Zwicker E, Scharf B. A model of loudness summation. Psychological review. 1965;72(1):3. [DOI] [PubMed] [Google Scholar]
  • 39.Evans E, Pratt S, Spenner H, Cooper N. Comparisons of physiological and behavioural properties: auditory frequency selectivity. In: Auditory physiology and perception. Elsevier; 1992. p. 159–169. [Google Scholar]
  • 40.Pickles J. Auditory-nerve correlates of loudness summation with stimulus bandwidth, in normal and pathological cochleae. Hearing research. 1983;12(2):239–250. doi: 10.1016/0378-5955(83)90109-0 [DOI] [PubMed] [Google Scholar]
  • 41.Heinz MG, Issa JB, Young ED. Auditory-nerve rate responses are inconsistent with common hypotheses for the neural correlates of loudness recruitment. Journal of the Association for Research in Otolaryngology. 2005;6(2):91–105. doi: 10.1007/s10162-004-5043-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Delgutte B. Physiological mechanisms of psychophysical masking: observations from auditory-nerve fibers. The Journal of the Acoustical Society of America. 1990;87(2):791–809. doi: 10.1121/1.398891 [DOI] [PubMed] [Google Scholar]
  • 43.Wegel R, Lane C. The auditory masking of one pure tone by another and its probable relation to the dynamics of the inner ear. Physical review. 1924;23(2):266. doi: 10.1103/PhysRev.23.266 [DOI] [Google Scholar]
  • 44.Heil P. Auditory cortical onset responses revisited. I. First-spike timing. Journal of neurophysiology. 1997;77(5):2616–2641. doi: 10.1152/jn.1997.77.5.2642 [DOI] [PubMed] [Google Scholar]
  • 45.Ehret G. Frequency and intensity difference limens and nonlinearities in the ear of the housemouse (Mus musculus). Journal of comparative physiology. 1975;102(4):321–336. doi: 10.1007/BF01464344 [DOI] [Google Scholar]
  • 46.Dai H. On measuring psychometric functions: A comparison of the constant-stimulus and adaptive up–down methods. The Journal of the Acoustical Society of America. 1995;98(6):3135–3139. doi: 10.1121/1.413802 [DOI] [PubMed] [Google Scholar]
  • 47.Dai H, Micheyl C. Psychometric functions for pure-tone frequency discrimination. The Journal of the Acoustical Society of America. 2011;130(1):263–272. doi: 10.1121/1.3598448 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Bermudez P, Zatorre RJ. A distribution of absolute pitch ability as revealed by computerized testing. Music Perception. 2009;27(2):89–101. doi: 10.1525/mp.2009.27.2.89 [DOI] [Google Scholar]
  • 49.Thurlow W, Bernstein S. Simultaneous Two-Tone Pitch Discrimination. The Journal of the Acoustical Society of America. 1957;29(4):515–519. doi: 10.1121/1.1908946 [DOI] [Google Scholar]
  • 50.Thurlow WR, Rawlings IL. Discrimination of number of simultaneously sounding tones. The Journal of the Acoustical Society of America. 1959;31(10):1332–1336. doi: 10.1121/1.1907630 [DOI] [Google Scholar]
  • 51.Stoelinga CN, Lutfi RA. Discrimination of the spectral density of multitone complexes. The Journal of the Acoustical Society of America. 2011;130(5):2882–2890. doi: 10.1121/1.3647302 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Xiong XR, Liang F, Li H, Mesik L, Zhang KK, Polley DB, et al. Interaural level difference-dependent gain control and synaptic scaling underlying binaural computation. Neuron. 2013;79(4):738–753. doi: 10.1016/j.neuron.2013.06.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Epstein M, Florentine M. Binaural loudness summation for speech presented via earphones and loudspeaker with and without visual cues. The Journal of the Acoustical Socient of America. 2012;131:3981–3988. doi: 10.1121/1.3701984 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Wang J, Caspary D, Salvi RJ. GABA-A antagonist causes dramatic expansion of tuning in primary auditory cortex. Neuroreport. 2000;11(5):1137–1140. doi: 10.1097/00001756-200004070-00045 [DOI] [PubMed] [Google Scholar]
  • 55.LeBeau FE, Malmierca MS, Rees A. Iontophoresis in vivo demonstrates a key role for GABAA and glycinergic inhibition in shaping frequency response areas in the inferior colliculus of guinea pig. Journal of Neuroscience. 2001;21(18):7303–7312. doi: 10.1523/JNEUROSCI.21-18-07303.2001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Rose HJ, Metherate R. Auditory thalamocortical transmission is reliable and temporally precise. Journal of neurophysiology. 2005;94(3):2019–2030. doi: 10.1152/jn.00860.2004 [DOI] [PubMed] [Google Scholar]
  • 57.Schiff ML, Reyes AD. Characterization of thalamocortical responses of regular-spiking and fast-spiking neurons of the mouse auditory cortex in vitro and in silico. Journal of neurophysiology. 2012;107(5):1476–1488. doi: 10.1152/jn.00208.2011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Cao XJ, Shatadal S, Oertel D. Voltage-sensitive conductances of bushy cells of the mammalian ventral cochlear nucleus. Journal of neurophysiology. 2007;97(6):3961–3975. doi: 10.1152/jn.00052.2007 [DOI] [PubMed] [Google Scholar]
  • 59.Reyes AD. Synaptic short-term plasticity in auditory cortical circuits. Hearing research. 2011;279(1-2):60–66. doi: 10.1016/j.heares.2011.04.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Jeffress LA. A place theory of sound localization. Journal of comparative and physiological psychology. 1948;41(1):35. [DOI] [PubMed] [Google Scholar]
  • 61.De Boer E. On the residue and auditory pitch perception. In: Auditory System. Springer; 1976. p. 479–583. [Google Scholar]
  • 62.Ruggero MA, Robles L, Rich NC. Two-tone suppression in the basilar membrane of the cochlea: Mechanical basis of auditory-nerve rate suppression. Journal of neurophysiology. 1992;68(4):1087–1099. doi: 10.1152/jn.1992.68.4.1087 [DOI] [PubMed] [Google Scholar]
  • 63.Olson ES, Duifhuis H, Steele CR. Von Békésy and cochlear mechanics. Hearing research. 2012;293(1-2):31–43. doi: 10.1016/j.heares.2012.04.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Narayan SS, Temchin AN, Recio A, Ruggero MA. Frequency tuning of basilar membrane and auditory nerve fibers in the same cochleae. Science. 1998;282(5395):1882–1884. doi: 10.1126/science.282.5395.1882 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Stiebler I, Ehret G. Inferior colliculus of the house mouse. I. A quantitative study of tonotopic organization, frequency representation, and tone-threshold distribution. Journal of Comparative Neurology. 1985;238(1):65–76. doi: 10.1002/cne.902380106 [DOI] [PubMed] [Google Scholar]
  • 66.Luo F, Wang Q, Farid N, Liu X, Yan J. Three-dimensional tonotopic organization of the C57 mouse cochlear nucleus. Hearing research. 2009;257(1-2):75–82. [DOI] [PubMed] [Google Scholar]
  • 67.Stiebler I, Neulist R, Fichtel I, Ehret G. The auditory cortex of the house mouse: left-right differences, tonotopic organization and quantitative analysis of frequency representation. Journal of Comparative Physiology A. 1997;181(6):559–571. doi: 10.1007/s003590050140 [DOI] [PubMed] [Google Scholar]
  • 68.Guo W, Chambers AR, Darrow KN, Hancock KE, Shinn-Cunningham BG, Polley DB. Robustness of cortical topography across fields, laminae, anesthetic states, and neurophysiological signal types. Journal of Neuroscience. 2012;32(27):9159–9172. doi: 10.1523/JNEUROSCI.0065-12.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Issa JB, Haeffele BD, Agarwal A, Bergles DE, Young ED, Yue DT. Multiscale optical Ca2+ imaging of tonal organization in mouse auditory cortex. Neuron. 2014;83(4):944–959. doi: 10.1016/j.neuron.2014.07.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Winkowski DE, Kanold PO. Laminar transformation of frequency organization in auditory cortex. Journal of Neuroscience. 2013;33(4):1498–1508. doi: 10.1523/JNEUROSCI.3101-12.2013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Rothschild G, Nelken I, Mizrahi A. Functional organization and population dynamics in the mouse primary auditory cortex. Nature neuroscience. 2010;13(3):353. doi: 10.1038/nn.2484 [DOI] [PubMed] [Google Scholar]
  • 72.Bandyopadhyay S, Shamma SA, Kanold PO. Dichotomy of functional organization in the mouse auditory cortex. Nature neuroscience. 2010;13(3):361. doi: 10.1038/nn.2490 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Zwislocki J. What is the cochlear place code for pitch? Acta oto-laryngologica. 1991;111(2):256–262. doi: 10.3109/00016489109137384 [DOI] [PubMed] [Google Scholar]
  • 74.Sutter M, Schreiner C, McLean M, O’connor K, Loftus W. Organization of inhibitory frequency receptive fields in cat primary auditory cortex. Journal of Neurophysiology. 1999;82(5):2358–2371. doi: 10.1152/jn.1999.82.5.2358 [DOI] [PubMed] [Google Scholar]
  • 75.Chaudhuri R, Fiete I. Computational principles of memory. Nature neuroscience. 2016;19(3):394. doi: 10.1038/nn.4237 [DOI] [PubMed] [Google Scholar]
  • 76.Barral J, Wang XJ, Reyes AD. Propagation of temporal and rate signals in cultured multilayer networks. Nature communications. 2019;10(1):1–14. doi: 10.1038/s41467-019-11851-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Zilany MS, Bruce IC, Carney LH. Updated parameters and expanded simulation options for a model of the auditory periphery. The Journal of the Acoustical Society of America. 2014;135(1):283–286. doi: 10.1121/1.4837815 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Rudnicki M, Schoppe O, Isik M, Völk F, Hemmert W. Modeling auditory coding: from sound to spikes. Cell and tissue research. 2015;361(1):159–175. doi: 10.1007/s00441-015-2202-z [DOI] [PMC free article] [PubMed] [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009251.r001

Decision Letter 0

Lyle J Graham

1 Oct 2020

Dear Dr. Reyes,

Thank you very much for submitting your manuscript "Mathematical framework for place coding in the auditory system" (PCOMPBIOL-D-20-01254) for consideration at PLOS Computational Biology. As with all papers peer reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent peer reviewers. Based on the reports, we regret to inform you that we will not be pursuing this manuscript for publication at PLOS Computational Biology.

The reviews are attached below this email, and we hope you will find them helpful if you decide to revise the manuscript for submission elsewhere. We are sorry that we cannot be more positive on this occasion. We very much appreciate your wish to present your work in one of PLOS's Open Access publications.

Thank you for your support, and we hope that you will consider PLOS Computational Biology for other submissions in the future.

Sincerely,

Lyle J. Graham

Deputy Editor

PLOS Computational Biology

**************************************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The author presents a mathematical framework to address the place coding of frequency and intensity in the auditory system. The analysis focusses on the coding of frequency and intensity for brief stimuli where neural responses can be viewed as binary. The author defines the acoustic and neural spaces mathematically and considers conditions that support a place code. The manuscript is clearly written. However, I have a few questions about the model formulation.

General comments

The work aims to illustrate the conditions that support a place code for frequency and intensity. However, the results do not show the code performing in a way that matches or predicts psychophysical measurements. In describing equation 4 (line 120) it is stated that the mapping can be made to be bijective by adjusting the number of elements in each space. This can certainly be done if there are no constraints. Is it possible to start from psychophysical measurements and predict something about the neural representation, or vice versa?

A fundamental aspect of the analysis is the partitioning of the frequency and pressure spaces into a discrete set of equivalence classes. However, I have trouble understanding the definition of the equivalence classes in relation to the notion of perceptual discriminability. For example, in the equivalence classes f_min and f_min + df - epsilon are equivalent for any 0 < epsilon <= df, but f_min + df – epsilon and f_min+df are different for any epsilon > 0. This says that the JND is essentially 0 for frequencies at the boundary of the intervals. This would also suggest that JND measurements should show a strong asymmetry for test frequencies above or below a reference frequency. Please clarify how the defined equivalence classes relate to perceptual discriminability.

The nature of noise in the model is unclear. The input to the network is described as Gaussian distributed. Are the inputs drawn from a Gaussian distribution, or does the strength of the input follow the shape of a Gaussian curve? Please discuss the nature of noise in the model and how the model may be limited if noise is not considered.

The simulation with spiking neurons tests the predictions of the addition and multiplication operations on an adaptive exponential integrate and fire model with parameters that are based on pyramidal neuron responses. Given the diversity of response types in the auditory system, the simulation does not appear to be a sufficient test of the model as a general place coding model that would apply throughout the auditory system. Is it intended that the model applies specifically to a population of neurons with the parameters used in the simulation? It would strengthen the results to show that the spiking simulation is robust to changes in neuron parameters.

Specific comments

Line 92: Please discuss why dx is constant. Shouldn’t dx depend on x if different frequency ranges are differentially represented in the neural space?

Lines 107-109: It is stated that frequency/pressure takes on an uncountable number of values, but the audible range is bounded. A bounded interval is still uncountable. Did you mean unbounded instead of uncountable, or is the discretization of frequency/pressure what is being referenced here?

Line 113: Please discuss the motivation for df and dp being independent of f and p, respectively, when the just noticeable difference frequency and pressure depend on the reference frequency and pressure.

Lines 248-251: It is stated that the prediction was for no change when sigma_a was smaller than sigma_b, but in parentheses this is written as (sigma_a/sigma_b < 0.5). Should the condition (sigma_a/sigma_b < x) be equivalent to the condition that is written?

Supplementary material equation 1, equation 4a: It appears that the intervals h_f and h_p have f and p as arguments, respectively. The author might consider different notation for the frequency and pressure intervals because it is not clear whether h_2 is a frequency or pressure interval, for example. There is a similar ambiguity for H_f and H_p. It is typically clear from the context what is being referred to, but the author might consider revising the notation.

Supplementary material equation 18 bottom: I believe that a_f should be a_f_alpha or a_f_beta.

Reviewer #2: I do not believe that the modelling framework of this paper is adequate to draw the stated conclusions.

Firstly, there appear to be some fundamental problems in terms of the definition of the acoustic space. As far as I understand, the acoustic space is defined as a discrete 2d grid where a given pure tone of frequency f and amplitude p is mapped to a point in this grid. A pure tone is then represented essentially by a matrix where all elements are 0 except a single element which is 1 (I state it like this because it seems much simpler to get your head round that than the version in the paper, and entirely equivalent). If we were only talking about a single pure tone, then the stimulus space would be the set of all matrices with this property. However, it seems that combinations of tones are allowed, and an addition operation is defined on the acoustic space in the supplementary material (eqs 17-18), so that simultaneously playing two tones at different frequencies and amplitudes gives you a matrix with two 1s. However, there is a major problem here that can be illustrated with the very simple example of adding two tones at the same frequency.

Two tones with the same frequency f and amplitude p added together will give a new tone with the same frequency f and depending on their phase relationship any amplitude between 0 (out of phase) to 2p (in phase). Equation 18 only allows for an output amplitude of p. I am not sure therefore what the addition operation in the acoustic space is supposed to represent, but it isn't sounds played simultaneously. This alone seems to fatally undermine the rest of the paper. If something as basic as adding two tones together isn't handled correctly, we would need a very strong argument and clear reasoning that this framework has anything to say about real auditory perception, and there is no such argument in the paper.

The discussion (L311-319) states that three key findings, but none of them are findings, they are rather the definitions of the model. There is no 'finding' that neural space is a set of contiguous neurons, it's the definition of the model and it is not an accurate model. There is no 'finding' that operations in neural space must be union and set difference, it's another definition of the model.

I'm generally very sympathetic to what I see as the aim of this paper, which is to make a mathematical model which - even though it oversimplifies - is able to throw light on some fundamental aspect of the problem. However, I don't see any of that here. I see a framework which is not only oversimplified but not even clearly defined, and "conclusions" drawn which are just restatements of the incorrect assumptions of the model.

Reviewer #3: In the manuscript by A. Reyes, the author deals with the challenge imposed by the observation that animals can decode both the frequency and the amplitude of sound within a short period of time after exposure. This observation is seemingly incompatible with rate and temporal codes. The author considers a different (previously proposed) model of a “place code” in which the frequency response is defined by tonotopy within a network and amplitude response is defined by the extent of activation within the same network. The author proposes an algebraic formalist to describe the mapping between the sound and the neural space. Borrowing from set theory, the author introduces operations of addition and multiplication on this place code. Finally, the author compares results of these operations with network simulations.

Major critique

It is somewhat unclear what the main findings are, and there is substantial confusion between what the author calls “findings” of the study and what are actually just the definitions and assumptions of the model that the author proposes.

Key finding are outlined in lines 311-319 of the discussion. There are three. The first finding about the “function unit” with flexible borders (lines 311-314) is not truly a finding, but a definition of the mathematical framework proposed by the author. The second finding about the range and resolution dictated by the network architecture (lines 314-316) is a trivial consequence of using a place code: It is a given that network architecture in which different neurons respond to different frequencies will dictate the limitations of the code. In fact, the author outlines the intuition behind this conclusion in the Introduction. Finally, the operations of addition and multiplication (lines 316-317) are again not a finding, but part of the framework proposed by the author. It is not clear whether these operations are simply a formalism to describe the abilities and limitations of the place code, or whether these algebraic operations provide a new insight into these abilities and limitations.

Perhaps the manuscript is not intended to outline findings about neural networks, but is rather intended to introduce a formalism that may be useful for describing place coding. In this case, it is unclear in what way this formalism is useful. Does it allow describing phenomena that could not be described before? Does it reveal insights or make predictions about auditory circuits?

The author appears to make two predictions. One is that the place code is “sufficient for representing sounds” (line 329). It is unclear how this conclusion is derived, and what “sufficient” even means. Clearly, the manuscript itself states that there are limitations of the place code with regard to the resolution of sounds and ambiguity of complex sounds (lines 139-143).

A second prediction is that the response to a complex sound is not simply a combination of responses to the simple sounds that it consists of. In other words, the manuscript states that responses are nonlinear. This appears to be a trivial consequence of using a nonlinear process of inhibition, as defined by the manuscript itself.

In summary, I was left with major confusion about the main premise of the manuscript, the main results, and the predictions.

Minor comments

- The assumption that rate coding does not work on very short timescales is questionable. Even if an individual neuron is binary (spike or no spike) within a short time interval, the firing rate can be measured in a large enough population of neurons as the probability of spiking – i.e., the fraction of neurons with a certain preferred frequency that responded within a short time window.

- It is unclear why the manuscript defines frequency as the leftmost endpoint of an interval. Shouldn’t it be the center of an interval? I.e., louder sounds should activate frequencies both to the left and to the right of a preferred frequency.

- Line 41: “unlikely” should actually be “likely”

- Line 137: < (less than) should be less than or equal to

- Line 215: the use of subscripts is confusing because two variables with the same subscript actually have no relationship to one another (in particular h_beta^I and h_beta^E)

- Fig.6 has confusing usage of diameter and projection length, which both appear to be related to the extent of activation.

--------------------

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

Reviewer #2: No: There is no code or data given. There's no experimental data reported, so that doesn't need to be provided, but I'm not sure if code needs to be made available or not.

Reviewer #3: Yes

--------------------

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009251.r003

Decision Letter 1

Lyle J Graham

3 May 2021

Dear Dr. Reyes (Alex),

Thanks for your patience. I was obligated to recruit a new reviewer for this round, but I think that she/he took into account the process from the beginning.

It's clear that both reviewers like what you are trying to do. At the same time I think that their comments pose a challenge, which is why this is a Major Revision, and I leave it to you to decide if you would like to submit a revision.  The rest is the form letter.

Best,

Lyle

Thank you very much for submitting your manuscript "Mathematical framework for place coding in the auditory system" for consideration at PLOS Computational Biology.

As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Lyle J. Graham

Deputy Editor

PLOS Computational Biology

Lyle Graham

Deputy Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: I appreciate the goal of developing a mathematical approach to studying place coding in the auditory system. The author improved the manuscript with the revision, but there remain points that should be clarified.

Comments

1. Although my comment 3 from the first review was addressed in the manuscript, the issue continues to be a point of confusion. The author now states that experiments should be performed to test the hypothesis that the JND should range from being infinitesimally small to being equal to ∆f. While it would be helpful to directly test the hypothesis, it seems unlikely that such a striking phenomenon would not have been noted in the many existing studies of frequency JND. Is it possible to revise the framework so that the JND would not be sensitive to where the test frequencies fall within the intervals?

2. The mathematical framework should make statements about coding as precise as possible. With this view, the notion that df and dp are loosely related to the JND is unsatisfying. Can the nature of this relationship be made explicit?

3. Please clarify the definition of the neural and acoustic spaces. It appears that the spaces are defined as the collection of half-open intervals. If this is the case, then the space is not closed under the addition operation (union) because the union of two disjoint half-open intervals is not a half-open interval. It is not clear whether the spaces are the collection of half-open intervals or the collection of unions of all such intervals.

Reviewer #4: This is a review of a "Mathematical framework for place coding in the auditory system" by A. Reyes. This is a resubmission, but it my first time reading this manuscript (I did not review the 1st submission).

The author outlines a novel way to describe neural encoding, specifically a place code of sound frequency & intensity, using formal language of topology and set theory. The development of the theory is thoughtful and thought-provoking. Implications of the theory are pursued, specifically: how tones and tone combinations would be represented as spatial patterns of activation under the rules stipulated by the mathematical formulation. These outcomes of the formal model are then compared to simulations of an auditory cortex-inspired model, adapted from the author's previous work.

*** General comments ***

I enjoyed the formal mathematical perspective and appreciate that that author has attempted to tackle a fundamental question of neural representation in a novel way. It seems to me that there are a number of limitations of the approach that make me question whether it can be usefully applied to the auditory system. I expect the author can address these, and I think that readers of PLoS CompBio will enjoy reading this creative theoretical work. Major concerns are (some of these are related detailed comments below):

* this framework explicitly leaves out all temporal information of inputs and neural responses, and yet we know the auditory system is highly specialized to extract temporal information from inputs. It would be helpful for the author to expand on this in the Introduction (see some detailed comments below)

* this framework does not account for nonlinearities that are well-known to arise for even the most basic combinations of tones (some examples: perceptual phenomena such as combination tones and missing fundamental; two-tone suppression at the the cochlear level), and also does not account for complications of phase (as raised by Reviewer 2 in the previous round). The author should discuss how the model should be interpreted in light of these phenomena

* what is the utility (or anticipated utility) of the mathematical formalism & topological machinery introduced? It seems that that the model is defined by set operations (union, exclusion) on intervals. These are topics that could be familiar and intuitive to non-mathematically-sophisticated readers (and thus the model could be presented in a more mathematically "gentle" way, without much reference to topological terms). Of course, if the topological formulation leads to insights gained, for example, from "deeper" theorems or tools of topology, then the approach would be justified, but I don't see substantial uses of the topological approach in this manuscript. There is also mention of more esoteric algebraic structures in the supplemental material (monoid, magma) but no discussion of the significance of these structures in the current study.

* The formal model is compared to a spiking model. It is not stated clearly what this model represents. I take it to be a model for auditory cortex (based on the author's previous work, and the mention of "auditory cortical circuits" on line 244). The author should describe the rationale for this model. In particular, could the author expand on this choice as opposed to, say, comparing the formal predictions to activation patterns of auditory nerve simulations. Several well-known auditory nerve models are freely-available (see e.g. https://github.com/mrkrd/cochlea), the elements of tonotopy and place code are already established at this stage of auditory processing, and coding of the "simple" stimuli considered here (tones, tone combinations) would likely begin in the cochlea/ANF (whereas auditory cortex may be engaged in encoding "higher-level" auditory features).

*** Detailed comments ***

Line 8 (abstract): I think I was confused by the mention of the bijective mapping in the abstract (and also line 128). From this statement, I expected to see stronger results about what this framework would imply for decoding. But, as is discussed (line 146, also Fig 3) there is ambiguity in decoding. I would suggest some language to clarify that the bijection is on single tones, but not combinations of tones, or something similar?

Lines 11-12 (abstract): suggest removing "predicted", and just say "outcomes of these operations" or "resulting outcomes of these operations" [echoing some of Referee 2 and Referee 3's concerns from previous round]

Lines 40-45: Couldn't a counterargument be made here that acoustic information can possibly be carried by first-spike latency (Heil1997, Bizley et al 2010, etc)?

More generally, it seems that some discussion should be given to the fact that there are areas of the auditory system appear highly-specialized for temporal features of sounds and temporal precision in synaptic transmission, so a "place only" code (all temporal information ignored) may be an extreme way to view encoding in the auditory system.

Line 44-46: agree that "temporal or ‘volley’ schemes [are] difficult to implement at the level of cortex", but an alternate view is that temporal or volley information is used at some earlier stage and transformed for use by the auditory cortex, so AC does not have to implement a place code for tones. Does this matter for interpretation of the model?

Line 54-56. Could add: it is also a challenge to understand how place codes can work when inputs (sounds) are dynamic and temporally complex. [but then should acknowledge that temporal dynamics are not considered in this study]

General comment on Intro: would be helpful to be more explicit that the author's definition of "place code" means place ONLY, no temporal information of any kind considered in inputs or neural responses. and then comment on the meaningfulness & limitations of this approach for understanding the auditory system.

Line 62 - "Math model" - change to mathematical

Eq. 3: The statements in lines 504-507 should also be included near here to clarify that f and p are not defined on a logarithmic scale, as might be expected

Line 129-131 - please provide reference for statement that E and I receptive fields differ. this seems to refer to the question of "co-tuning" that has been considered by the author before, more context/explanation could be helpful here.

Line 136: closure is said to be "important". Not clear to me on what grounds. Important in terms of the ensuring the model is a sensible description of neural processes? Or important because of the author's goal to adhere to formal mathematical requirements?

Line 174: suggest a more specific statement: "because the set union operation has no inverse" [instead of "because there is no inverse"]

Fig 7: Could dashed lines for theory predictions be included here, as was done in Fig. 6, or are they too distinct?

Line 322: typo: "applied [to] a well-known"

"Relation to Difference Limens" section. This is indeed an interesting prediction based on how the model is constructed, but doesn't this prediction require the equivalence classes (partitions of acoustic and neural space) to be very "rigid"... sensitive to infinitesimally small changes that cross the boundaries of intervals. This would seem to contradict the idea (line 115) that this construction is "loose". Indeed I would agree that it more reasonable to think of the partition of acoustic and neural space as more of qualitative/approximate way of thinking about how the place code construction.

Line 355 style: fDL could be $f_{DL}$ (subscript)

Line 371: "overlapping *non-empty* intervals" [so inequality strict}

Line 375: FRA defined in supplemental but not defined in text

Line 376: typo: "equal to [the] larger of two"

Line 394-397. Multiplication not right distributive. Can this be expanded on? For instance, are there certain local connectivity circuits/motifs for which this approach is more or less relevant

Line 401-403: "response of neurons to pure tones and white noise differ". Can something more specific be said here as to how the model is "in line" with the data. The results as presented did not specifically consider the case of broadband (noise) stimuli.

Line 404: typo. "it [is] not trivially"

Line 406 - 413. "Transmission of acoustic info" section could be clarified. The terms homeomorphism & homomorphism should be defined for non-specialists.

the phrase "would necessitate that the topology of the source and target brain regions be the same" is loose here... "the same" in a formal sense that there is a homeomorphism between the two? or "the same" in some way that can be interpreted in terms of the source space and neural space.

Lines 412-413 are also vague as they do not specify what constraints the author has in mind.

GENERAL: Nonlinearities such as missing fundamental, combination tones

Line 414. Algebra of loudness summations. This could be presented as a result (and moved to Results).

Line 479: Could say tonotopy starts in the cochlea (von Bekesy) and auditory nerve, i.e. prior to auditory brainstem.

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #4: No

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #4: No: I may have missed it, but I don't see any information regarding availability or distribution of code used for network simulations.

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009251.r005

Decision Letter 2

Lyle J Graham

22 Jun 2021

Dear Alex, I think you are there, but I would appreciate if you could address the last comments of Reviewer 4.

Best,

Lyle

Dear Dr. Reyes,

Thank you very much for submitting your manuscript "Mathematical framework for place coding in the auditory system" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations.

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Lyle J. Graham

Deputy Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The author addressed my comments.

Reviewer #4: The author has revised the manuscript and addressed my concerns. In particular, several of the additions (including lines 46, 61, lines 475-488) usefully define the scope and limitations of the study. The technical approach is clearly developed in the supporting documents. This material will be much appreciated by readers who want to dive deeper into the interesting and novel results presented in the paper.

The paper includes simulation results from a computational model. I do not see it stated if/where that code is available. Apologies if I missed it.

Two comments that I expect the author can address with some further minor revisions are:

* lines 113-115

"Theoretically, frequencies and pressures are unbounded and can take an uncountable number of values but under physiological conditions, the audible range is likely bounded by minimum and maximum values and is finite."

A previous reviewer flagged an earlier version of this passage, noting that finite intervals can contain uncountably many elements (for real numbers, e.g.). I think this wording could still be adjusted to avoid confusion. One simple fix could be to delete "and can take an uncountable number of values"

However, my sense is that the author is purposeful in using the word "uncountable" in order to draw a contrast with the fact that the model has finitely-many elements in acoustic space. In this case, maybe say something like

Theoretically, frequencies and pressures are unbounded and can take an uncountable number of values but under physiological conditions, the audible range of tones is likely bounded by minimum and maximum values and can be partitioned (due to JNDs / difference limens) into finitely-many discriminably-different frequencies and amplitudes

...

or worded differently, to the author's liking

* First spike latency comment. Please note: I am not an expert on this particular issue of first spike latency and have no connection to the relevant work. I am following up on this point simply because I am trying to think through the correct interpretation of the model.

In the manuscript the author writes:

[lines 40-44]: This is of some significance because in this short time interval, neurons can fire only bursts of 1-2 action potentials [14, 15], indicating that neurons essentially function as binary units. Therefore, it seems likely that neither frequency nor intensity can be encoded via the firing rate of individual cells since the dynamic range would be severely limited. "

In the letter to reviewers the author wrote:

"the Bizley study showed that first spike latency can be used to discriminate between higher and lower frequency stimuli"

so perhaps it would be appropriate to to modify the phrase "it seems likely that neither frequency nor intensity can be encoded via the firing rate of individual cells" could be modified to say something like.... single spikes may carry some information -- for instance first-spike spike latency may carry some frequency info -- but this frequency info is ambiguous (not invariant to sound pressure level) and likely restricted to narrow ranges of sound pressure level, so for simplicity we exclude this possible source of info in our construction of a place-code model

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #4: No: The paper includes simulation results from a computational model. I do not see it stated if/where that code is available. Apologies if I missed it.

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #4: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

References:

Review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript.

If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009251.r007

Decision Letter 3

Lyle J Graham

6 Jul 2021

Dear Alex, I'm very glad this is in, and thanks for your responses throughout the process. Best, Lyle

Dear Dr. Reyes,

We are pleased to inform you that your manuscript 'Mathematical framework for place coding in the auditory system' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Lyle Graham

Deputy Editor

PLOS Computational Biology

***********************************************************

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009251.r008

Acceptance letter

Lyle J Graham

28 Jul 2021

PCOMPBIOL-D-20-01254R3

Mathematical framework for place coding in the auditory system

Dear Dr Reyes,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Zsofi Zombor

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. Detailed description of mathematical analyses and proofs.

    Fig. A: Projections of multiple layers of staggered neurons on tonotopic axis decreases Δx. Fig. B: Algebra of loudness summation applied to stimuli consisting of 4 tones with equally spaced frequencies.

    (PDF)

    Attachment

    Submitted filename: Response_to_Reveiwers.pdf

    Attachment

    Submitted filename: Responses to Referees_6_6.pdf

    Attachment

    Submitted filename: responses_final.pdf

    Data Availability Statement

    The code for the simulations can be downloaded at https://github.com/AlexDReyes/ReyesPlosComp.git.


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES