Abstract
Bayesian inference and bounded rational decision-making require the accumulation of evidence or utility, respectively, to transform a prior belief or strategy into a posterior probability distribution over hypotheses or actions. Crucially, this process cannot be simply realized by independent integrators, since the different hypotheses and actions also compete with each other. In continuous time, this competitive integration process can be described by a special case of the replicator equation. Here we investigate simple analog electric circuits that implement the underlying differential equation under the constraint that we only permit a limited set of building blocks that we regard as biologically interpretable, such as capacitors, resistors, voltage-dependent conductances and voltage- or current-controlled current and voltage sources. The appeal of these circuits is that they intrinsically perform normalization without requiring an explicit divisive normalization. However, even in idealized simulations, we find that these circuits are very sensitive to internal noise as they accumulate error over time. We discuss in how far neural circuits could implement these operations that might provide a generic competitive principle underlying both perception and action.
Keywords: Analog circuits, Competition, Integration, Bayesian inference, Free energy
Introduction
The competition for limited resources is a central theme in biology. In evolutionary theory, the competition for limited resources enforces the process of natural selection, where differential reproductive success of different genotypes lets some genotypes increase their share in the overall population, while others are driven to extinction [5]. This process can be modeled by the replicator equation that quantifies how the proportion of a particular genotype evolves over time depending on the fitness of all other genotypes, such that genotypes achieving more than the average fitness proliferate, and genotypes that perform below average recede [53, 78]. From a mathematical point of view, each genotype can be considered as a hypothesis that accumulates evidence and where different hypotheses compete for probability mass, since the probabilities of all hypotheses must always sum to unity [66].
Also, ontogenetic processes underlying action and perception are governed by competition for limited resources. Well-known examples of competition include binocular rivalry [72], bistable perception [38], attention [20, 22] or affordance competition for action selection [17]. In particular, the process of perception is often understood as an inference process where sensory ambiguity is resolved by competing “hypotheses” that accumulate evidence on the time scale of several hundred milliseconds [37, 42]. A quantitatively well-studied example is the random-dot motion paradigm [11, 29], where subjects observe a cloud of randomly moving dots with a particular degree of motion coherency before they have to decide whether the majority of dots moved to the right or to the left. Depending on the degree of coherency, this evidence accumulation process proceeds faster or slower. Moreover, in this paradigm, neural responses in sensory cortical areas have been shown to be consistent with encoding of log-odds between different hypotheses, thereby reflecting the competitive nature of the evidence accumulation process [29]. Intriguingly, it can be shown that such a process of competitive evidence accumulation is formally equivalent to natural selection as modeled by the replicator dynamics [31, 66].
The problem of acting can be conceptualized in a similar way as the inference process [25, 36, 57, 70, 71, 73, 76]. An actor chooses between different competing actions and wants to select the action that will bring the highest benefit. Even in the absence of any perceptual uncertainty, an actor with limited information processing capabilities might not be able to select the best action—for example, when planning the next move in a chess game—because the number of possibilities exceeds what the decision-maker can consider in a given time frame. Such bounded decision-makers can sample the action space according to some prior strategy during planning and can only realize strategies that do not deviate too much from their prior [56, 59]. If this deviation is measured by the relative entropy between prior and posterior strategy, then the competition between actions is determined by the accumulated utility of each action in the planning process. In this framework, action and perception can be described by the same variational principle that takes the form of a free energy functional [10, 27, 55, 58].
In this study, we investigate how such competitive accumulation processes could be physically implemented. In particular, we are interested in the design of bio-inspired analog electric circuits that are made of components that are interpretable in relation to possible neural circuits. The components one typically finds in equivalent circuit diagrams of single neurons in textbooks are capacitors, resistors, voltage-dependent conductances and voltage sources [19, 51]. In order to allow for relay of currents between different neurons, we also allow for copy elements implemented by voltage- or current-controlled current sources that have fixed input-output relationships. In the following, we are interested in bio-inspired analog circuit designs that implement free energy optimizing dynamics, but whose components are restricted to this biologically motivated set of building blocks Fig 1. From a biological point of view, the appeal of such circuits is that they intrinsically perform normalization and do not require an explicit computational step for divisive normalization [14]. In particular, we assume in the following that there are a finite number of incoming input streams that are represented by time-dependent physical signals. These signals are accumulated competitively over time by a finite number of integrators that represent a free energy optimizing posterior distribution over the integrated inputs. The aim of the paper is to investigate the biological plausibility of circuit designs for such competitive evidence accumulation where integration and competition are implemented in the same process without the need for a separate process for explicit normalization.
Fig. 1.
Schematic element representations. The set of biologically interpretable building blocks includes capacitors, resistors, controlled current sources and voltage-dependent conductances. The circled V indicates a voltmeter. Arrows on currents indicate polarity. The function f in the variable conductance can represent different mappings depending on the context
Results
The frequency-independent replicator equation
Both Bayesian inference [7] and decision-making with limited information processing resources [58] may be written as an update equation of the following form
| 1 |
where is required for normalization and is a temperature parameter. Equation (1) describes the update from a prior to a posterior . This update can also be formalized as a variational principle in the posterior probability, where
| 2 |
extremizes a free energy functional. In the case of inference, the distribution indicates the prior probability of hypothesis x at time t and represents new evidence that comes in at time t given by the log-likelihood . In this case, the optimization implicit in Eq. (2) underlies approximate Bayesian inference, and in particular variational Bayes methods. In the case of acting, the distribution indicates a prior strategy of sampling actions x and represents the utility gain of choosing action x. In either case, the total utility or total evidence of x changes from a previous state of absolute utility or evidence to a new value of . The subtraction of the free energy difference leads to competition between the different hypotheses or actions x, if the hypothesis or action x gains probability mass, if it loses probability mass; in the case of , the probability mass remains constant.
One of the problems with Eq. (1) is that it is not straightforward how to implement the computation of the normalization . However, in continuous time this computation simplifies to computing an expectation value. In the limit of infinitesimally small time steps in Eq. (1), one arrives at the continuous update equation
| 3 |
Equation (3) is a special case of the replicator equation used in evolutionary game theory to model population dynamics. In evolutionary game theory, the probability p(x, t) indicates the frequency of type x in the total population at time t and corresponds to a fitness function that quantifies the survival success of type x. Types with higher-than-average fitness will proliferate; types with lower-than-average fitness will decline [53]. In contrast to Eq. (3), the general form of the replicator equation has a frequency-dependent fitness function that is the fitness is a function of p(x, t). However, in the following we will only consider the restricted case given by Eq. (3).
In both theoretical and experimental neuroscience, there is an ongoing debate as to whether the brain directly represents uncertainty as probabilities or as log-probabilities [39, 45, 48, 82]. We are therefore also interested in the logarithmic version of Eq. (3). Introducing the new variable
| 4 |
Equation (3) can be written as
| 5 |
Here we have a simple accumulation process U(x, t) with inputs . The advantage of Eq. (5) is that it does not require the explicit computation of the product between p(x, t) and . However, it still requires computing the expectation value that corresponds to the sum over all the products. The probability p(x, t) can be obtained from U(x, t) through the exponential transform at any point in time.
Equivalent analog circuits
In the following, we investigate two classes of bio-inspired analog circuits that implement Eqs. (5) and (3), respectively. In particular, we study differences in circuit design and the robustness properties of the circuits depending on whether uncertainty is represented in probability space or log-space. For each of the two implementations, we consider two input signal scenarios. The input signals can either be represented as currents or voltages to model different kinds of sensory encoding and to study in how far this difference in representation might lead to different circuit designs. Therefore, we consider four different circuits in the following: log-space current input, log-space voltage input, p-space current input and p-space voltage input.
A diagram of the first circuit with log-space representation and current input is shown in Fig. 2. The critical elements in the circuit are the different capacitive accumulators that integrate Eq. (5). As can be seen in Fig. 2a, each accumulator receives two input currents. The first input current I(x, t)—corresponding to in Eq. (5)—is specific for each accumulator. The second input current is the same for all accumulators and is given by corresponding to in Eq. (5). While the input current simply runs through the accumulator unaltered, the input current I(x, t) is fed into a current divider (CD—see Sect. 4 for details) with two branches, one of which connects to ground through a resistor with fixed resistance , while the other branch directs current through a voltage-dependent resistor to generate the weighted output current
| 6 |
Figure 2b shows an exemplary full circuit with three capacitative accumulators. In the full circuit, the output currents of all accumulators as described by Eq. (6) are merged and added up to the total current
| 7 |
The total current is directed as a baseline through all accumulators. Comparing Eq. (7) and the average in Eq. (5) reveals that the probability weights are given by the fraction of resistances determining the non-leaked current.
Fig. 2.

Schematic diagram of the log-space circuit with current inputs. a Capacitive accumulator subcircuit consisting of a primary circuit (black wiring) with a variable resistor that regulates the output current and a secondary circuit (gray wiring) that accumulates the current difference through a capacitor whose voltage adjusts the variable resistor . The input currents of the primary circuit are copied via current-controlled current sources to the secondary circuit. b Complete example circuit for three different accumulators. The individual output currents are combined into . Schematic element representations are shown in Fig. 1
Inside the accumulator, the difference between the two input currents has to be integrated. In order to ensure that the integration process does not alter the input currents themselves by putting an extra load on the input, the integration process has to be electrically isolated, which can be achieved by generating copies of the input currents into a separate circuit. These copies can be generated by two current-controlled current sources that generate copies of the input currents and I(x, t), respectively. The difference between the two currents is then integrated by a capacitor with capacitance , such that
| 8 |
The voltage corresponds to U(x, t) in Eqs. (5) and (4) and the capacitance corresponds to . In line with Eq. (4), the voltage-dependent resistors depend on this voltage through an exponential characteristic line
| 9 |
As long as the voltage represents log-probabilities and therefore assumes values between zero and negative infinity, the resistance is non-negative and well defined. Such a voltage-dependent resistor could be realized by a potentiometer with an exponential characteristic or by using the exponential relationship between current and voltage of a varistor or a transistor.
A diagram of the second circuit with log-space representation and voltage input is shown in Fig. 3. As shown in Fig. 3a, each accumulator is operated between two voltages given by the voltage V(x, t)—corresponding to in Eq. (5)—that is specific for the accumulator x and the voltage that is the same for all accumulators and corresponds to the weighted average in Eq. (5). As the integration has to be performed by a capacitor due to the bio-inspired constraints, the voltages have to be translated into currents, which can be achieved by voltage-controlled current sources. This will also ensure that the integrator circuit is isolated as in the previous circuit and follows the same dynamics as in Eq. (8). For illustrative purposes, the diagram in Fig. 3b shows a complete circuit for three different accumulators. Essentially, the circuit corresponds to a passive averager (PA—see Sect. 4 for details) that combines multiple voltages each in series with a voltage-dependent conductance into a common voltage given by
| 10 |
A comparison between Eq. (10) and the average in Eq. (5) implies that the probability weights p(x, t) are given by the relative conductance. To fit with Eq. (4), the voltage-dependent conductances must therefore have an exponential characteristic line
| 11 |
In contrast to the previous circuit, the conductance is well defined for any value of the integrated voltage. In fact, changing all conductances by the same multiplicative factor does not affect the operation of the circuit.
Fig. 3.

Schematic diagram of the log-space circuit with voltage inputs. a Capacitive accumulator subcircuit consisting of a primary circuit (black wiring) with a variable conductance as part of a passive averager and a secondary circuit (gray wiring) that accumulates the voltage difference through a capacitor and adjusts the conductance accordingly. The voltages of the primary circuit are transformed via voltage-controlled current sources into the currents I(x, t) and of the secondary circuit. b Complete example circuit for three different accumulators. The input voltages V(x, t) drive the passive averager of the primary circuit to produce the weighted average voltage . Schematic element representations are shown in Fig. 1
The third and the fourth circuits represent uncertainty directly in the probability space. Accordingly, they only differ from the previous circuits in terms of the inner workings of the accumulators, as the inputs and the weighted sum are the same in p-space and log-space. Figure 4a shows an accumulator in p-space with external current inputs I(x, t). As in the first circuit, each accumulator has two inputs given by I(x, t) and and one output given by . As in the log-space accumulator, the output corresponds to a weighted input current p(x, t) I(x, t) and this weighting is implemented by a current divider. Identical to the circuit diagram in Fig. 2b, the output currents of all accumulators are merged into the total current that is fed back as an input into the accumulators. In contrast to the log-space accumulator of the first circuit, inside the p-space accumulator the integral has to be taken over the weighted difference between the two input currents , where the weighting with p(x, t) is implemented by another current divider. The voltage-dependent conductances in both current dividers have to be adjusted according to
| 12 |
as the voltage now directly represents p(x, t) and therefore only assumes values in the unit interval. Note that the leak resistances in the two current dividers of each accumulator do not have to be identical, but the variable conductances always have to be adjusted such that the equality
| 13 |
holds for each current divider. The weighted current of the inner current divider is copied by a current-controlled current source and then integrated as a voltage over . The same voltage over the capacitance has to drive both voltage-dependent conductances in both current dividers in order to implement Eq. (3).
Fig. 4.

Schematic diagram of the probability space circuits. Only the accumulators are shown; example circuits are identical to Figs. 2b and 3b, respectively. a Capacitive accumulator subcircuit for current inputs. The accumulator consists of a primary circuit (black wiring) with a variable resistor and a secondary circuit that accumulates the weighted current difference . The additional weighting (compared to the log-space circuit) is accomplished by an inner current divider that operates identical to the outer current divider of the primary circuit. The capacitor integrates the weighted current difference and adjusts the resistors of the outer and inner current dividers accordingly. Another current-controlled current source is required to copy the output of the inner current divider in order to isolate the accumulation process from the rest of the circuit. b Capacitive accumulator subcircuit for voltage inputs. The accumulator consists of a primary circuit (black wiring) with a variable conductance that is critical for the passive averager and a secondary circuit (gray wiring) that accumulates the weighted current difference . The additional weighting (compared to the log-space circuit) is accomplished by an inner current divider that operates identical to the current input circuit in panel A. The capacitor integrates the weighted current difference and adjusts the conductance of the primary circuit and the resistance of the inner current divider accordingly. Schematic element representations are shown in Fig. 1
Figure 4b shows an accumulator in p-space where the external inputs are given as voltages V(x, t). As in the second circuit, each accumulator is operated between the two voltages V(x, t) and . Identical to the circuit diagram in Fig. 3b, voltage is determined by a passive averager across the different accumulators. As in the second circuit, the voltages are transformed into currents when they enter the accumulators by voltage-controlled current sources. The important difference to the log-space accumulator of the second circuit is again that in p-space the integral has to be taken over the weighted difference between the two currents . This weighting is implemented by a current divider in an identical fashion as in the previous circuit shown in Fig. 4a. In this case, the voltage-dependent conductance of the current divider follows Eq. (12) and the voltage-dependent conductance of the passive averager follows a simple proportionality characteristic given by .
Simulations
To test the noise robustness of the circuits shown in Figs. 2, 3, 4, we simulated their dynamics in a Simulink environment with idealized components, which we could selectively perturb by band-limited white noise. Put simply, these simulations are trying to reproduce competition between different streams encoding the evidence for alternative hypotheses without violating the obvious requirement that the resulting probabilities should sum to one. In our examples, we simulated three different time-varying inputs indexed by . The first input was a rectangular pulse of 5 s with an amplitude of A in the circuits with current-based inputs and V in the circuits with voltage-based inputs. The second input was a rectangular pulse of 2.5 s with the same amplitude. The third input was a cosine with amplitude A (or V) and a frequency of 0.19 Hz. The first two inputs mimic the more usual scenario where evidence is increased over a particular time window at a constant rate, whereas the third input is the more unusual scenario with waxing and waning evidence. The first two inputs are integrated into a ramp with different plateaus, and the third input integrates into a sine wave. The input signals and their integrals are shown in Fig. 5.
Fig. 5.
Input signals for simulation. a Time course of the three integrated signals W(x, t) indexed by . The signals represent the evidence of a particular hypothesis or the utility of a particular action. For and , the evidence or utility grows with a constant rate until saturation is reached. For , the evidence or utility is waxing and waning, following a sine function. Note that higher values correspond to more probable hypotheses or more desirable actions. A rational decision-maker following Eq. (19) should thus initially favor . After the onset of and , the decision-maker should be indifferent between these two options but should disfavor . After reaches saturation, should be favored. b Inputs which are fed into the circuits as either currents or voltages and drive the competitive integration process
We simulated all four circuits shown in Figs. 2, 3, 4 under three noise conditions. As a baseline, we first simulated all circuits without noise and plotted the probability encoded by the voltage of the integrating capacitors. In case of the log-space circuits, this corresponds to the exponential of the voltage. In the p-space circuits, the voltage directly encodes probability. This can be seen in the first column of Fig. 6. In the first 2 s, the cosine signal has the highest amplitude and therefore the highest probability weight, before it is overtaken by the onset of the two pulses. Eventually, the longer lasting pulse dominates the probability weighting.
Fig. 6.

Robustness simulation for all circuits. The top two rows show the simulated probabilities for the log-space circuits—first row current input, second row voltage input. The bottom two rows show the simulated probabilities for the probability space circuits—first row current input, second row voltage input. The first column shows the results for a noise-free simulation, where all circuits perform identically and consistent with the replicator equation. The second column shows the results where band-limited white noise was injected into the copy elements, that is, the current- or voltage-controlled current sources. The magnitude of the noise was identical for all circuits. The errors with respect to the corresponding noise-free simulation are shown in the third column. The fourth column shows the results where band-limited white noise was injected into the voltage-dependent resistors. Again the magnitude of the noise was identical for all circuits. The errors with respect to the corresponding noise-free simulation are shown in the last column
In the second noise condition, we added band-limited white noise on the output currents of all copying elements, that is, all voltage- or current-controlled current sources. The standard deviation of the noise was and roughly corresponded to three orders of magnitude below the maximum input signal. As can be seen from the simulation in the second and third column of Fig. 6, the noise has very different effects in the p-space and log-space circuits. While the log-space circuits show errors on the order of percentages in probability space, the p-space circuits fail and completely leave the range of permissible probability values. The circuit element in the p-space circuit that is responsible for this failure is the current-controlled current source that directly feeds into the integrator. The ultimate reason for this difference is of course that the integrator voltage in the p-space circuit is confined between zero and one, whereas the integrator voltage in the log-space circuit can take any negative value.
In the third noise condition, we added band-limited white noise on the resistance of the voltage-dependent conductances. The standard deviation of the noise was . According to Eqs. (9) and (12), the voltage-dependent resistance in the circuits with current inputs can decrease to almost zero for dominant inputs, but can take on values up to in our simulations. In contrast, the voltage-dependent resistance in the circuits with voltage inputs do not have to regulate their resistance down to zero, but to an arbitrary baseline resistance—because this baseline resistance cancels out in the passive averager. Accordingly, one would expect the most disruptive effects of noise for dominant inputs with high probability weighting, but less so in the case of passive averager circuits that operate on voltage inputs. In Fig. 6, it can be seen that the noise-corrupted probabilities in the passive averager circuits are much smoother for high probability weightings than in the current divider circuits. However, there seems to be no difference in the magnitude of the errors.
Implications for possible neural circuits
As already mentioned, there is an ongoing debate about whether the brain directly represents uncertainty as probabilities or as surprise, that is, log-probabilities. In the previous section, we have considered both possibilities in different circuit designs. As illustrated in Fig. 7, these bio-inspired analog circuit designs can serve as abstract templates for schematic neural circuits. Figure 7a shows a free energy optimizing neural circuit operating in log-space—compare Eq. (5). Input signals are excitatory and integrated by accumulator neurons that are inhibited at the same time by a pooled inhibition signal. To establish this inhibition signal, copies of all inputs are summed up by an inhibitory neuron that sends its signal to all accumulator neurons. The most critical operation in this circuit would require that the output signal U of the accumulator neuron modulates the weighting of the input signal before it enters the inhibitory unit. Moreover, this modulation of the input signal would have to correspond to a multiplicative weighting where the weighting factor is characterized by an exponential dependency on the excitatory output signal, such that the modulated input to the inhibitory neuron is given by . As U is the log-probability and therefore always negative, the weighting factor could also be interpreted in terms of a synaptic transmission probability that is modulated by the signal U.
Fig. 7.
Neural circuit diagrams for competitive signal integration. White triangles represent excitatory units corresponding to different accumulators x. Gray triangles correspond to inhibitory units. a Replicator dynamics in log-space according to Eq. (5). The little boxes with arrows denote a multiplicative modulation with an exponential characteristic. b Replicator dynamics in probability space according to Eq. (3). The little boxes with arrows denote a multiplicative modulation. c Mutual inhibition circuit. The little boxes with arrows denote a fixed weighting. d Pooled inhibition circuit. The little boxes with arrows denote a fixed weighting. e Feed-forward inhibition. The little boxes with arrows denote a fixed weighting
Figure 7b shows a schematic of a free energy optimizing neural circuit in probability space—compare Eq. (3). The basic principle of the circuit is the same as in Fig. 7a. The important differences between the p-space and log-space neural circuits are the following. First, the output of the accumulator neurons represents a probability p instead of the log-probability U. Second, each accumulator modulates its own inputs by a multiplicative factor given by the output activity p—this concerns both the excitatory input and the inhibitory input . Third, all multiplicative modulations are characterized by weighting factors that are proportional in p and not exponential as in the log-space case. Overall, the p-space circuit is more complex with nested recurrencies that require the simultaneous modulation of multiple sites in dependence of the same signal p.
In the literature, the dynamics of neural circuits for competitive signal integration are often modeled by drift diffusion processes [8, 9, 12, 35]. In these models, momentary evidence modulates the drift in a Brownian motion process. Mainly, four different kinds of drift diffusion models are distinguished: race models [75], mutual inhibition models [74, 79], feed-forward inhibition models [47, 65] and pooled inhibition models [77, 80]. Race models consist of independent accumulators without any inhibitory interactions and can therefore be disregarded in this context. We consider the other three inhibition models in the following. Linearized mutual inhibition models may be described by the dynamics
| 14 |
where denotes activity of accumulator i, k is a self-inhibition factor, w is the inhibitory weighting factor between the different neurons and is the input signal. The corresponding circuit is shown in Fig. 7c. Similarly, one can express the simplified dynamics of a pooled inhibition model as
| 15 |
where all neurons contribute equally to the global inhibitory signal. The corresponding circuit is shown in Fig. 7d. In contrast, feed-forward inhibition models only modulate their activity depending on the inputs I, such that
| 16 |
where w indicates the inhibitory effect of input on accumulator i. In this case, each input has connections with all accumulators, of which all but one are inhibitory. The corresponding circuit is shown in Fig. 7e.
As the input only enters additively in Eqs. (14)–(16), it makes sense to compare these models to the log-space circuit in Fig. 7a. The most obvious difference of the log-space circuit in Fig. 7a from all the other circuits listed above is that inhibition depends on both the inputs I and the neural activity U such that
| 17 |
There is no self-inhibition in these dynamics, as the introduction of a decay term would compromise the normalization of . Note that none of the other accumulators is normalized and therefore a separate normalization step is required. Comparing Eqs. (17)–(14), (15) and (16) raises the question of how accumulator dynamics of Eqs. (14)–(16) could approximate dynamics of the form of Eq. (17) that are required for Bayesian inference and bounded rational decision-making.
Important differences between the dynamics of Eqs. (14)–(17) can be illustrated by considering constant inputs I. For constant inputs, Eqs. (14)–(16) reach steady-state attractors where for all i. In contrast to these three inhibition models, update Eq. (17) does not reach a steady state unless all inputs are the same, that is, . This is the case, for example, when all hypotheses have the same likelihood or when all actions lead to the same increase in utility and therefore the posterior probabilities simply equal the prior probabilities. If the inputs are not the same in Eq. (17), we have the limit behavior if otherwise . The exponential is always bounded by one of the asymptotes 0 or 1. This difference between the models with respect to their limit behavior originates from the presence or absence of the decay term . If this term is omitted in the other models, the can also grow without bound both in the positive and negative direction. However, there are also important differences between the models even if we disregard the decay term . For constant inputs, the mutual inhibition model exhibits exponential growth. In contrast, the feed-forward inhibition model always has a constant growth rate and the pooled inhibition model converges to constant growth rates. The free energy update Eq. (17) also converges to constant growth rates and is therefore qualitatively most similar to the pooled inhibition model. However, both their modulation of the growth rates through the dynamics of and before convergence and the limit values of and differ.
Here we have focused on evidence accumulation with a finite number of accumulators where each accumulator corresponds to a different hypothesis. This corresponds to the scenario that is usually considered by evidence accumulation models based on drift diffusion processes [9, 49]. An obvious question is of course how to generalize this kind of setup to continuous hypothesis spaces. For particular families of distributions, like for example Gaussian distributions, one can replace an infinite number of accumulators by a finite number of sufficient statistics, for example mean and variance in the case of the Gaussian. This is exploited for example in Kalman filters and some predictive coding models [3, 26, 62]. Other possibilities include representing uncertainty through gain encoding or through convolutional codes with a finite number of basis functions [39]. Due to the many possibilities how one could think about a continuous generalization, we restrict ourselves to discrete states in the current study.
Discussion
In this study, we have described four bio-inspired analog circuit designs implementing the frequency-independent replicator equation. The frequency-independent replicator equation optimizes a free energy functional and can be used to describe both competitive evidence accumulation for perception and utility accumulation for action. The bio-inspired circuits differed in whether they implemented the frequency-independent replicator equation directly in probability space or in log-probability space and in whether the input signal was given as a voltage or as a current. The circuits were designed under the constraint that they should only consist of a restricted set of electrical components that are biologically interpretable in the sense that such components are commonly used when neural circuitry is schematized by equivalent electrical circuits. Accordingly, we sketch how the two basic circuit designs for free energy optimization in probability and log-probability space might translate into neural wiring in Fig. 7. Here we discuss the biological plausibility of these circuits.
Biological plausibility
In standard textbooks [51], neurons are usually modeled as capacitors that integrate currents over time and that have synapses and ion channels that can change their conductance depending on voltage. Also the neural integrators in our circuits are modeled as capacitors. The basic design of the circuits in Fig. 7 implies that each neural integrator receives both excitatory and inhibitory inputs. As all neural integrators receive the same inhibitory input, it is natural to assume that the inhibitory signals stem from a single inhibitory unit that pools copies of all the excitatory inputs and feeds the resulting inhibitory signal back to the neural integrators. This would imply, however, that inhibitory neurons mainly perform a spatial integration, whereas excitatory neurons would mainly perform a temporal integration. Accordingly, the inhibitory neurons would have to compute their output quasi-instantaneously compared to the time scales of the input. This very same problem is faced by all pooled inhibition models. Here the particular challenge of the circuit diagrams shown in Fig. 7 is the temporal dependence of the weights for averaging, as the probability weights would have to change on the same time scale as the inhibitory output activity that is quasi-instantaneously with respect to the time scales of the input signal.
As already described in the previous section, the critical operation in the free energy circuits is performed by the voltage-dependent conductances that regulate how much of any particular input signal reaches the inhibitory unit. In particular, in the log-space circuits it would be required that there is an exponential relation between the voltage signal and the resulting conductance or transmission probability. This would be a very particular property to look for in possible neural substrates. In contrast, this exponential relationship is not required in the p-space circuits. However, their biological plausibility suffers from two other deficiencies. First, the circuit design is considerably more complex than the log-space circuit design in that it requires multiple replications of the same voltage-dependent conductances that not only modulate the inputs to the inhibitory units, but also the inputs to the excitatory neural integrators. Second, as is evident from the simulations, the p-space circuits are extremely susceptible to noise.
Another implementation challenge of Eq. (5) is that the most unlikely hypotheses or actions require the accumulation signal U with the strongest magnitude that is the highest currents and the highest voltages. This is a natural consequence of operating in the log-domain, where unexpected events are assigned the most resource-intense encoding, such that expected events can be encoded more efficiently [46]. However, when implementing Eq. (5) in a real physical system this problem could be solved naturally, as any physical signal will have a natural minimum and maximum that is technically feasible. For example, if the physical signal that is used for representation of has a natural limit between zero and a minimum , that is , then the probability is confined to the interval with a minimum nonzero probability 1 / L assigned to any hypothesis or action. In order to deal with exclusively positive signal ranges, one can also redefine the representation as which implies . This redefined representation has the convenient effect that improbable hypotheses or actions are no more associated with the highest signal magnitude, but with the lowest. Similar cutoffs in the precision of probability representations are ubiquitous in Bayesian statistics, for example in the context of Cromwell’s rule or Occam’s Window.
Divisive normalization
As an alternative to modeling a single process that accomplishes signal integration and competition simultaneously, one could imagine a model where signal integration and competition are dealt with in separate stages of the process or even as two separate processes or mechanisms. The integration process does not pose any particular problem, but simply corresponds to independent integration processes of individual excitatory signals. In an analog circuit, this would correspond as usual to a capacitor that integrates currents into a voltage. The competition between the different integrated signals can then be introduced after integration by the application of a softmax function
| 18 |
where is the same temperature parameter as in Eq. (1) and W(x, t) is the integrated signal . This is the mathematical operation of divisive normalization. For example, Bayesian inference could be achieved in log-space by such a two-step process, where first log-likelihoods are integrated or added up over time and in a second step the summed or integrated signals are squashed through a softmax function [48]. Importantly, Eq. (18) optimizes the free energy functional of Eq. (2) under uniform priors. Nonuniform priors can be included to yield
| 19 |
where can be interpreted as the initial state of the accumulator x. While there exist analog implementations of the softmax function [24, 40, 83], these implementations have a circuit design that is not easily interpretable in terms of equivalent neural circuits. For example, the circuits in [24] enforce a constant output current that is additively composed of drain currents from multiple transistors that are controlled by exponentially weighted gate voltages. The softmax function is computed by the individual drain currents. However, in a biological setting a constant output current that drives the integration process is not plausible. Nevertheless, other implementations might be possible.
In neuroscience, divisive normalization has been advanced as a fundamental neural computation over the last two decades [14]. It has been suggested as a normalization mechanism to regulate stimulus sensitivity in the invertebrate olfactory system [54], the mammalian retina [52], primary visual cortex [13, 15] and other cortical areas [32]. However, the biophysical mechanisms and possible circuit designs that would support divisive normalization are still under debate [14]. One of the earliest proposed mechanisms for divisive normalization is shunting inhibition mediated by synapses that cause a change in membrane conductance without a major change in current flow [63]. For constant input, however, shunting only has a divisive effect on the membrane potential in integrate-and-fire models of neural activity, but not on the firing rate of these neurons [33]. This has led to the more recent proposal that shunting might be achieved by temporally varying changes in conductance [16, 69]. However, physiological evidence for this mechanism remains mixed [14]. Other proposed physiological mechanisms that could mimic divisive normalization at least for some experimental data are synaptic depression and modulation of ongoing activity to keep membrane potentials closer or further from the spiking threshold [1, 64]. As divisive normalization seems to play such a prominent role in biological information processing, our circuits might inspire an interesting alternative that does not require a separate mechanism for normalization, but a single process that automatically generates normalized signals. However, as discussed in the previous section the biological plausibility of these circuits is certainly also open for debate.
Circuits for Bayesian integration
Several hardware implementations of inference processes have been proposed in the recent past [41, 50, 81]. The implementation of continuous-time Bayesian inference in analog CMOS circuits, for example, has been recently discussed by Mroszczyk and Dudek [50]. The authors investigate message passing inference schemes in Bayesian networks that consist of multiple variables that factorize. The analog implementation they propose is based on the Gilbert multiplier that is seconded by transistor circuits such that the overall multiplier circuit can normalize incoming current signals. While these circuits are technologically optimized for accuracy and scalability, the building blocks of these circuits make a biological interpretation difficult. At the other end of the spectrum of biological realism, VLSI implementations of spiking neural networks for real-time inference have been recently proposed [18]. In contrast, the current study does not reach the neuromorphic realism of spiking networks, but starts out by addressing the question of how free energy optimizing dynamics could be implemented in circuits that allow for some degree of biological interpretation. While the direct implementation of such circuits seems to have received little attention so far, some special cases of the general replicator equation that correspond to the Lottka–Volterra equation have been implemented in VLSI to better understand competitive neural networks [2]. However, the equivalence between the replicator equation and the Lottka–Volterra equations does not hold for the frequency-independent replicator equation and therefore does not concern our results.
Bayesian inference has been proposed as a fundamental theory of perception and a considerable number of different neural implementations of inference processes have been proposed in the recent past [4, 21, 39, 43, 44, 61]. However, one might regard Bayesian inference as a particular instantiation of a more abstract optimization principle given by the free energy difference in Eq. (2), when the utility is given by a log-likelihood [25, 58]. Intriguingly, the same principle can be generalized to the problem of acting. A decision-maker starts out with a prior strategy and considers different options with different utilities. When the set of options is large, it might be impossible to consider all of them, such that the decision-maker has to make a decision after sampling a few possibilities [56, 59]. Making a decision based on these samples, the decision-maker effectively follows a probabilistic strategy that can be described by the posterior distribution in Eq. (1) optimizing a trade-off between utility gain and computational cost. The computational cost is measured by the relative entropy between prior and posterior strategy. Compared to a perfect decision-maker, such a decision-maker is bounded rational since he can only afford a limited amount of information processing. The principle issues of the proposed circuitries might therefore be applicable both to perception and action.
One of the main problems of implementing Bayesian integration is the issue of tractability, which often arises due to the computation of the normalization constant, especially when integrating over high-dimensional parameter spaces, but also, for example when summing over discrete states in larger size undirected graphical models. One way to deal with this kind of problem is to investigate stochastic and sampling-based approximations of probabilistic update schemes [21, 28, 34, 60, 67, 68]. Here we were not primarily interested in such stochastic implementations, because like many previous studies we were interested in circuits that integrate a finite number of given inputs and do not probabilistically ignore some inputs. Naturally, our circuits then do not provide a generic solution to Bayesian inference in arbitrary networks, but rather we have restricted ourselves to the special case of competitive evidence accumulation with a finite number of given inputs. If such input streams are given in terms of physical signals, then computing a weighted average by summing over these signals is certainly a tractable operation. Even though such competitive signal integration is equivalent to a Bayesian inference process [8], if one were interested in generic Bayesian inference in possibly continuous and high-dimensional parameter spaces, one would certainly need to consider some kind of approximation to Eqs. (1) and (3)—see for example [56] for a sampling-based implementation.
Circuits for free energy optimization
Free energy optimization has been studied previously in Hopfield networks in the context of memory retrieval and in Boltzmann machines in the context of learning generative probabilistic models. Both Hopfield networks and Boltzmann machines can be described by the same kind of energy function; only the dynamics of the latter are stochastic. The energy function
specifies the desirability of the binary state of all neurons i in the network with under given parameters and . In both networks, the dynamics minimize this energy function, which corresponds to a relaxation process into an equilibrium distribution over states s. Thus, the free energy does not play a direct role in the dynamics. However, if one restricts the class of equilibrium distributions to special classes of parameterized separable distributions , then one can optimize the variational free energy
to find the distribution that most closely matches the equilibrium distribution . In the case of Hopfield networks, this leads for example to a mean field approximation—compare Chapter 42 in [46].
Apart from the dynamics that govern the state evolution in these networks, there are also update rules that determine the parametric weights and of the networks during learning [6]. In Boltzmann machines with hidden units h, the equilibrium distribution over observable states x is given by where and is a normalization constant. Using the free energy
the equilibrium distribution can be expressed as a Boltzmann distribution with normalization constant . Learning a generative model for x can then be achieved by updating the parameters with the gradient
and similarly for the parameters . Crucially, however, such learning updates change the energy function itself by optimizing the log-likelihood of the data. Challenges of physical implementations of Boltzmann machines have been discussed in [23].
Both kinds of free energy updates are not directly relevant to our study, as neither the Hopfield network nor the Boltzmann machine can be used to optimize arbitrary free energy functions for competitive evidence accumulation. Both network types have been designed to solve completely different problems—i.e., memory retrieval and generative model learning. The energy function in these networks describes a particular recurrent network dynamics and does not constitute an external signal that is integrated over time. In contrast, in our circuits we study possible implementations of the evolution of the posterior distribution for decision-making or inference resulting from the temporal integration of a time-dependent external input signal. Unlike the distributions in Hopfield and Boltzmann machines that relax to equilibrium, the posterior in our implementations is an equilibrium distribution at any point in time as long as it follows the dynamics of Eq. (3). In any physical implementation, this can of course only be approximately true as long as the time scale of the input signal is slow compared to physical delays, etc. In the future, it might therefore also be interesting to study non-equilibrium systems for decision-making and inference [30].
Materials and methods
All simulations were performed using the Simscape library of MATLAB R2012b Simulink. We simulated all circuits using the numerical solver ode15s.
In the log-space circuit with current inputs shown in Fig. 2, the weighting with p(x, t) is performed by using current dividers (CD) with variable resistors. A schematic diagram of the basic current divider principle is shown in Fig. 8b. To compute the weighted average as given by Eq. (7), each input current I(x, t) in Fig. 2b is fed into a current divider that outputs the weighted current . These currents are then summed up by connecting the current dividers into a common point which produces . To ensure proper operation of the circuit, the voltage-dependent resistors of the current dividers have to be precisely set according to Eq. (9). To simulate the log-space circuit with current inputs shown in Fig. 2, we used the following components. We set F for the integrators which corresponds to given the magnitudes of the input currents shown in Fig. 5. The integrator capacitors are initialized with V corresponding to . The voltage-dependent resistor is simulated as
| 20 |
with for the fixed resistor of the current divider.
Fig. 8.
a Passive averager (PA) circuit. The output voltage is the weighted average of the voltages and , where the weights are given by the resistors and . b Current divider (CD) circuit. The input current is divided into two currents over the two resistors and where the magnitude of the current that flows through each branch is proportional to the conductance of each branch
The probability space circuit with current inputs shown in Fig. 4a is very similar to the log-space implementation of Fig. 2. The major distinction is that in the probability space circuit the difference is weighted with p(x, t) to form the accumulated current , whereas in the log-space circuit the difference is directly integrated without an additional weighting. The additional weighting in the probability space circuit is accomplished with an inner current divider that operates identical to the outer current divider, both of which are adjusted according to Eq. (12):
| 21 |
with for the fixed resistors of the current dividers. Note that in the probability space circuits directly corresponds to p(x, t). Therefore, the integrator capacitors are initialized with corresponding to . The capacitance F is set as in the previous circuit.
In the log-space circuit with voltage inputs shown in Fig. 3, the weighting with p(x, t) and summation over all x are performed simultaneously by using a passive averager (PA) with variable conductances. A schematic diagram of the basic passive averager principle is shown in Fig. 8a. To compute the weighted average as given by Eq. (10), the input voltages V(x, t) in Fig. 3b are combined through a passive averager that produces the weighted voltage . To ensure proper operation of the circuit, the voltage-dependent conductances of the passive averager have to be precisely set according to Eq. (11). To simulate the log-space circuit with voltage inputs shown in Fig. 3, we used the following components. We set F for the integrators which corresponds to given the magnitudes of the input currents shown in Fig. 5. The integrator capacitors are initialized with V corresponding to . The voltage-dependent conductances are simulated as
| 22 |
The probability space circuit with voltage inputs shown in Fig. 4b is very similar to the log-space implementation of Fig. 3. The major distinction is that in the probability space circuit the difference is weighted with p(x, t) to form the accumulated current , whereas in the log-space circuit the difference is directly integrated without an additional weighting. The additional weighting in the probability space circuit is accomplished with an inner current divider that operates identical to the inner current divider of the probability space circuit with current input and is adjusted according to Eq. (12):
| 23 |
with for the fixed resistor of the current divider. The outer weighting operation is performed with a passive averager, and thus, the voltage-dependent conductances have to be set according to:
| 24 |
Note that in the probability space circuits directly corresponds to p(x, t). Therefore, the integrator capacitors are initialized with corresponding to . The capacitance F is set as in the previous circuit.
In order to illustrate the robustness of the circuits against perturbations, we injected noise into the copying elements—i.e., the voltage- and current-controlled current sources—and into the voltage-dependent resistors or conductances. As a noise source, we used the Simulink band-limited white noise block which allows to introduce band-limited white noise into a continuous system. We set the parameters of the block to the following values: Noise Power and Sample Time (see Simulink documentation for more information). Additionally, we scaled the output of the white noise source with a constant multiplicative factor. In order to inject noise into the copy elements we controlled a current source with the white noise block and a scaling factor of and injected the output as an additive current to the current output of all controlled current sources. In order to inject noise into the voltage-dependent resistors, we used the white noise block and a scaling factor of 50 and injected the output as an additive component to the setting of the resistance value. Additionally, we limited the minimum value of the resistances to .
Acknowledgments
Open access funding provided by Max Planck Society (Administrative Headquarters of the Max Planck Society).
Footnotes
This study was supported by the DFG, Emmy Noether Grant BR4164/1-1.
Contributor Information
Tim Genewein, Email: tim.genewein@tuebingen.mpg.de.
Daniel A. Braun, Email: daniel.braun@tuebingen.mpg.de
References
- 1.Abbott L, Varela J, Sen K, Nelson S. Synaptic depression and cortical gain control. Science. 1997;275(5297):221–224. doi: 10.1126/science.275.5297.221. [DOI] [PubMed] [Google Scholar]
- 2.Asai T, Ohtani M, Yonezu H. Analog integrated circuits for the Lotka–Volterra competitive neural networks. IEEE Trans Neural Netw. 1999;10(5):1222–1231. doi: 10.1109/72.788661. [DOI] [PubMed] [Google Scholar]
- 3.Bastos AM, Usrey WM, Adams RA, Mangun GR, Fries P, Friston KJ. Canonical microcircuits for predictive coding. Neuron. 2012;76(4):695–711. doi: 10.1016/j.neuron.2012.10.038. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Beck JM, Ma WJ, Kiani R, Hanks T, Churchland AK, Roitman J, Shadlen MN, Latham PE, Pouget A. Probabilistic population codes for bayesian decision making. Neuron. 2008;60(6):1142–1152. doi: 10.1016/j.neuron.2008.09.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Begon M, Townsend CR, Harper JL. Ecology: from individuals to ecosystems. Hoboken: Wiley; 2006. [Google Scholar]
- 6.Bengio Y. Learning deep architectures for AI. Found Trends Mach Learn. 2009;2(1):1–127. doi: 10.1561/2200000006. [DOI] [Google Scholar]
- 7.Bishop C. Pattern recognition and machine learning. New York: Springer; 2006. [Google Scholar]
- 8.Bitzer S, Park H, Blankenburg F, Kiebel SJ (2014) Perceptual decision making: drift-diffusion model is equivalent to a bayesian model. Frontiers in Human Neuroscience 8(102) . doi:10.3389/fnhum.2014.00102 [DOI] [PMC free article] [PubMed]
- 9.Bogacz R, Brown E, Moehlis J, Holmes P, Cohen J. The physics of optimal decision making: a formal analysis of models of performance in two-alternative forced-choice tasks. Psychol Rev. 2006;113:700–765. doi: 10.1037/0033-295X.113.4.700. [DOI] [PubMed] [Google Scholar]
- 10.Braun DA, Ortega PA, Theodorou E, Schaal S (2011) Path integral control and bounded rationality. In: IEEE Symposium on adaptive dynamic programming and reinforcement learning, pp 202–209
- 11.Britten K, Shadlen M, Newsome W, Movshon J. The analysis of visual motion: a comparison of neuronal and psychophysical performance. J Neurosci. 1992;12:4745–4767. doi: 10.1523/JNEUROSCI.12-12-04745.1992. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Busemeyer JR, Diederich A. Survey of decision field theory. Math Soc Sci. 2002;43(3):345–370. doi: 10.1016/S0165-4896(02)00016-1. [DOI] [Google Scholar]
- 13.Carandini M, Heeger DJ. Summation and division by neurons in primate visual cortex. Science. 1994;264(5163):1333–1336. doi: 10.1126/science.8191289. [DOI] [PubMed] [Google Scholar]
- 14.Carandini M, Heeger DJ. Normalization as a canonical neural computation. Nat Rev Neurosci. 2011;13(1):51–62. doi: 10.1038/nrn3136. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Carandini M, Heeger DJ, Movshon JA. Linearity and normalization in simple cells of the macaque primary visual cortex. J Neurosci. 1997;17(21):8621–8644. doi: 10.1523/JNEUROSCI.17-21-08621.1997. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Chance FS, Abbott L, Reyes AD. Gain modulation from background synaptic input. Neuron. 2002;35(4):773–782. doi: 10.1016/S0896-6273(02)00820-6. [DOI] [PubMed] [Google Scholar]
- 17.Cisek P. Cortical mechanisms of action selection: the affordance competition hypothesis. Philos Trans R Soc B Biol Sci. 2007;362(1485):1585–1599. doi: 10.1098/rstb.2007.2054. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Corneil D, Sonnleithner D, Neftci E, Chicca E, Cook M, Indiveri G, Douglas R (2012) Real-time inference in a vlsi spiking neural network. In: IEEE International symposium on circuits and systems (ISCAS), 2012, pp 2425–2428. doi:10.1109/ISCAS.2012.6271788
- 19.Dayan P, Abbott LF. Theoretical neuroscience. Cambridge: MIT Press; 2001. [Google Scholar]
- 20.Desimone R, Duncan J. Neural mechanisms of selective visual attention. Annu Rev Neurosci. 1995;18:193–222. doi: 10.1146/annurev.ne.18.030195.001205. [DOI] [PubMed] [Google Scholar]
- 21.Doya K, editor. Bayesian brain: probabilistic approaches to neural coding. Cambridge: MIT Press; 2007. [Google Scholar]
- 22.Driver J. A selective review of selective attention research from the past century. Br J Psychol. 2001;92:53–78. doi: 10.1348/000712601162103. [DOI] [PubMed] [Google Scholar]
- 23.Dumoulin V, Goodfellow IJ, Courville A, Bengio Y (2013) On the challenges of physical implementations of rbms. arXiv preprint arXiv:1312.5258
- 24.Elfadel IM, Wyatt Jr JL 1993 The “softmax” nonlinearity: derivation using statistical mechanics and useful properties as a multiterminal analog circuit element. In: Advances in neural information processing systems 6 (NIPS 1993), Denver, Colorado, USA, pp. 882–887
- 25.Friston K. The free-energy principle: a rough guide to the brain? Trends Cognit Sci. 2009;13:293–301. doi: 10.1016/j.tics.2009.04.005. [DOI] [PubMed] [Google Scholar]
- 26.Friston K, Kiebel S. Predictive coding under the free-energy principle. Philos Trans R Soc Lond B Biol Sci. 2009;364(1521):1211–1221. doi: 10.1098/rstb.2008.0300. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Genewein T, Leibfried F, Grau-Moya J, Braun DA. Bounded rationality, abstraction and hierarchical decision-making: an information-theoretic optimality principle. Front Robot AI. 2015 [Google Scholar]
- 28.Gershman SJ, Vul E, Tenenbaum JB. Multistability and perceptual inference. Neural Comput. 2012;24(1):1–24. doi: 10.1162/NECO_a_00226. [DOI] [PubMed] [Google Scholar]
- 29.Gold JI, Shadlen MN. Neural computations that underlie decisions about sensory stimuli. Trends Cognit Sci. 2001;5(1):10–16. doi: 10.1016/S1364-6613(00)01567-9. [DOI] [PubMed] [Google Scholar]
- 30.Grau-Moya J, Braun DA (2013) Bounded rational decision-making in changing environments. In: NIPS 2013 workshop on planning with information constraints arXiv:1312.6726 [DOI] [PMC free article] [PubMed]
- 31.Harper M (2009) The replicator equation as an inference dynamic. arXiv preprint arXiv:0911.1763
- 32.Heeger DJ, Simoncelli EP, Movshon JA. Computational models of cortical visual processing. Proc Natl Acad Sci. 1996;93(2):623–627. doi: 10.1073/pnas.93.2.623. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Holt GR, Koch C. Shunting inhibition does not have a divisive effect on firing rates. Neural Comput. 1997;9(5):1001–1013. doi: 10.1162/neco.1997.9.5.1001. [DOI] [PubMed] [Google Scholar]
- 34.Hoyer PO, Hyvarinen A (2002) Interpreting neural response variability as monte carlo sampling of the posterior. In: Advances in neural information processing systems 15 (NIPS2002), Vancouver, British Columbia, Canada, pp. 277–284
- 35.Insabato A, Dempere-Marco L, Pannunzi M, Deco G, Romo R. The influence of spatiotemporal structure of noisy stimuli in decision making. PLoS Comput Biol. 2014;10(4):e1003,492. doi: 10.1371/journal.pcbi.1003492. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Kappen H, Gómez V, Opper M. Optimal control as a graphical model inference problem. Mach Learn. 2012;1:1–11. [Google Scholar]
- 37.Kersten D, Mamassian P, Yuille A. Object perception as Bayesian inference. Annu Rev Psychol. 2004;55:271–304. doi: 10.1146/annurev.psych.55.090902.142005. [DOI] [PubMed] [Google Scholar]
- 38.Kim C, Blake R. Psychophysical magic: rendering the visible ‘invisible’. Trends Cognit Sci. 2005;9:381–388. doi: 10.1016/j.tics.2005.06.012. [DOI] [PubMed] [Google Scholar]
- 39.Knill DC, Pouget A. The bayesian brain: the role of uncertainty in neural coding and computation. Trends Neurosci. 2004;27(12):712–719. doi: 10.1016/j.tins.2004.10.007. [DOI] [PubMed] [Google Scholar]
- 40.Liu SC, Liu SC, Liu SC (1999) A winner-take-all circuit with controllable soft max property. In: NIPS pp 717–723
- 41.Loeliger HA, Lustenberger F, Helfenstein M, Tarkoy F. Probability propagation and decoding in analog VLSI. IEEE Trans Inf Theory. 2001;47(2):837–843. doi: 10.1109/18.910594. [DOI] [Google Scholar]
- 42.Ma WJ. Organizing probabilistic models of perception. Trends Cognit Sci. 2012;16:511–518. doi: 10.1016/j.tics.2012.08.010. [DOI] [PubMed] [Google Scholar]
- 43.Ma WJ, Beck JM, Latham PE, Pouget A. Bayesian inference with probabilistic population codes. Nat Neurosci. 2006;9(11):1432–1438. doi: 10.1038/nn1790. [DOI] [PubMed] [Google Scholar]
- 44.Ma WJ, Beck JM, Pouget A. Spiking networks for bayesian inference and choice. Curr Opin Neurobiol. 2008;18(2):217–222. doi: 10.1016/j.conb.2008.07.004. [DOI] [PubMed] [Google Scholar]
- 45.Ma WJ, Jazayeri M. Neural coding of uncertainty and probability. Annu Rev Neurosci. 2014;37(1):205–220. doi: 10.1146/annurev-neuro-071013-014017. [DOI] [PubMed] [Google Scholar]
- 46.MacKay D. Information theory, inference, and learning algorithms. Cambridge: Cambridge University Press; 2003. [Google Scholar]
- 47.Mazurek ME, Roitman JD, Ditterich J, Shadlen MN. A role for neural integrators in perceptual decision making. Cereb Cortex. 2003;13(11):1257–1269. doi: 10.1093/cercor/bhg097. [DOI] [PubMed] [Google Scholar]
- 48.McClelland JL (2013) Integrating probabilistic models of perception and interactive neural networks: a historical and tutorial review. Front Psychol 4(503). doi:10.3389/fpsyg.2013.00503 [DOI] [PMC free article] [PubMed]
- 49.McMillen T, Holmes P. The dynamics of choice among multiple alternatives. J Math Psychol. 2006;50(1):30–57. doi: 10.1016/j.jmp.2005.10.003. [DOI] [Google Scholar]
- 50.Mroszczyk P, Dudek P (2014) The accuracy and scalability of continuous-time bayesian inference in analogue cmos circuits. In: IEEE international symposium on circuits and systems (ISCAS), 2014, pp 1576–1579. doi:10.1109/ISCAS.2014.6865450
- 51.Nicholls JG, Martin AR, Wallace BG, Fuchs PA. From neuron to brain. Sunderland: Sinauer Associates Sunderland; 2001. [Google Scholar]
- 52.Normann RA, Perlman I. The effects of background illumination on the photoresponses of red and green cones. J Physiol. 1979;286(1):491–507. doi: 10.1113/jphysiol.1979.sp012633. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Nowak MA. Evolutionary dynamics. Cambridge: Harvard University Press; 2006. [Google Scholar]
- 54.Olsen SR, Bhandawat V, Wilson RI. Divisive normalization in olfactory population codes. Neuron. 2010;66(2):287. doi: 10.1016/j.neuron.2010.04.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Ortega P, Braun D. Information, utility and bounded rationality. Lect Notes Artif Intell. 2011;6830:269–274. [Google Scholar]
- 56.Ortega P, Braun D, Tishby N (2014) Monte carlo methods for exact and efficient solution of the generalized optimality equations. In: IEEE international conference on robotics and automation (ICRA) pp. 4322–4327
- 57.Ortega PA, Braun DA. A minimum relative entropy principle for learning and acting. J Artif Intell Res. 2010;38:475–511. [Google Scholar]
- 58.Ortega PA, Braun DA. Thermodynamics as a theory of decision-making with information-processing costs. Proc R Soc A Math Phys Eng Sci. 2013 [Google Scholar]
- 59.Ortega PA, Braun DA. Generalized Thompson sampling for sequential decision-making and causal inference. Complex Adapt Syst Model. 2014;2:2. doi: 10.1186/2194-3206-2-2. [DOI] [Google Scholar]
- 60.Rao RP. Bayesian computation in recurrent neural circuits. Neural Comput. 2004;16(1):1–38. doi: 10.1162/08997660460733976. [DOI] [PubMed] [Google Scholar]
- 61.Rao RP. Bayesian brain: probabilistic approaches to neural coding, chap. Neural models of Bayesian belief propagation. Cambridge: MIT Press; 2007. [Google Scholar]
- 62.Rao RP, Ballard DH. Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects. Nat Neurosci. 1999;2(1):79–87. doi: 10.1038/4580. [DOI] [PubMed] [Google Scholar]
- 63.Reichardt W, Poggio T, Hausen K. Figure-ground discrimination by relative movement in the visual system of the fly. Biol Cybern. 1983;46(1):1–30. doi: 10.1007/BF00595226. [DOI] [Google Scholar]
- 64.Ringach DL. Spontaneous and driven cortical activity: implications for computation. Curr Opin Neurobiol. 2009;19(4):439–444. doi: 10.1016/j.conb.2009.07.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Shadlen MN, Newsome WT. Neural basis of a perceptual decision in the parietal cortex (area lip) of the rhesus monkey. J Neurophysiol. 2001;86(4):1916–1936. doi: 10.1152/jn.2001.86.4.1916. [DOI] [PubMed] [Google Scholar]
- 66.Shahlizi C. Dynamics of Bayesian updating with dependent data and misspecified models. Electron J Stat. 2009;3:1039–1074. doi: 10.1214/09-EJS485. [DOI] [Google Scholar]
- 67.Shi L, Griffiths TL (2009) Neural implementation of hierarchical bayesian inference by importance sampling. In: Advances in neural information processing systems 22 (NIPS 2009). Vancouver, British Columbia, Canada pp 1669–1677
- 68.Shon AP, Rao RP. Implementing belief propagation in neural circuits. Neurocomputing. 2005;65:393–399. doi: 10.1016/j.neucom.2004.10.035. [DOI] [Google Scholar]
- 69.Silver RA. Neuronal arithmetic. Nat Rev Neurosci. 2010;11(7):474–489. doi: 10.1038/nrn2864. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Tishby N, Polani D. Information theory of decisions and actions. In: Vassilis HT, editor. Perception-reason-action cycle: models, algorithms and systems. Berlin: Springer; 2011. [Google Scholar]
- 71.Todorov E. Efficient computation of optimal actions. Proc Natl Acad Sci USA. 2009;106:11,478–11,483. doi: 10.1073/pnas.0710743106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Tong F, Meng M, Blake R. Neural bases of binocular rivalry. Trends Cognit Sci. 2006;10:502–511. doi: 10.1016/j.tics.2006.09.003. [DOI] [PubMed] [Google Scholar]
- 73.Toussaint M, Harmeling S, Storkey A (2006) Probabilistic inference for solving (PO)MDPs
- 74.Usher M, McClelland JL. The time course of perceptual choice: the leaky, competing accumulator model. Psychol Rev. 2001;108(3):550–592. doi: 10.1037/0033-295X.108.3.550. [DOI] [PubMed] [Google Scholar]
- 75.Vickers D. Evidence for an accumulator model of psychophysical discrimination. Ergonomics. 1970;13(1):37–58. doi: 10.1080/00140137008931117. [DOI] [PubMed] [Google Scholar]
- 76.Vijayakumar S, Rawlik K, Toussaint M (2012) On stochastic optimal control and reinforcement learning by approximate inference. In: Proceedings of robotics: science and systems
- 77.Wang X, Sandholm T. Reinforcement learning to play an optimal Nash equilibrium in team Markov games. Cambridge: MIT Press; 2002. [Google Scholar]
- 78.Weibull J (1995) Evolutionary game theory. MIT Press, Cambridge
- 79.Wong KF, Huk AC, Shadlen MN, Wang XJ (2007) Neural circuit dynamics underlying accumulation of time-varying evidence during perceptual decision making. Front Comput Neurosci 1(6) . doi:10.3389/neuro.10.006.2007 [DOI] [PMC free article] [PubMed]
- 80.Wong KF, Wang XJ. A recurrent network mechanism of time integration in perceptual decisions. J Neurosci. 2006;26(4):1314–1328. doi: 10.1523/JNEUROSCI.3733-05.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Zaveri M, Hammerstrom D. CMOL/CMOS implementations of Bayesian polytree inference: digital and mixed-signal architectures and performance/price. IEEE Trans Nanotechnol. 2010;9(2):194–211. doi: 10.1109/TNANO.2009.2028342. [DOI] [Google Scholar]
- 82.Zhang H, Maloney LT (2012) Ubiquitous log odds: a common representation of probability and frequency distortion in perception, action, and cognition. Front Neurosci 6 [DOI] [PMC free article] [PubMed]
- 83.Zunino R, Gastaldo P (2002) Analog implementation of the softmax function. In: IEEE international symposium on circuits and systems, 2002. ISCAS 2002, vol 2, pp II–117. IEEE




