Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2021 Sep 3;17(9):e1009344. doi: 10.1371/journal.pcbi.1009344

The color phi phenomenon: Not so special, after all?

Lars Keuninckx 1,¤,*,#, Axel Cleeremans 1,#
Editor: Daniele Marinazzo2
PMCID: PMC8445478  PMID: 34478441

Abstract

We show how anomalous time reversal of stimuli and their associated responses can exist in very small connectionist models. These networks are built from dynamical toy model neurons which adhere to a minimal set of biologically plausible properties. The appearance of a “ghost” response, temporally and spatially located in between responses caused by actual stimuli, as in the phi phenomenon, is demonstrated in a similar small network, where it is caused by priming and long-distance feedforward paths. We then demonstrate that the color phi phenomenon can be present in an echo state network, a recurrent neural network, without explicitly training for the presence of the effect, such that it emerges as an artifact of the dynamical processing. Our results suggest that the color phi phenomenon might simply be a feature of the inherent dynamical and nonlinear sensory processing in the brain and in and of itself is not related to consciousness.

Author summary

Well known visual illusions are used to investigate how objective physical stimuli translate to subjective conscious experiences and are therefore believed to offer clues at how human consciousness works. One such illusion, the color phi phenomenon is marked by an anomalous time reversal of what is shown to test subjects and what they report to have seen. The color phi phenomenon has been a prime example to illustrate that conscious experience, instead of being merely a recording of the facts as they happen, involves both pre- and post-constructive processes resulting in a coherent narrative. We provide an alternative interpretation by showing that all aspects of this phenomenon can be adequately modeled in small neural networks. These networks are composed of simple leaky integrator neurons, emphasizing the importance of taking into account dynamical behavior. We illustrate how anomalous temporally-reversed perception of stimuli can appear based exclusively on bottom-up dynamics. Since the color phi phenomenon emerges as a feature of dynamical processing in our model, we conclude that it may explain less about human consciousness than previously thought and that it could be nothing more than a feature of dynamical processing in the brain.

Introduction

The color phi phenomenon (CPP), Fig 1, is perhaps one of the most baffling visual illusions known [1]. In the experiment, two different colored dots are shown in rapid succession and at some distance from one another. For certain intervals and spacings, which depend on the viewer, the perception is that of the first dot moving to the position of the second dot and changing to the color of the second dot somewhere along the perceived path. The perception of a mix of the colors of the dots is also possible. An interactive demonstration of the effect can be found online [2]. Clearly, the CPP is comprised of two separate effects. First, there is the phi phenomenon of perceived movement, which has been described since at least 1912 [3]. A swiftly moving “ghost” presence is perceived, located in between the positions of the discretely presented dots. The phi phenomenon has been the basis for moving pictures. The second part and the more striking anomaly of the CPP is that the color change is perceived before the second dot at the final position is. In the words of Herzog et al. [4]: “The conscious percept cannot have formed in a time-ordered fashion, but must have been constructed retrospectively”. The importance of the CPP is that it shows us that our conscious perception does not necessarily preserve the chronological order of the stimuli, nor does it need to correspond to actual applied stimuli.

Fig 1. The color phi phenomenon.

Fig 1

Two dots of different color are shown in rapid succession. For certain spatial and temporal distances, the perception is that of the first dot moving to the position of the second dot while abruptly changing color somewhere along the path.

The study of visual illusions such as Hermann grids and Mach bands which are direct illustrations of the existence of receptive fields, can offer important insights into our neural architecture [5]. Visual phenomena that are time-involved rather than static can hint at what kind of causal connections, feedforward, horizontal or feedback, exist between different brain areas [6]. In Ref. [7], it is argued that the receptive field of neurons in the visual cortex is mainly determined by feedforward processing, while feedback connections can modulate responses in V1 corresponding to perceptual grouping. In this way, the activity of feedback connections may correlate to perceptual awareness. However, as Bachmann [8] argues, it is difficult to distinguish experimentally whether top-down signaling to lower earlier activated feature-encoding neurons is due to higher level activation that is specific to target features or originates from non-specific activation linked to general arousal or expectation modulation that was initiated by feedforward non-feature-encoding reticulo-thalamic feedforward connections. In the majority of cases, top-down processing may be influenced by both specific and non-specific feedback.

Many illusions similar to the CPP exist that demonstrate that our perception is flawed when time is involved. For example, in the flash-lag illusion, the perception of the position of a moving object at an instant in time that another object is briefly shown, is systematically advanced relative to the true position [9]. The cutaneous rabbit illusion [10] is a tactile version of the phi phenomenon: taps or electrical pulses near the wrist and elbow in rapid succession create the feeling of a sequence moving up the arm, without there being any stimulus applied in between. The cutaneous rabbit illusion is accurately modeled by assuming perception follows a Bayesian inference process, based on a prior expectation of the stimulus having a low speed and uncertainty about its position [11, 12]. This results in an apparent distance contraction between the stimuli, perceptually leading to the rabbit hops. Given their success, these statistically based models add strength to the view of the brain as a (Bayesian) predictive inference machine [13], however they do not incorporate or highlight the underlying physiological connectionist basis.

In this article, we follow an approach that is more in line with O’Brien [14], in that perceptual awareness is seen as coinciding with at least temporally stable states or activation patterns existing in a connectionist system. Such an approach requires that phenomena such as the CPP are demonstrated and explained in a connectionist fashion, that is as existing in a neural network built up from elementary (artificial) neural units which have properties that are understood or accepted as being of similar nature as those of actual neurons in the brain. Most importantly, all these units operate continuously in parallel, hence the naming parallel distributed models (PDPs). Cohen et al. [15] presented a PDP account of the Stroop effect in a strictly feedforward dynamical model. Interestingly, their model had “attention” modulating inputs to steer the network’s task to respond to either the word or the color it was presented. Such attention-modulating input would in reality clearly be a part of a top-down feedback path, even if it is not explicitly derived from output activations present in this network. Since this was a feedforward network, it does not have memory obtained from attractor states. In contrast, Dehaene et al. [16] present a dynamical model that relies heavily on both short and long distance feedback between individual simulated thalamo-cortical columns. Interaction between these columns, which consist of integrate-and-fire units, allows them to simulate the attentional blink phenomenon. Conscious reportability is seen as caused by the synchronization or locking of several simulated columns in a global activity pattern. The apparent similarity across modalities between the cutaneous rabbit illusion and the phi phenomenon suggests a common underlying mechanism which, in a connectionist view, could be supported by a common neuronal microcircuit. This motivates why the study of archetypal (dynamical) microcircuits that demonstrate comparable abstractions of these effects is important.

In this article, we offer fully detailed and testable connectionist models which demonstrate stimulus-response order-of-arrival reversals, as well as “ghost” perceptions. We do so by employing a rather crude neuron model, that nevertheless has all the necessary properties an artificial neuron requires to be able to perform general computation. By choosing to position ourselves at a well-chosen distance from more biologically informed models, we are able to illustrate general principles.

With subjective experience changing from moment to moment, it is clear that consciousness itself takes time. Therefore, dynamical systems are an appropriate basis for modeling that offers insight in how things might be the way they are as, argued by Cleeremans and French [17], that is all modeling does. Differential equations (DEs) and their discrete-time counterparts difference equations form the natural language of dynamical systems [18]. Coupling of the same type of DE that describes an aspect of neural activity is how dynamical connectionist networks are built.

The CPP received some fame for being the centerpiece in an argument formulated by Dennet against the existence of a “Cartesian theater”, an informational finish line in the brain where the order-of-arrival coincides exactly with the order of conscious perception [19]. However, we argue, no matter in which manner the anomalous perception such as that of the CPP comes about, at some point, somewhere in the brain, there must exist an activation that reflects this subjective perception, if one is to accept that perception itself is physical. Therefore there must exist some neuronal circuit that transforms the applied stimuli in objective activations that in turn either are or cause the anomalous perception. Our goal here is to demonstrate that these phenomena can be modeled in the most simple way as generic dynamical effects. Note that we choose in this article to focus on mechanisms. Therefore we specifically do not correlate our models with results of behavioral experiments yet. Also, we remain agnostic as to the existence of feature vectors that include time labels as proposed in Herzog et al. [4]; our models do not require them as far as the CPP goes, but could be made compatible with that idea.

The remainder of this article is organized as follows. In the next section we introduce the dynamical neuron model we employ. Thereafter we describe a four-neuron network that illustrates the stimulus-response arrival order reversal. We introduce a model for the phi phenomenon, the operation of which depends on long distance feedforward “primer” connections. We then model the complete CPP using an echo state network (ESN), a recurrent dynamical neural network built from a large number of randomly connected toy model neurons. Finally, we summarize our results and propose ways in which we see this work continue.

Toy model neuron

Dynamical neuron models fall into two broad categories: spiking [20, 21] and non-spiking. Although spiking neuron models, which are inherently time-dependent, come closer to biological reality, they also form an extra hurdle since neural coding, apart from rate coding, is not completely understood and still actively researched. We choose here to go with a dynamical version of the simplest of all non-spiking neuron models, the perceptron [22]. Specifically, we aim to develop our arguments starting from three well accepted basic prepositions:

  • The computational power of a neuronal systems (brains) stems from the connection of many, nearly identical and relatively simple units, called neurons.

  • Neuronal signals are limited in amplitude and rate. Neurons and neuronal systems can saturate and thus are inherently nonlinear. If stimuli a and b elicit responses f(a) and f(b) respectively, then the simultaneous application of a and b does in general not result in response f(a + b).

  • Neurons and neuronal systems have a certain slowness to them. They cannot respond immediately to a change in input stimulus.

The first two properties give traditional artificial feedforward neural networks their general computational abilities, under certain mathematical conditions. The third property brings time and dynamical systems into the discussion. We start from the well-known perceptron neuron:

y=f(i=1Nwixi+b), (1)

in which the synaptic weights wi determine the input strength of input xi and f is a nonlinear activation function. The weights wi can be positive (excitatory) or negative (inhibitory). In general, a constant bias b is also included. A straightforward dynamical extension is given by the following ordinary differential equation:

τdydt=y+f(i=1Nwixi+b), (2)

which is also known as the leaky-integrator neuron [23], not to be confused with the spiking leaky integrate-and-fire neuron. Here, we will just call it a toy model neuron, since that is what it is: an abstract entity with minimal neuron-like properties. The derivative dy/dt is the velocity at which the output y changes. Here τ is a timescale, with higher τ’s meaning slower neurons. Note that for τ → 0 the response y(t) of Eq (2) equals that of Eq (1). The same is true for t → ∞, with a constant input x, which can be seen from setting dy/dt to zero. Thus Eq (2) represents a “slow” perceptron. The nonlinear activation function is usually a hyperbolic tangent, with output values ranging from −1 to 1 or a sigmoid f(x) = 1/(1 + expx), having a range of 0…1. Both are monotonic functions: increasing the stimulus strength leads to an increased response, up to saturation. Eq (2) can be interpreted as the average activity or firing rate of a single neuron as well as the activity of a larger neuronal (sub)structure. Our argument is that any phenomenon that can exist in a small network of these crude dynamical neurons, can also be present in the vastly more complex brain. Models built using this neuron then show which computational pathways might exist. Note that in Eq (2) all variables, including time t, are considered de-dimensionalized, as this simplifies notation [18].

Example: Backward temporal masking

Merely as a starting point and to illustrate what is possible with this model, consider the single-neuron circuit shown in Fig 2A. We will show it can serve as an elegant model for backward temporal sensory masking. The reader who is familiar with nonlinear dynamics can skip the remainder of this section. The circuit in Fig 2 is represented by:

τdydt=y+f(w1y+w2x) (3)

in which the parameters are chosen as: τ = 1 and w1 = w2 = 1.5. Here, for simplicity, the activation function is chosen to be a piece-wise linear function, of which the shape is reminiscent of a hyperbolic tangent:

f(x)={1ifx<1,xif1x1,1ifx>1. (4)

Plotting the right hand side of Eq (3), Fig 2B (top trace), for input x = 0, shows three points where dy/dt is zero. In other words, this system has three equilibrium points. A small positive push from the origin y = 0, will result in dy/dt > 0, thus the system evolves to more positive y values, until it reaches the right-hand side equilibrium y* = 1. Here it remains, because a small positive nudge towards values y > 1 results in a negative change dy/dt < 0. The treatment for the left hand equilibrium at y* = −1 is similar, showing it is also stable. Thus the system has a single unstable equilibrium in the center y* = 0 (indicated by an empty dot) and two stable equilibria at y* = ±1 (filled dots). The latter are the results of the above-unity positive feedback weight w1.

Fig 2. Backward temporal masking with a single dynamical neuron.

Fig 2

(A) Neuron circuit with excitatory self-feedback, Eq (3), illustrates backward temporal sensory masking. w1 = w2 = 1.5, τ = 1. (B) Positive feedback creates two stable equilibria at y* = ±1 (top trace). These form the minima of a gradient-descent function V(y) (bottom trace). (C) Depending on the interval between a priming and a countering stimulus of opposite polarity, masking can occur. (D) The effectiveness of the masking stimulus also depends on its amplitude.

As discussed in the Introduction, stable attractor states in the brain, at least temporary ones, have been hypothesized as the basis for conscious percepts [14, 24]. We emphasize that although the example given in this section is original work that to the best of our knowledge has not been shown earlier, it is hardly original in spirit and only serves as an introductory example. Note that technically, besides stable states, attractors can also include stable oscillations known as limit cycles and chaotic oscillations, which take in extended regions rather than isolated fixed points, in the phase space of a dynamical system. Mathis and Mozer [24] were able to correlate several priming experiments to a discrete-time model. Their model is based on a radial basis function (RBF) attractor network, instead of a specific type of (networked) artificial neuron, and falls into the category of gradient descent systems. Formally, in a (multivariable) gradient descent system, the change of the system state y(t)=(y1(t),,yK(t)) can be written as:

dydt=V(y)+x(t), (5)

where is the Nabla operator: V(y)=(Vy1,,VyK). When undriven by a stimulus x(t), the state evolves downward towards a minimum of a scalar function V(y):RKR, much like a golf ball rolling downward to the lowest point of a valley it happens to find itself in. The influence of the input stimulus is such that it can drive the ball, representing the state of the system at every point in time t, closer or farther away from specific local minima of the V(y), corresponding to influencing both which perception the stimulus elicits and the time it takes to do so. Eq (3) is actually also a gradient system, albeit a one-dimensional one. This is shown in Fig 2B (lower trace). The piece-wise linear activation function Eq (4) allows integration of the top trace, so the undriven system can be written as dy/dt = −dV(y)/dy. This leads to the piece-wise parabolic gradient function V(y), shown in Fig 2B (bottom trace). The local minima of V(y) are at the stable equilibria y* = ±1 of the system. The maximum at y* = 0 is an unstable equilibrium. An animation in the supporting information, S1 Animation, shows how the stable “pockets” or valleys are created as the positive feedback w1 increases.

In above setting, conscious perception coincides with a state sufficiently close to an equilibrium of a, not necessarily explicitly described, gradient system. Thus this exemplary system is capable of two percepts y = +1 and y = −1, when started from y = 0 at t = 0. The outcome depends on the stimulus x(t) during the time in which the dynamical state y(t) changes. Note that short-lived or transient states can also represent distinguishable conscious percepts, if they can cause distinguishable stable states in subsequent processing units exhibiting longer, possibly infinite, memory.

In Fig 2C (top trace), we show how this simple one-neuron system is useful in a computational version of a stimulus onset asynchrony experiment. Here, a priming stimulus of unit strength and duration 0.5 starts the system on a path to experience the positive percept (bottom trace). Some time later, during the transient, a single negative masking stimulus of amplitude −1.5 is administered. Depending on the time elapsed between the priming and masking stimulus, the system ends up in either one or the other equilibrium. The self-reinforcing feedback can be interpreted as the system making a top-down prediction: in the absence of new information, it will move towards the state that coincides with the direction that the initial input pushed it towards. Contrasting information, in the form of input steering the state in the opposite direction can be taken into account, however it must overcome the already accumulated “proof” which, especially in later stages, becomes more difficult to do. That is why in this example, even though the masking stimuli are stronger than the prime, only the one closest in succession to the primary stimulus is able to mask it successfully. In Fig 2D we repeat this simulation for different amplitudes of the masking stimulus but for a fixed time post-stimulus. As expected, the strength of the mask influences its effectiveness. Note that even if a masking stimulus is not effective in blocking the prime, it can still delay the response. Thus this simple model predicts that effective backward sensory masking depends on both amplitude and elapsed time since the primary stimulus. We have included an animation in the supporting information, S2 Animation, showing how the state moves towards a minimum of the gradient function. This model, given merely to exemplify the merits of dynamical systems, is of course too simple to catch all scenarios. For example blanking, leading to no perception, is nearly impossible here since the mask must cancel the primary stimulus exactly. This is due to the unstable equilibrium in the origin. Adding noise to Eq (3), such as is present in a drift model, randomizes the outcome and leads to a probability distribution that can be compared against results from actual experiments. Note that a non-dynamical sum-saturate neuron can be used to show (discrete-time) dynamical effects such as memory when it is used in the context of a recurrent network. An example is given by the simple recurrent network [25], where due to discrete state updates previous inputs play a role in determining the current output.

In the next section, we show a small network of toy model neurons that illustrates the first part of the CPP, namely how perceptual states and their respective causal stimuli can appear to be chronologically reversed under certain circumstances.

A dynamical model for stimulus-response order of arrival reversals

In Fig 3A, we show a neuronal circuit with four neurons of the type described by Eq (2). Two inputs xa and xb, feed into two chains each having two neurons, top: a1, a2, bottom: b1,b2. The respective outputs of the chains, ya2, and yb2, are taken to coincide with the perception of the corresponding stimulus, when above a chosen threshold value. Inhibitory (negative) “long distance” connections cross couple the inputs to the opposite chain. Formally, the circuit is represented by:

τa1dya1dt=ya1+f(wxa,a1xa+wa1,a1ya1+ba1),τa2dya2dt=ya2+f(wa1,a2xa+wa2,a2ya2+wxb,a2+ba2),τb1dyb1dt=yb1+f(wxb,b1xa+wb1,b1yb1+bb1),τb2dyb2dt=yb2+f(wb1,b2xa+wb2,b2yb2+wxa,b2+bb2), (6)

where we chose the parameters as: τa1 = τb1 = 2, τa2 = τb2 = 10, wxa,a1 = wxb,b1 = 1, wa1,a2 = wb1,b2 = 1, wxa,b2 = wxb,a2 = −2, ba1 = ba2 = bb1 = bb2 = −0.5. As activation function we choose:

f(x)={0ifx<0,xif0x1,1ifx>1. (7)

The threshold of perception was chosen as 0.5, equal to midway of the range of the nonlinear function f. Conceptually, the existence of this arbitrarily chosen threshold does not have to be problematic, as subsequent structures that take the outputs of the circuit may be tuned to have an all-or-nothing response based on this threshold. In Fig 3B, we show how differently spaced input stimuli (of width δ = 10), cause different response times RTa (red) and RTb (blue). For both large positive and negative input intervals, Δtin = tin,btin,a, the response arrival order is unaffected, i.e. first in first out. However, for Δtin ≈ −12 to 12, we see Δtout = tout,btout,a (black) has the opposite sign of Δtin, i.e. last in first out. Without the reversal effect, the green dashed line would be followed. Note that if both inputs are activated simultaneously, then both RTs are increased, however the interval between the responses remains zero.

Fig 3. A four-neuron model for demonstrating stimulus-response arrival order reversals.

Fig 3

(A) The model. See below Eq (6) for parameter values. (B) A Scan over stimuli intervals Δtin = tin,btin,a showing the influence on response times RTa = tout,atin,a (red), RTb = tout,btin,b (blue) and the response time interval Δtout = tout,btout,a (black). For a certain central region of the input interval, there exists a reversal of the order in which responses happen, compared to the order in which the stimuli were applied. (C) Scan over stimuli intervals and amplitude ratios, illustrating how temporal and amplitude-related stimuli characteristics influence their temporal perception. (D) Detailed signals for tin,btin,a such that the output order corresponds to the input order. (E) Detailed signals in the input-output reversal regime, where Δtin > 0 and Δtout < 0.

Fig 3C shows how the model behaves if we vary both the amplitudes of the input stimuli and the interval between them. Here, the sum of the amplitudes is kept at 1.5 and their ratios are varied. In Fig 3C, the lighter colored central part marks output intervals that are temporally reversed relative to the input interval. There is a tendency, within certain limits, that the more stronger a later arriving stimulus is, the easier it can overtake an earlier one. Thus the model demonstrates how both temporal and amplitude-related characteristics of complex stimuli influence their temporal perception in a non trivial way.

Fig 3D and 3E illustrate this process in detail. The input stimuli trigger a latching value at the output of the first neurons, ya1 and yb1, in both chains. These in turn drive the subsequent stages. This however takes some time due to the slower timescale of neurons a2 and b2. If the input xb is activated while the output value ya2 is still accumulating, Fig 3E, then the inhibitory synapse slows down the a-branch response. This leads to yb2 crossing the threshold before ya2, even if stimulus xa was present before xb was. Since the circuit is symmetrical along the branches, swapping the input stimuli leads to swapped output signals.

The circuit of Fig 3A clearly illustrates how dynamics analogous to anomalous order-reversed perception can be found in small dynamical networks. The key ingredient is the set of cross coupled inhibitory connections which jump forward in the chains, such that information that arrives at a later time can intercept partially processed information at the opposing chain as it flows from input to output. Loosely speaking, we make an exchange of time for space.

In the next section, we illustrate how the second aspect of the CPP, an apparent movement or “ghost” perception, can be found in a similar small network motif.

A dynamical model for the phi phenomenon

As mentioned, the color phi phenomenon has two aspects. In this section we explain how the appearance of a ghost percept, spatially and temporally placed in between percepts that correspond to actual stimuli, might be understood as a dynamical priming effect. In Fig 4A, we show a toy model neuron circuit consisting of 15 units, divided in three chains a, b and c. It represents a very crude one-dimensional visual system having three pixels. Stimuli on the respective inputs xa, xb and xc, Fig 4B (top), lead after some propagation time to perceptions as indicated by the outputs ya, yb and yc (bottom) being above a fixed threshold value. Each chain couples to the adjacent chain(s) via weak long distance feedforward synapses, shown as dotted connections in Fig 4A. These are driven by primer neurons pa, pb and pc, Fig 4B (middle), that have a fast activation and a slow decay, the latter due to the strong self-feedback and lack of negative bias. In Fig 4B (middle), the dotted line shows the combined input of the primers into the final neuron of the b-chain. As long as the interval between the stimuli on the a and c chain is sufficiently large, their combined priming signal does not exceed the activation threshold of the b-chain, Fig 4C. However, when the stimulus on c swiftly follows the one on a, as in Fig 4D, priming exceeds the threshold and elicits a response at yb. As in the preceding section, we find a timing anomaly: b is initially “perceived” after a and before c, here close to t = 115. An animation showing the situation for intermediate a-c stimulus intervals can be found in the supporting information, S3 Animation. Since the network is symmetric, the same effect exists when the stimuli are applied in reverse.

Fig 4. An abstract one-dimensional three-pixel visual system, illustrating the phi phenomenon.

Fig 4

(A) The circuit. Outputs ya, yb and yc separately respond to inputs xa, xb and xc respectively, corresponding to perception when above a certain threshold. Weights and τ-values of the a and c-chains are identical as those indicated in the b-chain. (B) Long distance weakly coupled feedforward priming units pa, pb and pc (middle graph) with fast activation and slow decay enable sensitivity to recent inputs to be carried to adjacent output units. (C) If the input pulses on xa and xc are temporally sufficiently separated, the decayed priming signals pa and pc are unable to evoke a response from the middle output yb. (D) When inputs xa and xc are stimulated adequately fast after one another, the priming signals combine and a ghost percept on yb appears. This ghost happens after ya and before yc cross the threshold in opposite directions. The ghost appears without the presence of an input stimulus on xb.

One may wonder how this simple model can be interpreted within a biological context. If a moving object, e.g. a prey animal, is seen at some position, say b, it is natural to assume that it will be moving to an adjacent position, here either a or c, next. The priming signals thus perform the function of expectation. Priming for the adjacent positions helps reduce response times, and increases the chance of catching that prey. Likewise, if an object is perceived as quickly changing its location from a to c, Fig 4D, then by assuming movement and thereby inferring that a path exists between the positions, the organism is potentially given useful information about the environment. If the objects are perceived with a greater temporal interval at positions that are not adjacent, as in Fig 4C, they are likely not the same and the existence of a pathway cannot be inferred. Thus this simple model demonstrates that perception, even when caused mainly by feedforward neuronal pathways, can in part be prediction [13]. Note that these interpretations are meant to place the abstract model of Fig 4 in a more understandable context, and by no means do we want to state physiological facts.

Clearly, this simple hand assembled model fails in two ways. First, in Fig 4D, there is overlap between the b and c percepts. Second, if xa and xc are stimulated at the same time, yb will also be seen, while in an actual experiment only the two outer position dots would be perceived simultaneously. These shortcomings could perhaps be alleviated at the expense of increased model complexity, by using inhibitory connections from the inputs to the outputs. Similar to the case of the order reversal model, as shown in Fig 3C, the stimulus characteristics influence whether or not a phi phenomenon happens. Our code S1 Data contains a script that plots the duration of the ghosts as a function of stimulus interval and amplitude ratio.

In the next section, we show a different approach to finding such a model and demonstrate that the CPP could be understood as nothing more than a side effect of dynamical processing.

The color phi phenomenon in an echo state network

The models presented in the previous sections illustrate both aspects of the CPP as discussed. Unfortunately, there seems to be no straightforward way to combine their structure into a single CPP model. Therefore, we turned to neural network paradigms to find such a model.

Specifically, we employ an echo state network (ESN) [26]. For completeness, we briefly explain the philosophy behind the ESN, after which we formally introduce our model. ESNs where envisioned to circumvent the problem of slow backpropagation learning found in traditional multilayer neural networks [27, 28] and to be able to exploit the computational abilities of high-dimensional nonlinear physical substrates in which the constituents have a fixed and often unknown interconnectivity. More in-depth discussions of ESNs, part of the larger group of reservoir computing methods, can be found in Refs [2931].

Instead of adjusting individual weights between neurons, such that the network itself performs a certain prescribed function, in an ESN, Fig 5, one starts from a large recurrent network, also called a reservoir, in which the weights between the network’s neurons or nodes are chosen at random, within certain constraints, and kept fixed throughout the experiment. The presence of recurrent connections implies the existence of time scales or memory on top of those offered by the neurons’ individual dynamical behavior. An input signal is connected, usually also via randomly chosen weights, and excites the network. For our purposes, the ESN will have 3 × 2 inputs corresponding to left, middle and right positions in two colors, blue and red, similar to the abstract one-dimensional visual system described in the previous section. The complicated time-dependent response or “echo” in the network caused by the time-varying input is then recorded. Mathematically, the input signal is non-linearly projected into the high-dimensional phase space of the dynamical network. The training phase of the ESN consists of constructing the desired output response from a linear combination of the response signals of the individual neurons. The weight values or coefficients of this linear combination are collectively known as the readout layer. Mathematically, the readout layer forms a down-projection from the phase space of the network onto a lower dimensional output space. The main idea behind ESNs is that as long as the network shows what is loosely referred to as “rich dynamical behavior” in response to the input, chances are good that a projection exists that maps the dynamical response of the network sufficiently close to the desired output. The calculation of the readout layer is a simple linear regression, which is computationally relatively cheap and fast. In this case, we have five outputs: three for the position and two for color. As such, the quite easy task of the ESN is to classify a single active input. Furthermore, when several inputs are active at the same time, we demand that all outputs must remain inactive. Since the ESN is a dynamical system, it will show transition effects between different input stimuli. After training, the ESN will be tested for the CPP as further outlined below. It is crucial to remark here that the network is not trained to show the CPP explicitly. As explained in detail in the next section, the ESN consists of a discrete-time version of the toy model neurons, Eq (2).

Fig 5. Illustration of the color phi phenomenon via an echo state network.

Fig 5

A large randomly connected network of neurons, also called a reservoir, interacts in a complicated and unknown way in response to an input stimulus. The connections between the neurons in the network remain fixed for the duration of the experiment. The desired output response is constructed from a linear combination of the responses of the individual neurons that form the network. After training, the ESN can generalize to novel input data.

Interpreted as generating perceptions, the readout layer is the place where “it all comes together”, i. e. the Cartesian theater. Note that it is not correct to interpret the readout layer as simply being a reporting tool for the network, a voice box so to speak, since without it, the computations that happen in the network do not get assembled into an active representation and remain inert. Alternatively, the outputs of the readout layer could be seen as a behavioral response. The reservoir itself could be extended to have specialized readout neurons, the outputs of which perform exactly as the readout layer.

ESN outline

In this section we formally describe the ESN we use to demonstrate the CPP. Starting from a discretization with timestep Ts, setting t = nTs, we replace differentials with differences as:

dydty(nTs)y(nTsTs)Ts, (8)

such that the toy model neuron Eq (2) takes on the form of a recurrence equation:

yn+1=(1α)yn+αf(iNwixi+b), (9)

in which n is the discrete timestep and the neuron timescale is given by α = (1 + τ/Ts)−1. Compared to the artificial neurons used in Cohen [15], in Eq (9) the differentiating action and the saturating function are swapped, however their steady-state output remains the same. The ESN state update equation is functionally equivalent to what is found in Ref [23]:

xn+1=(IA)xn+Af(Wresxn+Winxin,n+b), (10)

in which xn=(x1,n,,xN,n) is the network state vector at timestep n. In Eq (10), vectors are to be interpreted as column matrices. Furthermore, Wres is the N × N network connectivity matrix, Win is the N × 6 input weights matrix and b=(b1,,bN) is a constant random bias vector. Vector xin represents the six inputs as shown in Fig 5. In our experiments, we set the network size to N = 200 neurons. Network hyperparameters were chosen as suggested in Ref. [32]. The network connectivity matrix is constructed by drawing random numbers from a normal distribution N(μ=0,σ=1) in which 80% of the entries is set to zero (sparsity 0.8). Then the matrix is scaled to obtain a spectral radius of 0.9. Similarly, the input matrix Win is constructed from a normal distribution N(μ=0,σ=0.5) with sparsity set to 0.8. The timescales of the individual neurons are drawn from a uniform distribution αiU(low=0.1,high=0.3), and collected in a diagonal matrix Ai,i = αi. Here, f is the sigmoidal activation function. Both during training and testing of the ESN, the network states in response to the input signal are collected in a state matrix X of dimensions K × (N + 1). Each row of X consists of the network state xn at time n = 1, …, K. The last column of X is constant. The readout layer consists of a (N + 1) × 5 matrix Wout. The output of the ESN is calculated as:

Y=XWout, (11)

where Y is a K × 5 matrix, of which each column corresponds to an output signal as shown in Fig 5. The n-th row of Y corresponds to the output vector at timestep n, yn=(y1,n,y5,n). The readout layer Wout is determined using a linear regression, such that the output Ytrain in response to a training input that causes network states Xtrain comes close, in a least-squares sense, to a predefined desired output Ydesired:

Wout=(XtrainTXtrain)1XtrainTYdesired (12)

A well-trained ESN can generalize to previously unseen input data. We refer the reader to Refs [2931], as further details about the ESN are not important for the remainder of this article.

ESN experimental results

In this section we show how the CPP surfaces in the ESN we described above. As shown in Fig 6, the training input (top graph) consists of valid one-position/one-color stimuli, for which the desired output (third graph) is the corresponding position and color. From n = 13000 onward, the training stimulus consists of invalid input conditions in which multiple color and/or position inputs are active simultaneously. For these confusing stimuli, we require the network to not respond (keep all outputs at zero). A shift of 20 timesteps between input and desired output was allowed to give the network time to respond. After training, i.e. calculating the readout layer Wout, we can show the resulting actual output in response to the training input (bottom graph of Fig 6). The output is assembled from the complex network dynamics (first ten neurons shown in the second graph of Fig 6). Ideally, it should be identical to the desired output. An initial transient and some noise-like deviation is seen, however qualitatively the output comes close to what is desired.

Fig 6. ESN during training.

Fig 6

(top) Training input, (second graph) states of the first ten neurons, (third graph) desired output used to calculate the readout layer, (bottom graph) actual output calculated by the readout layer from the neuron states after training.

It is vital for this discussion to emphasize that we do not train with input conditions that could lead to the CPP: transitions between the conditions left-to-right or right-to-left of opposite color without passing via an activated middle input are explicitly excluded from the valid training examples. Furthermore, input conditions are sufficiently spaced as to allow the network’s internal dynamics to settle to a stable output. We are now ready to demonstrate that this neural network can undergo a color phi illusion. In Fig 7A (top), we apply several CPP input conditions, i.e. a stimulus that jumps from left-red to right-blue, with decreasing interval. Note that the left-red and right-blue input pulses never overlap. Ideally, the middle output would remain low (middle graph), corresponding to having not perceived anything anomalous. However, as the left-red to right-blue jumps get closer to one another, at n ≈ 3490, as shown in the zoomed graph of Fig 7B (bottom graph), the middle output is briefly active above the threshold of 0.5 (dashed horizontal line), while the right output is still below this threshold. This corresponds to a ghost or phi phenomenon as elaborated on in the previous sections. More so, at the same time the blue output is also active, thus the output of the ESN corresponds with the perception of this ghost in the middle position and in the final color before correctly perceiving the second applied stimulus of that color.

Fig 7. Echo state network during testing for the color phi phenomenon.

Fig 7

(A) (top) Stimuli jumping between left-red and right-blue with decreasing intervals, (middle graph) ideal response with low middle output, (bottom graph) actual output showing CPP near n ≈ 3490. (B) Zoom-in of (A). The vertical dashed green line in the bottom graph indicates the time at which the CPP occurs.

The dashed vertical line in Fig 7B (bottom graph) indicates the point in time where this condition first occurs. Thus the system portrays all the characteristics of undergoing a color phi illusion. The supporting information contains a short clip S4 Animation produced directly from the timeseries of Fig 7A, showing how the ESN “perceives” the CPP. The perception of the final transition from left red to right blue briefly shows a blue dot in the middle.

Analysis

Since the model is essentially a system of N = 200 coupled nonlinear difference equations, we cannot explain in detail how this behavior arises. It is obvious, however, that simple dynamical interactions give rise to the existence of the CPP here. Nevertheless, it is important to ask if there are any discernible differences between ESNs that do and those that do not show the CPP effect. In Fig 8 we present four network characteristics that we specifically checked to identify such differences. The plots are based on N = 100000 randomly assembled ESNs in total. In general, only a low fraction of the ESNs show the CPP. In our experiments this was 1.87%. Fig 8A shows the distribution of α-parameters of the neurons. In Fig 8B we plot the magnitude of the eigenvalues of the reservoir matrix Wres. Both of these metrics govern the timescales at which the dynamics evolve. There is no clear difference between networks with and without the CPP. In Fig 8C we plot the ratio of excitatory (positive) to inhibitory (negative) synapses. Finally, Fig 8D shows the number of positive long distance feedforward connections (LDFF) in the networks. Here, a connection from neuron i to neuron j is counted as LDFF, if i positively influences j via both a direct and an indirect path. This test was suggested by our phi model, where priming was caused by LDFF connections. Again, no differences between networks that show the CPP and those that do not come forward. These negative results suggest -but of course do not prove- that it truly is the detailed internal structure that matters, and the ESNs are otherwise macroscopically indistinguishable. We suspect only a small portion of the neurons in the ESNs assists in generating the CPP.

Fig 8. Some characteristics of the ESNs that do (Nwith = 1867) and do not (Nwithout = 98133) show the CPP.

Fig 8

(A) Distribution of neuron α-timescales, (B) magnitude of the reservoir matrix eigenvalues, (C) ratio of excitatory to inhibitory synapses, (D) number of positive long distance feedforward connections.

In Fig 9, we show a brief exploration of how robust the results are with respect to the choices of network parameters. Here, we scan over reservoir and input sparsity, spectral radius and input scaling, the latter being the σ-value with which the input matrix Win is constructed. The non-varied parameters are as described in the previous section. Each point is the fraction p out of n = 200 ENSs that showed the CPP. The error bars are estimated at the 95% confidence interval using Wald’s method: p±zp(1p)/n, with z = 1.96.

Fig 9. Fraction of ESNs that show the CPP as a function of several network parameters.

Fig 9

(A) Reservoir sparsity, (B) input sparsity, (C) spectral radius, (D) input scaling. Each point represents n = 200 tested networks. The error bars are estimated at the 95% confidence interval. Non-varied parameters kept as described in the text.

It must be remarked that invariably such parameters are highly interdependent, such that for the purpose of optimization they must be probed together. We did not explore this any further as this is was outside of the scope of this article.

In the next and final section we summarize our results and look at how they may be interpreted in light of the importance of the CPP.

Discussion

Our goal in this article was to show that small connectionist networks built from crude toy model neurons can capture the main aspects of the color phi phenomenon (CPP). The CPP was described by Dennett [19] as a landmark example of the constructive character of conscious experience: rather than merely recording events as they unfold over time, subjective experience involves both preconstructive and reconstructive mechanisms that constitute a coherent narrative for the subject whose experiences it is. There is thus no central “theater” in which unfolding events play out; rather, says Dennett, consciousness builds on “multiple drafts” out of which one narrative will eventually win out.

With this in mind, we simply asked which kinds of neural architectures might naturally lead to the striking reversals observed in the CPP. Our approach was based on the following assumption: if consciousness is something physical (as it must be), then there must be mechanisms through which objective stimuli are mapped into objectively measurable neuronal activations which either constitute or drive conscious perception in a way that is congruent with subjective reports. The core aspect of what needs to be explained in the case of the CPP is the chronological anomaly it presents us with: how can the color of the second stimulus be experienced before its position is? Here, we set out to explore which kinds of neural network architecture might produce such effects without further assumptions, and we showed that the dynamical properties of relatively simple elementary networks naturally produce CPP-like effects. Our models are relatively simple circuits that are assembled from toy model neurons. They nevertheless have three essential properties in common with real biological neural systems: modularity, nonlinearity and time dependence. Our work is connected to the ideas recently developed by Herzog et al. [4], who proposed an explanation of visual perception that operates in two stages. The first stage consists of a continuous and unconscious integration of sensory information into features, best representing the stimulus. These features include time labels. During processing, all features are continuously updated yet not reportable. In a second stage, the features are rendered into consciousness all at once. We demonstrated stimulus-response order reversal in a four-neuron network that uses local recurrence, showing that, at least in this scenario, such time labels are unnecessary.

In line with Herzog et al., our simple model shows continuous feature integration and a by definition discrete, threshold-like rendering of features into conscious percepts. Additionally, timing labels may be constructed as neuronal activaty, the value of which is related to the time since the onset of a (not yet reportable) stimulus. A thresholded signal derived from the stimulus as a whole acts as a gatekeeper. Naturally, a neural timing signal would have to be relative to the larger perceptual context or mental state, rather than present any objective wall clock time. In summary, the experience of time is just like any other percept: a construction. Our four-neuron time reversal model illustrates how little machinery is needed to falsify this experience. Thus we conclude that our models, if needed, could incorporate time label features as proposed by Herzog et al. [4].

Following the breakdown of possible types of temporal consciousness in Herzog et al. [33], we consider our models to come closest to the discrete retentional class, albeit with percepts of finite non-zero duration which is at least in part governed by the stimuli. In our ESN model, not all feature perceptions are updated simultaneously and this is where the CPP-like effect originates from: the “color” feature is updated (meaning; crosses a perceptual threshold) asynchronously from the“position” feature, see Fig 7B. Thus features can be updated all at once, without noticeable timing differences, but this is not a given, as it generally depends on the characteristics of the stimulus and probably also on the larger perceptual context. What our ESN model indicates is that the CPP is the result of dispersion between features that stimulus-wise belong to each other, but are processed by different pathways. Under the right conditions, subtle timing differences caused by the different velocities with which the features traverse the neural substrate cause a noticeable binding error.

Fekete et al. [34] have argued at length against a perception that is tied down in a “discrete straightjacket”, on the basis of both empirical physiological data and philosophical arguments. One strong argument against a fully discrete awareness is the asynchronicity of inter area communications of the brain.

It is interesting to note that there are common elements in the structure of the order reversal and the phi phenomenon networks we presented. Both have local recurrences and long-distance feedforward connections. Local recurrences evidently may induce long lasting states without the need for large buffers and feedforward priming connections, as discussed in the section on the phi phenomenon model, can be interpreted as a form of prediction. Sadly, these simple models do not offer a clear path back into how they might be represented in the actual neural architecture of the brain, although functionally similar circuits must exist at some level.

We see several directions in which we could expand this work. First, it is clear that the interactions that take place in the networks we proposed are influenced by the amplitude, duration and interval characteristics of the stimuli, something we did not fully explore, except briefly for the introductory masking example and in Fig 3C. In the latter it was shown that the relative strength of the stimuli influences the temporal perception of their intervals. Results from such explorations could be compared to empirical data with the goal of finding similar scaling laws.

Second, we did not offer any explanation as to how the proposed neuronal structures could have emerged in the first place. In other words: under which conditions might the relevant weights be learned instead of merely being hand-tuned? It is straightforward to equip the leaky integrator neurons with a learning rule that mimics spike-timing-dependent-plasticity, by setting w˙ij=ηxiy˙j, which can be interpreted as follows: when input xi is active while output yj increases, then this input predicts the increased output value, hence the associated weight should increase. This learning rule was recently proposed by Bengio et al. [35, 36]. It would be interesting to investigate under which circumstances the CPP emerges in a plastic neural network, as opposed to the ESN which has fixed randomly chosen internal weights.

In our models, we (naively?) interpreted the end of a percept as coinciding with crossing the same threshold as its onset but in the opposite direction. The animation S4 Animation in the supplementary information of how the ESN “perceives” the CPP, based on the timeseries of Fig 7, is constructed in this way. However, as many experiments have shown, there clearly is an asymmetry in the sensitivity of when something emerges vs. when it disappears from our perception, even in very simple visual scenarios. Such hysteresis effects have been hypothesized by Poltoratski and Tong [37] to disambiguate visual perception during the processing of scene transitions. In other words: the hysteretic thresholds are themselves a function of the scene in its entirety. Since a continuous signal -the precursor of a conscious percept- cannot pass two unequal values instantaneously, the existence of hysteresis naturally implies percepts of non-zero duration. Hysteresis effects could be explored further using the modeling approach shown in this article.

We stress again the crucial point that during the training phase of the ESN we explicitly excluded input conditions that could invoke a CPP: the readout layer was constructed from the network’s responses to input stimuli that did not contain direct left-to-right or right-to-left jumps of opposite color. We found 1.87% of N = 100000 networks showed the CPP, with the network parameters we chose. Kim and Senjowski [38] showed that for a spiking recurrent neural network performing a working memory task, network characteristics that strongly influence the emerging timescales, and which are thus good predictors of performance, can be found. We similarly analyzed several characteristics of our ESNs, but failed to find markers that predicted the presence or absence of the CPP.

Although the ESN, or more general reservoir computing, has been described as a “brain-inspired” machine learning paradigm, one could argue that the readout layer, of which the construction is fundamentally based on a matrix inversion, is all but biologically plausible. Thus we can interpret the ESN simply as a function approximator which in itself has no relation to consciousness or the brain whatsoever. Stated otherwise: the ESN shown here exhibits sensitivity to input conditions that invoke the CPP, but remains mute concerning conscious perception. We conclude that the CPP, while an intriguing visual illusion in its own right, might be nothing more than a consequence of the inherent dynamics that characterize information processing in the brain.

Supporting information

S1 Animation. An animation showing the creation of stable states by feedback in a single toy model neuron network used for demonstrating priming.

(WMV)

S2 Animation. An animation of a bistable system with positive input stimulus masked by negative stimuli with decreasing interval.

(WMV)

S3 Animation. An animation of the phi phenomenon, the appearance of a “ghost” percept in dynamical neuronal network representing an abstract three pixel visual system.

(WMV)

S4 Animation. An animation of the color phi phenomenon based on the timeseries found in Fig 7A as “perceived” by the ESN during testing.

(WMV)

S1 Data. Contains the Python scripts that directly generate the graphs of this article.

(ZIP)

Acknowledgments

AC is a Research Director with the Fonds de la Recherche Scientifique F.R.S.-FNRS (Belgium).

Data Availability

All data underlying our results stem from model simulations in the form of Python scripts and are included in the Supporting Material.

Funding Statement

AC and LK were supported by the European Research Council under the Advanced Grant titled "Consciousness: The Radical Plasticity Thesis", SH4, ERC-2013-ADG. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Kolers PA, von Grünau M. Shape and color in apparent motion. Vision Research. 1976;16:329–335. doi: 10.1016/0042-6989(76)90192-9 [DOI] [PubMed] [Google Scholar]
  • 2.Bach M. Colour Phi Phenomenon; 2008. https://michaelbach.de/ot/col-colorPhi/index.html.
  • 3.Wertheimer M. Experimentelle Studien über das Sehen von Bewegung. Zeitschrift für Psychologie. 1912;61:161–265. [Google Scholar]
  • 4.Herzog MH, Kammer T, Scharnowski F. Time Slices: What is the Duration of a Percept? PLOS Biology. 2016;14. doi: 10.1371/journal.pbio.1002433 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Eagleman DM. Visual illusions and neurobiology. Nature Reviews Neuroscience. 2001;2:920–926. doi: 10.1038/35104092 [DOI] [PubMed] [Google Scholar]
  • 6.Kafaligonul H, Breitmeyer BG, Öǧmen H. Feedforward and feedback processes in vision. Frontiers in Psychology. 2015;6:279. doi: 10.3389/fpsyg.2015.00279 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Lamme VA, Supèr H, Spekreijse H. Feedforward, horizontal, and feedback processing in the visual cortex. Current Opinion in Neurobiology. 1998;8:529–535. doi: 10.1016/S0959-4388(98)80042-1 [DOI] [PubMed] [Google Scholar]
  • 8.Bachmann T. A hidden ambiguity of the term “feedback” in its use as an explanatory mechanism for psychophysical visual phenomena. Front Psychol. 2014;5:780:507–508. doi: 10.3389/fpsyg.2014.00780 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.MacKay DM. Perceptual Stability of a Stroboscopically Lit Visual Field containing Self-Luminous Objects. Nature. 1959;181 (4607):507–508. doi: 10.1038/181507a0 [DOI] [PubMed] [Google Scholar]
  • 10.Geldard FA, Sherrick CE. The Cutaneous “Rabbit”: A Perceptual Illusion. Science. 1972;178:178–179. doi: 10.1126/science.178.4057.178 [DOI] [PubMed] [Google Scholar]
  • 11.Goldreich D. A Bayesian Perceptual Model Replicates the Cutaneous Rabbit and Other Tactile Spatiotemporal Illusions. PLOS ONE. 2007;2(3):1–11. doi: 10.1371/journal.pone.0000333 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Goldreich D, Tong J. Prediction, Postdiction, and Perceptual Length Contraction: A Bayesian Low-Speed Prior Captures the Cutaneous Rabbit and Related Illusions. Frontiers in Psychology. 2013;4:221. doi: 10.3389/fpsyg.2013.00221 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Hohwy J. The Predictive Mind. Oxford; 2013. [Google Scholar]
  • 14.O’Brien G, Opie J. A connectionist theory of phenomenal experience. TBehav Brain Sci. 1999;22:127–48; discussion 148. doi: 10.1017/S0140525X9900179X [DOI] [PubMed] [Google Scholar]
  • 15.Cohen JD, Dunbar K, Mc Clelland JL. On the control of automatic processes: A parallel distributed processing model of the Stroop effect. Psychological Review. 1990;97(3):332–361. doi: 10.1037/0033-295X.97.3.332 [DOI] [PubMed] [Google Scholar]
  • 16.Dehaene S, Sergent C, Changeux JP. A Neuronal Network Model Linking Subjective Reports and Objective Physiological Data during Conscious Perception. PNAS. 2003;100(14):8520–8525. doi: 10.1073/pnas.1332574100 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Cleeremans A, French RM. From chicken squawking to cognition: Levels of description and the computational approach in psychology. Psychologica Belgica. 1996;36. doi: 10.5334/pb.892 [DOI] [Google Scholar]
  • 18.Strogatz SH. Nonlinear Dynamics And Chaos: With Applications To Physics, Biology, Chemistry, And Engineering. Perseus books; 2001. [Google Scholar]
  • 19.Dennet DC. Consciousness Explained. Little Brown & Co; US; 1991. [Google Scholar]
  • 20.Izhikevich ME. Neural excitability spiking and bursting. Int J Bifurcat Chaos. 2000;10:1171–1266. doi: 10.1142/S0218127400000840 [DOI] [Google Scholar]
  • 21.Gerstner W, Kistler WM. Spiking Neuron Models. 1st ed. Cambridge University Press; 2002. [Google Scholar]
  • 22.Picton P. Neural Networks. 2nd ed. Palgrave, New York; 2000. [Google Scholar]
  • 23.Jäger H, Lukoševičius M, Popovici D. Optimization and Applications of Echo State Networks with Leaky Integrator Neurons. Neural Networks. 2007;20:335–352. doi: 10.1016/j.neunet.2007.04.016 [DOI] [PubMed] [Google Scholar]
  • 24.D Mathis W, Mozer M. Conscious and unconscious perception: A computational theory. In: Proceedings of the Eighteenth Annual conference of the Cognitive Science Society. Hillsdale, NJ; 1996. p. 324–328.
  • 25.Cleeremans A, Servan-Schreiber D, Mc Clelland J. Finite State Automata and Simple Recurrent Networks. Neural Computation—NECO. 1989;1:372–381. doi: 10.1162/neco.1989.1.3.372 [DOI] [Google Scholar]
  • 26.Jäger H. The “echo state” approach to analysing and training recurrent neural networks. German National Research Center for Information Technology; 2001.
  • 27.LeCun Y, Bottou L, Orr GB, Müller KR. Efficient BackProp. In: Orr GB, Müller KR, editors. Neural Networks: Tricks of the Trade. Berlin, Heidelberg: Springer Berlin Heidelberg; 1998. p. 9–50. Available from: 10.1007/3-540-49430-8_2. [DOI] [Google Scholar]
  • 28.Rojas R. Neural Networks—a systematic introduction. 1st ed. Springer-Verlag, New York; 1996. [Google Scholar]
  • 29.Verstraeten D, Schrauwen B, d’Haene M, Stroobandt D. An experimental unification of reservoir computing methods. Neural Networks. 2007;20:391–403. doi: 10.1016/j.neunet.2007.04.003 [DOI] [PubMed] [Google Scholar]
  • 30.Lukoševičius M, Jäger H, Schrauwen B. Reservoir computing trends. Künstliche Intelligenz. 2012;26:365–371. [Google Scholar]
  • 31.Appeltant L, Soriano MC, Van der Sande G, Danckaert J, Massar S, Dambre J, et al. Information processing using a single dynamical node as complex system. Nat Commun. 2011;2:468. doi: 10.1038/ncomms1476 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Lukoševičius M. In: Montavon G, Orr GB, Müller KR, editors. A Practical Guide to Applying Echo State Networks. Berlin, Heidelberg: Springer Berlin Heidelberg; 2012. p. 659–686. Available from: 10.1007/978-3-642-35289-8_36. [DOI] [Google Scholar]
  • 33.Herzog MH, Drissi-Daoudi L, Doerig A. All in Good Time: Long-Lasting Postdictive Effects Reveal Discrete Perception. Trends Cogn Sci. 2020Oct;24(10):826–837. doi: 10.1016/j.tics.2020.07.001 [DOI] [PubMed] [Google Scholar]
  • 34.Fekete T, Van de Cruys S, Ekroll V, van Leeuwen C In the interest of saving time: a critique of discrete perception Neuroscience of Consciousness, Volume 2018, Issue 1, 2018, niy003 Available from: 10.1093/nc/niy003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Bengio Y, Lee DH, Jorg B, Mesnard T, Lin Z. Towards Biologically Plausible Deep Learning. arXiv:1502.04156. Available from: https://arxiv.org/abs/1502.04156.
  • 36.Bengio Y, Mesnard T, Fischer A, Zhang S, Wu Y. STDP-Compatible Approximation of Backpropagation in an Energy-Based Model. Neural Computation. 2017;29(3):555–577. doi: 10.1162/NECO_a_00934 [DOI] [PubMed] [Google Scholar]
  • 37.Poltoratski S, Tong F. Hysteresis in the dynamic perception of scenes and objects. J Exp Psychol Gen. 2014Oct;143(5):1875–92. doi: 10.1037/a0037365 [DOI] [PubMed] [Google Scholar]
  • 38.Kim R, Sejnowski TJ. Strong inhibitory signaling underlies stable temporal dynamics and working memory in spiking neural networks. Nature Neuroscience. 2021Jan;24(1):129–39. doi: 10.1038/s41593-020-00753-w [DOI] [PubMed] [Google Scholar]
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009344.r001

Decision Letter 0

Daniele Marinazzo

27 Apr 2021

Dear Dr. Keuninckx,

First of all let me apologise once again for the delay. The current circumstances contribute to the further stretch of the queues of the turnover time distribution.

Thank you very much for submitting your manuscript "The color phi phenomenon: not so special, after all?" for consideration at PLOS Computational Biology.

As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The paper was overall very well received, but some issues need to be addressed. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: Review is uploaded as an attachment.

Reviewer #2: The authors present an echo state model, which can capture the color phi phenomenon. The results are interpreted within the broader context of the philosophical questions about the time course of consciousness. The research questions is timely and interesting. The presentation, the specific research question, and the model in action were not always clear to me.

Majors.

It is not easy to get what the model exactly does because the presentation is sometimes quite wooly and at other, important instances very short. For example, why presenting a model that does not work? Why having this lengthy detailed description of integrate and fire neurons? Why not explaining the echo state network right away and more clearly.

If I got it correctly, the model is more or less a low pass filter, which translates the input into a perceptual space producing some neural activity in the middle between, say the right blue and left red input, which mimics the phi phenomenon because there is activity at a location where nothing is produced. OK this is interesting.

However, what are the implications? It seems that the authors want to attack certain notions about the time course of consciousness by Dennett and us by showing that a simple model can do the job. It remains however unclear to me what exactly they want to attack. Or maybe they do not want to criticize things? At other instances it reads like that the authors want to naturalize philosophical thoughts. Please clarify.

A simple motion detector can also detect apparent motion and even the phi phenomenon when equipped with a color detector. Thus, even more simple mechanisms may exist. What do we learn from the authors model then?

It seems to me also that the network does not perceive motion, it just creates a signal in the middle of the trajectory, isn't it?

There seems to be also a misunderstanding about our labels. These are not Dennett's time stamps. Our model claims that motion and other temporal aspects of a stimulus are rendered conscious at one short moment as a vector of feature labels, such as (Shape=line, duration=100ms, motion=l-r, color=red). Hence, the motion is not perceived during the 100ms of the presentation, not even for 100ms, just as a number, namely, the output of a duration detector.

In this respect, it seems that the model is different from our model since it outputs a continuous stream. I have the feeling the authors want to defend a extensionalism (Dainton, Stanford Encycl.) and is a bit in the spirit of other models for example by Piper. Here, there could by a real controversy. These things are spelled out in Herzog et al. (2020).

I think also Dennett would agree with the model. His time stamps are for representation of time, a very different questions than the authors focus on.

In any case, I think the authors should make clearer what they are aiming for.

Our model is mainly inspired by post-dictive effects as they occur in feature fusion, feature fusion with TMS, the sequential metacontrast paradigm (SQM), and other where a later presented stimulus integrates with a previously presented one. Hence, the question is to what extent the model can address such phenomena, for example, a simple feature fusion percept where a red disk followed by a green disk produces are yellow disk.

Minors:

"The core aspect of what needs to be explained in the case of the CPP is the subjectively experienced reversal of the order in which stimuli occur." Why reversal? There is no reversal in app motion. The first disk is perceived before the second one.

"We conclude that the CPP, while an intriguing visual illusion in its own right, might be nothing more than a consequence of the inherent dynamics that characterize information processing in the brain." The CPP is not an inherent dynamic. It may be explained by one. What is the argument here? That Dennett has claimed that it needs sophisticated processing?

What do we learn from the subplot "neuronal state" in fig 6?

English is not always clear: For example: "When undriven by a stimulus ~x(t), the state \\falls" downward towards a local minimum of and scalar function V (~y) : R^K -> R, much like a golf ball rolling downward to the lowest point of a valley it happens to find itself in." Undriven = no stimulus presented. States don't fall. I think state is the wrong word here because at each time point of "falling" is a state. I think a reader who understands the Nabla Operator knows what an attractor state is and thus needs no metaphors. For those who do not know, I am not sure whether the metaphor helps. Mathematically: why is a global minimum reached and not a local one?

"Since it is not straightforward to manually construct a neuronal circuit that shows the complete CPP, we turn to neural network paradigms to find one." Rather void sentence.

Lines 166 ff could be removed. This should be clear to readers of Plos Comp.

Figures are low resolution. Please present them in the text together with the captions.

"merely artifacts of the inherent dynamical and nonlinear behavior of such systems." Why artifacts? Again, it seems that authors want to say "look we do not need complicated machinery" but is unclear who they think claimed otherwise.

"In Ref. [7]," spell out please.

Mathis and Mozer, no year is given.

Signed

Michael Herzog

Dainton, B. (2018) Temporal consciousness. In The Stanford

Encyclopedia of Philosophy Winter 2018.

Michael H. Herzog, Leila Drissi-Daoudi, and Adrien Doerig (2020).

All in Good Time: Long-Lasting Postdictive Effects Reveal Discrete Perception. TICS.

Piper, M.S. (2019) Neurodynamics of time consciousness: an

extensionalist explanation of apparent motion and the specious

present via reentrant oscillatory multiplexing. Conscious. Cogn.

Reviewer #3: The authors address what is on the face of it an intriguing mystery, the so-called color-phi illusion which has been argued to evidence postdictive processing or even discrete perception by some. Rather than offering yet another conceptual analysis, the authors take a much more sensible approach to the problem, and set out to show that this phenomenon could result from generic facets of the dynamics of artificial neural networks. The authors do well in presenting the rational of their (and similar) modeling approach, stressing the logic in first demonstrating the general nature of an effect in abstract modeling.

As a modeling device, the authors choose to utilize leaky integrator neurons, which are exhibit several of the properties one would want from neural building blocks not the least of which is a temporal aspect to their response. They proceed to present three simple models in increasing order of complexity, a model for temporal reversal, a model for the phi phenomenon, and a model incorporating both.

The authors modelling efforts and analysis are a timely addition to the recent discussion triggered by Herzog et al’s discrete model of perception. As such they expose a possible weakness in the original argument, namely that time reversal has to be construed as postdiction. However, I think the paper would benefit from some additional analyses as listed below before it is published:

Major comments:

I really like the later parts of the discussion. I think they highlight some major issues with the notion of imposing discreetness on neural dynamics, and the conceptual link between discreteness and consciousness. I recommend that this somewhat agnostic view be represented also in the earlier discussions.

Order reversal model: The authors say that the parameters had to be handpicked to achieve reversal – but why couldn’t they use simulation, and parametrization to produce a phase map to demonstrate the robustness of the effect?

The authors acknowledge the limitations of their phi phenomenon model, but again do not directly assess the robustness of even this limited result (the movie is a step in the right direction).

Again in the CPP model we are not offered insight as to robustness – e.g. what is the effect of the sparseness parameter? The spectral radius? Etc. (i.e. do we still get CPP in 1.2% of random networks?)

Minor comments

“A four toy model neuron model” a four neuron toy model?

I’m not sure eq. 5 helps illustrating the point – it just complicates a very straight forward example.

Why does the output layer have to be conceived as a cartesian theater? Can’t it be taken to be as behavioral output, thus implying that sensory motor mapping is the substrate of representation?

Why is the ESN constrained not to respond to simultaneous stimuli?

As an aside - our recent paper offers some analysis and points to existing evidence from VSDI to the proposed mechanism for phi illusions (In the interest of saving time: a critique of discrete perception)

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

Reviewer #3: None

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Jake Gavenas

Reviewer #2: Yes: Michael Herzog

Reviewer #3: Yes: Tomer Fekete

**********

Have all data underlying the figures and results presented in the manuscript been provided?

Large-scale datasets should be made available via a public repository as described in the PLOS Computational Biology data availability policy, and numerical data that underlies graphs or summary statistics should be provided in spreadsheet form as supporting information.

Reviewer #1: No: The article was a simulation study, so no data was collected. They say they will include the python code in a zip file in the supplementary material, but that is not done with this manuscript.

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Attachment

Submitted filename: CPP_review.docx

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009344.r003

Decision Letter 1

Daniele Marinazzo

13 Jul 2021

Dear Dr. Keuninckx,

Thank you very much for submitting your manuscript "The color phi phenomenon: not so special, after all?" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we will accept this manuscript for publication.

The reason why this is still technically a "minor revision" decision is that we would like to give you the chance to address the residual comments by Dr. Herzog, which would likely bring to a clearer final product.

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

***********************

A link appears below if there are any accompanying review attachments. If you believe any reviews to be missing, please contact ploscompbiol@plos.org immediately:

[LINK]

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #2: The authors have clarified and improved many things. Still, I am not really happy with the presentation. The main problem is the imbalance of context, which makes it hard to get what is important. Basic, well known stuff is explained in great detail and other things, such as the Nabla operator (I guess most readers are not familiar with), are just mentioned en passant (the fix "readers familiar..." does not really work). I will not urge the authors to change it, if they do not want to. It is their paper.

The second thing it is still with the interpretation what the implications are. For example, in the abstract it is written that the "color phi phenomenon is...an artifact of inherent dynamics". This does not make to much sense too me. Why an artifact and not a feature? It seems the authors what to attack certain positions, but which ones? The term postdiction is indeed often used for mechanism rather than a term for psychophysical effects. What do the authors have in mind?

I have some small comments in the attached pdf.

Signed

Michael Herzog

Reviewer #3: The authors did a thorough job in addressing the reviewers comments, thus I gladly endorse it for publication.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

Reviewer #3: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: Yes: Michael Herzog

Reviewer #3: Yes: Tomer Fekete

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

References:

Review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript.

If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Attachment

Submitted filename: PCOMPBIOL-D-21-00262_R1_reviewer.pdf

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009344.r005

Decision Letter 2

Daniele Marinazzo

12 Aug 2021

Dear Dr. Keuninckx,

We are pleased to inform you that your manuscript 'The color phi phenomenon: not so special, after all?' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

Daniele Marinazzo

Deputy Editor

PLOS Computational Biology

***********************************************************

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1009344.r006

Acceptance letter

Daniele Marinazzo

31 Aug 2021

PCOMPBIOL-D-21-00262R2

The color phi phenomenon: not so special, after all?

Dear Dr Keuninckx,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Andrea Szabo

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Animation. An animation showing the creation of stable states by feedback in a single toy model neuron network used for demonstrating priming.

    (WMV)

    S2 Animation. An animation of a bistable system with positive input stimulus masked by negative stimuli with decreasing interval.

    (WMV)

    S3 Animation. An animation of the phi phenomenon, the appearance of a “ghost” percept in dynamical neuronal network representing an abstract three pixel visual system.

    (WMV)

    S4 Animation. An animation of the color phi phenomenon based on the timeseries found in Fig 7A as “perceived” by the ESN during testing.

    (WMV)

    S1 Data. Contains the Python scripts that directly generate the graphs of this article.

    (ZIP)

    Attachment

    Submitted filename: CPP_review.docx

    Attachment

    Submitted filename: Reply_AC_LK.pdf

    Attachment

    Submitted filename: PCOMPBIOL-D-21-00262_R1_reviewer.pdf

    Attachment

    Submitted filename: Reply_AC_LK_3.pdf

    Data Availability Statement

    All data underlying our results stem from model simulations in the form of Python scripts and are included in the Supporting Material.


    Articles from PLoS Computational Biology are provided here courtesy of PLOS

    RESOURCES