Skip to main content
PLOS Computational Biology logoLink to PLOS Computational Biology
. 2023 Nov 7;19(11):e1011574. doi: 10.1371/journal.pcbi.1011574

Inferring causal connectivity from pairwise recordings and optogenetics

Mikkel Elle Lepperød 1,2,*, Tristan Stöber 2,3,4, Torkel Hafting 1, Marianne Fyhn 2,5, Konrad Paul Kording 6
Editor: Marcus Kaiser7
PMCID: PMC10656035  PMID: 37934793

Abstract

To understand the neural mechanisms underlying brain function, neuroscientists aim to quantify causal interactions between neurons, for instance by perturbing the activity of neuron A and measuring the effect on neuron B. Recently, manipulating neuron activity using light-sensitive opsins, optogenetics, has increased the specificity of neural perturbation. However, using widefield optogenetic interventions, multiple neurons are usually perturbed, producing a confound—any of the stimulated neurons can have affected the postsynaptic neuron making it challenging to discern which neurons produced the causal effect. Here, we show how such confounds produce large biases in interpretations. We explain how confounding can be reduced by combining instrumental variables (IV) and difference in differences (DiD) techniques from econometrics. Combined, these methods can estimate (causal) effective connectivity by exploiting the weak, approximately random signal resulting from the interaction between stimulation and the absolute refractory period of the neuron. In simulated neural networks, we find that estimates using ideas from IV and DiD outperform naïve techniques suggesting that methods from causal inference can be useful to disentangle neural interactions in the brain.

Author summary

Understanding the interactions between neurons is essential for grasping brain function. Neuroscientists often face a challenge when stimulating multiple neurons simultaneously: determining which neuron influenced another. The use of optogenetics, where neurons are controlled with light, has improved precision, but when several neurons are activated at once, it’s difficult to discern the specific source of influence. In this study, we discuss the potential pitfalls of overlapping influences. We introduce techniques from econometrics, namely instrumental variables (IV) and difference in differences (DiD), as a method to better identify neuron-to-neuron interactions. Our tests on simulated neural networks indicate that these techniques can be more effective than some traditional methods. By integrating approaches from neuroscience and econometrics, we aim to enhance methods for understanding neural connections in the brain.

Introduction

A central goal of neuroscience is understanding the neural mechanisms or causal chains giving rise to brain activity underlying perception, cognition, and action. Although simultaneously occurring processes at many spatial and temporal scales contribute to brain activity, one of the defining features of neural networks is their connectivity. Connectivity directs neuronal communication. This communication is causal; connection strength controls the causal effect when one neuron signals to another neuron. The human brain contains about 86 billion neurons [1] with countless connections. A complex system of this scale is incredibly hard to understand because of the numerous ways the contributing elements may interact internally [2]. While observing correlations within such systems is relatively easy, transitioning from observed correlations to a causal or mechanistic understanding is hard. For example, suppose we observe a given correlation between neurons A and B. In that case, this may be because of a causal interaction between A and B, because of common input (confounder bias), or as a side effect of correcting for other effects (collider bias). After all, the same activities may emerge from different and distinct causal chains [3, 4].

In today’s typical studies in systems neuroscience, we only record from a small subset of all the neurons [5]. The data obtained from such recordings, e.g., from electrophysiology or in vivo imaging, is observational, as opposed to interventional involving perturbations. From observational data we generally cannot know how much of the observed activity was caused by observed or unobserved confounding. Unobserved neural activity confounds estimates of causal interactions and makes it challenging to reveal underlying mechanisms [6]. Relationships estimated from observational data in the presence of confounders may contain significant errors and lead to incorrect conclusions [7].

Confounding is the big threat to causal validity [8] irrespective of using simple regressions or advanced functional connectivity techniques [912]. Inference techniques, such as maximum entropy-based inverse Ising inference [13], and generalized linear models (GLMs) [14] may give solutions to confounding from observed variables as they support “explaining away” background activity [9]. However, this is only a meaningful strategy if most neurons are included in the recordings. Even a fully observed system is only identifiable under certain assumptions about nonlinearity or noise [15, 16]. For example, systematic errors can occur due to confounding emergent network states [17]. Moreover, fully observable scenarios are rare in experimental settings, especially in the mammalian brain with its many millions of neurons. Brain data will, therefore, rarely satisfy the criteria necessary for causal identifiability. The general conclusion from causal inference is that naïve regressions in partially observed systems will not reveal causality. We aim to circumvent this by developing new methods to identify effective connectivity between arbitrary neuron pairs.

In this paper, we restrict our focus to causal effects related to measuring effective connectivity. This is not just the direct influence of one neuron on a downstream neuron. Instead, it is defined as the change in postsynaptic spike probability caused by an extra presynaptic spike. Therefore, effective connectivity includes direct effects, bi-synaptic and other effects from the network. During perturbations, this effect is directly measurable. On the other hand, estimating the direct causal effect will generally require some level of modeling. Estimating the direct causal effect from the effective connectivity is an interesting and, to our knowledge, unresolved problem in computational neuroscience. However, in many cases, effective connectivity is meaningful, e.g., when reconstructing a circuit with equivalent information flow. After all, a neuron will exert its effect through many paths. This paper thus focuses on effective connectivity, which we term causal influence.

To estimate the causal influence of one neuron on another neuron, stimulating the presynaptic neuron is the gold standard. Indeed, a standard definition of causality is the effect of an intervention changing one variable in the system, independently of all the other variables [8]. If we can experimentally stimulate and fully control single neurons, the ability to estimate causal relationships by regression is possible. However, gold-standard perturbations are experimentally challenging (it is challenging to stimulate just one neuron and get complete control) and may yield low cell counts because it requires intra- or juxtacellular [18, 19], or two-photon stimulation [2026]. Therefore, it would be highly desirable to obtain causality from widefield one-photon optogenetic [27, 28] stimulation combined with neural recordings of large neuron populations. We focus on widefield one-photon optogenetic stimulation and refer to this simply as optogenetic stimulation.

Interpreting results from optogenetic stimulation in terms of causal interactions is often difficult. In most experimental settings, optogenetic stimulation will affect multiple neurons simultaneously. Hence, the stimulus will produce a distributed pattern of activity which then percolates through the network of neurons. Postsynaptic activity induced by stimulation could thus, in principle, come from any of the stimulated neurons, introducing problematic confounding.

For inspiration for resolving the confounding problem induced by optogenetic stimulation, we look to other disciplines addressing the confounding problem or, more precisely, confounded regressors. The inference of causality from observational data has been addressed in the fields of statistics [8], machine learning [4], and econometrics [7]. These fields have extensively worked on methods to estimate causality in the face of unobserved confounding and may provide clues for resolving these challenges in neuroscience.

A commonly used approach towards causal inference in economics is instrumental variables (IV), which has a genuinely long and exciting intellectual history [2931]. The IV can be considered to emulate the treatment assignment in a randomized controlled trial. It is a measurable variable that directly affects one of the variables of interest and only indirectly affects the rest.

Let us consider a common problem in neuroscience, the estimation of the influence of a neuron with spike activity X onto another neuron with activity Y. We want to estimate the causal effect β from one neuron onto another. One approach would be to estimate the conditional mean function E[Y|X] with regression. For simplicity, we can assume the linear model Y = βX + U. Here U represents other factors in the brain other than X and Y, that are measured. Because of the many neurons and high degree of connectivity in the brain, the regressor X is confounded; it can be correlated to U through some unknown functions p, f (p(X) = f(U)). This relationship implies that the regression estimate of β will not estimate the magnitude of causation from presynaptic to postsynaptic, but its association. Regression can thus generally not provide causal insights.

IVs are a way of getting around this problem. Let us say we have a stimulation signal Z, which we will call an instrument and that this instrument affects X. However, the setting is generally not that no stimulation means X = 0 and stimulation X = 1. Instead, realistic stimulation protocols would rather increase the probability of spiking E(X|Z = 1) > E(X|Z = 0). If stimulation only adds say 1/10th of the probability of a spike (e.g. p = .3 instead of p = .2) then our estimates of causal effect sizes would be off by a factor of 10. The intuition behind the IV approach is that it estimates how much the stimulation affects X and then only uses this stimulation-related aspect of X to identify the influence on Y. So, if the stimuli are random in time and uncorrelated with U, we may use Z as an IV and obtain the true causal effect using the so-called Wald estimator [32] given by

β^IV=E[Y|Z=1]-E[Y|Z=0]E[X|Z=1]-E[X|Z=0]. (1)

This formula is standard for IV identification found in textbooks [7, 33]; see S1 Appendix for proof of identification, both parametrically and non-parametrically. In other words, IVs are a trick to use upstream randomness to identify a causal system.

For an instrument to be “good”, it must be unaffected by other variables, which is the case for single-neuron stimulation. However, the stimulation cannot be used as an IV for multi-neuron stimulation because it will drive parts of the rest of the brain dynamics, it will affect U, violating the key assumption of IV. In the brain, almost everything is affected by the network state. However, certain variables can be more or less affected. For example, the overall activity of the network is due to slow and strongly non-random dynamics. In contrast, the temporal pattern of when a neuron is in a refractory state may be in good approximation random. For example, if neurons are spiking according to conditional Poisson distributions, their exact timing conditioned on the network state will be random. While refractoriness may not be perfectly random, the exact spike-times are notoriously difficult to predict [9], suggesting that refractoriness is quite random. The absolute refractory period may serve as an IV because a neuron is randomly unaffected by stimuli during refractoriness.

When dealing with temporal data, we can also use an additional method commonly used for causal inference: difference in differences (DiD). The idea is that we have two time-series that are identical (or more specifically, parallel) apart from a perturbation affecting only one of them. For example, we may have two neural trajectories, one with perturbation and one without. By subtracting out the unperturbed one we can get rid of biases introduced by the ongoing dynamics. DiD in this sense is a way of producing a synthetic estimation problem that is less affected by biases (or unaffected if its assumptions are given).

For our stimulation problem we can formalize the effect of treatment X in a longitudinal study with two time points Y*, Y. We can calculate the differences (E[Y|X = 1] − E[Y*|X = 1]) − (E[Y|X = 0] − E[Y*|X = 0]) where the two differences in parenthesis denote a treatment group X = 1 and a control group X = 0, respectively. For example, we may measure baseline firing rates Y* in the primary visual cortex during standardized visual stimulation in control and treatment groups of animals over several trials to estimate E[Y*|X = 0] and E[Y*|X = 1]. Then, after giving the treatment animal group X = 1 a drug, you measure both groups to obtain E[Y|X = 0] and E[Y|X = 1]. The estimates obtained from the control group correct for any changes over time due to other factors than the treatment itself, e.g. animals have different hormone levels during pre and post-treatment; see Fig 1g for an intuitive sketch. It is also possible to separate trials from a single individual into control and treatment blocks and adapt the DiD framework. When estimating connectivity, we can compare pre- and post-spikes in otherwise matched brain states using DiD to improve causality estimates.

Fig 1. Instrumental variables corrects spurious correlations induced by optogenetic stimulation.

Fig 1

a) Simple network containing three neurons shows stimulation configuration with blue laser light and the connections illustrated by axon touching dendrites. b) For each stimulation pulse, three spike detection windows are placed relative to stimulus onset time, pre, during, and post, which is stored as Z, X, and Y, respectively, where Z and X represent either A or B while Y represents C. c) Cross-correlation histogram (CCH) with the vertical axis showing the conditional expected value E[Mi = 1|shift(Mj, τ) = 1], i, j ∈ {A, B, C}, τ = 0, 1, …, 9 with shifts in timesteps τ. d) Transmission probability estimates as a function of ground truth between AC, and BC. e) Directed acyclic graph (DAG) showing the relations between variables. The OLS method only uses X to estimate Y, which fails because of the confounding factor U; the refractoriness denoted Z affects X without being affected by U and can therefore be used as an IV. f) Conditional expectancies where the true weight = 7 rightmost values in d. The upper panel is mirrored around zero to aid visualization. The estimates in the lower panel represent differences between solid and dashed lines in the upper panel. The shaded gray area depicts the estimation window, and cyan shading indicates the stimulation window, which in this example also depicts DiD reference (Y*). g) A sketch of the intuition behind DiD.

Here we show that combining the IV and DiD techniques allows for estimating the effective connectivity between neuron pairs under optogenetic stimulation. We first show how conventional optogenetic stimulation of neurons introduces confounding. This confounding effect is then simulated in a simple network of three neurons using a Binomial GLM approach. With this simple model, we show that by using the refractory period as an IV, we can distinguish between connected and unconnected neuron pairs. Combining the IV estimate with DiD, we correct for biased neuronal activity due to state differences before and after stimuli. We compare these estimates with a naïve, although widely used, cross-correlation histogram (CCH) method that fails to distinguish respective pairs. Furthermore, we infer effective connectivity in a simulated network of randomly connected excitatory and inhibitory neurons with distributed synaptic weights. With this data, we compute errors of the IV and DiD methods and show that they are robust to different simulated network states. Furthermore, we analyze false positive and false negative errors and goodness of fit on synaptic weights with pairwise assessments using ordinary least squares fitting (OLS) and our IV/DiD combination.

The observed differences between the IV/DiD and the CCH estimates underline the importance of considering potential confounding when estimating connections based on neural activity measurements.

Results

Confounding

When we simultaneously stimulate many neurons and observe the activity of a postsynaptic neuron after stimulation, it is hard to know which subset of the stimulated neurons caused the activity in the putative postsynaptic neuron—after all they are simultaneously stimulated. To test inference methods and build intuitions in such situations we design a simulation tool as there are no, to our knowledge, any clean dataset containing both recordings and ground truth available.

We employ a linear-nonlinear cascade model, similar to spike response models with soft threshold [14, 3436] in the form of a binomial GLM given by

pi,t+1=P(Mi,t+1=1|Mi,t,Mj,t,Mi,t-1,Mj,t-1,...,Mi,t-H,Mj,t-H)pi,t+1=σ(τ=t-Ht(Mi,τr(τ)+jN(i)WijMj,τc(τ))-b+Ui(t))Mi,t+1Bernoulli(pi,t+1). (2)

These equations describe a probabilistic model for a higher-order discrete-time Markov process, where the state of node i at time t + 1, denoted by Mi,t+1, depends on the previous states of the node itself and its neighbors jN(i) up to a history length H, with timesteps of 1ms. The probability pi,t+1 is calculated using a logistic (sigmoid) function σ and Bernoulli-distributed random variables. The function r represents refractoriness which is set to a large negative value after a spike representing absolute refractoriness followed by an exponential term representing relative refractoriness. The constant b reflects baseline activity, and the coupling function c represents response dynamics modeled by an exponential term; see Binomial GLM simulation for further details. We thus have a simulation that we can use to test our methods; for parameters see Table 1.

Table 1. Simulation parameters.

Name Fig 1 Fig 2 Fig 3 Fig 4 Fig 5 Fig 6
γ S 5.0 6.0 6 [0, 8] [0, 8] 3.0
λS 50.0 50.0 50.0 50.0 50.0 50.0
λSmin 10.0 10.0 10.0 10.0 10.0 10.0
λSmax 200.0 200.0 200.0 200.0 200.0 200.0
τ ex 10.0 10.0 - 10.0 10.0 10.0
γ ex 2.0 [0, 7] - 2.0 2.0 2.0
λex 100.0 100.0 - 100.0 100.0 100.0
λexmin 30.0 30.0 - 30.0 30.0 30.0
λexmax 400.0 400.0 - 400.0 400.0 400.0
τ in 10.0 10.0 - 10.0 10.0 10.0
γ in -5.0 [0,-7] - -5.0 -5.0 -5.0
λin 100.0 100.0 - 100.0 100.0 100.0
λinmin 30.0 30.0 - 30.0 30.0 30.0
λinmax 400.0 400.0 - 400.0 400.0 400.0

To illustrate confounding effects, we first simulated a network comprised of three neurons (i ∈ {A, B, C}) shown in Fig 1a. The neurons have a baseline activity and are additionally driven by Poisson-distributed excitatory and inhibitory inputs of a timescale of 10ms. Together, this simulates slow network effects in the background noise U. The interactions between neurons create a simple system to explore causality: A and B receive stimulation, but only B drives C (WCA = 0, WCB > 0). We can then ask how to statistically resolve that only B drives C, a task called causal discovery, and how strongly B drives C, a task called causal inference [4, 7].

To estimate the actual causal influence of a stimulated neuron on postsynaptic neurons, we need to distinguish the influence of one stimulated neuron from the influence of other stimulated neurons. This procedure is challenging since the two neurons receive the same inputs.

One intuition is to use unsuccessful stimulations, i.e. instances where one neuron is not responding due to membrane potential fluctuations, maybe a first suggestion which is fine if neurons are independent outside stimulation periods. However, membrane potentials can be correlated on long timescales such as with brain oscillations [37] or up-down states [38]. During these long timescale states, the network state confounds the effect of stimulations. Moreover, a correlation-based measure fails to discover the correct network structure. The CCH and OLS estimates described below show the effect on the stimulated neurons (Fig 1c). Since the stimulation affects both A and B simultaneously, it induces a strong correlation between A and B. This further confounds the system by rendering the correlation between B and C nonzero. When comparing the peak cross-correlations with baseline as described in Cross correlation histogram, specifically Eq (18), β^CCH estimates the connectivity of A and C to be nonzero (Fig 1d). A naïve reading of this result may suggest causal influences of both A and B on C [6]. This raises the question of whether there is any way to avoid this problem.

Instrumental variables

Central to our approach is the concept of counterfactuals or potential outcomes, which allows us to compare what actually happened with what could have happened under different circumstances. Put simply, this means imagining what would have occurred if a subject had received a different treatment or intervention than what they actually received. We use this approach to estimate the causal effect of a given treatment or intervention on a specific outcome. Using this idea the causal effect is the difference Yi|do(T=1)Yi|do(T=0) where Yi|do(T=1), shortened to Yi(T = 1), denoting the value Y of subject i given that we force some treatment T = 1, denoted by the do operator. For example, the treatment T can indicate that a patch electrode gives a pulse i, and Yi indicates the increase in membrane potential. However, the causal effect is typically estimated by the average treatment effect (ATE) defined by the expected difference E[Y(T = 1) − Y(T = 0)] since we can not observe the same identical trial with and without a given treatment.

In what follows, X, Y, Z are random variables with binary values, and indexing refers to a trial, where we assume all trials are independent. Further, X(Z = 1) refers to the value X when Z = 1. We use Z = 1 to denote if an upstream spike preceded stimulus, and thus Z = 1 indicates that X = 0 or, in other words, X did comply with treatment. This is opposite to conventional notation, where Z = 1 → X = 1, which might confuse readers with previous knowledge from causal inference. Here we want to obtain an estimate of E[Y(X = 1) − Y(X = 0)] which we can write as E[Y(X = 1)] − E[Y(X = 0)] when they are independent.

The IV technique has been used extensively in econometrics and can provide a provable causal estimate of β given three main assumptions [33].

  • (i)

    Relevance: The instrument (Z) has a causal effect on the the regressor (X).

  • (ii)

    Exclusion Restriction: There must be no direct influence of the instrument (Z) on the outcome variable (Y); its effect must be fully mediated through the regressor (X).

  • (iii)

    Instrumental Independence: The instrument (Z) and the outcome variable (Y) must not share causes.

The validity of these assumptions is central when using the IV approach.

A final assumption (iv) is required to prove the causal validity of the estimator Eq (4), either a linear parametric assumption or a non-parametric assumption with effect homogeneity or monotonicity [33]; see S1 Appendix for proofs of identification following these assumptions.

Refractoriness as an IV

One way to avoid confounding may be a timescale separation—using something that is random on a short timescale. Arguably, a neuron’s refractoriness is such a variable. Suppose a neuron is in its refractory period (typically about 1–4ms, [39]). In that case, it will be less likely to fire, and during the absolute refractory period, no amount of stimulation could make it spike. Refractoriness gives us an interesting way of inferring causality by comparing stimulation of when a neuron can spike and when the neuron is muted. We would effectively not use the spikes but the missing ones for causal inference.

To think through confounding effects let us introduce some meaningful notation. Let X,Y,Z[0,1]NS denote random variables that indicate whether a neuron spiked relative to the stimulus during trials i = 0, 1, 2, …, NS. We evaluate pairwise relations: Z and X reference the spiking of one specific stimulated neuron at two time points, and Y reference one specific downstream neuron. In other words, X and Z are extracted from one row (Mi) and Y from another (Mj) given in Eq (2) using windows relative to stimulus onset. The window sizes were 2 ms with stimulus relative time given by -2,0,2 for Z, X, Y respectively as shown in Fig 1b. As further depicted in Fig 1b, Z0 = 0 indicates that the putative upstream neuron did not fire within a set time window before the first optogenetic stimulus, where subscripts indicate trial numbers. Furthermore, X0 = 1 and Y0 = 1 indicate the upstream and downstream neurons fired in this first trial. We seek to estimate whether a response Y0 = 1 was caused by X0 = 1. The putative upstream neuron is refractory if Z0 = 1; in our simulations, the neuron fired 1–2ms before stimulus onset. With this notation we can now meaningfully address the problem.

We can be statistically precise about the causal reasoning. Consider the directed acyclic graph (DAG) [8] representing variables and their relations (Fig 1e). Here, X represents a putative upstream neuron e.g., A or B, and Y represents the putative downstream neuron e.g. C. The set U contains unobserved network activity and the stimulus. During IV estimation, we use refractoriness which is (at least in first-order approximation) assumed to be random and influences stimuli response to infer the causal interaction between neurons. This approach is known to satisfy Pearl’s criteria for causal inference [40], and IVs can reveal real causality.

In the following, we aim to estimate the effective connectivity defined as the difference in conditional probability βijP(Mi = 1|do(Mj = 1)) − P(Mi = 1|do(Mj = 0)). This means that the causal effect is defined as the difference in the post-synaptic neurons firing probability between a case where we set the pre-synaptic activity to one and the case where we set it to zero. Let us consider the case where there is no confounding, the neurons are conditionally independent and we can simplify:

βij=P(Mi=1|Mj=1)-P(Mi=1|Mj=0)

Let us go through a concrete example. Let there be a network of two neurons i and j. Let us assume i is not refractory (r = 0) and a constant coupling c = 1. With only one spike from neuron j after one time step and no noise (U = 0), we can derive βij = σ(Wijb) − σ(−b). In this sense βij is the causal effect induced by the underlying weight Wij and we can directly see that these two will be monotonously related. With refractoriness, however, we fit a linear model to represent the ground truth βij and compute errors as the mean absolute error to estimate E[|β^-β|]; see Computing errors for further details. The case with refractoriness will be more complex, and sets up the problem we will be working on in the rest of this paper.

Let us start with the naïve estimate of β: the OLS estimate. For binary variables (allowing only X = 0 and X = 1), it can be written as the difference of conditional expectations [7]

β^OLS=E[Y|X=1]-E[Y|X=0]. (3)

Note that here we implicitly condition on the stimulus due to stimuli onset selection when building the variables X, Y, Z. When X is confounded (endogenous), i.e. E[X|U] ≠ 0, the estimate β^OLS will be biased. What kind of biasing activities do we expect? Brains have slow dynamics, such as up-down states. Through those activities, pre- and post-synaptic neurons will be strongly correlated regardless of the actual synaptic connectivity. This consideration already presages the use of DiD below, which aims to minimize such biases. Importantly, OLS is an important baseline, used in many neuroscience settings.

The idea behind IV is to use a source of randomness that is local as a means of obtaining randomness. Here we want to use refractoriness Z as an instrument. The IV-estimator for our binary case (see Wald estimator [32]) is given by

β^IV=E[Y|Z=0]-E[Y|Z=1]E[X|Z=0]-E[X|Z=1] (4)

Notice the flipped Z compared to Eq (1) as it now represents refractoriness, not stimuli that further lead to E[X|Z = 1] = 0 by construction. If refractoriness would be entirely random we would eliminate bias and we may hope that, even for slight violations of this assumption, it will lower biases.

We can now investigate if the use of an IV (Eq (4)) gives a better estimate of connectivity strength compared to Eq (3) or simply analyzing the lagged correlations employing CCH calculated with Eq (18). Note that this CCH estimator can be considered an OLS estimate not conditioned on the stimulus. We use the IV estimator on the three-neuron system simulated with Eq (2) illustrated in Fig 1a. Interpreting the estimates of β, the IV converges to the ground truth for βCB but displays a slight negative bias in estimating βCA (Fig 1d red dashed lines). Contrary, the OLS and CCH methods both falsely conclude that βCA > 0 (Fig 1d blue dashed lines represent Eq (3), and yellow lines Eq (18)). The technique based on refractoriness reduces bias but, in practice, does not abolish it.

When inspecting the stimulation response as shown in Fig 1f, we can see after splitting up Eq (4) (with E[X|Z = 1] = 0) and Eq (3) that the baseline values before stimulus-response are non-zero. The network dynamics in U affect both neural refractoriness and downstream activities. To correct for these effects, we combine the approaches with a DiD correction [41]; see Difference in differences. We introduce superscript * to indicate that the response window (Fig 1b) is shifted one window size backward in time. The OLS/DiD estimate is then given by Eq (3) with YYY* i.e.

β^OLS,DiD=E[Y|X=1]-E[Y*|X=1]-E[Y|X=0]+E[Y*|X=0] (5)

and similarly, for the IV/DiD estimate given by

β^IV,DiD=E[Y|Z=1]-E[Y*|Z=1]-E[Y|Z=0]+E[Y*|Z=0]E[X|Z=1]-E[X*|Z=1]-E[X|Z=0]+E[X*|Z=0] (6)

The DiD corrects for both the ongoing change and the network state and we should expect it to further lower bias relative to the IV. The IV/DiD method converges to the correct causal conclusions that βCB > 0 and βCA ∼ 0 (Fig 1d solid red lines), with improved accuracy compared to the IV method. The IV/DiD method produces good estimates of the causal interactions between neurons for the simple three-neuron system.

The strength of neural dynamics characterized by the condition number affects errors

Interacting neurons in a biological network exhibit inhibition and excitation, and their interplay produces nontrivial dynamics. When estimating the influences between neurons, we effectively solve an inverse problem. Such inverses are typically ill-posed problems, in the Hadamard sense [42] meaning a solution 1) does not exist, 2) is not unique, or 3) is unstable. Assuming the two first conditions to be untrue, i.e., a unique solution exists, the remaining lack of stability is referred to as an ill-conditioned problem. In statistics, this indicates whether small changes in (input) data have a comparatively small change in the (output) statistic and is meaningfully measured by its variance. When a statistical problem is ill-conditioned, the computed estimate may be far from the true estimate, regardless of our choice of estimator that minimizes the variance. This problem is typically due to multicollinearity. For example, in the linear model, Y = βX + U, multicollinearity is defined as near-linear dependence between column vectors in the design matrix X. In terms of neural measurements, X would be a N-by-time matrix with columns representing population states at different timepoints, and multicollinearity would represent correlations between these population states. In the inverse problem X = XY, with ′ denoting transpose, the condition number κ(XX) is a measure of the existence of multicollinearity [43]. This condition number is given by the ratio of singular values of the Gram matrix XX by κ(XX)=σmaxσmin. Large condition numbers imply that the inverse problem is ill-conditioned. Effectively, problems with large condition numbers, amplify a small amount of noise associated with small singular values, making them unstable. Multicollinearity may occur when a common, unobserved confounder drives neurons, as is very common in brains. Representing neurons as variables in the design matrix M ∈ [0, 1]N×T (Eq (2)), the condition number was computed from the covariance of M as the ratio of extreme singular values.

To assess how the proposed estimators performed under such circumstances, we simulated a fully connected random neural network of 200 neurons with an equal number of excitatory and inhibitory neurons (Fig 2a). Network dynamics were simulated with the Binomial GLM framework Eq (2). By stimulating five neurons, we examined the connection strength from these to the rest of the network and produced the results in Fig 2. The stimulus-trial onset times had a temporal Poisson distribution with a period 50 ms and were further clipped between 10–200 ms; see Binomial GLM simulation. The connectivity was given by a Gaussian distribution normalized by the square root of its size [44].

Fig 2. IV is more robust to multicollinearity for moderate condition numbers.

Fig 2

a) Sketch of the neural network configuration. b) Condition number is more affected by excitatory drive than inhibitory drive. c) Error in IV/DiD and CCH for w ≥ 0 and varying condition numbers as a function of drive strength. d) Error in IV/DiD and OLS/DiD for w ≥ 0 and varying condition numbers as a function of drive strength. e) Errors as a function of the excitatory drive. Error bars represent one standard deviation of errors. f) Errors as a function of the inhibitory drive. Error bars represent one standard deviation of errors. g) Scatter plot of IV/DiD versus ground truth. h) Scatter plot of CCH versus ground truth. i) AUROC as a function of the excitatory drive strength. j) AUROC as a function of the inhibitory drive strength.

To compute errors we estimated a ground truth using a two-neuron network and computed the mean absolute error; see Computing errors. The error is thus interpretable as the expected absolute error performed by the estimator. Condition numbers were higher for larger drive strengths but less sensitive to inhibitory drive (Fig 2b). With higher condition numbers, errors increased in both the IV/DiD, and the CCH method for ground truth β ≥ 0 where low condition number (dark points) means meaningful error for CCH but almost no error for IV (Fig 2c). The errors for IV/DiD were lower for moderate condition numbers and increased with increasing condition numbers (Fig 2d). The errors for IV/DiD were lower for moderate excitatory confounding (Fig 2e) but increased with confounding along with increased standard deviation measured across neurons illustrated with shading around the solid mean line. The errors for IV/DiD were constantly low with increasing inhibitory confounding (Fig 2f), with low standard deviations across neurons. Counter-intuitively, the error in CCH seems to get smaller with increased excitatory confounding, therefore we plotted the estimated connectivity against ground truth in Fig 2g and 2h. We observe that while the bias and variance increase for the IV/DiD estimator, for the CCH estimator the variance got smaller and the slope of increase also got lower, indicating that it is close to zero for all estimates, which would explain the artificial decrease in error.

To assert how the estimators performed as binary classifiers, classifying whether there is a connection or not, we computed the area under the receiver operating characteristic (AUROC) curve. This procedure measures the quality of classification at various levels of thresholds. It indicates how well the estimator will function in a classification task where AUROC = 1 indicates a perfect classifier, and AUROC = 0.5 indicates chance level. Classification accuracy was generally higher for IV/DiD compared to other estimators (Fig 2i) with excitatory confounding. Note also that the AUROC is consistently lower for CCH, which indicates that the low error seen for high confounding is indeed artificially induced. Classification accuracy was higher for IV/DiD compared to other estimators (Fig 2j) with inhibitory confounding. The network state can act as a confounder, and the condition number is a major way of indexing the confounding level.

Stimulation, connection strength, and network size affects condition number and errors

To evaluate the IV/DiD method in a different setting, we simulated 480 networks with alternative stimulation strengths, number of neurons, and connectivity strength while keeping the excitatory and inhibitory drive constant. The connectivity was given by a Gaussian distribution normalized by the square root of its size [44] to keep the largest singular value constant to changes in network size.

By stimulating five neurons and examining the connection strength from these to the rest of the network, we produced the results in Fig 3. The condition numbers now correspond to network size, stimulus strength, and connection strength (Fig 3d). With higher condition numbers, the IV/DiD method outperformed the CCH method for ground truth β ≥ 0 (Fig 3a). The IV/DiD method was more robust and was generally better than OLS/DiD (Fig 3b). The errors for IV/DiD were significantly lower than all other methods (Fig 3c). Condition numbers were higher for stronger stimulation and connection strengths and were more sensitive to these variables in larger networks (Fig 3d).

Fig 3. Errors and corresponding condition numbers are affected by network size, stimulation and connection strengths.

Fig 3

a) Error in IV/DiD and CCH for w ≥ 0 and varying condition numbers determined by stimulus strength, connection strength, and the number of neurons. b) Error in IV/DiD and OLS for w ≥ 0 and varying condition numbers determined by stimulus strength, connection strength, and the number of neurons. c) Errors are significantly lower in IV/DiD than all other estimators; each line represents a comparison with the estimator corresponding to the left edge. d) Condition number as a function of stimulus strength, connection strength, and the number of neurons, color-coded according to the color bar given in (b). e) Errors from IV/DiD get lower with stronger stimuli than those in OLS, and CCH. f) Errors from IV/DiD increase with stronger connections, although less than OLS, OLS/DiD and CCH. g) AUROC from IV/DiD gets higher with stronger stimuli than those in OLS, and CCH. h) AUROC from IV/DiD gets higher with stronger connections, whereas those in OLS, and CCH varies less. Significance values are encoded as ns not significant, * p < 0.05, ** p < 0.01, **** p < 0.0001 computed with Mann Whitney U test.

To see how the quality of the estimates varied with stimulus and connection strength, we plotted their relations with error bars showing a 95% confidence interval over the non-visualized parameters, i.e. for stimulus strength error bars come from connection strength and network size, in contrast to Fig 2. The IV/DiD method yields lower error for higher stimulation strengths (Fig 3e). Intuitively, we need relatively strong stimuli for the downstream neuron to be affected by the instrument, this might explain the smaller errors with stronger stimuli seen in Fig 3e for IV/DiD. The IV/DiD method gave better classification results as measured with the AUROC, with stronger stimulus and in general for stronger connection strengths (Fig 3g and 3h).

Connectivity affects statistical power, stimulus percentage affects error

Causal Inference techniques obtain better causal validity at the cost of statistical power. When the number of synapses per neuron increases, estimating the weights should become harder even if we could record from all neurons. To quantify this, we simulated ten networks with 200 neurons without additional confounding factors where we reduced connectivity or equivalently increased sparsity (0% meaning fully connected 90% meaning 10% connectivity). We found that the error decreased as a function of sparsity in IV/DiD, and OLS (Fig 4a). To further quantify robustness, we also simulated ten networks with 200 neurons, where we increased the percentage of stimulated neurons. The error increased with an increased percentage of stimuli (Fig 4b) as would be expected when considering that the condition number increases with increasing stimulus percentage.

Fig 4. Error diminishes with sparsity while is increased by increasing stimulus percentage.

Fig 4

a) Error as a function of sparsity. b) Error as a function of percentage of stimuluated neurons. c) IV/DiD estimates as a function of true weight and sparsity. d) IV/DiD estimates as a function of true weight and stimulus percentage. e) CCH estimates as a function of true weight and sparsity. f) CCH estimates as a function of true weight and stimulus percentage. g) Error as a function of the number of stimulus trials, sparsity is denoted in the panel titles. h) Error as a function of the number of stimulus trials, stimulus percentage is denoted in the panel titles.

By plotting the estimates as a function of the ground truth and sparsity, it is evident that the error is due to a more extensive spread which for the IV/DiD method decreases with sparsity (Fig 4c) while not so much for CCH (Fig 4e). When increasing the stimulus percentage, the spread increases for IV/DiD, and the center of the estimates at β = 0 is shifted upwards. However, the slope seemingly remains steady (Fig 4d). For the CCH method, the spread increased with increased stimulus percentage upon reducing again, the slope went towards zero, and the center at β = 0 is shifted upwards (Fig 4f).

To reach asymptotic convergence for IV/DiD, a large number of trials is needed (∼ 106) as seen in Fig 4g, for larger sparsity, more trials are needed, but with the result of a smaller error. When increasing the stimulus percentage, the convergence curve flattens out while being shifted towards larger errors (Fig 4h).

The IV/DiD method introduced here for the estimation of causal effects requires considerable amounts of data. The applicability of the proposed method may be limited to cases where weights are sufficiently sparse, and the stimulation effect is sparsely distributed among neurons with strong stimuli responses. Based on these findings, the right domain of application for the IV/DiD method would not be densely stimulated in vivo mammalian cortex but rather systems that are smaller, more sparse, or scenarios where only a small number of neurons are stimulated.

Estimation under distributed optogenetic stimulation strength

Results presented in Fig 4b show that a uniform distribution of strong stimuli across the excitatory population is detrimental for the estimation of connectivity, although estimation is better with IV/DiD compared to OLS,(DiD), and CCH. However, a uniform distribution of stimuli strength does not reflect the distribution of light intensity obtained with optogenetic stimulation. Optogenetic stimulation is generally seen as a perturbation that, by and large, affects neurons in proximity (< 1mm) of the light source, decreasing the effect with distance. However, this needs to be more accurate in terms of conceptualizing the spatial effect of stimulation, as it depends on multiple factors. Light intensity and opsin density are essential as more light and ion channels will cause a more substantial effect on each cell. Moreover, the number of potentially stimulated neurons is critical as more neurons will have a larger impact on the overall population activity. Finally, the physiological properties of the cells are important, as light may have a stronger effect on spiking activity when the cell’s membrane potential is sufficiently close to the firing threshold. A relation between four parameters should give the induced effect of optogenetic stimulation as a function of distance: light intensity, the spatial distribution of neurons, distributions of membrane potential across neurons, and the distribution of induced photo-currents.

To estimate light intensity, we calculated the spatial extent of laser light delivered by fiber-optics under plausible experimental conditions according to [45]. To approximate an optogenetic experiment we modeled the transmission of light through brain tissue with the Kubelka-Munk model for diffuse scattering in planar, homogeneous media [46] given by

T=1Sr+1. (7)

Here T denotes a transmission fraction, S is the scattering coefficient for mice [45], and r is the distance from a light source. Further, we combined diffusion with geometric loss assuming that absorption is negligible as in [45] and computed the intensity as presented in Fig 5a by

I(r)I(r=0)=ρ2(Sr+1)(r+ρ)2 (8)

where r is the distance from the optical fiber and

ρ=d2(nNA)2-1. (9)

Here d is the diameter of the optical fiber, NA is the numerical aperture of the optical fiber and n is the refraction index for gray matter [46]; see numerical values for parameters in Table 2.

Fig 5. Spatial extent of optogenetic stimulus.

Fig 5

a) Due to scattering and geometric loss the light intensity (I, cyan line) with an intensity of 10mW/mm2 exiting the optogenetic fiber follows approximately an inverse square law r−2 where r is the distance from the fiber. If neurons are uniformly distributed, the number of affected neurons in a spherical slice increases by r2 (N, black line). The total photocurrent (P, blue line) summed over neurons in a spherical slice increased with distance due to the nonlinear relation between light intensity and photocurrent, depicted as a percentage of maximum. b) Experimental data for verification [52]. c) Hit rate increases relative to stimulus strength with network size. d) Condition numbers get severely large with increasing network size. e) Errors increase with network size. f) AUROC decrease with network size. g) Scatter of estimands to ground truth for N = 500.

Table 2. Stimulation parameters.

name value units
density ∼ 1960 Nmm−3
S 10.3 mm−1
NA 0.37
r 0.1 μm
n 1.36
n Hill 0.76
K 0.84
depth 0.7 mm

While the experiments in [45] uses blue light, their estimates assume no absorption making the equations somewhat general. This modeling of light intensity yields an approximately 1/r2 reduction with distance r from the stimulation site (Fig 5a, cyan line). This is explained by the surface of a 3D shell growing with 4πr2 and photons will be roughly isotropic (but see [47]) beyond the scattering length as depicted in the inset of Fig 5a. The same number of photons must cross each sphere around the stimulation location unless absorbed or scattered inwards. As a result, the density of photons decreases with distance.

The number of illuminated neurons at a given distance will, however, increase with distance to the stimulation site, given that neurons are roughly uniformly distributed in brain tissue (Fig 5a, black line). It will increase by approximately r2 with distance. This distance dependence derives from the same surface scaling for the 3D shell as the photon flow. Thus, the number of neurons that can be activated increases rapidly with distance.

To estimate the stimulation effect, the mechanism by which light affects spiking activity needs to be considered. The distribution of membrane potentials across neurons can largely characterize this. Surprisingly, this distribution has been observed to be symmetrically distributed and relatively flat [4850]. The expected response from a pulse of light that induces a charge Q should be proportional to the number of affected neurons whose membrane potential sits within a Q/C range of the threshold (C is the capacitance). Given that the distribution of membrane potentials is relatively flat (the density close to the threshold is generally within an order of magnitude of the density of its mode) suggests that the spiking response to a perturbation for any neuron is roughly proportional to the induced photo-current.

The peak amplitude of the photo-current relates approximately logarithmically to the light intensity. To estimate the peak amplitude photocurrent, we used the Hill equation fitted by parameters found in [51] given by

P=ImaxInKn+In (10)

Here, Imax = 642pA is the maximum current, n = 0.76 is the Hill coefficient, and K = 0.84mW/mm2 represents the half-maximal light sensitivity of the ChR2. We used the light intensity I given by Eq (8) multiplied by an initial intensity of 10mW/mm2.

Assuming that opsins are evenly distributed across neurons, the induced photo-current will not be proportional to light intensity—it will fall slower. Based on this, we calculate the overall stimulation effect as the product of the number of neurons in a spherical slice and the peak amplitude photo-current. This product increases with distance (up to the distance where absorption becomes important) (Fig 5a), blue line. In other words, there is more activation at 500um than at 100um—when neurons densely express opsins, optogenetics is not a local technique.

To verify this intuition with experimental data, we extracted data with WebPlotDigitizer(apps.automeris.io/) from [52] where an optic fiber was retracted while measuring average spiking activity (Fig 5b, blue line). By multiplying the distance squared with this activity, we get an experimentally valid estimate of the scaled population spiking activity (Fig 5b), orange line), which increases with distance. Thus, experimentally, optogenetic stimulation utilizing single photon activation does not produce a localized effect.

With this knowledge, we simulated more realistic distributions of stimuli strength to evaluate its effect on the proposed estimators. We simulated five networks of sizes ranging from 100–500 neurons with additional confounding factors and balanced excitation and inhibition. We gave each neuron a random spatial distance from the simulated optogenetic stimulus for stimulation. The stimulus intensity was then set according to Eq (10), and was constant throughout the trials. The trial onset had a temporal Poisson distribution with a period of 50 ms and was further clipped between 10–200 ms. To evaluate the stimulation effect, we measured the hit rate as the percentage of successful stimulation trials (E[X]). The hit rate was more prominent for larger network size and stimulus strength (Fig 5c). The effect of a larger hit rate with network size is likely due to more neurons being stimulated with high stimulus strength creating a cascade effect throughout the network. By increasing the maximum stimulus strength, we further observed a non-linear increase in hit rate across the population where at some point, almost all neurons had a hit rate at ∼ 90%. Condition numbers increased with larger network size (Fig 5d). Errors, as measured by the combined error across all hit rates, increased with larger network size (Fig 5e). Classifications were less accurate with larger network size (Fig 5f). Inspecting estimands as a function of ground truth, we observed that OLS/DiD, and CCH, to a large extent, predict the hit rate (Fig 5g).

Inhibitory networks

In previous sections, we have only explored excitatory connections. However, the estimators could also be used with inhibitory connections. We simulated a network of 120 neurons with balanced excitation and inhibition while stimulating five inhibitory neurons to assess the effectiveness of inhibitory estimates. By examining the connectivity of these stimulated neurons, we generated the results presented in Fig 6. The IV/DiD and CCH are overall similar (Fig 6a) but are again outperformed by the OLS/DiD method (Fig 6b) for most variations of condition numbers. Comparing overall results, the OLS/DiD method was significantly better than the other method. Notably, contrary to estimating excitatory connections, stimuli benefited all estimators. Condition numbers were mainly affected by connection strength (Fig 6c). Errors were smaller for OLS/DiD for all connections, and stimulation strengths (Fig 6e and 6f). Despite giving poor effective connectivity estimates, classification was comparable across methods except for IV/DiD, which performed comparably worse (Fig 6g and 6h). Estimating the existence of inhibitory connections is relatively easy, but estimating their strengths appears harder.

Fig 6. IV is better than naive, but not OLS/DiD.

Fig 6

a) Error in IV/DiD and CCH for w ≤ 0 and varying condition numbers as a function of stimulus strength, connection strength and number of neurons. b) Error in OLS/DiD and CCH for w ≤ 0 and varying condition numbers as a function of stimulus strength, connection strength, and the number of neurons. c) Errors are significantly lower in OLS/DiD than all other estimators; each line represents a comparison with the estimator corresponding to the left edge. d) Condition number as a function of stimulus strength and connection strength, color-coded according to the color bar given in (a,b). e) Errors from OLS/DiD are generally lower with varying stimulation strength. f) Errors from OLS/DiD are generally lower with varying connection strength. g) AUROC from are comparable across methods with varying stimulation strength. h) AUROC from are comparable across methods with varying connection strength. Error bars represent 95% confidence across network size. Significance values are encoded as ns not significant, * p < 0.05, ** p < 0.01, **** p < 0.0001 computed with Mann Whitney U test.

Discussion

Here we have asked if the refractory period of neurons can be used as an IV to reverse engineer the causal flow of activity in a network of simulated neurons. We have found that this approach performs considerably better than the naïve approaches. Moreover, we have found that neither a naïve linear regression model nor a naïve cross-correlation method produces reliable estimates of connectivity between neuron pairs. The IV approach improves the inference process by looking at the missing responses because refractoriness effectively allows better estimates of causal effects.

One might worry that if a downstream cell needs multiple simultaneous inputs to spike, and some of the stimulated cells converge onto the same downstream neuron, the estimated effective connectivity of upstream cell X is biased outside of this stimulus regime. However, the stimuli are considered as part of the confounding error. It activates both the presynaptic neuron of interest and potentially many other presynaptic neurons. IV addresses this issue by removing the interaction between stimuli and postsynaptic neurons. Thus, the estimated effect of X on Y reflects the causal effect regardless of the stimuli. This is reflected in the error measure; zero error means that the causal effect is regardless of stimuli.

Moreover, every neuron is influenced by all of its presynaptic neurons. If the neuron is roughly linear then the causal effects βij will approximate the weights Wij. However, if neurons are very nonlinear then the real causal influences can not meaningfully be approximated by pairwise causal influences. As such, this approach, along with the bulk of the rest of the literature, is predicated on a largely untested assumption of neurons being approximately linear.

Currently, we have no ground-truth data set to test our technique and compare it with other approaches to actual neural activities. Ideally, we would have known causal effects from single-cell stimulation (e.g., from two-photon optogenetics) to establish causal effects. Such data should contain many randomly distributed, short, and intensive stimulation trials combined with traditional optogenetics, designed in a way where refractoriness matter. Since most optogenetic protocols use set stimulation frequency and light intensity, such a dataset, to the best of our knowledge, is currently not available and prevents us from testing how well our estimator would work on experimental data. Future experiments are needed to obtain reliable insights.

For the refractory period to be a good instrument, it must not be overly affected by the network activity. This criterion will be problematic in many cases. After all, network activity affects neuron activity and hence refractoriness. However, there are multiple scenarios where refractoriness will be a good instrument. For example, if we have balanced excitation and inhibition, we may expect largely independent refractory states of individual neurons. If a neuron biophysically implements something like conditional Poisson spiking, its refractory states will be random conditioned on the network state. Notably, the phase of a neuron may be far more random than the activity of the network as a whole.

The randomness of refractory times is the one factor that makes or breaks the IV approach we introduced here. Even if neurons’ refractory states are strongly correlated during regular network operation, there may be ways of randomizing refractoriness. First, it would help to use a task and situation where neurons are as uncorrelated as possible. Second, we may use a set of conditioning pulses of stimulation to increase the independence of refractory states. Giving one burst of stimulation strong enough to elicit multiple spikes from each neuron may effectively randomize their phases [53]. Third, we may utilize chemical, behavioral, or molecular perturbations to produce a good IV. For example, we can construct intracellular oscillators unaffected by neural activities or constructs that force a neuron into quiescence at random times. There has yet to be any effort to produce good IVs in neuroscience, so there may be many possibilities for improvements.

One popular way of estimating causal effects is fitting generalized linear models (GLMs) to simultaneously recorded neuron activities [14, 54]. GLMs are multiple nonlinear regressions and require multiple neurons to perform well. However, even if activity from all neurons were recorded, GLMs might fail to estimate causal connections [17]. However, complete recordings are not possible in the mammalian brain, especially not in primates, where recordings include only a minimal subset of the neurons involved in the actual computation. When using GLMs, one may accurately estimate latency distributions and sequences of spikes from individual neurons. These ideas should be merged with IV approaches. One of the strengths of the IV estimator presented here is that it only requires one pair to be recorded because we can utilize the randomness of the refractory periods along with random stimulations. Under those assumptions, the IV estimator can produce meaningful causal estimates.

The main problem with optogenetic stimulation, when used to infer connectivity, is its non-local property. This is due to the inverse relationship between changes in light intensity, and the affected number of neurons combined with a logarithmic relation between light intensity and photocurrent [51]. In addition, the distribution of membrane potentials across neurons is relatively flat [4850], making neurons highly sensitive to perturbations. One could imagine situations where optogenetic activation was more local. If, for example, the membrane potential distributions were skewed with the mode far from the threshold, a powerful stimulus would be required for a neuron to elicit spikes. There could also be other ways of making optogenetic stimulation more local. For example, if one engineered opsins or brain tissue that are more light absorbent (e.g. by ubiquitously producing melanin), one could stimulate more locally. Having melanin under a ubiquitously expressed promotor in the brain would dramatically make optogenetics more local and could probably be a target for the construction of transgenic animals. How to engineer more localized stimulation is a significant problem when causally interrogating a system.

Very weak laser pulses in noisy networks might mainly elicit spikes in very few close-by neurons in each trial [55]. However, the stimulus will still affect the membrane potential of many neurons further away, some of which will spike. Therefore, more than weak stimulation is needed to remove the principal problem of correlation-based techniques. After all, the network still acts as a confounder, and, if anything, the weak stimulation will reduce the statistical power of the approach. Lowering stimulation amplitudes is not a way of obtaining meaningful causal estimates.

We acknowledge the potential advantages of two-photon optogenetic stimulation of arbitrary subgroups of cells in vivo; e.g. [56]. This technique can in some cases give more favorable and realistic conditions compared to blanket 1-photon optogenetic stimulation and holds the potential for application to a vast number of presynaptic and postsynaptic neurons across various brain regions [57]. Our intention is therefore not to undermine the utility of two-photon optogenetic stimulation. Rather, we aim to underscore some of its existing limitations, such as its inapplicability in freely moving animals or across different animal species for example due to the necessity for head fixation. Although limitations may be surmounted in the future, they currently present considerable challenges that constrain the applicability of this technology in specific contexts. Consequently, we posit the necessity to persist in the exploration and enhancement of alternative approaches, such as 1-photon methods, which may be more apt for certain applications. This perspective underscores the necessity for a multifaceted approach to technological advancement in neuroscience, recognizing the strengths and weaknesses of each method to optimize their application in diverse research contexts.

There are many techniques for causal inference, most of which have yet to be applied to the field of neuroscience, and are based on approximating randomness in a seemingly regular world. In many cases, one could use regression discontinuity designs in a spiking system [58, 59]. Moreover, one could use a difference-in-difference approach [41]. Matching approaches [60, 61], can be used when comparing similar network states and their evolution over time. In general, neuroscience is in a quest for causal interpretations. Thus, we could benefit considerably by utilizing established techniques in the field of causal inference.

In the presence of ill-conditioning, the estimates of causal effects will be sensitive (unstable) to small changes in the underlying data. As outlined in [43], ill-conditioning can affect statistical analyses including the estimation of causal effects in three ways. First, measurement errors, such as a temporal shift in spike time estimate, e.g., low sampling frequency, inaccurate spike sorting, or general noisy measurement due to animal movement, can lead to significant estimation errors. Second, inference can give misleading results in ill-conditioning caused by bad design or sampling. There will always be natural variability in the observations, necessitating assessing ill-conditioning before any statistical analysis. Third, rounding errors can lead to small changes in data. Under ill-conditioning, this numerical problem is often not considered in neuroscience but will become evermore relevant when large-scale recordings require large-scale inferences. Causal inference approaches are not immune to the general problems coming from the ubiquitous ill-conditioning of neuroscience data.

The finding that inhibitory connections are very hard to estimate surprised us. After all, we had generally thought of excitation and inhibition as two sides of the same medal. However, we consistently find it hard to estimate inhibitory connection strengths. There may be some statistical reasons for this. An excitatory neuron can make another neuron fire at any point in time. An inhibitory neuron, on the other hand, can only prevent another neuron from spiking if it actually would be spiking. Neurons spend a vast amount of time not spiking. Similarly, neurons spend far more time being just below the firing threshold than being just above it. As such, the power to detect inhibition may be much higher than the power to detect excitation. Incidentally, this may also relate to the finding that there are far fewer inhibitory neurons in the brain than excitatory neurons: it is simply easier to learn how to excite than how to inhibit because there is more data.

Notably, practical short-term use of the techniques proposed exists. It is possible to stimulate just a local cluster of neurons optogenetically, say 10. This setup puts us into the situation of Fig 4b, where errors can be quite low, at least in the limit of relatively long recordings. As such, applying the methods introduced here to such experimental data, ideally with measured ground truth connectivity, is a promising and realistic next step.

Methods

Difference in differences

When dealing with temporal data we can use an additional method commonly used for causal inference, namely DiD. The effect of treatment in a longitudinal study with two time points Y*, Y is given by the average treatment effect of the treated (ATT) E[Y(1) − Y(0)|X = 1]. The ATT can be estimated by taking the differences (E[Y|X = 1] − E[Y*|X = 1]) − (E[Y|X = 0] − E[Y*|X = 0]) where the two differences in parenthesis denote a treatment group and a control group respectively. The DiD estimate is valid given two assumptions: (i) consistency and (ii) parallel trends; see [7].

To sample the stimulus trials, let ti denote stimulus onset time for stimulus trials i = 0, 1, 2, …, NS, and spike response windows [τ0V,τ1V] for variables V ∈ {X, Y, Z} which are associated with spikes tkV for upstream neurons X, Z and downstream neurons Y. The variables are given by

Vi={1τ0VtkV-tiτ1V,tkV0else.

The spike windows used for DiD [τ0V*,τ1V*] are given by [τ0V-τV,τ1V-τV] for window size τV=τ1V-τ0V where τ0V<τ1V.

Binomial GLM simulation

To model neuronal populations, we use a GLM framework similar to [35, 36]. This simulator and others alike are now implemented in Spikeometric spikeometric.readthedocs.io However, instead of using the Poisson GLM, we employed the binomial GLM, and instead of fitting the filters to neural data, we hand-tuned them. As discussed in the blogpost [62] the Poisson GLM comes with an exponential (unbounded) inverse link function which can produce nonphysical high spike rates. The binomial GLM comes with a (bounded) logistic inverse (sigmoid) link function which is ideal for spike simulations given as

σ(x)=exp(x)exp(x)+1 (11)

The generative model used throughout this paper is given in Eq (2).

Here the N × time matrix M ∈ {0, 1}N×T indicated spikes at time tΔ = Δtt where time bins were of size Δt = 1ms. Further, H = 10 denoted spike-history as the number of time steps backward. The bias b = 5 was hand-tuned such that average spike rates were ∼ 10Hz.

The spike-indicator matrix M was given by the sampled number of successes from the Bernoulli distribution with probability p.

The coupling matrix W ∈ RN×N contained cross-coupling weights. The cross-coupling filter was given by

c(t)={exp(-αt)for1t50for6tH, (12)

where α = 0.2. The refractory filters representing the absolute and relative refractory period were given by

r(t)={-100fort=1,2,3(absolute)-30exp(-12(t+4))for4tH(relative) (13)

A normal distribution gave the connection strength WN(0,σ/N) with zero mean and standard deviation scaled by the square root of the number of neurons. This scaling had two purposes: to keep the spike variance constant through time [44] and to keep the largest singular value constant as a function of network size [63].

The noise term U was employed to impose perturbations upon the system. Noise terms were stimuli and excitatory and inhibitory external confounding given as U(t) = US(t) + Upositive(t) + Uin(t) with respective strengths given by γS, γpositive, γin. The simulated optogenetic stimulation was given by

US(t)={γSfortSttS+20else. (14)

The excitatory external confounding was given by

Uex(t)={γexfortexttex+τex0else (15)

and the inhibitory external confounding was given by

Uin(t)={γinfortinttin+τin0else (16)

Stimulation onset times tS, tex, tin were drawn from a clipped Poisson distribution with average period λS, λex, λin ms maximum period λSmax,λexmax,λinmax ms and minimum period λSmin,λexmin,λinmin ms.

The parameters in Table 1 were used to produce the resulting figures.

Inhibitory and excitatory neurons

To follow Dale’s law we split networks into equal amounts of excitatory and inhibitory neurons with the following trick. First, we drew random weights W ∈ RN/2×N/2 with zero mean and normalized by the square root of half of the number of neurons. Then we combined four variants of the weight matrix in the following way

Wi,j={Wi,jifWi,j>0else0Wi,2j={Wi,jifWi,j>0else0W2i,j={Wi,jifWi,j<0else0W2i,2j={Wi,jifWi,j<0else0

The reasoning behind this procedure was to ensure we had a maximum singular value of one in the connectivity matrix which gives stable network dynamics.

Distributed optogenetic perturbation intensity (Fig 5)

To estimate the distribution of light intensity on stimulated neurons, we distributed neurons uniformly in 15 spherical slices in the range [0, 1mm], which had a radius given by the cone-shaped light; see Fig 1a inset.

Since the model neurons do not have membrane potentials, we scaled the maximum value of input current relative to the hit rate close to 100%.

Cross correlation histogram

The statistical tests giving the probabilities pdiff and pfast were done according to [55, 64]. We employed two tests to test whether the cross-correlation histogram (CCH) peak was significant. By using the Poisson distribution with a continuity correction [64] given by Eq (17), we calculated pdiff by comparing the peak in positive time lag with the maximum peak in a negative time lag, called pcausal in [55]. The probability pfast represents the difference between CCH and its convolution with a hollow Gaussian kernel [64]. These two measures of significance were required to be < 0.01 and given by

p(N|λ(m))=1-k=0N-1e-λ(m)λ(m)kk!-e-λ(m)λ(m)N2N!. (17)

Here λ represents the counts at bin m, and N is the number of bins considered. To naively estimate the effective connectivity between pairs, we used the “spike transmission probability” defined in [55] as

ptrans=1nm=3ms6msCCH(m)-λGauss(m), (18)

n is the number of spikes detected in the presynaptic neuron and λGauss(m) is the CCH count convolved with a hollow Gaussian kernel at bin m.

Computing errors

We computed errors according to

ϵij=β^ij-βijβij=0.9477(σ(Wij-b)-σ(-b))σ(x)=11+exp(-x) (19)

The reported error was then given by the mean absolute error over all estimated weights

Error=1N|ϵ| (20)

We further separated error measures into excitatory and inhibitory connections denoted as Error(w ≥ 0) and Error(w ≤ 0), respectively. To estimate the ground truth β we ran two blocks of simulations where two connected neurons; one with refractoriness given by Eq (13) and one with r = 0. In each block, we simulated 106 time steps. With this setup we could verify that with r = 0 βOLS gave the correct estimate of β = σ(Wb) − σ(−b). This gave us the confidence that we could use the OLS estimate with refractoriness to define the ground truth given in Eq (19). To do this we fitted a linear curve y = α(σ(Wb) − σ(−b)) to the OLS estimate to obtain α = 0.9477.

Supporting information

S1 Appendix

The following proofs follow the book from Brady Neal [65]. Text A. Identification proof of ATE with linear parametric form of IV. Text B. Non-parametric identification of CACE with potential outcomes.

(PDF)

Data Availability

All code used to produce data and results can be found at https://github.com/lepmik/causal-optoconnectics.

Funding Statement

This research was partly funded by the National Institutes of Health (NIH) grant R01MH103910 to KK. In addition, the Research Council of Norway Grant 231248, 300504 to TH and MF. KK received salary from the University of Pennsylvania, MEL, MF, TH TS received salary from the University of Oslo, MEL, TH, MF received salary from Simula Research Laboratory. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Azevedo FA, Carvalho LR, Grinberg LT, Farfel JM, Ferretti RE, Leite RE, et al. Equal numbers of neuronal and nonneuronal cells make the human brain an isometrically scaled-up primate brain. Journal of Comparative Neurology. 2009;513(5):532–541. doi: 10.1002/cne.21974 [DOI] [PubMed] [Google Scholar]
  • 2. Jonas E, Kording KP. Could a Neuroscientist Understand a Microprocessor? PLoS Comput Biol. 2017;13(1):1–24. doi: 10.1371/journal.pcbi.1005268 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Drton M, Foygel R, Sullivant S. Global identifiability of linear structural equation models. The Annals of Statistics. 2011; p. 865–886. [Google Scholar]
  • 4. Peters J, Janzing D, Schölkopf B. Elements of causal inference: foundations and learning algorithms. MIT Press; 2017. [Google Scholar]
  • 5. Stevenson IH, Kording KP. How advances in neural recording affect data analysis. Nature neuroscience. 2011;14(2):139–142. doi: 10.1038/nn.2731 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Mehler DMA, Kording KP. The lure of causal statements: Rampant mis-inference of causality in estimated connectivity. arXiv preprint arXiv:181203363. 2018;.
  • 7. Angrist JD, Pischke JS. Mostly harmless econometrics: An empiricist’s companion. Princeton university press; 2008. [Google Scholar]
  • 8. Pearl J. Causality. Cambridge university press; 2009. [Google Scholar]
  • 9. Stevenson IH, Rebesco JM, Miller LE, Körding KP. Inferring functional connections between neurons. Current opinion in neurobiology. 2008;18(6):582–588. doi: 10.1016/j.conb.2008.11.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Honey C, Sporns O, Cammoun L, Gigandet X, Thiran JP, Meuli R, et al. Predicting human resting-state functional connectivity from structural connectivity. Proceedings of the National Academy of Sciences. 2009;106(6):2035–2040. doi: 10.1073/pnas.0811168106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Aitchison L, Lengyel M. With or without you: predictive coding and Bayesian inference in the brain. Current opinion in neurobiology. 2017;46:219–227. doi: 10.1016/j.conb.2017.08.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Pfau D, Pnevmatikakis EA, Paninski L. Robust learning of low-dimensional dynamics from large neural ensembles. In: Advances in neural information processing systems; 2013. p. 2391–2399. [Google Scholar]
  • 13. Schneidman E, Berry MJ, Segev R, Bialek W. Weak pairwise correlations imply strongly correlated network states in a neural population. Nature. 2006;440(7087):1007–1012. doi: 10.1038/nature04701 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Pillow JW, Shlens J, Paninski L, Sher A, Litke AM, Chichilnisky E, et al. Spatio-temporal correlations and visual signalling in a complete neuronal population. Nature. 2008;454(7207):995. doi: 10.1038/nature07140 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Daniusis P, Janzing D, Mooij J, Zscheischler J, Steudel B, Zhang K, et al. Inferring deterministic causal relations. arXiv preprint arXiv:12033475. 2012;.
  • 16. Shimizu S, Hoyer PO, Hyvärinen A, Kerminen A. A linear non-Gaussian acyclic model for causal discovery. Journal of Machine Learning Research. 2006;7(Oct):2003–2030. [Google Scholar]
  • 17. Das A, Fiete IR. Systematic errors in connectivity inferred from activity in strongly recurrent networks. Nature Neuroscience. 2020;23(10):1286–1296. doi: 10.1038/s41593-020-0699-2 [DOI] [PubMed] [Google Scholar]
  • 18. Pinault D. A novel single-cell staining procedure performed in vivo under electrophysiological control: morpho-functional features of juxtacellularly labeled thalamic cells and other central neurons with biocytin or Neurobiotin. Journal of neuroscience methods. 1996;65(2):113–136. doi: 10.1016/0165-0270(95)00144-1 [DOI] [PubMed] [Google Scholar]
  • 19. Couey JJ, Witoelar A, Zhang SJ, Zheng K, Ye J, Dunn B, et al. Recurrent inhibitory circuitry as a mechanism for grid formation. Nature Neuroscience. 2013;16(3):318–324. doi: 10.1038/nn.3310 [DOI] [PubMed] [Google Scholar]
  • 20. Lerman GM, Gill JV, Rinberg D, Shoham S. Two Photon Holographic Stimulation System for Cellular-Resolution Interrogation of Olfactory Coding. In: Optics and the Brain. Optical Society of America; 2017. p. BrM3B–5. [Google Scholar]
  • 21. Nikolenko V, Poskanzer KE, Yuste R. Two-photon photostimulation and imaging of neural circuits. Nature methods. 2007;4(11):943. doi: 10.1038/nmeth1105 [DOI] [PubMed] [Google Scholar]
  • 22. Emiliani V, Cohen AE, Deisseroth K, Häusser M. All-optical interrogation of neural circuits. Journal of Neuroscience. 2015;35(41):13917–13926. doi: 10.1523/JNEUROSCI.2916-15.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Baker CA, Elyada YM, Parra A, Bolton MM. Cellular resolution circuit mapping with temporal-focused excitation of soma-targeted channelrhodopsin. eLife. 2016;5. doi: 10.7554/eLife.14193 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Shemesh OA, Tanese D, Zampini V, Linghu C, Piatkevich K, Ronzitti E, et al. Temporally precise single-cell-resolution optogenetics. Nature Neuroscience. 2017;20(12):1796–1806. doi: 10.1038/s41593-017-0018-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Daie K, Svoboda K, Druckmann S. Targeted photostimulation uncovers circuit motifs supporting short-term memory. Nature Neuroscience. 2021;24(2):259–265. doi: 10.1038/s41593-020-00776-3 [DOI] [PubMed] [Google Scholar]
  • 26. Hage TA, Bosma-Moody A, Baker CA, Kratz MB, Campagnola L, Jarsky T, et al. Synaptic connectivity to L2/3 of primary visual cortex measured by two-photon optogenetic stimulation. eLife. 2022;11. doi: 10.7554/eLife.71103 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Boyden ES, Zhang F, Bamberg E, Nagel G, Deisseroth K. Millisecond-timescale, genetically targeted optical control of neural activity. Nature neuroscience. 2005;8(9):1263. doi: 10.1038/nn1525 [DOI] [PubMed] [Google Scholar]
  • 28. Zemelman BV, Lee GA, Ng M, Miesenböck G. Selective photostimulation of genetically chARGed neurons. Neuron. 2002;33(1):15–22. doi: 10.1016/S0896-6273(01)00574-8 [DOI] [PubMed] [Google Scholar]
  • 29. Wright PG. Tariff on animal and vegetable oils. Macmillan Company, New York; 1928. [Google Scholar]
  • 30. Marinescu IE, Lawlor PN, Kording KP. Quasi-experimental causality in neuroscience and behavioural research. Nature human behaviour. 2018;2(12):891–898. doi: 10.1038/s41562-018-0466-5 [DOI] [PubMed] [Google Scholar]
  • 31. Liu T, Ungar L, Kording K. Quantifying causality in data science with quasi-experiments. Nature computational science. 2021;1(1):24–32. doi: 10.1038/s43588-020-00005-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Wald A. The fitting of straight lines if both variables are subject to error. The Annals of Mathematical Statistics. 1940;11(3):284–300. doi: 10.1214/aoms/1177731868 [DOI] [Google Scholar]
  • 33. Hernan MA, Robins JM. Causal Inference: What If. Boca Raton: Chapman & Hall/CRC; 2020. [Google Scholar]
  • 34. Gerstner W, Kistler WM, Naud R, Paninski L. Neuronal Dynamics. Cambridge University Press; 2014. Available from: 10.1017/cbo9781107447615. [DOI] [Google Scholar]
  • 35. Chichilnisky E. A simple white noise analysis of neuronal light responses. Network: computation in neural systems. 2001;12(2):199. doi: 10.1080/713663221 [DOI] [PubMed] [Google Scholar]
  • 36. Paninski L, Pillow JW, Simoncelli EP. Maximum Likelihood Estimation of a Stochastic Integrate-and-Fire Neural Encoding Model. Neural Computation. 2004;16(12):2533–2561. doi: 10.1162/0899766042321797 [DOI] [PubMed] [Google Scholar]
  • 37. Buzsaki G. Rhythms of the Brain. Oxford university press; 2006. [Google Scholar]
  • 38. Wilson C. Up and down states. Scholarpedia. 2008;3(6):1410. doi: 10.4249/scholarpedia.1410 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Koch C. Biophysics of computation: information processing in single neurons. Oxford university press; 2004. [Google Scholar]
  • 40. Pearl J. Causal diagrams for empirical research. Biometrika. 1995;82(4):669–688. [Google Scholar]
  • 41. Abadie A. Semiparametric difference-in-differences estimators. The Review of Economic Studies. 2005;72(1):1–19. doi: 10.1111/0034-6527.00321 [DOI] [Google Scholar]
  • 42. Hadamard J. Sur les problèmes aux dérivées partielles et leur signification physique. Princeton university bulletin. 1902; p. 49–52. [Google Scholar]
  • 43. Öztürk F, Akdeniz F. Ill-conditioning and multicollinearity. Linear Algebra and Its Applications. 2000;321(1-3):295–305. doi: 10.1016/S0024-3795(00)00147-6 [DOI] [Google Scholar]
  • 44.Glorot X, Bengio Y. Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings; 2010. p. 249–256.
  • 45. Aravanis AM, Wang LP, Zhang F, Meltzer LA, Mogri MZ, Schneider MB, et al. An optical neural interface: in vivo control of rodent motor cortex with integrated fiberoptic and optogenetic technology. J Neural Eng. 2007;4(3):S143–S156. doi: 10.1088/1741-2560/4/3/S02 [DOI] [PubMed] [Google Scholar]
  • 46. Ho AHP, Kim D, Somekh MG. Handbook of photonics for biomedical engineering. Springer; Netherlands; 2017. [Google Scholar]
  • 47. Thunemann M, Ness TV, Kilic K, Ferri CG, Sakadzic S, Dale AM, et al. Does Light Propagate Better Along Pyramidal Apical Dendrites in Cerebral Cortex? In: Optics and the Brain. Optical Society of America; 2018. p. JW3A–56. [Google Scholar]
  • 48. Paré D, Shink E, Gaudreau H, Destexhe A, Lang EJ. Impact of spontaneous synaptic activity on the resting properties of cat neocortical pyramidal neurons in vivo. Journal of neurophysiology. 1998;79(3):1450–1460. doi: 10.1152/jn.1998.79.3.1450 [DOI] [PubMed] [Google Scholar]
  • 49. Destexhe A, Paré D. Impact of network activity on the integrative properties of neocortical pyramidal neurons in vivo. Journal of neurophysiology. 1999;81(4):1531–1547. doi: 10.1152/jn.1999.81.4.1531 [DOI] [PubMed] [Google Scholar]
  • 50. Rudolph M, Destexhe A. On the use of analytical expressions for the voltage distribution to analyze intracellular recordings. Neural computation. 2006;18(12):2917–2922. doi: 10.1162/neco.2006.18.12.2917 [DOI] [PubMed] [Google Scholar]
  • 51. Wang H, Peca J, Matsuzaki M, Matsuzaki K, Noguchi J, Qiu L, et al. High-speed mapping of synaptic connectivity using photostimulation in Channelrhodopsin-2 transgenic mice. Proceedings of the National Academy of Sciences. 2007;104(19):8143–8148. doi: 10.1073/pnas.0700384104 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Han X, Chow BY, Zhou H, Klapoetke NC, Chuong A, Rajimehr R, et al. A High-Light Sensitivity Optical Neural Silencer: Development and Application to Optogenetic Control of Non-Human Primate Cortex. Frontiers in Systems Neuroscience. 2011;5. doi: 10.3389/fnsys.2011.00018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Ermentrout GB, Galán RF, Urban NN. Reliability, synchrony and noise. Trends in neurosciences. 2008;31(8):428–434. doi: 10.1016/j.tins.2008.06.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Roudi Y, Tyrcha J, Hertz J. Ising model for neural data: model quality and approximate methods for extracting functional connectivity. Physical Review E. 2009;79(5):051915. doi: 10.1103/PhysRevE.79.051915 [DOI] [PubMed] [Google Scholar]
  • 55. English DF, McKenzie S, Evans T, Kim K, Yoon E, Buzsáki G. Pyramidal Cell-Interneuron Circuit Architecture and Dynamics in Hippocampal Networks. Neuron. 2017;96(2):505–520. doi: 10.1016/j.neuron.2017.09.033 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Packer AM, Russell LE, Dalgleish HW, Häusser M. Simultaneous all-optical manipulation and recording of neural circuit activity with cellular resolution in vivo. Nature methods. 2014;12(2):140. doi: 10.1038/nmeth.3217 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57. Fişek M, Herrmann D, Egea-Weiss A, Cloves M, Bauer L, Lee TY, et al. Cortico-cortical feedback engages active dendrites in visual cortex. Nature. 2023;617(7962):769–776. doi: 10.1038/s41586-023-06007-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58. Lansdell B, Kording K. Spiking allows neurons to estimate their causal effect. bioRxiv. 2018; p. 253351. [Google Scholar]
  • 59. Imbens GW, Lemieux T. Regression discontinuity designs: A guide to practice. Journal of econometrics. 2008;142(2):615–635. doi: 10.1016/j.jeconom.2007.05.001 [DOI] [Google Scholar]
  • 60. Stuart EA. Matching methods for causal inference: A review and a look forward. Statistical science: a review journal of the Institute of Mathematical Statistics. 2010;25(1):1. doi: 10.1214/09-STS313 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.King G, Nielsen R. Why propensity scores should not be used for matching. Copy at http://j.mp/1sexgVw Download Citation BibTex Tagged XML Download Paper. 2016;378.
  • 62.Mineault P. Using the binomial GLM instead of the poisson for spike data; 2011. Available from: https://xcorr.net/2011/10/03/using-the-binomial-glm-instead-of-the-poisson-for-spike-data/.
  • 63. Geman S. A limit theorem for the norm of random matrices. The Annals of Probability. 1980; p. 252–261. [Google Scholar]
  • 64. Stark E, Abeles M. Unbiased estimation of precise temporal correlations between spike trains. J Neurosci Methods. 2009;179:90–100. doi: 10.1016/j.jneumeth.2008.12.029 [DOI] [PubMed] [Google Scholar]
  • 65.Neal B. Introduction to Causal Inference. 2015;.
PLoS Comput Biol. doi: 10.1371/journal.pcbi.1011574.r001

Decision Letter 0

Lyle J Graham, Marcus Kaiser

27 Aug 2023

Dear Dr. Lepperød,

Thank you very much for submitting your manuscript "Inferring causal connectivity from pairwise recordings and optogenetics" for consideration at PLOS Computational Biology.

As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments.

We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts.

Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Marcus Kaiser, Ph.D.

Academic Editor

PLOS Computational Biology

Lyle Graham

Section Editor

PLOS Computational Biology

***********************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #1: The authors suggest the use of the idea of Instrumental Variables from economics to facilitate the inference of synaptic connectivity (effective connectivity?) from neural data where optogenetics can be employed as a perturbation. The analysis of the proposed method is done on simulated neural data showing the positive effect it has on the inference process.

I find the general idea of this paper quite interesting and important. Using perturbation (optogenetics or otherwise) as a tool for facilitating inference of connectivity is a very promising tool. Also, the concept of instrumental variables is a welcome novel addition to efforts for inferring connectivity. However, I find the results themselves weak and the exposition of the results quite unclear.

Regarding the results, my criticism are as follows

-the authors perform their analysis on simulated data. Although the authors do study the effect of various features of the simulated network (such as type external field), important features are unexplored. For example the network size is rather small (200 neurons: 100 excitatory, 100 inhibitory), the connectivity is stated to be excitatory. I would have excepted the simulations to be done with larger networks, and in particular with more realistic model neurons, e.g. Hodgkin-Huxely or Integrate-and-Fire. There many biologically realistic neural network models, e.g. of the visual cortex, and the authors come from one of the major centres for building such models. So I am a bit surprised that such analyses are not performed.

- The results are section 2.4 are based on stimulating 5 neurons. How does this change if the analysis is performed on a larger fraction of neurons being stimulated? Is it important to know a priori what the fraction of activated neurons are? if so, how can one have a good estimate of that in a real data?

- The authors do not explore their approach on any real dataset. It is true that for real datasets, the ground truth is not known. Still one can compare the results of different methods and see if they indeed do give different results, and how big the difference is.

Regarding the presentation, although I really liked the introduction, I found the presentation of the Results right after the Introduction without having the author know the main methodological approaches made the paper difficult to read. The Methods section at the end is also very cluttered and jumps between definitions, proofs and material (e.g. 4.3.2) that can also be described in the results section. Some of the material (e.g. proofs) can be moved to a supplemental information/apprendices. I think the paper needs a major rewrite even if not further analyses are described.

Reviewer #2: This study is interesting and creative, but unlike the authors I would not discount the usefulness of two-photon optogenetic stimulation of arbitrary subgroups of cells in vivo (Packer et al. 2015 https://www.nature.com/articles/nmeth.3217) to infer functional connectivity – it creates far more favourable and realistic conditions than blanket 1-photon optogenetic stimulation, and can already be applied to hundreds of presynaptic and thousands of postsynaptic neurons across brain areas (Fisek et al. 2023 https://www.nature.com/articles/s41586-023-06007-6) and more in the future. It would be useful to compare blanket 1-photon and patterned 2-photon stimulation when inferring functional connectivity.

My main concern, however, is that the simulated model system (eq. 2) used by the authors to apply their statistical methods to may not exhibit sufficiently realistic types of noise that are found in cortical networks in vivo, leading to an underestimation of confounds. London et al. (2010) showed that cortical networks are chaotic because small perturbations in the spiking history grow. This is because the probability of a presynaptic spike evoking an extra postsynaptic spike (which is on the order df a percent) multiplied by the fan-out of the presynaptic neuron (the number of its postsynaptic targets, which in the cortex is on the order of thousands) is much larger than one. Two spiking histories of the same network in response to the same stimulus will therefore be massively different if looked at in small time bins, which may interfere with the statistical analysis methods presented in this manuscript. Are the networks used in this manuscript sufficiently chaotic, and do perturbations in these models grow as fast (e.g. with a gain of 28, London et al. 2010) as they do in vivo? The authors should ensure and demonstrate that their model system generates sufficiently realistic irreproducibility of spiking histories of the network.

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1011574.r003

Decision Letter 1

Lyle J Graham, Marcus Kaiser

4 Oct 2023

Dear Dr. Lepperød,

We are pleased to inform you that your manuscript 'Inferring causal connectivity from pairwise recordings and optogenetics' has been provisionally accepted for publication in PLOS Computational Biology.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. 

Best regards,

Marcus Kaiser, Ph.D.

Academic Editor

PLOS Computational Biology

Lyle Graham

Section Editor

PLOS Computational Biology

***********************************************************

Reviewer's Responses to Questions

Comments to the Authors:

Please note here if the review is uploaded as an attachment.

Reviewer #2: The rebuttals of the reviewers' comments by the authors are well argued, and the changes they made to their manuscript are good. Yet I'm torn because the changes are relatively minor and the authors did not include tests with larger networks as suggested by Reviewer 1, in which neurons have more realistic connectivity and synaptic weights. Future studies should involve random (Brunel) networks of integrate-and-fire neurons with realistic distributions of synaptic weights, as well as structured networks in which connectivity and connection strength represent similarity of sensory coding (see Cossell et al., 2015 for example).

**********

Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

PLoS Comput Biol. doi: 10.1371/journal.pcbi.1011574.r004

Acceptance letter

Lyle J Graham, Marcus Kaiser

30 Oct 2023

PCOMPBIOL-D-23-01040R1

Inferring causal connectivity from pairwise recordings and optogenetics

Dear Dr Lepperød,

I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript.

Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work!

With kind regards,

Zsofi Zombor

PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix

    The following proofs follow the book from Brady Neal [65]. Text A. Identification proof of ATE with linear parametric form of IV. Text B. Non-parametric identification of CACE with potential outcomes.

    (PDF)

    Attachment

    Submitted filename: Causal connectomics plos biology 2023 response to reviewers.pdf

    Data Availability Statement

    All code used to produce data and results can be found at https://github.com/lepmik/causal-optoconnectics.


    Articles from PLOS Computational Biology are provided here courtesy of PLOS

    RESOURCES