Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2019 Oct 15;200:174–190. doi: 10.1016/j.neuroimage.2019.06.031

A guide to group effective connectivity analysis, part 1: First level analysis with DCM for fMRI

Peter Zeidman a,, Amirhossein Jafarian a, Nadège Corbin a, Mohamed L Seghier b, Adeel Razi c,a, Cathy J Price a, Karl J Friston a
PMCID: PMC6711459  PMID: 31226497

Abstract

Dynamic Causal Modelling (DCM) is the predominant method for inferring effective connectivity from neuroimaging data. In the 15 years since its introduction, the neural models and statistical routines in DCM have developed in parallel, driven by the needs of researchers in cognitive and clinical neuroscience. In this guide, we step through an exemplar fMRI analysis in detail, reviewing the current implementation of DCM and demonstrating recent developments in group-level connectivity analysis. In the appendices, we detail the theory underlying DCM and the assumptions (i.e., priors) in the models. In the first part of the guide (current paper), we focus on issues specific to DCM for fMRI. This is accompanied by all the necessary data and instructions to reproduce the analyses using the SPM software. In the second part (in a companion paper), we move from subject-level to group-level modelling using the Parametric Empirical Bayes framework, and illustrate how to test for commonalities and differences in effective connectivity across subjects, based on imaging data from any modality.

Highlights

  • This guide walks through a group effective connectivity study using DCM and PEB.

  • Part 1, presented here, covers first level analysis using DCM for fMRI.

  • It clarifies the specific neural and haemodynamic models in DCM and their priors.

  • An accompanying dataset is provided with step-by-step analysis instructions.

1. Introduction

Neural models enable us to make inferences about brain circuitry using downstream measurements such as functional magnetic resonance imaging (fMRI). Just as the behaviour of a gas can be described by kinetics equations, which do not require knowing the position of every particle, so neural models can capture the mean activity of large numbers of neurons in a patch of brain tissue (Deco et al., 2008). A common application of these models in neuroimaging is to assess effective connectivity – the directed causal influences among brain regions – or more simply the effect of one region on another. This characterisation can be distinguished from the analysis of functional connectivity, which concerns statistical dependencies (e.g., the correlation or transfer entropy) between measurements, and structural connectivity, which concerns the physical architecture of the brain in terms of white matter tracts and synaptic connections. Effective connectivity cannot typically be observed directly, so models are used to traverse multiple spatial and temporal scales: the microscopic activity of neural populations, the meso- or macroscopic resolution of measurements (for example, LFP, EEG, MEG, ECoG or functional MRI) and population-level effects that are apt for characterising individual subjects.

Dynamic Causal Modelling (DCM) is a framework for specifying models of effective connectivity among brain regions, estimating their parameters and testing hypotheses. It is primarily used in human neuroimaging, but it has also successfully been applied with a range of species including rodents (Papadopoulou et al., 2017) and zebrafish (Rosch et al., 2017). A DCM forward (generative) model can be conceptualized as a procedure that generates neuroimaging timeseries from the underlying causes (e.g., neural fluctuations and connection strengths). The generated timeseries depend on the model's parameters, which generally have some useful interpretation; for example, a parameter may represent the strength of a particular neural connection. Having specified a forward model, one can then simulate data under different models (e.g. with different connectivity architectures), and ask which simulation best characterises the observed data. Practically, this is done in two stages: first, model inversion (i.e., estimation) is the process of finding the parameters that offer the best trade-off between accuracy (the fit of the predicted timeseries to the data) and the complexity of the model (how far the parameters had to move from their prior values to explain the data). This trade-off between accuracy and complexity is quantified by the model evidence. In the second stage, hypotheses are tested by comparing the evidence for different models (e.g. with different network architectures), either at the single-subject or the group level. These two stages are known as Bayesian model inversion and comparison, respectively. To evaluate the evidence for a model one needs to average over the unknown parameters, which means model inversion is usually needed prior to model comparison. This averaging or marginalisation is why model evidence is sometimes called the marginal likelihood of a model.

A variety of biologically informed forward models have been implemented for DCM. These range from simple mathematical descriptions of the gross causal influences among brain regions (Friston et al., 2003) to detailed models of cortical columns, which require temporally rich data afforded by electromagnetic recordings (Moran et al., 2013). In the context of fMRI, the objective of DCM is to explain the interactions among neural populations that show experimental effects. In other words, having identified where in the brain task-related effects are localised – usually using a mass-univariate (SPM) analysis – DCM is used to ask how those effects came about, in terms of (changes in) the underlying neural circuitry. Fig. 1 illustrates the forward model typically used with task-based fMRI experiments. Experimental stimuli drive a neural circuitry model, which predicts the resulting change in neural activity over time. Neural activity is tuned by a vector of parameters θ(n), which includes the strength of connections and the extent to which the connections are influenced by experimental conditions. The generated neural activity drives a model of neurovascular coupling and haemodynamics, which predicts the resulting change in blood volume and deoxyhaemoglobin level, tuned by the haemodynamic parameters θ(h). The final part of the model predicts the fMRI timeseries including noise (e.g. due to thermal variations in the scanner), that one would expect to measure given the neural activity and haemodynamics. This is configured by parameters θ(ε). By specifying this forward model and estimating the parameters θ=(θ(n),θ(h),θ(ε)), the variance in the observed timeseries is partitioned into neural, haemodynamic and noise contributions.

Fig. 1.

Fig. 1

The forward (generative) model in DCM for fMRI. This is split into three parts: neural, observation (subsuming neuro vascular, haemodynamic, BOLD signal components) and measurement (the addition of observation noise). The neural model is driven by experimental stimuli, specified as short events (delta functions). The resulting neural activity causes a change in blood flow (haemodynamics), mediated by neurovascular coupling, and consequently the generation of the BOLD signal. The addition of observation noise gives the fMRI timeseries. Image credits: Image credits: “Brain image” by parkjisun and “CT Scan” by Vectors Market from the Noun Project.

To illustrate the methodology – and detail the theory behind it – we analysed data from a previously published fMRI study on the laterality of semantic processing (Seghier et al., 2011). Language is typically thought to be left lateralised; however, the right hemisphere also responds in language tasks. This experiment asked how the left and right frontal lobes interact during semantic (relative to perceptual) processing. We do not attempt to offer any new insights into laterality or semantic processing here; rather we use these data to work through each step of a DCM analysis in detail. In the main text, we survey the current implementation and the specific models used for fMRI. In the appendices, we provide additional technical detail on the models and their implementation in the SPM software package. To help associate methods with their implementation in the SPM software (http://www.fil.ion.ucl.ac.uk/spm/software/), MATLAB function names are provided in bold text, such as (spm_dcm_fit.m), and these functions are listed in Table 1. We hope this worked example and tutorial-style overview of the theory will complement and expand on previous reviews and tutorials on DCM (Stephan, 2004; Seghier et al., 2010; Stephan et al., 2010; Kahan and Foltynie, 2013). The example data and a step-by-step guide to running these analyses can be found at https://github.com/pzeidman/dcm-peb-example.

Table 1.

DCM for fMRI software functions in SPM.

Function name Description
spm_dcm_specify Creates a DCM for fMRI Matlab structure for a single subject. (Details of the DCM structure can be found in the help text of spm_dcm_estimate.m)
spm_dcm_fit Fits (estimates or inverts) a GCM arraya of DCMs. This in turn calls spm_dcm_estimate to fit each DCM.
spm_dcm_fmri_check Provides basic validation statistics (e.g. explained variance) for a DCM or GCM array.
spm_dcm_fmri_priors Specifies priors on parameters for fMRI DCMs.
spm_dcm_review Graphical user interface for reviewing the contents of a DCM.
spm_fx_fmri Neural and haemodynamic model for fMRI DCMs.
spm_gx_fmri Observation model for fMRI DCMs.
spm_int The integrator used to generate predicted timeseries from the DCM.
spm_nlsi_GN The model estimation scheme used when fitting DCMs (variational Laplace).
a

A GCM (Group DCM) array is a Matlab cell array of DCM structures or filenames, with one row per subject and one column per model. For most group analyses, the first column of the GCM is expected to contain each subject's ‘full’ model, which includes all parameters of interest, and subsequent columns contain reduced models with certain parameters fixed at their prior expectation (typically zero).

2. Experimental design

DCM is a hypothesis-driven approach, the success of which depends on having an efficient experimental design. First, hypotheses need to be clearly articulated, which may relate to effects at the within-subject level, the between-subject level, or both. Here, the within-subject hypothesis was that processing the meaning of familiar words (i.e., their semantic content) would induce greater responses in left frontal cortex than right frontal cortex. The between-subject hypothesis was that this difference in hemispheric responses, quantified by the ‘Laterality Index’ (LI), would vary across subjects and could be explained by the strength of specific connections.

The hypotheses determine the experimental design. An efficient design at the within-subject level typically involves varying at least two experimental factors independently. Commonly, one factor will be a manipulation of the stimuli that drive neural responses, and another factor will be a manipulation of the task demands or context that modulates these responses. The distinction between driving and modulatory effects will be made explicit in the DCM analysis that follows. Here, we had two independent factors at the within-subject level: stimulus type (Words or Pictures) and task (Semantic or Perceptual reasoning), forming a balanced factorial design with four experimental conditions (words ​+ ​semantic, words ​+ ​perceptual, pictures ​+ ​semantic, pictures ​+ ​perceptual). An interaction between these two factors was hypothesised; namely, a greater response to words than picture stimuli, specifically in the context of the semantic task. Here, we will identify the connections underlying this interaction using DCM.

3. Region selection and fMRI timeseries extraction

DCM is used to model the connectivity between brain regions of interest (ROIs), and the criteria for defining ROIs varies across studies. For resting state experiments, there are no experimental effects, so ROIs are typically selected using an Independent Components Analysis (ICA), or using stereotaxic co-ordinates or masks from meta-analyses or the literature. For task-based experiments, such as that used here, ROIs are usually selected based on an initial mass-univariate SPM analysis, where the objective of DCM is to find the simplest possible functional wiring diagram that accounts for the results of the SPM analysis. Seghier et al. (2011) evaluated an SPM contrast for the main effect of task and identified four ROIs in frontal cortex: 1) left ventral, lvF, 2) left dorsal, ldF, 3) right ventral, rvF, 4) right dorsal, rdF. Relevant timeseries were extracted, pre-processed and summarised within each ROI by their first principal component (see Appendix 1: Timeseries extraction). Fig. 2 illustrates the experimental timing and timeseries from an example subject.

Fig. 2.

Fig. 2

Prerequisites for DCM analysis of task fMRI data: the design (U) and data (Y). Left: Experimental inputs U. White areas indicate times during the experiment when experimental stimuli were shown to the subject. There were three conditions: ‘Task’ comprised all semantic decision trials, ‘Pictures’ and ‘Words’ comprised the subset of trials for each condition. Right: fMRI timeseries Y for each of the four brain regions to be modelled from a typical subject. These are concatenated vertically to give data vector y specified in Equation (1).

4. Neural model specification

DCM partitions the variability in a subject's timeseries into neural and non-neural (i.e. haemodynamic and noise) sources. This necessitates a two-part model, which can be written1 as follows:

z˙=f(z,U,θ(n))y=g(z,θ(h))+X0β0+ε (1)

The first line describes the change in neural activity due to experimental manipulations. The level of neural activity within all the modelled brain regions are encoded by a vector z. These are the hidden states, which cannot be directly observed using fMRI. The function f is the neural model (i.e., a description of neuronal dynamics), which specifies how the change in neural activity over time z˙ is caused by experimental stimuli U, current state z, and connectivity parameters θ(n). On the second line of Equation (1), the function g is the haemodynamic model, which specifies the biophysical processes that transform neural activity z into the Blood Oxygen-Level Dependent (BOLD) response with parameters θ(h). The remainder of the second line comprises the measurement or noise part of the model. A General Linear Model (GLM) with design matrix X0 and parameters β0 captures known uninteresting effects such as the mean of the signal. Finally, zero-mean I.I.D. observation noise ε is modelled, the variance of which is estimated from the data (see Appendix 2: Observation noise specification).

The choice of neural f and observation model g depends on the desired level of biological realism and the type of data available. Here, we used the default neural and haemodynamic models for fMRI data, first introduced in Friston et al. (2003), which capture slow emergent dynamics that arise from coupled neural populations. Models are specified by answering a series of questions (Q1-Q8) in the DCM software, which are described in the following sections.

4.1. Input specification

The first question, when specifying a DCM, is which experimental conditions to include. These form the columns in matrix U (Fig. 2, left). We had three conditions – Task (all semantic decision trials), Pictures (the subset of trials in which subjects made semantic judgements based on picture stimuli) and Words (the subset of trials with written stimuli). Trials of the perceptual control task and incorrect trials were not modelled and so formed the implicit baseline. Having selected these conditions, SPM imports the onset times of the trials automatically from the initial GLM analysis (Note that when trials have a positive duration – i.e. they are blocks – the corresponding columns of U have value one during stimulus presentation and zero when stimuli are absent. In the special case where all the trials are events with zero duration, U is scaled by the number of time bins per second.).

4.2. Slice timing

Whereas the neural state z is continuous, fMRI data are discrete, with a volume acquired every 3.6 s in our data (the repetition time, TR). A strategy is therefore needed to align the acquisition of the fMRI data to the model. Most fMRI data are acquired in sequential slices, meaning that measurements from different brain regions (located in different slices) will be separated in time. DCM has a slice timing model (Kiebel et al., 2007) that enables the acquisition time of each region to be accounted for, which may particularly benefit models with widely spaced brain regions. However, this assumes that we know the time at which each slice was acquired, which is generally not the case – because the brain is rotated and deformed during spatial normalisation. Furthermore, MRI sequences that do not acquire slices in sequential order (e.g. interleaved or multi-band sequences) would not be properly represented by the slice timing model. If in doubt, the typical approach is to minimise slice timing effects by using the middle slice of the volume. Here, we set the slice timing model to use the last slice of the volume (3.6s for all regions) to be consistent with the original publication of these data.

4.3. Bilinear or nonlinear

The third question when specifying the DCM is which neural model to use; i.e., how to approximate function f in Equation (1). The default neural model in DCM for fMRI (spm_fx_fmri.m) uses a Taylor approximation to capture the effective connectivity among brain regions, and the change in effective connectivity due to experimental inputs. As detailed in Appendix 3: Derivation of the fMRI neural model, any function (satisfying certain requirements) can be represented by a Taylor series, which is a mathematical expression consisting of an infinite sum of terms. By truncating the series after the first few terms, a simple expression can be derived that provides a close approximation of the true function. In DCM for fMRI, the neural model is a Taylor approximation of function f, as follows:

z˙=Jz+Cu(t)J=(A+kB(k)uk(t))u(t)=Ut,:Tuk(t)=Ut,k (2)

where z˙ is the change in neural activity (the neural response) per unit time. The first line says that neural response z˙ depends on connectivity matrix J. The columns of this matrix are the outgoing connections and the rows are the incoming connections, so element Jmn is the strength of the connection from region n to region m. (This is also the Jacobian matrix – in which each element Jmn is the partial derivative of neural activity in region m with respect to region n). Parameter matrix C is the sensitivity of each region to driving inputs, where element Cpq is the sensitivity of region p to driving input from experimental condition q. This is multiplied by u(t), the row of U corresponding to all the experimental inputs at time t.

The second line of Equation (2) specifies the connectivity matrix J, which is configured by two sets of parameters: A and B. Parameter matrix A specifies the average or baseline effective connectivity (see Section 4.6: Centre input) and B(k) specifies the modulation of effective connectivity due to experimental condition k=1K. Each matrix B(k) is multiplied by experimental inputs uk(t) relating to condition k at time t. In this experiment, we had three B-matrices corresponding to K=3 experimental conditions or inputs: Task (the onsets of all trials), Pictures (blocks in which the stimuli were pictures) and Words (blocks in which the stimuli were words).

The parameters in matrices A, B and C are in units of hertz (Hz) because they are rates of change (rate constants). Matrix A is the rate of change in neural response due to neural activity z, i.e. the effective connectivity. Matrix B is the rate of change in the effective connectivity (matrix A) due to the modulatory inputs. Finally, matrix C is the rate of change of the neural response due to the driving inputs. For more detail on the units and interpretation of the parameters, see Appendix 3: Derivation of the fMRI neural model and Appendix 4: The neural parameters.

Importantly, each brain region in this model is equipped with an inhibitory self-connection, specified by the elements on the leading diagonal of the average connectivity matrix A and modulatory input matrices B(k). These parameters control the self-inhibition in each region, or equivalently, their gain or sensitivity to inputs. Biologically, they can be interpreted as controlling the region's excitatory-inhibitory balance, mediated by the interaction of pyramidal cells and inhibitory interneurons (cf. Bastos et al., 2012). These parameters are negative and preclude run-away excitation in the network. This is implemented by splitting the average connectivity matrix A and modulatory input matrices B(k) into two parts: intrinsic within-region self-inhibition (AI,BI) and extrinsic between-region connectivity (AE,BE). These parts are combined as follows:

J=0.5exp(AI)exp(kBI(k)uk(t))Intrinsic(selfinhibition)+(AE+kBE(k)uk(t))Extrinsic(betweenregion) (3)

where 0.5Hz is the default strength of the self-connections. AI and BI(k) are diagonal matrices, i.e.

AI=[AI1000AI2000AI3],BI(k)=[BI1(k)000BI2(k)000BI3(k)] (4)

and AE and BE(k) are off-diagonal matrices as follows:

AE=[0AE1,2AE1,3AE2,10AE2,3AE3,1AE3,20],BE(k)=[0BE1,2(k)BE1,3(k)BE2,1(k)0BE2,3(k)BE3,1(k)BE3,2(k)0] (5)

Equations (3)–(5) specify the same model as in Equation (2), except the self-connections are constrained to be negative. The self-connections AI and BI(k) are unitless log scaling parameters that scale (multiply up or down) the default value of 0.5Hz. This furnishes them with a simple interpretation: the more positive the self-connection parameter, the more inhibited the region, and so the less it will respond to inputs from the network. Conversely, the more negative the self-connection parameter, the less inhibited the region. Matrices AE and BE(k) are the extrinsic connectivity among regions, in units of Hz. For example, AE3,1 is the strength of the connection from region 1 to region 3, or equivalently the rate of change in region 3 effected by region 1. If it is positive, then the connection is excitatory (region 1 increases activity in region 3) and if it is negative then the connection is inhibitory. Similarly, BE3,1(k) is the increase or decrease in connectivity from region 1 to region 3 due to experimental condition k.

In summary, the neural model in DCM for fMRI captures directed interactions between brain regions, with connection strengths encoded in matrices of parameters. Matrices AI and BI(k) are the self-connections, which are unitless log scaling parameters. Matrices AE and BE(k) are the between-region connections, in units of Hz. Care needs to be taken, therefore, to correctly report the different units of each type of parameter. In the software implementation of this model in SPM (spm_fx_fmri.m), the diagonal elements of the connectivity matrices are the self-connections and the off-diagonal elements are the between-region connections.

Returning to the model specification for the example experiment, the question asked by the DCM user interface is whether to use a bilinear or non-linear model. The model described above is referred to as ‘bilinear’, because the modulatory parameters B are the interaction between the neural activity z and experimental inputs U (switching the B parameters on and off is akin to switching between different linear models). The bilinear model was later extended (Stephan et al., 2008) to include a nonlinear term, enabling brain regions to modulate the effective connectivity between other brain regions. Here, we did not need to consider nonlinear effects, so we selected the default bilinear model. We next asked how the activity in each brain region should be modelled.

4.4. States per region

The ‘one-state’ bilinear DCM for fMRI model, described above, represents the level of activity of each brain region i at time t as a single number zi(t). A subsequent development was two-state DCM (Marreiros et al., 2008) that generates richer neural dynamics, for modelling each brain region as a pair of excitatory and inhibitory neural populations. This has been used, for example, to model changes to the motor cortico-striato-thalamic pathway in Parkinson's disease (Kahan et al., 2014). The two-state model requires the use of positivity and negativity constraints on all connections, which needs to be taken into account when interpreting the results (for details, see https://en.wikibooks.org/wiki/SPM/Two_State_DCM). Here, for simplicity, we selected the one-state DCM.

4.5. Stochastic effects

The model described in equations (2), (3), (4), (5)) is deterministic, meaning that the experimental stimuli drive all the neural dynamics. Stochastic DCM (Li et al., 2011) estimates time-varying fluctuations on both neural activity (hidden states) and the measurements. This means that stochastic DCM can be used to model resting state (e.g. Bastos-Leite et al., 2014; Goulden et al., 2014; Dirkx et al., 2016) as well as task-based fMRI studies where endogenous fluctuations are important (e.g. Bernal-Casas et al., 2013; Jung et al., 2018). However, stochastic DCM poses a challenging model estimation problem, as both the connectivity parameters and trajectory of the hidden states need to be inferred. For resting state fMRI, a more recent technology, DCM for Cross-Spectral Densities (Friston et al., 2014), offers a simpler and more efficient solution, by modelling the data in the frequency domain (see Section 4.7). By modelling the data features in terms of spectral power, the stochastic fluctuations above become spectral components that are much easier to parameterise and estimate. Here, we elected not to include stochastic effects.

4.6. Centre input

The next question is whether to mean-centre input matrix U. If experimental input is mean-centred, then the parameters in matrix A represent the average effective connectivity across experimental conditions and modulatory parameters B(k) add to or subtract from this average. If U is not mean-centred, then A is the effective connectivity of the unmodelled implicit baseline (akin to the intercept of a linear model), onto which each modulatory input adds or subtracts. Mean-centring can increase the model evidence, by enabling the connectivity parameters to stay closer to their prior expectation (of zero) during model inversion. Furthermore, it ensures that excursions from baseline activity are reduced; thereby eluding nonlinear regimes of the haemodynamic model. Finally, mean-centring affords the matrix A a simpler interpretation (the average connectivity). Here, we chose to mean-centre the inputs, giving positive values in U when stimuli were presented and negative values elsewhere.

4.7. Timeseries or cross-spectral density (CSD)

DCM for Cross Spectral Densities (CSD), also called Spectral DCM (sDCM), is used for modelling fMRI data in the frequency domain, rather than the time domain. Also, unlike the original DCM for fMRI that models the timeseries directly, DCM for CSD models the functional connectivity (statistical dependencies) among the timeseries – more specifically, second order statistics like the cross-spectral density (Friston et al., 2014). This provides an efficient method for analysing resting state data (e.g. Park et al., 2017; Almgren et al., 2018; Zhou et al., 2018). It uses the same neural model as described above, but without modulatory inputs, as it is assumed that the connection strengths remain the same throughout the acquisition. Unlike stochastic DCM, this method does not try to model the neural state fluctuations in the time domain. By fitting data features in the frequency domain, estimation is significantly quicker, more efficient, and more sensitive to group differences (Razi et al., 2015). Here, we chose to fit timeseries rather CSD, because we were interested in condition specific, time-varying connectivity due to the task.

4.8. Connections

Having selected the form of the model, the next step is to configure it by specifying which parameters should be switched on (i.e., informed by the data) and which should be switched off (fixed at their prior expectation of zero). It is this sparsity structure that defines the architecture or model in question. Fig. 3 illustrates the network architecture we specified for each subject's DCM (spm_dcm_specify.m). We will refer to this as the ‘full model’, because all parameters of interest were switched on. Extrinsic or between-region connectivity parameters (matrix AE) were enabled between dorsal and ventral frontal regions in each hemisphere, and between homologous regions across hemispheres. Heterotopic connections were switched off, in line with previous findings: see the discussion in Seghier et al. (2011).

Fig. 3.

Fig. 3

The network architecture implemented for this analysis. Top: Schematic of the network indicating which parameters were switched on. These were the average connections over experimental conditions (intrinsic self-connections AI and extrinsic between-region connections AE), modulation of self-connections by pictures and/or words (BI) and driving input by Task (C matrix). This is a simplification of the architecture used by Seghier et al. (2011). Middle and bottom rows: The matrices corresponding to this network, indicating which parameters were estimated from the data (switched on, white) and which were fixed at zero (switched off, black). The regions of frontal cortex were left ventral, lvF, left dorsal, ldF, right ventral, rvF, right dorsal, rdF. The experimental conditions in matrix C were T = task, P = pictures, W = words.

DCM distinguishes two types of experimental input: driving and modulatory. Driving inputs are usually brief events that ‘ping’ specific regions in the neural network at the onset of each stimulus. The resulting change in neural activity reverberates around the network. Modulatory inputs up- or down-regulate specific connections and represent the context in which the stimuli were presented. They are typically modelled as blocks (box-car functions). We set Task (the onset of all Semantic trials) as the driving input to all regions (matrix C) and we set the context of being in Pictures blocks or Words blocks as modulatory inputs on the self-connection of each region (the diagonal elements of matrices BI(2) and BI(3) respectively). Limiting modulatory effects to the self-connections, rather than including the between-region connections, adds biological interpretability (as changes in the excitatory-inhibitory balance of each region) and generally improves parameter identifiability.

5. Haemodynamic model specification

The DCM haemodynamic model predicts the fMRI timeseries one would expect to measure, given neural activity. This does not require specification on a per-experiment basis, so here we just provide a brief summary of the pathway from neural activity to fMRI timeseries. Technical details are given in Appendix 5: Haemodynamic and BOLD signal model.

Following experimental stimulation, the temporal evolution of the BOLD signal can be divided into deoxygenated, oxygenated and sustained response phases, each of which can be linked to interactions of neuronal activity, neurovascular coupling, and blood vessel dynamics as summarized in Fig. 4. The baseline level of the BOLD signal is determined by the net oxygen extraction exchange between neurons and blood vessels, as well as cerebral blood flow. In response to experimental stimulation, neurons consume oxygen, increasing the ratio of deoxygenated to oxygenated blood. This is reflected by a lag in the BOLD response (the deoxygenated phase). In response to stimulation, neural activity drives astrocytes, releasing a vasodilatory signal (e.g., nitric oxide), which causes an increase in cerebral blood inflow. As a result, the oxygen level, blood volume, and blood outflow are all increased, which is accompanied by a rise in BOLD signal (oxygenated phase) after stimulation. In the absence of further stimulation, the activity of neurons return to their resting state, accompanied by a gradual decrease in the BOLD signal (sustained response phase). (Note that an initial dip in the BOLD signal and a post-stimulus undershoot may also be obesrved, not shown in Fig. 4). The dynamic interactions between cerebral blood flow, deoxyhemoglobin and blood volume are captured by the haemodynamic model (spm_fx_fmri.m) and the BOLD signal model (spm_gx_fmri.m), the parameters of which are estimated on a per-region basis. These parameters are concatenated with those of the neural model and estimated using the fMRI data.

Fig. 4.

Fig. 4

BOLD signal divided into deoxygenated, oxygenated and sustained response phases. The DCM forward model captures the biophysical processes that give rise to this signal. In the deoxygenated phase, neurons consume oxygen while blood flow is not altered. The blood inflow, outflow, and oxygen level increase in response to the neural activity, up to the peak of the BOLD signal at 5–6s post stimulation. BOLD signal exhibits a gradual decay to its baseline in the absence of further stimulation.

6. Model estimation

Having specified the forward model, the next step is to invert the model for each subject (spm_dcm_fit.m). Estimation or inversion is the process of finding the parameters (e.g. connection strengths) that offer the best trade-off between explaining the data and minimizing complexity (i.e. keeping the parameters close to their prior or starting values). Because there are multiple settings of the parameters that could potentially explain the observed data, DCM uses Bayesian inference, which involves quantifying uncertainty about the parameters before and after seeing the data. This starts with specifying priors that constrain the parameters. Model estimation combines the priors with the observed fMRI data to furnish updated posterior beliefs (i.e. after seeing the data). The priors and posteriors have the form of probability densities. Below, we detail the priors used in DCM, which are configured by the DCM software when model estimation is performed. We will then briefly explain the model estimation procedure itself, known as Variational Laplace.

6.1. Priors

The priors over parameters in DCM form a multivariate normal density, which is specified by its mean and covariance. Practically, these densities are expressed as a vector of numbers (the mean or expected values of the parameters) and a covariance matrix. Elements on the leading diagonal of the covariance matrix are the prior variance (uncertainty) for each parameter, and the off-diagonal elements are the covariance between the parameters. The choice of priors for each connectivity parameter depends on whether the connection was ‘switched on’ or ‘switched off’. Each switched on parameter has expectation zero and non-zero variance (Fig. 5, left). This says that in the absence of evidence to the contrary, we assume there is no connectivity or experimental effect, but we are willing to entertain positive or negative values if the data support it. The width of this distribution (its variance) determines how uncertain we are that the parameter is zero. The prior for each ‘switched off’ parameter has expectation zero and variance close to zero (Fig. 5, right). This says that we are certain that the parameter is zero, regardless of the data. Both of these are called ‘shrinkage priors’, because, in the absence of evidence, the posterior density shrinks to zero. For this experiment, we selected the connections to switch on and off (Fig. 3), and the DCM software translated these choices into priors for each parameter (spm_dcm_fmri_priors.m). Note that by default, in order to decrease the time required for model estimation, if more than eight brain regions are included then DCM automatically constrains the model by using sparsity-inducing priors based on the functional connectivity (Seghier and Friston, 2013). This was not the case here, and the priors for all free parameters are listed in Table 3.

Fig. 5.

Fig. 5

Illustration of priors in DCM. Left: the prior for a ‘switched on’ parameter is a Gaussian probability density with zero mean and non-zero variance. Right: the prior for a ‘switched off’ parameter has zero or close-to-zero variance, meaning the parameter is fixed at the prior expectation, which is typically zero.

Table 3.

Free parameters and their priors.

Name Parametrizationa Prior expectation Prior variance (uncertainty) 90% CI
AE AE 0 1/64 [-0.21 0.21]
AI 0.5Hzexp(AI) 0 1/64 [-0.21 0.21]
BE BE 0 1 [-1.65 1.65]
BIk 0.5Hzexp(BI(k)) 0 1 [-1.65 1.65]
C 116C 0 1 [-1.65 1.65]
ε exp(ε) 0 1/256 [-0.10 0.10]
κ 0.64Hzexp(κ) 0 1/256 [-0.10 0.10]
λi exp(λi) 6 1/128 [5.85 6.14]
τh 2.00sexp(κ) 0 1/256 [-0.10 0.10]
a

Log scaling parameters have no units - they are exponentiated and then multiplied by fixed default values, listed in the Parametrization column.

6.2. Variational Laplace

Model inversion (i.e., parameter estimation) is the process of finding the parameters that enable the model to best explain the data; i.e. maximize the log model evidence lnpy|m. This is the log of the probability of having observed the data y given the model m. Generally, model evidence cannot be calculated or derived analytically (because it involves marginalization over very high dimensional integrals); so instead an approximation called the negative variational free energy F (Friston et al., 2007) can be used. The free energy is a lower bound on the model evidence (in machine learning, an Evidence Lower Bound or ELBO). It is useful because it scores how well the model achieved a trade-off between accuracy and complexity:

lnp(y|m)F=accuracy(y,m)complexity(m) (6)

The accuracy term quantifies how closely the predicted timeseries corresponds to the observed data. The complexity term is the Kullback-Leibler divergence between the priors and the posteriors; i.e., the difference between the two distributions. If the parameters had to move far from their prior expectation in order to explain the data, then the complexity of the model will be high. This measure of complexity also distinguishes parameters that are independent from those that co-vary (making less individual contribution to explaining the data). When selecting among several models of the same data, the best model is the one with the highest (most positive) free energy, because it offers the most accurate and least complex explanation for the data. We used the DCM software to invert each subject's model, obtaining estimates of their free energy F and the posterior probability density over the parameters that maximised F. This completes a description of the first-level (within subject) analysis.

7. Results

7.1. Diagnostics

A basic diagnostic of the success of model inversion is to look at the estimated parameters and the percentage variance explained by the model. Fig. 6 (top) and Table 4 show the neural parameters from a randomly selected subject (subject 37), which we will use to exemplify an interpretation of the parameters (spm_dcm_review.m). Many of the neural parameters (A,B,C) moved away from their prior expectation of zero, with 90% credible intervals (pink bars) that did not include zero. Fig. 6 (bottom) shows the modelled timeseries and residuals from this subject. There were clearly dynamics (solid lines) related to the onsets of the task (grey boxes). The explained variance for this subject was 18.85% and the mean across subjects was 17.27% (SD 9.37%), computed using spm_dcm_fmri_check.m. It is unsurprising that the explained variance was quite low, because we did not model the control conditions (perceptual matching) or the baseline rest periods. Nevertheless, most of the subjects evinced nontrivial neural parameters, with 90% credible intervals that excluded zero; so we could be confident that there was useful information in the data pertaining to our experimental effects.

Fig. 6.

Fig. 6

Example DCM neural parameters and model fit for a single subject. Top: The parameters corresponding to Equation (3). The error bars are 90% credible intervals, derived from the posterior variance of each parameter, and the vertical dotted lines distinguish different types of parameter. Note this plot does not show the covariance of the parameters, although this is estimated. The parameters are: the average inhibitory self-connections on each region across experimental conditions (AI), the average between-region extrinsic connections (AE), the modulation of inhibitory self-connections by pictures (BI(2)) and by words (BI(3)), and the driving inputs (C). For a full list of parameters, please see Table 4. Bottom: Example subject's predicted timeseries (solid lines) with one line per brain region. The dotted lines show the model plus residuals. Underneath, blocks showing the timing of the word and picture trials.

Table 4.

Example subject's neural parameters.

Parametera Description Units Expectation Precision Probability†
1 AI11 Self-connection on lvF None −0.16 66.94 0.91
2 AI22 Self-connection on ldF None −0.04 68.64 0.62
3 AI33 Self-connection on rvF None −0.04 75.39 0.62
4 AI44 Self-connection on rdF None −0.18 93.87 0.96
5 AE21 lvF → ldF Hz 0.42 233.16 1.00
6 AE31 lvF → rvF Hz 0.06 406.70 0.88
7 AE12 ldF → lvF Hz −0.02 291.40 0.58
8 AE42 ldF → rdF Hz 0.57 145.30 1.00
9 AE13 rvF → lvF Hz 0.43 149.48 1.00
10 AE43 rvF → rdF Hz 0.10 102.21 0.86
11 AE24 rdF → ldF Hz −0.03 483.41 0.73
12 AE34 rdF → rvF Hz −0.21 858.90 1.00
13 BI11(2) Pictures on lvF self None −0.47 41.73 1.00
14 BI22(2) Pictures on ldF self None 2.12 3.52 1.00
15 BI33(2) Pictures on rvF self None 0.13 16.78 0.70
16 BI44(2) Pictures on rdF self None −0.16 19.21 0.68
17 BI11(3) Words on lvF self None 2.80 1.98 1.00
18 BI22(3) Words on ldF self None 0.27 9.98 0.81
19 BI33(3) Words on rvF self None 0.24 6.40 0.73
20 BI44(3) Words on rdF self None 0.11 13.41 0.71
21 C11 Driving: task on lvF Hz −0.07 910.27 0.99
22 C21 Driving: task on ldF Hz 0.10 909.84 1.00
23 C31 Driving: task on rvF Hz 0.26 811.03 1.00
24 C41 Driving: task on rdF Hz 0.08 474.01 0.96
a

Region names: 1 = lvF, 2 = ldF, 3 = rvF, 4 = rdF. Condition names (superscript on matrix BI): 2 = Pictures, 3 = Words. †Probability that the posterior estimate of the parameter is not zero. For a parameter with marginal posterior density N(μ,σ2) this is given by 1NCDFabsμ,σ2, where NCDF is the normal cumulative density function.

7.2. Interpretation of parameters

We will use the same subject's model to interpret key parameters. The B parameters are the most interesting experimentally; these are the modulations of connections by each experimental condition (Pictures and Words). Positive parameter estimates indicate increased self-inhibition due to the experimental condition, and negative values meant disinhibition. We allowed picture and word stimuli to modulate each of the self-connections, and three of these parameters, numbered 13, 14 and 17 in Fig. 6 (top), deviated with a high degree of posterior confidence from their prior expectation of zero. These are illustrated in green and red text in Fig. 7. Picture stimuli increased self-inhibition on ldF and decreased self-inhibition on lvF, thereby shifting responses from the dorsal to ventral frontal cortex, specifically in the left hemisphere. Word stimuli increased self-inhibition in lvF, making it less sensitive to input from the other modelled regions.

Fig. 7.

Fig. 7

Estimated parameters from a single subject. Between-region (extrinsic) parameters are in units of Hz, where positive numbers indicate excitation and negative numbers indicate inhibition. Self-connection parameters have no units and scale up or down the default self-connection of −0.5 Hz (see Equation (3)). Positive numbers for the self-connections indicate increased self-inhibition and negative numbers indicate disinhibition. For clarity, only parameters with 90% probability of being non-zero are displayed (see Table 4 for details). Colours and line styles as for Fig. 3.

It is sufficient to report the estimated parameters and make qualitative statements about their meaning, as above (e.g., that the strength of a particular connection was increased or decreased by an experimental condition). However, what is the quantitative interpretation of these parameters? Taking region lvF as an example, we can write out Equation (3) in full, to express the rate of change in lvF's neural activity:

z˙1=(0.5exp(AI11)Averageexp(BI11(2)u2(t))Picturesexp(BI11(3)u3(t))Words)z1Selfconnection+AE12z2ldFlvF(A)+AE13z3rvFlvF(A)+C11u1(t)Driving(C)u1(t)={0.6,task0.4,otherwiseu2(t)={0.8,pictures0.2,otherwiseu3(t)={0.8,words0.2,otherwise (7)

This says that the response in region lvF was governed by the strength of its self-connection (line 1 of Equation (7)) as well as incoming connections from regions ldF, rvF and the driving input (line 2 of Equation (7)). The values for the experimental inputs u1(t), u2(t) and u3(t) at time t were set during the specification of the model, due to mean-centring of the regressors (see Section 5.6: Centre input). Plugging in the estimated parameters from Table 4, the self-inhibition in lvF during picture trials was 0.5×exp(0.16)×exp(0.47×0.8)×exp(2.8×0.2)=0.17Hz. The self-inhibition of lvF during word trials was far stronger: 0.5×exp(0.16)×exp(0.47×0.2)×exp(2.8×0.8)=4.40Hz. Therefore, region lvF was more sensitive to inputs from the rest of the network when the stimuli were pictures than words. These task effects can also be expressed as a change in the time constant τ of region lvF: τ= 5.88s in the context of pictures and τ= 0.23s in the context of words (see Appendix 4). Rewriting this as the half-life of region lvF; neural activity decayed to half its starting level 4.08s after the onset of picture stimuli and 0.16s after the onset of word stimuli. Picture stimuli therefore elicited a far more sustained response than word stimuli in lvF. The other key factor influencing lvF was the incoming connection from region rvF (0.43 Hz), and the positive sign indicates this connection was excitatory.

Inspecting the parameters in this way provides insight into the sign and magnitude of the connection strengths and experimental effects. However, this does not constitute a formal test of any hypotheses. There are various strategies for testing hypotheses at the group (between-subject) level, using classical or Bayesian statistics, and we detail these in the second part of the tutorial (please see the companion paper).

8. Discussion

This paper reviews the current implementation of DCM for fMRI by stepping through the analysis of a factorial fMRI experiment. This first level (within-subject) analysis started by identifying brain regions evincing experimental effects, for which we extracted representative fMRI timeseries. We then specified a DCM, by selecting which connections should be ‘switched on’ and which should be ‘switched off’. This specified the priors for the connectivity parameters. Inverting each subject's model provided a probability density over the connections strengths (A), the change in connections due to each experimental condition (B) and the sensitivity of each region to external input (C), as well as the free energy approximation to the log model evidence F. The appendices provide the technical detail of each of these steps.

A common question from DCM users is: what assumptions are made by DCM? As a Bayesian model, most assumptions are stated up-front as priors. The key assumptions for the basic (deterministic 1-state bilinear) neural model are as follows:

  • The pre-processed fMRI timeseries used for DCM have been selected because they show experimental effects. The signals are averaged over voxels and nuisance effects are regressed out, therefore, the signal-to-noise ratio is high – the prior expectation of the variance of the noise is 1exp(6)=0.0025 (see Appendix 2). This expresses the prior belief that most of the variance is interesting and where possible, we would like the variance to be ascribed to the model rather than to observation noise. Furthermore, the variance of the observation noise is assumed to be independent of the neural/haemodynamic parameters.

  • The neural response due to intrinsic (within-region) activity is expected to decay over a period of seconds following experimental stimulation. The prior on the self-connection parameters says that an isolated brain region's time constant τ will be between 1.63s and 2.46s with 90% probability, and between 0.38s and 10.49s in the context of modulation by an experimental condition (Appendix 4). This response can be increased or decreased by incoming connections from other regions.

  • The priors for the parameters of the haemodynamic, BOLD signal and observation models are consistent with empirical measurements using animal models and human subjects (c.f. Buxton et al., 1998; Stephan et al., 2007). In DCM for fMRI, three of these parameters are estimated from the data, because there is particular variability/uncertainty associated with them, and the priors are listed in Table 3. Values for fixed parameters, which are not estimated from the data, can be found in the Matlab functions spm_fx_fmri.m and spm_gx_fmri.m.

  • The free energy is assumed to serve as a good proxy for the log model evidence. This is exactly true for linear models (where the free energy becomes log model evidence) and has been validated for weakly non-linear models like DCM for fMRI using sampling methods (Chumbley et al., 2007). Caution needs to be taken with highly nonlinear models, where local optima pose a challenge; one method for addressing this is to use a multi-start estimation algorithm which re-initializes subject-level inversions using group-level estimated parameters (Friston et al., 2015).

The next step in our analysis was to test which neural effects were conserved over subjects, and which differed due to brain Laterality Index – the between-subjects factor that was the focus of this experiment. These analyses are detailed in the companion paper, where we cover Bayesian model comparison (i.e., hypothesis testing) at the within and between subject level.

Conflicts of interest

None.

Acknowledgements

The Wellcome Centre for Human Neuroimaging is supported by core funding from Wellcome [203147/Z/16/Z]. We thank Wiktor Olszowy for helpful feedback on the manuscript.

Footnotes

1

Vectors are denoted by lower case letters in bold italics (a) and matrices by upper case letters in bold italics (A). Other variables and function names are written in plain italics (f). The dot symbol (%) on its own means multiplication and when positioned above a variable (e.g. z˙) denotes the derivative of a variable with respect to time. An element in row m and column n of matrix A is denoted by Amn. All variables and theiar dimensions are listed in Table 2.

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.neuroimage.2019.06.031.

Appendix 1. Timeseries extraction

Before a DCM can be specified, Regions of Interest (ROIs) need to be selected and representative timeseries extracted from each. The fMRI data for a subject can be considered a large 4D matrix Yˆ where the first three dimensions are space and the fourth dimension is time (in scans). By extracting timeseries, we seek to reduce this to a smaller matrix Y where there are a small number of ROIs that define the brain network. There are various strategies for selecting the voxels that contribute to each ROI – indeed, questions pertaining to this are among the most common from DCM users on the SPM Mailing List. The most important consideration is that DCM is intended to explain the coupling between neural populations that show experimental effects. An initial GLM analysis is therefore normally used to identify voxels that show a response to each experimental factor. To reduce noise, only voxels that exceed some liberal statistical threshold for a contrast of interest are usually retained.

For the data presented here, the following steps were applied by (Seghier et al., 2011), which may provide a useful recipe for preparing DCM studies:

  • 1.

    Statistical Parametric Mapping (SPM). A General Linear Model (GLM) was specified for each subject, and T-contrasts were computed to identify brain regions that showed a main effect of each factor and an interaction between factors. Additionally, an F-contrast was calculated to identify all ‘Effects of Interest’ – to later regress out any uninteresting effects such as head motion or breathing from the timeseries. This F-contrast was an identity matrix of dimension n, where the first n columns in the design matrix related to interesting experimental effects.

  • 2.

    Group-level region selection. Contrast maps from each subject were summarized at the group level using one-sample t-tests. These group-level results were used to select the peak MNI coordinates of the ROIs. Different contrasts could have been used to select each ROI; however, in this case, the main effect of task (semantic > perceptual matching) was used to identify all four ROIs.

  • 3.

    Subject-level feature selection. Having identified the ROI peak coordinates at the group level, the closest peak coordinates for each individual subject were identified. This allowed for each subject to have slightly different loci of responses. Typically, one would constrain each subject-level peak to be within a certain radius of the group-level peak, or alternatively, to be within the same anatomical structure (e.g. using an anatomical mask). Here, subject-level peaks were constrained to be a maximum of 8 mm from the group level peak, and had to exceed a liberal statistical threshold of p < 0.05 uncorrected. At this stage, Seghier et al. (2011) excluded any subjects not showing experimental effects in every brain region above the statistical threshold. We suggest that with the development of hierarchical modelling of connectivity parameters, detailed in the companion paper, removing subjects with noisy or missing data in certain brain regions may be unnecessary. A subject who lacks a strong response in one brain region or experimental condition, for whatever reason, may still contribute useful information about other brain regions or conditions (and indeed useful information about intersubject variability). Therefore, when an ROI contains no voxels showing a response above the selected threshold, we recommend dropping the threshold until a peak voxel coordinate can be identified.

  • 4.

    ROI definition. Having identified the peak coordinates for each ROI, timeseries were extracted. Each ROI was defined as including all the voxels that met two criteria: 1) located within a sphere centred on the individual subject's peak with 4 mm radius and 2) exceeded a threshold of p < 0.05 uncorrected, for the task contrast at the single-subject level. Note that applying a threshold at this stage is not to ensure statistical significance (this happens in step 2). Rather, the threshold is simply used to exclude the noisiest voxels from the analysis.

  • 5.

    ROI extraction. SPM was used to extract representative timeseries from each ROI, which invoked a standard series of processing steps (spm_regions.m). The timeseries are pre-whitened (to reduce serial correlations), high-pass filtered, and any nuisance effects not covered by the Effects of Interest F-contrast are regressed out of the timeseries (i.e. ‘adjusted’ to the F-contrast). Finally, a single representative timeseries is computed for each ROI by performing a principal components analysis (PCA) across voxels and retaining the first component (or principal eigenvariate). This approach is used rather than taking the mean of the timeseries, because calculating the mean would cause positive and negative responses to cancel out (and further that means are effected by extreme values). That could pose a problem due to centre-surround coding in the brain, where excitatory responses are surrounded by inhibitory responses – and would cancel if averaged.

Finally, prior to DCM model estimation, the software automatically checks whether the fMRI data are within the expected range (spm_dcm_estimate.m). If the range of the fMRI data exceeds four (in the units of the data), DCM rescales the data to have a range of four, on a per-subject basis. This was the case for our data.

These steps produced one timeseries per region, for each subject, which were then entered into the DCM analysis. The complete pipeline above can be performed in the SPM software semi-automatically, using the steps described in the accompanying practical guide.

Appendix 2. Observation noise specification

DCM separately estimates the precision (inverse variance) of zero-mean additive white noise for each brain region (spm_nlsi_gn.m). The white noise assumption is used because the preliminary general linear model estimates serial correlations, which are used to whiten principal eigenvariates from each region. From Equation (1) we have the model:

Y=g(z,θ(h))+X0Β0+ε (8)

To simplify the implementation, Y is vectorised (the timeseries from each region are stacked on top of one another) to give y=vec(Y). The observation noise ε is specified according to a normal density:

εN(0,Σy) (9)

In practice, DCM uses the precision matrix Πy which is the inverse of the covariance matrix Σy. It is specified by a multi-component model, which is a linear mixture of precision components Qi with one component per brain region i=1R. Each precision matrix is weighted by a parameter λi which is estimated from the data:

Πy=iexp(λi)Qi (10)

The diagonal elements of the precision matrix Qi equal one for observations associated with brain region i and zero elsewhere. Taking the exponential of parameter λi ensures that the estimated precision cannot be negative. In total, in the experiment presented here, we had 792 observations per subject (V=198 fMRI volumes times R=4 brain regions), and the corresponding precision components are illustrated in Fig. A1.

From Table 3, the prior density for parameter λi was N(6,1128). This means that scaling factor exp(λi) had a lognormal prior density: Lognormal(6,1128). The resulting prior expected precision was exp(6)=403.43 with 90% credible interval [348.84 466.56]. This prior says that the data are expected to have a high signal-to-noise ratio, because fMRI data are highly pre-processed and averaged, and are selected from brain regions that are known to show experimental effects. Model inversions are therefore preferred which ascribe a high level of variance to the model rather than to noise.

Fig. A1.

Fig. A1

Illustration of the observation noise model in DCM for fMRI. Each precision component Qi was a matrix with VR=792 elements, where V is the number of volumes (observations) and R is the number of brain regions. There was one precision component per brain region. Log scaling parameter λi was estimated from the data and scaled up or down the corresponding component Qi.

Appendix 3. Derivation of the fMRI neural model

Neural responses may be written generically as follows:

z˙=dzdt=f(z,u) (11)

where vector z is the state or level of activity in each region, z˙ is the rate of change in each brain region – called the neural response - and f is a function describing the change in brain activity in response to experimental inputs u. The ‘true’ function f would be tremendously complicated, involving the nonlinear, high dimensional dynamics of all cell types involved in generating a neural response. Instead, we can approximate f using a simple mathematical tool – a Taylor series. The more terms we include in this series, the closer we get to reproducing the true neural response. The definition of the Taylor series T up to the second term, with two variables z and u evaluated at z=m and u=n is:

T(z,u)=f(m,n)+(zm)fz+(un)fu+12((zm)2fzz+2(zm)(un)fzu+(un)2fuu) (12)

where fz and fu are the partial derivatives of f with respect to z and u, fzz and fuu are the second order derivatives and fzu is the mixed derivative (i.e. the derivative of f with respect to z of the derivative with respect to u, or vice versa). Each of these partial derivatives is evaluated at (z=m,u=n). Setting m=0 and n=0, defining the baseline neural response f(m,n)=0, and dropping the higher order terms we get the simpler expression:

T(z,u)=0+zfz+ufu+12(z2fzz+2zufzu+u2fuu) (13)

By re-arrangement of the final term:

T(z,u)=zfz+ufu+zufzu+12(z2fzz+u2fuu) (14)

Finally, factorizing z and dropping the final term (as z2 and u2 will be very small around the origin) gives:

T(z,u)=(fz+ufzu)z+ufu=(A+Bu)z+Cu (15)

here, we have assigned letters to the three derivative terms A=fz,B=fzu,C=fu, which gives the expression for the neural model used in the DCM literature (Equation (2)). With multiple brain regions, these derivative terms become matrices. As introduced in the main text, matrix A is the rate of change in neural response due to the other neural responses in the system – i.e. the effective connectivity. B is the rate of change in effective connectivity due to the inputs and is referred to as the bilinear or interaction term. Finally, C is the rate of change in neural response due to the external input, referred to as the driving input. In the DCM framework, A, B and C become parameters which are estimated from the data. (To apply negativity constraints on the self-connections, A and B are sub-divided into intrinsic and extrinsic parts, see Equation (3).)

Appendix 4. The neural parameters

Whereas Appendix 3 motivated the DCM neural model as a function approximated by a Taylor series, here we consider it from the perspective of a simple dynamical system, to help gain an intuition for the parameters. Consider a DCM with a single brain region, driven by a brief stimulus at time t=0. The neural equation can be simplified to the following:

z˙=az (16)

Where self-connection or rate constant a has units of Hz and is negative. The solution to this equation, the neural activity at any given time, is an exponential decay (under the constraint that a is negative):

z(t)=z(0)exp(at) (17)

where z(0) is the initial neuronal activity. This function is plotted in Fig. A2 (left) with parameter =0.5Hz , which is the default value in DCM.

Fig. A2.

Fig. A2

Illustration of neural response as an exponential decay. Left: the neural response under the default prior of a = −0.5Hz to an instantaneous input at time zero. Also plotted are the corresponding neural time constant τn and half-life. Middle: The resulting prior over time constant τn in the absence of modulation. The median is τn=2 seconds with 90% credible interval [1.63s 2.46s]. Right: The prior over τn in the presence of modulation. The median is τn=2 seconds with 90% credible interval [0.38s 10.49s]. Dashed lines in the middle and right panels show the median.

Fig. A2 (left) also illustrates two common ways of characterizing the rate of decay. The neural time constant τn is defined as:

τn=1a (18)

This is the time in seconds taken for the neural activity to decay by a factor of 1e (36.8% of its peak response). Given a=0.5Hz the time constant is τn=2s. This inverse relationship between the rate constant a and time constant τn is why the connectivity parameters in DCM are in units of Hz (Hz is 1/seconds). It can be more intuitive to express the rate of decay as the half-life, which is the time at which the activity decays to half its starting value. Given the self-connection of −0.5 Hz, the half-life is:

t12=τnln2=1.39s. (19)

In DCM a prior probability density is specified over each self-connection parameter a, which in turn specifies our expectation about a typical region's time constant. Fig. A2 (centre) shows the resulting prior time constant in DCM for fMRI. The median is 2s with 90% of the probability mass (the credible interval) between 1.63s and 2.46s.

In our analyses, we allowed self-connections to be modulated by experimental conditions. The rate constant a was therefore supplemented to give a+bu (see Equation (2) of the main text). The modulatory parameter b, multiplied by the experimental input u, could increase or decrease the region's rate of decay. Fig. A2 (right) shows the prior time constant for connections with modulation switched on (where u=1), giving 90% credible interval [0.38s 10.49s]. These plots make clear that DCM for fMRI does not model the activity of individual neurons, which typically have time constants on the order of milliseconds. Rather, it models the slow emergent dynamics that evolve over seconds and arise from the interaction of populations of neurons. For details of how these plots were generated, please see the supplementary text.

Appendix 5. Haemodynamic and BOLD signal model

The translation of neural activity z, predicted by the DCM neural model, to observed BOLD response y, is described by a three-part model illustrated in Fig. A3. We will summarise each of the three parts in turn.

Fig. A3.

Fig. A3

The model used to translate from neural activity to the BOLD signal in DCM. This is split into three parts. i. Neural activity z(t) triggers a vasoactive signal s (such as nitric oxide) which in turn causes an increase in blood flow. ii. The flow inflates the blood vessel like a balloon, causing a change in both blood volume v and deoxyhaemoglobin (dHb) q. iii. These combine non-linearly to give rise to the observed BOLD signal. A key reference for each part of the model is given - see text for further details. Symbols outside the boxes are parameters and those in bold type are free parameters that are estimated from the data: decay κ, transit time τh and ratio of intravascular to extravascular contribution to the signal εh. See Table 2 for a full list of symbols. Adapted from Friston et al. (2000).

Table 2.

Symbols.

Variable Dimension Units Meaning
A R×R Hz Effective connectivity (average or baseline)
AE R×R Hz Extrinsic average or baseline effective connectivity
AI R×R Log scaling parameters on average or baseline intrinsic connections
α 1×1 Grubb's exponent (stiffness of blood vessels)
Bk R×R Hz Modulatory input parameters for condition k
BE(k) R×R Hz Modulation of extrinsic connections by condition k
BI(k) R×R Log scaling parameters on modulation of intrinsic connections by condition k
Β0 C0×R Parameters for null effects
C R×J Hz Driving input parameters
C0 1×1 Number of first level covariates of no interest
E0 1×1 Resting oxygen extraction fraction
ε V×R Observation noise
εh 1×1 Fraction of intravascular to extravascular signal
F 1×1 Nats Negative variational free energy for a given model
f Neural model
fin 1×1 Hz Rate of blood inflow
fout 1×1 Hz Rate of blood outflow
g Observation model
γ 1×1 Hz Rate of decay of feedback to vasodilatory signal
J R×R Hz Effective connectivity or Jacobian matrix
K 1×1 Number of experimental conditions
kx 1×1 Coefficient within the BOLD signal model
κ 1×1 Hz Rate of vasodilatory signal decay
λi 1×1 Log scaling parameter for covariance component i
PH(1) 1×1 Total haemodynamic parameters per DCM
PN(1) 1×1 Total neural parameters per DCM
Pε(1) 1×1 Total observation parameters per DCM
Πy (VR)×(VR) Precision of observations (measurements)
Qi (VR)×(VR) Covariance component i
q 1×1 Level of deoxyhaemoglobin normalized to rest
R 1×1 Number of modelled brain regions
R 1×1 Total voxels (and timeseries) in the MRI volume
R2E 1×1 Hz Extravascular transverse relaxation rate
R2I 1×1 Hz Intravascular transverse relaxation rate
r0 1×1 Constant relating R2I to oxygen extraction rate
s 1×1 Vasodilatory signal
S 1×1 Modelled BOLD signal
S0 1×1 Modelled BOLD signal at rest
SE 1×1 Extravascular contribution to S
SI 1×1 Intravascular contribution to S
SE0 1×1 Extravascular effective spin density
SI0 1×1 Intravascular effective spin density
Σy (VR)×(VR) Covariance of the observations (measurements)
TU 1×1 Total time points in the inputs U
TE 1×1 Secs Echo time
τn 1×1 Secs Example neural time constant
τh 1×1 Secs Haemodynamic transit time
ϑ0 1×1 Frequency offset - outer surface of magnetized values
θh PH(1)×1 All first level haemodynamic parameters
θn PN(1)×1 All first level neural parameters
θε Pε(1)×1 All first level observation parameters
U TU×K All experimental inputs
ut J×1 All experimental inputs at time t
uk(t) 1×1 Experimental input by condition k at time t
ϑ0 1×1 Hz Frequency offset at the outer surface of magnetised vessels
V 1×1 Total measurements (volumes) per subject
v 1×1 Blood volume normalized to rest
V0 Resting venous blood volume fraction
V1 Blood volume fraction following neural activity
Vh 1×1 Fraction of intravascular blood volume
X0 V×C0 Design matrix for null effects
Y V×R Observed timeseries from all regions of interest
Yˆ V×R All timeseries from the acquired MRI volume
z R×1 Neural activity in each region

Regional cerebral blood flow (rCBF)

Neural activity z(t) drives the production of a chemical signal s that dilates local blood vessels. This causes oxygenated blood to flow into the capillaries, where oxygen is extracted. As a result, partially deoxygenated blood flows into the veins (venous compartment). The vasodilatory signal s decays exponentially and is subject to feedback by the blood flow fin that it induces:

f˙in=ss˙=z(t)κsγ(fin1) (20)

where parameter κ is the rate of decay for the signals, and γ is the time constant controlling the feedback from blood flow. Empirical estimates have shown κ to have a half-life of around 1 s (Friston et al., 2000), placing it in the correct range to be mediated by nitric oxide (NO). Adjusting κ primarily changes the peak height of the modelled BOLD response, whereas adjusting γ primarily changes the duration of response. Both parameters also modulate the size of the post-stimulus undershoot.

Venous balloon

Increased blood flow causes a local change in the volume of blood v in the blood vessel, as well as the proportion of deoxyhaemoglobin q (dHb). This process is captured by the Balloon model of Buxton et al. (1998). It treats the venous compartment as a balloon, which inflates due to increased blood flow and consequently expels deoxygenated blood at a greater rate. The change in blood volume v, normalized to the value at rest, depends on the difference between blood inflow and outflow:

τhv˙=fin(t)fout(v,t)fout(v,t)=v(t)1α (21)

Where the time constant τh is the mean transit time of blood, i.e. the average time it takes for blood to traverse the venous compartment. Grubb's parameter α controls the stiffness of the blood vessel (Grubb et al., 1974) and adjusting it has the effect of changing the peak height of the modelled BOLD response.

The increase in blood volume following neural activity is accompanied by an overall decrease in relative dHb content, the rate of which depends on the delivery of dHb into the venous compartment minus the amount expelled:

τhq˙=fin(t)1(1E0)1fin(t)E0fout(v,t)q(t)v(t) (22)

The first expression on the right hand side of Equation (22) approximates the fraction of oxygen extracted from the inflowing blood, which depends on the inflow fin and the resting oxygen extraction fraction E0 (the percentage of the oxygen removed from the blood by tissue during its passage through the capillary network). The second term relates to the outflow, where the ratio q/v is the dHb concentration.

BOLD signal

Finally, the change in blood volume v and dHb q combine to cause the BOLD signal S, measured using fMRI. For the purpose of this paper we define S as the signal acquired using a gradient echo EPI readout. The model used in DCM is due to Buxton et al. (1998) and Obata et al. (2004), which were extended and re-parameterised by Stephan et al. (2007). In the following paragraphs we provide a recap of the basic mechanisms of MRI and functional MRI (fMRI), in order to motivate the form of the BOLD signal model. Readers familiar with MR physics may wish to skip this introduction.

We will use classical mechanics to describe the way the MR signal is generated. When entering an MRI scanner, the subject is exposed to the main magnetic field b0. This magnetic field is always on and its axis is aligned with the tunnel of the scanner. All the hydrogen protons of the body can be thought of as acting like tiny magnets whose strength is measured by their magnetic moments μ. In what follows, the coordinate system x,y,z is used where z corresponds to the b0 axis, y and x are the orthogonal vectors forming the transverse plane. When submitted to the magnetic field B0, two phenomena occur:

1/All of the proton magnetic moments precess about the b0 axis at the Larmor frequency, which is proportional to the amplitude of the b0 field strength (e.g. 123  MHz at 3T).

2/The proton magnetic moments orient themselves such that their vector sum is a net magnetization vector, m, aligned with the b0 field axis and pointing in the same direction (Fig. A4.i). The net magnetization vector m can be decomposed into two components, the longitudinal component mz along the z axis and the transverse component mxy, which is the projection of m into the transverse plane. The transverse component is zero when the system is at equilibrium; i.e., when the net magnetization is aligned with the z axis, yet it is only the transverse component that can be measured in MRI.

In order to disturb the equilibrium state and thereby create a transverse component mxy that can be detected, a rotating magnetic field b1 is applied orthogonal to the b0 axis for a short period of time. This is termed ‘excitation’ and results in the tilting of the net magnetization towards the transverse plane (Fig. A4.ii).

Once the b1 field is turned off, the net magnetization has a transverse component and continues to precess around the main magnetic field, b0. Since the precession frequency is proportional to the amplitude of the magnetic field, any spatial variation of the magnetic field amplitude across one voxel will induce a difference in precessional frequency for the protons. For this reason, the protons accumulate a delay relative to each other and so have differential phase orientation (Fig. A4.iii). Over time the delays, or relative phase difference, increase (Fig. A4.iv). As a result their vector sum; i.e., the transverse componentmxy decreases. This process, whereby the transverse component of the net magnetization decreases, is called effective transverse relaxation. It is characterized by an exponential decay with a time constant T2 (Fig. A4.v), or alternatively a relaxation rate R2, whereby R2=1T2.

Crucially, for functional MRI, dHb and oxyhaemoglobin (Hb) molecules have different magnetic susceptibility (i.e. a different response to being placed in a magnetic field). Unlike Hb, which exhibits a weak, diamagnetic response to the main magnetic field, dHb exhibits a stronger, paramagnetic response. At the boundaries between two tissues with different magnetic susceptibilities, the magnetic field is distorted, increasing the local spatial inhomogeneity in the amplitude of the magnetic field. The decrease in dHb following neural activity makes the blood less paramagnetic, and more similar to the surrounding tissue in terms of magnetic susceptibility. As a result, the magnetic field around the blood vessel becomes less distorted, with a smaller range of precessional frequencies of protons in the voxel. As a consequence, less differential phase accumulates between the proton magnetic moments and the amplitude of the transverse component of the net magnetization vector mxy decreases less rapidly. This corresponds to a shorter R2 (or equivalently a longer T2). Therefore, at the time the data are acquired, TE (Echo Time), the signal will be higher if it follows a period of neural activity.

Fig. A4.

Fig. A4

Generation of the Magnetic Resonance (MR) signal. i. When exposed to a strong magnetic field, proton magnetic moments μ add together to create a net magnetization m, aligned with the B0 axis pointing in the same direction as this main field. ii After excitation with a flip angle (α) imparted by a rotating B1 field, applied orthogonal to B0 for a short period of time, the net magnetization is composed of a longitudinal and a transverse component, which precesses about the B0 axis. iii Inhomogeneity in the magnetic field within a voxel causes the protons' magnetic moments to precess with different frequency leading to differential phase (orientation) between the protons, reducing the transverse component mxy which is the vector sum of all the protons. iv. The differential phase accumulated by the protons increases over time. v. As a result, the amplitude of the transverse component of the net magnetization (i.e. the detectable MR signal) further decreases, following an exponential decay characterized by the effective transverse relaxation rate R2* ​= ​1/T2*. The MR signal is acquired at an echo time TE. vi. DCM assumes that there are two contributions to the measured signal – intravascular (Si) and extravascular (Se) – each with their own R2 relaxation rates.

Having revised the fundamentals of fMRI, we now return to the BOLD signal model in DCM. It follows Ogawa et al. (1993), in treating the tissue within a voxel as consisting of many small cubes, each with a cylinder running through the centre (Fig. A4vi). There are two compartments – the extravascular tissue outside the cylinder and the blood vessel (intravascular venous compartment) that is filled with blood. The BOLD signal at rest S0 is modelled as a linear mixture of these extravascular and intravascular contributions (Buxton et al., 1998):

S0=(1V0)Se+V0Si (23)

Where V0 is the resting venous blood volume fraction. From Obata et al. (2004), each compartment's resting BOLD signal at the time of measurement, TE in seconds, is modelled by:

Se=Se0exp(R2eTE)Si=Si0exp(R2iTE) (24)

where R2e and R2i are the effective transverse relaxation rates for the extravascular and intravascular compartments respectively, in units of Hz, and Se0 and Si0 are the maximal signals originating from each compartment before any signal decrease due to differential dephasing of the protons. Following neural activation, there will be an altered BOLD signal, S, compared to S0, written ΔS=SS0. A linear approximation of this change is as follows:

ΔSS0(ΔR2eTE)(V0εhΔR2iTE)+(V0V1)(1εh) (25)

where ΔR2e and ΔR2i is the change in each compartment's R2 between activation and rest, εh=SiSe is the ratio of intra-to extra-vascular signal contributions and V1 is the venous blood volume fraction following activation (Obata et al., 2004; Stephan et al., 2007). Obata et al. (2004) derived approximations for the changes in extravascular and intravascular signal decay:

ΔR2e=4.3ϑ0V0E0(q1)ΔR2i=r0E0(qv1) (26)

where ϑ0 is the frequency offset at the outer surface of magnetised vessels in units of Hz, E0 is the fraction of oxygen extraction at rest, and r0 is a constant relating ΔR2i to the oxygen extraction rate. Plugging these into Equation (25) and re-arranging gives the final expression (Stephan et al., 2007):

ΔSS0V0(k1(1q)+k2(1qv)+k3(1v))k1=4.3ϑ0E0TEk2=εhr0E0TEk3=1εh (27)

The three terms inside the brackets, weighted by coefficients k1, k2 and k3, relate to the extravascular contribution to the BOLD signal, the intravascular contribution and the ratio of extravascular and intravascular signals respectively.

Summary and implementation

The implementation of haemodynamic model in SPM is split into two parts: the rCBF/Balloon model (spm_fx_fmri.m) and the BOLD signal model (spm_gx_fmri.m). They include five hidden states: the net neural activity z, vasoactive signal S, log rCBF ln(fin), log venous volume ln(v) and log dHb ln(q). The logs of these states are taken in order to enforce positivity constraints, requiring the differential equations to be supplemented accordingly:

dln(fin)dt=f˙infindln(v)dt=v˙vdln(q)dt=q˙q (28)

There are nine parameters overall, labelled in Fig. A2. To keep estimation of the model tractable, only three of these are estimated from the fMRI data, and the log of their values are estimated to ensure positivity: the rate of signal decay ln(κ) from the rCBF model, the transit time ln(τh) from the balloon model, and the ratio of intra-to extra-vascular signals ln(εh) from the BOLD signal model. The priors for these parameters are listed in Table 3.

Appendix A. Supplementary data

The following is/are the supplementary data to this article:

Multimedia component 1
mmc1.docx (17.8KB, docx)

References

  1. Almgren H., Van de Steen F., Kühn S., Razi A., Friston K., Marinazzo D. Variability and reliability of effective connectivity within the core default mode network: a multi-site longitudinal spectral DCM study. Neuroimage. 2018;183:757–768. doi: 10.1016/j.neuroimage.2018.08.053. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Bastos-Leite A.J., Ridgway G.R., Silveira C., Norton A., Reis S., Friston K.J. Dysconnectivity within the default mode in first-episode schizophrenia: a stochastic dynamic causal modeling study with functional magnetic resonance imaging. Schizophr. Bull. 2014;41:144–153. doi: 10.1093/schbul/sbu080. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Bastos A.M., Usrey W.M., Adams R.A., Mangun G.R., Fries P., Friston K.J. Canonical microcircuits for predictive coding. Neuron. 2012;76:695–711. doi: 10.1016/j.neuron.2012.10.038. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bernal-Casas D., Balaguer-Ballester E., Gerchen M.F., Iglesias S., Walter H., Heinz A., Meyer-Lindenberg A., Stephan K.E., Kirsch P. Multi-site reproducibility of prefrontal–hippocampal connectivity estimates by stochastic DCM. Neuroimage. 2013;82:555–563. doi: 10.1016/j.neuroimage.2013.05.120. [DOI] [PubMed] [Google Scholar]
  5. Buxton R.B., Wong E.C., Frank L.R. Dynamics of blood flow and oxygenation changes during brain activation: the balloon model. Magn. Reson. Med. : official journal of the Society of Magnetic Resonance in Medicine / Society of Magnetic Resonance in Medicine. 1998;39:855–864. doi: 10.1002/mrm.1910390602. [DOI] [PubMed] [Google Scholar]
  6. Chumbley J.R., Friston K.J., Fearn T., Kiebel S.J. A Metropolis-Hastings algorithm for dynamic causal models. Neuroimage. 2007;38:478–487. doi: 10.1016/j.neuroimage.2007.07.028. [DOI] [PubMed] [Google Scholar]
  7. Deco G., Jirsa V.K., Robinson P.A., Breakspear M., Friston K. The dynamic brain: from spiking neurons to neural masses and cortical fields. PLoS Comput. Biol. 2008;4 doi: 10.1371/journal.pcbi.1000092. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Dirkx M.F., den Ouden H., Aarts E., Timmer M., Bloem B.R., Toni I., Helmich R.C. The cerebral network of Parkinson's tremor: an effective connectivity fMRI study. J. Neurosci. 2016;36:5362–5372. doi: 10.1523/JNEUROSCI.3634-15.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Friston K., Zeidman P., Litvak V. Empirical Bayes for dcm: a group inversion scheme. Front. Syst. Neurosci. 2015;9:164. doi: 10.3389/fnsys.2015.00164. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Friston K., Mattout J., Trujillo-Barreto N., Ashburner J., Penny W. Variational free energy and the Laplace approximation. Neuroimage. 2007;34:220–234. doi: 10.1016/j.neuroimage.2006.08.035. [DOI] [PubMed] [Google Scholar]
  11. Friston K.J., Harrison L., Penny W. Dynamic causal modelling. Neuroimage. 2003;19:1273–1302. doi: 10.1016/s1053-8119(03)00202-7. [DOI] [PubMed] [Google Scholar]
  12. Friston K.J., Mechelli A., Turner R., Price C.J. Nonlinear responses in fMRI: the Balloon model, Volterra kernels, and other hemodynamics. Neuroimage. 2000;12:466–477. doi: 10.1006/nimg.2000.0630. [DOI] [PubMed] [Google Scholar]
  13. Friston K.J., Kahan J., Biswal B., Razi A. A DCM for resting state fMRI. Neuroimage. 2014;94:396–407. doi: 10.1016/j.neuroimage.2013.12.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Goulden N., Khusnulina A., Davis N.J., Bracewell R.M., Bokde A.L., McNulty J.P., Mullins P.G. The salience network is responsible for switching between the default mode network and the central executive network: replication from DCM. Neuroimage. 2014;99:180–190. doi: 10.1016/j.neuroimage.2014.05.052. [DOI] [PubMed] [Google Scholar]
  15. Grubb R.L., Raichle M.E., Eichling J.O., Ter-Pogossian M.M. The effects of changes in PaCO2 cerebral blood volume, blood flow, and vascular mean transit time. Stroke. 1974;5:630–639. doi: 10.1161/01.str.5.5.630. [DOI] [PubMed] [Google Scholar]
  16. Jung K., Friston K.J., Pae C., Choi H.H., Tak S., Choi Y.K., Park B., Park C.A., Cheong C., Park H.J. Effective connectivity during working memory and resting states: a DCM study. Neuroimage. 2018;169:485–495. doi: 10.1016/j.neuroimage.2017.12.067. [DOI] [PubMed] [Google Scholar]
  17. Kahan J., Foltynie T. Understanding DCM: ten simple rules for the clinician. Neuroimage. 2013;83:542–549. doi: 10.1016/j.neuroimage.2013.07.008. [DOI] [PubMed] [Google Scholar]
  18. Kahan J., Urner M., Moran R., Flandin G., Marreiros A., Mancini L., White M., Thornton J., Yousry T., Zrinzo L. Resting state functional MRI in Parkinson's disease: the impact of deep brain stimulation on ‘effective’connectivity. Brain : J. Neurol. 2014;137:1130–1144. doi: 10.1093/brain/awu027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Kiebel S.J., Kloppel S., Weiskopf N., Friston K.J. Dynamic causal modeling: a generative model of slice timing in fMRI. Neuroimage. 2007;34:1487–1496. doi: 10.1016/j.neuroimage.2006.10.026. [DOI] [PubMed] [Google Scholar]
  20. Li B., Daunizeau J., Stephan K.E., Penny W., Hu D., Friston K. Generalised filtering and stochastic DCM for fMRI. Neuroimage. 2011;58:442–457. doi: 10.1016/j.neuroimage.2011.01.085. [DOI] [PubMed] [Google Scholar]
  21. Marreiros A.C., Kiebel S.J., Friston K.J. Dynamic causal modelling for fMRI: a two-state model. Neuroimage. 2008;39:269–278. doi: 10.1016/j.neuroimage.2007.08.019. [DOI] [PubMed] [Google Scholar]
  22. Moran R., Pinotsis D.A., Friston K. Neural masses and fields in dynamic causal modeling. Front. Comput. Neurosci. 2013;7:57. doi: 10.3389/fncom.2013.00057. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Obata T., Liu T.T., Miller K.L., Luh W.-M., Wong E.C., Frank L.R., Buxton R.B. Discrepancies between BOLD and flow dynamics in primary and supplementary motor areas: application of the balloon model to the interpretation of BOLD transients. Neuroimage. 2004;21:144–153. doi: 10.1016/j.neuroimage.2003.08.040. [DOI] [PubMed] [Google Scholar]
  24. Ogawa S., Menon R.S., Tank D.W., Kim S.G., Merkle H., Ellermann J.M., Ugurbil K. Functional brain mapping by blood oxygenation level-dependent contrast magnetic resonance imaging. A comparison of signal characteristics with a biophysical model. Biophys. J. 1993;64:803–812. doi: 10.1016/S0006-3495(93)81441-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Papadopoulou M., Cooray G., Rosch R., Moran R., Marinazzo D., Friston K. Dynamic causal modelling of seizure activity in a rat model. Neuroimage. 2017;146:518–532. doi: 10.1016/j.neuroimage.2016.08.062. [DOI] [PubMed] [Google Scholar]
  26. Park H.J., Pae C., Friston K., Jang C., Razi A., Zeidman P., Chang W.S., Chang J.W. Hierarchical dynamic causal modeling of resting-state fMRI reveals longitudinal changes in effective connectivity in the motor system after thalamotomy for essential tremor. Front. Neurol. 2017;8:346. doi: 10.3389/fneur.2017.00346. [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Razi A., Kahan J., Rees G., Friston K.J. Construct validation of a DCM for resting state fMRI. Neuroimage. 2015;106:1–14. doi: 10.1016/j.neuroimage.2014.11.027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Rosch R.E., Hunter P., Baldeweg T., Friston K., Meyer M. 2017. Imaging and Dynamic Causal Modelling Reveal Brain-wide Changes in Effective Connectivity and Synaptic Dynamics during Epileptic Seizures. (bioRxiv [PREPRINT]) [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Seghier M.L., Friston K.J. Network discovery with large DCMs. Neuroimage. 2013;68:181–191. doi: 10.1016/j.neuroimage.2012.12.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Seghier M.L., Josse G., Leff A.P., Price C.J. Lateralization is predicted by reduced coupling from the left to right prefrontal cortex during semantic decisions on written words. Cerebr. Cortex. 2011;21:1519–1531. doi: 10.1093/cercor/bhq203. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Seghier M.L., Zeidman P., Neufeld N.H., Leff A.P., Price C.J. Identifying abnormal connectivity in patients using dynamic causal modeling of FMRI responses. Front. Syst. Neurosci. 2010;4 doi: 10.3389/fnsys.2010.00142. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Stephan K.E. On the role of general system theory for functional neuroimaging. J. Anat. 2004;205:443–470. doi: 10.1111/j.0021-8782.2004.00359.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Stephan K.E., Weiskopf N., Drysdale P.M., Robinson P.A., Friston K.J. Comparing hemodynamic models with DCM. Neuroimage. 2007;38:387–401. doi: 10.1016/j.neuroimage.2007.07.040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Stephan K.E., Penny W.D., Moran R.J., den Ouden H.E., Daunizeau J., Friston K.J. Ten simple rules for dynamic causal modeling. Neuroimage. 2010;49:3099–3109. doi: 10.1016/j.neuroimage.2009.11.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Stephan K.E., Kasper L., Harrison L.M., Daunizeau J., den Ouden H.E., Breakspear M., Friston K.J. Nonlinear dynamic causal models for fMRI. Neuroimage. 2008;42:649–662. doi: 10.1016/j.neuroimage.2008.04.262. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Zhou Y., Friston K.J., Zeidman P., Chen J., Li S., Razi A. The hierarchical organization of the default, dorsal attention and salience networks in adolescents and young adults. Cerebr. Cortex. 2018;28:726–737. doi: 10.1093/cercor/bhx307. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia component 1
mmc1.docx (17.8KB, docx)

RESOURCES