Abstract
A central question in neuroscience is how self-organizing dynamic interactions in the brain emerge on their relatively static structural backbone. Due to the complexity of spatial and temporal dependencies between different brain areas, fully comprehending the interplay between structure and function is still challenging and an area of intense research. In this paper we present a graph neural network (GNN) framework, to describe functional interactions based on the structural anatomical layout. A GNN allows us to process graph-structured spatio-temporal signals, providing a possibility to combine structural information derived from diffusion tensor imaging (DTI) with temporal neural activity profiles, like that observed in functional magnetic resonance imaging (fMRI). Moreover, dynamic interactions between different brain regions discovered by this data-driven approach can provide a multi-modal measure of causal connectivity strength. We assess the proposed model’s accuracy by evaluating its capabilities to replicate empirically observed neural activation profiles, and compare the performance to those of a vector auto regression (VAR), like that typically used in Granger causality. We show that GNNs are able to capture long-term dependencies in data and also computationally scale up to the analysis of large-scale networks. Finally we confirm that features learned by a GNN can generalize across MRI scanner types and acquisition protocols, by demonstrating that the performance on small datasets can be improved by pre-training the GNN on data from an earlier study. We conclude that the proposed multi-modal GNN framework can provide a novel perspective on the structure-function relationship in the brain. Accordingly this approach appears to be promising for the characterization of the information flow in brain networks.
Subject terms: Dynamical systems, Network models, Computational science
Introduction
Brain connectivity comes in different flavors, either resting on the structural anatomical layout, as derived from diffusion tensor imaging (DTI) or based on temporally resolved activity patterns, like observed in functional MRI (fMRI)1. White matter tracks reconstructed from DTI provide a foundation for structural connectivity (SC) and can be used to quantify the (static) anatomical connection strength between brain regions. On the other hand fMRI enables us to map out dynamic neural activity distributions across the brain, whereas the coherence of fluctuations is usually referred to as functional connectivity (FC). Such functional states can alternate very rapidly in contrast to changes in the structural connectome, alterations of which are mainly related to the natural development of the brain, aging or disease2,3. Therefore the brain structure can be considered as static during a fMRI measurement, in comparison to the fast functional fluctuations. Intuitively one might follow then the motto “structure determines function”, but it has been shown that the relationship between brain structure and function is quite complex and still a focus of intense research4–8. For instance, brain regions with robust SC usually show also high FC, but the inverse is not necessarily true9. While FC is a statistical measure with no information concerning the directionality of the relation, effective connectivity and directed functional connectivity measures try to infer causal dependencies in functional imaging data10. Thus connectivity measures derived from different modalities can provide distinct, but complementary aspects of brain connectivity11–13. Still, studying their relations is challenging mainly due to the complex spatio-temporal dependencies and inherent difficulty in long term forecasting.
In this paper we propose a data driven model, which combines information from fMRI and DTI to infer causal dependencies between brain regions. Temporal activity patterns of neuron pools, interconnected by the spatial anatomical layout, can be interpreted as time-varying graph structured signals. For such applications, graph neural networks (GNN) have shown to be useful, providing a possibility to process data with graph-like properties in the framework of artificial neural networks (ANN)14. Motivated by their success in computer vision15,16, convolution operations were recently extended to the graph domain17,18. Training such convolution filters in ANN enables us to capture inherent spatial dependencies in the non-Euclidean geometry of graphs, which are used in our context to integrate spatial relations of brain networks, based on their structural anatomical connections. Further, temporal dependencies in a dynamic system can be acquired by recurrent neural networks (RNNs) that have proven to be well suited for processing data with sequential structure. In our study, RNNs learn temporal characteristics of brain dynamics, like those observed in resting-state fMRI. A certain type of GNN architecture denoted as diffusion convolution recurrent neural network (DCRNN)19, provides the possibility to integrate spatio-temporal information of graph-structured signals. By combining fMRI with DTI data, the idea is to replicate brain dynamics more accurately, to get an improved understanding of functional interactions between brain regions, which are physically constrained by their structural backbone20.
Causal relationships between brain regions can be revealed by directed functional connectivity and effective connectivity. Two prominent and distinct approaches have been established in recent years10. The first one is based on a simple idea taken up by the British econometrician Clive Granger21. If one event A causes another event B, then A would precede B, and information on the occurrence of A should contribute to the prediction of the occurrence of event B. Such temporal dependencies between multivariate processes are typically described in the framework of a multivariate vector auto regressive (VAR) model, building a foundation for Granger causality (GC). By trying to make accurate predictions of temporal neural profiles, GC tests if adding information about neural activity in brain region B helps to improve the prediction of the activity in region A (and vice versa). This provides an exploratory measure for directed causal dependencies between segregated brain areas.
The second popular approach is methodologically different: Dynamic causal modeling (DCM) relies on a mechanistic input-state-output model of neuron pools, describing the effective connectivity strength between brain areas22. Experimental conditions and stimuli are encoded in input functions, and the model output can be related to empirically observed electromagnetic or hemodynamic responses. In a Bayesian framework, effective couplings of neural populations are estimated, providing a neurophysiological perspective on causal relationships between different regions in the brain. However due to its relatively high computational complexity, the analysis with DCM is usually limited to a few pre-defined regions in the brain only, what could neglect relevant components for the analysis23.
Here we present a data-driven machine learning approach that combines structural and functional information of neuron pools in a predictive framework for brain dynamics. By studying spatio-temporal dependencies between brain areas which were learned by the DCRNN model from DTI and fMRI data, we deduce the information flow between segregated areas in the brain. This provides us with a multi-modal data-driven perspective on causal relationships within brain networks. Currently, for investigating causal structures from an information theoretic perspective, a VAR is most often used as the underlying predictive model for Granger causality inference10,24. In our study we compare the VAR model to the multi-modal DCRNN model for this application. We test the capabilities of the two models to replicate empirically observed fMRI signals, in order to assess how well they can capture the underlying functional dynamics. First we show that the DCRNN is able to make more accurate long-term predictions in fMRI data. While a classical VAR model has to fit a parameter for each possible pair of the N brain areas in the network, which parameters then grow with an order , the DCRNN learns localized filters on the structural network, also making its number of parameters independent of the network size18. This is especially useful for the analysis of large brain networks, when only sparse imaging data is available. Moreover, as white matter tracks build the physical substrate for the propagation of neural signals, a greater neurophysiological plausibility is provided by the DCRNN, because neural interactions are related to their anatomical connections in this model as well. This property makes such a GNN architecture also suitable for an explorative study of large-scale brain networks, unlike classical DCM, which is often limited to a few predefined brain regions only23.
Finally such a GNN framework, which integrates anatomical and functional neuroimaging data, can provide a novel perspective on the general relationship between brain structure and function. Until now different computational models have contributed valuable insights into the complex structure–function relationship, by simulating how empirically observed FC patterns can emerge from the structural backbone4,6,9,25. Furthermore, by relying on methods from graph theory, functional networks have been derived from a mapping of the underlying structural graphs26–28. Also approaches from machine learning have been employed to predict the strength of functional connections from the anatomical connectivity29,30, and more recently some models have been proposed that additionally considered dynamic FC patterns for studying this relation20,31.
So far most of the methods for investigating the structure–function relationship try to predict only the temporal coherence patterns of functional fluctuations (FC) from their structural connectivity profile (SC), but such a correlation based FC might be limited in its ability to characterize the rich nature of functional brain dynamics. Instead of replicating only FC patterns, our approach directly models the measured neural activity in different brain areas, thereby capturing more of the original information in the empirically observed functional dynamics. By modeling neural interactions on the structural backbone, this method allows us to reconstruct the amount of information on activity distributions that occurs in structurally connected brain regions. Accordingly, such a spatio-temporal GNN model can provide a novel possibility to investigate the complex structure-function relation under a different lens.
Usually for a good performance more complex machine learning models require a larger amount of data, but it is not economical feasible in MRI to perform studies with very large sample sizes. To account for these issues we demonstrate that also in our context transfer learning32 can enhance the model accuracy of small datasets. We pre-train the DCRNN on a large-scale dataset of 100 resting-state fMRI (rs-fMRI) sessions provided by the Human Connectome Project33 (HCP). We then show that the pre-trained model considerably improves the predictive performance on a smaller independent dataset of 10 sessions compared to standard training. This points to the ability of the DCRNN to generalize across scanner types and acquisition protocols to a certain extent, enabling the possibility for transfer learning.
Results
Model description
In this study we use the DCRNN model19 architecture to explore the spatio-temporal relationships of brain dynamics in resting state fMRI. An overview of the model structure is provided in Fig. 1. To learn the temporal dependencies of the BOLD signal, recurrent neural networks (RNNs) with sequence-to-sequence learning are employed34. In such an architecture the encoder network maps information from an input sequence into a hidden representation, which is used by the decoding part to sequentially generate outputs, based on this encoded information. In the context of brain dynamics, the input sequence corresponds to measurements of the BOLD signal in N brain regions at time points, while the objective is to predict the signal at subsequent time points.
In addition to temporal, also spatial dependencies between brain regions are incorporated via diffusion convolution operations19. Consider the network of regions of interest (ROIs) as a graph , where denotes a set of vertices (nodes), represents a set of edges and is a weighted adjacency matrix. The latter represents the spatial connectivity of the nodes, i.e. the ROIs on the neuronal network, which are adjacent to each other, i.e. connected by an edge. Also the weights result from DTI, reflecting the axonal connection strength between the connected regions. Goal of the DCRNN model is to learn a function h(...) which maps past activity states , to future states:
1 |
The encoder, as well as the decoder of the DCRNN consist of gated recurrent units36, modified with graph convolutions18, and for training the model scheduled sampling was applied37. A detailed description of the model architecture is provided in the methods section.
Data description
For the first part of our evaluation, resting-state fMRI data from the S1200 release provided by the Human Connectome Project33 (HCP) was employed38. Further the multi-model parcellation proposed by Glasser et. al39 was applied to divide each hemisphere into 180 segregated regions. The BOLD signal in each region was averaged, so for each resting state session, time courses were obtained. During each session images were acquired, so the data can be arranged in a matrix . For the following analysis, we filter the data with a 0.04–0.07 Hz narrow band bandpass filter, as it has shown to be reliable and functionally relevant for gray matter activity20,40–43. We additionally present results in supplement IV, when employing a more liberal bandpass filter with cutoff frequencies between 0.02 and 0.09 Hz.
The input and output (label) samples for the DCRNN model were generated from the data in , by defining windows of length to obtain input sequences of neural activity states , and respective target sequences of length . The index t was propagated through each resting-state fMRI session, so in total input-output pairs were generated per session. The first of those time window samples of each fMRI session were used for training the DCRNN model, the subsequent for validation, and the last for testing. In total 4 resting-state fMRI sessions from 25 different subjects were employed for the evaluations in the following sections. The input and output length was chosen to be , what would correspond to a time span of roughly , based on the sampling with a repetition time 44. But note that in general the sequence-to-sequence architecture employed would be able to deal with arbitrary input and output signal lengths34.
In addition to temporal brain dynamics, also structural information was incorporated into the model, described by the anatomical connection strength between brain regions deduced from DTI. Therefore the DTI dataset provided in the S1200 release38 was further processed by employing multi-shell, multi-tissue constrained spherical deconvolution45, implemented in the MRtrix3 software package35. White matter tractography was performed to estimate whole brain structural connectivity between the regions of the multi-modal parcellation atlas39. The number of generated streamlines connecting two brain regions were used to define the edge strengths in the graph adjacency matrix . A more detailed description of the datasets and preprocessing involved can be found in the methods subsection ‘HCP data’.
Model performance
In a first step we assess the capabilities of the DCRNN model to learn temporal activity patterns in neuron pools, and their relationships across the spatial layout. As a first baseline we compare the DCRNN to the performance of a linear vector autoregressive (VAR) model21, further described in the subsection ‘Autoregressive models’. A common way to estimate causal relations between different regions of interest (ROIs) in a brain network, is to fit a multivariate VAR model to neural temporal activity patterns, like those observed in different neuroimaging modalities10,24,46. Evaluating the fitted VAR allows us to infer, if one spatial brain region, contains additional information about future activity profiles of other regions, indicating a causal dependency between them. The accuracy in replicating empirically observed neural activity profiles can indicate how well a model has learned the underlying process of neural dynamics, including the interactions and dependencies among brain regions. In this comparison we incorporated two different optimization methods for the estimation of the VAR coefficients. The first one employs an ordinary least squares (OLS) fit on the neural activity timecourses from individual rs-fMRI sessions47. The second approach, in analogy to the DCRNN, follows a gradient-descent based optimization48 on the windowed neural activity samples as outlined in the subsection ‘Data description’. For this evaluation we rely on the latter one, as it could improve the performance of the VAR, as described in the subsection ‘Autoregressive models’.
The evaluations were performed on test data from 100 rs-fMRI sessions (4 sessions from 25 subjects), using the last of the time window samples of each session, corresponding to 114 test samples per session. Within a time window length of both models can make relatively reasonable predictions, but also the difference in the prediction accuracy becomes apparent. A representative example of the accuracy of the two approaches is shown in Fig. 2, as well as their average performance on the complete testing data set. Figure 2a illustrates that a linear VAR model can generate in a few cases also correct long term predictions, but most often after ( s) the error starts to accumulate and the predictions become less accurate. The predictions of the DCRNN (Fig. 2b) remain stable over much longer forecasting horizons, and the average mean absolute error is considerably lower than the of the VAR.
To further test the significance of the performance difference across subjects, the overall MAE between the predicted and true BOLD signals were computed for each of the 25 subjects individually, as an average across sessions, brain regions and test samples. A paired t-test was applied and with considering a significance threshold of 0.05, the difference in forecasting accuracy between the models showed to be highly significant with across subjects. Due to the oscillations of the BOLD signal, the predicted signals tend to intersect the true ones at some point in time, what can be seen for example in Fig. 2a. After the truth and prediction have diverged from each other at previous timepoints, at this point of intersection the MAE tends to be smaller again. Therewith these intersections are reflected in some flatspots occurring in the error bars along the forecasting horizon, like observed in Fig. 2c. The forecasting error in dependence of different model output horizons is provided in Table 1. The difference in the performance between the DCRNN and VAR model already becomes apparent within the first few predicted timesteps, and the margin tends to further increase for larger prediction horizons.
Table 1.
Model | Forcasting horizon (TRs) | |||||
---|---|---|---|---|---|---|
5 | 10 | 15 | 20 | 25 | 30 | |
VAR | 0.0321 | 0.0589 | 0.0751 | 0.1099 | 0.1449 | 0.1786 |
DCRNN | 0.0018 | 0.0028 | 0.0065 | 0.0115 | 0.0163 | 0.0279 |
Additional analysis of the performance of the DCRNN model can be found in the Supplementary Information. At first in supplement I we reproduced the evaluations by testing the prediction accuracy on a cohort of new and unseen subjects. Further in Supplementary I we have discussed the impact of the training dataset size, as well as the role of the input horizon length on the model’s prediction performance. Additionally we studied the consistency of the model performance across subjects and we investigated, how the prediction accuracy depends on the different brain regions, to examine if there are areas with more or less complex temporal dynamics. To test the efficiency of neural network architecture implemented by the DCRNN, we then compared it to different baseline models in Supplementary II. Further in Supplementary III we compared the VAR and DCRNN on a smaller dataset collected at a different imaging site. To study the impact of a more liberal frequency filtering within the 0.02–0.09 Hz range, the equivalent evaluation is provided in Supplementary IV. In Supplementary V we evaluate the different approaches employing the volumetric AAL parcellation49 and performing an alternative method for reconstructing the anatomical connectivity50. Finally the effects of the timeseries input length and optimization methods on the VAR model performance are discussed in Supplementary VI.
Impact of spatial modeling
For this application of the DCRNN model, the anatomical connectivity was used to characterize spatial relations between nodes in the brain network, shaping the transition of activity between brain regions. To illustrate that the DCRNN indeed has learned relevant spatial interactions between different ROIs, we evaluate this recurrent neural network model, without employing graph (diffusion) convolution layers. This restriction considers only self-couplings (filters of order ) of nodes on the structural graph. Figure 3a shows the test MAE in dependence of the incorporated walk order K. The increase in computational time per epoch in dependence of included transition orders K is depicted in Fig. 3b. A more detailed comparison of the prediction MAE between the sequence-to-sequence model without graph convolutions (), and including spatial transitions up to order is illustrated in Fig. 3c.
These results show, that the vast amount of the information about future activity in one region comes from the region itself. But by including first order transitions on the structural network () the error can already be decreased by . Filters of higher orders only slightly improve the predictions further, as shown in Fig. 3a, but the computational load increases linearly with order K, like illustrated in Fig. 3b. The role of such transitions within the anatomical network can tell us something about the general structure–function relationship in the human brain, by pointing out how much information about functional dynamics comes from structurally connected regions. The comparison between and shows, that roughly up to of the predictive performance can be attributed to information from regions that are anatomically connected with each other. A paired t-test was applied to test the difference in the models accuracy between and across subjects, which turned out to be significant with a value of .
Causal connectivity
In this section the objective is to study the principle of information passing between different ROIs the DCRNN has learned from the neuroimaging data. As shown in the previous subsection ‘Impact of spatial modeling’, propagating information on the anatomical network can improve the predictions of the temporal evolution of the BOLD signal, displaying a dependence among structurally connected brain regions. Such a dependency might go beyond simple coherency based functional connectivity, as the latter usually assess only the temporal similarity of two signals. Observing that the past activity in some regions contains additional information about the future activity in other regions, beyond information retrieved from their own past, could indicate some flow of information among them and could provide a first indication for a causal dependency structure. Now to derive such a measure of causal connectivity strength, by following the idea of Granger21, the goal is to reconstruct how information about the activity in ROI A contributes to the prediction of the activity in ROI B. To reveal relationships inside the data by directly looking at the learned parameters is often challenging when ANN models become more complex. One simple strategy used to account for this problem is to induce perturbations in the models input space and then observe how these perturbations are propagated to the models outputs51,52.
In our context, the DCRNN first learns a function h(...), mapping the original input sequences of neural activity states to a predicted output sequences of future states . Then the information about the activity in a ROI is removed, by simply replacing the values in the input sequence with the mean value of the data distribution . Next the input sequence with the artificial perturbation in is projected by the model h(...) to an output sequence . Finally the differences of the models predictions with the perturbation in the input space in ROI , and the predictions with the original input can be compared. A measure of influence of the information in ROI on the predictions in other ROIs can then be defined as:
2 |
with describing the impact of region on region n. Here and denote the predictions in region n with and without the perturbation of in the input space respectively, of a test sample s at time step t.
To visualize this measure of influence of on each individual region n, values of can be projected onto the cortical surface. In the following we studied the impact of the parieto-insular vestibular cortex (PIVC) on all other brain regions. Here PIVC in the right hemisphere is characterized as a conjunction of ROIs R_OP2-3 and R_Ig, as defined by Glasser et al.39. Previous results show that this location coincides with the average location of PIVC across human subjects53,54. The perturbation was induced in R_OP2-3 and R_Ig simultaneously, and Fig. 4 illustrates the strength of influence on all other regions (encoded in red) of the target region PIVC (marked in blue).
The results of this analysis show that PIVC exhibited an interrelationship with the Sylvian fissure, the perisylvian cortex and the insula. Similar connectivity patterns have been observed using diffusion weighted imaging55,56 and resting state functional connectivity57 in human subjects as well as in non-human primates using tracer techniques58. Several separate regions of the vestibular cortex are located within this Sylvian network, including the posterior insular cortex area (PIC), a region critical to the integration of visual and vestibular cues (for human subjects59,60; for non-human primates the region is referred to as VPS58,61). The information flow within this Sylvian network is not fully understood yet. Current theories assume that vestibular and visual cues about self motion are combined within PIVC and PIC and are then further processed to the temporoparietal junction (TPJ), a larger cortical region located at the junction of the temporal and parietal cortices, where visual-vestibular signals are integrated into a representation of the self in space54. The results of the current analysis support this view by providing first evidence for a potentially causal relationship with the supramarginal gyrus, which is part of the TPJ. Further functional connectivity between PIVC was observed with the visual cortex. This result is interesting, since several studies have shown inhibitory interactions between the visual system and PIVC57,62–64, such that PIVC is inhibited when visual cues are processed attentively and vice versa. As shown by magnetic resonance spectroscopy, this inhibition of PIVC is reflected by a decrease of excitatory neurotransmitter (glutamate and its precursor glutamine) within PIVC, concomitant with an increase in negative BOLD signal in PIVC65. These inhibitory interactions are assumed to be modulated in magnitude by attention networks located in the visual and parietal cortices57.
Model generalization
Often one problem is the availability of a sufficient amount of data, in order to fully train and take advantage of machine learning models with large parameter spaces. Especially in MRI studies it is usually time-consuming and costly to acquire such large data sets. To account for this limitation, the concept of transfer learning was proposed in machine learning32. The basic idea behind transfer learning is that if only sparse data are available to learn a certain task, one can pretrain the model on a large-scale dataset of a similar task. In a next step, the feature representations learned on the large database can be used as an initialization for learning the desired target task. The goal is to transfer knowledge of one source domain to a target domain, by re-using the pretrained models weights. If the feature representation of the source domain is diverse enough, this can improve the model performance in comparison to starting the training without any prior knowledge, e.g. relying on a random initialization of the model weights32.
To investigate if transfer learning can also be suitable for our application, we studied the capabilities of the DCRNN to generalize across different datasets. Therefore we pretrained the DCRNN using the data provided by the HCP33, as described in the methods subsection ‘HCP data’. The model was pretrained for 70 epochs on in total 100 resting-state fMRI sessions, including the anatomical connectivity as reconstructed from DTI. Next we used a dataset acquired with a Siemens Magnetom Prisma 3T at the University of Regensburg (UR), where 10 different subjects participated in a resting-state fMRI session, including a DTI session, to acquire the corresponding structural imaging data. Each resting-state session of the UR dataset had a duration of 7.3 min, whereby 600 fMRI images were collected per scan. The acquisition parameters and the preprocessing involved are outlined in more detail in the subsection ‘UR data’. In analogy to the HCP dataset, the UR data was further processed by windowing the average BOLD signals in the regions defined by Glasser et al.39, thereby obtaining windows with an input and output length of timepoints. The first of these samples were used as a training data set, the subsequent for validation and the final for testing. We fine tuned the DCRNN, pretrained on the HCP data, by training it for 70 more epochs on the UR dataset, and initialized the second training with a lower learning rate of 0.001. This pretrained model was compared to the DCRNN, only trained on the UR dataset, and with weight parameters initialized randomly with Xavier/Glorot initialization66.
The comparison between relying on standard training, and utilizing transfer learning is illustrated in Fig. 5. Figure 5a shows the training and validation error during learning when starting with a random initialization of the weights in red. This model was trained in total for 140 epochs on the UR dataset only. In blue the training and validation error is depicted of the model, initially pretrained on the HCP dataset for 70 epochs, and fine tuned on the UR dataset for the subsequent 70 epochs. Figure 5a illustrates that at onset, the training error on the UR data is relatively high, but as the pretrained model adapts to the new dataset the MAE becomes considerably smaller than without pretraining. In Fig. 5b the test MAE in dependence of the prediction horizon is depicted. In total 540 test samples from 10 different subjects were used for the evaluation. The average test error could be reduced by from 0.0388 to 0.0284 by encompassing transfer learning. Accordingly, the model performance on the small UR dataset, containing 10 sessions a 7.3 min becomes comparable to the performance on the large HCP dataset with 100 sessions a 14.4 min with a . Finally, to evaluate the significance across subjects, the test MAE with and without pre-training the model was computed for each of the 10 subjects. A paired t-test was applied and the difference was significant with .
Discussion
We introduced a multi-modal framework for inferring causal relations in brain networks, based on a graph neural network architecture, uniting structural and functional information observed with DTI and fMRI. First this model provides a data-driven perspective on a fundamental question in neuroscience, namely how the function of the brain is related to its structure. Moreover, by modeling dynamic interactions on the structural anatomical substrate, this framework accounts for non-linear spatio-temporal dependencies between segregated brain regions, allowing us to reconstruct a multi-modal measure of causal influence strength.
First, we evaluated the performance of the DCRNN by studying its capabilities to reproduce empirically observed neural activity patterns, and compared it to a VAR model, like that typically used for the analysis of brain connectivity with Granger causality21,46. We showed that the DCRNN can also capture temporal long-term dependencies in fMRI data, enabling it to make accurate predictions up to ( s) in the 0.04–0.07 Hz frequency range, which could reduced the overall test MAE considerably in comparison to a linear VAR. Note that results in subsection ‘Model performance’ demonstrate, that despite its simplicity, a VAR can make quite reliable predictions within the first . Also its linearity allows for various possibilities for statistical inference of causal relations between different time courses, making it a feasible and fast tool for the estimation of Granger-causal connectivity24. But in the future it could be of interest to also consider non-linear and long-term relationships in neuroimaging data, in order to get a more complete picture of functional interactions between areas in the brain. The improved accuracy of the DCRNN reveals that it can better learn inherent characteristics of brain dynamics, and might therefore be more able to characterize causal relations than simple linear models. We further reproduced the analysis on a different dataset in Supplementary III, which could reveal that especially on small datasets it is beneficial to model transitions in brain networks with localized graph filters18. With this technique the predictions of the DCRNN remain stable also on large brain networks, even when only sparse data are available to fit this model to complex network structures. We additionally verified the results by employing a more liberal bandpass filter with cutoff frequencies 0.02–0.09 Hz in Supplementary IV. By including more frequency components, the BOLD signal becomes more complex and is accordingly harder to predict. The same analysis has been carried out relying on a volumetric brain atlas49, and using an alternative tractography method to reconstruct the structural connectivity in Supplementary V. In all cases the difference between the VAR and DCRNN in the prediction performance is apparent, especially for larger horizons. Also the DCRNN does not require stationarity of time series data, therefore avoiding potentially distorting pre-pocessing steps in order to achieve the latter. Another aspect that improves the plausibility of the estimated causal relations between brain regions is the integration of structural information into the graph neural network model. As the propagation of neural signals is physically constrained by the layout of white matter connections, propagating information via graph convolutions along anatomically connected regions is in agreement with prior knowledge about the anatomy of the brain.
The impact of this structural modeling was further investigated in the subsection ‘Impact of spatial modeling’. In the DCRNN the propagation of information is realized as a stationary diffusion process in the notion of a diffusion convolutions (DCs)19. Results show that diffusion steps of order already contribute most to the improvement of prediction accuracy, while higher order terms of only have a nominal further impact on the performance. The influence of structural modeling on the predictive performance provides additional insight into the general structure–function relation in the brain, by pointing out, how much additional information about the functional activity in a certain region can be gained from the inclusion of structurally connected regions. By including filters up to order , the predictions could be improved by in comparison to when information from anatomically connected regions has been neglected. Note that for each time step t the DCRNN already applies multiple DCs to the multi-variate time series data, thereby inherently capturing the influence of higher order structural connections. Therefore low orders of diffusion walks seem to be already sufficient to account for indirect transitions. A good trade-off between computational load and model accuracy could be achieved with a maximum walk order of , as the computational complexity linearly increases with K. Learning localized filters characterized by polynomial coefficients renders it possible to efficiently analyze large scale networks18, which allowed us to conduct an analysis with regions simultaneously on a single GPU. So unlike classical DCM, this model can also be applied to study interactions across the whole brain, making it suitable for an exploratory analysis.
The results demonstrated that propagating information across anatomical connections improves the model accuracy, pointing towards functional dependencies between different brain regions. In the spirit of explainable artificial intelligence (XAI), we proposed a method to reconstruct such dependencies, which the DCRNN has learned from the data in subsection ‘Causal connectivity’. Inducing perturbations in the model’s input space allowed us to study how the activity in a certain region influences other regions. This influence would quantify the importance of temporal information on the activity in a certain ROI for predicting the activity in other ROIs. Following the philosophy of Granger causality, this indicates a causal dependency between ROIs, thereby providing a measure of directed influence among each other. This kind of relation is referred to as directed functional connectivity or causal connectivity, as such information theoretic measures are dependent on causal mechanisms, but are not necessarily identical with them67,68, which distinguishes them from explicit model-based approaches like DCM for effective brain connectivity10. For our approach we used the more general notion of causal connectivity, as we do not only incorporate functional data, but also structural information to describe such causal dependencies between different regions. To demonstrate an application of our proposed approach, we evaluated the influence of PIVC on other brain regions. The derived connectivity network indicated a causal relationship between PIVC with brain regions in the Sylvian fissure, the perisylvian cortex and the insula, but also with the visual cortex.
In a final step, we proposed an approach to improve the model performance on smaller data sets. We demonstrated that the concept of transfer learning32 finds also an application in our context of detecting intrinsic patterns in fMRI time-series and structural connectivity data. Features learned from the data of the HCP repository33 could be well transferred to a smaller dataset, acquired with a Siemens Magnetom Prisma 3T. This made it possible to achieve almost the same accuracy on a small dataset with 10 sessions (each 7.3 min in duration) as with a large dataset of 100 sessions (each 14.4 min in duration). The acquisition and preprocessing protocols of the two fMRI datasets were relatively comparable in our study, so in other cases with larger differences in the temporal resolutions of the data, downsampling one dataset could be necessary in order for the model to better learn transferable feature representations.
Note that by integrating the structural information into the model, the functional interactions learned by this model also depend on the predefined anatomical layout. Therefore the quality of DTI data is additionally relevant for the results, but it is known that DTI has problems to accurately reconstruct long-range white matter tracks69. In our study we incorporated data from young healthy subjects and we computed a group SC matrix for the whole subject cohort, in order to model neural transitions on the anatomical substrate. Such transitions were characterized by local graph filters, which are optimized specifically for the underlying structural layout used in the model. In case of subject cohorts with very different SC profiles, like in studies including diseased patients or when comparing younger and older subjects, such graph filters would not generalize across cohorts and therefore have to be learned for the SC of every group individually. Also fMRI comes with its limitations for studying neural interactions, as the sampling rate is considerably lower than the underlying neural responses, and the neural activity is only indirectly measured based on the observed hemodynamic response46. An interdependence between the temporal information of two brain areas therefore only provides an indication for a causal relationship, and it should not necessarily be assumed to be identical with the latter. So the interpretation of the results should consider the informative content of the neuroimaging data used in this model. In our study we investigated possible applications of the GNN model using two different MRI data sets and different approaches for white matter tractography and frequency filtering (as outlined in the Supplementary Information). But for future studies, alternative imaging modalities and preprocessing schemes, could also be interesting for studying the structure–function relationship under a different light, for example by employing a log-transformed SC matrix to modify the influence of long-range structural connections, or observing functional dynamics in higher temporal resolution with electroencephalography (EEG) or magnetoencephalography (MEG).
In conclusion we think that GNN architectures can provide an interesting novel approach to combine complex non-linear temporal and spatial patterns as observed in fMRI and DTI data. Currently GNNs already show very promising applications for classification tasks in MRI based on brain connectivity networks70–73. In our study we showed that they can be also suitable to characterize the non-Euclidean spatial relationship of segregated brain regions when analyzing dynamic functional interactions on the structural network. Beyond the investigation of causal relations, this data-driven approach to brain dynamics could also be of interest for other applications. While many current approaches dealing with the structure–function relationship in the human brain focus on inferring the overall functional coherence patterns from their SC6,26,28,29,31,74, this framework allows us to directly relate temporally resolved activity profiles to their anatomical substrate. Further this whole-brain model could be interesting for clinical research, by studying dynamics in the diseased brain and observing how functional interactions between different areas might be affected. This multi-modal brain model could also be used to simulate the impact of a structural lesion to investigate the effects on the brain functions75. For detecting functional dependencies among different brain areas, we studied the signal changes in all other areas of the network, caused by a perturbation in a certain target area of interest. Alternative ways to look at such dependencies among the models input variables could be provided by approaches, like sensitivity analysis76 or layer-wise relevance propagation77, what might be of interest for future investigations in this area.
Methods
DCRNN
In the context of neuroimaging, neural activity patterns can be interpreted as a graph structured spatio-temporal signal distribution. The nodes in this graph represent ROIs in a human brain, while the edges reflect the connection strengths between these ROIs in the anatomical neuronal network, which forms a structural scaffold for the flow of information. This connection strength is given by the axonal connection strength as determined from DTI measurements. The activity dynamics on such networks can be modeled by a random walk on a graph, where a diffusion convolution operation is invoked to capture the spatial dependencies18,19. A diffusion–convolution recurrent neural network (DCRNN) is designed to integrate diffusion convolution, a sequence-to-sequence architecture and a scheduled sampling technique19. The model, as it is applied in the current study, is described in detail below.
When considering voxel time series of brain activity maps, we collect all data into a data matrix with . Given N ROIs, taken from a brain atlas and each represented by a meta-voxel, and considering T time points for each meta-voxel time series, which represents the activation time course of one of the ROIs, then we have
3 |
Note that the columns of the data matrix describe the activation of all ROIs at any given time point , while its rows represent the meta-voxel time course of every ROI .
Now consider a network of ROIs (brain areas, neuron pools) as an undirected graph , where denotes a set of vertices (nodes), represents a set of edges and is a weighted adjacency matrix. The latter represents the structural connectivity of the nodes, i.e. the ROIs on the neuronal network, which are adjacent to each other and connected by an edge. Such undirected graphs can be deduced from diffusion tensor imaging (DTI) data, which also provide the edge weights . The latter reflect the anatomical connection strengths between the vertices. Note that DTI alone cannot determine the direction of information flow, what makes it necessary to incorporate functional imaging data.
The flow of activity observed on is expressed as a time-dependent graph signal . It represents the feature of each ROI, which here is the BOLD signal amplitude. Forecasting the flow of activity on amounts to learning a function h(...) that maps past graph signals to future graph signals:
4 |
Spatial dependencies
Information flow on is considered a stochastic random walk process modeled by
a re-start probability
a state transition matrix
Here we have with and
5 |
where the denote normalized edge strengths. Here state transitions are modeled as a diffusion process on a graph. Note that because the DTI cannot obtain directed graphs, its diffusion matrix is symmetric, i.e. . Thus an eigen-decomposition exists according to
6 |
Further the state transition matrix is proportional to a normalized graph Laplacian
7 |
representing a random walk on the graph. Now consider the set of eigenvectors of the diffusion Laplacian matrix as a set of basis vectors. Then the graph signal can be transformed to the conjugate domain and vice versa, hence we have78
8 |
9 |
Finally invoking the convolution theorem, the graph convolution operator can be defined as
10 |
where denotes a filter parametrized by and denotes the Hadamar product in the conjugate domain. The transformed vector summarizes the filter parameters into a parameter vector in the conjugate frequency domain. If it is replaced by a diagonal feature matrix, i.e. , it represents a convolution kernel. Thus we have for the output signal
11 |
Now expanding the filter kernel into a power series with respect to the eigenvalue matrix of the transition matrix , unfolding the bi-quadratic form into a sum of rank one outer product forms , which can be considered elementary filter kernel, and finally keeping only terms up to order K, we obtain
12 |
Note that this diffusion convolution operation includes the inverse diffusion process, represented by the transpose state transition matrix as well, since DTI can only yield undirected graphs. Thus, as has been shown by19, diffusion convolution is intimately related to spectral graph convolution (SGC)18. More precisely, GDC is equivalent to SGC up to a similarity transformation19.
Considering a CNN architecture and using the diffusion convolution operation, the output of each of the diffusion convolution layers (DCL) is then given as follows:
13 |
Hereby denotes the input at time t, the corresponding output of the qth convolution layer, Q the number of filters employed, any suitable activation function, and parameterizes the q-th convolutional kernel of order k. The DCL learns to represent graph structured data and can be trained with gradient descent based optimization techniques.
Note that this random walk on a graph represents a Markov process. At the limit it converges to a stationary distribution , which for finite can be approximated by79
14 |
The i-th row of this matrix represents the likelihood of diffusion starting from ROI , hence the proximity of any other ROI with respect to ROI .
Temporal dependencies
Given the graph convolution operation, temporal dynamics on the graph can be modeled using gated recurrent units (GRU)36. The trick is to replace the matrix multiplications in GRU by diffusion convolutions , as derived in Eq. (12). This leads to the diffusion convolutional gated recurrent unit (DCGRU)19
15 |
16 |
17 |
18 |
where denote the input and output states of the GRU at time t and denotes their concatenation. Also represent reset and update gates at time t, and , respectively, denote bias terms. Furthermore, denote the parameter sets of the corresponding filters. An illustration of a single DCGRU cell is provided in Fig. 6.
Similar to GRUs, also DCGRUs can be employed to build layers of recurrent neural networks, which can be trained by backpropagation through time (BPTT)80,81. If multiple step ahead forecasting is intended, a sequence-to-sequence architecture can be used. In this architecture, both the encoder and the decoder are composed of DCGRU layers forming a diffusion convolution recurrent neural net (DCRNN) (see Fig. 1). During training, a time series of past events is fed into the encoder and its final states form the input to the decoder. The latter then generates predictions, which can be compared to available ground truth observations. For later testing, such ground truth observations are replaced by predictions generated by the model itself. Given BOLD signal voxel activation time series, segments of an observed voxel time series are used to train a DCRNN to predict future activations.
Training the DCRNN
The network is trained by maximizing the likelihood of generating the target future time series using BPTT learning. Hence, DCRNN can capture spatio-temporal dependencies between time series. After the Bandpass filtering of the BOLD signal in each region, the data of each fMRI session was scaled between 0 and 1 before starting the training. The DCRNN19 was implemented using the TensorFlow82 library for machine learning, and computations were performed on an Nvidia Quadro K6000 GPU, running on a desktop PC with an Intel(R) Xeon(R) CPU E5-1620 v4 CPU under Linux Debian 9. Scheduled sampling37 is invoked during training to account for the fact that the distribution of input stimuli during testing might differ from the distribution of training stimuli. During scheduled sampling reference observations are fed to the model with probability , while predictions released by the model are fed in with probability at the i-th iteration. During supervised training, instances to be predicted are, of course, known. An inverse sigmoidal function determines the sampling probability decay:
19 |
It was found to be sufficient to train the model for 70 epochs, and the scheduled sampling parameter can be set to . As an objective function the mean absolute error (MAE) was used to describe the overall difference between true activity and predicted activity :
20 |
For this optimization problem, the ADAM algorithm83 was employed. The samples in the training data set were randomly permuted and the gradient was derived from mini-batches of 32 samples. To achieve a good convergence and to avoid too strongly growing gradients, it was found useful to use an annealing learning rate, initialized with . The learning rate was decreased by a factor of 0.1 at epochs 20, 40 and 60, or if the validation error did not improve for more than 10 epochs. Before lowering the learning rate, the weights with lowest validation error were restored, in order to avoid getting stuck in local optima. The encoder and decoder of the sequence-to-sequence architecture consist to two diffusion convolution GRU layers each, and the hidden state size is set to . The training performance is illustrated in Fig. 7. Due to the curriculum learning strategy, within the first few epochs the probability of the decoder receiving a true label is very high, like that illustrated by the gray line in Fig. 7. Therefore at onset of the training, the model has to make only correct short term predictions and the training MAE is already relatively small, indicated by the solid dark blue line. For validation and testing, the decoder always receives its own previous prediction. The validation MAE is illustrated with the dashed light blue line, and it can be seen that the model gradually learns to make also accurate long term predictions. Finally, for the evaluations the whole population dataset was scaled to unit variance and zero mean.
Autoregressive models
As Granger causality21 is typically based on linear vector autoregressive (VAR) models for stochastic time series data, we evaluated a VAR as one baseline method. The idea of an autoregressive process (AR) is that a time series x(t) can be described by a linear function of the first of its lagged values84
21 |
with coefficients , intercept and an error term u(t). This expression can be extended to a multivariate VAR model with N time series as84
22 |
where coefficients are stored in matrices , and intercepts and errors are described by vectors and . In the context of this study, time series reflect the BOLD signal of N brain regions, measured at different times t.
For the estimation of coefficients and intercepts various methods exist24, and in this study we rely on two different strategies. The first is based on a typical ordinary least squares (OLS) estimation24,47 on individual subject sessions, implemented in the statsmodel python package85. The first of each fMRI session were used to fit the model to the data, and the subsequent were used for validation and the last were employed as a test set. To check for stationarity of the analyzed time series an augmented Dickey–Fuller test for unit roots was performed47,86, with a p-value of .
Additionally, in order to render the comparison to the DCRNN more accurate, we implemented a gradient descend based optimization for a VAR model in TensorFlow82, to verify that the differences in predictive performance can be related to the models, and not solely to the optimization strategies. In analogy to the DCRNN training, input-output samples of neural activities were generated from the data like described in subsection ‘Data description’, which were used to minimize the MAE between the model’s prediction and groundtruths . The convergence could be optimized by employing stochastic gradient descent (SGD) optimization with a batch size of 1, using an annealing learning rate with a start value of . The VAR model was trained for 100 epochs, and the learning rate was reduced by a factor of 0.1 after epoch 70 and 90. A comparison of the error on the test set between the two different optimization strategies can be found in supplement VI.
Best performance could be achieved employing a SGD based optimization in combination with a lag order of . Note that with such a high lag order, around of the time courses do not fulfill the stationarity criteria of the augmented Dickey–Fuller test anymore (). Yet the prediction accuracy could still be improved by including lags up to , like shown in supplement VI. As the objective criterion of the evaluation was to assess the capabilities of replicating empirically observed neural activity patterns, we chose the VAR model with best accuracy for comparison with the DCRNN in ‘Model performance’.
Datasets
HCP data
The first data set used in this study is provided by the HCP data repository33,87. The S1200 release includes data from subjects which participated in four resting state fMRI sessions, lasting 14.4 min each and collecting 1200 volumes per session. Customized Siemens Connectome Skyra magnetic resonance imaging scanners with a field strength of T were employed for data acquisition, using multi-band (factor 8) acceleration88–91. The data was collected by gradient-echo echo-planar imaging (EPI) sequences with a repetition time ms and an echo time ms. The field of view was and slices with a thickness of mm were obtained, containing voxels with a size of . The preprocessed version, including motion-correction, structural preprocessing and ICA-FIX denoising was chosen38,92–97. Next a multi-model parcellation scheme was applied to divide the cortical gray matter hemisphere into 180 regions39, and the BOLD signal inside each region was averaged, to obtain the temporal activity evolution for each area. For our study we found it appropriate to apply global signal regression, firstly because it showed to effectively reduce movement artifacts in HCP datasets98. Also in this study of causal relations, the goal was to extract the additional information, which certain brain regions contain about the activity of other regions, whereby local interactions rather than global modulations were of interest for us. Those time courses were bandpass filtered, first performing the evaluations on a noise reduced narrowband in ‘Model performance’, employing a filter with cutoff frequencies 0.04–0.07 Hz40–43, and additionally implementing a more liberal bandpass filter with cutoff frequencies 0.02–0.09 Hz, as displayed in Supplementary IV.
Diffusion MRI data was collected in 6 runs, whereby approximately 90 directions were sampled during each run, employing three shells of and , including 6 images99. A Spin-echo EPI sequence was employed with repetition time , echo time , and multi band factor 3. In total slices were obtained, with field of view and an isotropic voxel size of . The preprocessing included intensity normalization across runs, EPI distortion correction, eddy-current corrections, removing motion artifacts, and gradient non-linearity corrections38,100–103. To obtain the structural connectivity strengths between regions defined by Glasser et al.39, the MRtrix3 software package was employed35. Briefly multi-shell multi-tissue constrained spherical deconvolution45 was used to obtain the response functions for fiber orientation distribution estimation104,105. Furher 10 million streamlines were created with anatomical constrained tractography106 and spherical-deconvolution informed filtering was applied107, reducing the number of streamlines to 1 million. To quantify the strength of the structural connections, the number of streamlines connecting two brain regions were computed, and normalized by the region volumes. A detailed description of the workflow can be found in: https://osf.io/fkyht/. The group structural connectome was computed as an average across the first 10 subjects, as the variance in the structural connectivity strength is relatively low across subjects108, while probabilistic tractography methods are relatively computationally demanding. For this dataset, including only young healthy subjects, the similarity of the SC across subjects was relatively high, and the correlation coefficient between SC values of different subjects was on average 0.91. But note that when comparing very different subject cohorts, like healthy and diseased subjects, the SC matrix should be computed for every studied group individually.
UR data
The second dataset was acquired with a Siemens Magnetom Prisma with field strength T at the University of Regensburg (UR). The data of 10 different subjects were used, whereby resting state fMRI data were collected during a scanning time of 7.3 min. All subjects provided written informed consent and the study was approved by the local ethics committee of the University of Regensburg. All methods were performed in accordance with the relevant guidelines and regulations. An EPI sequence was employed using multi-band (factor 8) acceleration, sampling 600 volumetric images per run with a repetition time of ms and an echo time of ms. The field of view was and slices with thickness of were collected, containing voxels with a size of . For preprocessing the HCP pipeline (version 4.0.0) was employed, as described by Glasser et al.38. To achieve good correspondence between the two datasets, the further preprocessing was also performed as outlined in subsection ‘HCP data’. The fMRI time courses were averaged within each brain region of the multi-modal parcellation scheme39, and again global signal regression was applied. Finally those time courses were bandpass filtered within the noise reduced range of 0.04–0.07 Hz.
To reconstruct the anatomical connectivity, diffusion MRI data was collected in 4 runs, sampling approximately 90 directions, employing two shells with and , and also including 7 images. The repetition time of the Spin-echo EPI sequence was ms with an echo time ms, employing a multi-band (factor 4) acceleration. Overall slices were collected, with a field of view , containing voxels with a size of . Preprocessing of the diffusion MRI data was based on the HCP guidelines38, and finally the anatomical connectivity matrices were obtained like in the previous subsection ‘HCP data’, using constrained spherical deconvolution as provided in the MRtrix package35. The group structural connectivity was computed as an average over the 10 subjects.
Supplementary Information
Acknowledgements
The work was supported by the DFG projects GR988/25-1 and ISNT89/393-1 granted to M.W.G. and WU392/9-1 granted to S.Wü., the Harris Foundation of Dartmouth College granted to M.W.G., and the Hanns Seidel Foundation granted to G.-I.H. Data were provided in part by the Human Connectome Project, WU-Minn Consortium (Principal Investigators: David Van Essen and Kamil Ugurbil; 1U54MH091657) funded by the 16 NIH Institutes and Centers that support the NIH Blueprint for Neuroscience Research; and by the McDonnell Center for Systems Neuroscience at Washington University.
Author contributions
Research was conducted by S.W., W.M.M. and E.W.L., with feedback from M.W.G. and A.M.T. Data was collected by G.-I.H. under supervision of S.Wü. The manuscript was written by S.W., with feedback from E.W.L., S.M.F., M.W.G. and A.M.T. The study was supervised by E.W.L. and M.W.G. All authors revised the manuscript.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
The online version contains supplementary material available at 10.1038/s41598-021-87411-8.
References
- 1.Lang E, Tomé A, Keck I, Gorriz J, Puntonet C. Brain connectivity analysis: A short survey. Comput. Intell. Neurosci. 2012 doi: 10.1155/2012/412512. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Bartzokis G, et al. White matter structural integrity in healthy aging adults and patients with Alzheimer disease: A magnetic resonance imaging study. Arch. Neurol. 2003;60:393–8. doi: 10.1001/archneur.60.3.393. [DOI] [PubMed] [Google Scholar]
- 3.Hagmann P, et al. White matter maturation reshapes structural connectivity in the late developing human brain. Proc. Natl. Acad. Sci. 2010;107:19067–19072. doi: 10.1073/pnas.1009073107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Deco G, Senden M, Jirsa V. How anatomy shapes dynamics: A semi-analytical study of the brain at rest by a simple spin model. Front. Comput. Neurosci. 2012;6:68. doi: 10.3389/fncom.2012.00068. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Hermundstad AM, et al. Structural foundations of resting-state and task-based functional connectivity in the human brain. Proc. Natl. Acad. Sci. U. S. A. 2013;110(15):6169–74. doi: 10.1073/pnas.1219562110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Messé, A., Rudrauf, D., Benali, H. & Marrelec, G. Relating structure and function in the human brain: Relative contributions of anatomy, stationary dynamics, and non-stationarities. PLoS Comput. Biol.10, (2014). [DOI] [PMC free article] [PubMed]
- 7.Abdelnour F, Voss HU, Raj A. Network diffusion accurately models the relationship between structural and functional brain connectivity networks. NeuroImage. 2014;90:335–347. doi: 10.1016/j.neuroimage.2013.12.039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Bettinardi RG, et al. How structure sculpts function: Unveiling the contribution of anatomical connectivity to the brain’s spontaneous correlation structure. Chaos Interdiscip. J. Nonlinear Sci. 2018;27:047409. doi: 10.1063/1.4980099. [DOI] [PubMed] [Google Scholar]
- 9.Honey CJ, et al. Predicting human resting-state functional connectivity from structural connectivity. Proc. Natl. Acad. Sci. U. S. A. 2009;106(6):2035–40. doi: 10.1073/pnas.0811168106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Friston K, Moran R, Seth AK. Analysing connectivity with granger causality and dynamic causal modelling. Curr. Opin. Neurobiol. 2013;23:172–178. doi: 10.1016/j.conb.2012.11.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Amico E, Goni J. Mapping hybrid functional–structural connectivity traits in the human connectome. Netw. Neurosci. 2018;2:306–322. doi: 10.1162/netn_a_00049. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Xue W, Bowman F, Pileggi A, Mayer A. A multimodal approach for determining brain networks by jointly modeling functional and structural connectivity. Front. Comput. Neurosci. 2015;9:22. doi: 10.3389/fncom.2015.00022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Chu S, Parhi K, Lenglet C. Function-specific and enhanced brain structural connectivity mapping via joint modeling of diffusion and functional MRI. Sci. Rep. 2018;8:1–9. doi: 10.1038/s41598-018-23051-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Wu, Z. et al. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 1–21 (2020). [DOI] [PubMed]
- 15.Fukushima K. A neural network model for the mechanism of selective attention in visual pattern recognition. Syst. Comput. Jpn. 1987;18:102–113. doi: 10.1002/scj.4690180110. [DOI] [Google Scholar]
- 16.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521:436–44. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
- 17.Bruna, J., Zaremba, W., Szlam, A. & Lecun, Y. Spectral networks and locally connected networks on graphs. In International conference on learning representations (ICLR2014) (CBLS, 2014).
- 18.Defferrard, M., Bresson, X. & Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. NIPS. 3837–3845 (2016).
- 19.Li, Y., Yu, R., Shahabi, C., & Liu, Y (Data-driven traffic forecasting, Diffusion convolutional recurrent neural network, 2018).
- 20.Deco G, Kringelbach ML, Jirsa VK, Ritter P. The dynamics of resting fluctuations in the brain: Metastability and its dynamical cortical core. Sci. Rep. 2017;7:1–14. doi: 10.1038/s41598-017-03073-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Granger CWJ. Investigating causal relations by econometric models and cross-spectral methods. Econometrica. 1969;37:424–438. doi: 10.2307/1912791. [DOI] [Google Scholar]
- 22.Friston KJ, Harrison L, Penny W. Dynamic causal modelling. Neuroimage. 2003;19:1273–1302. doi: 10.1016/S1053-8119(03)00202-7. [DOI] [PubMed] [Google Scholar]
- 23.Daunizeau J, David O, Stephan K. Dynamic causal modelling: A critical review of the biophysical and statistical foundations. NeuroImage. 2009;58:312–22. doi: 10.1016/j.neuroimage.2009.11.062. [DOI] [PubMed] [Google Scholar]
- 24.Barnett L, Seth A. The MVGC multivariate granger causality toolbox: A new approach to granger-causal inference. J. Neurosci. Methods. 2013;223:50–68. doi: 10.1016/j.jneumeth.2013.10.018. [DOI] [PubMed] [Google Scholar]
- 25.Chen X, Wang Y. Predicting resting-state functional connectivity with efficient structural connectivity. EEE/CAA J. Autom. Sin. 2018;5:1079–1088. doi: 10.1109/JAS.2017.7510880. [DOI] [Google Scholar]
- 26.Liang H, Wang H. Structure–function network mapping and its assessment via persistent homology. PLoS Comput. Biol. 2017;13:e1005325. doi: 10.1371/journal.pcbi.1005325. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Abdelnour F, Dayan M, Devinsky O, Thesen T, Raj A. Functional brain connectivity is predictable from anatomic network’s laplacian eigen-structure. NeuroImage. 2018;172:728–739. doi: 10.1016/j.neuroimage.2018.02.016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Becker C, et al. Spectral mapping of brain functional connectivity from diffusion imaging. Sci. Rep. 2018 doi: 10.1038/s41598-017-18769-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Deligianni F, Carmichael D, Zhang HG, Clark C, Clayden J. Noddi and tensor-based microstructural indices as predictors of functional connectivity. PLoS One. 2016;11:e0153404. doi: 10.1371/journal.pone.0153404. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Rosenthal G, et al. Mapping higher-order relations between brain structure and function with embedded vector representations of connectomes. Nat. Commun. 2018;9:1–12. doi: 10.1038/s41467-018-04614-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Surampudi SG, et al. Multiple kernel learning model for relating structural and functional connectivity in the brain. Sci. Rep. 2018;8:1–14. doi: 10.1038/s41598-018-21456-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Pan S, Yang Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2010;22:1345–1359. doi: 10.1109/TKDE.2009.191. [DOI] [Google Scholar]
- 33.Van Essen D, et al. The WU-Minn human connectome project: An overview. NeuroImage. 2013;80:62–79. doi: 10.1016/j.neuroimage.2013.05.041. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Sutskever, I., Vinyals, O. & Le, Q. V. Sequence to sequence learning with neural networks. CoRRabs/1409.3215 (2014).
- 35.Tournier J-D, et al. MRtrix3: A fast, flexible and open software framework for medical image processing and visualisation. NeuroImage. 2019 doi: 10.1101/551739. [DOI] [PubMed] [Google Scholar]
- 36.Chung, J., Gulcehre, C., Cho, K. & Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling (2014).
- 37.Bengio, S., Vinyals, O., Jaitly, N. & Shazeer, N. Scheduled sampling for sequence prediction with recurrent neural networks. NIPS. 171–1179 (2015).
- 38.Glasser M, et al. The minimal preprocessing pipelines for the human connectome project. NeuroImage. 2013;80:105–124. doi: 10.1016/j.neuroimage.2013.04.127. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Glasser M, et al. A multi-modal parcellation of human cerebral cortex. Nature. 2016;536:171–178. doi: 10.1038/nature18933. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Glerean E, Salmi J, Lahnakoski J, Jääskeläinen I, Sams M. Functional magnetic resonance imaging phase synchronization as a measure of dynamic functional connectivity. Brain Connect. 2012;2:91–101. doi: 10.1089/brain.2011.0068. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Buckner R, et al. Cortical hubs revealed by intrinsic functional connectivity: Mapping, assessment of stability, and relation to alzheimer’s disease. J. Neurosci. 2009;29:1860–73. doi: 10.1523/JNEUROSCI.5062-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Biswal BB, Yetkin FZ, Haughton V, Hyde JS. Functional connectivity in the motor cortex of resting human brain using echo-planar MRI. Magn. Reson. Med. 1995;34(4):537–41. doi: 10.1002/mrm.1910340409. [DOI] [PubMed] [Google Scholar]
- 43.Achard S, Salvador R, Whitcher B, Suckling J, Bullmore E. A resilient, low-frequency, small-world human brain functional network with highly connected association cortical hubs. J. Neurosci. 2006;3:e17. doi: 10.1523/JNEUROSCI.3874-05.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Uǧurbil K, et al. Pushing spatial and temporal resolution for functional and diffusion MRI in the human connectome project. NeuroImage. 2013;80:80–104. doi: 10.1016/j.neuroimage.2013.05.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Jeurissen B, Tournier J-D, Dhollander T, Connelly A, Sijbers J. Multi-tissue constrained spherical deconvolution for improved analysis of multi-shell diffusion MRI data. NeuroImage. 2014;103:411–426. doi: 10.1016/j.neuroimage.2014.07.061. [DOI] [PubMed] [Google Scholar]
- 46.Seth A, Barrett A, Barnett L. Granger causality analysis in neuroscience and neuroimaging. J. Neurosci. 2015;35:3293–7. doi: 10.1523/JNEUROSCI.4399-14.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Hamilton J. Time Series Analysis. Princeton***: Princeton University Press; 1994. [Google Scholar]
- 48.Kiefer J, Wolfowitz J. Stochastic estimation of the maximum of a regression function. Ann. Math. Stat. 1952;23:462–466. doi: 10.1214/aoms/1177729392. [DOI] [Google Scholar]
- 49.Tzourio-Mazoyer N, et al. Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. NeuroImage. 2002;15:273–89. doi: 10.1006/nimg.2001.0978. [DOI] [PubMed] [Google Scholar]
- 50.Behrens T, Berg H, Jbabdi S, Rushworth M, Woolrich M. Probabilistic diffusion tractography with multiple fibre orientations: What can we gain? NeuroImage. 2007;34:144–55. doi: 10.1016/j.neuroimage.2006.09.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Zeiler, M. & Fergus, R. Visualizing and understanding convolutional neural networks. ECCV 2014, Part I, LNCS 8689. 10.1007/978-3-319-10590-1_53 (2013).
- 52.Samek W, Binder A, Montavon G, Lapuschkin S, Müller K-R. Evaluating the visualization of what a deep neural network has learned. IEEE Trans. Neural Netw. Learn. Syst. 2017;28:2660–2673. doi: 10.1109/TNNLS.2016.2599820. [DOI] [PubMed] [Google Scholar]
- 53.Lopez C, Blanke O, Mast F. The human vestibular cortex revealed by coordinate-based activation likelihood estimation meta-analysis. Neuroscience. 2012;212:159–79. doi: 10.1016/j.neuroscience.2012.03.028. [DOI] [PubMed] [Google Scholar]
- 54.Frank S, Greenlee M. The parieto-insular vestibular cortex in humans: More than a single area? J. Neurophysiol. 2018;120:1438–1450. doi: 10.1152/jn.00907.2017. [DOI] [PubMed] [Google Scholar]
- 55.Wirth A, Frank S, Greenlee M, Beer A. White matter connectivity of the visual-vestibular cortex examined by diffusion-weighted imaging. Brain Connect. 2018;8:235–244. doi: 10.1089/brain.2017.0544. [DOI] [PubMed] [Google Scholar]
- 56.Indovina I, et al. Structural connectome and connectivity lateralization of the multimodal vestibular cortical network. NeuroImage. 2020;222:117247. doi: 10.1016/j.neuroimage.2020.117247. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Frank SM, et al. Attention networks in the parietooccipital cortex modulate activity of the human vestibular cortex during attentive visual processing. J. Neurosci. 2020;40:1110–1119. doi: 10.1523/JNEUROSCI.1952-19.2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Guldin WO. Is there a vestibular cortex? Trends Neurosci. 1998;21:254–9. doi: 10.1016/s0166-2236(97)01211-3. [DOI] [PubMed] [Google Scholar]
- 59.Frank S, Baumann O, Mattingley J, Greenlee M. Vestibular and visual responses in human posterior insular cortex. J. Neurophysiol. 2014;112:2481–2491. doi: 10.1152/jn.00078.2014. [DOI] [PubMed] [Google Scholar]
- 60.Frank S, Wirth A, Greenlee M. Visual-vestibular processing in the human sylvian fissure. J. Neurophysiol. 2016;116:263–271. doi: 10.1152/jn.00009.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Chen A, Deangelis G, Angelaki D. Convergence of vestibular and visual self-motion signals in an area of the posterior sylvian fissure. J. Neurosci. 2011;31:11617–27. doi: 10.1523/JNEUROSCI.1266-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Wenzel R, et al. Deactivation of human visual cortex during involuntary ocular oscillations—a pet activation study. Brain J. Neurol. 1996;119:101–10. doi: 10.1093/brain/119.1.101. [DOI] [PubMed] [Google Scholar]
- 63.Brandt T, Bartenstein P, Janek A, Dieterich M. Reciprocal inhibitory. Visual-vestibular interaction visual motion stimulation deactivates the parieto-insular vestibular cortex. Brain J. Neurol. 1998;121:1749–58. doi: 10.1093/brain/121.9.1749. [DOI] [PubMed] [Google Scholar]
- 64.Frank SM, Sun L, Forster L, Tse PU, Greenlee MW. Cross-modal attention effects in the vestibular cortex during attentive tracking of moving objects. J. Neurosci. 2016;36:12720–12728. doi: 10.1523/JNEUROSCI.2480-16.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Frank S, et al. Visual attention modulates glutamate-glutamine levels in vestibular cortex: Evidence from magnetic resonance spectroscopy. J. Neurosci. 2021 doi: 10.1523/JNEUROSCI.2018-20.2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Glorot X, Bengio Y. Understanding the difficulty of training deep feedforward neural networks. J. Mach. Learn. Res. Proc. Track. 2010;9:249–256. [Google Scholar]
- 67.Seth A, Chorley P, Barnett L. Granger causality analysis of FMRI bold signals is invariant to hemodynamic convolution but not downsampling. NeuroImage. 2012;65:540–55. doi: 10.1016/j.neuroimage.2012.09.049. [DOI] [PubMed] [Google Scholar]
- 68.Bressler S, Seth A. Wiener-granger causality: A well established methodology. NeuroImage. 2010;58:323–9. doi: 10.1016/j.neuroimage.2010.02.059. [DOI] [PubMed] [Google Scholar]
- 69.Thomas C, et al. Anatomical accuracy of brain connections derived from diffusion MRI tractography is inherently limited. Proc. Natl. Acad. Sci. U. S. A. 2014;111(46):16574–9. doi: 10.1073/pnas.1405672111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Ktena SI, et al. Metric learning with spectral graph convolutions on brain connectivity networks. NeuroImage. 2018;169:431–442. doi: 10.1016/j.neuroimage.2017.12.052. [DOI] [PubMed] [Google Scholar]
- 71.Arslan, S., Ktena, S. I., Glocker, B. & Rueckert, D. Graph saliency maps through spectral convolutional networks: Application to sex classification with brain connectivity. In GRAIL/Beyond-MIC@MICCAI (2018).
- 72.Li, X. et al.Graph neural network for interpreting task-fMRI biomarkers. 10.1007/978-3-030-32254-0_54 (2019). [DOI] [PMC free article] [PubMed]
- 73.Kim B-H, Ye JC. Understanding graph isomorphism network for RS-fMRI functional connectivity analysis. Front. Neurosci. 2020;14:630. doi: 10.3389/fnins.2020.00630. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Luisa Saggio M, Ritter P, Jirsa KV. Analytical operations relate structural and functional connectivity in the brain. PLoS One. 2016;11:e0157292. doi: 10.1371/journal.pone.0157292. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Alstott J, Breakspear M, Hagmann P, Cammoun L, Sporns O. Modeling the impact of lesions in the human brain. PLoS Comput. Biol. 2009;5:e1000408. doi: 10.1371/journal.pcbi.1000408. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Simonyan, K., Vedaldi, A. & Zisserman, A. Deep inside convolutional networks: Visualising image classification models and saliency maps. In 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Workshop Track Proceedings (2014).
- 77.Bach S, et al. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS One. 2015;10:1–46. doi: 10.1371/journal.pone.0130140. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78.Shuman DI, Narang SK, Frossard P, Ortega A, Vandergheynst P. The emerging field of signal processing on graphs: Ex-tending high-dimensional data analysis to networks and other irregular domains. IEEE Signal Process. Mag. 2013;30:83–98. doi: 10.1109/MSP.2012.2235192. [DOI] [Google Scholar]
- 79.Teng S. Scalable algorithms for data and network analysis. Found. Trends Theoret. Comput. Sci. 2016;12:1–274. doi: 10.1561/0400000051. [DOI] [Google Scholar]
- 80.Werbos P. Backpropagation through time: What it does and how to do it. Proc. IEEE. 1990;78:1550–1560. doi: 10.1109/5.58337. [DOI] [Google Scholar]
- 81.Lillicrap TP, Santoro A. Backpropagation through time and the brain. Curr. Opin. Neurobiol. 2019;55:82–89. doi: 10.1016/j.conb.2019.01.011. [DOI] [PubMed] [Google Scholar]
- 82.Abadi, M. et al. Tensorflow: A system for large-scale machine learning. In 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), 265–283 (2016).
- 83.Kingma, D. & Ba, J. Adam: A method for stochastic optimization (2014).
- 84.Luetkepohl H. The New Introduction to Multiple Time Series Analysis. New York***: Springer; 2005. [Google Scholar]
- 85.Seabold, S. & Perktold, J. Statsmodels: Econometric and statistical modeling with python (2010).
- 86.Mackinnon J. Approximate asymptotic distribution functions for unit-root and cointegration tests. J. Bus. Econ. Stat. 1994;12:167–76. doi: 10.1080/07350015.1994.10510005. [DOI] [Google Scholar]
- 87.Hodge M, et al. ConnectomeDB—Sharing human brain connectivity data. NeuroImage. 2015 doi: 10.1016/j.neuroimage.2015.04.046. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Moeller S, et al. Multiband multislice GE-epi at 7 tesla, with 16-fold acceleration using partial parallel imaging with application to high spatial and temporal whole-brain fMRI. Magn. Reson. Med. 2010;63(5):1144–53. doi: 10.1002/mrm.22361. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89.Feinberg D, et al. Multiplexed echo planar imaging for sub-second whole brain fMRI and fast diffusion imaging. PLoS One. 2010;5:e15710. doi: 10.1371/journal.pone.0015710. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90.Setsompop K, et al. Blipped-controlled aliasing in parallel imaging for simultaneous multislice echo planar imaging with reduced g-factor penalty. Magn. Reson. Med. 2012;67(5):1210–24. doi: 10.1002/mrm.23097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Xu, J. et al. Highly accelerated whole brain imaging using aligned-blipped-controlled-aliasing multiband EPI. Proceedings of the 20th Annual Meeting of ISMRM 2036 (2012).
- 92.Jenkinson M, Bannister P, Brady M, Smith S. Improved optimization for the robust and accurate linear registration and motion correction of brain images. NeuroImage. 2002;17:825–841. doi: 10.1006/nimg.2002.1132. [DOI] [PubMed] [Google Scholar]
- 93.Jenkinson M, Beckmann CF, Behrens TE, Woolrich MW, Smith SM. FSL. NeuroImage. 2012;62:782–790. doi: 10.1016/j.neuroimage.2011.09.015. [DOI] [PubMed] [Google Scholar]
- 94.Fischl B. Freesurfer. NeuroImage. 2012;62:774–781. doi: 10.1016/j.neuroimage.2012.01.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Smith SM, et al. Resting-state fMRI in the human connectome project. NeuroImage. 2013;80:144–168. doi: 10.1016/j.neuroimage.2013.05.039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Salimi-Khorshidi G, et al. Automatic denoising of functional MRI data: Combining independent component analysis and hierarchical fusion of classifiers. NeuroImage. 2014;90:449–468. doi: 10.1016/j.neuroimage.2013.11.046. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Griffanti L, et al. ICA-based artefact removal and accelerated fMRI acquisition for improved resting state network imaging. NeuroImage. 2014;95:232–247. doi: 10.1016/j.neuroimage.2014.03.034. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Burgess G, et al. Evaluation of denoising strategies to address motion-correlated artifact in resting state fMRI data from the human connectome project. Brain Connect. 2016;6:669–680. doi: 10.1089/brain.2016.0435. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Sotiropoulos S, et al. Advances in diffusion MRI acquisition and processing in the human connectome project. NeuroImage. 2013;80:125. doi: 10.1016/j.neuroimage.2013.05.057. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Sotiropoulos S, et al. Effects of image reconstruction on fibre orientation mapping from multi-channel diffusion MRI: Reducing the noise floor using SENSE. Magn. Reson. Med. 2013 doi: 10.1002/mrm.24623. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Andersson J, Skare S, Ashburner J. How to correct susceptibility distortions in spin-echo echo-planar images: Application to diffusion tensor imaging. NeuroImage. 2003;20:870–88. doi: 10.1016/S1053-8119(03)00336-7. [DOI] [PubMed] [Google Scholar]
- 102.Andersson J, Sotiropoulos S. Non-parametric representation and prediction of single- and multi-shell diffusion-weighted MRI data using gaussian processes. NeuroImage. 2015;122:166–76. doi: 10.1016/j.neuroimage.2015.07.067. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 103.Andersson J, Sotiropoulos S. An integrated approach to correction for off-resonance effects and subject movement in diffusion MR imaging. NeuroImage. 2015;125:1063–1078. doi: 10.1016/j.neuroimage.2015.10.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.Tournier J-D, Calamante F, Gadian D, Connelly A. Direct estimation of the fiber orientation density function from diffusion-weighted MRI data using spherical deconvolution. NeuroImage. 2004;23:1176–85. doi: 10.1016/j.neuroimage.2004.07.037. [DOI] [PubMed] [Google Scholar]
- 105.Tournier J-D, Calamante F, Connelly A. Robust determination of the fibre orientation distribution in diffusion MRI: Non-negativity constrained super-resolved spherical deconvolution. NeuroImage. 2007;35:1459–72. doi: 10.1016/j.neuroimage.2007.02.016. [DOI] [PubMed] [Google Scholar]
- 106.Smith R, Tournier J-D, Calamante F, Connelly A. Anatomically-constrained tractography: Improved diffusion MRI streamlines tractography through effective use of anatomical information. NeuroImage. 2012;62:1924–38. doi: 10.1016/j.neuroimage.2012.06.005. [DOI] [PubMed] [Google Scholar]
- 107.Smith R, Tournier J-D, Calamante F, Connelly A. Sift: Spherical-deconvolution informed filtering of tractograms. NeuroImage. 2013;67:298–312. doi: 10.1016/j.neuroimage.2012.11.049. [DOI] [PubMed] [Google Scholar]
- 108.Zimmermann, J., Griffiths, J., Schirner, M., Ritter, P. & McIntosh, A. R. Subject-specificity of the correlation between large-scale structural and functional connectivity. Netw. Neuroscie. 1–35 (2019). [DOI] [PMC free article] [PubMed]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.