Abstract
The increasing availability of brain imaging technologies has led to intense neuroscientific inquiry into the human brain. Studies often investigate brain function related to emotion, cognition, language, memory, and numerous other externally induced stimuli as well as resting-state brain function. Studies also use brain imaging in an attempt to determine the functional or structural basis for psychiatric or neurological disorders and, with respect to brain function, to further examine the responses of these disorders to treatment. Neuroimaging is a highly interdisciplinary field, and statistics plays a critical role in establishing rigorous methods to extract information and to quantify evidence for formal inferences. Neuroimaging data present numerous challenges for statistical analysis, including the vast amounts of data collected from each individual and the complex temporal and spatial dependence present. We briefly provide background on various types of neuroimaging data and analysis objectives that are commonly targeted in the field. We present a survey of existing methods targeting these objectives and identify particular areas offering opportunities for future statistical contribution.
Keywords: Neuroimaging, fMRI, DTI, connectivity, prediction, activation
1 Introduction
Neuroimaging utilizes powerful noninvasive techniques to capture properties of the human brain in vivo. Imaging studies reveal insights about normal brain function and structure, neural processing and neuroanatomic manifestations of psychiatric and neurological disorders, and neural processing alterations associated with treatment response. There are several widely used imaging modalities, including magnetic resonance imaging (MRI), functional MRI, diffusion tensor imaging (DTI), positron emission tomography (PET), electroencephalography (EEG), and magnetoencephalography (MEG), among others. These modalities leverage different physiological characteristics to reflect properties of either brain structure or function. This review will largely focus on the commonly used fMRI, which captures correlates of neural activity, but some of the ideas presented incorporate or extend to other modalities.
1.1 Imaging Modalities
Functional MRI quantifies brain activity by measuring correlates of blood flow and metabolism. A fundamental concept behind fMRI is that neural activity is associated with localized changes in metabolism. As a brain area becomes more active, for example to perform a memory task, there is an associated localized increase in oxygen consumption. To meet this increased demand, there is an increase in oxygen-rich blood flow to the active brain area. There is a relative increase in oxyhemogloblin and decrease in deoxyhemoglobin in activated brain areas, since the increased supply of oxygen outpaces the increased demand.
The most common form of fMRI works by leveraging magnetic susceptibility properties of hemoglobin in capillary red blood cells, which delivers oxygen to neurons. Hemoglobin is diamagnetic when oxygenated and paramagnetic when deoxygenated. The MR scanner produces blood oxygenated level dependent (BOLD) signals that vary according to the degree of oxygenation. Thus, fMRI can be used to produce distributed maps of localized brain activity (see Figure 1(a)). It is important to keep in mind that although we use the phrase brain activity, what we actually measure with BOLD fMRI is several steps removed from the actual neuronal activity. MRI works in a conceptually similar manner except that the MR signal varies according to tissue type, enabling the production of structural (anatomical) images that distinguish gray matter, white matter, and cerebral spinal fluid (see Figure 1(b)).
DTI is an MRI technique that provides information regarding the white matter structure in the brain. Neurons are the basic unit of the brain, and humans amazingly have an approximate 86 billion neurons (Suzana Herculano-Houzel, 2012), with longstanding estimates reaching up to 100 billion (Society for Neuroscience, 2008). Axons are neuron fibers that serve as lines of transmission in the nervous system, and they form (millions of) bundles of textured fibers in the white-matter of the brain. This extensive system of white-matter bundles directly links some brain structures, with association bundles joining cortical areas within the same hemisphere, commissural bundles linking cortical areas in separate hemispheres, and projection fibers uniting areas in the cerebral cortex to subcortical structures (Hendelman, 2005). DTI non-invasively maps the diffusion of water molecules, which reveals the presence, integrity, and direction of white-matter fibers since water molecules are more likely to diffuse in the direction of white-matter fibers than perpendicular to these fibers. Further, white-matter fiber tracking techniques portray axonal fibers and structural brain connectivity (see Figure 1(c)) (Behrens et al., 2003 and 2007). There is widespread interest in mapping the structural and functional connectivity of the human brain as evidenced by the National Institutes of Health’s (NIH’s) Human Connectome Project (NIH).
1.2 Common Analysis Objectives
Targeting insights about normal brain function and structure, neural manifestations of mental and neurological disorders, and neural plasticity associated with treatment response, statistical analyses often center on objectives targeting localization, brain connectivity, and prediction/classification. Localization is predicated on the theory of functional specialization – that is, the notion that different areas in the brain are specialized for different functions. From properly designed studies, one can perform statistical analyses to identify localized changes in the brain that correspond to changes in tasks performed in the scanner. These analyses produce images highlighting statistically significant (or highly probable) task-related changes in neural activity as illustrated in Figure 2(a). Localization studies, also referred to here as activation (or neuroactivation) studies, can be extended to identify localized differences in brain function between groups of subjects, for example schizophrenia patients and healthy controls, and differences between scanning sessions, which may reflect treatment-related alterations. For example, Figure 2(a) highlights brain regions exhibiting high probabilities of increased inhibitory control related neural activity from a baseline to a post-treatment period for cocaine addicts relative to corresponding changes in control subjects. The brain regions include the middle frontal gyrus and left and right thalamus.
Functional connectivity studies seek to determine multiple brain areas that exhibit similar temporal activity profiles, either task-related or at resting state. These studies may determine links to selected seed brain regions, dissociate particular brain networks, or generate whole brain complex brain networks (see Figure 2(b)). Functional connectivity properties may be compared between subgroups of subjects and between different scanning sessions. Whereas functional connectivity merely targets associations in brain activity between distinct brain regions, effective connectivity seeks to establish a stronger relationship reflecting the influence that one brain region exerts on another.
Prediction or classification analyses stand to have a significant translational impact. For example, one can use baseline imaging and clinical data to generate maps forecasting metabolic activity in the brain of an Alzheimer’s disease patient at a six month follow up visit (see Figure 2(c)). Another example involves the use of imaging and other clinical or biologic data to blindly classify individuals into one or more groups, for instance as either a treatment responder on non-responder. Such models would have important clinical applications to aid in treatment decisions, predict disease progression, and as a diagnostic tool when costs are not prohibitive.
Numerous statistical tools have been developed to address these common objectives in brain imaging studies. We provide an overview of select approaches. This review is not intended to offer a complete description of existing approaches in the field, but rather it will provide the reader with select methods that are used to address central substantive issues, highlight analytic challenges that statisticians face in applying existing methods and in developing new ones, and discuss important areas for future research that will benefit from statistical thinking. We begin by describing the data collected in fMRI studies and highlighting attributes that are important for statistical modeling.
2 Data Descriptions, Analytic Challenges, and Preprocessing
Functional MRI yields dynamic three-dimensional (3-D) maps giving second to second depictions of distributed brain activity patterns. Studies commonly acquire scans every two to three seconds and may yield hundreds of scans in a single session, but the acquisition speed and duration varies according to study objectives. At the time of statistical analysis, each scan is often arranged in a 91×109×91 array, with each volume element (or voxel) in the array containing a localized measure of neural activity. Figure 2(c) displays a 91×109 array for a selected axial slice across the z-dimension. The temporal evolution of brain activity at a single location (voxel), denoted Yi(v) (S × 1), forms a time series as illustrated for two distinct locations in Figure 3. Thus, we may regard fMRI data as either a collection of hundreds of thousands of time series arising from spatially distinct sources or alternatively as a movie of dynamic 3-D brain maps. Either of these perspectives reveals the massive amount of data produced in an imaging study with tens of millions of spatio-temporal neural activity measures for each subject and billions of measures across all subjects for many studies. The enormity of the data poses challenges for statistical modeling and computation.
Incorporating known biologic information into statistical models is often beneficial, but the mere complexity of the brain presents challenges. One challenge stems from the voluminous and intricate systems of networks in the brain, which render correlations that do not necessarily decay with increasing distance. Figure 4(a) shows correlations between the fMRI profile for the selected voxel (at the cross hair) and the profiles from all other voxels in the image. Note that high correlations exist between the selected voxel and many neighboring voxels, bilaterally in the opposite hemisphere, and in some distant areas. There are obvious departures from an assumption that the strengths of associations decrease with increasing distances (see Figure 4(b)), which poses a major challenge for modeling spatial dependence. Another analytic challenge arising from the ultra high dimensionality is that many objectives seek to make inferences at each voxel. One has to cope with multiplicity issues, since this often amounts to tens or hundreds of thousands of statistical tests.
The data proceed through a series of steps from the time that they are retrieved from the scanner to the time of statistical analysis, and we generally refer to these steps as the preprocessing pipeline. Detailed coverage of preprocessing is beyond the scope of this review; however, it is important for the reader to have knowledge of these steps as they may substantially impact subsequent statistical analysis. Our brief remarks omit data processing that occurs prior to retrieving data from the scanner. Typical preprocessing steps include motion correction to adjust for head movement, slice timing correction because each 3-D scan representing a single time point actually consists of several 2-D slices acquired at slightly different times, registration of the fMRI scans to an anatomical MRI scan, normalization to warp each individual’s set of scans to a standard space for group analysis, temporal filtering to address temporal correlations and to remove non-physiologic trends such as scanner drift, and spatial smoothing, e.g. using convolution with a Gaussian kernel, to adjust for residual between-subject neuroanatomic differences that persist following normalization. Another source of motivation for spatial smoothing is that it helps to support the assumptions underlying random field theory (discussed later), which is a popular technique to address multiple testing. These preprocessing steps are covered in more detail by Strother (2006), and they are implemented in several neuroimaging software packages, some of which are freely available (82).
3 Survey of Existing Methods
3.1 Methods for Localization
3.1.1 THE GENERAL LINEAR MODEL
The general linear model (GLM) has been a cornerstone of neuroimaging analyses targeting localization (40). A linear mixed model is conceptually very well suited for neuroactivation analyses to incorporate subject-specific effects, group-level parameters, and correlations between repeated measures obtained from each individual. The massive amount of data, however, precludes routine use of the mixed model due to heavy computational demands. As an alternative, a two-stage modeling approach is employed, with the first stage single-subject GLM given by
(1) |
where Yi(v) (S × 1) is a vector of S serial brain activity (BOLD) measures for subject i at voxel v, Xiv (S × q) is the design matrix containing q independent variables, βi(v) (q × 1) represents the parameter vector linking experimental tasks to the fMRI responses, Hiv (S × m) contains m additional covariates that are not of substantive interest (e.g. high-pass filtering to remove low-frequency signal drift), and εi(v) (S × 1) is random error about the ith subject’s mean (93). We assume , where is unknown, and V reflects the correlations between serial BOLD measures.
The BOLD response to neuronal activity is governed by properties of a hemodynamic response function (HRF) (see Figure 5). Following a stimulus evoking a neuronal response, there is a period of increased blood flow and oxygenation, which peaks after roughly 5 or 6 seconds, and then falls back toward baseline. There is often a transient dip below baseline characterizing the post-stimulus undershoot. To accommodate the HR properties, it is standard to convolve the design matrix with the HRF using . Often a single HRF model is specified for all voxel locations, but some methods more flexibly allow for spatially varying HRFs across voxels (88).
Next, one models the individualized experimental effects in terms of group-level parameters using a second-stage GLM:
(2) |
where βi(v) contains regression coefficients from (1), μ(v) is the group-level mean vector, dj(v) contains random errors, and . One often considers linear contrasts Cβi(v), rather than modeling the entire vector. It is straightforward to show that this two-stage procedure implies the following linear mixed model
(3) |
with fixed effects Xivμ(v), random subject-specific effects Xivdi(v), and random error introduced by εi(v). The two-stage approach substantially reduces the computational burden, since estimation in both GLMs uses least-squares and avoids use of the Newton Raphson procedure or alternative iterative algorithms required for (3). In practice, one replaces βi(v) with estimates, say , which sacrifices some efficiency relative to fitting the linear mixed model (3) directly.
The matrix V incorporating covariances between serial BOLD responses is rarely known in practice, and two strategies are used to address it. Prewhitening obtains an initial estimate of the temporal autocorrelation from the data and subsequently transforms Yi(v) to remove this correlation (87). Precoloring, on the other hand, introduces known autocorrelations through a linear transformation or temporal filtering, say WYi(v) (87,40,92,16,69). Temporal filtering introduces autocorrelations, which are deemed to dominate the existing correlations in V so that the resulting covariance structure is well approximated by , and estimation proceeds using weighted least-squares.
3.1.2 SPATIAL MODELING
One apparent limitation of the GLM framework is that the models are estimated separately for each of the v = 1,…, V voxels, which assumes independence between voxels. Several approaches have been developed that incorporate spatial correlations between neural activity arising from different voxels. The most conceptually straightforward approaches incorporate correlations between a voxel and its contiguous first-order neighbors. Penny et al. (2005) incorporate correlations between in-plane neighboring voxels in a Bayesian model with priors for regression parameters relying on a user specified spatial kernel matrix. Their approach assumes that BOLD responses are spatially homogeneous and locally contiguous within each slice of an image. Katanoda et al. (2002) addressed spatial correlations by including data (for a given voxel) from the six physically contiguous voxels in three orthogonal directions (53). Woolrich et al. (2004) propose a spatio-temporal Bayesian framework that uses simultaneous autoregressive models for neighboring voxels allowing both separable and nonseparable models (89).
As illustrated in Figure 4, spatial correlations extend beyond first order neighbors, may not decrease with increasing distances, and often include select long-range associations. Bowman (2007) proposed a linear mixed model, an extended version of 3, which incorporates temporal (repeated measures) correlations using random effects and captures spatial correlations between voxels with the strengths of these correlations depending on a measure of functional distance dij between the neural activity in voxels i and j (11). Specifically, the model is given by
(4) |
where Yi = [Yi(1)′,…, Yi(V)]′, with Yi(V) as in (1), and Zi = (IV ⊗ 1S). The model decomposes the measured BOLD signal into localized mean components Xiβ, individualized mean-zero random deviations αi that induce temporal correlations between scans, and random errors that exhibit functionally based spatial correlations. The model simultaneously incorporates temporal and spatial correlations via Var(Yi) = (Φi + IS) + σt(IV ⊗ JS), where JS denotes a unit matrix, and Φi handles parametric covariance structures of the form (Φ)ij = σ2f(dij). Despite its flexibility, this model was proposed for region of interest (ROI) studies, which focus on particular neuroanatomic structures, and intensive computations may limit applicability to whole-brain studies.
Other methods offer broader spatial coverage for correlations by linking parameters both within and between defined brain regions. Bowman (2005) uses a simultaneous autoregressive model to capture exchangeable spatial correlations between all pairs of voxels within functionally defined networks. Using a known parcellation of the brain (see Figure 6), Derado et al. (2010) extend this autoregressive model to additionally incorporate repeated measures associations between multiple scanning sessions, e.g. before and after treatment (10, 29). Bowman et al. (2008) also leverage the neuroanatomic parcellation to establish a Bayesian framework for incorporating shorter and longer range correlations by pooling the jth effect from model (2) across all voxels in brain region g (14). The second-stage likelihood function for βigj = (βigj(1),…, βigj(Vg))′ follows from
(5) |
where μgj = (μgj1,…, μgjVg)′. Spatial correlations are introduced by modeling the random effects αigj collectively for all G brain regions, αij = (αi1j,…, αiGj)′, through the following prior probability distributions:
(6) |
Similar to (10,29), this model includes exchangeable correlations between voxels within a neuroanatomic region, and Γi incorporates between region correlations. The model provides an excellent compromise between the necessary sophistication to address several aspects of spatial and temporal correlations present in the data and simplicity to facilitate computational demands as it can be readily implemented with user-friendly software making use of the Gibbs sampler (95). The model does not, however, account for dependence between different scanning sessions or multiple effects obtained on each individual and includes a relatively simple intra-regional correlation model. To overcome these shortcomings, Derado et al. (2012) propose an approach that augments model (5). Their model has useful predictive capabilities, and we discuss the approach in section 3.4.
Xu et al. (2009) present an alternative spatial modeling framework, where the goal is to address variability in activation locations across individuals (94). Collectively, all of the aforementioned spatial modeling extensions offer clear advantages over the standard two-stage GLM. Notable advantages are that the modeling assumptions are better suited to the properties of the data and underlying neurophysiology, increased precision for estimation, increased statistical power, and expanded interpretations concerning the associations, hence interplay, between different brain regions. Spatial smoothing of the data, performed prior to statistical analysis, is a standard preprocessing step for fMRI data. The analyst should carefully consider the influence of spatial smoothing, and we recommend either forgoing spatial smoothing during preprocessing or performing very focal spatial smoothing to limit the impact on subsequent spatial modeling and estimation. Despite the substantial progress in spatial modeling, this remains an important area for involvement by statisticians. Some potential areas of future development include the development of unified (one-stage) spatio-temporal modeling, perhaps with non-separable covariance models and multimodal modeling, integrating supplementary information from other imaging modalities, e.g. regarding the underlying structural connectivity. Note that our use of the term multimodal is consistent with the neuroimaging literature and describes data from two or more imaging techniques that are combined for analysis, rather than the conventional use of the term in statistics to describe multiple modes of a distribution.
3.1.3 UNSPECIFIED ONSETS OF STIMULI
Occasionally studies are designed with unspecified onset times of stimuli to prompt changes in neural activity. For example, if subjects trained in Zen meditation are instructed to focus their awareness by concentrating on breathing while in the scanner, the time at which a subject achieves a deep meditative state may be unknown (66). Robinson et al. (2010) present a model to identify unknown change points in the data (72). Their work estimates the onset distributions evoking a change in neural activity from baseline and present a hidden Markov random field model to cluster voxels based on characteristics of their onset, duration, and anatomical location. Independent component analysis (ICA) is another approach that does require a design matrix a priori, and we discuss ICA in section 3.3.
3.1.4 SPECTRAL MODELING
There are alternatives to modeling fMRI data in the time domain, which may be based on Fourier or wavelet transformations of the data. A primary benefit of performing such transformations is that it simplifies analyses in the transformed space, e.g. because Fourier coefficients are approximately uncorrelated across frequencies, and similarly wavelets have decorrelating properties. The previously discussed approach by Katanoda et al. (2002) conducts a Fourier domain analysis for their model that captures spatial correlations between six nearest neighbor voxels (53). Ombao et al. (2008) present a spatio-spectral model to account for spatial and temporal correlations using spatially varying temporal spectra to characterize the underlying spatiotemporal processes (65).
Kang et al. (2012) proposed a spatio-spectral mixed-effects model that is conceptually similar to the model in equations (5) and (6) (51). The linear mixed effects model is specified for a given frequency band as follows:
(7) |
where g = 1,…, G represents ROIs, v = 1,…, Vg indexes voxels in ROI g, and bgv, , and are mutually independent and normally distributed. With analysis performed in the frequency domain, there are associated components for real and imaginary parts, indexed by j ∈ {R, I}. Correspondingly, , with ; for stimulus p; and , with , where is the spectrum at frequency band . The matrix introduces correlations between the Vg voxels within ROI g and assumes that the covariance elements are determined as a function the Euclidean distances between the corresponding voxels within the ROI, models correlations between the G ROIs, and the simplified covariance structure resulting from the frequency domain model are reflected by .
Rowe (2005) presents a model for complex fMRI data that describes both the magnitude and phase and can be used to test for task related changes in the magnitude, the phase, or both the magnitude and phase (74). Zhu et al. (2009) also present a Rician regression model to characterize noise contributions in fMRI along with associated estimation and diagnostic procedures (96).
3.2 Statistical Inferences
Statistical inferences in neuroactivation studies seek to identify localized task related alterations in brain activity, localized differences in neural activity between groups of subjects, and treatment (or other session) related changes in localized activity. One may target these objectives via a set of null hypotheses H0 = {H01,… ,H0V}, addressed by linear combinations of group-level effects, e.g. Cμ. For frequentist approaches, one proceeds by calculating an appropriate test statistic at each voxel, Tv, often a t-statistic or an F-statistic based on the underlying modeling assumptions. One may also consider nonparametric alternatives such as permutation testing for localized inferences (64). The goal for any of these approaches, whether frequentist or Bayesian, is to produce activation maps as shown in Figure 2(a), which reveal locations exhibiting significant or highly probable changes (or differences) in neural activity. Given the massive number of tests performed, typically hundreds of thousands, it is desirable to establish some control over the collective testing errors.
For a given search region , controlling familywise error at a significance level α involves selecting a threshold tα such that . Common methods for addressing multiplicity in neuroimaging include uncorrected approaches that specify an arbitrarily small significance level such as α = 0.005, Bonferroni-type procedures, random field theory, permutation testing, and false discovery rate (FDR). Of these, the Bonferroni, random field theory, and permutation testing control familywise error.
The widely used Bonferroni approach selects a threshold tα such that for each voxel , which by Boole’s inequality ensures that the familywise error control is achieved. The number of voxels V is extremely larger for whole-brain studies and the test statistics at each voxel are not statistically independent, making the Bonferroni procedure highly conservative. This approach would rarely yield statistical significance, and in practice, Bonferroni corrections are often adapted to consider the size of an activated cluster.
Random field theory is a mathematically elegant approach that regards a map of localized test statistics as a continuous random field, e.g. Gaussian, t, F, or χ2. Any threshold tα applied to the test statistic map has an associated excursion set containing the voxels where the localized statistic exceeds the threshold. Random field theory uses a topological property called the Euler characteristic to summarize this set, which Worsley et al. (1992) heuristically describe as follows: “the number of isolated parts of the excursion set, irrespective of their shape, minus the number of ’holes’ (91),” although the formal definition involves the curvature of the boundary of the excursion set at tangent planes (1). As simple examples, the Euler characteristic is 1 for a solid ball and 0 for a doughnut. If no holes are present, then it counts the number of isolated regions of activation in an image above the threshold. Figure (7) illustrates the appearance of both isolated regions and holes for lower thresholds applied to the t-statistic map (see (a)) and only isolated regions remaining for higher thresholds.
Under the null hypothesis for a given search region , the critical value tα satisfies . If tα is large, then the exceedence probability for the maximum is approximated by the expected value of the Euler characteristic, E(χtα), and is given by
(8) |
where Λ is a matrix of partial derivatives of the random field with respect to the dimensions x, y, and z, which (under a set of assumptions) can be approximated as |Λ|1/2 = (FWHMx × FWHMy × FWHMz)−1(4 loge 2)3/2 using the full widths at half maximum (FWHM) of the Gaussian smoothing kernel, and R = V/(FWHMx × FWHMy × FWHMz) is a measure of the number of resolution elements (resels) in the search volume (91).
The random field theory approach is widely used in the neuroimaging community and is easily implemented using available software packages. Criticisms largely target the required assumptions, which are clearly delineated by Nichols and Hayasaka (2003) (63). Considering the Gaussian case, one assumes that under the null hypothesis, the test statistic image can be modeled by a smooth, homogeneous, mean-zero, unit variance Gaussian random field. The level of smoothness should be sufficient to reflect properties of a continuous random field. The random field theory approach works well with extensive smoothing, for example 10 voxels FWHM, but low smoothness may yield conservative results. The spatial autocorrelation function (ACF) must be twice differentiable at the origin. The random field theory approach assumes that the data are stationary or are stationary after a deformation of space (90). Also, the roughness/smoothness is assumed to be known without appreciable error. Random field theory becomes more conservative with decreasing sample sizes and was shown to yield thresholds comparable to the Bonferroni procedure for data analyses with low degrees of freedom (63).
Resampling testing procedures have been proposed as alternatives for statistical inferences for neuroimaging data, including wavelet-based procedures and traditional permutation testing (19,17,18,64). Nichols and Holmes (2001) use permutation testing to construct an empirical distribution for the maximum statistic Tmax. Implementation for fMRI group-level analyses involves permuting labels across individuals defining contrasts or subgroups, computing the voxel test statistics, and determining for each permutation sample b = 1,…, B, which yields the permutation distribution for the maximum statistic. At significance level α, the critical value is determined from this empirical distribution by the smallest b* such that (1 − b*/B) ≤ α, and any voxel with is declared statistically significant, and exact p-values may be determined directly from this distribution. A key advantage of permutation testing is that it does not require strong assumptions about the distribution of the data, in contrast to the random field theory approach. Nichols and Holmes (2002) showed that permutation tests may yield increases in power with small sample sizes over other testing procedures.
FDR has received attention in several large-scale data areas. In contrast to familywise error control, FDR protects against the expected rate of false discoveries (or false positives) among the significant tests (9,43) and is defined to be 0 when no tests are rejected. FDR offers control according to the following: E(FDR) ≤ π0ω, where π0 is the unknown proportion of null hypotheses that are true and ω is the user-specified level of control. In many neuroimaging applications, π0 ≈ 1 since there will be no effect in the vast majority of voxels. So setting π0 = 1 is often reasonable, yet it may be conservative when π0 is substantially smaller than 1. Adaptive FDR procedures seek less conservative approaches by estimating the unknown quantity π0. Reiss et al. (2012) reveal vulnerabilities with adaptive FDR, including astonishing paradoxical cases where adaptive FDR yields more liberal results than not making any correction for multiplicity (70). FDR is easily implemented, for a specified rate ω, by first ordering the p-values p(1) ≤ p(2) ≤ … ≤ p(V) for the V voxels and then determining the largest i, say i*, such that
(9) |
The procedure declares voxels corresponding to p(1),…, p(i*) as significant.
FDR is a flexible approach for addressing multiplicity, since it is implemented using p-values, which can be produced by a range of models and testing frame-works. FDR does not require smoothed data and, in fact, is more powerful for unsmoothed data. This is a relative strength compared to random field theory, which often requires aggressive smoothing for good performance. Some view FDR as a technique that is best suited for applications targeting preliminary discovery, where one seeks to control the number of discoveries that prove to be false at a subsequent validation phase. In neuroimaging, however, activation results from whole-brain analyses tend to be the final goal and are not routinely designed with plans for subsequent validation analyses on discovered findings.
3.3 Connectivity
3.3.1 FUNCTIONAL CONNECTIVITY
Functional connectivity refers to the temporal coherence in neural activity between spatially remote brain regions (39). It is examined using various undirected measures of association including Pearson’s correlation coefficient, partial correlation, mutual information, and spectral coherence, among others. Seed-based methods represent one simple approach to determine functional connectivity. Seed-based techniques simply identify a select number of brain regions, usually via hypothesis driven selection, and calculate the associations between these seeds and every other brain region considered, producing correlation images such as that depicted in Figure 4(a). To mitigate the impact of varying hemodynamics for different brain regions, one may accommodate time lags between brain regions i and j, e.g. using the following:
(10) |
An area of growing interest for seed-based (and other) connectivity approaches concerns the potential dynamic nature of functional associations or networks, whereby connectivity patterns change dynamically over time. Seed-based procedures remain popular due to their simplicity in implementation and interpretation. A major drawback is that they may miss important findings since they are inherently limited in scope by considering a small number of seed regions. These methods are typically hypothesis driven, but the exact seed locations may heavily influence the findings.
A related approach to determine connectivity is to select a set of brain regions to serve as nodes and then calculate associations between all nodal pairs. The nodes may target hypothesis driven brain regions, such as the putative depression-related regions in Figure 8(a), or may represent subregions drawn from an exhaustive parcellation of the brain, with an example shown in Figure 8(b). In either case, associations are computed between every pair of nodes, which produces a complete functional connectivity matrix. The connectivity matrix in Figure 8(c) corresponds to the 90 whole-brain ROIs shown in Figure 8(b). Depending on the objectives, one may scale up to include a larger number of regions to generate whole-brain connectivity networks (see Figure 2(b)). As the ultimate goal, functional connectivity studies often seek to determine group or treatment related differences in connectivity patterns.
Partitioning approaches organize the brain into collections of voxels that have shared properties, with larger differences between groupings. Two such procedures are independent component analysis (ICA) and cluster analysis. ICA is often motivated by the classical cocktail party problem, in which a group of people are talking simultaneously in a room, and one is interested in isolating one of the voices (or discussions). The human brain is quite adept at handling such tasks. Signal processing approaches regard the problem as one of blind source separation, in which the goal is to dissociate the mixture of signals into the originating sources without a priori information about the sources or the mixing process, while assuming that the sources are statistically independent.
Letting Y(T × V) represent an individual’s fMRI data from T scans, ICA decomposes the data into linear combinations of spatio-temporal source signals:
(11) |
where M(T × q) is the nonsingular mixing matrix whose columns contain latent time series for each of the q independent components, S(q × V) has rows containing statistically independent spatial signals, with columns assumed to be non-Gaussian, and E(T × V) contains noise or variability not explained by the independent components, with each column assumed to follow a multivariate normal distribution (5, 56). Noise free ICA omits the error term E (56). From the noise free formulation, S = M−1Y, which clearly reveals that M−1 operates as an unmixing matrix yielding statistically independent signals from the fMRI data. ICA has been extended for group analyses, with approaches developed by Calhoun et al. (2001), Beckmann and Smith (2005), Guo and Pagnoni (2008), Guo (2011), and Eloyan et al. (2013) (6, 21, 34, 48, 47).
ICA decomposes observed fMRI data into spatially independent sources, with each component having an associated latent time series. Putative attributions about the nature of the components may be made, for example, consistently task-related, transiently task-related, and other physiologic and non physiologic sources. These task-related attributions stem from relating the latent time series to the stimuli, but the interpretation of many components is often challenging and can not be established with certainty. One noted advantage of ICA relative to modeling approaches such as the GLM is that it does not require specification of a design matrix. Yet one can still determine brain regions that are associated with the experimental stimuli. ICA does not, however, provide a comprehensive framework for inferences concerning the relationship between the experimental stimuli and the BOLD response. Another advantage of ICA is that noise related signals revealed by the procedure can be removed from the BOLD response prior to subsequent analyses.
Cluster analysis is another set of techniques for partitioning the brain into networks of voxels or regions that exhibit similar temporal or task-related dynamics. There are well established methods in the statistical literature for performing cluster analysis. These data-driven methods have been successfully applied in brain imaging, including K-means (2, 44, 45), fuzzy clustering (3, 35, 36, 80), hierarchical clustering methods (45, 12, 13, 26, 81), a hybrid hierarchical K-means approach (37), and dynamical cluster analysis (DCA) (4), among others. Hierarchical clustering generally begins with each brain region (or node) as a single cluster, calculates functional distances between all pairs of brain regions, e.g. fij = 1 − ρij, where ρij is the partial correlation between regions i and j, and iteratively joins the most similar regions or clusters (recalculating the functional distances at each step), with the process ceasing when only a single cluster remains. The procedures then examine each stage of the resulting clustering tree to determine the optimal number of clusters, often based on criteria assessing both within and between cluster variability. Bowman et al. (2012) present a multimodal approach that combines fMRI with structural connectivity information derived from DTI to determine functional connectivity via cluster analysis (15). They defined a distance measure
(12) |
where πij ∈ [0,1) is the probability of structural connectivity between brain regions i and j determined from the DTI and λ ∈ [0, ∞) is an attenuation parameter that is optimized empirically. This distance incorporates structural connectivity, but still permits clustering of regions (or clusters) based on fij in the absence of structural connectivity, i.e. when πij → 0. This method generally improves network coherence, particularly with increased noise in the fMRI signal.
There has been a recent emergence of complex functional brain network analyses, borrowing from other applications of network science. Such methods work in conjunction with approaches that specify a set of brain regions to serve as nodes and then quantify associations between the temporal fMRI profiles for every pair of regions. One may view the resulting map of associations as a system of interacting regions, for instance shown in Figure 2(b). Network analyses attempt to summarize various characteristics of these whole-brain networks and then to conduct hypothesis testing about these properties. Typical summaries include graph metrics to reflect communication ability (either local or global), such as clustering coefficient, path length, and efficiency; centrality metrics such as degree, betweenness, closeness, eigenvector centrality; and community structure including whole-brain topological properties such as small-worldness (76).
Most methods address undirected networks and thus convey information on whole-brain functional connectivity. There are current needs for statistical input for proper inference based on network metrics. For example, the metrics are estimated and have associated sampling distributions, but most approaches do not take this variability into account. Comparing brain network properties between groups of subjects also requires statistical thinking for formal inferences.
Simpson et al. (2011) develop a multivariate approach that applies exponential random graph models (ERGMs) to functional brain networks (75). The approach represents global network structure by locally specified explanatory metrics. Let W(G × G) denote a random, symmetric, connectivity matrix with the (i, j)th element Wij = 1 if regions i and j pass a minimum connection threshold, resulting in a connecting edge in the graphic representation, and Wij = 0 otherwise, where i = 1,…, G and j = 1,…, G. ERGMs specify
(13) |
where h(w, X) is a pre-specified network feature, possibly consisting of covariates that are functions of the network (e.g., number of paths of specified length) and nodal covariates X (e.g. location of the node), θ is a parameter vector linking the network feature to the connectivity matrix after accounting for the contribution of other network features in the model, and κ(θ) is a normalizing constant. The model yields inferences about whether certain local network properties are observed more than would be expected by chance.
3.3.2 EFFECTIVE CONNECTIVITY
Some neuroimaging analyses seek to determine stronger relationships than the undirected associations describing functional connectivity. As an example, effective connectivity targets the influence that one region exerts on another (39). Patel et al. (2006) present an approach that uses a Bayesian model to quantify functional connectivity based on the relative difference between the marginal probability that a voxel, v1, is active and the probability that v1 is active conditional on elevated activity in voxel v2. Larger differences between these conditional and marginal probabilities reflect voxel pairs exhibiting stronger functional connections (67). Their approach then investigates the existence of a stronger hierarchical relationship between each pair of functionally connected voxels using a measure called ascendancy. v1 is ascendant to v2 when the marginal activation probability of v1 is larger than that of v2. The model yields measures of the degree of functional connectivity between a voxel pair and the degree of ascendancy of one voxel relative to the other.
Structural equation models (SEM) have been applied to fMRI and PET to determine causal associations between brain regions. SEM focuses on the covariance structure that reflects associations between the variables. Parameter estimation in SEM minimizes differences between the observed covariances and those implied by a user-defined path (or structural) model. The parameters of the SEM represent the connection strengths between the brain activity measurements in different regions and correspond with measures of effective connectivity.
Dynamic causal modeling (DCM) regards the brain as a deterministic nonlinear dynamic system that receives inputs and produces outputs and uses a Bayesian modeling framework to estimate effective connectivity (41, 42). DCM seeks to estimate parameters at the neuronal level so that the modeled BOLD signals are maximally similar to the observed BOLD signals. The approach parameterizes effective connectivity in terms of coupling, representing the influence of one brain region on another (42).
Granger causality (GC) has recently gained attention in the neuroimaging literature as a method to establish directional relationships between the neural activity from two spatially distinct regions. Let Yv = {Yv(t), t = 1,…, T} denote a stationary time series reflecting the measured brain activity in voxel v. Consider the model
(14) |
where Λj(2 × 2) is a matrix of unknown coefficients and εv represents model error. This model regresses the current value of a time series for one voxel, say v1, on the history of both v1 and v2. Additionally consider the model
(15) |
where γj is an unknown scalar coefficient relating the current neural activity in voxel v1 to its own history. GC is defined as
(16) |
which gives a measure of the extent to which the past values from v1 and v2 assist in predicting the current value of v1 over and above the extent to which the past values of v1 predict the current value of v1 (46). Causality is inferred if there is a significant improvement in model fit by including the cross-autoregressive terms.
Critics of GC have raised issues about its utility in practice for functional neuroimaging (38, 84, 85, 86, 60, 62). Questions center on whether there is sufficient temporal resolution in fMRI data, with repetition times often two seconds or more, to ascertain causality from the lagged association models. Solo (2011) notes that GC found on a slow time-scale, e.g. based on fMRI, does not necessarily hold on a faster time-scale (79). He suggests that ms time-scale measurements, e.g. MEG/EEG, are necessary to pursue dynamic causality. This critique is of course applicable to other effective connectivity approaches based on fMRI data. Also, haemodynamic variations across the brain are likely to swamp any causal lag in the underlying neural time series (38, 73). Measurement noise can reverse the estimation of GC direction, and temporal smoothing can induce causal relationships (77). Smith et al. (2011) empirically evaluated the performances of several functional connectivity approaches as well as directional measures. They observed that GC and other lagged methods generally performed poorly, but the approach by Patel et al. (2006) performs reasonably well.
3.4 Prediction
There are an increasing number of neuroimaging analyses targeting prediction and classification. This area has perhaps the greatest potential for translational impact for functional neuroimaging. There have been a range of prediction objectives targeted. We consider methods that involve the use of neuroimaging data, perhaps coupled with other data, to either forecast future neural activity or to predict or blindly classify a clinical outcome or behavioral response. For example, prediction and classification methodology could potentially be used to define imaging markers of depression subtypes, to identify neural patterns of individuals in an undifferentiated aging population who have a high probability of developing Parkinson’s disease, and to determine distinct neural profiles of patients who respond to a particular therapeutic treatment.
We begin with methods aimed at forecasting future neural activity. Guo et al. (2008) developed a Bayesian model that uses a patient’s pretreatment scans, along with relevant characteristics, to predict the patient’s brain activity after a specified treatment regimen (49). The predicted post-treatment neural activity maps provide objective and clinically relevant information that may be incorporated into the treatment selection process. Let Yi1(v)(S1 × 1) denote the vector of baseline scans for subject i corresponding to voxel v, Yi2(v) (S2 × 1) represent scans from a post-baseline follow-up period, and Yi(v) = (Yi1(v), Yi2(v))′. Guo et al. fit a GLM for Yi(v) (S × 1) analogous to that specified in (1), but extended by assuming a linear covariance structure for the error term of the second stage model (analogous to (2)) to capture the correlations between pre- and post-baseline neuroimaging data. Prediction proceeds using the conditional distribution of [βi2(v) | βi1, μ(v)]. They demonstrated that their prediction framework accurately forecasted post-treatment neural processing in both a PET study of working memory among individuals with schizophrenia and an fMRI study of inhibitory control among cocaine-dependent subjects.
Derado et al. (2012) developed an extended Bayesian spatial hierarchical framework for predicting follow-up neural activity based on an individual’s baseline functional neuroimaging data (30). Their approach increases precision by borrowing strength from the spatial correlations present in the data, while handling temporal correlations between different scanning sessions. Let βig(v) = (βig1(v), βig2(v)) represent parameters for both sessions one and two from a first stage GLM. Extending models (2) and (5), Derado et al. (2012) propose
(17) |
where ϕg(v) is a spatial dependence parameter for local correlations, αig is a random effects vector, and Ψg is a variance-covariance matrix associated with the repeated scanning sessions. Figure 9 depicts the framework for temporal and local spatial correlations. Priors addressing spatio-temporal correlations include
(18) |
Hence, their approach addresses spatial correlations between defined neuroanatomical regions via Γj using an unstructured model, between all voxel pairs within each of the defined brain region using an exchangeable covariance structure, and via a multivariate conditional autoregressive model for the immediate 26 neighbors of each brain voxel (see Figure 9). Correlations due to repeated scanning sessions are captured by the variance-covariance matrix Ψg. They demonstrated good performance using PET data from a study of Alzheimer’s disease to predict disease progression, and the method is also applicable to fMRI data.
An increasingly common goal is to use neuroimaging data to predict a behavioral or clinical response. One begins with training data D = {X, y}, where X(n × p) is a matrix of p image-derived independent variables (also called features), and y is a response vector, with yi ∈ {+1, −1} for binary prediction and for regression. One uses the training data to develop a model that yields accurate prediction of a separate sample y* from input variables x*. We consider the binary case for simplicity, and we include both settings in which observation of y* follows that of x* (prediction) and those in which a temporal sequence cannot be established but one seeks to blindly distinguish subgroups (classification). Prediction and classification are often based on . These studies face challenges due to the commonly described curse of dimensionality caused by having substantially greater number of measures p on each subject (possibly hundreds of thousands) than the number subjects (n ≤ 50, in many cases).
Support vector machines (SVM) or support vector classifiers (SVC) have perhaps been the most popular tool for prediction or classification in neuroimaging analyses (22, 27, 31, 32, 54, 59). In its simplest form, the SVC determines linear boundaries in the feature space to distinguish two classes. This framework can be extended to identify nonlinear boundaries and extended to multi-class or regression settings for continuous outcomes. SVM methods do not perform variable selection, but they are able to cope with high-dimensional neuroimaging data structures. The existing literature, coupled with our experiences applying SVM, suggests that these techniques usually exhibit good performance in practice. Chen et al. (2011) developed methodology to apply SVM techniques to longitudinal neuroimaging data, where prediction is based on linear combinations of features from different scanning sessions (24).
Related methods for prediction and classification have been proposed, which incorporate a loss function and a penalty term on the model complexity, e.g. penalized least-squares. Chu et al. (2011) apply two kernel regression techniques, specifically kernel ridge regression (KRR) and relevance vector regression (RVR) (25). The KRR method implements the dual formulation of ridge regression to facilitate computations, and RVR considers a set of linear basis functions as the kernel and uses sparse Bayesian methods for estimation. Michel et al. (2011a) also present a sparse Bayesian regression approach for multiclass prediction, which initially groups features into several classes and then applies class-specific regularization, attempting to adapt the amount of regularization to the available data (57). Michel et al (2011b) apply an norm of the image gradient, also called the total variation (TV), as regularization (58). Their method tends to determine block structure, assuming that the spatial layout of the neural processing is sparse and spatially structured in groups of connected voxels. Bunea et al. (2011) apply the widely used LASSO and elastic net procedures to neuroimaging data for prediction and implement a bootstrap based extension to aid parsimony of variable selection (20). Marquand et al. (2010) present a prediction model that uses Gaussian processes to forecast pain intensity from whole-brain fMRI data (55).
Despite the number of studies pursuing prediction and classification, there is a need for more principled applications of statistical methods and for the development of new methods. Examples of more principled use of existing techniques include assessments of the sensitivity of various approaches to tuning parameters, evaluating variability involved in assessing generalization accuracy, and incorporating precision into variable selection techniques, among others. In addition, there are opportunities to apply or possibly develop techniques to leverage the vast and complicated structure present in imaging-derived variables and to incorporate known structure from auxiliary information. Statistics may also play a role in encouraging biologic plausibility in the variable selection process.
3.5 Software
There are several software that are available for neuroimaging analysis, reflecting a major strength in the field. Many of these software packages are freely accessible and serve as major assets to applied researchers. Several packages are fairly comprehensive, implementing various preprocessing steps, statistical analyses, and advanced visualization techniques. Numerous specialized software packages are also available. We list some popular software tools to aid readers, but we do not attempt to give a comprehensive summary of available packages.
FMRIB Software Library (FSL), developed by is a comprehensive library of tools for fMRI, MRI and DTI. Statistical Parametric Mapping (SPM) is a collection of MATLAB functions equipped with a graphical user interface, which offers broad preprocessing and analysis capabilities for fMRI and MRI data. Similarly, AFNI (often interpreted as an acronym for Analysis of Functional NeuroImages) is a set of C programs for processing, analyzing, and displaying fMRI data. Brain Voyager is a commercial package containing tools for the analysis of fMRI, DTI, EEG, MEG and Transcranial magnetic stimulation (TMS) data. Beyond these comprehensive packages, numerous software tools are available to implement specialized analytic methods or analyses of data from a particular modality. Examples of such tools include Bayesian Spatial Model for Activation and Connectivity (BSMac), DTI Studio, Meta-analysis Toolbox, Brain Connectivity Toolbox (BCT), Connectome Workbench, and medInria.
4 Discussion
Neuroimaging is an exciting and rapidly expanding field that is advancing our understanding of the brain, impacting neuroscience, psychology, psychiatry, and neurology. This paper provides a survey of major substantive objectives and existing analytic methods, but by no means is our summary comprehensive. Brain imaging research is inherently interdisciplinary, and the role of statistics is critical as it contributes to defining rigorous methodology for extracting information and for quantifying statistical evidence.
One area of growing interest is multimodal imaging, which has the potential to incorporate imaging modalities reflecting physiologically distinct, yet complementary, information, for instance related to brain function, localized properties of tissue density, local diffusion properties, structural connections between regions, and electrophysiological measures of neural activity. These multimodal imaging data may lead to more accurate and reliable analytic approaches than considering data from a single modality and also may expand the information content and possible interpretations. Multimodal analyses also have the benefit of more complete information by pooling data with physiologically similar objectives, such as brain function, but with distinct bases for measurement such as blood oxygenation, blood flow, metabolism, and electrical activity. Moreover, various imaging modalities incorporates information across different temporal and spatial scales. Despite the many potential advantages, multimodal data present challenges for statistical analysis. The dimensionality may become unwieldy. Alignment of information across different temporal scales, say milliseconds to seconds, may cause analytic and interpretative issues. Similarly, integrating information across spatial scales may prove difficult. Adding to the potential sources of information, multimodal data may include non-imaging data such as genomic, clinical, demographic, and biologic information, which also exacerbates some the aforementioned challenges.
A second area for statistical contribution involves developing methodology that is able to incorporate biologic information. Given the rapid expansion and intense interest in neuroimaging research, more information will become available to incorporate in future modeling efforts. To give examples, the Human Connectome Project, funded by the National Institutes of Health, is an enormous initiative aimed at advancing our knowledge about the human brain and its functional and structural connectivity properties. Also, United States President Barack Obama announced plans for the Brain Research through Advancing Innovative Neurotechnologies (BRAIN) initiative, which promises to be another substantial national investment to aid our understanding of the brain. As these large-scale initiatives, combined with the rapid expansion of the field more broadly, provide insights about the brain, statisticians will have an opportunity to develop more functionally and structurally informed models, e.g. via Bayesian modeling frameworks through incorporation of physiologically-based prior distributions.
Many imaging datasets have both nested spatial and temporal structures, which make Bayesian methodology appealing. Bayesian modeling has proven to be beneficial for neuroimaging data by distributing the overall complexities of the data across various hierarchical levels and enabling flexible posterior inferences. The models, however, often involve a very large number of parameters, which may bring about computational issues related to Markov chain Monte Carlo methods for posterior simulations and concerning assessments of properties of simulations such as convergence and dependence of posterior draws for model parameters. Carefully constructed models, for example those enabling the use of Gibbs sampler, facilitate computations and are often reasonable to implement in practice. For additional modeling flexibility, there may be a need to consider the use of alternative posterior approximation strategies such as variational Bayes to reduce the computational demands.
Some neuroimaging studies are moving from traditional cross-sectional designs to longitudinal designs. The growing number of longitudinal studies and the acquisition of multimodal images raises the issue of missing data. Subjects may have data missing from a single measurement occasion or from a single modality, so the common practice of discarding all data from subjects who have any missing data is inefficient. Some immediate gains may be made by applying existing methods in the statistical literature for handling missing data to neuroimaging, but the enormity and complexity of imaging data prompts the need for additional methodological development.
There is substantial variability in fMRI data, both within and between subjects, and group studies are often based on limited sample sizes, prompting the need to consider analytic techniques and study design considerations that lead to reliable findings. As one strategy, meta-analyses are important to determine both the consistency in task-related changes in brain activity across studies and the consistency of simultaneously activated pairs or networks of brain regions (23, 33, 52, 72). Conducting studies with larger sample sizes would also help to yield more precise estimates and more powerful tests of statistical hypotheses, and such larger studies including hundreds or even thousands of subjects are beginning to emerge in the field. In summary, neuroimaging is an exciting and rapidly expanding field, which presents numerous challenges for quantifying evidence, and statistics should play a critical role in the growth of this interdisciplinary field.
Acknowledgments
This paper was partially supported by NIH grant U18 NS082143-01.
LITERATURE CITED
- 1.Adler RJ. The geometry of random fields. New York: Wiley; 1981. p. 89. [Google Scholar]
- 2.Balslev D, Nielsen FA, Frutiger SA, et al. Cluster analysis of activity-time series in motor learning. Human Brain Mapping. 2002;15:135–45. doi: 10.1002/hbm.10015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Baumgartner R, Ryner L, Richter W, et al. Comparison of two exploratory data analysis methods for fMRI: fuzzy clustering vs. principal component analysis. Magnetic Resonance Imaging. 2000;18:89–94. doi: 10.1016/s0730-725x(99)00102-2. [DOI] [PubMed] [Google Scholar]
- 4.Baune A, Sommer FT, Erb M, et al. Dynamical cluster analysis of cortical fMRI activation. NeuroImage. 9:477–489. doi: 10.1006/nimg.1999.0429. [DOI] [PubMed] [Google Scholar]
- 5.Beckmann CF, Smith SM. Probabilistic Independent Component Analysis for Functional Magnetic Resonance Imaging. IEEE Transactions on Medical Imaging. 2004;23(2):137–152. doi: 10.1109/TMI.2003.822821. [DOI] [PubMed] [Google Scholar]
- 6.Beckmann CF, Smith SM. Tensorial extensions of independent component analysis for multisubject FMRI analysis. NeuroImage. 2005;25:294–311. doi: 10.1016/j.neuroimage.2004.10.043. [DOI] [PubMed] [Google Scholar]
- 7.Behrens TEJ, Woolrich MW, Jenkinson M, Johansen-Berg H, Nunes RG, Clare S, Matthews PM, Brady JM, Smith SM. Characterization and propagation of uncertainty in diffusion-weighted MR imaging. Magn Reson Med. 2003;50:1077–1088. doi: 10.1002/mrm.10609. [DOI] [PubMed] [Google Scholar]
- 8.Behrens TEJ, Johansen-Berg H, Jbabdi S, Rushworth MFS, Wool-rich MW. Probabilistic diffusion tractography with multiple fibre orientations: what can we gain? NeuroImage. 2007;34:144–155. doi: 10.1016/j.neuroimage.2006.09.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Benjamini Y, Hochberg Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society, Series B. 1995;57:289–300. [Google Scholar]
- 10.Bowman FD. Spatio-Temporal Modeling of Localized Brain Activity. Biostatistics. 2005;6:558–575. doi: 10.1093/biostatistics/kxi027. [DOI] [PubMed] [Google Scholar]
- 11.Bowman F. Spatio-temporal models for region of interest analyses of functional neuroimaging data. J Am Stat Assoc. 2007;102(478):442–453. [Google Scholar]
- 12.Bowman FD, Patel R, Lu C. Methods for detecting functional classifications in neuroimaging data. Human Brain Mapping. 2004;23(2):109–119. doi: 10.1002/hbm.20050. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Bowman FD, Patel R. Identifying spatial relationships in neural processing using a multiple classification approach. NeuroImage. 2004;23:260–68. doi: 10.1016/j.neuroimage.2004.04.022. [DOI] [PubMed] [Google Scholar]
- 14.Bowman FD, Caffo B, Bassett SS, Kilts CD. A Bayesian hierarchical framework for spatial modeling of fMRI data. NeuroImage. 2008;39:146–156. doi: 10.1016/j.neuroimage.2007.08.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Bowman FD, Zhang L, Derado G, Chen S. Determining Functional Connectivity using fMRI Data with Diffusion-Based Anatomical Weighting. NeuroImage. 2012;62:1769–1779. doi: 10.1016/j.neuroimage.2012.05.032. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Bullmore ET, Brammer MJ, Williams SCR, Rabe-Hesketh S, Janot N, David AS, Mellers JDC, Howard R, Sham P. Statistical Methods of Estimation and Inference for Functional MR Image Analysis. Magnetic Resonance in Medicine. 1996;35:261–277. doi: 10.1002/mrm.1910350219. [DOI] [PubMed] [Google Scholar]
- 17.Bullmore E, Fadili J, Breakspear M, Salvador R, Suckling J, Brammer M. Wavelets and statistical analysis of functional magnetic resonance images of the human brain. Stat Methods Med Res. 2003;12:375–399. doi: 10.1191/0962280203sm339ra. [DOI] [PubMed] [Google Scholar]
- 18.Bullmore E, Fadili J, Maxim V, Xendur L, Whitcher B, Suckling J, Brammer M, Breakspear M. Wavelets and functional magnetic resonance imaging of the human brain. NeuroImage. 2004;23:S234–S249. doi: 10.1016/j.neuroimage.2004.07.012. [DOI] [PubMed] [Google Scholar]
- 19.Bullmore E, Long C, Suckling J, Fadili J, Calvert B, Zelaya F, Carpenter TA, Brammer M. Colored Noise and Computational Inference in Neurophysiological (fMRI) Time Series Analysis: Resampling Methods in Time and Wavelet Domains. Human Brain Mapping. 2001;12:61–78. doi: 10.1002/1097-0193(200102)12:2<61::AID-HBM1004>3.0.CO;2-W. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Bunea F, She Y, Ombao H, Gongvatana A, Devlin K, Cohen R. Penalized least squares regression methods and applications to neuroimaging. NeuroImage. 2011;55:1519–1527. doi: 10.1016/j.neuroimage.2010.12.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Calhoun VD, Adali T, Pearlson GD, Pekar JJ. A method for making group inferences from functional MRI data using independent component analysis. Human Brain Mapping. 2001;14:140–151. doi: 10.1002/hbm.1048. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Casanova R, Hsu FC, Mark A. Espeland, for the Alzheimer’s Disease Neuroimaging Initiative: Classification of structural MRI images in Alzheimer’s disease from the perspective of ill-posed problems. PLoS One. 2012;7:e44877. doi: 10.1371/journal.pone.0044877. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Cauda F, Cavanna AE, D’Agata F, Sacco K, Duca S, Geminiani GC. Functional connectivity and coactivation of the nucleus accumbens: a combined functional connectivity and structure-based meta-analysis. Journal of cognitive neuroscience. 2011;23:2864–2877. doi: 10.1162/jocn.2011.21624. [DOI] [PubMed] [Google Scholar]
- 24.Chen S, Bowman FD. A Novel Support Vector Classifier for Longitudinal High-dimensional Data and Its Application to Neuroimaging Data. Statistical Analysis and Data Mining. 2011;4(6):604–611. doi: 10.1002/sam.10141. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Chu C, Ni YZ, Tan G, Saunders CJ, Ashburner J. Kernel regression for fMRI pattern prediction. NeuroImage. 2011;56(2):662–673. doi: 10.1016/j.neuroimage.2010.03.058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Cordes D, Haughton V, Carew JD, et al. Hierarchical clustering to measure connectivity in fMRI resting-state data. Magnetic Resonance Imaging. 2002;20:305–317. doi: 10.1016/s0730-725x(02)00503-9. [DOI] [PubMed] [Google Scholar]
- 27.Cox D, Savoy R. Functional magnetic resonance imaging (fMRI) brain reading: detecting and classifying distributed patterns of fMRI activity in human visual cortex. NeuroImage. 2003;19:261270. doi: 10.1016/s1053-8119(03)00049-1. [DOI] [PubMed] [Google Scholar]
- 28.Daubechies I, Roussos E, Takerkart S, Benharrosh M, Golden C, DArdenne K, Richter W, Cohen JD, Haxby J. Independent component analysis for brain fMRI does not select for independence. PNAS. 2009;106(26):10415–10422. doi: 10.1073/pnas.0903525106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Derado G, Bowman FD, Kilts C. Modeling the spatial and temporal dependence in fMRI data. Biometrics. 2010;66:949–957. doi: 10.1111/j.1541-0420.2009.01355.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Derado G, Bowman FD, Zhang L. Predicting Brain Activity using a Bayesian Spatial Model. Statistical Methods in Medical Research. 2012 doi: 10.1177/0962280212448972. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Doehrmann O, Ghosh SS, Polli FE, Reynolds GO, Horn F, Keshavan A, Triantafyllou C, Saygin ZM, Whitfield-Gabrieli S, Hofmann SG, Pollack M, Gabrieli JD. Predicting treatment response in social anxiety disorder from functional magnetic resonance imaging. JAMA Psychiatry. 2013;70(1):87–97. doi: 10.1001/2013.jamapsychiatry.5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Dosenbach NU, Nardos B, Cohen AL, Fair DA, Power JD, Church JA, Nelson SM, Wig GS, Vogel AC, Lessov-Schlaggar CN, Barnes KA, Dubis JW, Feczko E, Coalson RS, Pruett JR, Jr, Barch DM, Petersen SE, Schlaggar BL. Prediction of individual brain maturity using fMRI. Science 10. 2010;329(5997):1358–61. doi: 10.1126/science.1194144. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Eickhoff SB, Laird AR, Grefkes C, Wang LE, Zilles K, Fox PT. Coordinate-based activation likelihood estimation meta-analysis of neuroimaging data: A random-effects approach based on empirical estimates of spatial uncertainty. Human Brain Mapping. 2009;30:2907–2926. doi: 10.1002/hbm.20718. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Eloyan A, Crainiceanu CM, Caffo BS. Likelihood-based population independent component analysis. Biostatistics. 2013:1–14. doi: 10.1093/biostatistics/kxs055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Fadili MJ, Ruan S, Bloyet D, et al. A multistep unsupervised fuzzy clustering analysis of fMRI time series. Human Brain Mapping. 2000;10:160–78. doi: 10.1002/1097-0193(200008)10:4<160::AID-HBM20>3.0.CO;2-U. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Fadili MJ, Ruan S, Bloyet D, et al. On the number of clusters and the fuzziness index for unsupervised FCA application to BOLD fMRI time series. Medical Image Analysis. 2001;5:55–67. doi: 10.1016/s1361-8415(00)00035-9. [DOI] [PubMed] [Google Scholar]
- 37.Filzmoser P, Baumgartner R, Moser E. A hierarchical clustering method for analyzing functional MR images. Magnetic Resonance Imaging. 1999;17:817–826. doi: 10.1016/s0730-725x(99)00014-4. [DOI] [PubMed] [Google Scholar]
- 38.Friston K. Causal modelling and brain connectivity in functional magnetic resonance imaging. PLoS Biology. 2009;7(2):e1000033. doi: 10.1371/journal.pbio.1000033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Friston KJ, Frith CD, Liddle PF, Frackowiak RSJ. Functional connectivity: the principal component analysis of large (PET) data sets. Journal of Cerebral Blood Flow and Metabolism. 1993;13:5–14. doi: 10.1038/jcbfm.1993.4. [DOI] [PubMed] [Google Scholar]
- 40.Friston KJ, Holmes AP, Poline JB, Grasby PJ, Williams SCR, Frackowiak RSJ, Turner R. Analysis of fMRI Time-Series Revisited. NeuroImage. 1995;2:45–53. doi: 10.1006/nimg.1995.1007. [DOI] [PubMed] [Google Scholar]
- 41.Friston KJ. Bayesian estimation of dynamic systems: an application to fMR. NeuroImage. 2002;16:513–30. doi: 10.1006/nimg.2001.1044. [DOI] [PubMed] [Google Scholar]
- 42.Friston KJ, Harrison L, Penny W. Dynamic causal modeling. NeuroImage. 2003;19(4):1273–302. doi: 10.1016/s1053-8119(03)00202-7. [DOI] [PubMed] [Google Scholar]
- 43.Genovese CR, Lazar NA, Nichols TE. Thresholding of Statistical Maps in Functional Neuroimaging Using the False Discovery Rate. NeuroImage. 2002;15:870–878. doi: 10.1006/nimg.2001.1037. [DOI] [PubMed] [Google Scholar]
- 44.Goutte C, Toft P, Rostrup E, et al. On clustering fMRI time series. NeuroImage. 1999;9:298–310. doi: 10.1006/nimg.1998.0391. [DOI] [PubMed] [Google Scholar]
- 45.Goutte C, Nielsen FA, Liptrot MG, et al. Featurespace clustering for fMRI meta-analysis. Human Brain Mapping. 2001;13(3):165–83. doi: 10.1002/hbm.1031. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Granger C. Investigating causal relations by econometric models and cross-spectral methods. Econometrica. 1969;37:424–438. [Google Scholar]
- 47.Guo Y. A General Probabilistic Model for Group Independent Component Analysis and Its Estimation Methods. Biometrics. 2011;67:1532–1542. doi: 10.1111/j.1541-0420.2011.01601.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Guo Y, Pagnoni G. A unified framework for group independent component analysis for multi-subject fMRI data. NeuroImage. 2008;42:1078–1093. doi: 10.1016/j.neuroimage.2008.05.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Guo Y, Bowman FD, Kilts CD. Predicting the Brain Response to Treatment using a Bayesian Hierarchical Model with Application to a Study of Schizophrenia. Human Brain Mapping. 2008;29(9):1092–1109. doi: 10.1002/hbm.20450. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Hendelman W. Atlas of Functional Neuroanatomy. 2. CRC Press; 2005. [Google Scholar]
- 51.Kang H, Ombao H, Linkletter C, Long N, Badre D. Spatio-Spectral Mixed Effects Model for Functional Magnetic Resonance Imaging Data. Journal of the American Statistical Association. 2012;107(498):568–577. doi: 10.1080/01621459.2012.664503. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Kang J, Johnson TD, Nichols TE, Wager TD. Meta analysis of functional neuroimaging data via Bayesian spatial point processes. Journal of the American Statistical Association. 2011;106:124–134. doi: 10.1198/jasa.2011.ap09735. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Katanoda K, Matsuda Y, Sugishita M. A Spatio-Temporal Regression Model for the Analysis of Functional MRI Data. NeuroImage. 2002;17:1415–1428. doi: 10.1006/nimg.2002.1209. [DOI] [PubMed] [Google Scholar]
- 54.LaConte S, Strother S, Cherkassky V, Anderson J, Hu XP. Support vector machines for temporal classification of block design fMRI data. NeuroImage. 2005;26:317329. doi: 10.1016/j.neuroimage.2005.01.048. [DOI] [PubMed] [Google Scholar]
- 55.Marquand A, Howard M, Brammer M, Chu C, Coen S, Mouro-Miranda J. Quantitative prediction of subjective pain intensity from whole-brain fMRI data using Gaussian processes. NeuroImage. 2010;49(3):2178–2189. doi: 10.1016/j.neuroimage.2009.10.072. [DOI] [PubMed] [Google Scholar]
- 56.McKeown MJ, Makeig S, Brown GG, Jung TP, Kindermann SS, Bell AJ, Sejnowski TJ. Analysis of fMRI Data by Blind Separation Into Independent Spatial Components. Human Brain Mapping. 1998;6:160–188. doi: 10.1002/(SICI)1097-0193(1998)6:3<160::AID-HBM5>3.0.CO;2-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Michel V, Eger E, Keribin C, Thirion B. Multiclass sparse Bayesian regression for fMRI-based prediction. International Journal of Biomedical Imaging. 2011 doi: 10.1155/2011/350838. Article ID 350838. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Michel V, Gramfort A, Varoquaux G, Eger E, Thirion B. Total Variation Regularization for fMRI-Based Prediction of Behavior. IEEE Transactions on Medical Imaging. 2011;30(7):1328–1340. doi: 10.1109/TMI.2011.2113378. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Mouro-Miranda J, Bokde ALW, Born C, Hampel H, Stetter M. Classifying brain states and determining the discriminating activation patterns: Support Vector Machine on functional MRI data. NeuroImage. 2005;28:980995. doi: 10.1016/j.neuroimage.2005.06.070. Find this article online. [DOI] [PubMed] [Google Scholar]
- 60.Nalatore H, Ding M, Rangarajan G. Mitigating the effects of measurement noise on Granger causality. Physical Review E. 2007;75(3):31123.1–31123.10. doi: 10.1103/PhysRevE.75.031123. [DOI] [PubMed] [Google Scholar]
- 61.National Institutes of Health. The Human Connectome Project. NIH Blueprint for Neuroscience Research. http://www.neuroscienceblueprint.nih.gov/connectome/index.htm, retrieved 03-25-2013.
- 62.Nolte G, Ziehe A, Nikulin V, Schlgl A, Krmer N, Brismar T, Mller K. Robustly estimating the flow direction of information in complex physical systems. Physical Review Letters. 2008;100(23):234101.1–234101.4. doi: 10.1103/PhysRevLett.100.234101. [DOI] [PubMed] [Google Scholar]
- 63.Nichols TE, Hayasaka S. Controlling the familywise error rate in functional neuroimaging: a comparative review. StatisticalMethods in Medical Research. 2003;12:419–446. doi: 10.1191/0962280203sm341ra. [DOI] [PubMed] [Google Scholar]
- 64.Nichols TE, Holmes AP. Nonparametric Permutation Tests For Functional Neuroimaging: A Primer with Examples. Human Brain Mapping. 2001;15:1–25. doi: 10.1002/hbm.1058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Ombao H, Shao X, Rykhlevskaia E, Fabiani M, Gratton G. Spatio-Spectral Analysis of Brain Signals. Statistica Sinica. 2008;18:1465–1482. [Google Scholar]
- 66.Pagnoni G, Cekic M, Guo Y. Thinking about Not-Thinking: Neural Correlates of Conceptual Processing during Zen Meditation. PLOS One. 2008;3(9):e3083. doi: 10.1371/journal.pone.0003083. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Patel R, Bowman FD, Rilling JK. A Bayesian approach to determining connectivity of the human brain. Human Brain Mapping. 2006;27:267–76. doi: 10.1002/hbm.20182. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Penny WD, Trujillo-Bareto N, Friston KJ. Bayesian fMRI time series analysis with spatial priors. NeuroImage. 2005;24(2):350–362. doi: 10.1016/j.neuroimage.2004.08.034. [DOI] [PubMed] [Google Scholar]
- 69.Purdon PL, Solo V, Weisskoff RM, Brown E. Locally Regularized Spatiotermporal Modeling and Model Comparison for Functional MRI. NeuroImage. 2001;14:912–923. doi: 10.1006/nimg.2001.0870. [DOI] [PubMed] [Google Scholar]
- 70.Reiss PT, Schwartzman A, Lu F, Huang L, Proal E. Paradoxical results of adaptive false discovery rate procedures in neuroimaging studies. NeuroImage. 2012;63:1833–1840. doi: 10.1016/j.neuroimage.2012.07.040. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Robinson JL, Laird AR, Glahn DC, Lovallo WR, Fox PT. Metaanalytic connectivity modeling: delineating the functional connectivity of the human amygdala. Human brain mapping. 2010;31:173–184. doi: 10.1002/hbm.20854. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Robinson LF, Wager TD, Lindquist MA. Change point estimation in multi -subject fMRI studies. Neuroimage. 2010;49(2):1581–92. doi: 10.1016/j.neuroimage.2009.08.061. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Roebroeck A, Formisano E, Goebel R. Mapping directed influence over the brain using Granger causality and fMRI. Neuroimage. 2005;25(1):230–242. doi: 10.1016/j.neuroimage.2004.11.017. [DOI] [PubMed] [Google Scholar]
- 74.Rowe D. Modeling both the Magnitude and Phase of Complex-Valued fMRI Data. NeuroImage. 2005;25:1310–1324. doi: 10.1016/j.neuroimage.2005.01.034. [DOI] [PubMed] [Google Scholar]
- 75.Simpson SL, Hayasaka S, Laurienti PJ. Exponential Random Graph Modeling for Complex Brain Networks. PLoS ONE. 2011;6(5):e20039. doi: 10.1371/journal.pone.0020039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Simpson SL, Bowman FD, Laurienti PJ. Analyzing complex functional brain networks: fusing statistics and network science to understand the brain. Statistics Surveys. 2013 doi: 10.1214/13-SS103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Smith SM, Miller KL, Salimi-Khorshidi G, Webster M, Beckmann CF, Nichols TE, Ramsey JD, Woolrich MW. Network modelling methods for fMRI. NeuroImage. 2011;54:875–891. doi: 10.1016/j.neuroimage.2010.08.063. [DOI] [PubMed] [Google Scholar]
- 78.Society for Neuroscience. Brain Facts: A Primer on the Brain and Nervous System. 2008. [Google Scholar]
- 79.Solo V. The Dangers of Granger: What They Didnt Tell you about Granger Causality (in fMRI); 17th Annual Meeting of the Organization for Human Brain Mapping; Qubec City, Canada. 2011. [Google Scholar]
- 80.Sommer FT, Wichert A. Exploratory analysis and data modeling in functional neuroimaging. Cambridge (MA): The MIT Press; 2003. pp. 17–42. [Google Scholar]
- 81.Stanberry L, Nandy R, Cordes D. Cluster analysis of fMRI data using dendrogram sharpening. Human Brain Mapping. 2003;20:201–219. doi: 10.1002/hbm.10143. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Strother S. Evaluating fMRI Preprocessing Pipelines: Review of Preprocessing Steps for BOLD fMRI. IEEE Engineering in Medicine and Biology. 2006 Mar-Apr;:27–41. doi: 10.1109/memb.2006.1607667. [DOI] [PubMed] [Google Scholar]
- 83.Suzana Herculano-Houzel, 2012
- 84.Tiao G, Wei W. Effect of temporal aggregation on the dynamic relationship of two time series variables. Biometrika. 1976;63(3):513–523. [Google Scholar]
- 85.Wei W. The effect of temporal aggregation on parameter estimation in distributed lag model. Journal of Econometrics. 1978;8(2):237–246. [Google Scholar]
- 86.Weiss A. Systematic sampling and temporal aggregation in time series models. Journal of Econometrics. 1984;26(3):271–281. [Google Scholar]
- 87.Woolrich MW, Ripley BD, Brady M, Smith SM. Temporal autocorrelation in univariate linear modeling of FMRI data. Neuroimage. 2001;14(6):1370–86. doi: 10.1006/nimg.2001.0931. [DOI] [PubMed] [Google Scholar]
- 88.Woolrich MW, Behrens TE, Smith SM. Constrained linear basis sets for HRF modelling using Variational Bayes. NeuroImage. 2004;21(4):1748–61. doi: 10.1016/j.neuroimage.2003.12.024. [DOI] [PubMed] [Google Scholar]
- 89.Woolrich M, Jenkinson M, Brady J, Smith S. Fully Bayesian spatiotemporal modeling of fMRI data. IEEE Trans Med Imag. 2004;23(2):213–231. doi: 10.1109/TMI.2003.823065. [DOI] [PubMed] [Google Scholar]
- 90.Worsley KJ, Andermann M, Koulis T, MacDonald D, Evans AC. Detecting changes in nonisotropic images. Human Brain Mapping. 1999;8:98–101. doi: 10.1002/(SICI)1097-0193(1999)8:2/3<98::AID-HBM5>3.0.CO;2-F. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Worsley K, Evans A, Marrett S, Neelin P. A three-dimensional statistical analysis for CBF activation studies in human brain. Journal of Cerebral Blood Flow and Metabolism. 1992;12:900–918. doi: 10.1038/jcbfm.1992.127. [DOI] [PubMed] [Google Scholar]
- 92.Worsley KJ, Friston KJ. Analysis of fMRI Time Series RevisitedAgain. NeuroImage. 1995;2:173–181. doi: 10.1006/nimg.1995.1023. [DOI] [PubMed] [Google Scholar]
- 93.Worsley K, Liao C, Aston J, Petre V, Duncan G, Moralese F, Evans A. A general statistical analysis for fMRI data. NeuroImage. 2002;15:1–15. doi: 10.1006/nimg.2001.0933. [DOI] [PubMed] [Google Scholar]
- 94.Xu L, Johnson TD, Nichols TE, Nee DE. Modeling Inter-Subject Variability in fMRI Activation Location: A Bayesian Hierarchical Spatial Model. Biometrics. 2009;65:1041–1051. doi: 10.1111/j.1541-0420.2008.01190.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Zhang L, Agravat S, Derado G, Chen S, Bowman FD. BSMac: A MATLAB toolbox Implementing a Bayesian Spatial Model for Brain Activation and Connectivity. Journal of Neuroscience Methods. 2012;204:133–143. doi: 10.1016/j.jneumeth.2011.10.025. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Zhu H, Li Y, Ibrahim JG, Shi X, An H, Chen Y, Gao W, Lin W, Rowe DB, Peterson BS. Regression Models for Identifying Noise Sources in Magnetic Resonance Images. Journal of the American Statistical Association. 2009;104:623–637. doi: 10.1198/jasa.2009.0029. [DOI] [PMC free article] [PubMed] [Google Scholar]