Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2019 Dec;203:116187. doi: 10.1016/j.neuroimage.2019.116187

Spatial confidence sets for raw effect size images

Alexander Bowring a, Fabian Telschow b, Armin Schwartzman b,c, Thomas E Nichols a,d,e,
PMCID: PMC6854455  PMID: 31533067

Abstract

The mass-univariate approach for functional magnetic resonance imaging (fMRI) analysis remains a widely used statistical tool within neuroimaging. However, this method suffers from at least two fundamental limitations: First, with sufficient sample sizes there is high enough statistical power to reject the null hypothesis everywhere, making it difficult if not impossible to localize effects of interest. Second, with any sample size, when cluster-size inference is used a significant p-value only indicates that a cluster is larger than chance. Therefore, no notion of confidence is available to express the size or location of a cluster that could be expected with repeated sampling from the population.

In this work, we address these issues by extending on a method proposed by Sommerfeld et al. (2018) (SSS) to develop spatial Confidence Sets (CSs) on clusters found in thresholded raw effect size maps. While hypothesis testing indicates where the null, i.e. a raw effect size of zero, can be rejected, the CSs give statements on the locations where raw effect sizes exceed, and fall short of, a non-zero threshold, providing both an upper and lower CS.

While the method can be applied to any mass-univariate general linear model, we motivate the method in the context of blood-oxygen-level-dependent (BOLD) fMRI contrast maps for inference on percentage BOLD change raw effects. We propose several theoretical and practical implementation advancements to the original method formulated in SSS, delivering a procedure with superior performance in sample sizes as low as N=60. We validate the method with 3D Monte Carlo simulations that resemble fMRI data. Finally, we compute CSs for the Human Connectome Project working memory task contrast images, illustrating the brain regions that show a reliable %BOLD change for a given %BOLD threshold.

Graphical abstract

Image 1

1. Introduction

Over the last three decades, the statistical parametric mapping procedure (Friston et al., 1994a) for inference of task-fMRI data has prevailed as the international standard within the field of neuroimaging. Incorporating a mass-univariate statistical approach, functional data at each voxel is described in terms of experimental conditions and residual variability included as parameters in a general linear model. From this model, a group-level statistical parametric map (SPM) of t-statistic’s contrasting a specified experimental condition relative to a baseline condition is formed. Using a corrected significance level based on the theory of random fields to account for the multiple-comparison problem (Friston et al., 1994b), hypotheses are tested at each voxel independently and the SPM is finally thresholded to localize brain function. While simple by nature, this technique has proven immensely powerful and provided us with the tools to gain deep insight into cognitive function.

There is, however, information that is not captured using the current fMRI approach to inference. Specifically, for clusterwise inference, the cluster-level p-value only conveys information about a cluster’s spatial extent under the null-hypothesis. Since no information is provided regarding the statistical significance of each voxel comprising a significant cluster, the most we can say is that significant activation has occurred somewhere inside the cluster (Woo et al., 2014). An implication of this is that when a large, sprawling cluster covers many anatomical regions, the precise spatial specificity of the activation is in fact poor. While a recent effort has attempted to solve this problem by ‘drilling down’ to find the exact source of activation (Rosenblatt et al., 2018), this can come at the cost of lower statistical power. A related problem of cluster inference is that no information is provided about the spatial variation of significant clusters. For example, if a given fMRI study were to be repeated many times with new sets of subjects, there would of course be variation in the size and shape of clusters found, yet the current statistical results have no way to characterize this variability.

A more pressing issue, perhaps, stems from an age-old paradox caused by the ‘fallacy of the null hypothesis’ (Rozeboom, 1960). The paradox is that while statistical models conventionally assume mean-zero noise, in reality all sources of noise will never cancel, and therefore improvements in experimental design will eventually lead to statistically significant results. Thus, the null-hypothesis will, eventually, always be rejected (Meehl, 1967). The recent availability of ambitious, large-sample studies (e.g Human Connectome Project (HCP), N=1,200; UK Biobank, N=30,000 and counting) have exemplified this problem. Analysis of high-quality fMRI data acquired under optimal noise conditions has been shown to display almost universal activation across the entire brain after hypothesis testing, even with stringent correction (Gonzalez-Castillo et al., 2012).

For the reasons discussed above, alongside further concerns about misconceptions and the misuse of p-values in statistical testing (Nuzzo, 2014; Wasserstein et al., 2016), there has been a growing consensus among sections of the neuroimaging community that the statistical results commonly reported in the literature should be supplemented by effect estimates (Chen et al., 2017; Nichols et al., 2017). The main argument put forward supporting raw effect sizes is that they increase interpretability of the statistical results, highlighting the magnitude of statistically significant differences and providing another layer of evidence to support the overall scientific conclusions inferred from an fMRI study. This may also help tackle reproducibility concerns that have become prominent within the field due to failed attempts in replicating published neuroimaging results (Poldrack et al., 2017), a problem aggravated by the ubiquity of underpowered studies in the fMRI literature where traditional statistical inference methods are unlikely to detect the majority of meaningful effects (Cremers et al., 2017; Turner et al., 2018).

In this work, we seek to address all of these issues by applying and extending a spatial inference method initially proposed by Sommerfeld et al. (2018) (SSS) to obtain precise confidence statements about where activation occurs in the brain. Unlike hypothesis testing, our spatial Confidence Sets (CSs) allow for inference on non-zero raw effect sizes. While the method can be applied to any parameter in a mass-univariate general linear model, here we will focus inference on the mean percentage BOLD change raw effect. For a cluster-forming threshold c, and a predetermined confidence level 1α, the CSs comprise of two sets: the upper CS (denoted Aˆc+, red voxels in Fig. 1), giving all voxels we can assert have a percentage BOLD raw effect size truly greater than c; and the lower CS (Aˆc, blue voxels overlapped by yellow and red in Fig. 1), for which all voxels outside this set we can assert have a percentage BOLD raw effect size truly less than c. The upper CS is smaller and nested inside the lower CS, and the assertion is made with (1α)100% confidence holding simultaneously for both regions. Fig. 1 provides an illustration of the schematic we will use to display the CSs, also showing the point estimate set (Aˆc, yellow voxels overlapped by red) obtained by thresholding the data at c.

Fig. 1.

Fig. 1

Schematic of the colour-coded regions used to visually represent the Confidence Sets (CSs) and point estimate set. CSs displayed in the glass brain were obtained by applying the method to 80 participants contrast data from the Human Connectome Project working memory task, using a a c=2.0% BOLD change threshold at a confidence level of 1α=95%.

With this interpretation, the CSs can be linked to traditional statistical voxelwise thresholding with control of the familywise error rate (FWE): In a one-sided t-test, for the set of level α FWE-significant voxels we have (1α)100% confidence that the signal is greater than zero. Put another way, we have (1α)100% confidence that the voxelwise level α FWE results are all true positives. The CSs can be viewed as a generalisation of these methods, except that the confidence statement is no longer relative to a signal of zero, but to a non-zero signal magnitude c. Users may choose the threshold c based on prior knowledge of raw effect sizes reported in previous similar studies to their own; since computation of the CSs is quick, users may also report results for a variety of cluster-forming thresholds as we do in this work.

The motivating data in SSS were longitudinal temperature data in North America, and the goal was to infer on areas at risk of climate change. In this work, we are motivated by subject-level fMRI contrast of a parameter estimate maps, and we seek to infer brain areas where a substantial raw effect is present in units of percentage BOLD change. In SSS, the CSs were referred to as ‘Coverage Probability Excursion sets’ – shortened to ‘CoPE sets.’

The main contributions of this work are modifications to the SSS method for computing CSs that improve the method’s finite-sample performance in the context of neuroimaging. In particular, we propose a combination of the Wild t-Bootstrap method and the use of Rademacher variables (instead of Gaussian variables) for multiplication of the bootstrapped residuals, which we find substantially improves performance of the method in moderate sample sizes (e.g. N=60). We also develop a linear interpolation method for computing the boundary over which the bootstrap is applied, and a novel approach for assessing the empirical coverage of the CSs that reduces upward bias in how the simulation results are measured. Another contribution here is that we assess the finite-sample accuracy of the method on synthetic 3D signals that are representative of fMRI activation clusters, whereas SSS only considered 2D images. Altogether, we carry out a range of 3D simulations alongside smaller 2D simulations to evaluate our proposed methodological modifications and compare our results to the simulations conducted in SSS. Finally, we apply the method to the Human Connectome Project working memory task dataset, operating on the subject-level percentage BOLD change raw effect maps, where we obtain CSs for a variety of cluster-forming thresholds. Here, the method localizes brain activation in cognitive regions commonly associated to working memory, determining with 95% confidence a raw effect of at least 2% BOLD change.

The remainder of this paper is organized as follows. First of all, we summarize the key theory of CSs before detailing our proposed modifications. We then describe the settings used for our simulations, and provide background information about the HCP dataset analyzed in this work. Finally, we report the results of our simulations before presenting the CSs computed for the HCP data.

2. Theory

2.1. Overview

A comprehensive treatment of the original method, including proofs, can be found in SSS. Here we develop the method specifically for the general linear model (GLM) and describe our own enhancements to the method. While the method can be performed for subject-level inference, we will motivate the method in the context of a group-level analysis, describing how the method can be applied to subject-level %BOLD estimate maps in order to obtain group-level CSs making confidence statements about %BOLD effect sizes relating to the entire population from which the participants were drawn.

For a compact domain SRD, e.g. D=3, consider the GLM at location sS,

Y(s)=Xβ(s)+ε(s) (1)

where Y(s) is a N×1 vector of observations at s, X is a N×p design matrix, β(s) is a p×1 vector of unknown coefficients, and ε(s) a N×1 vector of mean-zero errors, independent over observations, and with each εi having common variance σ2(s) and some unspecified spatial correlation. (Throughout we use boldface to indicate a vector- or matrix-valued variable.) In the context of a task-fMRI analysis, Y(s) is a vector of subject-level %BOLD response estimate maps obtained by applying a first-level GLM to each of the N participants functional data.

For a p×1 contrast vector w, we seek to infer regions of the brain where a contrast of interest wTβ has exceeded a fixed threshold c. Particularly, we are interested in the noise-free, population cluster defined as:

Ac={sS:wTβ(s)c}. (2)

Since we are unable to determine this excursion set in practice, our solution is to find spatial CSs: an upper set Aˆc+ and lower set Aˆc that surround Ac for a desired confidence level of, for example, 95%. We emphasize that these clusters regard the raw units of the signal. Going forward, we assume that the design matrix X and contrast w have been carefully chosen so that wTβˆ has the interpretation of mean %BOLD change across the group. For example, in a one-sample group fMRI model where data Y have %BOLD units, choosing X as a column of 1’s and w=(1) would ensure that wTβˆ has units of %BOLD change.1 In this setting, we wish to obtain an upper CS, Aˆc+, such that we have 95% confidence all voxels contained in this set have a population raw effect size greater than, for example, c=2.0% BOLD change, and a lower CS, Aˆc, such that we have 95% confidence all voxels outside of this set have a population raw effect size less than 2.0% BOLD change. Moreover, we desire that the 95% confidence statement holds simultaneously across both CSs at once. SSS show that a construction of such CSs is possible within the general linear model framework using the following key result.

2.1.1. Result 1

Consider the general linear model setup described in (1). Let βˆ denote the ordinary least squares estimator of β, βˆ(s)=(XTX)1XTY(s), and define vw2=wT(XTX)1w to be the normalised variance of the contrast estimate.

Then for a constant k, and for upper and lower CSs defined as

Aˆc+:={s:wTβˆ(s)c+kσˆ(s)vw},Aˆc:={s:wTβˆ(s)ckσˆ(s)vw},

the limiting coverage of the CSs is

limnP[Aˆc+AcAˆc]=P[supsAc|G(s)|k],

where Ac denotes the boundary of Ac, and G is a smooth Gaussian field on S with mean zero, unit variance, and with the same spatial correlation as each εi.

Result 1 is subject to continuity of the relevant fields and some basic conditions on the increments and moments of the error field ε. A list of these assumptions, as well as a proof of Result 1, are itemized in SSS.

For a pre-determined confidence level 1α (e.g. 1α=95%), by choosing k such that

P[supsAc|G(s)|k]1α, (3)

Result 1 ensures with asymptotic probability of 1α that Aˆc contains the true Ac, and Aˆc+ is contained within Ac. In practice, k is determined as the (1α)100 percentile of the maximum distribution of the asymptotic absolute error process |G(s)| over the true boundary set Ac=s:wTβs=c (see Fig. 2). The upper CS taken away from the lower CS (Aˆc(Aˆc+)c) can be interpreted analogously to a standard confidence interval: with a confidence of 1α, we can assert the true boundary Ac lies within this region. Here, we allude to the classical frequentist interpretation of confidence, where stated precisely, there is a probability of 1α that the region (Aˆc(Aˆc+)c) computed from a future experiment fully encompasses the true set boundary Ac.

Fig. 2.

Fig. 2

A demonstration of how the CSs are computed for a realization of the GLM Y(s)=Xβ(s)+ε(s) in one dimension, for each location s. The yellow voxels Aˆc are obtained by thresholding the observed group contrast map at threshold c; this is the best guess of Ac, the set of voxels whose true, noise-free raw effect surpasses c. The red upper CS Aˆc+ and blue lower CS Aˆc are computed by thresholding the signal at c+kσˆ(s)vw and ckσˆ(s)vw, respectively. We have (1α)100% confidence that Aˆc+AcAˆc, i.e. that Aˆc+ (red) is completely within the true Ac, and Ac is completely within Aˆc- (blue). We find the critical value k from the (1α)100 percentile of the maximum distribution of the absolute error process over the estimated boundary Aˆc (green ’s) using the Wild t-Bootstrap; σˆ is the estimated standard deviation and vw is the normalised contrast variance.

Application of Result 1 presents us with two challenges: that the boundary set Ac and the critical value k are both unknown. To solve the first problem, SSS propose using Aˆc as a plug-in estimate of Ac. There remain, however, technicalities at to how the boundary is determined in any non-abstract setting, and in particular in a 3D image. In Section 2.3 we propose our own novel method for boundary estimation. Before that, we address the second problem, finding the critical value k via a Wild t-Bootstrap resampling scheme.

2.2. The Wild t-Bootstrap method for computation of k

To apply Result 1, we require knowledge of the tail distribution of the limiting Gaussian field G along the boundary Ac. However, the distribution of this field is unknown, because it is dependent on the unknown spatial correlation of the errors εi. We can approximate the maximum distribution of G using the Gaussian Wild Bootstrap (Chernozhukov et al., 2013), also known as the Gaussian Multiplier Bootstrap, which multiplies residuals by random values to create surrogate instances of the random errors.

SSS construct G as follows: The standardized residuals,

ε˜(s)=Y(s)Xβˆ(s)σ(s), (4)

are multiplied by i.i.d. Gaussian random variables, r1*,,rN*, summed and scaled,

G*(s)=1Ni=1Nri*ε˜i(s), (5)

producing a field G* with approximately the same covariance as each error εi, where the superscript asterisk (*) indicates these are just one of many bootstrap realizations. With B bootstrap samples G*, we choose k as the (1α)100 percentile of the B suprema supsAˆc|G*(s)| to approximate the LHS of (3) and apply Result 1 to obtain the CSs.

Up to this point, we have summarized the Gaussian Wild Bootstrap methodology as proposed in SSS. However, when applying this method to our own simulations, we consistently found that our coverage results fell below the nominal level. This was particularly severe for 3D simulations we conducted using a small sample size (N=60), where our results in some cases suffered from under-coverage of 40% or more below the nominal level (see Fig. 8). Hence we made two alterations: First, while SSS used Gaussian multipliers, we found improved performance using Rademacher variables, where each ri takes on 1 or -1 with probability 1/2; others have also reported improved performance with Rademacher variables as well (Davidson and Flachaire, 2008). Second, we implemented a Wild t-Bootstrap (Telschow and Schwartzman, 2019) method, normalizing the bootstrapped residuals ε˜i(s) by their standard deviation σˆ*. This detail was omitted in the proof of Result 1 provided in SSS, where the true standard deviation was assumed to be known. By taking into account the estimation of the standard deviation via the Wild t-Bootstrap, we found improved performance for moderate sample sizes. The Wild t-Bootstrap version of G is

G˜*(s)=1Ni=1Nri*ε˜i(s)σˆ*(s), (6)

where σˆ*(s) is the standard deviation of the present realization of the bootstrapped residuals ri*ε˜i(s). We then determine k as described above but using G˜* instead of G*. Going forward, we refer to this method as the “Wild t-Bootstrap”, to be contrasted with the original “Gaussian Wild Bootstrap” method proposed in SSS.

Fig. 8.

Fig. 8

Coverage results for the 3D large spherical signal (Signal 2. in Fig. 6) simulation with homogeneous Gaussian noise. Empirical coverage results are presented for implementations of the CS method with and without the Wild t-Bootstrap we propose in Section 2.2, and the interpolation schema for assessing simulations results we propose in Section 2.4. Once again, all simulations using the SSS assessment method quickly converge to close to 100%. Using our proposed assessment method, the Gaussian Wild bootstrap had severe under-coverage for small sample sizes, while the Wild t-Bootstrap results hover slightly above the nominal level for all sample sizes.

With these two alterations we found a dramatic increase in performance for small sample sizes in 3D simulations. Notably, in contrast to the Gaussian Wild Bootstrap, our simulation results presented in Section 4 suggest that empirical coverage rates for this modified procedure remain valid, i.e. stay above the nominal level.

2.3. Approximating the boundary on a discrete lattice

In the previous section, we described the ideal Wild t-Bootsrap procedure used to obtain the maximum distribution of G along the boundary Ac in order to apply Result 1. However, in any practical application, data will be observed on a discrete grid of lattice points at a fixed resolution. Therefore, a key challenge is how to appropriately approximate the true continuous boundary Ac from the lattice representation of the data.

In SSS, spline-interpolation was used to estimate a 1D boundary at a resolution greater than their 2D sampled field (SSS, Section 4.1). However, to apply the method to fMRI data we will work with 3D images, and estimating a 2D spline boundary for a 3D field is more involved, requiring careful tuning of the spline basis to accommodate the structure of the 3D signal. Instead, we choose to use a first-order weighted linear interpolation method to approximate the signal values at estimated locations along the true, continuous boundary Ac, providing a method of boundary estimation that is less computationally intensive than spline interpolation. Consider two adjacent points on the lattice, sO and sI, such that sO lies outside of Ac, while sI lies inside Ac. By the definition of Ac, wTβsO<c, and wTβ(sI)c. Under the assumption that the component of the signal between sO and sI increases linearly, we can find the location s* between sO and sI such that wTβ(s*)=c, our estimate of where the true continuous boundary Ac crosses between sO and sI. We can then construct a linear interpolant for the location s*, using weights

m1=wTβ(sI)cwTβ(sI)wTβ(sO),m2=cwTβ(sO)wTβ(sI)wTβ(sO), (7)

for locations sO and sI, respectively. By construction, applying m1 and m2 for locations sO and sI, respectively. By construction, applying m1 and m2 to the contrast image returns the threshold: m1wTβ(sO)+m2wTβ(sI)=wTβ(s*)=c. Applied to standardized residuals ε˜(sO) and ε˜(sI), we can likewise obtain the residuals at the estimated continuous boundary point ε˜(s*)=m1ε˜(sO)+m2ε˜(sI).

By repeating this procedure for all adjacent points sO and sI that lie on the lattice either side of Ac, we are able to estimate the standardized residual values at locations that should approximately sample the true continuous boundary Ac, and thus we can apply the ideal Wild t-Bootstrap procedure outlined in Section 2.2. Of course, in practice we apply this interpolation method on the observed, noisy data, using the plug-in estimated boundary Aˆc.

In the simulation results in Section 4, we assess performance of the method when the bootstrap procedure is carried out over the true boundary Ac, and the plug-in estimated boundary Aˆc that must be used in practice.

2.4. Assessment of continuous coverage on a discrete lattice

In testing the finite-sample validity of our method through simulation, it is imperative that we are able to accurately measure when violations of the subset condition Aˆc+AcAˆc occur. While this may seem a trivial task, as touched on in the previous section, the boundaries of each of these three sets can become ambiguous when data are collected on a discrete lattice.

To illustrate this point, consider the configuration of sets displayed in Fig. 3a. In this instance, suppose the right half of the image corresponds to Ac (green pixels overlapped by yellow), and yellow pixels belong to Aˆc+. We wish to determine whether the condition Aˆc+Ac has been violated or not. One may argue that at the resolution for which the data have been acquired, all pixels that belong to Aˆc+ also belong to Ac, and therefore no violation has occurred. However, the example presented in Fig. 3a has in fact been derived from a 2D simulation conducted at a higher resolution: this 50×50 simulation was obtained by down-sampling a 100×100 grid by dropping every other pixel. Fig. 3a displays the sets Ac and Aˆc+ from the down-sampled, low resolution simulation, while Fig. 3b shows the same set of results at the original resolution. In Fig. 3b we see that there has been an upcrossing of the yellow pixels belonging to Aˆc+ over the boundary of the green, and therefore the subset condition Aˆc+Ac has been violated. From this simulation, it is clear that when we conclude that no violation has occurred in situations like Fig. 3a, our empirical coverage will miss violations and be positively biased. By an analogous argument the same issue occurs when testing violations of AcAˆc.

Fig. 3.

Fig. 3

Demonstrating the resolution issue for testing the subset condition Aˆc+AcAˆc.

Fig. 3a: Here Ac is comprised of the right half of the image (all green and yellow pixels), and Aˆc+ is shown as yellow pixels. It appears that Aˆc+Ac.

Fig. 3b: The same configuration as Fig. 3a at double the resolution. Here, we have enough detail to see that Aˆc+ has crossed the boundary Ac (yellow seeping into blue), and the subset condition Aˆc+Ac has been violated.

In SSS this direct comparison of the lattice representation of the three sets was used to assess coverage in the simulations. While they observed this phenomenon of missed violations leading to over-coverage, the proposed solution was to sequentially increase the resolution of the data. We instead again make use of interpolation.

Since, in simulation, we know the true continuous mean image and Ac, following the method described in Section 2.2 we can obtain weights m1 and m2 to interpolate between points sO and sI either side of the true, continuous boundary Ac, in order to find a location s that approximately lies on the boundary (if the true mean is linear, it would be exactly on the boundary). To determine if sAˆc+, we then re-apply the weights m1 and m2 and assess whether

wTβˆ(s)kσˆ(s)vw=m1(wTβˆ(sO)kσˆ(sO)vw)+m2(wTβˆ(sI)kσˆ(sI)vw)c. (8)

If the inequality holds, then by definition sAˆc+. Otherwise, sAˆc+, and therefore we can conclude that the subset condition Aˆc+Ac has been violated. By checking whether wTβˆ(s)+kσˆ(s)vwc, we can similarly test for a violation of AcAˆc.

By applying this interpolation scheme to all pairs of lattice points with one point inside, one outside, the lattice representation of the boundary, we have devised a method to more accurately assess violations of the subset condition Aˆc+AcAˆc for configurations similar to Fig. 3a. We applied this method for testing the subset condition in our simulations alongside a direct comparison of the lattice representations of the three sets of interest as was done in SSS. The addition of the weighted interpolation method caused a considerable decrease in the empirical coverage results towards the nominal level in all of our 3D simulations. Using the direct comparison of the three sets on its own here essentially determined total empirical coverage (Aˆc+AcAˆc for all simulation runs), even when using small sample sizes and a low nominal coverage level. This is likely to be because the discrete lattice of observed data points is relatively less dense inside the true continuous process for larger, 3D settings, and therefore more violations of the subset condition are missed if only a direct comparison of the lattice representation of the CSs is carried out.

3. Methods

3.1. Simulations

In this section we describe the settings used in order to evaluate the CSs obtained for synthetic data. As a simplified instance of the general linear model setup described in Section 2.1, we simulate 3000 independent samples of the signal-plus-noise model

Yi(s)=μ(s)+εi(s),i=1,,N

using a range of signals μ(s), Gaussian noise structures εi(s) with stationary and non-stationary variance, in two- and three-dimensional regions S. We compute the critical value k, applying the Wild t-Bootstrap method outlined in Section 2.2 with B=5000 bootstrap samples to both the true boundary Ac and the plug-in boundary Aˆc that would be used in practice. The boundaries were obtained using the interpolation method outlined in Section 2.3. We then compare the empirical coverage – the percentage of trials that the true thresholded signal is completely contained between the upper and lowers CSs (i.e. the number of times for which Aˆc+AcAˆc) – across the two sets of results, using the assessment method outlined in Section 2.4. In each simulation, we apply the method for sample sizes of N=60,120,240 and 480, and using three nominal coverage probability levels 1α=0.80,0.90 and 0.95.

3.2. 2D simulations

We analyzed the performance of the CSs on a square region of size 100×100. For the true underlying signal μ(s) we considered two different raw effects: First, a linear ramp that increased from a magnitude of 1 to 3 in the x-direction while remaining constant in the y-direction (Fig. 4a). Second, a circular effect, created by placing a circular phantom of magnitude 3 and radius 30 in the centre of the search region, which was then smoothed using a 3 voxel FWHM Gaussian kernel (Fig. 4b). If we were to assume that each voxel had a size of 2 mm3, we note that this would amount to applying smoothing with a 6 mm FWHM kernel, a fairly typical setting used in fMRI analyses.

Fig. 4.

Fig. 4

Linear ramp and circular signals μ(s).

Fig. 4a: Signal 1. A linear ramp signal that increases from magnitude of 1–3 in the x-direction.

Fig. 4b: Signal 2. A circular signal with magnitude of 3 and radius of 30, centred within the region and convolved with a 3 voxel FWHM Gaussian kernel.

To each of these signals we added subject-specific Gaussian noise εi, also smoothed using a 3 voxel FWHM Gaussian kernel, with homogeneous and non-homogeneous variance structures: The first noise field had a spatially constant standard deviation of 1 (Fig. 5a), the second field had a linearly increasing standard deviation structure in the y-direction from 0.5 to 1.5 while remaining constant in the x-direction (Fig. 5b). Thus, the variance of this noise field spatially increased in the y-direction from 0.5 to 1.5 in a non-linear fashion.

Fig. 5.

Fig. 5

Stationary and non-stationary standard deviation fields of the noise εi(s).

Fig. 5a: Standard Deviation 1. Stationary variance of 1 across the region.

Fig. 5b: Standard Deviation 2. Non-stationary (linear ramp) standard deviation field increasing from 0.5 to 1.5 in the y-direction.

Altogether, the two underlying signals and two noise sources gave us four separate trials; across all of the simulations, we obtained Confidence Sets for the noise-free cluster Ac at a cluster-forming threshold of c=2.

3.3. 3D simulations

Four signal types μ(s) were considered to analyze performance of the method in three dimensions. The first three of these signals were generated synthetically on a cubic region of size 100×100×100: Firstly, a small spherical effect, created by placing a spherical phantom of magnitude 3 and radius 5 in the centre of the search region, which was then smoothed using a 3 voxel FWHM Gaussian kernel (Fig. 6a). Secondly, a larger spherical effect, generated identically to the first effect with the exception that the spherical phantom had a radius of 30 (Fig. 6b). Lastly, we created an effect by placing four spherical phantoms of magnitude 3 in the region of varying radii and then smoothing the entire image using a 3 voxel FWHM Gaussian (Fig. 6c). For each of these signals, the final image was re-scaled to have a maximum intensity of 3.

Fig. 6.

Fig. 6

The four 3D signal types μ(s), from top-to-bottom: small sphere, large sphere, multiple spheres, and the UK Biobank full mean image. Note that the colormap limits for the first three signal types are from 0 to 3, while the colormap limits for the UK Biobank mean image is from −0.4 to 0.5.

Similar to the two-dimensional simulations, for the three signals described above we added 3-voxel smoothed Gaussian noise of homogeneous and heterogeneous variance structures. The first noise field had a spatially constant standard deviation of 1, while the second field had a linearly increasing standard deviation in the z-direction from 0.5 to 1.5, while remaining constant in both the x- and y-directions. For all three effects, we obtained Confidence Sets for the threshold c=2.

For the final signal type, we took advantage of big data that has been made available through the UK Biobank in an attempt to generate an effect that replicated the true %BOLD change induced during an fMRI task. We randomly selected 4000 subject-level contrast of parameter estimate result maps from the Hariri Faces/Shapes task-fMRI data collected as part of the UK Biobank brain imaging study. Full details on how the data were acquired and processed is given in Miller et al. (2016), Alfaro-Almagro et al. (2018) and the UK Biobank Showcase; information on the task paradigm is given in Hariri et al. (2002). From these contrast maps, we computed a group-level full mean (Fig. 6d) and full standard deviation image. In the final simulation, we used the group-level Biobank mean image as the true underlying signal μ(s) for each subject, and the full standard deviation image was used for the standard deviation of each simulated subject-specific Gaussian noise field εi(s) added to the true signal. Because of the considerably large sample size of high-quality data from which these maps have been obtained, we anticipate that both of these images are highly representative of the true underlying fields that they approximate. Both images were masked using an intersection of all 4000 of the subject-level brain masks.

Once again, we smoothed the noise field using a 3 voxel FWHM Gaussian kernel; since the Biobank maps were written with voxel sizes of 2 mm3, this is analogous to applying 6 mm FWHM smoothing to the noise field of the original data. We obtained Confidence Sets for a threshold of c=0.25% BOLD change.

3.4. Application to Human Connectome project data

For a real-data demonstration of the method proposed here, we computed CSs on 80 participants data from the Unrelated 80 package released as part of the Human Connectome Project (HCP, S1200 Release). We applied the method to subject-level contrast maps obtained for the 2-back vs 0-back contrast from the working memory task results included with the dataset. To compare the CSs with results obtained from standard fMRI inference procedures, we also performed a traditional statistical group-level analysis on the data. A one-sample t-test was carried out in SPM, using a voxelwise FWE-corrected threshold of p<0.05 obtained via permutation test with SPM’s SnPM toolbox. We chose to use the HCP for its high-quality task-fMRI data, the working memory task specifically picked for its association with cognitive activations in subcortical networks that can not be distinguished by the anatomy. Full details of the task paradigm, scanning protocol and analysis pipeline are given in Barch et al. (2013) and Glasser et al. (2013), here we provide a brief overview.

For the working memory task participants were presented with pictures of places, tools, faces and body parts in a block design. The task consisted of two runs, where on each run a separate block was designated for each of the image categories, making four blocks in total. Within each run, for half of the blocks participants undertook a 2-back memory task, while for the other half a 0-back memory task was used. Eight EVs were included in the GLM for each combination of picture category and memory task (e.g. 2-back Place); we compute CSs on the subject-level contrast images for the 2-back vs 0-back contrast results that contrasted the four 2-back related EVs to the four 0-back EVs.

Imaging was conducted on a 3T Siemans Skyra scanner using a gradient-echo EPI sequence; TR = 720 ms, TE = 33.1 ms, 208×180 mm FOV, 2.0 mm slice thickness, 72 slices, 2.0 mm isotropic voxels, and a multi-band acceleration factor of 8. Preprocessing of the subject-level data was carried out using tools from FSL and Freesurfer following the ‘fMRIVolume’ HCP Pipeline fully described in Glasser et al. (2013). To summarize, the fundamental steps carried out to each individual’s functional 4D time-series data were gradient unwarping, motion correction, EPI distortion correction, registration of the functional data to the anatomy, non-linear registration to MNI space (using FSL’s Non-linear Image Registration Tool, FNIRT), and global intensity normalization. Spatial smoothing was applied using a Gaussian kernel with a 4 mm FWHM.

Modelling of the subject-level data was conducted with FSL’s FMRIB’s Improved Linear Model (FILM). The eight working task EVs were included in the GLM, with temporal derivatives terms added as confounds of no interest, and regressors were convolved using FSL’s default double-gamma hemodynamic response function. The functional data and GLM were temporally filtered with a high pass frequency cutoff point of 200s, and time series were prewhitened to remove autocorrelations from the data.

In comparison to a typical fMRI study, the 4 mm FWHM smoothing kernel size used in the HCP preprocessing pipeline is modest. Because of this, we applied additional smoothing to the final contrast images to emulate maps smoothed using a 6 mm FWHM Gaussian kernel.

4. Results

4.1. Methodological comparisons

In this work we have proposed two fundamental methodological changes to the procedures carried out in SSS: in Section 2.2 we suggested the Wild t-Bootstrap instead of the Gaussian Wild Bootstrap used for SSS, and in Section 2.4 we introduced the interpolation method for assessing empirical coverage alongside the direct comparison methods used for SSS. Here, we show the impact of these methodological innovations on the empirical coverage results from simulations carried out using two different synthetic signals, the 2D circular signal (Signal 2. in Fig. 4b) and the 3D large spherical signal (Signal 2. in Fig. 6). The standard deviation of the subject-specific Gaussian noise fields εi(s) had a stationary variance of 1 across the region in both simulations (for the 2D case, this corresponds to Standard Deviation 1. in Fig. 5).

Empirical coverage results for each of the three confidence levels 1α=0.80,0.90 and 0.95 are presented for the 2D circular signal in Fig. 7 and for the 3D large spherical signal in Fig. 8. In both simulations, for all methods the bootstrap procedure was carried out over the estimated boundary Aˆc (as must be done with real data). In each figure, the green curves highlight the results for the Gaussian Wild Bootstrap and coverage assessment method that were applied in SSS. The red curves highlight the results for the Wild t-Bootstrap and interpolation assessment method that we have proposed.

Fig. 7.

Fig. 7

Coverage results for the 2D circular signal simulation with homogeneous Gaussian noise (Signal 2., Standard deviation 1. in Fig. 5). Empirical coverage results are presented for implementations of the CS method with and without the Wild t-Bootstrap we propose in Section 2.2 and the interpolation schema for assessing simulations results we propose in Section 2.4. All empirical coverage results for simulations using the SSS assessment method are close to 100%, suggesting that this assessment substantially biases the results upwards. Using our proposed assessment method, while both the Wild t-Bootstrap and Gaussian Wild bootstrap converge to the nominal level, the Wild t-Bootstrap performed better for small sample sizes.

In Figs. 7 and 8, all simulations using the direct comparison assessment (SSS Simulation Assessment) produced results substantially above the nominal level, converging to almost 100% for both the Gaussian Wild Bootstrap (green curves) and Wild t-Bootstrap (blue curves) methods across all three confidence levels. We suspect this is due to the resolution issue described in Section 2.4, suggesting that this assessment method missed violations of the coverage condition Aˆc+AcAˆc causing a considerable positive bias in all of these results. Further evidence of this is suggested by the empirical coverage obtained for simulations using the interpolation assessment method (BTSN Simulation Assessment, pink and red curves), which appear to be converging much closer to the nominal level as is theoretically expected by Result 1.

Considering only the results using the interpolation assessment, in both figures empirical coverage for the Wild Bootstrap method (pink curves) came below the nominal level for small sample sizes. For the 2D circle simulation, the empirical coverage result for 60 subjects was 84.7% for the nominal target of 1α=0.95 (right plot in Fig. 7). For the 3D spherical simulation this under-coverage was even more severe, where the corresponding empirical coverage result was 54.9% (right plot in Fig. 8). In comparison, coverage performance for the Wild t-Bootstrap method (red curves) was much improved, staying close to the nominal level in both the 2D and 3D simulations across all sample sizes. While for the 3D spherical signal the empirical coverage remained slightly above the nominal target, for the circular signal almost all results lie within the 95% confidence interval of the nominal coverage level. For these reasons, in the remaining simulation results presented in this section we only consider the Wild t-Bootstrap method with our proposed interpolation assessment.

4.2. 2D simulations

Empirical coverage results for each of the three confidence levels 1α=0.80,0.90 and 0.95, are presented for the linear ramp signal (Signal 1. in Fig. 4a) in Fig. 9, and for the circular signal (Signal 2. in Fig. 4b) in Fig. 10. Results are also presented in tabular format in Table S1. In both plots, results obtained for simulations applying the bootstrap procedure over the estimated boundary Aˆc are displayed with a solid line, while results for simulations using the true boundary Ac are displayed with a dashed line. We emphasize that when computing CSs for real data, only the estimated boundary can be used.

Fig. 9.

Fig. 9

Coverage results for Signal 1., the 2D linear ramp signal. While the true boundary coverage results (dashed curves) fall under the nominal level, results for the estimated boundary method (solid curves) that must be applied to real data remain above the nominal level. Performance of the method improved for larger confidence levels, and in particular, the estimated boundary results for a 95% confidence level seen in the right plot hover slightly above nominal coverage for all sample sizes.

Fig. 10.

Fig. 10

Coverage results for Signal 2., the 2D circular signal. Coverage performance was close to nominal level in all simulations. The method was robust as to whether the subject-level noise had homogeneous (red curves) or heterogeneous variance (blue curves), or as to whether the estimated boundary (dashed curves) or true boundary (solid curves) method was used; in all plots, all of the curves lie practically on top of each other.

For the linear ramp, across all confidence levels we observed valid, over-coverage for the estimated boundary method, and under-coverage for the true boundary method. In both cases, the degree of agreement between our empirical results and the nominal coverage level improved for larger confidence levels, and as the sample size increased. For instance, while our estimated boundary empirical results were around 88% when the nominal target level was set at 80% (Fig. 9, left), corresponding empirical coverage results hovered around 97% for a nominal target of 95% (Fig. 9, right). Comparing the differences between the solid and dashed curves, there is also greater harmonization between the estimated and true boundary results for higher confidence levels. The method performed similarly regardless of whether homogeneous or heterogeneous noise was added to the model, evidenced by the minimal differences between the red and the blue curves for each of the two boundary methods seen in the plots.

For the circular signal the method performed remarkably well, with almost all our empirical coverage results lying within the 95% confidence interval of the nominal coverage rate (red and blue curves sandwiched between black dashed lines for all three plots in Fig. 10). Once again, the use of homogeneous or heterogeneous noise in the model had minimal difference on the method’s empirical coverage performance, and in this setting, our results were virtually identical whether the estimated boundary or true boundary was used for the bootstrap procedure. This has made the dashed curves hard to distinguish in the plots, as the solid curves lie practically on top of them.

4.3. 3D simulations

Empirical coverage results for each of the three confidence levels 1α=0.80,0.90 and 0.95, are presented in Fig. 11, Fig. 12, Fig. 13, Fig. 14 respectively for each of the four signal types (small sphere, large sphere, multiple spheres, Biobank full mean) displayed in Fig. 6. Results are also presented in tabular format in Table S2. Once again, results obtained for simulations applying the bootstrap procedure over the estimated boundary Aˆc are displayed with a solid line, and results for simulations using the true boundary Ac are displayed with a dashed line.

Fig. 11.

Fig. 11

Coverage results for Signal 1., the 3D small spherical signal. For all confidence levels, coverage remained above the nominal level in all simulations, and for a 95% confidence level (right plot), coverage hovered slightly above the nominal level for all sample sizes. The method was robust as to whether the subject-level noise had homogeneous (red curves) or heterogeneous variance (blue curves), or as to whether the estimated boundary (dashed curves) or true boundary (solid curves) method was used.

Fig. 12.

Fig. 12

Coverage results for Signal 2., the large 3D spherical signal. Coverage results here were very similar to the results for the small spherical signal shown in Fig. 11, suggesting that the method is robust to changes in boundary length.

Fig. 13.

Fig. 13

Coverage results for Signal 3., the multiple spheres signal. Once again, for all confidence levels, coverage remained above the nominal level in all simulations. Here, the true boundary method (dashed curves) performed slightly better than the estimated boundary method (solid curves) in small sample sizes, although the choice of boundary made less of a difference for a higher confidence level. For a 95% confidence level (right plot), all results hover slightly above nominal coverage for all sample sizes.

Fig. 14.

Fig. 14

Coverage results for Signal 4., the UK Biobank full mean signal, where the full standard deviation image was used as the standard deviation of the subject-level noise fields. Coverage results here were similar to the results for the multiple spheres signal shown in Fig. 13; in small sample sizes, coverage was slightly improved for the true boundary method (dashed curves) compared to the estimated boundary method (solid curves), however, for a 95% confidence level (right plots), all results hover slightly above nominal coverage for all sample sizes.

Overall, the results for all four signal types were consistent: In general, empirical coverage always came above the nominal target level, and the extent of over-coverage diminished when a higher confidence level was used. Particularly, for a nominal target of 1α=0.95, all of our 3D empirical coverage results lie between 95% and 98%. The method was robust as to whether the bootstrap procedure was applied over the true or estimated boundary, or as to whether the variance of the noise field was homo- or heterogeneous. The similarity of the empirical coverage results, in spite of differences in these specific settings, is exhibited in all of the plots by the uniformity of the red and blues curves (indicating minimal differences in performance whether the noise had homogeneous or heterogeneous variance), and agreement between the solid and dashed curves (indicating minimal differences in performance whether the true boundary or estimated boundary was used). In the empirical coverage plots for the small and large spherical signals shown in Figs. 11 and 12, all of these curves lie virtually on top of each other.

While performance with the multiple spheres and Biobank signals presented in Figs. 13 and 14 was slightly better when using the true boundary, the true- and estimated boundary performance converged as the sample size increased.

4.4. Human Connectome Project

Confidence Sets obtained from applying the method to 80 subjects contrast data from the Human Connectome Project working memory task are shown in Fig. 15 and Fig. 16.

Fig. 15.

Fig. 15

Slice views of the Confidence Sets for 80 subjects data from the HCP working memory task for c=1.0%,1.5% and 2.0% BOLD change thresholds. The upper CS Aˆc+ is displayed in red, and the lower CS Aˆc displayed in blue. In yellow is the point estimate set Aˆc, the best guess from the data of voxels that surpassed the BOLD change threshold. The red upper CS has localized regions in the frontal gyrus, frontal pole, anterior insula, supramarginal gyrus and cerebellum for which we can assert with 95% confidence that there has been (at least) a 1.0% BOLD change raw effect.

Fig. 16.

Fig. 16

Further slice views of the Confidence Sets. Here, we see that the red upper CS has also localized regions in the anterior cingulate, superior front gyrus, supramarginal gyrus, and precuneous for which we can assert with 95% confidence that there has been (at least) a 1.0% BOLD change raw effect.

In both Figs. 15 and 16, the red upper CS localized brain regions within the frontal cortex commonly associated to working memory. This included areas of the middle frontal gyrus (left and right; Fig. 15, sagittal and coronal slices), superior frontal gyrus (left and right, Fig. 16, coronal slice) anterior insula (left and right; Fig. 15, sagittal and axial slices), as well as the anterior cingulate (Fig. 16, all slices). In all of the above regions, the method identified clusters of voxels for which we can assert with 95% confidence there was a percentage BOLD change raw effect greater than 2.0% (Figs. 15 and 16, bottom plots).

Further brain areas localized by the upper CS were the frontal pole (left and right; Fig. 15, sagittal and axial slices), supramarginal gyrus (left and right; Fig. 15, sagittal slice and Fig. 16, coronal and axial slices), precuneous (Fig. 16, sagittal slice) and cerebellum (Fig. 15, sagittal slice). While for these areas we can assert with 95% confidence there was a percentage BOLD change raw effect greater than at least 1.0% (Figs. 15 and 16, top plots), on-the-whole the method only localized areas where there was a BOLD change of at least 2.0% in parts of the frontal cortex. This can be observed by the ‘disappearance’ of the red CSs in brain regions located in the ocipital lobe for the 2.0% BOLD change plots when compared with the corresponding 1.0% and 1.5% BOLD change plots in Figs. 15 and 16.

As the percentage BOLD change threshold increases between plots, there is a shrinking of both the blue lower CSs and red upper CSs. By using a larger threshold, there are less voxels we can confidently declare have surpassed this higher level of percentage BOLD change, and thus the volume of the red upper CSs decreases (in some cases, vanishing). At the same time, there are more voxels we expect to be able to confidently declare have fallen below the threshold. Since these are precisely the (grey background) voxels that lie outside of the lower blue CSs, the volume of the blue lower CSs also decreases.

Finally, in Fig. S1 and Fig. S2 the red upper CSs are compared with the thresholded t-statistic map (green-yellow voxels) obtained from applying a traditional one-sample t-test group-analysis to the 80 subjects working memory task contrast data, using a voxelwise FWE-corrected threshold of p<0.05. Differences here highlight how statistical significance may not translate to practical significance; while over 28,000 voxels were declared as active in the thresholded t-statistic results, only 4,818 voxels were contained in the upper CS indicating a percentage BOLD change of at least 1.0%.

5. Discussion

5.1. Spatial inference on %BOLD raw effect size

Thorough interpretation of neuroimaging results requires an appreciation of the practical (as well as statistical) significance of differences through visualization of raw effect magnitude maps with meaningful units (Chen et al., 2017). In this work, we have presented a method to create Confidence Sets for raw effect size maps, providing formal confidence statements on regions of the brain where the %BOLD response magnitude has exceeded a specified activation threshold, alongside regions where the %BOLD response has not surpassed this threshold. Both of these statements are made simultaneously, and across the entire brain. This not only enables researchers to infer brain areas that have responded to a task, but also allows for inference on areas that did not respond to the task. In this sense, the method goes beyond statistical hypothesis testing, where the null-hypothesis of no activation can ‘fail to be rejected’, but never accepted. By operating on percentage BOLD change units, instead of t-statistic values, the confidence set maps present a clear and more direct interpretation of the biophysical changes that occur during a neuroimaging study, which can be distorted by the thresholded statistic maps commonly reported at the end of an investigation (Engel and Burton, 2013). In essence, the CSs synthesize information that is usually provided separately in a raw effect size and t-statistic map, determining practically significant effects in terms of effect magnitude, that are also reliable in terms of statistical significance traditionally given by p-values in a statistic image. While in this work we have focused on BOLD fMRI, the methods presented here are applicable to any neuroimaging measure that can be fit in a group-level GLM.

The choice of threshold c is ultimately up to the user, and may depend on the aims of the investigation. Researchers may choose a threshold based on prior knowledge of raw effect sizes observed in previous similar studies, and it is likely that localization of larger raw effects will be possible as sample sizes increase. Obtaining the CSs for the Human Connectome Project contrast data in this work was computationally quick, each analysis taking no longer than a couple of minutes. Therefore, one possible strategy is to evaluate a variety of different c’s on pilot or historical data before fixing a value to use on a study of interest.

5.2. Analysis of HCP data and simulation results

In our analysis of the HCP working memory task-fMRI dataset, we have primarily focused on activated areas localized by the red upper CS. However, the confidence set maps in Figs. 15 and 16 also quantify the spatial precision of the point estimate ‘best guess from the data’ activation clusters. While so far we have described the confidence sets in terms of the red and blue upper and lower CSs, we now highlight that the set difference between the upper and lowers CSs acts as a confidence region itself; with 95% confidence, we can assert that the boundary of the point estimate set (raw effect size > threshold; yellow voxels overlapped by red in Figs. 15 and 16) is completely contained within this region. The set difference region, visualized by blue and yellow voxels (but not red) in Figs. 15 and 16, therefore anticipates how the point estimate clusters may fluctuate if the experiment was to be repeated again. From this perspective, the vast areas of the brain covered by blue in Figs. 15 and 16 demonstrate the high level of uncertainty in localizing a raw effect size of, for example, 1.0% BOLD change, despite the large sample size of N=80 used for the HCP. The regions of greatest uncertainty were sub-cortical areas, covered by expansive clusters of blue as seen in the axial slices displayed in Fig. 15 and sagittal slices in Fig. 16. Large intersubject variability here may be explained by the high multi-band acceleration factor used in the HCP scanning protocol, which is generally more suited for scanning the cortex (Smith et al., 2013).

'For the 2D simulations, the method achieved close to nominal coverage for the circular signal, but performed less well for the ramp signal, obtaining under-coverage for the true boundary method and over-coverage for the estimated boundary method. We believe differences in the circle and ramp results are not due to changes in the signal shape per se, but instead are caused by differences in the slope of each shape close to the true boundary Ac. Since the linear ramp signal has a shallower gradient at the true boundary compared to the circle, local changes in the observed signal around the boundary are dominated by changes in the noise. Since the noise is more wavey than the signal, the linear interpolation method for obtaining the boundary is likely to be less accurate for the ramp, causing too many violations of the subset condition, which may explain the under-coverage for the true boundary results seen here.

For the 3D simulations, the method obtained over-coverage in all of our results. Here, the degree of over-coverage was consistently larger for the smaller confidence level of 1α=0.80 in comparison to the larger confidence level of 1α=0.95. Notably, the over-coverage was also more severe for signals with a longer boundary, such as the multiple spheres and Biobank signals, when compared to the Small Sphere signal that had a shorter boundary length. One possible reason for this is that our proposed method for assessing coverage may still be missing instances where violations of the subset condition Aˆc+AcAˆc occur, causing the results to be slightly positively biased. While our assessment method reduces the influence of grid coarseness by sampling locations on the true continuous boundary Ac, ultimately we can still only assess coverage at a discrete set of points on a continuous process. For signals with a longer boundary length, the set of sampled locations obtained with the interpolation method is relatively less dense within the true continuous boundary, and thus it is more likely violations of the subset condition are missed. Over-coverage for smaller confidence levels may also be explained by this, as theoretically more violations should occur here, but these may be missed due to inaccuracies caused by the discreteness of the lattice. This line of reasoning is consistent with Section 4.4 of SSS, where it was shown that coverage approached the nominal level as the resolution of the grid was increased.

5.3. Methodological innovations

In this work, we have advanced on the original methods applied in SSS. From a theoretical standpoint, we have proposed a Wild t-Bootstrap method (dividing bootstrap residuals by bootstrap standard deviation) to compute the critical quantile value k. We have also introduced an interpolation scheme for obtaining the boundary and assessing the simulation coverage results to reduce the influence of grid coarseness. In Section 4.1, we demonstrated that applying the assessment method in SSS could lead to empirical coverage of close to 100%, suggesting that this method may considerably bias the simulation results upwards. When using our proposed assessment, the Wild Bootstrap method suffered from under-coverage, most severely for small sample sizes in the 3D setting of the large spherical signal presented in Fig. 8. This was greatly remedied by the Wild t-Bootstrap method, for which empirical results stayed close to the nominal target independent of sample size.

Our simulations using the original procedures may not seem consistent with the simulation results published in Fig. 5 of SSS, where empirical coverage stayed close to the nominal target. However, the signal-plus-noise models investigated to test the performance of the CSs in SSS were much smoother than the synthetic signals considered to emulate fMRI data with this effort. By applying a larger degree of smoothing, the signals used in SSS effectively had a much higher resolution. Because of this, it is likely the resolution issue presented in Fig. 3 was less critical, reducing the positive bias in empirical coverage induced from using the original simulation assessment procedure. Further evidence for this is provided in Fig. 7 of SSS, where they observed an increase in coverage after repeating their simulations on a coarser lattice. In our simulation results in Section 4.1, the scale of under-coverage from using the Gaussian Wild bootstrap method was much more severe for the 3D simulation on the spherical signal in Fig. 8 compared to the 2D circular signal in Fig. 7. This may explain why the Gaussian Wild bootstrap method performed relatively well in SSS, as only 2D signals were considered there.

5.4. Limitations & future work

The principal limitation of this work is one that is intentional and explicit: Our method is for spatial inference on maps of raw and not standardized effects, such as Cohen’s d or partial R2 (t- or F-statistics, which scale with sample size, do not estimate population quantities and are not suitable for making confidence statements). Even when scaled to percentage BOLD change, it has been shown that raw effects can modulate with acquisition parameters such as the scanner field strength or echo time (UIudag et al., 2009). Users should therefore be cautious when combining effect estimates from studies using heterogeneous acquisition setups, and clearly specify such differences when reporting the results of any meta-analysis on raw effects. It is also known that inhomogeneities in the vasculature of the brain is a cause of variation in the BOLD response. Therefore, we recommend that any interpretation of %BOLD change inferred from the CSs is referenced against a variance map or similar image that indicates the most venous brain regions. We note that each of these points are general complications of raw effect sizes within fMRI, rather than issues with the method proposed in this effort per se. Nonetheless, the use of standardized effect estimates may help to remedy these problems in the future. The statistical characteristics of standardized effect maps are fundamentally different to the raw effect images motivating the method here, and the topic of our current work is to develop CSs for standardized effect size images.

The need for resampling to conduct inference is another limitation of this effort, especially given the big data motivation of this work. However, the bootstrap is only conducted on the estimated boundary, Aˆc, not the whole 3D volume, which substantially reduces the computational burden. For very large datasets, techniques for approximating empirical distributions can be used to improve the accuracy of the estimation of k based on a smaller number (e.g. B=500) of bootstrap samples (Winkler et al., 2016).

6. Contributions

A.B. drafted the manuscript. All authors edited and revised the manuscript. A.B. proposed use of Rademacher variables for the bootstrap and developed the method for assessing simulations. F.T. proposed use of the Wild t-Bootstrap method and developed the linear interpolation method for approximating the boundary. A.B. conducted simulations and data analysis, with additional contributions from F.T. to the simulation and simulation figures code. A.S. and T.E.N. oversaw the project in all its intellectual, methodological and computational aspects.

Data availability

We have used data from The Human Connectome Project and UK Biobank. All code used for the simulations and analysis of HCP data are available at: https://github.com/AlexBowring/Spatial_Confidence_Sets_Raw.

Acknowledgements

TEN is supported by the Wellcome Trust (100309/Z/12/Z). AS and FT were partially supported by NIH grants (R01CA157528 and R01EB026859). The authors would like to thank Max Sommerfeld for his early involvement in the project, and Samuel Davenport for helping to process the Biobank Data used in the Simulations section.

Footnotes

1

For examples of how to set up more complex designs and contrasts, see Figure A.2. in Appendix A section of (Poldrack et al., 2011).

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.neuroimage.2019.116187.

Appendix A. Supplementary data

The following is the Supplementary data to this article:

Multimedia component 1
mmc1.pdf (3.7MB, pdf)

References

  1. Hariri Ahmad R., Tessitore Alessandro, Mattay Venkata S., Fera Francesco, Weinberger Daniel R. The amygdala response to emotional stimuli: a comparison of faces and scenes. Neuroimage. September 2002;17(1):317–323. doi: 10.1006/nimg.2002.1179. [DOI] [PubMed] [Google Scholar]
  2. Alfaro-Almagro Fidel, Jenkinson Mark, Bangerter Neal K., Andersson Jesper L.R., Griffanti Ludovica, Douaud Gwenaëlle, Sotiropoulos Stamatios N., Jbabdi Saad, Hernandez-Fernandez Moises, Vallee Emmanuel, Vidaurre Diego, Webster Matthew, McCarthy Paul, Rorden Christopher, Daducci Alessandro, Alexander Daniel C., Zhang Hui, Dragonu Iulius, Matthews Paul M., Miller Karla L., Smith Stephen M. Image processing and quality control for the first 10,000 brain imaging datasets from UK biobank. Neuroimage. February 2018;166:400–424. doi: 10.1016/j.neuroimage.2017.10.034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Barch Deanna M., Burgess Gregory C., Harms Michael P., Petersen Steven E., Schlaggar Bradley L., Corbetta Maurizio, Glasser Matthew F., Curtiss Sandra, Dixit Sachin, Feldt Cindy, Nolan Dan, Bryant Edward, Tucker Hartley, Owen Footer, Bjork James M., Poldrack Russ, Smith Steve, Johansen-Berg Heidi, Snyder Abraham Z., Van Essen David C., WU-Minn HCP Consortium Function in the human connectome: task-fMRI and individual differences in behavior. Neuroimage. October 2013;80:169–189. doi: 10.1016/j.neuroimage.2013.05.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Chen Gang, Taylor Paul A., Cox Robert W. Is the statistic value all we should care about in neuroimaging? Neuroimage. February 2017;147:952–959. doi: 10.1016/j.neuroimage.2016.09.066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Chernozhukov Victor, Chetverikov Denis, Kato Kengo. 2013. Gaussian Approximations and Multiplier Bootstrap for Maxima of Sums of High-Dimensional Random Vectors. [Google Scholar]
  6. Davidson Russell, Flachaire Emmanuel. The wild bootstrap, tamed at last. J. Econom. September 2008;146(1):162–169. [Google Scholar]
  7. Engel Stephen A., Burton Philip C. Confidence intervals for fMRI activation maps. PLoS One. December 2013;8(12) doi: 10.1371/journal.pone.0082419. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Friston K.J., Holmes A.P., Worsley K.J., Poline J.P., Frith C.D., Frackowiak R.S.J. Statistical parametric maps in functional imaging: a general linear approach. Hum. Brain Mapp. 1994;2(4):189–210. [Google Scholar]
  9. Friston K.J., Worsley K.J., Frackowiak R.S., Mazziotta J.C., Evans A.C. Assessing the significance of focal activations using their spatial extent. Hum. Brain Mapp. 1994;1(3):210–220. doi: 10.1002/hbm.460010306. [DOI] [PubMed] [Google Scholar]
  10. Glasser Matthew F., Sotiropoulos Stamatios N., Wilson J Anthony, Coalson Timothy S., Fischl Bruce, Andersson Jesper L., Xu Junqian, Jbabdi Saad, Webster Matthew, Polimeni Jonathan R., Van Essen David C., Jenkinson Mark, WU-Minn HCP Consortium The minimal preprocessing pipelines for the human connectome project. Neuroimage. October 2013;80:105–124. doi: 10.1016/j.neuroimage.2013.04.127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Gonzalez-Castillo Javier, Saad Ziad S., Handwerker Daniel A., Inati Souheil J., Brenowitz Noah, Bandettini Peter A. Whole-brain, time-locked activation with simple tasks revealed using massive averaging and model-free analysis. Proc. Natl. Acad. Sci. U.S.A. April 2012;109(14):5487–5492. doi: 10.1073/pnas.1121049109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Cremers Henk R., Wager Tor D., Yarkoni Tal. The relation between statistical power and inference in fMRI. PLoS One. November 2017;12(11) doi: 10.1371/journal.pone.0184923. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Meehl Paul E. Theory-Testing in psychology and physics: a methodological paradox. Philos. Sci. June 1967;34(2):103–115. [Google Scholar]
  14. Miller Karla L., Alfaro-Almagro Fidel, Bangerter Neal K., Thomas David L., Yacoub Essa, Xu Junqian, Bartsch Andreas J., Jbabdi Saad, Sotiropoulos Stamatios N., Andersson Jesper L.R., Griffanti Ludovica, Douaud Gwenaëlle, Okell Thomas W., Weale Peter, Dragonu Iulius, Garratt Steve, Hudson Sarah, Collins Rory, Jenkinson Mark, Matthews Paul M., Smith Stephen M. Multimodal population brain imaging in the UK biobank prospective epidemiological study. Nat. Neurosci. November 2016;19(11):1523–1536. doi: 10.1038/nn.4393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Nichols Thomas E., Das Samir, Eickhoff Simon B., Evans Alan C., Glatard Tristan, Hanke Michael, Kriegeskorte Nikolaus, Milham Michael P., Poldrack Russell A., Poline Jean-Baptiste, Proal Erika, Bertrand Thirion, Van Essen David C., White Tonya, Thomas Yeo B.T. Best practices in data analysis and sharing in neuroimaging using MRI. Nat. Neurosci. February 2017;20(3):299–303. doi: 10.1038/nn.4500. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Nuzzo Regina. Scientific method: statistical errors. Nature. February 2014;506(7487):150–152. doi: 10.1038/506150a. [DOI] [PubMed] [Google Scholar]
  17. Poldrack Russell A., Mumford Jeanette A., Nichols Thomas E. Cambridge University Press; August 2011. Handbook of Functional MRI Data Analysis. [Google Scholar]
  18. Poldrack Russell A., Baker Chris I., Durnez Joke, Gorgolewski Krzysztof J., Matthews Paul M., Munafò Marcus R., Nichols Thomas E., Poline Jean-Baptiste, Vul Edward, Yarkoni Tal. Scanning the horizon: towards transparent and reproducible neuroimaging research. Nat. Rev. Neurosci. February 2017;18(2):115–126. doi: 10.1038/nrn.2016.167. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Rosenblatt Jonathan D., Finos Livio, Weeda Wouter D., Solari Aldo, Goeman Jelle J. All-Resolutions inference for brain imaging. Neuroimage. November 2018;181:786–796. doi: 10.1016/j.neuroimage.2018.07.060. [DOI] [PubMed] [Google Scholar]
  20. Rozeboom William W. The fallacy of the null-hypothesis significance test. Psychol. Bull. 1960;57(5):416–428. doi: 10.1037/h0042040. [DOI] [PubMed] [Google Scholar]
  21. Smith Stephen M., Beckmann Christian F., Andersson Jesper, Auerbach Edward J., Bijsterbosch Janine, Douaud Gwenaëlle, Duff Eugene, Feinberg David A., Griffanti Ludovica, Harms Michael P., Kelly Michael, Laumann Timothy, Miller Karla L., Moeller Steen, Petersen Steve, Power Jonathan, Salimi-Khorshidi Gholamreza, Snyder Abraham Z., Vu An T., Woolrich Mark W., Xu Junqian, Yacoub Essa, Uğurbil Kamil, Van Essen David C., Glasser Matthew F., WU-Minn HCP Consortium Resting-state fMRI in the human connectome project. Neuroimage. October 2013;80:144–168. doi: 10.1016/j.neuroimage.2013.05.039. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Sommerfeld Max, Sain Stephan, Schwartzman Armin. Confidence regions for spatial excursion sets from repeated random field observations, with an application to climate. J. Am. Stat. Assoc. July 2018;113(523):1327–1340. doi: 10.1080/01621459.2017.1341838. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Telschow Fabian J.E., Schwartzman Armin. January 2019. Simultaneous Confidence Bands for Functional Data Using the Gaussian Kinematic Formula. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Turner Benjamin O., Paul Erick J., Miller Michael B., Barbey Aron K. Small sample sizes reduce the replicability of task-based fMRI studies. Commun. Biol. June 2018;1:62. doi: 10.1038/s42003-018-0073-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. UIudag K., Müller-Bierl B., Ugurbil K. 2009. An Integrative Model for Neuronal Activity-Induced Signal Changes for Gradient and Spin Echo Functional Imaging. [DOI] [PubMed] [Google Scholar]
  26. Wasserstein Ronald L., Lazar Nicole A., Others The ASA’s statement on p-values: context, process, and purpose. Am. Stat. 2016;70(2):129–133. [Google Scholar]
  27. Winkler Anderson M., Ridgway Gerard R., Douaud Gwenaëlle, Nichols Thomas E., Smith Stephen M. Faster permutation inference in brain imaging. Neuroimage. November 2016;141:502–516. doi: 10.1016/j.neuroimage.2016.05.068. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Woo Choong-Wan, Krishnan Anjali, Wager Tor D. Cluster-extent based thresholding in fMRI analyses: pitfalls and recommendations. Neuroimage. May 2014;91:412–419. doi: 10.1016/j.neuroimage.2013.12.058. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia component 1
mmc1.pdf (3.7MB, pdf)

Data Availability Statement

We have used data from The Human Connectome Project and UK Biobank. All code used for the simulations and analysis of HCP data are available at: https://github.com/AlexBowring/Spatial_Confidence_Sets_Raw.

RESOURCES