Abstract
High angular resolution diffusion imaging (HARDI) is a type of diffusion magnetic resonance imaging (dMRI) that measures diffusion signals on a sphere in q-space. It has been widely used in data acquisition for human brain structural connectome analysis. To more accurately estimate the structural connectome, dense samples in q-space are often acquired, potentially resulting in long scanning times and logistical challenges. This paper proposes a statistical method to select q-space directions optimally and estimate the local diffusion function from sparse observations. The proposed approach leverages relevant historical dMRI data to calculate a prior distribution to characterize local diffusion variability in each voxel in a template space. For a new subject to be scanned, the priors are mapped into the subject-specific coordinate and used to help select the best q-space samples. Simulation studies demonstrate big advantages over the existing HARDI sampling and analysis framework. We also applied the proposed method to the Human Connectome Project data and a dataset of aging adults with mild cognitive impairment. The results indicate that with very few q-space samples (e.g., 15 or 20), we can recover structural brain networks comparable to the ones estimated from 60 or more diffusion directions with the existing methods.
Keywords: Diffusion MRI, Diffusion vector selection, Sparse q-space samples, Structural connectome
I. INTRODUCTION
Diffusion Magnetic Resonance Imaging (dMRI) is an advanced MRI methodology that measures the probabilistic motion of water molecules in order to probe tissue microstructure [1]. The earliest applications of dMRI focus on studying the anisotropic diffusion of water in the white matter (WM) using the diffusion tensor imaging (DTI) model [2]. However, DTI has significant limitations, since it can only describe one major fiber direction within each voxel [3], presenting a significant obstacle for efforts to trace WM pathways in crossing fiber regions [4]. To overcome this limitation, advanced acquisition methods, such as high angular resolution diffusion imaging (HARDI) [5], have been proposed to better characterize the underlying WM structure by estimating an object known as the diffusion orientation distribution function (ODF), denoted here as f, or its sharper version, the fiber ODF (fODF) obtained by deconvolving the ODF [6]. To estimate the structural connectome, tractography algorithms [7, 8] are applied to the estimated fODF field to compute streamlines connecting different brain regions. Therefore, accurate estimation of the fODF is paramount for performing quality structural connectome analysis.
With HARDI data, the ODF or fODF can be estimated from the diffusion signal using the Funk-Radon transform (FRT) [9], making it the most popular dMRI acquisition framework for structural connectome analysis. Commonly, HARDI data are acquired using uniformly distributed diffusion gradient vectors densely sampled on spheres, known as shells, which inhabit q-space. The distribution of these gradient vectors has been a subject of study. Frequently an electrostatic repulsion model is used to generate them [10]. As shown in this paper, such gradient setting can be sub-optimal with respect to recovering the diffusion signal and fODF. Moreover, in practice, scans are often time limited and, consequently, parameters are altered for fast acquisition, e.g., only 15 or less directions are sampled in some studies [11, 12]. Therefore, there are important questions to consider in HARDI: if we are restricted to a limited number of gradient directions to sample, where in q-space should we sample, and given the sparse samples, how do we recover the fODF for connectome analysis? In this paper we address these questions by (a) optimizing the diffusion sampling scheme for each individual’s brain, and (b) estimating diffusion signal from sparse samples.
The methods proposed in this paper leverage the assumption that high-quality dMRI data are available for the population of interest, e.g. healthy subjects in a certain age range. For simplicity, we only consider diffusion signal on a single shell. The diffusion signal at the v-th voxel can be considered as a realization of a random function , i.e., a measurable function with realizations in the space of square integrable functions over the 2-sphere. In particular, we assume the distribution of Xv can be modeled as a Gaussian process. Note that Gaussian process models have previously been used to model the diffusion signal in dMRI [13]. Employing an empirical Bayesian approach, the historical dMRI data is used to estimate the mean and covariance functions defining Xv, characterizing the prior distribution of the diffusion signal across the brain. This prior is used to derive an estimator for a new subject’s diffusion signal in a manner that is robust even in the sparsely sampled case. Additionally, we propose a greedy diffusion direction selection algorithm to sequentially select the best directions to acquire dMRI data. The proposed method shows significant advantages over the current dMRI methods. In particular, we demonstrate that our method can produce a much better structural connectome reconstruction with sparse diffusion samples.
Incorporating prior information in the design of q-space sampling has been investigated before. For example, [14] proposes an optimal sampling scheme in the context of DTI and historical data are used to estimate the probability density function of the diffusion tensor field. [15] uses diffusion kurtosis imaging (DKI) to form a parametric approximation of the diffusion signal and prior information is incorporated via the DKI parameters estimated from historical subjects. These methods are derived in the context of finite dimensional parametric models for diffusion signal estimation (i.e., DTI and DKI) and can be problematic for connectome analysis, owing to their assumptions of the underlying diffusion process.
In our framework, we consider the diffusion signal as a random function in which is capable of representing a far richer class of local diffusion patterns than DTI or DKI. Accordingly, the optimal q-space acquisition problem for the reconstruction of the diffusion signal at a particular voxel is best formulated as a problem of optimal experimental design for sampling and recovering functional data in . Although there are a few existing papers concerning optimal experimental design for functional data, most of them deal with functions on [16, 17]. Our approach makes repeated use of special properties of , especially to incorporate relevant prior information, thus distinguishing our problem from these existing ones. We summarize our contributions as follows:
The diffusion signal is modeled as a random function in , allowing for more flexibility in the types of signals that can be accommodated than do parametric models such as DTI and DKI. This is particularly important for modeling complex diffusion signals in fiber crossing regions. To our knowledge, this is the first work for optimal design in dMRI which both incorporates historical data for constructing priors and does not induce a fixed finite dimensional parametric model for the diffusion signal.
The proposed method uses historical data (e.g., data from large brain imaging studies such as the Human Connectome Project (HCP) [18]) to construct a spatially varying prior distribution of diffusion signals over all voxels in a template space. These template space priors can be ported to the coordinate space of a new subject of interest to facilitate real-time optimal sampling design and sparse sample signal reconstruction, amongst other dMRI tasks of interest.
We propose a computationally efficient approximate optimal design algorithm that avoids costly high-dimensional optimization or stochastic techniques such as simulated annealing. Further, we give a theoretical performance bound between our approximate solution and the true optimal design.
The rest of the paper is organized as follows. In Section II we present the proposed methodology. Section III shows the results of a simulation study comparing the performance of our method with existing ones. In Section IV, we show real data examples for several tasks related to dMRI processing and structural connectome analyses. Conclusions and future directions are presented in Section V.
II. METHODS
A. Statistical Model and ODF Estimation
Under the HARDI framework, for any subject, the diffusion signal at voxel v at q-space direction is modeled as a noisy observation from a voxel specific random function
| (1) |
where the random function Xv is assumed to be a Gaussian process (GP) with voxel specific mean and covariance function and , respectively, which we will denote as GP(μv, Cv). Xv describes the distribution of diffusion signal in the local brain region v for a population of interest, e.g. healthy adults from age 20–35. The measurement error models the noise introduced by the scanner and is assumed to be independent of diffusion direction pm and brain region v. Although measurement error in MRI generally follows a Rician distribution, it is not uncommon to use GP models to model the diffusion signal [13]. In this work, we consider the case of single shell acquisition, i.e., all diffusion data is collected using a single b-value. For notational clarity, we denote Sv(pm) = Svm.
The primary object of interest in tractography, the ODF, can be computed from the diffusion signal on a single sphere in q-space by the Funk-Radon transform (FRT):
where is the FRT, δ is the Dirac delta function and fv is the ODF [9]. The authors of [19] show the Funk-Hecke theorem (FKT) implies that the ODF can be constructed by a simple linear transformation applied to the coefficients of the basis expansion of the diffusion signal with respect to a set of modified spherical harmonic functions. As this representation has a tendency to be overly smooth, the FKT can be applied again to obtain a linear transformation representing a deconvolution operator to obtain the fODF [6]. Therefore, we focus on optimizing the diffusion signal estimation with respect to the spherical harmonics, denoted as {ϕ1, …, ϕJ}, as this will result in optimal (f)ODF reconstruction, owing to their simple linear relationship.
B. Sparse Sample Diffusion Signal Estimation
In this work, we are concerned with selecting the optimal q-space directions for the estimation of the diffusion signal function under a sparsity constraint on the total number of samples, i.e., we collect data in the form {(p1, Sv1), …, (pM, SvM))} with M as a small integer. Let the vector sM = (Sv1, …, SvM)T denote the observed diffusion signals and let PM = (p1, …, pM)T be the corresponding gradient directions in q-space.
1). Optimal Reduced Rank Basis:
As discussed, to facilitate future connectome extraction it is important to obtain a continuous representation of the estimated diffusion function with respect to basis functions {ϕ1, …, ϕJ}. However, using the standard regularized least squares based estimation of the coefficients of ϕj, e.g. from [19], with sparse and irregular observations can be problematic, as illustrated in [20] where a reduced rank solution is promoted. Therefore, we propose a low rank estimator which is robust to the case of sparse samples.
By the Karhunen-Loéve theorem, the optimal rank K basis system for representing the mean centered random function Xv − μv, in terms of minimum expected integrated squared error, is given by the first K eigenfunctions of the covariance function Cv, denoted as {ψv1, …, ψvK}, which are defined by the integral equation
where ρvk ⩾ 0 is the eigenvalue associated with the k’th eigenfunction ψvk(p). Due to this optimality property, the eigenfunctions are a good candidate for sparse sample modeling. Furthermore, if we impose the restriction that ψvk lies in the linear span of {ϕ1, …, ϕJ}, then there exists a simple linear relationship between the ψvk’s and ϕj’s, and thus we can still leverage the analytic FRT for ODF estimation. This restriction translates the infinite dimensional integral equation above to the following finite dimensional eigenvalue problem:
with orthonormality restriction , where the eigenfunctions are approximated by (bvkj is the j-th element in bvk) and Σv is the covariance matrix defined according to
| (2) |
where ϕ(p) = (ϕ1(p), …, ϕJ(p))T.
2). Sparse Estimator:
We now present the proposed estimator for the diffusion signal function given sparse q-space samples. For notational convenience, we temporarily drop the subscript v, with the understanding that the estimates are formed per voxel. From the discussion above, we would like to obtain an estimate of X in the form , where are the coefficients of the diffusion signal associated with the k’th eigenfunction. Due to the assumption of sparsity in the current set-up, estimating these coefficients directly using numerical quadrature is problematic. Following [21], which proposes an estimator in the setting of sparsely observed longitudinal data, a stable estimate of ξk can be constructed by deriving the conditional distribution of P(ξK|sM, PM), where ξK = (ξ1, …, ξK)T.
Given the model for the observed data (1) and Gaussian assumptions, it is easy to show that the joint distribution on (ξK, sM)|PM is normal with mean vector (0K, μM)T and covariance matrix
where μM = (μ(p1), …, μ(pM))T, ΛK = Diag(ρ1, …, ρK), and such that [ΨM,K]mk = ψk(pm) for m = 1, …, M and k = 1, …, K. Therefore, P(ξK|sM, PM) is a normal distribution with mean
| (3) |
where IM is the M × M identity matrix. Denote the vector of eigenfunctions ψK(p) = (ψ1(p), …, ψK(p))T, based on , we propose to estimate X using it’s conditional mean:
| (4) |
where is the vector of coefficients defining the approximation of the mean function μ over the spherical harmonic basis {ϕ1, …, ϕJ} and is the matrix whose columns are the bk’s.
It is notable that (3) is both the best estimate of the coefficients under the Gaussian assumption and the best linear predictor even when the Gaussian assumption is violated. In theory, this provides robustness to some deviations from the Gaussian assumption, a property that is verified in our simulations. Alternatively, the derivation above can be motivated through a fully Bayesian framework under the prior that X ~ GP(μ, C), or equivalently , where it can be shown that the mean of the posterior predictive distribution for X(p) is exactly (4). Further consideration of these points is provided in the supplemental material.
C. Optimal q-Space Sampling
Given the estimator (4), we formulate the diffusion sampling, the process of obtaining sM in q-space, as an optimization problem: for a fixed total budget of M samples, our goal is to select a set of directions PM = (p1, …, pM)T such that the expected integrated squared error of the estimated diffusion signal, conditional on PM, is minimized. Denote the optimal locations as , which will be referred to as the optimal q-space design. We first consider this problem for a single voxel and then extend it to a set of voxels in a region of interest.
1). Optimal Diffusion Sampling for One Voxel:
At a particular voxel, we identify the optimal q-space design as the solution to the following optimization problem:
| (5) |
where is the estimator from Equation (4) which relies on PM, and the expectation is with respect to the distribution of X at this voxel. Expanding the square and swapping the integral and the expectation operators, it is straightforward to show that (5) can be equivalently formulated as the following maximization problem:
| (6) |
where, under the rank K assumption on C, we have cM(p) = ΨM,KΛKψK(p) and . Substituting these definitions into (6) and integrating, we arrive at the following finite dimensional optimization problem:
| (7) |
In practice, it is often sufficient to constrain the candidate design points to be in a large but finite set , e.g. a dense equispaced grid on to measure diffusion-weighted signals. In this scenario, the problem (7) can be written as
| (8) |
where and .
The global problem (8) is an integer program and is known to be NP-hard. To circumvent this computational bottleneck, we propose a sequential search strategy for producing a greedy approximation . Assuming that we have selected in the previous rounds, we propose solving the conditional optimization problem to obtain
| (9) |
where is the eigenfunction evaluation matrix over and .
Algorithm 1 presents an efficient procedure for sequentially solving (9) by making use of a rank one update scheme to compute that leverages the block matrix inversion formula to avoid a series of costly matrix inversions, especially when m gets large. Specifically, it is easy to see that Γm can be partitioned as
where and with ψK(pm) = (ψ1(pm), …, ψK(pm))T, and pm is a candidate point in for . As a result, the inverse of Γm can be computed efficiently through
| (10) |
where , and is precomputed based on the iteration of finding .
A question that naturally arises is: how good is the approximation ? In Theorem 1, we provide a bound for the ratio of the performance between the greedy approximation from Algorithm 1 and and the global optimum from (8).
Theorem 1:
Let be the approximate design obtained after m iterations of Algorithm 1. Define , where λmax is the function which returns the maximum eigenvalue of its argument. Then
| (11) |
where is the solution to (8).
For a proof of Theorem 1 and an interpretation of (11) in terms of the distribution of the random function X, see the supplemental material.
Algorithm 1.
Algorithm for greedy approximation to (5)
| 1: | Input: Diagonal matrix of eigenvalues ΛK, eigenfunctions {ψ1, …, ψk}, possible design points , measurement error variance σ2, and total budget . |
| 2: | Output: Vector of indices defining the design i |
| 3 | Initialize: i ← zeros(M), list of indices of |
| 4: | for m = 1 : M do |
| 5: | voptim ← 0 |
| 6: | for n in a do |
| 7: | ψn ← (ψ1(pn))⊤ for |
| 8: | if m = 1 then ▷ handle boundary case |
| 9: | ; |
| 10: | else |
| 11: | compute using formula (10) and |
| 12: | let |
| 13: | |
| 14: | if vn > voptim then |
| 15: | ioptim ← n; voptim ← vn; ; |
| 16: | i(m) ← ioptim and remove i(m) from a |
| 17: | ; |
2). Extension to Multiple Voxels:
The previous formulation considers the optimal design for recovery of the diffusion signal function for a single voxel. We now discuss how to extend our construction to the case when a single q-space design is needed for a collection of voxels in a region of interest , e.g. the whole white matter or all voxels covering certain fiber bundles of interest. Extending the objective function in (5) to perform joint optimization may be accommodated by minimizing the sum over the region
If no a-priori weighting is assumed, we can let , and by similar derivations as in Section II-C1, the sequential optimization problem in (9) now becomes
| (12) |
where the superscript v indicates the voxel specific data. Simple augmentation of Algorithm 1 can be made to accommodate this situation.
D. Prior Summarization From Historical Data
Notice that both the estimator (4) and Algorithm 1 depend on unknown parameters σ2, u, BK, and ΛK, the latter two of which can be computed from the decomposition of Σ (defined in equation (2)) along with any standard technique to select the rank K, e.g. the proportion of variance explained. σ characterizes the signal to noise ratio and is related to the scanner used to acquire the data. u and Σ characterize the mean and variance of the diffusion signal in a voxel.
We propose to estimate σ from b = 0 (non-diffusion weighted) data with the assumption that we have access to n (> 2) such images for the subject to be scanned, e.g. collected at some pre-screening step (see the supplemental material for further elaboration and evaluation of this estimator). In the rest of this section, we discuss how to leverage high-quality historical diffusion data to estimate u and Σ.
We assume that we have access to a relevant collection of densely sampled high-resolution historical dMRI data from the population of interest, denoted as , and where pim is a q-space acquisition direction, Mi is the number of samples, svim is the observed diffusion signal for the ith subject in the mth direction at the vth voxel and N represents the number of subjects. Such data are ubiquitous and easy to access due to recent efforts from the HCP and the UK Biobank. We leverage to estimate u and Σ across the brain. This proves challenging in that each subject’s data in is gathered in a different coordinate system and the subjects have different brain morphology. We circumvent this issue by registering the historical data to a common space (also referred to as the template space). The mean and covariance functions are estimated at each voxel in the template space. The warping function used to perform registration is constrained to be a diffeomorphism. As a result, for any subject of interest that has been registered to the template space, the inverse warping can be used to map the estimated priors in template space to the subject space smoothly. We now discuss this pipeline in detail.
1). Diffusion Signal Estimation in Subject Space:
Given the fact that data in are acquired with dense samples, we can confidently estimate the smoothed diffusion signal at each voxel independently using the popular regularized least squares estimator [19]. In brief, the diffusion signal is represented using a basis expansion over the ϕj’s. For voxel v in historical subject i, its diffusion signal is estimated by solving the following regularized least squares problem
where ϕ(p) = (ϕ1(p), …, ϕJ(p))T, denotes the Laplace-Beltrami operator on and penalty λi > 0 is selected by minimizing the generalized cross validation (GCV) criteria. The resulting smoothed estimate of diffusion signal is given by . More details on this estimation procedure can be found in the supplementary materials.
2). Mean and Covariance in Template Space:
Let γi be the diffeomorphic warping function from the historical subject i to the desired template space, e.g., based on the subject’s T1 or fractional anisotropy image [22]. The warping function deforms the regular voxel grid in subject space to align the subject’s data to the template, and a re-gridding procedure is needed in the template space to reconstruct a regular voxel grid. We can obtain an estimate of the subject’s diffusion function at each of the voxels in the template space through interpolation of . That is, for any voxel in the template space, we have where is computed by a linear interpolation scheme applied to for v’s in the subject space mapped near by γi. We use the popular Advanced Normalization Tools (ANTs) to perform this alignment and refer the reader to [23] for more details on the interpolation procedure.
Once the aligned diffusion signal has been computed for all subjects in , the mean and covariance functions at in the template space can be estimated by the empirical estimators: and , where and are the sample mean and covariance matrix of , respectively. We assume a large historical sample, N > J and therefore we expect these estimators to be stable.
3). Mean and Covariance in Subject Space:
Let γ be the warping function to align the new subject to be scanned to the template space. characterizes the distribution of the diffusion signal at voxel in the template space. Our goal is to map the GPs in the template space to the subject space to facilitate q-space sampling (Algorithm 1) and diffusion signal estimation (estimator (4)). To do this, we estimate the mean and covariance functions at each voxel in subject space using linear interpolation guided by the inverse warping transformation γ−1. This process is described in detail below.
The mean functions are parameterized by , a vector in , and therefore for subject space voxel v can be formed by standard multivariate linear interpolation. The covariance functions are parameterized by symmetric positive definite (SPD) matrices and thus the estimate of for subject space voxel v requires an interpolation of SPD matrices. Considering the non-Euclidean structure of the SPD matrix manifold, we propose the following interpolation scheme.
Denote the manifold of J × J SPD matrices and let M1, …, Mn with each . The weighted Karcher mean of the sample is defined as
| (13) |
where d(·, ·) is induced by a selected Riemannian metric. If we let the weights wi be proportional to the Euclidean distance between and the voxel of interest v in the subject space, the weighted Karcher mean can be used as an interpolation scheme.
There are several valid metrics that can be used to compute the distance between elements in [24]. It is easy to show , so using the standard Euclidean (Frobenius) norm as a metric for the interpolation appears as a reasonable choice. However, interpolation of SPD matrices under the Euclidean norm can result in a “swelling” effect, in which the determinant of the interpolated value is larger than the determinant of any of the original SPD matrices [25]. This would result in an undesirable amplification of the estimated variability. To overcome this limitation, we use the log-Euclidean metric proposed in [24]
where log(·) represents the matrix logarithm. For any , the matrix logarithm always exists and is defined by log(M) = Vlog(D)V⊤, where V is the matrix whose columns are the eigenvectors of M and D is the diagonal matrix of eigenvalues. Since D is diagonal, the matrix logarithm has an especially simple form: log(D) = Diag(log(D11), log(D22), …, log(DJJ)). The subject space covariance function for voxel v is defined by the solution to the interpolation (13) applied to the set of , for mapped near v by γ−1, with .
With the estimated σ, u and Σ for the subject to be scanned, we now can apply Algorithm 1 to acquire sparsly sampled dMRI data, and estimator (4) to recover the diffusion signal (and fODF with the FRT).
III. SIMULATION STUDY
In this study, we compare the performance of the proposed method with the penalized linear regression approach using spherical harmonics from [19] (denoted SHLS) and the spherical finite rate of innovation and sparse regularization estimator from [26] (denoted SFR) over a series of increasing budgets. For a given budget, the samples for SHLS and SFR were selected using the popular electrostatic repulsion model (ESR). The ESR algorithm constructs a sampling design by modeling the acquisition directions as points on and attempting to find a configuration which minimizes the electrostatic energy between point pairs. It has been widely used for constructing acquisition schemes over , e.g., in the HCP. The proposed greedy approach (Algorithm 1), denoted GDS (greedy design selector), was used to select the sampling design for our sparse estimator (4), denoted CU (conditional expectation update).
We simulate the diffusion signal at a particular voxel with both single and crossing fibers as follows. Denote gm,κ the density function of the von Mises-Fisher (VMF) distribution with mean direction and concentration parameter κ. A sample fODF is simulated according to the generative model
| (14) |
where ν1 = (1, 0, 0),, and w1 = w2 = 0.5. Note that we must include to respect the symmetry of the diffusion process. Some examples of generated fODFs from (14) are shown in the top left panel of Figure 1. Note that single fibers are simulated from (14) when the samples m1i and m2i are sufficiently close. The sampled fODF is then represented using the real, symmetric spherical harmonics {ϕ1, …, ϕJ}. Under this representation, the FRT is invertible (i.e. it is a bijection for symmetric differentiable functions on ) and has a simple closed form which can be applied to the coefficients of {ϕ1, …, ϕJ} to obtain the corresponding diffusion signal Xi.
Fig. 1.

Simulation study results. The top left panel displays a sample of fODFs generated from model (14). The top right figure compares the MISE of different methods. The bottom two panels compare the angular accuracy, where the bottom left panel displays the angular error and the bottom right gives the percentage of false peaks for a sequence of increasing budgets.
N = 200 diffusion signals were simulated according to this procedure and evaluated at 90 locations on selected by the ESR algorithm. Each observation was contaminated with i.i.d. measurement error with σ = 0.01. An independent test set of size Ntest = 100 was also constructed using the same method. The reconstruction techniques were applied to the test set for a sequence of increasing observations (budgets). SFR estimates the fODFs directly, so an estimate of the diffusion signal function was obtained by applying the inverse FRT. In contrast, SHLS and CU estimate the diffusion signal function and therefore we apply the FRT and then the spherical deconvolution [6] to obtain the fODF estimates.
Performance was compared using three evaluation metrics: the mean integrated squared error (MISE) of the estimated diffusion signal, the percentage of false peaks (PFP), and the estimated angular error (EA). The PFP and EA are popular measures quantifying the angular fidelity of the estimated fODF. The PFP is computed simply as the proportion of estimated fODFs in the test set which have the same number of peaks as the true fODF, i.e.
where np is a function that counts the number of local maximum of a function. We defined the estimated angular error (EA) as
where θ is the function defined as follows: if only a single peak is detected, θ = 0; if two or more peaks are detected, θ is set to the angle between the two highest peaks. This is a natural extension of the EA metric used in [27].
The top right panel of Figure 1 shows the results of the MISE comparison. The proposed method is uniformly outperforming both of the competitors over all budgets considered. The boost in performance is especially pronounced in the sparse (⩽ 20) sample case. The bottom left panel of Figure 1 displays the EA and the bottom right shows the PFP for all estimators as a function of budget. The same pattern in the MISE evaluation is observed in both the angular assessments, indicating that the boost performance for the sparse sample diffusion signal estimation carries through to quality fODF reconstruction. Additional simulation studies investigating the performance in Non-Gaussian and low signal to noise ratio scenarios result in similar conclusions to those presented. Further, we found the computational complexity of CU-GDS fitting to be comparable with that of SHLS-ESR. For these results as well as a discussion on the selection of an adequate training sample size N, we refer the interested reader to the supplemental material.
IV. REAL DATA EXPERIMENTS
A. Data Acquisition and Preprocessing
We evaluated our method using diffusion data from two datasets: HCP and a local aging dataset consisting of older adults with mild cognitive impairment (MCI). The full HCP imaging acquisition and minimal diffusion MRI image preprocessing steps for HCP data are documented in [28]. Briefly, all imaging was conducted on the 3T Siemens Connectom scanner (Erlangen, Germany). High-resolution T1w anatomical images were acquired with the 3D MPRAGE (magnetization prepared rapid gradient echo) sequence with a slice acceleration factor of 2 using 0.7 mm isotropic resolution. Diffusion imaging was performed using a 2D spin-echo EPI (echo planar imaging) sequence with approximately 90 diffusion directions at three non-zero b-values (1,000, 2,000, and 3,000 s/mm2 each) and 18 b = 0 reference scans at 1.25 mm isotropic resolution. A full diffusion MRI run includes 6 runs of about 9 mins 50 seconds each, representing 3 acquisitions of different b-vectors, each acquired once with right-to-left (RL) and left-to-right (LR) phase encoding polarities; these are used to correct for susceptibility-induced distortion.
For our study, we randomly selected 240 healthy HCP subjects (age range: 20–35 years) to be used as training data, i.e. high-resolution historical data . An additional 10 subjects from the same demographic were used as independent test data. A template of the diffusion data was created from the FA images of 20 randomly selected training subjects. Each subject’s FA image was then warped to the template using the nonlinear registration algorithm in ANTs [23]. The training data were used to estimate two template space priors according to the procedure outlined in Section II-D using the b = 1, 000 and b = 2, 000 shells, respectively. The inverse warping function for each subject in the test data was used to map the template space mean and covariance functions into their subject space.
To study the generalizability of our method, we utilized a local dataset consisting of 16 older adults (> 60 years of age) with MCI. The T1 and dMRI data were collected using a Siemens 3T Prisma scanner (Erlangen, Germany) at the University of Rochester Imaging Center. A total of 14 b = 0, 26 b = 1, 000 and 102 b = 2, 000 images were obtained with q-space locations selected using the ESR algorithm. All acquisitions were acquired with the same 2 mm isotropic resolution. Susceptibility distortion and motion correction were performed using FSL. The T1 image was processed using ANTs and FreeSurfer, registered to the dMRI and a white matter mask was extracted. For additional details about the study population and scanning protocol, see [29].
B. Global Estimation and Tractography
All experiments in this subsection are based on the HCP data. For a sequence of increasing total budgets, we applied the GDS algorithm to select diffusion directions from the set of 90 b = 1, 000 vectors for each HCP test subject independently. We determined an optimal set of directions to sample 1) for each voxel, and 2) for an ROI. For the second case, we randomly selected 10, 000 voxels within the white matter mask to define the ROI , i.e. approximating the optimal directions over the whole white matter. In analysis not reported here, we observed that the final results were insensitive to resamplings of , so long as the sample was sufficiently large. For each design selected by GDS, we estimated the diffusion signal using the proposed estimator (4), denoted again as CU. Throughout this section, we compare our method to SHLS-ESR for various tasks of interest, noting that it displayed largely similar performance to SFR-ESR in the simulation study and was found to be much faster for large datasets.
Since in general SHLS-ESR reconstruction enjoys good performance for dense samples, we first evaluated the diffusion signal estimation for both techniques at various sparse budgets by comparing them to the SHLS-ESR applied to the full 90 directions, denoted as SHLS-ESR90. Figure 2 (a) displays boxplots of the MISE between the sparse sample reconstructions and SHLS-ESR90 over all of the white matter voxels, where CU-GDS-PV (per voxel) represents sampling scenario 1), and CU-GDS-COM (common, per subject) represents sampling scenario 2). The CU-GDS framework produced reconstructions which better approximate the signal estimates of SHLS-ESR90 than does SHLS-ESR for each of the sparse budgets considered. The performance boost becomes more pronounced in the sparser cases, e.g., when we only have the budget to sample fewer than 20 directions. The per voxel design resulted in lower average MISE than the per subject design, as expected since the latter is equivalent to the former with an additional constraint, but the difference is minimal relative to the performance of SHLS-ESR. Figure 2 (b) displays the optimal per-budget q-space sampling schemes constructed from the GDS-COM algorithm for two randomly selected HCP test subjects. The corresponding ESR schemes are also included for comparison (red crosses). We note two important features. First, for a given budget, the two GDS-COM designs select relatively few q-space samples in common with ESR. This indicates that, under the CU estimator, the ESR method produces sub-optimal designs for diffusion signal estimation. Second, the two GDS-COM designs have a relatively large number of design points in common (the cyan color). This is not surprising, as there exist large similarities in the diffusion profiles of healthy young adult brains. That said, the designs are not identical, indicating there is some residual subject-specific variability driving the selection of the locations that differ.
Fig. 2.

(a) Boxplots of the MISE between the sparse sample reconstructions (SHLS-ESR, CU-GDS-COM, and CU-GDS-PV) and SHLS-ESR90, where CU-GDS-COM represents CU fit based on per subject design, and CU-GDS-PV represents CU fit based on per voxel design. (b) Comparison of two q-space sampling schemes, ESR and our GDS-COM, for two randomly selected test subjects from HCP. The sample location for ESR are denoted by red crosses. The cyan points indicate sample locations selected by the GDS-COM for both of the test subjects, while the blue and green samples are locations selected for only one of the subjects. Each panel shows the selected design under a different budget constraint.
For each total budget considered, fODFs were constructed from the diffusion signal estimates by applying first the FRT and subsequently the spherical deconvolution [6]. Figure 3 displays the fODFs from a coronal slice of a randomly selected test subject. Using only 5 directions, CU-GDS-COM estimation already results in fODFs that very closely approximate those based on the full set of 90 directions. For instance, the single fiber region in the corpus callosum (bottom left pink area) as well as the crossing fibers in the centrum semiovale (center blue area) are nearly identical between the budget = 5 and budget = 90 plots. This is not the case for SHLS-ESR, which struggles in both areas for sparse budgets. Boxplots of the per-voxel MISE between the sparse budget and full data reconstructions for both methods are displayed below. For each sparse budget considered, the CU-GDS-COM produces estimates which more faithfully represent the full data reconstructions than does the SHLS-ESR.
Fig. 3.

fODFs from a coronal slice of a randomly selected test subject reconstructed by SHLS-ESR (left column) and CU-GDS-COM (right column) for several budgets considered. The boxplots display the per voxel MISE between the sparse budgets and full data reconstructions, for both methods.
The motivating application for this paper is to develop a sparse diffusion weighted signal sampling framework for brain structural connectome analysis. Therefore, we investigated the quality of the tractogram and structural connectivity resulting from the reconstructed fODFs at various sampling budgets. The tractogram was computed from the fODFs using the particle filtering tractography algorithm [8] that is implemented in dipy. Connectome matrices were reconstructed using the Desikan atlas [30] created by FreeSurfer [31]. The top plot in Figure 4 (a) shows a comparison of the tractography results from SHLS-ESR and CU-GDS-COM reconstructions of a randomly selected test subject for several budgets. We can see that with as few as 5 directions, the proposed methodology is able to produce meaningful streamlines. The bottom left plot in Figure 4 (a) displays the corresponding difference maps between the sparse budget and full data connectome matrices, and the bottom right plot in Figure 4 (a) shows the test set average per-budget normalized Frobenius distance between the sparse and full data connectomes for both methods. We see that on average, the proposed method produces connectomes closer to the full data reconstruction than does the SHLS-ESR based framework for each sparse budget considered. To further quantify the structural connectivity matrix, we extracted and compared several graph topological metrics including network density, global efficiency, transitivity, and characteristic path length. Detailed definitions of these metrics as well as their interpretation within the context of connectome analysis can be found in [32]. Figure 4 (b) displays the average and 95% error bars for each of these metrics computed over the 10 subjects in the test set at different sampling budgets. We can see that the network properties of the connectomes restored using the CU-GDS-COM method converge to the full data results faster than those of the SHLS-ESR method. The SHLS-ESR generally needs 30 or more directions to obtain a connectivity matrix that is comparable with the one obtained from 90 directions, while the CU-GDS-COM only needs around 10.
Fig. 4.

(a, Top) WM fiber streamlines and connectivity matrices from a randomly selected test subject. Tractography was run on fODFs constructed from diffusion signal function estimates from SHLS-ESR (left column) and CU-GDS-COM (right column) for a sequence of increasing total budgets. (a, Lower left) The corresponding connectivity difference maps between the sparse budget and full data reconstructions. (a, Lower right) The line plot shows the normalized Frobenius distance between the budget limited reconstruction and full data connectome, on the log scale. (b) Line plots of several graph theoretic metrics of brain networks for SHLS-ESR and CU-GDS-COM methods over different budgets computed on the test set. Means and 95% error bars are displayed.
C. Reproducibility Analysis
We studied the ability of the proposed method to capture meaningful variability across subjects. Due to the incorporation of a prior in the sampling direction selection and signal fitting, it is a valid concern that our method could be “shrinking” the signal estimates too much towards the “average brain”, especially in the sparse sample case. To investigate this, we leveraged the test and retest data in HCP, i.e., those subjects that have multiple scans collected at different times. In total, 12 non training-set subjects with their test and retest diffusion data (24 scans) were involved in our analyses.
We compared the reproducibility (the inter- v.s. intra-subject variabilities) of the individual fODFs computed by either method. In particular, for each subject, we fit fODFs using the complete data (all 90 directions) collected at the first scan (scan-1). The data from the second scan (scan-2) was used to estimate the fODFs under a variety of budget constraints. All of the fODFs were warped to the template space to allow joint comparisons. For each budget and both reconstruction methods, within- and between-subject errors were computed for the scan-2 fODFs using the complete data scan-1 fODFs as ground truth. The reproducibility at each voxel was summarized using the [33], where and are the mean between- and within-subject errors at voxel v, respectively. dICC takes values ranging from 0 to 1, where a higher value indicates better reproducibility, i.e., large inter-subject variance and small intra-subject variance. In our case, a higher dICC can indicate that the priors are not dominating the estimation and that our method can properly model subject level heterogeneity from individual dMRI data. As a baseline, we also computed fODFs using just the prior mean. The per-voxel distance between the mean fODF and scan-1 “ground truth” was calculated and averaged across subjects, denoted here as .
Figure 5 (a) shows the dICCv at all white matter voxels in a coronal slice of the template space for several budgets. The proposed method displays better reproducibility, especially at sparse budgets, for nearly all of the voxels shown. The left panel of Figure 5 (b) shows the per-budget average dICCv, i.e.. Our method results in better average reproducibility for all the budgets considered. The red-dashed line highlights the reproducibility of SHLS-ESR-90, i.e. SHLS-ESR on the complete data. We see that with as few as 20 acquisitions, the proposed method exhibits stronger reproducibility than the competitor method applied to the complete set of 90 b=1, 000 acquisitions. The right panel of Figure 5 (b) shows the per-budget within- (WS) and between-subject (BS) distances averaged across the white matter voxels, i.e., , for both reconstruction methods. The prior mean result, i.e., is also displayed. As a Bayesian method, in the absence of much data, the estimates from CU-GDS are shrunk towards the prior, which is manifested in the relative proximity of the between- (green solid) and within-subject (green dashed) curves at low vs high budgets. Still, even for the budget=5 case, the proposed method exhibits significantly lower within-subject distance and higher between-subject distance compared to the prior mean (purple dashed line). Thus, CU-GDS is able to capture meaningful subject specific variability in the very sparse regime. This is in contrast to SHLS-ESR (red lines), whose performance in the budget< 10 case is very poor to the point where it is inferior to the naive strategy of simply using the prior mean for each subject.
Fig. 5.

(a) dICCv using the fODFs computed from the test-retest HCP data for several budgets. (b) (Left) Average dICCv for a variety of budgets for CU-GDS and SHLS-ESR. The average dICCv for SHLS-ESR-90 is displayed for reference (red-dashed). (Right) per-budget within- (WS) and between-subject (BS) distances averaged across the white matter voxels. The horizontal line identifies the prior mean’s result, i.e.,.
D. Out of Distribution Analysis
One potential disadvantage of the proposed method is its requirement of high-quality dMRI data from the population of interest. This could restrict the potential impact of the method given that such data are not always available. In this section we investigate how well the priors learned from one population generalize to another population. In particular, we used the priors estimated from the HCP young adults and applied the proposed methodology to a test dataset consisting of older adults with MCI. Since this out of distribution (OOD) data are predominantly on the b = 2, 000 shell, we used the HCP priors constructed from this shell for analysis. The FA image for each OOD subject was registered to the template space and the resulting warping function was used to map the priors into the corresponding subject spaces. CU-GDS and SHLS-ESR estimators were applied to the b = 2, 000 data for a set of increasing budgets. Again, the fODFs were constructed from the estimated diffusion signal functions by applying the FRT and spherical deconvolution from [6]. Due to the absence of a real “ground truth” for comparison, we adopt the same tactic as before and allow the SHLS-ESR reconstructed fODFs from the first 90 directions to serve this role. The fODFs constructed from the prior mean diffusion signal function were also computed. To allow direct comparison of the results to those from an “in distribution” (data from HCP) sample, we also computed the fODFs using CU-GDS applied to the HCP test subjects b = 2, 000 shell.
For each test subject, the difference between the sparse budget fODFs and high cost ground truth was measured at each voxel using both distance and angular error. The per-subject averages of both metrics over the entire OOD sample are displayed in the box plots in Figure 6. For comparison, the median performance of the prior mean for both OOD and HCP samples is shown, along with the results of the CU-GDS method applied to the in distribution HCP b = 2, 000 test data. Although it does not perform as well as the in distribution case, the CU-GDS method still better approximates the SHLS-ESR90 fODFs than do those constructed from SHLS-ESR for all of the sparse budgets considered, as measured by both error (panel (a)) and angular error (panel (b)). Notice that even in the OOD sample, the CU-GDS method performs significantly better than the prior mean fODF reconstruction, indicating that the boost in performance is not driven by the prior mean. Comparing the in vs. out of distribution results of the CU-GDS, it is clear that access to priors learned using the same population as the test data results in better reconstructions. Still, even when trained on a historical sample from a population that is ostensibly very different than the target, our method’s sparse sample estimates better approximate the full data fODFs when compared to the standard technique.
Fig. 6.

Average per out of distribution (OOD) subject sparse budget fODF reconstruction performance for CU-GDS (green) and SHLS-ESR (red). The median of the prior mean fODF performance is shown for reference (purple line). As measured by both distance (a) and estimated angular error (b). For OOD data, although not as good as “in distribution” (the HCP case) data, CU-GDS can still better approximate the directional information computed from SHLS-ESR90 for each spares budget considered.
V. CONCLUSION AND DISCUSSION
This paper introduces a novel methodology which incorporates high-resolution historical dMRI data to optimize diffusion weighted directions in dMRI acquisition and sparse sample diffusion signal estimation. Historical data is used to construct a prior distribution of diffusion signal at each voxel in a template space through a low-rank GP model. Given a new subject of interest for dMRI acquisition, the GP priors in the template space are mapped to the subject space and then used to define the subject-specific acquisition directions that minimize the expected integrated squared error of the proposed diffusion signal estimator. A computationally efficient greedy algorithm is proposed to select the directions and is shown to approximate the performance of the true optimal set of directions within some computable bounds. Using extensive simulation and real data experiments, we show that the combination of the proposed diffusion sampling and signal estimation generally outperforms traditional techniques. The proposed framework is significantly better in the sparse sample case, i.e, when we have limited budget in the number of diffusion directions to acquire. In applying the proposed method to analyze HCP data, we show the boost in performance of our method in diffusion signal estimation carries through to fODF reconstruction and ultimately to quality tractography estimation.
We now provide several remarks on the practical application of our method. Given the public availability of large brain imaging data sets with densely sampled high-resolution dMRI, e.g., the different types of HCP sets, the Adolescent Brain Cognitive Development (ABCD) set, and the UK Biobank set, the requirement of finding a sample data to serve in constructing the prior is easy to accomplish for many populations of interest. Even if an exact reference population cannot be acquired, results from Section IV-D indicate that, at least in some cases, our method still performs well when applying the priors learned from one population to the analysis of another. In order to construct quality empirical priors, it is important to include an adequate number of subjects in the historical training sample. In additional simulation results reported in Section S5 of the supplemental material, we found that the performance of our method was stable for N ⩾ 50. In practice, this will depend on the population of interest and, if necessary, a resampling scheme can be used to select N. In order to use our methodology for data acquisition of a new subject, a small number of b = 0 scans must first be collected to provide the subject’s coordinate for prior injection and to estimate the measurement error variance σ2. In this work, the registration between template and subject space was conducted using the FA image, but it is possible to instead use the b = 0 image for this purpose in a real application. For estimating σ2, although the historical sample could be used, it is important to do so for each new subject since it captures the noise introduced by the scanner, which can vary substantially between machines and scanning sites (see the supplemental material Section S2 for further discussion).
Moreover, an important question that arises in applying the proposed method to real data collection is how to correct eddy current-induced distortions and subject movements for sparsely sampled dMRI. In practice, the most popular method for motion and distortion correction is FSL’s eddy toolbox, which makes several assumptions about the acquired data. For instance, it is assumed that the diffusion encoding directions span the entire sphere. In analysis not reported we observed that, for a given budget, the GDS designs had similar electrostatic energy as the ESR and typically much smaller energy than a design by random uniform sampling on . For example, notice that the GDS designs in Figure 2 are relatively well dispersed, even compared to the “optimally dispersed” ESR designs. We therefore expect the GDS designs to satisfy the dispersion requirement of FSL’s eddy toolbox. Additionally, we note the current eddy toolbox in FSL uses a GP model [13] to predict diffusion signal, and that this model may perform poorly for sparse data. To make it work for sparse data, one can replace the prediction equation (7) in [13] with our prediction equation (4). Alternatively, to use the current eddy toolbox as is for the proposed acquisition method, we make the following two suggestions. 1) Acquire at least 20 directions. According to [27], we expect the current eddy toolbox to work for this sparsity level. 2) Acquire blip-up-blip-down volumes for each direction. The repetition of all diffusion gradient directions using forward and reverse phase encoding provides a means to resolve signal intensity recovery in compressed areas, and gives the best distortion correction in practice [34].
We conclude by offering several potential applications and extensions of the proposed framework. First, in clinical scans, long scanning times can be infeasible for a variety of reasons. The proposed method can be used to design optimal gradients to quickly acquire sparse q-space data for important clinical purposes such as surgical planning. Second, there are studies trying to increase the spatial resolution of diffusion MRI to more accurately detect white matter fiber tracts at a submillimeter isotropic resolution [12]. Such spatial resolution requires long TR, and thus only very few samples can be collected in a given time period. In this sparse setting, the proposed method shows the most prominent performance boost over the standard technique. One immediate extension is to augment the statistical model to accommodate for signal attenuation in the multi-shell situation. This would allow the derivation of a joint optimization scheme for both gradient direction and magnitude in the dMRI protocol. Another extension involves multi-voxel analysis and information borrowing from neighborhood regions. The white matter tracts often group in bundles that span many voxels, inducing correlations between the signal in neighboring voxels. Considering such correlation can increase the efficiency of the estimation procedure. Finally, it is likely that in some situations, an appropriate historical sample for the reference population does not exist and priors from healthy subjects may have poor generalization performance, e.g. in clinical populations where particular tracts are profoundly altered due to pathology such as lesions or tumors. That said, the priors can still be of assistance in such cases, namely for identifying “outlier regions” where the current subject differs substantially from the reference prior. Under the Bayesian perspective, due to conjugacy we have an explicit form of the full posterior of the diffusion signal, given the observed data. In practice, this allows methods from Bayesian outlier detection, e.g. [35, 36], to be adapted for our purposes. There is also a literature on outlier detection in functional data analysis that could be used in this task, e.g. [37], but we leave the specifics as an interesting direction of future inquiry.
Supplementary Material
Contributor Information
William Consagra, Department of Biostatistics and Computational Biology, University of Rochester, Rochester, NY, 14642.
Arun Venkataraman, Department of Physics, University of Rochester, Rochester, NY, 14627.
Zhengwu Zhang, Department of Statistics and Opertaions Research, UNC Chapel Hill, Chapel Hill, NC, 27599.
REFERENCES
- [1].Stejskal EO and Tanner JE, “Spin diffusion measurements: spin echoes in the presence of a time dependent field gradient,” The Journal of Chemical Physics, vol. 42, pp. 288–292, 1965. [Google Scholar]
- [2].Basser P, Mattiello J, and Lebihan D, “Estimation of the effective self-diffusion tensor from the NMR spin echo,” Journal of Magnetic Resonance, Series B, vol. 103, pp. 247–254, 1994. [DOI] [PubMed] [Google Scholar]
- [3].Tuch DS, Reese TG, Wiegell MR, Makris N, Belliveau JW, and Wedeen VJ, “High angular resolution diffusion imaging reveals intravoxel white matter fiber heterogeneity,” Magnetic Resonance in Medicine, vol. 48, no. 4, pp. 577–582, 2002. [DOI] [PubMed] [Google Scholar]
- [4].Maier-Hein KH, Neher PF, Houde J-C, Côté M-A, Garyfallidis E, Zhong J, Chamberland M, Yeh F-C, Lin Y-C, Ji Q, et al. , “The challenge of mapping the human connectome based on diffusion tractography,” Nature Communications, vol. 8, no. 1, pp. 1–13, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [5].Alexander D, Barker G, and Arridge S, “Detection and modeling of non-gaussian apparent diffusion coefficient profiles in human brain data,” Magnetic Resonance in Medicine, vol. 48, no. 2, pp. 331–340, 2002. [DOI] [PubMed] [Google Scholar]
- [6].Descoteaux M, Deriche R, Knösche TR, and Anwander A, “Deterministic and probabilistic tractography based on complex fibre orientation distributions,” IEEE Transactions on Medical Imaging, vol. 28, pp. 269–286, 2009. [DOI] [PubMed] [Google Scholar]
- [7].Tournier J-D, Calamante F, and Connelly A, “Mrtrix: Diffusion tractography in crossing fiber regions,” International Journal of Imaging Systems and Technology, vol. 22, no. 1, pp. 53–66, 2012. [Google Scholar]
- [8].Girard G, Whittingstall K, Deriche R, and Descoteaux M, “Towards quantitative connectivity analysis: Reducing tractography biases,” NeuroImage, vol. 98, pp. 266–278, 2014. [DOI] [PubMed] [Google Scholar]
- [9].Tuch DS, “Q-ball imaging,” Magnetic Resonance in Medicine, vol. 52, pp. 1358–1372, 2004. [DOI] [PubMed] [Google Scholar]
- [10].Jones D, Horsfield M, and Simmons A, “Optimal strategies for measuring diffusion in anisotropic systems by magnetic resonance imaging,” Magnetic Resonance in Medicine, vol. 42, no. 3, pp. 515–525, 1999. [PubMed] [Google Scholar]
- [11].Chen N-K, Guidon A, Chang H-C, and Song AW, “A robust multi-shot scan strategy for high-resolution diffusion weighted MRI enabled by multiplexed sensitivity-encoding (MUSE),” Neuroimage, vol. 72, pp. 41–47, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [12].Chang H-C, Sundman M, Petit L, Guhaniyogi S, Chu M-L, Petty C, Song AW, and Chen N.-k., “Human brain diffusion tensor imaging at submillimeter isotropic resolution on a 3 Tesla clinical MRI scanner,” Neuroimage, vol. 118, pp. 667–675, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [13].Andersson JL and Sotiropoulos SN, “Non-parametric representation and prediction of single- and multi-shell diffusion-weighted MRI data using Gaussian processes,” NeuroImage, vol. 122, pp. 166–176, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [14].Caruyer E and Deriche R, “Adaptive design of sampling directions in diffusion tensor MRI and validation on human brain images,” in MICCAI Workshop on Diffusion Modelling and the Fiber Cup, (Londres, United Kingdom: ), Sept. 2009. [Google Scholar]
- [15].Poot DHJ, den Dekker AJ, Achten E, Verhoye M, and Sijbers J, “Optimal experimental design for diffusion kurtosis imaging,” IEEE Transactions on Medical Imaging, vol. 29, no. 3, pp. 819–829, 2010. [DOI] [PubMed] [Google Scholar]
- [16].Ji H and Müller H-G, “Optimal designs for longitudinal and functional data,” Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol. 79, no. 3, pp. 859–876, 2017. [Google Scholar]
- [17].Wu M, Diez-Roux A, Raghunathan TE, and Sánchez BN, “FPCA-based method to select optimal sampling schedules that capture between-subject variability in longitudinal studies,” Biometrics, vol. 74, no. 1, pp. 229–238, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [18].Glasser MF, Smith SM, Marcus DS, Andersson JL, Auerbach EJ, Behrens TE, Coalson TS, Harms MP, Jenkinson M, Moeller S, Robinson EC, Sotiropoulos SN, Xu J, Yacoub E, Ugurbil K, and Van Essen DC, “The Human Connectome Project’s neuroimaging approach,” Nature Neuroscience, vol. 19, no. 9, pp. 1175–1187, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [19].Descoteaux M, Angelino E, Fitzgibbons S, and Deriche R, “Regularized, fast, and robust analytical Q-ball imaging,” Magnetic Resonance in Medicine, vol. 58, pp. 497–510, 2007. [DOI] [PubMed] [Google Scholar]
- [20].James GM, Hastie TJ, and Sugar CA, “Principal component models for sparse functional data,” Biometrika, vol. 87, no. 3, pp. 587–602, 2000. [Google Scholar]
- [21].Yao F, Müller H-G, and Wang J-L, “Functional data analysis for sparse longitudinal data,” Journal of the American Statistical Association, vol. 100, no. 470, pp. 577–590, 2005. [Google Scholar]
- [22].Avants BB, Epstein CL, Grossman M, and Gee JC, “Symmetric diffeomorphic image registration with cross-correlation: evaluating automated labeling of elderly and neurodegenerative brain,” Medical Image Analysis, vol. 12, pp. 26–41, Feb 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [23].Avants BB, Tustison N, and Song G, “Advanced normalization tools (ants),” Insight Journal, vol. 2, no. 365, pp. 1–35, 2009. [Google Scholar]
- [24].Arsigny V, Fillard P, Pennec X, and Ayache N, “Geometric means in a novel vector space structure on symmetric positive-definite matrices,” SIAM Journal on Matrix Analysis and Applications, vol. 29, no. 1, pp. 328–347, 2007. [Google Scholar]
- [25].Dryden IL, Koloydenko A, and Zhou D, “Non-Euclidean statistics for covariance matrices, with applications to diffusion tensor imaging,” Ann. Appl. Stat, vol. 3, pp. 1102–1123, 09 2009. [Google Scholar]
- [26].Deslauriers-Gauthier S and Marziliano P, “Sampling signals with a finite rate of innovation on the sphere,” IEEE Transactions on Signal Processing, vol. 61, no. 18, pp. 4552–4561, 2013. [Google Scholar]
- [27].Ning L, Laun F, Gur Y, DiBella EV, Deslauriers-Gauthier S, Megherbi T, Ghosh A, Zucchelli M, Menegaz G, Fick R, St-Jean S, Paquette M, Aranda R, Descoteaux M, Deriche R, O’Donnell L, and Rathi Y, “Sparse Reconstruction Challenge for diffusion MRI: Validation on a physical phantom to determine which acquisition scheme and analysis method to use?,” Med Image Anal, vol. 26, pp. 316–331, Dec 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [28].Glasser MF, Sotiropoulos SN, Wilson JA, Coalson TS, Fischl B, Andersson JL, Xu J, Jbabdi S, Webster M, Polimeni JR, Essen DCV, Jenkinson M, and f. t. W.-M. H. Consortium, “The minimal preprocessing pipelines for the Human Connectome Project,” NeuroImage, vol. 80, pp. 105–124, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [29].Venkataraman A, Risk B, Qiu D, Zhong J, Lin F, and Zhang Z, “Quantitative evaluation of multiband diffusion mri data,” in Proceedings of the 2021 ISMRM & SMRT Annual Meeting & Exhibition, vol. 1347, ISMRM, 2021. [Google Scholar]
- [30].Desikan RS, Ségonne F, Fischl B, Quinn BT, Dickerson BC, Blacker D, Buckner RL, Dale AM, Maguire RP, Hyman BT, Albert MS, and Killiany RJ, “An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest,” NeuroImage, vol. 31, pp. 968–980, 2006. [DOI] [PubMed] [Google Scholar]
- [31].Fischl B, “FreeSurfer,” NeuroImage, vol. 62, pp. 774–781, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [32].Rubinov M and Sporns O, “Complex network measures of brain connectivity: Uses and interpretations,” NeuroImage, vol. 52, pp. 1059–1069, 2010. [DOI] [PubMed] [Google Scholar]
- [33].Zhang Z, Descoteaux M, Zhang J, Girard G, Chamberland M, Dunson D, Srivastava A, and Zhu H, “Mapping population-based structural connectomes,” Neuroimag, vol. 172, pp. 130–145, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [34].Brun L, Pron A, Sein J, Deruelle C, and Coulon O, “Diffusion mri: Assessment of the impact of acquisition and preprocessing methods using the brainvisa-diffuse toolbox,” Frontiers in Neuroscience, vol. 13, p. 536, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [35].Bayarri M and Morales J, “Bayesian measures of surprise for outlier detection,” Journal of Statistical Planning and Inference, vol. 111, no. 1, pp. 3–22, 2003. Special issue I: Model Selection, Model Diagnostics, Empirical Ba yes and Hierarchical Bayes. [Google Scholar]
- [36].Zaslavsky AM and Bradlow ET, “Posterior predictive outlier detection using sample reweighting,” Journal of Computational and Graphical Statistics, vol. 19, no. 4, pp. 790–807, 2010. [Google Scholar]
- [37].Gervini D, “Outlier detection and trimmed estimation for general functional data,” Statistica Sinica, vol. 22, no. 4, pp. 1639–1660, 2012. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
