Abstract
Tensor-valued and matrix-valued measurements of different physical properties are increasingly available in material sciences and medical imaging applications. The eigenvalues and eigenvectors of such multivariate data provide novel and unique information, but at the cost of requiring a more complex statistical analysis. In this work we derive the distributions of eigenvalues and eigenvectors in the special but important case of m×m symmetric random matrices, D, observed with isotropic matrix-variate Gaussian noise. The properties of these distributions depend strongly on the symmetries of the mean tensor/matrix, D̄. When D̄ has repeated eigenvalues, the eigenvalues of D are not asymptotically Gaussian, and repulsion is observed between the eigenvalues corresponding to the same D̄ eigenspaces. We apply these results to diffusion tensor imaging (DTI), with m = 3, addressing an important problem of detecting the symmetries of the diffusion tensor, and seeking an experimental design that could potentially yield an isotropic Gaussian distribution. In the 3-dimensional case, when the mean tensor is spherically symmetric and the noise is Gaussian and isotropic, the asymptotic distribution of the first three eigenvalue central moment statistics is simple and can be used to test for isotropy. In order to apply such tests, we use quadrature rules of order t ≥ 4 with constant weights on the unit sphere to design a DTI-experiment with the property that isotropy of the underlying true tensor implies isotropy of the Fisher information. We also explain the potential implications of the methods using simulated DTI data with a Rician noise model.
Keywords: eigenvalue and eigenvector distribution, asymptotics, sphericity test, singular hypothesis testing, DTI, spherical t-design, Gaussian orthogonal ensemble
AMS subject classifications: 60F05, 62K05, 62E20, 68U10
1. Introduction
Tensors of second and higher order are ubiquitous in the physical sciences. Some examples include the moment of inertia tensor; electrical, hydraulic, and thermal conductivity tensors; stress and strain tensors, etc. One key advance in the field of tensor measurement was the advent of diffusion tensor imaging (DTI), a magnetic resonance–based imaging technique that provides an estimate of a second-order diffusion tensor in each voxel within an imaging volume [5, 6]. This effectively provides discrete estimates of a continuous or piecewise continuous tensor field within tissue and organs. With the possibility of measuring tensors in millions of individual voxels within, for example, a live human brain, there is a clear need for a statistical framework to be developed to (a) design optimal DTI-experiments, (b) characterize central tendencies and variability in such data, and (c) provide a family of hypothesis tests to assess and compare tensors and the quantities derived from them.
1.1. Tensor-variate normal distribution
In DTI, a tensor D is represented by a symmetric matrix D = (Di,j : 1 ≤ i ≤ j ≤ 3), and it has been established that the measured tensor components Dij, over multiple independent acquisitions from the same subject in the same voxel, conform to a multivariate normal distribution [34]. We previously proposed a normal distribution for tensor-valued random variables that arise in DTI whose precision and covariance structures could be written as fourth-order tensors [10]:
where A is a fourth-order precision tensor, D̄ is the mean tensor, and “:” is a tensor contraction.
There are distinct advantages to analyzing tensor or tensor-field data in the laboratory coordinate system in which their components are measured, and using the tensor-valued variates with a fourth-order tensor precision tensor rather than writing the tensor as a vector and using a square covariance matrix. For example, by retaining the tensor form, it is easy to establish the condition that the statistical properties be coordinate independent, yielding an isotropic fourth-order precision tensor
which can be parameterized with only two constants, μ and λ. This form, if achieved, can greatly simplify statistical analysis and is the focus of this paper.
In the following sections, we switch from tensor to matrix notation [10], as the correspondence between the Gaussian tensor-variate and standard multivariate normal can be established using appropriate conversion factors [12]. The outline of the paper is as follows. First, in this section we state the properties for the m-dimensional isotropic Gaussian matrix. In section 2 we describe a spectral representation and change of variables applicable to general symmetric random matrices. In section 3 we derive distributions for the eigenvalues and eigenvectors of the isotropic Gaussian, while in section 4 we obtain the analytical expressions in the limit of small noise for different symmetries of the mean tensor D̄. In the remaining sections, we focus on the application of these results to DTI. In section 5 we develop a sphericity test, testing for the isotropy of the diffusion tensor; in section 6 we study the isotropy of the Fisher information and justify the use of spherical t-designs as gradient tables in DTI experimental design; and finally, in section 7 we test many of the mathematical results and predictions using Monte Carlo simulations of the DTI-experiment. The main theorems are proved in Appendix SM1 of the supplementary material.
1.2. Isotropic Gaussian matrix distribution
Given a fixed symmetric matrix D̄ ∈ ℝm×m, it is shown in [31, 10] that the probability distribution of an m×m symmetric Gaussian random matrix D = (Dij : 1 ≤ i ≤ j ≤ m) is isotropic around D̄ if and only if it has density of the form
| (1) |
| (2) |
with precision parameter μ > 0 and interaction parameter λ satisfying the constraint λm > −2μ. To fix the ideas, when m = 3 this corresponds to a Gaussian distribution for the vectorized matrix
| (3) |
with mean vec(D̄) and precision matrix
| (4) |
In particular (Dij : 1 ≤ i < j ≤ m) are independent, and (Dii : 1 ≤ i ≤ m) are negatively correlated for λ > 0, with covariance Σ(μ, λ) = A(μ, λ)−1, where
Remark 1.1
When D̄ = 0, λ = 0, and μ = 1 or, depending on the scaling convention, μ = 1/2, the random matrix distribution (1) is known in the literature as the Gaussian orthogonal ensemble (GOE). The connection between general isotropic Gaussian matrices and the GOE was first noticed in [37]. The fluctuations of the diagonal elements ((Dii − D̄ii) : 1 ≤ i ≤ m) are exchangeable and independent of the off-diagonal elements.
2. Spectral representation and change of variables
We summarize basic facts from the random matrix literature [3, 21, 25, 32, 22, 16, 24, 40]. A symmetric matrix D ∈ ℝm×m has spectral decomposition D = OGO⊤, where G is a diagonal matrix containing the m eigenvalues (γ1, γ2, … , γm) ∈ ℝm, and O = (O−1)⊤ is an orthogonal matrix with columns corresponding to the normalized eigenvectors. The orthogonal matrices form a compact group 𝒪(m) with respect to the matrix multiplication, which contains the special orthogonal group 𝒮𝒪(m) = {O ∈ 𝒪(m) : det(O) = 1} of rotations. The (m − 1)m/2 independent entries under the diagonal (Oij : 1 ≤ j < i ≤ m) determine O, and the eigenvalues are distinct for symmetric matrices outside a set of Lebesgue measure zero in ℝ (m+1)m/2. The spectral decomposition is not unique, since D = OGO⊤ = ROPGP⊤O⊤R⊤ for any permutation matrix P, and any R = (Rij = ±δij)1≤i≤j≤m, which form the subgroup ℛ(m) of reflections with respect to the Cartesian axes, isomorphic to {1,−1}m. In order to determine unique O and G, we sort the eigenvalues in descending order γ1 > γ2 > · · · > γm and impose, for each column vector (O1j,O2j, … , Omj)⊤, j = 1, … , m, the condition that the first encountered nonzero coordinate is positive, and denote by 𝒪(m)+ the set of such matrices. An O ∈ 𝒪(m)+ is a representative of the left coset Oℛ(m). The change of variables
| (5) |
has differential
where J is the Jacobian of the inverse map X ↦ D, which is evaluated by means of differential geometry. We consider a differentiable map Y : ℝm×m → ℝm×m. The matrix differential can then be written using the chain rule
and the wedge product acting on the transformed differentials is
Note that the wedge product is taken over the independent entries of the matrix, for example, if X is symmetric,
and when X is skew-symmetric,
The wedge product is also anticommutative, meaning that dx∧dy = −dy∧dx. However, when we compute volume elements, we always choose an ordering of the wedge product producing a nonnegative volume. The Jacobian calculation is based on the following result.
Proposition 2.1 (see [24, Prop. 1.2])
When A,D are m×m matrices and D is symmetric,
| (6) |
Since O⊤O = I, it follows that the matrix differential O⊤dO = −dO⊤O is skew-symmetric. We also have
and
where the differential matrix on the right-hand side has diagonal entries dγi and off-diagonal entries,
By using property (6), we obtain
| (7) |
is the Vandermonde determinant. The wedge product (O⊤dO)∧ defines a uniform measure on 𝒪(m) which is invariant under the group action, and the Haar probability measure given by
is obtained by normalizing with the volume measure (see Corollary 2.1.16 in [33])
We rewrite (7) as
| (8) |
3. Eigenvalue and eigenvector distribution
3.1. Zero-mean isotropic Gaussian matrix
We consider first a zero-mean symmetric random matrix D with isotropic Gaussian distribution (1), where D̄ = 0. This is an important special case to consider. While it does not satisfy the physical requirement that the eigenvalues of a diffusion (or other transport) tensor all must be nonnegative, it illustrates the mathematical machinery necessary to derive a closed-form expression for the resulting distribution of tensor eigenvalues. From the spectral decomposition D = OGO⊤, it follows by using the change of variables (5) in the density (1) that O is independent from G and represents a random rotation distributed according to the constrained probability
and the ordered D-eigenvalues have joint density on {γ ∈ ℝm : γ1 > · · · > γm},
| (9) |
with normalizing constant
| (10a) |
| (10b) |
Remark 3.1
The density (9) is not generally Gaussian, since the Vandermonde determinant induces repulsion between the eigenvalues, which are never independent, even in the case when λ = 0 and the diagonal elements Dii are independent. When λ = 0, after rescaling, (9) is the well-known GOE eigenvalue density, which plays a special role below (see Theorem 4.1). For m = 3, .
3.2. General case
Theorem 3.2
Let D̄ ∈ Rm×m be a symmetric matrix with a spectral decomposition D̄ = ŌḠŌ⊤, where Ḡ = diag(γ̄1, γ̄2, … , γ̄m), γ̄1 ≥ γ̄2 ≥ ·· · ≥ γ̄m, are the ordered eigenvalues of D̄ , and Ō ∈ 𝒪(m)+ (which is not uniquely determined when there are repeated eigenvalues), and let D be a symmetric m×m Gaussian matrix with density (1) isotropic around the mean value D̄. Then, the ordered D-eigenvalues γ1 > γ2 > · · · > γm have joint density
| (11) |
and ℐm is the spherical integral below known as the Harish–Chandra–Itzykson–Zuber (HCIZ) integral [41, 26]:
Conditionally on the eigenvalues (γ1, … , γm), the conditional probability of R = Ō⊤O has density
| (12) |
with respect to the Haar probability measure Hm(dR) on Ō⊤𝒪(m)+.
Proof
As in the zero-mean case, we start from the isotropic Gaussian matrix density (1) with mean D̄. By using the spectral representations D = OGO⊤ and D̄ = ŌḠŌ⊤, after the change of variables described in section 2, we find the joint density of (G,O) with respect to the product measure
given as
We change coordinates with O ↦ R = Ō⊤O ∈ 𝒪(m)+, and using the invariance property of the Haar measure we see that
which proves (11). In the new coordinates the random matrix has density
| (13) |
with respect to dγ × Hm(dR) on {γ ∈ ℝm : γ1 > γ2 > · · · > γm}× Ō⊤𝒪(m)+ which proves (12).
Remark 3.3
When Ḡ = γ̄Id we say that D̄ is spherical. In such a case, G is stochastically independent from O, which follows the Haar probability distribution. Equation (11) shows the density of the ordered eigenvalues. Often the random matrix literature deals with the density of the unordered eigenvalues on ℝm, which depends only on the order statistics and differs by a 1/m! factor. The HCIZ integral admits the series expansion
where the sum is over the set of partitions of k into at most m parts,
and Cα(z1, … , zm) is the homogeneous zonal polynomial corresponding to the partition α [28, 33, 39, 23]. Theorem 4.6 deals with the second-order asymptotics of ℐm(nγ, γ̄) as n→∞. When m = 3,
expressed in Euler angular coordinates.
4. Small noise asymptotics
4.1. Spectral grouping
Theorem 4.1
Let (D(n), n ∈ ℕ) be a sequence of random m×m symmetric matrices such that, for some deterministic limit D̄ and scaling sequence a(n) →∞,
| (14) |
where vec(X) is Gaussian with zero-mean and covariance Σ(1, λ) for some λ > −2/m as in (4).
Denoting by ( , 1 ≤ j ≤ m) and (γ̄j : 1 ≤ j ≤ m) the ordered eigenvalues of D(n) and D̄ , respectively, assume that D̄ has k distinct eigenvalues, i.e.,
with 1 ≤ k ≤ m, ℓ0 = 0, ℓk = m, corresponding to eigenspaces of respective dimensions mi = (ℓi − ℓi−1). Consider the clusters
formed by the ordered eigenvalues of D(n) corresponding to the eigenspaces of D̄ taken in the D̄-eigenvalue order, and define the corresponding cluster barycenters as
We also consider the eigenvalue fluctuations
and the cluster barycenter fluctuations
As n→∞, the following limiting distribution appears:
-
For the cluster barycenters, we havewhere
(15) have joint Gaussian density(16) with zero-mean and covariance -
For each cluster, the differences between the eigenvalues and their barycenterare asymptotically independent from their cluster barycenter and the other clusters, with limiting distribution
(17) where (γ1 > γ2 > · · · > γmi) are eigenvalues of the standard mi-dimensional GOE of symmetric Gaussian matrices with zero-mean and precision Ami (1, 0) with barycenterMoreover, the differences (γ1 − γ̃mi, … , γmi − γ̃mi) are independent from γ̃mi, with degenerate density(18) where δ0(z) denotes the Dirac distribution, which is also the conditional density of the GOE eigenvalues (γ1, … , γmi) conditioned on {γ1 + · · · + γmi = 0}.
-
In particular, for each cluster,
and these eigenvalue differences are asymptotically independent from the cluster barycenter and the other clusters.
Remark 4.2
The weak convergence hypothesis (14) implies
which means that and in probability. The asymptotic distribution in (17) depends only on mi (the size of the cluster) and not on the interaction parameter λ. When has an isotropic Gaussian distribution with covariance Σ(1, λ), and the mean D̄ = γ̄I is spherically symmetric, there is only one cluster, and the distributional equalities in Theorem 4.1 hold exactly without going to the limit in distribution. A related result is given in [44] for the joint asymptotic distribution of eigenvalues and eigenvectors. Similar results have been derived in the special cases of noncentral Wishart random matrices and sample covariance matrices which are asymptotically Gaussian [2], [33, Theorem 9.5.5].
Next, we illustrate the implications of Theorem 4.1 in the 3-dimensional situation, which is relevant for DTI.
Corollary 4.3
Let D be a 3 × 3 symmetric matrix with Gaussian density (1). As μ→∞ with λ > −2μ/3, we have four asymptotic regimes depending on the symmetries of the mean matrix D̄.
- γ̄1 > γ̄2 > γ̄3 (totally asymmetric tensor). The joint density of (γ1, γ2, γ3) is approximated by the Gaussian density of (D11,D22,D33), i.e.,
(19) -
γ̄1 > γ̄2 = γ̄3 (prolate tensor). Let γ̃23 = (γ2+γ3)/2. The joint distribution of (γ1, γ̃23) is approximated by the Gaussian distribution of (D11, (D22 + D33)/2, i.e.,
(20) Conditionally on (γ1, γ̃23), the asymptotic distribution of (γ2, γ3) is degenerate, with γ3 = (2γ̃23 − γ2) and(21) that is, , with τ exponentially distributed with rate 2μ and independent from the barycenter γ̃23.
-
γ̄1 = γ̄2 > γ̄3 (oblate tensor). This is similar to the prolate case. Let γ̃12 = (γ1+γ2)/2. Asymptotically the joint distribution of (γ̃12, γ3) is approximated by the Gaussian distribution of ((D11 + D22)/2,D33, with
(22) and the asymptotic conditional distribution of (γ1, γ2) given (γ̃12, γ3) is degenerate with γ2 = (2γ̃12 − γ1), and(23) i.e. , with τ exponentially distributed with rate 2μ, independent from γ̃12.
-
γ̄1 = γ̄2 = γ̄3 (isotropic tensor). The barycenter is Gaussian with mean γ̄1 and variance 1/(6μ + 9λ). Conditionally on γ̃123, (γ1, γ2, γ3) is degenerate, with γ2 = (3γ̃123 −γ1 −γ3), and the conditional density of (γ1, γ3) given γ̃123 is approximated as
(24) Asymptotically, the conditional distribution of the vector
coincides with the conditional distribution of the ordered eigenvalues of the 3-dimensional standard GOE, conditioned on having zero barycenter, and is independent of γ̃123.
Remark 4.4
For a totally anisotropic mean tensor D̄, the asymptotic Gaussian density (19) for the rescaled eigenvalue fluctuations around their barycenter coincides with the Gaussian eigenvalue density (18) of [10]. However, in [10] it was erroneously postulated that the map D = (OGO⊤) ↦ G was linear with constant Jacobian and (19) would be the eigenvalue density of a random tensor with isotropic Gaussian noise; in fact, in the nonasymptotic case the eigenvalue density is given by (11).
4.2. Axial and radial diffusivity marginals
Two eigenvalue statistics that are particularly relevant in DTI are axial diffusivity (AD), which corresponds to the largest D-eigenvalue γ1, is measured along the principal axis of the diffusion tensor, and is considered a putative axonal damage marker, and radial diffusivity (RD), which corresponds to γ̃23 = (γ2+γ3)/2, is measured perpendicular to the principal axis, and is thought to be sensitive to the degree of hindrance that diffusing water molecules experience due to the axonal membrane and myelin sheath. In this subsection we derive the distributions for AD and RD in dimension m = 3 when D has the density given in (1). When the mean matrix D̄ is prolate, we show in Corollary 4.3 that in the small noise limit the joint distribution of AD and RD is asymptotically Gaussian, as shown in (20).
In the case of D with spherical mean D̄ = γ̄Id, we can also derive the marginal densities of AD and RD. See also [15], which contains a recursive expression for the distribution of the largest GOE eigenvalue in arbitrary dimension. After changing variables in the joint conditional eigenvalue density (24), we see that zi = (γi − γ̃123) are independent of the barycenter γ̃123, z1 = (γ1− γ̃123) and (−z3) = (γ̃123−γ3) are identically distributed, with marginal density
and cumulative distribution function
where
denote the standard Gaussian density and cumulative distribution function, respectively. The cumulative distribution function of γ1 is obtained by taking convolution with the barycenter γ̃123 distribution 𝒩(γ̄, 1/(6μ + 9λ)), as
The joint density of AD and RD is given by
4.3. Eigenvector asymptotics
In the settings of Theorem 4.1, where D(n) and D̄ have respective spectral decompositions O(n)G(n)O(n)⊤ and ŌḠŌ⊤, we study the asymptotics of R(n) = Ō⊤O(n) ∈ 𝒪(m). Omitting the n superscript, we use the decomposition R = ŘR̂, where
| (25) |
is block diagonal with blocks Ř(j,j) ∈ 𝒪(mj) corresponding to the mj-dimensional eigenspaces of D̄.
These matrices form a subgroup 𝒦γ̄ ≃ 𝒪(m1)×𝒪(m2)×· ··×𝒪(mk) such that ŘD̄Ř⊤ = D̄ for all Ř ∈ 𝒦γ̄, and the conditional eigenvector density (12) is invariant under the action of 𝒦γ̄.
R̂ ∈ 𝒮𝒪(m) is a rotation with Lie matrix exponential representation
| (26) |
is skew-symmetric, with blocks for 1 ≤ j < l ≤ k, and zero (mj × mj)-blocks on the diagonal, with free parameters. The subgroup
is a complement subgroup of 𝒦γ̄ in 𝒪(m).
In dimension m = 3, R̂ = exp(Ŝ) is a clockwise rotation by an angle around the unit vector u = (Ŝ23,− Ŝ13, Ŝ12)/θ. The matrix exponential exp(dŜ) of an infinitesimal 3 × 3 skew-symmetric matrix is the composition of three infinitesimal rotations around the Cartesian axes x, y, z by the Euler angles dŜ23 (roll), dŜ13 (pitch), and dŜ12 (yaw), respectively, which commute up to infinitesimals of higher order.
Theorem 4.5
In the settings of Theorem 4.1, let
with Ř(n) ∈ 𝒦γ̄ and Ŝ(n) skew-symmetric. The blocks ,i = 1,…, k, corresponding to the D̄ eigenspaces are asymptotically distributed according to the product of the Haar measures on the respective orthogonal groups 𝒪(mi), with the constraint ŌŘ(n) ∈ 𝒪(m)+, and asymptotically independent from the eigenvalue fluctuations.
After rescaling, the entries ( ) are asymptotically mutually independent, and independent of Ř(n) and the eigenvalue fluctuations, with limiting Gaussian distribution
Remark
Theorem 4.5 extends Theorem 4.1 in [37] for D̄ with nonnegative distinct eigenvalues, given also in [35], to the case with repeated eigenvalues.
4.4. Second-order approximation of the HCIZ integral
Theorem 4.6
Let γ, γ̄ ∈ ℝm be ordered vectors such that the coordinates (γ1 > γ2 > · · · > γm) are distinct, while the γ̄ coordinates may coincide, with multiplicities mi = (ℓi− ℓi−1) and
for 0 = ℓ0 < ℓ1 < · · · < ℓk = m, 1 ≤ k ≤ m. Then, as n→∞,
| (27) |
Remark 4.7
Theorem 4.6 was proven by [2] (see also [33, Theorem 9.5.2]) in the case of nonnegative eigenvalues without multiplicities.
5. Testing the sphericity hypothesis
In DTI, it is often desirable to establish different symmetries of the underlying tensor field. One of the often-used tests is that of isotropy of the underlying mean diffusion tensor [6, 17]. Here we also develop one such test and call it a test of sphericity to avoid confusion with the “isotropy” of the precision tensor. Consider a sequence of random symmetric matrices D(n) such that , where the limit is a zero-mean Gaussian symmetric matrix, D̄ is deterministic, and a(n) →∞ is a scaling sequence. For example, in section 6 the scaling sequence is given by the number of gradients in the DTI measurement. In order to test the sphericity hypothesis
we introduce the sampled eigenvalue central moments
where γi are the eigenvalues of D.
Lemma 5.1
κr(D) is a homogeneous polynomial of degree r in the matrix entries, satisfying for all c ∈ ℝ
| (28) |
This implies that the derivatives satisfy ∇ℓκr(Id) = 0 for all 0 ≤ ℓ < r, while ∇rκr(D) = ∇rκr(0) are constant tensors such that
Corollary 5.2
Let D(n) be a sequence of m × m symmetric random matrices, and let X be a zero mean symmetric Gaussian matrix such that for some γ̄ ∈ ℝ and scaling sequence a(n) →∞,
Then
When the covariance of X is isotropic, (κr(X) : 2 ≤ r ≤ m) are stochastically independent from κ1(X).
Proof
For the first statement we apply the continuous mapping theorem together with (28). If X has zero-mean isotropic Gaussian distribution, the conditional distribution of (X −κ1(X)Id) given κ1(X) is also zero-mean isotropic Gaussian and does not depend on the value of κ1(X).
To test the sphericity hypothesis with γ̄ ≠ 0, it is natural to use statistics of the form
and calibrate the test against the distribution of
| (29) |
evaluated at c = κ1(D(n)). However, without additional assumptions on the covariance structure of X, the probability density functions of κr(X) for r ≥ 2 do not have closed-form expressions and can be computed only numerically, for example, by Monte Carlo simulations. Note also that, since ∇ℓκr(Id) = 0 for all r ≥ 2, 0 ≤ ℓ < r, we are dealing with a singular hypothesis testing problem [19, 20, 43], where the constraints {κr(D̄ ) = 0, r ≥ 2} for which we are testing are singular at the true parameter D̄ = γ̄ Id; consequently, any smooth sphericity statistics τ(n) will follow non-Gaussian higher order asymptotics. We proceed now in dimension m = 3, assuming that the Gaussian matrix limit X has zero-mean and isotropic precision matrix A(1, λ) with λ > −2/3, to explicitly compute the asymptotic density of some commonly used sphericity statistics based on eigenvalues sample mean, variance, and skewness.
Lemma 5.3
In the settings of Theorem 4.1, under the sphericity hypothesis H0, the test statistics
| (30) |
are asymptotically independent, with limiting distributions
| (31) |
In dimension m,
with asymptotically independent components.
Proof
We start from the asymptotic eigenvalue density (11), which under H0 is given by
and we apply the continuous mapping theorem [42] to the smooth bijection
By changing variables, the Vandermonde determinant cancels out, and the resulting joint central moments density is given by
It follows by an optimization argument that the support of the κ3 conditional distribution given κ2 is the interval [ ]. We do a further change of variables, setting κ′ = (κ1, κ2, τ3) with , obtaining
| (32) |
which factorizes as the distribution of independent random variables κ1 ~ 𝒩(κ1(γ̄), 1/(6μ + 9λ), and τ3 uniformly distributed on [−1, 1].
Related ellipticity and sphericity measures are fractional anisotropy [7]
relative anisotropy [10]
and volume ratio [36]
Corollary 5.4
In the settings of Theorem 4.1 with dimension m = 3, under the sphericity hypothesis H0, there are two possible asymptotic regimes:
-
When D̄ = 0, the sequence of statistics
(33) converges jointly in distribution to the random vectorwith independent and U ~ Uniform[−1, 1].
-
Otherwise, the rescaled statistics
(34) (35) (36) are asymptotically equivalent, within probability, and .
Remark 5.5
Corollary 5.4 generalizes Theorem 8.3.7 in [33] on VR asymptotics without positivity assumptions. In order to use the VR statistics to test the isotropy of the mean D̄, one should first test the hypothesis κ1(γ̄) = 0, under which
If this hypothesis is accepted, we assume that we are in the asymptotic regime (1) and construct a conditional sphericity test by using the conditional distribution of VR(γ(n)) given {a(n)κ1(γ(n))2 = t}, which converges in distribution to the law of
with independent from U ~ Uniform[−1, 1]. If the hypothesis κ1(γ̄) = 0 is rejected, we use the rescaled VR statistics in (35).
Eigenvalue central moment statistics have been considered earlier in the DTI literature, the distribution of Tr(D) for D isotropic Gaussian is derived in [11], the variance is discussed in [7, 44, 37], and skewness is explored in [8]. Note that under H0, the limit laws of are parameter free. However, evaluating requires knowledge of the scaling sequence normalization, while evaluating does not. can be used as two-sided test statistics, accepting the sphericity hypothesis with confidence level α when . The left-tail rejection region corresponds to the anomalous situation, with , and the right tail corresponds to or . We can test for symmetries with a sequence of confidence levels , with c(n) →∞ and c(n)/a(n) → 0, and construct an asymptotically superefficient eigenvalue estimator γ̂(n) as follows:
If κ2(γ(n))< c(n)/(6a(n)), accept the isotropy hypothesis and set ;
-
else if
accept the oblate tensor hypothesis and set ;
else ifaccept the prolate diffusion tensor hypothesis and set ;
otherwise, reject the hypothesis that the tensor has symmetries and use the unmodified estimator γ̂(n) = γ(n).
The situation with mean matrix D̄ = 0 arises in two-sample problems. Consider two m × m symmetric random matrices D′,D″, which are measured with independent and isotropic Gaussian noises, with precision matrices A(μ, λ) and A(μ, λ) and means D̄′, D̄″, respectively. Their difference D = (D′ − D″) is again symmetric Gaussian with mean D̄ = (D̄′ − D̄″) and isotropic precision matrix A(μ, λ), with parameters
In order to test the hypothesis D̄′ = D̄″, one could use the statistics
| (37) |
Testing equality in distribution of two sample matrix eigenvalues and eigenvectors separately was discussed in [37], under the hypothesis of asymptotically Gaussian and isotropic error, and was generalized in [38] to nonisotropic error covariances.
6. Asymptotic statistics in DTI under Rician noise
We consider an ideal DTI-experiment with measurements following the Rician likelihood,
| (38) |
where S is the signal, Y is the observation, η2 is the noise parameter, and Iℓ(z) is the modified Bessel function of the first kind of order ℓ. The signal is determined by the second-order tensor model
| (39) |
where D is the (symmetric) diffusion tensor, ρ is the unweighted reference signal, and g is the applied magnetic field gradient. The function g ↦ S(g,D)/ρ is interpreted as the Fourier transform of the displacement distribution of a water molecule undergoing Gaussian diffusion in a unit time interval, and the problem is to estimate the diffusion tensor D from the noisy spectral measurements Y. For fixed ρ and η2 we denote the log-likelihood of D as
The observed information with respect to the tensor parameter D is given by
and the Fisher information is obtained by integrating out the data Y with respect to (38) under the signal model (39) with tensor parameter D, as
| (40) |
depending on the signal to noise ratio (SNR) S/η of the complex Gaussian error model through the weight function
see [27]. Note that necessarily, Jij,ij(D) = 4Jii,jj(D) for all 1 ≤ j < i ≤ 3. By replacing the Rician density (38) with another likelihood which is a function of the SNR, we always obtain Fisher information of the form (40), with a different weight function.
We now consider a sequence of DTI-experiments, with measurements ( ) from respective signals ( ), corresponding to the gradients , and denote the scaled Fisher information as
| (41) |
Assume that M(n) → ∞ and that the sequence of discrete gradient distributions
converges weakly to a probability π on ℝ3, which implies
| (42) |
Let D(n) be a regular statistical estimator of the tensor parameter such as, for example, the maximum likelihood estimator (MLE), the penalized MLE, the Bayesian maximum a posteriori estimator (MAP), or the posterior mean, based on the data ( ) with gradients ( ). When 0 < det(J(∞)) < ∞, under the tensor model with true parameter D̄, all these regular estimators are consistent with asymptotically Gaussian error such that
| (43) |
6.1. Isotropic Gaussian limit error distribution
When J(∞)( D̄) = A(μ̄, μ̄) as in (4) for some μ̄ > 0, the Gaussian limit distribution (43) is isotropic. In such a case, Theorem 4.1, Corollary 4.3, and Lemma 5.3 apply with a(n) = μ̄M(n) and λ = 1. When the true tensor D̄ = γ̄ I is isotropic and the asymptotic gradient design distribution π(dg) is radially symmetric, asymptotic isotropy is achieved with
| (44) |
where b = ||g||2, referred to as the b-value, is integrated with respect to
and u = g/||g|| has uniform distribution σ(du) on the surface of the unit sphere 𝒮2 = {u ∈ ℝ3: ||u|| = 1}. A more general condition implying (44) is the following: the asymptotic gradient design distribution decomposes as
| (45) |
where for ν-almost all b-values, the conditional probability on 𝒮2 is such that
| (46) |
for all homogeneous polynomials f(u1, u2, u3) of degree t = 4.
Proposition 6.1
When the true diffusion tensor D̄ is isotropic, the uniform gradient distribution σ(du) maximizes det(J) among all probability distributions on the unit sphere.
Proof
When J is invertible, we have [30, Theorem 8.1]
| (47) |
which implies that the function J ↦ log det(J) ∈ ℝ ∪ {−∞} is concave, and that a local maximum is also a global maximum. Let ν(du) be a probability measure on 𝒮2, and consider a small perturbation of the uniform measure σ in the direction ν. By taking the differential using (47), we obtain
| (48) |
where, since J−1(σ) is also isotropic, for every u, v ∈ 𝒮2 we have
and the integrand in (48) is constant, which means that det (J(σ)) is a global maximum.
This shows that when the true tensor D̄ is isotropic, asymptotically uniform gradient designs are most informative, minimizing the Gaussian entropy of the asymptotic estimation error
In the next section, we introduce discrete gradient distributions which attain the same bound.
6.2. Spherical t-designs in diffusion tensor imaging
A spherical t-design ϒ ⊂ 𝒮m−1 is a finite subset of m-dimensional unit vectors with the property
| (49) |
for all polynomials f(u1, …, um) of degree r ≤ t, where σ is the uniform probability measure on 𝒮m−1, and #ϒ is the number of points in ϒ. In other words, a spherical t-design is a quadrature rule on Sm−1 with constant weights. The algebraic theory behind such designs is deep and beautiful [18]; for a recent survey see [4, 1]. In particular, in dimension m = 3, spherical t-designs of order t ≥ 4 satisfy (46). A database of spherical t-designs on 𝒮2 computed by Rob Womersley is available on his webpage http://web.maths.unsw.edu.au/~rsw/Sphere/EffSphDes/. Table 1 displays the sizes of these designs, and Figure 1 shows a spherical t-design of order 4 with 14 gradients from Womersley’s database.
Table 1.
Number na of points in some known antipodal spherical t-designs of order 4 ≤ t ≤ 17 in 𝒮2, computed by Rob Womersley, while n is for his nonantipodal spherical t-designs.
| Size of Spherical t-Designs | ||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| t | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 |
| na | - | 12 | - | 32 | - | 48 | - | 70 | - | 94 | - | 120 | - | 156 |
| n | 14 | 18 | 26 | 32 | 42 | 50 | 62 | 72 | 86 | 98 | 114 | 128 | 146 | 163 |
Figure 1.
A nonantipodal spherical t-design of order 4, with 14 gradients, by Rob Womersley.
When ϒ = −ϒ, we say that the spherical design is antipodal. Two well-known examples (see [10, 13]) are the regular icosahedron and its dual, the regular dodecahedron, whose vertices form antipodal spherical t-designs of order 5 with sizes 12 and 20, respectively. Note that any two antipodal gradients produce the same DTI-signal. Starting from an antipodal spherical t-design ϒ and selecting one gradient from each antipodal pair {u,−u} ⊂ ϒ, we obtain a design ϒ′ of size #ϒ′ = #ϒ/2 which satisfies (49) for all homogeneous polynomials of even degree ≤ t. Figures 2 and 3 show, respectively, the intersection of the northern hemisphere with the regular icosahedron and the dodecahedron, forming gradient designs of sizes 6 and 10, which satisfy (49) for all homogeneous polynomials of degrees 2 and 4.
Figure 2.
Gradient design based on the icosahedron with six gradients on the northern hemisphere.
Figure 3.
Gradient design based on the dodecahedron with 10 gradients.
In the DTI-experiment, for a finite subset of b-values and respective spherical t-designs of order , we construct the gradient set as the union of shells
The resulting gradient distribution
satisfies (45), and when the true tensor D̄ = γ̄ I is totally symmetric, we have
with
i.e., the Fisher information coincides with the precision matrix of an isotropic Gaussian matrix distribution. When ϒ ⊂ S2 is a spherical t-design and O ∈ SO(3) is a rotation matrix, the rotated design Oϒ is a spherical t-design as well. Since the true tensor D̄ is unknown (and possibly not isotropic), in practice it is advisable to choose the gradient directions covering 𝒮2 as uniformly as possible. To achieve this, different t-designs can be rotated with respect to one another in order to maximize the spread between gradient directions. Namely, starting from a collection of spherical t-designs of respective orders tk, 1 ≤ k ≤ n, we find the optimized design , 1 ≤ k ≤ n, where are rotation matrices maximizing
| (50) |
with dist(U, V ) = supu∈U,v∈V dist(u, v), and dist(u, v) is the geodesic distance on 𝒮2. The maximizer can be achieved by a greedy iterative algorithm, where in turn (50) is optimized with respect to each single Ok keeping fixed the other rotations until convergence to a fixed point. Figure 4 shows a gradient sequence obtained in such a way, with colors corresponding to spherical t-designs on different shells. The benefits of these gradient designs are illustrated in the next section.
Figure 4.

Gradient sequence based on combined antipodal spherical t-designs of orders 5 (black), 7 (red), 9 (green), and 11 (blue), of respective sizes 12, 32, 48, and 70. The spherical t-designs on different shells were rotated in order to maximize the minimal geodesic distance (50) between gradients.
7. Illustration of the methods
7.1. Monte Carlo study with isotropic Gaussian noise
Figure 5 shows the results from a Monte Carlo study with a sample of N = 10000 i.i.d. 3×3 symmetric random matrices from the isotropic Gaussian density (1) with precision parameters μ = 1/2, λ = 0 for the following various choices of the diagonal mean matrix:
Figure 5.
10000 pairs of distinct eigenvalues of i.i.d. symmetric random matrices with isotropic Gaussian noise (μ = 1/2, λ = 0) with various mean: zero, corresponding to the 3×3 GOE (a), isotropic (b), prolate (c), oblate (d). For comparison we show i.i.d. 2 × 2 GOE eigenvalue pairs (e), and i.i.d. standard Gaussian pairs (f). Within each pair the ordering is randomized to emphasize the repulsion effect around the diagonal.
D̄ = 0, corresponding to the 3 × 3 GOE;
D̄ isotropic, with γ̄1 = γ̄2 = γ̄3 = 15;
D̄ prolate, with γ̄1 = 15 > γ̄2 = γ̄3 = 3;
D̄ oblate, with γ̄1 = γ̄2 = 15 > γ̄3 = 3.
For comparison, in Figure 5e we show i.i.d. eigenvalue pairs from the 2 × 2 Gaussian orthogonal ensemble, and in Figure 5f we show i.i.d. pairs of independent standard Gaussian random variables. The empirical joint eigenvalue distribution avoids the diagonal, in agreement with (9). We see that the fluctuations of the eigenvalues corresponding to the same D̄ eigenspaces around their mean are distributed like the GOE corresponding to the dimension of the eigenspace. One can see also some differences between the GOE eigenvalue distribution in dimension 2 (in Figure 5e, sampled with precision parameters μ = 1/2, λ = 0, which agrees with Figures 5c and 5d), and dimension 3 (in Figure 5a, which agrees with 5b).
Figure 6 shows that in the case with prolate mean matrix, the empirical distribution of the cluster barycenter (γ2 + γ3)/2 fits the Gaussian distribution very well.
Figure 6.

Histogram and fitted Gaussian curve from 10000 i.i.d. realizations of the cluster barycenter (γ2 + γ3)/2 in the prolate mean tensor case.
Figure 7 shows the behavior of the sphericity test statistics τ2, τ4, τ5 under Gaussian matrix distributions, with the same isotropic precision matrix A(2, 2) and different means, namely, a spherical mean tensor and 15 prolate mean tensors, all with the same mean diffusivity κ1(D̄) = 15, and FA in (0.01, 0.15]. We can see that at this noise level, under the null hypothesis, the distributions of these three test statistics fit the asymptotic distribution very well, while under prolate alternatives the corresponding sphericity tests have approximately the same power at all significance levels.
Figure 7.
Probability densities (left) and cumulative probabilities (right) of the sphericity test statistics τ2(D), τ4(D), τ5(D), where the 3 × 3 symmetric random matrix D is Gaussian with isotropic precision A(2, 2), and there are 16 alternative mean tensors D̄, with fixed mean diffusivity κ1(D̄) = 15. Under the null hypothesis, D̄ is spherical, while the alternatives correspond to prolate mean tensors with FA in (0.0, 0.15]. For each test statistics, the probability density and cumulative probability curves are labeled by the FA values of the corresponding mean tensors. The broken curves display the limit distribution under the null hypothesis.
Figure 8b displays on the unit sphere the orthonormal eigenvector triples from the Gaussian model with isotropic noise parameters μ = 1/2, λ = 0, with N = 200 i.i.d. replications. On the left side of the figure, the mean tensor is diagonal and totally anisotropic with γ̄1 = 15, γ̄2 = 7.5, γ̄3 = 3. On the right, the mean tensor is diagonal and oblate, with γ̄1 = γ̄2 = 15, γ̄3 = 3, and the eigenvectors corresponding to the first two eigenvalues are uniformly distributed around the equator.
Figure 8.
200 i.i.d. orthonormal eigenvector triples from the Gaussian model with isotropic noise parameters μ = 1/2, λ = 0, with totally asymmetric (left) and oblate (right) diagonal mean tensor, using a graphical construction similar to the one introduced in [9].
7.2. Monte Carlo study of sphericity test statistics based on DTI data with Rician noise
In order to validate the asymptotic results of Lemma 5.3 and Corollary 5.4, we conducted another large Monte Carlo study, with DTI data simulated under the Rician noise model with ground truth parameters η2 = 64.056, ρ = 110.046 and isotropic diffusion tensor D̄ = 6.622×10−4×Id mm2/s. For each of the experimental designs 1–5 below, which have an increasing number of acquisitions, we simulated N = 50000 replications of the dataset, and for each replication we computed the MLE D(n) based on the simulated data by using the Expectation-Maximization algorithm from [29]. The empirical distribution of the sphericity statistics (30) and (35) with their theoretical limit distributions are displayed correspondingly in Figures 9–13.
Figure 9.
Scatterplot of the eigenvalue statistics ( ) in (a) and ( ) in (b), from a Monte Carlo study based on N = 50000 replications of a dataset generated under Design 1, where the true tensor and the Fisher information are isotropic. The histogram density estimators are compared with theoretical limit densities (black continuous curves), which are uniform on the vertical axes and on the horizontal axes. The best-fitting gamma densities (red broken curves) are also shown, with shape parameter 2.4238 and scale parameter 2.0627 in (a) and with shape parameter 2.4566 and scale parameter 2.0137 in (b).
Figure 13.
Scatterplot of the eigenvalue statistics ( ) in (a) and ( ) in (b), from a Monte Carlo study based on N = 50000 replications of a dataset generated under Design 5, with isotropic true tensor and anisotropic Fisher information. The histogram density estimators are compared with theoretical limit densities (black continuous curves), which are uniform on the vertical axes and on the horizontal axes. The best-fitting gamma densities (red broken curves) are also shown, with shape parameter 1.9576 and scale parameter 4.5494 in (a) and with shape parameter 1.9565 and scale parameter 4.2094 in (b).
-
Design 1: Spherical t-design of order 4 with 14 gradients computed by R. Womersley, shown in Figure 1, with b-value 996 s/mm2, and one acquisition at zero b-value, for a total of 15 acquisitions. The corresponding Fisher information is given byand the ML estimator vec(D(n)) has a Gaussian approximation with mean vec(D̄) and isotropic covariance
-
Design 2: This design is based on the icosahedron with the six gradients shown in Figure 2 for each b-value in the set {560, 778, 996, 1276, 1556, 1898, 2240} s/mm2, and one acquisition at zero b-value, for a total of 43 acquisitions. The corresponding Fisher information is given byand the ML estimator vec(D(n)) has a Gaussian approximation with mean vec(D̄) and isotropic covariance
-
Design 3: This design is based on the dodecahedron with the 10 gradients shown in Figure 3 for each b-value in the setand one acquisition at zero b-value, for a total of 71 acquisitions. The corresponding Fisher information is given byand the ML estimator vec(D(n)) has a Gaussian approximation with mean vec(D̄) and isotropic covariance
-
Design 4: Combination of spherical t-designs of orders 5, 7, 9, and 11 shown in Figure 4 on shells corresponding to the b-values {560, 996, 1556, 2240}, respectively, with one acquisition at zero b-value, for a total of 163 acquisitions. The corresponding Fisher information is given byand the ML estimator vec(D(n)) has a Gaussian approximation with mean vec(D̄) and isotropic covariance
-
Design 5: This design consists of three repetitions of the 32 gradients in Figure 14 for each b-value inand three acquisitions at zero b-value, for a total of 1443 acquisitions. The ML estimator vec(D(n)) has a Gaussian approximation with mean vec(D̄ ) and nonisotropic covariance
(51)
Figure 14.
The 32 gradients table used by default with the commercial 3T Philips Achieva MR-scanner.
All scatterplots in Figures 9–13 are consistent with the asymptotic independence of the sphericity statistics and from . When the experimental design is based on spherical t-designs of order t ≥ 4 (Designs 1–4), with isotropic Fisher information, the empirical distributions of and fit the theoretical limit distribution (Figures 9–12). Design 5 has the largest number of acquisitions and is the most informative of all; however, the Fisher information is not isotropic, and Figure 13 shows that the empirical distributions of and do not fit the distribution, with the consequence of underestimating the type I error probability of rejecting an isotropic true tensor. We conclude that the distribution of these sphericity statistics is sensitive to anisotropies of the estimation error distribution. As was shown in section 5, these sphericity test statistics should be calibrated against the law of τ (c + κ1(X), κ2(X), κ3(X)), evaluated at c = κ1(D(n)), where X is the zero-mean symmetric Gaussian matrix with covariance (51).
Figure 12.
Scatterplot of the eigenvalue statistics ( ) in (a) and ( ) in (b), from a Monte Carlo study based on N = 50000 replications of a dataset generated under Design 4, where the true tensor and the Fisher information are isotropic. The histogram density estimators are compared with theoretical limit densities (black continuous curves), which are uniform on the vertical axes and on the horizontal axes. The best-fitting gamma densities (red broken curves) are also shown, with shape parameter 2.4924 and scale parameter 1.9986 in (a) and with shape parameter 2.4993 and scale parameter 1.9896 in (b).
We also remark that in part (b) of Figure 9–11, compared with the uniform density, the histogram estimator of the density shows an increasing linear trend. This linear trend is not as evident in 12, which is based on a larger number of acquisitions, and the distribution of the MLE D(n) is presumably better approximated by a Gaussian than in the previous cases. By taking absolute value , the linear trend cancels out, and the histogram of in part (a) of Figures 9–13 robustly fits the uniform distribution in all the situations we have considered.
Figure 11.
Scatterplot of the eigenvalue statistics ( ) in (a) and ( ) in (b), from a Monte Carlo study based on N = 50000 replications of a dataset generated under Design 3, where the true tensor and the Fisher information are isotropic. The histogram density estimators are compared with theoretical limit densities (black continuous curves), which are uniform on the vertical axes and on the horizontal axes. The best-fitting gamma densities (red broken curves) are also shown, with shape parameter 2.4405 and scale parameter 2.0467 in (a) and with shape parameter 2.4526 and scale parameter 2.0298 in (b).
8. Conclusion
We have considered the problem of estimating the spectrum γ̄1 ≥ γ̄2 ≥ · · · ≥ γ̄m and the eigenvectors of a real symmetric m×m matrix D̄, possibly nonpositive, by the spectrum and the eigenvectors of a consistent and asymptotically Gaussian matrix estimator D(n), assuming that the covariance of the rescaled limit is isotropic. When D̄ has repeated eigenvalues, the delta method does not apply, and the spectrum of the matrix estimator has a non-Gaussian limit distribution. In the limit, the random eigenvalues of D(n) form clusters corresponding to the D̄ eigenspaces, with jointly Gaussian barycenters. Within each cluster, the differences between eigenvalues and barycenter are independent from the barycenter and the other clusters and follow the conditional law of GOE eigenvalues conditioned on having zero barycenter.
In many applications it is important to detect the symmetries of the true matrix parameter D̄, in particular to test whether D̄ is spherical, which leads to singular hypothesis testing problems. A statistical test against D̄-symmetries needs to be calibrated, taking into account the repulsion between the random eigenvalues of D(n) corresponding to the same D̄ -eigenspace. In dimension m = 3, we derived the asymptotic joint distribution of some commonly used sphericity statistics such as fractional anisotropy (FA), relative anisotropy (RA), and volume ratio (VR) under isotropy assumptions. We have also discussed the implications of these general results for the design and analysis of DTI measurements, and we showed that gradient designs based on spherical t-designs have isotropic Fisher information and are asympotically most informative when the true tensor is spherical. A direct application would be in denoising the FA maps derived from diffusion tensor estimates. Testing for sphericity at each volume element with a fixed confidence level corresponds to an FA cut-off threshold which is not constant over the voxels but depends locally on the estimated noise and mean diffusivity parameters. We have seen in the Monte Carlo study that the simulated sphericity statistics are well fit to their theoretical limit distribution when the Fisher information of the experiment was isotropic. However, there was a significant discrepancy under experimental Design 5, with nonisotropic Fisher information. We conclude that these findings give a strong theoretical argument in favor of using spherical t-designs in DTI, and we plan to conduct similar experiments with real DTI data in the near future. Finally, our work in progress is to generalize this theory to situations in which the covariance of the Gaussian limit matrix has symmetries without being fully isotropic.
Supplementary Material
Figure 10.
Scatterplot of the eigenvalue statistics ( ) in (a) and ( ) in (b), from a Monte Carlo study based on N = 50000 replications of a dataset generated under Design 2, where the true tensor and the Fisher information are isotropic. The histogram density estimators are compared with theoretical limit densities (black continuous curves), which are uniform on the vertical axes and on the horizontal axes. The best-fitting gamma densities (red broken curves) are also shown, with shape parameter 2.4103 and scale parameter 2.0842 in (a) and with shape parameter 2.4315 and scale parameter 2.0542 in (b).
Acknowledgments
We thank Konstantin Izyurov, Sangita Kulathinal, Antti Kupiainen, and Juha Railavo for insightful discussions, and we thank the two anonymous reviewers for their valuable questions and remarks.
References
- 1.An C, Chen X, Sloan IH, Womersley RS. Well conditioned spherical designs for integration and interpolation on the two-sphere. SIAM J Numer Anal. 2010;48:2135–2157. https://doi.org/10.1137/100795140. [Google Scholar]
- 2.Anderson GA. An asymptotic expansion for the distribution of the latent roots of the estimated covariance matrix. Ann Math Statist. 1965;36:1153–1173. [Google Scholar]
- 3.Anderson GW, Guionnet AA, Ofer Z. An Introduction to Random Matrices. Cambridge University Press; 2010. [Google Scholar]
- 4.Bannai E, Bannai E. A survey on spherical designs and algebraic combinatorics on spheres. European J Combin. 2009;30:1392–1425. [Google Scholar]
- 5.Basser PJ, Mattiello J, LeBihan D. MR diffusion tensor spectroscopy and imaging. Biophys J. 1994;66:259. doi: 10.1016/S0006-3495(94)80775-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Basser PJ, Mattiello J, LeBihan D. Estimation of the effective self-diffusion tensor from the NMR spin echo. J Magnetic Resonance Ser B. 1994;103:247–254. doi: 10.1006/jmrb.1994.1037. [DOI] [PubMed] [Google Scholar]
- 7.Basser PJ. Inferring microstructural features and the physiological state of tissues from diffusion weighted images. NMR Biomed. 1995;8:333–344. doi: 10.1002/nbm.1940080707. [DOI] [PubMed] [Google Scholar]
- 8.Basser PJ. New histological and physiological stains derived from diffusion-tensor MR-images. Imaging Brain Structure and Function, Ann New York Acad Sci. 1997;820:123–138. doi: 10.1111/j.1749-6632.1997.tb46192.x. [DOI] [PubMed] [Google Scholar]
- 9.Basser PJ, Pajevic S. Statistical artifacts in diffusion tensor MRI (DT-MRI) caused by background noise. Magnetic Resonance Med. 2000;44:41–50. doi: 10.1002/1522-2594(200007)44:1<41::aid-mrm8>3.0.co;2-o. [DOI] [PubMed] [Google Scholar]
- 10.Basser PJ, Pajevic S. A normal distribution for tensor-valued random variables: Applications to diffusion tensor MRI. IEEE Trans Med Imaging. 2003;22:785–794. doi: 10.1109/TMI.2003.815059. [DOI] [PubMed] [Google Scholar]
- 11.Basser PJ, Pajevic S. Dealing with uncertainty in diffusion tensor MR data. Israel J Chem. 2003;43:129–144. [Google Scholar]
- 12.Basser PJ, Pajevic S. Spectral decomposition of a 4th-order covariance tensor: Applications to diffusion tensor MRI. Signal Process. 2007;87:220–236. [Google Scholar]
- 13.Batchelor PG, Atkinson D, Hill DLG, Calamante F, Connelly A. Anisotropic noise propagation in diffusion tensor MRI sampling schemes. Magnetic Resonance Med. 2003;49:1143–1151. doi: 10.1002/mrm.10491. [DOI] [PubMed] [Google Scholar]
- 14.Chattopadhyay AK, Pillai KC. Asymptotic expansions for the distributions of characteristic roots when the parameter matrix has several multiple roots. Multivariate Analysis, III; Proc. Third Internat. Sympos., Wright State Univ; Dayton, OH. 1972; Academic Press; 1973. pp. 117–127. [Google Scholar]
- 15.Chiani M. Distribution of the largest eigenvalue for real Wishart and Gaussian random matrices and a simple approximation for the Tracy–Widom distribution. J Multivariate Anal. 2014;129:69–81. [Google Scholar]
- 16.Chikuse Y. Statistics on Special Manifolds. Springer Lecture Notes in Statistics. 2003;174 [Google Scholar]
- 17.Clement-Spychala ME, Couper D, Zhu H, Muller KE. Approximating the Geisser-Greenhouse sphericity estimator and its applications to diffusion tensor imaging. Stat Interface. 2010;3:81–90. doi: 10.4310/SII.2010.v3.n1.a7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Delsarte P, Goethals JM, Seidel JJ. Spherical codes and designs. Geom Dedicata. 1977;6:363–388. [Google Scholar]
- 19.Drton M. Likelihood ratio tests and singularities. Ann Statist. 2009;37:979–1012. [Google Scholar]
- 20.Drton M, Xiao H. Wald tests of singular hypothesis. Bernoulli. 2016;22:38–59. [Google Scholar]
- 21.Dyson FJ. A Brownian-motion model for the eigenvalues of a random matrix. J Math Phys. 1962;3:1191–1198. [Google Scholar]
- 22.Edelman A. PhD Thesis. MIT; 1989. Eigenvalue and Condition Numbers of Random Matrices. [Google Scholar]
- 23.Farrell RH. Multivariate Calculation Use of the Continuous Groups. Springer; 1985. [Google Scholar]
- 24.Forrester PJ. London Math Soc Monographs. Princeton University Press; 2010. Log-Gases and Random Matrices. [Google Scholar]
- 25.Guionnet A. Noncommutative Probability and Random Matrices at Saint-Flour. Springer; 2012. Large random matrices: Lectures on macroscopic asymptotics; pp. 169–463. [Google Scholar]
- 26.Hikami S, Brézin E. WKB-expansion of the Harish-Chandra-Itzykson-Zuber integral for arbitrary β. Progr Theoret Phys. 2006;116:441–502. [Google Scholar]
- 27.Idier J, Collewet G. Properties of Fisher Information for Rician Distributions and Consequences in MRI. 2014 preprint hal-01072813. [Google Scholar]
- 28.James AT. Distributions of matrix variates and latent roots derived from normal samples. Ann Math Statist. 1964;35:475–501. [Google Scholar]
- 29.Liu J, Gasbarra D, Railavo J. Fast estimation of diffusion tensors under Rician noise by the EM algorithm. J Neurosci Methods. 2016;257:147–158. doi: 10.1016/j.jneumeth.2015.09.029. [DOI] [PubMed] [Google Scholar]
- 30.Magnus JR, Neudecker H. Matrix Differential Calculus with Applications in Statistics and Econometrics. Wiley; 1999. [Google Scholar]
- 31.Mallows CL. Latent vectors of random symmetric matrices. Biometrika. 1961;48:133–149. [Google Scholar]
- 32.Mehta ML. Random Matrices. 3. Elsevier; 2004. [Google Scholar]
- 33.Muirhead RJ. Aspects of Multivariate Statistics. Wiley; 1984. [Google Scholar]
- 34.Pajevic S, Basser PJ. Parametric and non-parametric statistical analysis of DT-MRI data. J Magnetic Resonance. 2003;161:1–14. doi: 10.1016/s1090-7807(02)00178-7. [DOI] [PubMed] [Google Scholar]
- 35.Pajevic S, Basser PJ. A joint PDF for the eigenvalues and eigenvectors of a diffusion tensor. Proc Internat Soc Magnetic Resonance Med. 2010;18:303. [Google Scholar]
- 36.Pierpaoli C, Infante I, Mattiello J, Di Chiro G, Le Bihan D, Basser PJ. Diffusion tensor imaging of brain white matter anisotropy. Proc Internat Soc Magnetic Resonance Med. 1994;2:1038. [Google Scholar]
- 37.Schwartzman A, Mascarenhas WF, Taylor JE. Inference for eigenvalues and eigenvectors of Gaussian symmetric matrices. Ann Statist. 2008;36:2886–2919. [Google Scholar]
- 38.Schwartzman A, Dougherty RF, Taylor JE. Group comparison of eigenvalues and eigenvectors of diffusion tensors. J Amer Statist Assoc. 2010;105:588–598. doi: 10.1198/jasa.2010.ap07291. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Takemura A. Zonal Polynomials. Institute of Mathematical Statistics; 1984. IMS Lecture Notes Monogr. Ser. 4. [Google Scholar]
- 40.Tao T. Topics in Random Matrix Theory. American Mathematical Society; 2012. [Google Scholar]
- 41.Tao T. The Harish-Chandra-Itzykson-Zuber Integral Formula. 2013 https://terrytao.wordpress.com/2013/02/08/the-harish-chandra-itzykson-zuber-integral-formula/
- 42.Van Der Vaart AW. Asymptotic Statistics. Cambridge University Press; 2000. [Google Scholar]
- 43.Watanabe S. Cambridge Monogr Appl Comput Math. Vol. 25. Cambridge University Press; 2009. Algebraic Geometry and Statistical Learning Theory. [Google Scholar]
- 44.Zhu H, Zhang H, Ibrahim JG, Peterson BS. Statistical analysis of diffusion tensors in diffusion-weighted magnetic resonance imaging data. J Amer Statist Assoc. 2007;102:1085–1102. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.












