Abstract
Testing correlation structures has attracted extensive attention in the literature due to both its importance in real applications and several major theoretical challenges. The aim of this paper is to develop a general framework of testing correlation structures for the one-, two-, and multiple sample testing problems under a high-dimensional setting when both the sample size and data dimension go to infinity. Our test statistics are designed to deal with both the dense and sparse alternatives. We systematically investigate the asymptotic null distribution, power function, and unbiasedness of each test statistic. Theoretically, we make great efforts to deal with the non-independency of all random matrices of the sample correlation matrices. We use simulation studies and real data analysis to illustrate the versatility and practicability of our test statistics.
Keywords: MSC 2010 subject classifications, Primary 62H15, secondary 62H10, Dense alternatives, Global testing, Sample correlation matrices, Sparse alternatives
1. Introduction.
Consider random samples obtained from K independent populations. Let z(ℓ) be a p−dimensional random vector for ℓ = 1, …, K. We denote to be the nℓ independent samples of z(ℓ) for the ℓ−th population and as its sample mean. Then, the sample covariance matrix and sample correlation matrix of are, respectively, given by
where diag(Sn) is a diagonal matrix constructed from the diagonal elements of Sℓ. There has been growing interest in the development of methods and theory for hypothesis testing on correlation structures in different settings (Kullback, 1967; Aitkin, 1969; Jennrich, 1970; Schott, 1996; Browne, 1978; Cole, 1968; Schott, 2005; Gao et al., 2017; Zhou et al., 2015; Debashis and Alexander, 2014). See, for example Anderson (2003) and Cai (2017) for overviews of statistical challenges associated with such developments.
1.1. Existing Literature.
Under the classical setting with fixed p as minℓ{nℓ} → ∞, there are three major testing problems corresponding to K = 1, K = 2, and K > 2, respectively. As K = 1, it is one sample testing problem that focuses on testing H01 : R1 = R∗ against HA1 : R1 ≠ R∗, where R1 is the population correlation matrix and R∗ is a specific correlation matrix. An interesting asymptotic result is that the test statistic
is asymptotically distributed as a linear form in 0.5p(p−1) independent random variables, and not in general unless R∗ = Ip (Kullback, 1967; Aitkin, 1969; Bartlett and Rajalakshman, 1953), where Ip is the p×p identity matrix. This fact shows that testing the correlation matrix is a more difficult task than testing the covariance matrix. As K = 2, it is two sample testing problem that tests H02 : R1 = R2 against HA2 : R1 ≠ R2, where R1 and R2 are two population correlation matrices. Several test statistics as distances between and and their asymptotic distributions have been studied in the literature (Aitkin, 1969; Jennrich, 1970; Larntz and Perlman, 1985). As K > 2, it is multiple sample testing problem that tests H0K : R1 = … = RK against HAK : not H0K. Many test statistics and their asymptotic distributions have been extended from the case K = 2 to K > 2 (Kullback, 1967; Schott, 1996; Browne, 1978; Cole, 1968; Gupta et al., 2013).
Recently, ultra-high dimensional data arise from a variety of applications, including neuroimaging and genetics; that is, both p and minℓ{nℓ} converge to infinity. Testing correlation structures in this high-dimensional setting has attracted extensive attention in the past decade due to both its importance in real applications and two major theoretical challenges, including high dimensionality and dependency (Cai, 2017; Debashis and Alexander, 2014). In this case, the test statistics developed for the classical setting either do not perform well or are no longer applicable. Therefore, under the high-dimensional setting, a collection of new testing statistics have been developed in the last few years for both the one- and two-population testing problems (Cai, 2017; Zhou et al., 2015; Cai and Zhang, 2016; Bodwin et al., 2016; Schott, 2005; Gao et al., 2017). For the one-sample case, the existing results focus on the test of short-range dependence, which includes independency as a special case, since the standard random matrix theory results are not directly applicable for a composite null. Moreover, the existing testing statistics are particularly powerful under either a “sparse” alternative or a dense alternative. For instance, Zhou et al. (2015) proposed several extreme value statistics to test the equality of two large U-statistic based correlation matrices, which include the rank-based correlation matrices as special cases.
1.2. Our Contributions.
The aim of this paper is to provide a general framework of testing correlation structures for the one-, two-, and multiple sample testing problems as p → ∞. Compared with the existing literature discussed above, we make four major contributions as follows.
-
(I)
For the first time, we develop a set of test statistics to test correlation structures for the one-, two-, and multiple sample testing problems under the high-dimensional setting. Our test statistics are designed to deal with both the dense and sparse alternatives. Specifically, they are the sum or the maximum of two terms, including a term for the dense alternative and the other for the sparse alternative.
-
(II)
We propose the test statistics for testing H01: R1 = R∗ as K = 1 and then derive its asymptotic distribution and power function, even when R∗ is an arbitrary correlation matrix. We make great efforts to deal with the non-independent elements of population random vectors during the derivation. In contrast, the existing results based on the standard random matrix theory (Bai and Silverstein, 2004) are limited to the covariance matrix or independent correlation (Gao et al., 2017; Li and Xue, 2015; Shao and Zhou, 2014; Qiu and Chen, 2012).
-
(III)
Similar to testing H01 : R1 = R∗, we derive the asymptotic distribution of the test statistics and stride to deal with the non-independency of the two random matrices of the sample correlation matrices for testing H02 : R1 = R2.
-
(IV)
To the best of our knowledge, we propose the first test statistic for testing H0K : R1 = ··· = RK under the high-dimensional setting and then establish its asymptotic distribution under both H0K and HAK without assuming the normality. We also stride to deal with the non-independency of all random matrices of the sample correlation matrices.
The rest of this paper is organized as follows. Section 2 focuses on the one-sample problem, whereas Section 3 focuses on two- and multiple- sample testing problems. In each section, we propose the test statistics and establish its asymptotic distribution, power function, and unbiasedness. Section 5 will present simulation studies. We apply the test statistics to the ADHD data sets in Section 6. All proofs are collected in the Appendices.
2. Test Statistics for One Sample Testing Problem.
In this section, we focus on the one-sample problem of testing H01 : R1 = R∗ against HA1 : R1 ≠ R∗. This section consists of three parts. In Section 2.1, we describe two proposed test statistics. We characterize its asymptotic null distributions in Section 2.2 and its power properties in Section 2.3.
2.1. Test statistics.
We first introduce two terms as follows:
where with , , and δ{∙} is an indicator function. The first term Ln,1 is designed for the dense alternative, whereas Tn,1 is for the sparse alternative.
Based on Ln,1 and Tn,1, we propose a weighted test statistic Mn,1 as follows:
| (2.1) |
where the second term of Mn,1 is a hard thresholding, C0 is a large positive number and s∗(n1, p) is a scalar threshold depending on (n1, p). The choices of C0 and s∗(n1, p) will be given in the following Remark 2.1. For a given significance level α, we construct the acceptance region of Mn,1 to be
| (2.2) |
where q1−α is the (1 − α)100% quantile of N (0, 1) and µz0 will be specified below.
We also propose a maximum test statistic as follow:
| (2.3) |
where is a positive constant and different represents different contributions from Ln, 1 and Tn,1 to . For a given significance level α, we construct the acceptance region of to be
where cα is a critical value and the choices of cα and will be given in Remark 2.1.
2.2. Null distribution.
Our first theoretical result is to characterize the limiting null distribution of Ln,1. We introduce two assumptions that will be used later. Assumption (a) specifies the moment assumption of . Assumption (b) specifies the ratio of the dimension of to the sample size nℓ. We then introduce Assumption (a) as follows.
Assumption (a). has the independent component structure , where with independently and identically distributed (iid) elements , , and the kurtosis of is equal to . That is, are standardized iid random variables only requiring that the fourth moment exists. The spectral norm of Rℓ is bounded.
Assumption (a) imposes the independent component structure on , which has been commonly used in random matrix theory (Bai and Silverstein, 2004; Chen et al., 2010). It only requires the existence of moments until the fourth order. The identically distributed assumption is not critical for most theoretical developments below.
We state Assumption (b) as follows.
Assumption (b). The ratio of the dimension p to the sample size nℓ tends to a constant, that is, p/nℓ → yℓ ϵ (0,∞).
Assumption (b) gives the convergence regime of the data dimension and the sample sizes. It assumes that the data dimension increases proportionally with the sample size, even when the limit yℓ can be extremely small (or large). Therefore, the data dimension may be much smaller (or greater) than the sample size.
Our first theoretical result quantifies the limiting distribution of the statistic . Let denote the convergence in distribution.
Theorem 2.1. If Assumptions (a)-(b) hold for ℓ = 1 and under H01, we conclude that
(I.1).
(I.2). holds under some additional conditions, including s*(n1,p) – 4 log p→ + ∞ and (C1), (C2), and (C3) in Cai, Liu and Xia (2013) for R* and , where µz0 is defined as
with r*khj being the (h, j) entry of for k=1/2, 1, 3/2,and 2.
Remark 2.1. When R∗ = Ip, Cai and Jiang (2011) proved that converged to a type I extreme value distribution function F(t) = exp[−(8π)−1/2 exp(−t/2)] under H01. When R* ≠ Ip, similar to (22) of Cai and Zhang (2016), we conclude that converges to the type I extreme value distribution function under H01 and (C1), (C2) and (C3) in Cai, Liu and Xia (2013) for R* and . The choices of s*(n1,p), C0, and cr are given as follows:
- Choice of the threshold s∗(n1,p): The test statistic Mn,1 mainly targets at Ln,1. For simplicity, the threshold is taken to be
where u0 satisfies exp[−(8π)−1/2 exp(−u0/2)] = 0.99. The threshold ensures that even if n1 and p are small, the probability of the event {Tn,1 > s∗(n1,p)} is bounded by 0.01 under H01. The probability of the event {Tn,1 > s∗(n1, p)} becomes negligible under H01 when either n1 or p is relatively large.
Choice of the constant C0: The role of C0 is to ensure that the second term of Mn,1 acts as the main term in Mn,1 when Tn,1 > s∗(n1, p). It is enough to require that is far away from q1−α. For simplicity, let C0 be p2 throughout this paper.
Choice of the constant and the critical value cα: Theorem 2.1 shows that is asymptotically distributed as N (0, 1) under H01. To balance the contribution of Ln,1 and that of Tn,1, should be relatively small for extremely dense R1 − R∗, whereas should be large for extremely sparse R1 − R∗. However, it is unknown whether R1 − R∗ is dense or sparse, so we choose such that and have the same (1 − α/2)100% quantile, where α is the significance level. That is, we have and , where satisfies . Then, we have under H01
Proof. We will give the skeletons of the proof of Theorem 2.1. The details of the proof are placed in Appendices. The proof proceeds in three steps.
Skeleton of Step 1. To obtain the expansion of as follows
we assume without loss of generality. This step is mainly to use the Taylor expansions of and with .
Skeleton of Step 2. We want to derive the limits of the following four terms , , and in probability.
Skeleton of Step 3. We want to derive the limiting null distribution of , , , . Thus by the delta method, we obtain the central limit theorem (CLT) of . Because these terms involve diag(S1) − Ip, we cannot directly use the random matrix theory on linear spectral statistics of S1 to obtain the limiting distribution of these terms. To solve the problem, we construct four martingale difference sequences to establish the CLT of these terms. Especially, the derivation of the CLT for the case R∗ ≠ Ip is much more difficult than the derivation for the case R∗ = Ip.
Corollary 2.1. Under the assumptions of Theorem 2.1, we have the following results:
- If R∗ is the identity matrix Ip, then µz0 reduces to
- If the population is Gaussian, then µz0 reduces to
where r∗1jj' is the (j, j') entry of R∗.
Theorem 2.1 provides a unified framework of testing H01 : R1 = R∗ for an arbitrary R∗. Our test statistics account for both dense and sparse alternatives, and they work for satisfying the independent component structure specified in Assumption (a) and a ratio of p/n1 = y1 in Assumption (b). Technically, to prove Theorem 2.1, we develop a set of novel tools to deal with the dependence between S1 and diag(S1), which is technically nontrivial and is of independent interest for handling the sample correlation in more general settings. In contrast, Gao et al. (2017) only established the CLT of the sample correlation matrices of a high dimensional vector whose elements have an identity correlated structure R∗ = Ip. Moreover, their theoretical result involves some two-dimensional contour integrals, which can be difficult to compute.
2.3. Power properties and optimality.
We examine the power properties of Mn,1 and . We first establish the asymptotic distribution of the statistic under the alternative hypothesis HA1.
Theorem 2.2. Assuming that Assumptions (a) and (b) hold for ℓ = 1, we have
where µzA and σzA depend on the alternative population correlation matrix R1 and will be given in Appendix.
Given the result in Theorem 2.2, we can characterize the properties of the power functions, which is given by
In the following, we will study the properties of the power functions g1(R1, α) and .
Corollary 2.2. Assuming that Assumptions (a) and (b) hold for ℓ = 1, we have the following results:
If tr[(R1 − R∗)2] > c0 > 0, then g1(R1, α) > α when the sample size n1 is large enough and c0 is any given small constant;
If tr[(R1 − R∗)2] tends to infinity, then g1(R1, α) and g'1(R1, α) are close to one as n1 → ∞;
If the absolute value of at least one entry of R1−R∗ is greater than and the conditions (C1), (C2), and (C3) in Cai et al. (2013) hold for R1 and then g1(R1, α) and are close to one as n1 → ∞.
Corollary 2.2 shows that the proposed test Mn,1 is asymptotically unbiased. In Appendix, we will prove that (i) For the dense alternative tr[(R1 – R*)2] → ∞, the power functions tend to one; (ii). For the sparse alternative, if the absolute value of at least one entry of R1 − R∗ is greater than , then the power functions will be close to one.
Similar to Cai and Ma (2013), we define
where b1, b10 are positive constants, and with ei and ej being the ith column and jth column of the p × p identity matrix, respectively.
Theorem 2.3. Let 0 < α < β < 1. Suppose that as p/n1 → y1 > 0 as n1 → + ∞. Then there exist two constants 0 < b1, b10 < 1 such that for any test ϕ with the significance level α for testing H01 : R1 = R∗, we have
where is the expectation under the population correlation matrix being R1.
Theorem 2.3 shows that no level α test can distinguish the null hypothesis from the alternative hypothesis with the power tending to one as p/n1 → y1 > 0, or . Then, Theorem 2.3 gives the lower bound for the optimality of our proposed procedure.
3. Extensions to Two and Multiple Sample Testing Problems.
This section consists of two parts. In Section 3.1, we focus on the two-sample problem of testing H02 : R1 = R2 against HA2 : R1 ≠ R2. In Section 3.2, we consider the multiple sample testing problem.
3.1. Extension to Two Sample Testing Problem.
3.1.1. Test statistics and their null distributions for two-sample testing problem.
Let and for ℓ = 1, 2. We introduce two terms as follows:
where is defined as
The first term Ln,2 is introduced to deal with the dense alternative, whereas the second term Tn,2 is for the sparse alternative.
We propose a weighted test statistic Mn,2 as follows:
| (3.1) |
where C0,2 and the threshold s(n1, n2, p) will be given in Remark 3.1. For a given significance level α, we construct an acceptance region of Mn,2 to be
where and will be defined below.
We also propose a maximum test statistic as follows:
| (3.2) |
where is a positive constant. For a given significance level α, we construct an acceptance region of to be
where the positive constant and the critical value cα,2 will be given in Remark 3.1.
We establish the asymptotic null distribution of Ln,2 as follows.
Theorem 3.1. Let R be the common correlation matrix R = R1 = R2. Assuming that Assumptions (a) and (b) hold for ℓ = 1 and 2 and under H02, we conclude that
(II.1). ;
(II.2). holds under some additional conditions, including s(n1, n2, p) − 4 log p → + ∞ and (C1), (C2), and (C3) of Cai, Liu and Xia (2013) for R and , where µz12 is given by
with r0khj being the (h, j) entry of Rk for k = 1/2, 1, 2/3, and 2 and
Remark 3.1. Similar to (22) of Cai and Zhang (2016), we conclude that
converges to the type I extreme value distribution function under H02 and (C1), (C2), and (C3) in Cai, Liu and Xia (2013) for R and . The choices of s(n1, n2, p), C0,2, and cα,2 are given as follows:
- Choice of the threshold s(n1, n2, p): The test statistic Mn,2 mainly targets at Ln,2. For simplicity, we set s(n1, n2, p) as
Where satisfies . The threshold ensures that even for small n1, n2 and p, the probability of the event {Tn,2 > s(n1, n2, p)} is bounded by 0.01 under H02. The probability of the event {Tn,2 > s(n1, n2, p)} becomes negligible under H02 when either n1, n2 or p is moderately large.
Choice of C0,2, and cα,2: The constants C0,2, and cα,2 are the same as C0, and cα in Remark 2.1. Moreover, under H02.
Proof. We will give the skeletons of the proof of Theorem 3.1. The details of the proof are placed in Appendices. The proof proceeds in three steps.
Skeleton of Step 1. It is assumed that holds for ℓ = 1, 2 without loss of generality. We obtain the expansion of as follows
This step is mainly to use the Taylor expansions of and with for ℓ=1,2.
Skeleton of Step 2. We want to derive the limits of the following ten terms in probability: , , , , , , , , , .
Skeleton of Step 3. We want to derive the limiting null distribution of , , , , , , . Thus by the delta method, we obtain the CLT of . Because these terms involve the product of any two or three terms among S1, diag(S1) − Ip, S2 and diag(S2) − Ip, the CLT for Theorem 2.1 is not directly applicable. Thus, in order to derive the CLT of these terms, eight new martingale difference sequences are constructed. Especially, the derivation of the CLT for the two population case R1 = R2 is very different from and more difficult than the derivation for the one population case R1 = R∗.
Remark 3.2. Under the null hypothesis H02, we do not know the true R, so we have to estimate the terms related to R in the asymptotic mean and variance. Let
where with i = 1,…,nℓ. Then, we estimate a0 = p–1tr(R2), b0, c0, and dℓ as follows:
with letting if βℓ = 0, and . Finally, we can obtain an estimate of µz12 as follows:
Corollary 3.1. Under the same assumptions of Theorem 3.1, we concluded that
holds under some additional conditions, including s(n1, n2, p) − 4 log p → + ∞ and (C1), (C2), and (C3) in Cai, Liu and Xia (2013) for R and .
3.1.2. Power properties and optimality.
In the following, we will study the power properties of the statistics Mn,2 and .
Theorem 3.2. Under Assumptions (a) and (b), we have
where µA12 and σA12 depend on the alternative population correlation matrices R1 and R2 and will be given in Appendix.
Theorem 3.2 gives the asymptotic distribution of the statistic under the alternative hypothesis. The power function is given by
and . Then, g2(R1, R2, α) and have the following properties.
Corollary 3.2. Under the same assumptions of Theorem 3.2, we have the following results:
If tr[(R2 − R1)2] > c0 > 0, where c0 is a positive scalar, then g2(R1, R2, α) > α when the sample size is large enough;
If tr[(R1 − R2)2] → ∞, then g2(R1, R2, α) and are close to one as n1, n2 → ∞;
If the absolute value of at least one entry of R1 − R2 is greater than and the conditions (C1), (C2), and (C3) in Cai et al. (2013) hold for Rℓ and , ℓ = 1, 2, then g2(R1, R2, α) and are close to one as n1, n2 → ∞.
Corollary 3.2 shows that the proposed test Mn,2 is asymptotically unbiased. In Appendix, we will prove that for the dense alternative tr[(R1 − R2)2] → ∞, the power functions tend to one. For the sparse alternative, if the absolute value of at least one entry of R1 − R2 is greater than , then the power functions are close to one.
Similar to Cai and Ma (2013), we define
where b2, b20 are positive constants, and with ei and ej being the ith column and jth column of the p × p identity matrix, respectively.
Theorem 3.3. Let 0 < α < β < 1. Suppose that p/ni → yi > 0 as ni → ∞ for i = 1, 2. Then there exist two constants 0 < b2, b20 < 1 such that for any test ϕ with the significance level α for testing H02 : R1 = R2, we have
where is the expectation under the two population correlation matrix being R1 and R2.
Theorem 3.3 shows that no level α test can distinguish between the null hypothesis and all alternative hypotheses with the power tending to one as p/ni → yi > 0 for i = 1, 2 and or . Then, Theorem 3.3 gives the lower bound for the optimality of our proposed procedure.
3.2. Test statistic for multiple sample testing problem.
We extend the test statistic from two samples to K samples. The one weighted test statistic is constructed as
where
with being a vector of weights.
For simplicity, we focus on the asymptotic distribution of Mn,K.
We first present a key lemma as follows.
Lemma 3.1. Assume that Assumptions (a) and (b) hold for ℓ = 1,…, K. Then, are asymptotically distributed as a multivariate normal distribution. Moreover, we have
where with , and
Moreover, , and γAuu' have closed forms and will be given in Appendix.
Based on Lemma 3.1, we can establish the asymptotic null distribution of Mn,K as follows.
Theorem 3.4. Under the same assumptions of Lemma 3.1 and H0K, we conclude that
(III.1). ;
(III.2). under some additional conditions, including and (C1), (C2), and (C3) of Cai, Liu and Xia (2013) for R and , where νK = 4[tr(R2)]2uK is given by
and can be similarly defined as µz12.
Remark 3.3. There are two important issues associated with Mn,K. The first one is to determine the weights . Since the asymptotic variance of is equal to , a reasonable choice of , is for 1 ≤ ℓ1 < ℓ2 ≤ K. The second one is to estimate the asymptotic mean and variance under H0K, since we do not know what the true R is. A good estimate of p−1tr(R2) is , where was defined in Remark 3.2. Then the estimate of νK
Furthermore, the estimate can be obtained by replacing 1 and 2 by ℓ1 and ℓ2 in in Remark 3.2. The C0 is the same as C0,2 in Remark 3.1. The threshold is obtained by replacing 1 and 2 by ℓ1 and ℓ2 in s(n1, n2, p) in Remark 3.2.
4. Estimation of the kurtosis β1.
To estimate β1 in Theorems 2.1 and 3.1, we consider two cases as follows.
Case 1: When R1 is unknown, the covariance matrix Σ1 is unknown. We may use an estimator of β1 as follows:
where and
Case 2: For R1 = R∗ for a pre-specified correlation matrix R∗, we may estimate β1 as follows:
where , and
The following lemma gives the consistency of the estimator under the null hypothesis H01 : R1 = R∗.
Lemma 4.1. Suppose that holds for all ℓ = 1,…, p, where c is a positive constant. Under the null hypothesis H01 and assumptions (a)–(b), we have
Proof. Without loss of generality, assume . Let eℓ be the ℓ-th column of the p × p identity matrix. We can show the following results:
| (4.1) |
| (4.2) |
| (4.3) |
where op(1) is uniform for ℓ = 1,…, p. For instance, to prove (4.1), we have
where o(1) is uniform for all ℓ = 1,…, p. It follows that
Therefore, we have
which yields . This completes the proof of Lemma 4.1.
5. Simulation studies.
In this section, we carried out simulation studies to evaluate the finite-sample performance of the proposed test statistics in terms of the empirical test size and power. We consider both one sample testing problem and two sample testing problem. For the one sample testing problem, we set the dimension p to be p = 50, 100, 200, 500, and 1000 and the sample size n1 to be n1 = 100, 120, 200, and 300. The data were generated according to for i = 1,…, n1, where the elements of were independently and identically generated from Gaussian population N (0, 1) or Gamma(4, 2) − 2. For the two sample testing problem, we set p = 50, 100, 200, 500, and 1000 and (n1, n2) = (100, 100), (150, 150), (200, 200). The data were generated according to for i = 1,…, nℓ and ℓ = 1, 2, where the elements of were independently and identically generated from Gaussian population N (0, 1) or Gamma(4, 2) − 2. For the three sample testing problem, we set p = 50, 100, 200, 500, and 1000 and (n1, n2, n3) = (100, 100, 100), (100, 100, 100), (100, 100, 200) and (100, 200, 200). The data were generated according to for i = 1,…, nℓ and ℓ = 1, 2, 3, where the elements of were independently and identically generated from Gaussian population N (0, 1) or Gamma(4, 2) − 2. We set the nominal size to be 5%, run 2,000 replications for empirical sizes and 1000 replications for empirical powers for each setting.
We consider nine different sets of population correlation matrices for Rℓ. For the two sample testing problem, we compare our tests denoted as “FDS” for Mn,2 and “MAX” for with the extreme statistic test, denoted as “CZ” in Cai and Zhang (2016). However, for the one sample testing problem, we cannot find any competing method when R∗ is not an identity matrix, so we do not include any alternative method. When R∗ is an identity matrix, we compare our test “FDS” for Mn,1 and “MAX” for with “GHPY” in Gao et al. (2017) and “LX” in Li and Xue (2015). For the three-sample testing problem, since we cannot find any competing method, we do not include any alternative method. For the sake of space, we selectively present some key results in Tables 1–3 and include additional results in the supplementary document. The first three models are designed for the one sample testing problem, whereas the middle four ones are for the two sample testing problem and the last three ones are for the three-sample testing problem. The ten different models of population correlation matrices are summarized as follows.
Table 1.
Empirical sizes in Model 1.1 and empirical powers in Models 1.2–1.3 for H01 (in percentage)
| wij ~ N (0; 1) | wij ~ Gamma(4,2)–2 | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| ϵ | n | Methods | p=50 | 100 | 200 | 500 | 1000 | 50 | 100 | 200 | 500 | 1000 |
| Empirical sizes in Model 1.1 | ||||||||||||
| 0.0 | 100 | FDS | 6.20 | 4.75 | 5.90 | 5.75 | 6.25 | 5.45 | 6.25 | 6.60 | 6.55 | 5.90 |
| MAX | 4.25 | 3.25 | 3.50 | 3.10 | 3.30 | 4.60 | 5.00 | 4.30 | 4.60 | 4.75 | ||
| GHPY | 4.85 | 4.15 | 4.70 | 4.10 | 5.10 | 4.85 | 4.50 | 4.70 | 5.75 | 4.90 | ||
| LX | 3.15 | 1.90 | 1.70 | 0.90 | 0.90 | 4.35 | 4.55 | 3.55 | 3.10 | 3.10 | ||
| 200 | FDS | 6.15 | 5.15 | 5.90 | 6.05 | 5.95 | 5.70 | 5.55 | 6.35 | 5.55 | 5.85 | |
| MAX | 4.40 | 3.40 | 5.05 | 4.10 | 3.80 | 4.85 | 4.45 | 5.10 | 4.75 | 4.65 | ||
| GHPY | 5.80 | 4.80 | 4.90 | 4.85 | 4.95 | 5.20 | 4.35 | 5.35 | 4.80 | 5.10 | ||
| LX | 5.35 | 3.40 | 3.55 | 2.15 | 2.00 | 5.20 | 4.90 | 5.40 | 4.90 | 5.00 | ||
| 300 | FDS | 5.15 | 4.60 | 5.15 | 5.65 | 5.05 | 6.25 | 6.05 | 6.50 | 5.85 | 6.40 | |
| MAX | 3.90 | 3.80 | 3.80 | 4.45 | 3.50 | 4.55 | 4.40 | 5.10 | 4.15 | 4.75 | ||
| GHPY | 4.45 | 5.20 | 4.65 | 5.10 | 5.25 | 4.95 | 5.50 | 5.65 | 4.80 | 5.85 | ||
| LX | 5.25 | 4.25 | 3.90 | 2.80 | 2.75 | 5.15 | 5.20 | 6.05 | 5.15 | 5.65 | ||
| 0.5 | 100 | FDS | 5.50 | 5.25 | 5.90 | 5.95 | 6.35 | 5.75 | 6.00 | 5.70 | 5.80 | 5.95 |
| MAX | 5.20 | 4.80 | 5.50 | 5.65 | 5.15 | 5.10 | 4.45 | 4.55 | 4.75 | 4.10 | ||
| 200 | FDS | 5.25 | 6.00 | 5.30 | 5.05 | 6.25 | 5.60 | 5.40 | 5.75 | 5.45 | 5.10 | |
| MAX | 4.60 | 4.65 | 4.45 | 4.45 | 4.90 | 4.65 | 4.55 | 4.20 | 3.85 | 4.35 | ||
| 300 | FDS | 5.85 | 6.00 | 5.00 | 5.75 | 5.05 | 4.50 | 5.75 | 6.35 | 5.80 | 5.80 | |
| MAX | 5.05 | 4.25 | 4.60 | 4.20 | 4.00 | 3.70 | 4.65 | 5.10 | 4.30 | 4.30 | ||
| Empirical powers in Model 1.2 | ||||||||||||
| 1.0 | 100 | FDS | 28.9 | 59.4 | 90.8 | 99.9 | 100.0 | 28.1 | 58.1 | 92.1 | 100.0 | 100.0 |
| MAX | 21.0 | 49.8 | 86.8 | 99.8 | 100.0 | 20.4 | 49.6 | 88.3 | 99.9 | 100.0 | ||
| GHPY | 19.6 | 49.5 | 86.9 | 100.0 | 100.0 | 20.4 | 46.9 | 87.4 | 99.3 | 100.0 | ||
| LX | 7.10 | 7.6 | 8.9 | 15.7 | 23.3 | 10.7 | 13.4 | 16.1 | 24.4 | 34.3 | ||
| 200 | FDS | 58.9 | 93.3 | 100.0 | 100.0 | 100.0 | 55.8 | 92.9 | 99.9 | 100.0 | 100.0 | |
| MAX | 48.9 | 90.8 | 100.0 | 100.0 | 100.0 | 48.8 | 90.2 | 99.8 | 100.0 | 100.0 | ||
| GHPY | 51.6 | 90.4 | 99.9 | 100.0 | 100.0 | 50.9 | 89.8 | 100.0 | 100.0 | 100.0 | ||
| LX | 17.8 | 28.3 | 49.7 | 86.90 | 97.55 | 22.9 | 34.0 | 54.0 | 85.9 | 96.8 | ||
| 300 | FDS | 83.4 | 99.9 | 100.0 | 100.0 | 100.0 | 81.9 | 99.7 | 100.0 | 100.0 | 100.0 | |
| MAX | 77.4 | 99.5 | 100.0 | 100.0 | 100.0 | 75.6 | 99.5 | 100.0 | 100.0 | 100.0 | ||
| GHPY | 78.8 | 99.6 | 100.0 | 100.0 | 100.0 | 77.3 | 99.0 | 100.0 | 100.0 | 100.0 | ||
| LX | 33.6 | 59.0 | 90.6 | 99.8 | 100.0 | 37.2 | 62.8 | 89.6 | 99.4 | 100.0 | ||
| 0.0 | 100 | FDS | 75.0 | 79.0 | 85.2 | 90.0 | 94.6 | 74.5 | 81.6 | 85.5 | 92.1 | 95.3 |
| MAX | 81.6 | 84.4 | 90.1 | 93.3 | 96.8 | 82.2 | 87.0 | 90.1 | 94.9 | 97.2 | ||
| GHPY | 7.6 | 6.2 | 5.5 | 4.6 | 5.5 | 9.0 | 5.9 | 5.1 | 5.7 | 5.1 | ||
| LX | 85.4 | 88.1 | 92.2 | 94.7 | 97.4 | 87.3 | 90.0 | 92.8 | 95.8 | 97.9 | ||
| 200 | FDS | 71.4 | 78.7 | 79.9 | 84.6 | 88.2 | 72.0 | 76.4 | 80.3 | 84.5 | 88.4 | |
| MAX | 77.9 | 84.8 | 84.3 | 88.4 | 91.7 | 78.9 | 78.9 | 86. | 89.3 | 91.6 | ||
| GHPY | 10.4 | 6.2 | 5.5 | 5.3 | 4.9 | 10.1 | 5.6 | 6.4 | 5.1 | 5.5 | ||
| LX | 83.3 | 88.3 | 88.1 | 90.6 | 93.7 | 83.7 | 86.3 | 88.7 | 91.8 | 93.0 | ||
| 300 | FDS | 71.2 | 76.6 | 80.8 | 84.3 | 87.5 | 71.9 | 76.8 | 80.4 | 85.7 | 86.4 | |
| MAX | 77.2 | 81.6 | 86.0 | 88.8 | 90.5 | 77.5 | 82.7 | 85.4 | 88.9 | 90.4 | ||
| GHPY | 8.4 | 7.0 | 5.2 | 5.3 | 5.1 | 9.2 | 6.7 | 6.8 | 4.7 | 5.1 | ||
| LX | 83.2 | 85.0 | 88.2 | 90.7 | 92.7 | 82.6 | 86.6 | 88.4 | 91.6 | 92.5 | ||
| Empirical powers in Model 1.3 | ||||||||||||
| 0.09 | 100 | FDS | 44.8 | 44.1 | 47.0 | 46.8 | 47.2 | 38.4 | 39.7 | 38.5 | 38.0 | 38.6 |
| MAX | 53.8 | 53.4 | 55.0 | 54.2 | 54.5 | 46.3 | 47.3 | 46.1 | 44.3 | 44.4 | ||
| 200 | FDS | 99.2 | 99.8 | 99.8 | 99.9 | 99.9 | 97.0 | 97.3 | 97.4 | 97.5 | 98.1 | |
| MAX | 99.4 | 99.6 | 99.9 | 99.9 | 100.0 | 98.3 | 98.5 | 98.3 | 98.2 | 98.7 | ||
| 300 | FDS | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 99.9 | 99.9 | 99.9 | 100.0 | 100.0 | |
| MAX | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 99.9 | 99.9 | 99.9 | 100.0 | 100.0 | ||
| 0.12 | 100 | FDS | 93.6 | 94.6 | 96.0 | 96.0 | 96.3 | 84.9 | 84.5 | 82.6 | 78.8 | 78.5 |
| MAX | 96.2 | 96.7 | 97.5 | 97.5 | 97.3 | 89.4 | 88.5 | 84.4 | 82.5 | 82.3 | ||
| 200 | FDS | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 99.8 | 99.8 | 99.7 | 99.3 | 99.3 | |
| MAX | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 99.9 | 99.8 | 99.9 | 99.6 | 99.6 | ||
| 300 | FDS | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | |
| MAX | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | 100.0 | ||
Table 3.
Empirical test sizes in Model 3.1 and empirical powers in Models 3.2 and 3.3 for H03 (in percentage)
| wij ~ N (0, 1) | wij ~ Gamma(4,2)–2 | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| (n1,n2,n3) | p=50 | 100 | 200 | 500 | 1000 | 50 | 100 | 200 | 500 | 1000 | |
| Empirical test sizes in Model 3.1 | |||||||||||
| (50,100, 100) | FDS | 6.10 | 6.60 | 5.95 | 6.20 | 6.60 | 5.45 | 5.50 | 6.00 | 5.85 | 5.30 |
| (100,100, 100) | FDS | 6.30 | 5.90 | 5.20 | 5.85 | 5.35 | 5.90 | 4.85 | 5.15 | 5.90 | 5.60 |
| (100,100, 200) | FDS | 4.70 | 5.60 | 5.65 | 5.75 | 5.25 | 5.00 | 5.35 | 4.80 | 5.35 | 5.15 |
| (100,200, 200) | FDS | 5.85 | 5.45 | 5.60 | 5.70 | 4.55 | 5.75 | 4.85 | 5.40 | 4.90 | 5.25 |
| Empirical powers in Model 3.2 | |||||||||||
| (50,100, 100) | FDS | 18.5 | 44.9 | 86.4 | 100.0 | 100.0 | 14.4 | 36.3 | 81.6 | 100.0 | 100.0 |
| (100,100, 100) | FDS | 22.1 | 52.5 | 91.8 | 100.0 | 100.0 | 18.6 | 50.1 | 91.0 | 100.0 | 100.0 |
| (100,100, 200) | FDS | 28.8 | 74.1 | 99.6 | 100.0 | 100.0 | 26.8 | 71.2 | 98.9 | 100.0 | 100.0 |
| (100,200, 200) | FDS | 38.9 | 88.2 | 99.9 | 100.0 | 100.0 | 39.2 | 85.8 | 99.8 | 100.0 | 100.0 |
| Empirical powers in Model 3.3 | |||||||||||
| (50,100, 100) | FDS | 39.7 | 37.4 | 37.1 | 40.3 | 53.1 | 30.1 | 28.8 | 32.1 | 33.3 | 35.8 |
| (100,100, 100) | FDS | 47.2 | 45.9 | 47.7 | 50.4 | 50.1 | 40.6 | 43.2 | 41.5 | 45.1 | 46.8 |
| (100,100, 200) | FDS | 68.1 | 69.8 | 70.3 | 73.6 | 73.1 | 64.2 | 65.2 | 65.6 | 68.8 | 67.8 |
| (100,200, 200) | FDS | 84.5 | 87.5 | 87.2 | 89.3 | 88.8 | 83.2 | 85.0 | 85.1 | 86.3 | 86.1 |
Model 1.1: The population correlation matrix is set as where ρ is taken as 0.0 and 0.5.
Model 1.2: The population correlation matrix is set as , where R* = Ip, 1p is a p × 1 vector of ones and ek is the kth column of the p × p identity matrix. When ϵ = 1, the signal pattern of R1 − R∗ is dense. When ϵ = 0, the signal pattern of R1 − R∗ is sparse.
Model 1.3: The population correlation matrix is set as where and ϵ = 0.09 and 0.12. In this case, the signal pattern of R1 − R∗ is sparse.
Model 2.1: The population correlation matrices are set as with ρ = 0.00, 0.25, 0.50, and 0.75. The simulation results for ρ = 0.25 and 0.75 are included in the supplementary file.
Model 2.2: The population correlation matrices are set as and with ϵ = 0.05 and 0.08. In this case, the signal pattern of R2 − R1 is dense.
Model 2.3: The population correlation matrices are set as and with ρ = 0.05, 0.08, 0.10, and 0.12 and ϵp = exp(0.008p)/[1 + exp(0.008p)]. In this case, the signal pattern of R2 − R1 is sparse. The simulation results for ρ = 0.05, 0.10, and 0.12 are included in the supplementary file.
Model 2.4: The population correlation matrices are set as R1=Ip and with ϵ = 0.2, 0.225, 0.25, and 0.275. In this case, the signal pattern of R2 − R1 is between dense case and sparse case. The simulation results are included in the supplementary file.
Model 3.1: The three population correlation matrices are taken as R1 = R2 = R3 = Ip. The model is used for evaluating the empirical performance on Type I errors of the proposed test Mn,3.
Model 3.2: The three population correlation matrices are taken as R1 = R2 = Ip and . Here R3 − R1 or R3 − R2 is dense.
Model 3.3: The three population correlation matrices are taken as R1 = R2 = Ip and . Here R3 – R1 or R3 – R2 is a little sparse.
Overall, the Type I error rates for our tests “FDS” and “MAX” are relatively accurate for all sample sizes, for all dimensions, for all correlation matrices, and for the two different distributions of error terms. For the one sample testing problem, “FDS” and “MAX” can deal with an arbitrary correlation matrix R∗, whereas other test statistics “GHPY” and “LX” cannot. It seems that both ρ and p have some minor impact on its Type I error rates. The proposed tests ‘FDS” and “MAX” perform very well for both sparse and dense alternatives. Consistent with our expectations, the statistical powers for rejecting the null hypothesis increase as ϵ, n, and p increase. It seems that “MAX” has a little better performance than “FDS”.
For the two- and three- sample testing problems, “FDS” and “MAX” also can deal with arbitrary correlation matrices. It seems that ρ, p, and the error distribution have little impact on its Type I error rates. The proposed tests “FDS” and “MAX” perform reasonably well for sparse alternatives, dense alternatives, and between sparse and dense alternatives. It seems that “MAX” is slightly better than “FDS”.
6. Real data analysis.
6.1. Alzheimer’s Disease Neuroimaging Initiative (ADNI) data.
“Data used in the preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). The ADNI was launched in 2003 as a public-private partnership, led by Principal Investigator Michael W. Weiner, MD. The primary goal of ADNI has been to test whether serial magnetic resonance imaging (MRI), positron emission tomography (PET), other biological markers, and clinical and neuropsychological assessment can be combined to measure the progression of mild cognitive impairment (MCI) and early Alzheimer’s disease (AD). For up-to-date information, see www.adni-info.org.”1 We consider 749 T1 weighted images collected at the baseline of ADNI1, consisting of 206 normal subjects, 364 mild cognitive impairment (MCI) subjects and 179 Alzheimer’s disease (AD) subjects. These scans were performed on a 1.5T MRI scanners using a sagittal MPRAGE sequence and the typical protocol includes the following parameters: repetition time (TR) = 2400 ms, inversion time (TI) = 1000 ms, flip angle = 8°, and field of view (FOV) = 24 cm with a 256×256×170 mm3 acquisition matrix in the x, y, and z dimensions, which yields a voxel size of 1.25 × 1.26 × 1.2 mm3.
The T1-weighted images were processed using the Hierarchical Attribute Matching Mechanism for Elastic Registration (HAMMER) pipeline. The processing steps include anterior commissure and posterior commissure correction, skull-stripping, cerebellum removal, intensity inhomogeneity correction, and segmentation. We performed automatic regional labeling by labeling the template and by transferring the labels following the deformable registration of subject images. Finally, we labeled 93 regions of interest (ROIs) and computed their volumes for each subject.
6.2. Group comparisons.
We are interested in characterizing differences among the three correlation matrices of ROI volumes for normal subjects, MCI subjects and AD subjects, which are denoted as RNC, RMCI, and RAD, respectively. Statistically, we test three two sample testing problems, including RNC = RMCI, RNC = RAD, and RMCI = RAD, and one three sample testing problem, that is, RNC = RMCI = RAD.
We applied the test statistics Mn,2 and Mn,3 to carry out these tests as follows. First, for each ROI, we fitted a linear regression model with its ROI volume as response and age, gender and whole brain volume as covariates by using data obtained from all subjects. Second, for each group, we calculated its correlated matrix based on the residuals of all ROIs obtained from the first step. Figure 1 presents the correlation matrices corresponding to the three groups. Then, we clustered the 93 ROIs according to the correlation matrix of the normal control group. For example, Cluster 1 includes the large area of prefrontal cortex, and its functions span over the frontoparietal control network (orbitofrontal cortex, middle frontal gyrus), default node network and ventral attention network. This region has been implicated in decision making, complex cognitive behavior, processing of higher information, decision making, personal expression, social behavior moderating, attention, memory, recognizing faces, characters and etc. Third, we calculated the p–value of testing RNC = RMCI, that of RNC = RAD, and that of RMCI = RAD as 1.23 × 10−9, 0, and 1.78×10−11, respectively. Fourth, we calculated the p–value of testing RNC = RMCI = RAD as 0.
Fig 1.

Graphical display of correlation matrices of normal subjects, MCI subjects and AD subjects.
Supplementary Material
Fig 2.

Graphical display of difference between correlation matrices of normal subjects, MCI subjects and AD subjects.
Table 2.
Empirical sizes in Model 2.1 and empirical powers in Models 2.2–2.3 for H02 (in percentage)
| wij ~ N (0, 1) | wij ~ Gamma(4,2)–2 | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| ϵ | (n1,n2) | Methods | p=50 | 100 | 200 | 500 | 50 | 100 | 200 | 500 |
| Empirical sizes in Model 2.1 | ||||||||||
| 0.0 | (100, 100) | FDS | 5.85 | 6.00 | 6.20 | 6.85 | 4.85 | 4.65 | 5.15 | 5.50 |
| MAX | 4.65 | 5.15 | 4.90 | 5.05 | 3.50 | 3.45 | 3.85 | 4.10 | ||
| CZ | 5.00 | 5.10 | 5.85 | 5.25 | 3.25 | 3.70 | 3.55 | 3.10 | ||
| (150, 150) | FDS | 5.85 | 6.45 | 5.50 | 5.50 | 4.85 | 5.55 | 4.80 | 4.65 | |
| MAX | 4.70 | 5.55 | 5.10 | 5.15 | 4.00 | 4.35 | 4.25 | 3.55 | ||
| CZ | 4.40 | 4.45 | 4.80 | 5.05 | 3.35 | 3.95 | 4.25 | 2.80 | ||
| (200, 200) | FDS | 5.30 | 5.20 | 5.25 | 5.80 | 4.85 | 4.95 | 5.40 | 5.25 | |
| MAX | 4.25 | 4.55 | 4.45 | 4.75 | 3.60 | 3.70 | 4.00 | 4.20 | ||
| CZ | 4.20 | 5.45 | 4.50 | 4.70 | 3.60 | 4.00 | 3.00 | 3.35 | ||
| 0.5 | (100, 100) | FDS | 6.15 | 6.75 | 5.60 | 6.25 | 5.30 | 4.20 | 5.50 | 5.40 |
| MAX | 5.30 | 5.75 | 5.00 | 5.60 | 3.45 | 4.15 | 4.60 | 4.55 | ||
| CZ | 5.55 | 5.45 | 5.50 | 5.55 | 4.00 | 4.40 | 4.15 | 3.95 | ||
| (150, 150) | FDS | 6.10 | 5.95 | 6.70 | 6.25 | 5.40 | 5.75 | 5.70 | 5.05 | |
| MAX | 5.30 | 4.75 | 5.85 | 4.85 | 4.05 | 4.04 | 4.20 | 4.45 | ||
| CZ | 5.60 | 5.55 | 4.75 | 5.00 | 3.75 | 3.60 | 3.85 | 4.00 | ||
| (200, 200) | FDS | 5.20 | 6.30 | 6.30 | 6.90 | 5.45 | 4.90 | 6.10 | 5.15 | |
| MAX | 4.20 | 5.30 | 5.05 | 5.95 | 4.50 | 3.65 | 5.65 | 4.30 | ||
| CZ | 4.20 | 4.55 | 4.95 | 5.30 | 4.15 | 3.70 | 4.30 | 4.05 | ||
| Empirical sizes in Model 2.2 | ||||||||||
| 0.05 | (100, 100) | FDS | 49.6 | 83.6 | 99.2 | 100.0 | 48.6 | 82.8 | 99.9 | 100.0 |
| MAX | 42.5 | 78.8 | 98.7 | 100.0 | 41.1 | 77.6 | 99.9 | 100.0 | ||
| CZ | 9.8 | 10.4 | 12.6 | 14.3 | 9.7 | 9.8 | 10.3 | 10.7 | ||
| (150, 150) | FDS | 71.8 | 96.7 | 100.0 | 100.0 | 70.8 | 97.3 | 100.0 | 100.0 | |
| MAX | 64.7 | 94.9 | 99.9 | 100.0 | 62.2 | 95.5 | 100.0 | 100.0 | ||
| CZ | 14.4 | 15.0 | 16.8 | 16.6 | 11.5 | 11.8 | 14.1 | 14.3 | ||
| (200, 200) | FDS | 84.9 | 99.6 | 100.0 | 100.0 | 82.4 | 99.8 | 100.0 | 100.0 | |
| MAX | 79.5 | 99.2 | 100.0 | 100.0 | 77.3 | 99.5 | 100.0 | 100.0 | ||
| CZ | 14.9 | 17.5 | 19.5 | 23.6 | 14.6 | 15.5 | 17.9 | 20.6 | ||
| 0.08 | (100, 100) | FDS | 87.7 | 99.8 | 100.0 | 100.0 | 90.1 | 99.4 | 100.0 | 100.0 |
| MAX | 84.2 | 99.7 | 100.0 | 100.0 | 85.8 | 99.3 | 99.9 | 100.0 | ||
| CZ | 18.4 | 21.8 | 24.1 | 30.1 | 19.3 | 20.5 | 21.6 | 26.1 | ||
| (150, 150) | FDS | 98.5 | 100.0 | 100.0 | 100.0 | 97.9 | 100.0 | 100.0 | 100.0 | |
| MAX | 97.5 | 100.0 | 100.0 | 100.0 | 97.2 | 100.0 | 100.0 | 100.0 | ||
| CZ | 31.4 | 32.8 | 40.2 | 43.4 | 26.0 | 31.0 | 36.6 | 40.1 | ||
| (200, 200) | FDS | 99.7 | 100.0 | 100.0 | 100.0 | 99.8 | 100.0 | 100.0 | 100.0 | |
| MAX | 99.5 | 100.0 | 100.0 | 100.0 | 99.6 | 100.0 | 100.0 | 100.0 | ||
| CZ | 37.6 | 45.4 | 53.8 | 62.3 | 37.3 | 42.8 | 49.6 | 58.7 | ||
| Empirical powers in Model 2.3 | ||||||||||
| 0.08 | (100, 100) | FDS | 65.5 | 86.8 | 99.6 | 99.6 | 61.8 | 78.5 | 97.6 | 96.5 |
| MAX2 | 72.8 | 90.6 | 99.8 | 99.8 | 68.6 | 83.3 | 98.8 | 97.5 | ||
| CZ | 78.7 | 93.2 | 99.9 | 99.9 | 76.0 | 87.7 | 99.5 | 98.3 | ||
| (150, 150) | FDS | 94.1 | 99.4 | 100.0 | 100.0 | 92.6 | 98.5 | 99.9 | 100.0 | |
| MAX | 96.5 | 99.6 | 100.0 | 100.0 | 95.2 | 99.3 | 99.9 | 100.0 | ||
| CZ | 97.7 | 99.9 | 100.0 | 100.0 | 97.0 | 99.5 | 99.9 | 100.0 | ||
| (200, 200) | FDS | 99.5 | 100.0 | 100.0 | 100.0 | 99.1 | 99.9 | 100.0 | 100.0 | |
| MAX | 99.9 | 100.0 | 100.0 | 100.0 | 99.5 | 100.0 | 100.0 | 100.0 | ||
| CZ | 99.9 | 100.0 | 100.0 | 100.0 | 99.7 | 100.0 | 100.0 | 100.0 | ||
ACKNOWLEDGEMENT
The authors thank the editor, the associate editor and the referees for their constructive comments and suggestions that led to a significant improvement of this article.
APPENDIX A: SOME EXPRESSIONS
Let r∗khj be the (h, j) entry of and rℓkhj be the (h, j) entry of for k = 1/2, 1, 3/2, 2, 3. Let ej be the jth column of the p × p identity matrix.
A.1. Expressions of µzA, µz0 and for one population in Theorem 2.1 and 2.2. Expression of µzA and µz0:
When R1 = R∗, we have
Expression of
When R1 = R∗, we have .
A.2. Expressions of , , and Theorems 3.1–3.4.
where and is the (h, j) entry of for k = 1/2, 1, 3/2, 2, 3.
When , we have
where is the (h, j) entry of Rk for k= 1/2, 1, 3/2. 2, 3. We have where
and
When , we have , and
For we have
Where we have .
Footnotes
Data used in preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report. A complete listing of ADNI investigators can be found at: http://adni.loni.usc.edu/wpcontent/uploads/howtoapply/ADNI_Acknowledgement_List.pdf.
ADNI manuscript citation guidelines. https://adni.loni.usc.edu/wp-content/uploads//how to apply /ADNI DSP Policy.pdf
Contributor Information
Shurong Zheng, KLAS and School of Mathematics and Statistics, Northeast Normal University, Changchun, Jilin, China zhengsr@nenu.edu.cn.
Guanghui Cheng, School of Economics and Statistics, Guangzhou University, Guangzhou, China chenggh845@nenu.edu.cn.
Jianhua Guo, KLAS and School of Mathematics and Statistics, Northeast Normal University, Changchun, Jilin, China jhguo@nenu.edu.cn.
Hongtu Zhu, Department of Biostatistics, The University of North Carolina at Chapel Hill Chapel Hill, NC, USA htzhu@email.unc.edu.
REFERENCES
- Aitkin MA (1969), “Some tests for correlation matrices,” Biometrika, 56(2), 443–446. [Google Scholar]
- Anderson TW (2003), An Introduction to Multivariate Statistical Analysis, 3rd Edition, New York: Wiley. [Google Scholar]
- Bai ZD, and Silverstein JW (2004), “CLT for linear spectral statistics of large-dimensional sample covariance matrices,” The Annals of Probability, 32(1A), 553–605. [Google Scholar]
- Bartlett MS, and Rajalakshman DV (1953), “Goodness of fit tests for simultaneous autoregressive series,” Journal of the Royal Statistical Society (Series B), 15, 107–124. [Google Scholar]
- Bodwin K, Zhang K, and Nobel A (2016), “A testing-based approach to the discovery of differentially correlated variable sets,” arXiv:1509.08124v2,. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Browne MW (1978), “The likelihood ratio test for the equality of correlation matrices,” The British Journal of Mathematical and Statistical Psychology, 31(2), 209–217. [Google Scholar]
- Cai T, Liu WD, and Xia Y (2013), “Two-sample covariance matrix testing and support recovery in high-dimensional and sparse settings,” Journal of the American Statistical Association, 108(501), 265–277. [Google Scholar]
- Cai TT (2017), “Global Testing and Large-Scale Multiple Testing for High-Dimensional Covariance Structures,” Annual Review of Statistics and Its Application, 4, 423–446. [Google Scholar]
- Cai TT, and Jiang TF (2011), “Limiting laws of coherence of random matrices with applications to testing covariance structure and construction of compressed sensing matrices,” The Annals of Statistics, 39(3), 1496–1525. [Google Scholar]
- Cai TT, and Ma ZM (2013), “Optimal hypothesis testing for high dimensional covariance matrices,” Bernoulli, 19(5B), 2359–2388. [Google Scholar]
- Cai TT, and Zhang A (2016), “Inference for high-dimensional differential correlation matrices,” Journal of multivariate analysis, 143, 107–126. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen SX, Zhang LX, and Zhong PS (2010), “Tests for high-dimensional covariance matrices,” Journal of the American Statistical Association, 105(490), 810–819. [Google Scholar]
- Cole N (1968), “The likelihood ratio test of the equality of correlation matrices. 1968–65: The L. L. Thurstone Psychometric Laboratory, University of North Carolina, Chapel Hill, North Carolina,”. [Google Scholar]
- Debashis P, and Alexander A (2014), “Random matrix theory in statistics: a review,” Journal of Statistical Planning and Inference, 150, 1–29. [Google Scholar]
- Gao JT, Han X, Pan GM, and Yang YR (2017), “High-dimensional correlation matrices: the central limit theorem and its application,” Journal of the Royal Statistical Society (Series B), 79(3), 677–693. [Google Scholar]
- Gupta AK, Johnson BE, and Nagar DK (2013), “Testing equality of several correlation matrices,” Revista Colombiana de Estad´ıstica, 36(2), 237–258. [Google Scholar]
- Jennrich RI (1970), “An asymptotic χ2 test for the equality of two correlation matrices,” Journal of the American Statistical Association, 65(330), 904–912. [Google Scholar]
- Kullback S (1967), “On testing correlation matrices,” Journal of the Royal Statistical Society (Series C), 16(1), 80–85. [Google Scholar]
- Larntz K, and Perlman MD (1985), “A simple test for the equality of correlation matrices,” Technical Report, 63. [Google Scholar]
- Li DN, and Xue LZ (2015), “Joint limiting laws for high-dimensional independence tests,” arXiv: 1512.08819V1,. [Google Scholar]
- Qiu YM, and Chen SX (2012), “Test for bandedness of high-dimensional covariance matrices and bandwidth estimation,” The Annals of Statistics, 40(3), 1285–1314. [Google Scholar]
- Schott JR (1996), “Testing for the equality of several correlation matrices,” Statistics and probability letters, 27, 85–89. [Google Scholar]
- Schott JR (2005), “Testing for complete independence in high dimensions,” Biometrika, 92(4), 951–956. [Google Scholar]
- Shao QM, and Zhou WX (2014), “Necessary and sufficient conditions for the asymptotic distributions of coherence of ultra-high dimensional random matrices,” The Annals of Probability, 42(2), 623–648. [Google Scholar]
- Zhou C, Han F, Zhang XS, and Liu H (2015), “An Extreme-value approach for testing the equality of large U-statistic based correlation matrices,” arXiv:1502.03211,. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
