Abstract
In this work, we show that Spearman’s correlation coefficient test about found in most statistical software is theoretically incorrect and performs poorly when bivariate normality assumptions are not met or the sample size is small. There is common misconception that the tests about are robust to deviations from bivariate normality. However, we found under certain scenarios violation of the bivariate normality assumption has severe effects on type I error control for the common tests. To address this issue, we developed a robust permutation test for testing the hypothesis based on an appropriately studentized statistic. We will show that the test is asymptotically valid in general settings. This was demonstrated by a comprehensive set of simulation studies, where the proposed test exhibits robust type I error control, even when the sample size is small. We also demonstrated the application of this test in two real world examples.
Keywords: rank correlation, studentized, small sample, non-normality
1. Introduction
The concept of correlation and regression was originally conceived by Galton when studying how strongly the characteristics of one generation of living things manifested in the following generation (Stanton, 2001). The ideas prompting the development of more mathematically rigorous treatment of correlation were developed by Karl Pearson in 1896, which yielded the well-known Pearson Product Moment Correlation Coefficient (Pearson, 1896) given as
| (1) |
where and are two random variables from a non-degenerative joint distribution , denotes the covariance, and and , and are the population means and standard deviations, respectively. If we let denote paired i.i.d. observations, then the sample Pearson correlation coefficient is given as
| (2) |
where and are the sample means of and , respectively.
Shortly after Pearson’s work was published, Spearman introduced the rank correlation coefficient in 1904. The Spearman correlation has advantages of being robust to extreme values and disparities between the marginal distributions between two variables (Spearman, 1961). However, it should be noted that K. Pearson, in his biography of Galton, says that the latter "dealt with the correlation of ranks before he even reached the correlation of variates, i.e. about 1875", but Galton apparently did not publish anything explicitly about this result (Kendall and Stuart, 1979). Mathematically, Spearman’s correlation coefficient is defined as the Pearson correlation coefficient on the ranks of and denoted as
| (3) |
where and are the ranks of and , respectively, . In general, when discussing the Spearman correlation coefficient little attention is given to its population measure. However, if we consider that converges to , then one may consider the population measure linked to Spearman’s sample correlation coefficient as,
| (4) |
where and are the marginal cumulative distribution functions (CDFs) for and , respectively. The expectation is taken over the joint distribution . The sample estimator of can be obtained by replacing the original observations with their ranks in Equation 2,
| (5) |
For samples from a bivariate normal population, there is also a known relation between the Spearman and Pearson correlation coefficients (Moran, 1948), which is
| (6) |
While Pearson’s measures the linear relationship between two random variables it is often described that Spearman’s measures the strength and direction of association between and monotonically, thus it may be considered a more general measure of association, albeit it does measure the linear association between and . Spearman’s correlation coefficient is also less sensitive to extreme values because it is rank based. Due to these advantages, it is widely used as a measure of association between two measurements. It is often of interest to test whether two random variables are correlated, i.e. . The common methods include (1) a -distribution based test, (2) a test based on Fisher’s transformation, and (3) what we term the naive permutation test.
The -distribution based test is commonly used when the sample size is large, with the -statistic defined as
This statistic was first used for Pearson’s correlation coefficient. Under bivariate normality assumptions it approximately follows student’s distribution with degrees of freedom under (Edgell and Noon, 1984). When used for Spearman’s , this test incorrectly based on the approximate bivariate normality of the ranks. For the test based on Fisher’s transformation, the statistic is defined as
Under bivariate normality assumptions the transformed statistic approximately follows normal distribution under (Fieller et al., 1957).
Permutation tests have been applied in a broad range of scenarios (Pauly and Smaga, 2020). For small sample size cases, naive permutation tests are also often used for testing , where and are randomly shuffled separately and independently to simulate the sample distribution of under , which may be an invalid test for testing given does not imply , where denotes the joint CDF of , . For example, Marozzi proposed a permutation test for Kendall coefficient, which is a rank-based measure of concordance between multiple criteria (Marozzi, 2014). When there are only two judges, the Kendall is a linear transformation of , so the tests on these two coefficients are equivalent. However, similar to the naive permutation test, the method did not account for the difference between independence and zero correlations.
These tests are so widely used that they are often the default options in common statistical software packages such as R (Team et al., 2013) and SAS (Institute, 2015). Both software packages (cor.test function for R and CORR procedure for SAS) by default uses the -distribution based test, as shown in their documentations. However, there is little discussion that these tests rely on the untenable assumption that the underlying sample distribution of the ranks has a bivariate normal distribution, which is in fact an impossibility. Even among those who noted this assumption, there is a misconception that the above tests are robust to such deviations because Spearman’s is rank based. This is exemplified in a discussion by Fieller et al. in their article (Fieller et al., 1957),
Conversely, starting from any bivariate distribution we can always find monotonic transformations , to standardized normal variates and . The resulting bivariate distribution will not necessarily be bivariate normal, but we think it likely that in practical stations it would not differ greatly from this form. This is a field in which further investigation would be of considerable interest.
However, as we will show in Section 3, all the commonly used tests about as discussed above, including the naive permutation test, are not even asymptotically valid when the exchangeabilty assumptions are violated under . That is, even when , the type I error cannot be controlled at the desired level. In some cases, the type I error can severely drift away from the desired level as the sample size increases! An undesirable feature that is more notable in the era of “big-data”. Another variation of this approach is the Fisher-Yates coefficient, which transforms the original and to their corresponding normal quantiles before the testing (Fisher and Yates, 1938). Although the marginal distributions of the transformed variates take a pseudo normal form, the joint normality of these transformed values is not guaranteed.
In terms of our modified permutation test it is important to note the classic large sample result in Serfling where the “distribution free” large sample normal approximation for the sampling distribution for Pearson’s sample correlation coefficient is derived using the multivariate delta method (Kendall and Stuart, 1979). This method guarantees type I error converges to when given finite fourth moments. A straightforward way to obtain a similar result for Spearman’s correlation is given by replacing the with . The test is asymptotically valid because the ranks are asymptotically independent, as we will discuss in Section 2. Even though large sample approximations about these estimators are asymptotically valid they tend to suffer inflated type I errors in the small sample setting, e.g. .
Permutation tests provide a strong alternative testing approach. The permutation approach has been applied to a variety of univariate and multivariate problems and its properties have been extensively studied (Giancristofaro and Bonnini, 2008; Arboretti Giancristofaro et al., 2009; Pesarin and Salmaso, 2010a; Basso and Salmaso, 2011; Pesarin and Salmaso, 2012, 2013; Arboretti et al., 2014; Giancristofaro, 2014; Salmaso, 2015). For an overview of permutation tests, please see Salmaso and Pesarin (2010) and Pesarin and Salmaso (2010b). Recently, DiCiccio and Romano have shown that the permutation distribution of Pearson’s correlation coefficient does not converge to the sampling distribution when two random variables are dependent but uncorrelated (DiCiccio and Romano, 2017). Therefore, a a simple permutation test ignoring possible dependency structures can lead to invalid inference about the Pearson’s and related correlation coefficients (DiCiccio and Romano, 2017; Hutson and Yu, 2021).
In this work, we show that a naive permutation test of suffers a similar problem. To address this issue, we propose a studentized permutation test for Spearman’s correlation , which extends the work or Diccicio and Romano for Pearson’s correlation coefficient (DiCiccio and Romano, 2017). We will show that the proposed test is asymptotically valid under general assumptions and is exact under exchangeability assumptions when , i.e. more simply when and are independent. We show that our newly proposed test has robust Type I error control even when the exchangeability assumption does not hold and the sample distribution is non-normal. Importantly, even when the sample size is as small as 10, the type I error is still well controlled, which is advantageous over the tests based on large sample approximations. This will be illustrated by a set of simulation studies. Finally, we will demonstrate the application of this test in real world examples of transcriptomic data of TCGA breast cancer patients, as well as a data set of PSA levels and age.
2. Methods
2.1. Spearman’s permutation correlation test
Spearman’s coefficient is the Pearson correlation coefficient of the ranks of and , that is . When there are ties in the data, the tied ranks are typically taken as an average. Unlike Pearson’s correlation coefficient , which measures the linear relationship between two random variables, Spearman’s correlation coefficient measures a monotonic association, thus is far less restrictive. Note that Spearman’s correlation coefficient is also the linear measure between and . It is also less sensitive to non-normality or extreme values.
Despite the above advantages, it is a misconception that the tests of based on the bivariate normality assumptions underlying the original data will be robust to the deviation from this assumption. In fact, tests of typically suffer similar issue as for . We also emphasize that “normality” refers to the joint normality as opposed to marginal normality, because two random variables that are marginally normal can have a joint non-normal distribution. Therefore, the Fisher-Yates coefficient, which back transforms a variables rank through the normal quantile function does not provide what heuristically one may consider as a simple correction. In Section 3, we will empirically show that violation of the joint normality assumption will have severe effect on type I error control. In addition, it is in fact impossible for the joint distribution of the ranks to be bivariate normal.
Our approach is to replace the observations with their ranks in order to develop a Spearman’s correlation permutation test analog to the Pearson’s correlation permutation test, with some subtle differences. The studentized permutation test of the Pearson’s proposed by DiCiccio and Romano only requires finite fourth moments and that the paired observations are i.i.d (DiCiccio and Romano, 2017), i.e. and are independent and identically distributed when . However, it should be noted the ranks and are no longer independent pairs. For example, suppose we have , then after knowing we will immediately know .
In spite of this, we can show that and are asymptotically independent and follow identical distributions when . This result follows naturally from the convergence of empirical CDF to CDF. Without loss of generality, we start by considering a single variable , Since and are i.i.d. observations, we have , being i.i.d as well, so . Also we have and being the ranks of and , so and , where is the empirical CDF of . By the Strong Law of Large Numbers, we have almost surely, thus and converge almost surely to a pair of independent variables and , Therefore, and are asymptotically independent. Note that this result is on two observations of the same random variable, i.e. and . It does not concern the dependency between and , or between and .
It is obvious that and follow the same distribution, and the same conclusion applies to and . Therefore, two paired observations and are identically, and asymptotically independently distributed when . Consequently, the exchangeability condition of i.i.d observations holds asymptotically, so the test will be asymptotically exact.
With the above results, the one-sided studentized permutation test for testing versus is performed by the following steps. The test is implemented in the R perk (permutation tests of correlation c(k)oefficients) package, which will be available on CRAN (The Comprehensive R Archive Network, https://cran.r-project.org/) and GitHub (https://github.com/hyu-ub/perk).
For paired i.i.d. observations , calculate their ranks within each random variable, .
Estimate the Spearman’s using Equation 5 as .
- Estimate the variance of sample estimates by
Calculate the studentized statistic .
- Randomly shuffle for times. For each permutation, calculate the permuted studentized statistic , . Note that needs to be calculated for each permutation because is not permutation invariant. Specifically, for permutation , we have
- Calculate the p-value by
Reject if .
3. Simulations
We examined the Type I error control across all of the tests introduced above using distributions commonly found in the literature for these examinations across a wide range of settings (DiCiccio and Romano, 2017; Hutson, 2019). For our simulation study, we focused on testing versus , with sample sizes = 10, 25, 50, 100, 200. Each simulation utilized 10, 000 Monte Carlo replications and the number of permutations used is 1, 000. We compared the test, Fisher’s -transformation (Fisher’s ), Fisher-Yates method, Serfling’s large sample normal approximation (Asymp Norm), naive permutation test (Permute), and studentized permutation test (Stu Permute). The Type I error control for was examined. The simulation scenarios 1 through 5 from DiCiccio and Romano. Two additional distributions were studied as well:
Multivariate normal (MVN) with mean zero and identity covariance.
Exponential given as where , , and is uniformly distributed on the two dimensional unit circle.
where and , where and are iid random variables.
Circular given as the uniform distribution on a two dimensional unit circle.
Multivariate -distribution (MVT) with 5 degrees of freedom.
Mixture of two bivariate normal distributions given as where , , . We select a range of ’s: 0.1, 0.3, 0.6 and 0.9 to simulate different degrees of dependencies between and (MVN 1, MVN 3, MVN6, MVN 9).
Mixture of four bivariate normal distributions (MVN 45), given as where , . In addition, , where , , , .
The results in Table 1 and Figure 2 show that the large sample asymptotic normal approximation has inflated type I error rates for all distributions when . The test, Fisher’s test, Fisher-Yates, and naive permutation tests tend to be over-conservative for the exponential and circular distributions. While for the distribution, the type I error is consistently inflated. Note that, for these tests, such deviation cannot be corrected as sample size increases. Instead, they may converge to an arbitrary level, either lower or higher than .
Table 1:
Type I error rate of testing versus .
| Distribution | n | t test | Fisher’s Z | Fisher-Yates | Asymp Norm | Permute | Stu Permute |
|---|---|---|---|---|---|---|---|
| MVN | 10 | 0.0498 | 0.0420 | 0.0479 | 0.1223 | 0.0495 | 0.0457 |
| 25 | 0.0501 | 0.0444 | 0.0499 | 0.0811 | 0.0494 | 0.0487 | |
| 50 | 0.0512 | 0.0473 | 0.0509 | 0.0639 | 0.0508 | 0.0525 | |
| 100 | 0.0529 | 0.0484 | 0.0527 | 0.0596 | 0.0530 | 0.0518 | |
| 200 | 0.0479 | 0.0438 | 0.0493 | 0.0525 | 0.0487 | 0.0496 | |
| Exponential | 10 | 0.0707 | 0.0623 | 0.0813 | 0.1300 | 0.0715 | 0.0479 |
| 25 | 0.0744 | 0.0689 | 0.0914 | 0.0804 | 0.0741 | 0.0492 | |
| 50 | 0.0795 | 0.0724 | 0.1061 | 0.0680 | 0.0798 | 0.0509 | |
| 100 | 0.0772 | 0.0706 | 0.0975 | 0.0561 | 0.0774 | 0.0497 | |
| 200 | 0.0789 | 0.0735 | 0.0998 | 0.0549 | 0.0785 | 0.0509 | |
| 10 | 0.0625 | 0.0531 | 0.0693 | 0.1275 | 0.0606 | 0.0489 | |
| 25 | 0.0631 | 0.0574 | 0.0846 | 0.0768 | 0.0629 | 0.0480 | |
| 50 | 0.0631 | 0.0579 | 0.0910 | 0.0614 | 0.0628 | 0.0457 | |
| 100 | 0.0677 | 0.0611 | 0.1082 | 0.0567 | 0.0672 | 0.0491 | |
| 200 | 0.0728 | 0.0675 | 0.1149 | 0.0572 | 0.0731 | 0.0521 | |
| Circular | 10 | 0.0124 | 0.0095 | 0.0041 | 0.0907 | 0.0124 | 0.0514 |
| 25 | 0.0052 | 0.0040 | 0.0004 | 0.0620 | 0.0048 | 0.0474 | |
| 50 | 0.0031 | 0.0024 | 0.0001 | 0.0575 | 0.0030 | 0.0487 | |
| 100 | 0.0021 | 0.0015 | < 0.0001 | 0.0537 | 0.0024 | 0.0497 | |
| 200 | 0.0016 | 0.0009 | < 0.0001 | 0.0489 | 0.0020 | 0.0482 | |
| MVT | 10 | 0.0582 | 0.0502 | 0.0626 | 0.1305 | 0.0584 | 0.0497 |
| 25 | 0.0596 | 0.0543 | 0.0726 | 0.0825 | 0.0602 | 0.0496 | |
| 50 | 0.0542 | 0.0489 | 0.0700 | 0.0608 | 0.0553 | 0.0456 | |
| 100 | 0.0584 | 0.0535 | 0.0740 | 0.0566 | 0.0574 | 0.0496 | |
| 200 | 0.0590 | 0.0539 | 0.0852 | 0.0527 | 0.0590 | 0.0496 | |
| MVN 1 | 10 | 0.0541 | 0.0458 | 0.0499 | 0.1271 | 0.0531 | 0.0534 |
| 25 | 0.0484 | 0.0451 | 0.0490 | 0.0777 | 0.0482 | 0.0498 | |
| 50 | 0.0481 | 0.0428 | 0.0478 | 0.0631 | 0.0484 | 0.0492 | |
| 100 | 0.0497 | 0.0452 | 0.0520 | 0.0564 | 0.0498 | 0.0497 | |
| 200 | 0.0495 | 0.0462 | 0.0523 | 0.0529 | 0.0493 | 0.0488 | |
| MVN 3 | 10 | 0.0547 | 0.0467 | 0.0539 | 0.1341 | 0.0544 | 0.0497 |
| 25 | 0.0557 | 0.0514 | 0.0607 | 0.0814 | 0.0551 | 0.0516 | |
| 50 | 0.0571 | 0.0498 | 0.0607 | 0.0686 | 0.0562 | 0.0508 | |
| 100 | 0.0530 | 0.0473 | 0.0651 | 0.0561 | 0.0544 | 0.0497 | |
| 200 | 0.0529 | 0.0468 | 0.0603 | 0.0517 | 0.0508 | 0.0473 | |
| MVN 6 | 10 | 0.0648 | 0.0555 | 0.0740 | 0.1328 | 0.0643 | 0.0461 |
| 25 | 0.0694 | 0.0624 | 0.0880 | 0.0832 | 0.0686 | 0.0526 | |
| 50 | 0.0658 | 0.0606 | 0.0917 | 0.0638 | 0.0650 | 0.0494 | |
| 100 | 0.0706 | 0.0644 | 0.0988 | 0.0590 | 0.0694 | 0.0498 | |
| 200 | 0.0678 | 0.0615 | 0.1034 | 0.0527 | 0.0667 | 0.0496 | |
| MVN 9 | 10 | 0.0944 | 0.0843 | 0.1173 | 0.1405 | 0.0924 | 0.0573 |
| 25 | 0.0932 | 0.0865 | 0.1276 | 0.0829 | 0.0913 | 0.0502 | |
| 50 | 0.0965 | 0.0891 | 0.1414 | 0.0662 | 0.0948 | 0.0519 | |
| 100 | 0.0997 | 0.0928 | 0.1425 | 0.0593 | 0.0994 | 0.0498 | |
| 200 | 0.0950 | 0.0882 | 0.1531 | 0.0536 | 0.0953 | 0.0494 | |
| MVN 45 | 10 | 0.0564 | 0.0467 | 0.0487 | 0.1301 | 0.0547 | 0.0515 |
| 25 | 0.0492 | 0.0444 | 0.0480 | 0.0791 | 0.0491 | 0.0482 | |
| 50 | 0.0481 | 0.0445 | 0.0507 | 0.0634 | 0.0482 | 0.0512 | |
| 100 | 0.0492 | 0.0447 | 0.0498 | 0.0557 | 0.0497 | 0.0488 | |
| 200 | 0.0521 | 0.0468 | 0.0492 | 0.0552 | 0.0509 | 0.0519 |
Figure 2:
Type I error rate of testing versus .
For MVN 1-9, we simulated a range of dependency among uncorrelated and , where MVN 1 has the weakest and MVN 9 has the strongest dependency (Figure 1). The above four tests showed the type I error rate inflation becomes increasingly severe as the dependency increases. This demonstrates the failure in controlling type I error results from the data being dependent, which can occur when the underlying distribution is non-normal.
Figure 1:
Contour plots for the density of a synthetic data set (n = 10, 000) for each simulation distribution.
The MVN 45 is a case where the dependency of original data is remedied by using the ranks. In this case, the ranks will distribute as if it comes from a bivariate normal distribution, regardless of the distance between the centers of individual Gaussian subpopulations. Therefore, all four tests show well control of the type I error rate.
On the other hand, the studentized permutation test robustly control type I error for all distributions examined, even when the is as small as 10. This demonstrates a clear advantage of the proposed test over all other commonly used tests for Spearman’s correlation coefficient.
We also examined the power of different testing methods when or 0.6 under the scenario of multivariate normal distributions. For other distributions, most of the tests fail to control the type I error except for the proposed test, so the comparison of power is not meaningful. Table 2 shows that the power of the proposed test is generally lower than that of the test and naive permutation test, but the difference is generally less than 3%. The exception is when and , where the loss of power is around 7%. However, this difference quickly diminishes when increases to 25. On the other hand, the proposed test has a comparable power to Fisher’s test in almost all scenarios. These results demonstrate that the proposed test achieve a very robust type I error control at the cost of a small decrease in power.
Table 2:
Power of testing versus under bivariate normal distributions.
| n | t test | Fisher’s Z | Fisher-Yates | Asymp Norm | Permute | Stu Permute | |
|---|---|---|---|---|---|---|---|
| 10 | 0.2130 | 0.1848 | 0.2076 | 0.3909 | 0.2087 | 0.1824 | |
| 25 | 0.4402 | 0.4202 | 0.4502 | 0.5248 | 0.4397 | 0.4143 | |
| 50 | 0.7036 | 0.6861 | 0.7282 | 0.7434 | 0.6996 | 0.6870 | |
| 100 | 0.9339 | 0.9264 | 0.9489 | 0.9406 | 0.9305 | 0.9285 | |
| 200 | 0.9965 | 0.9961 | 0.9983 | 0.9968 | 0.9964 | 0.9965 | |
| 10 | 0.5821 | 0.5461 | 0.5794 | 0.7599 | 0.5774 | 0.5090 | |
| 25 | 0.9376 | 0.9313 | 0.9502 | 0.9587 | 0.9356 | 0.9215 | |
| 50 | 0.9981 | 0.9977 | 0.9995 | 0.9986 | 0.9979 | 0.9973 | |
| 100 | > 0.9999 | > 0.9999 | > 0.9999 | > 0.9999 | > 0.9999 | > 0.9999 | |
| 200 | > 0.9999 | > 0.9999 | > 0.9999 | > 0.9999 | > 0.9999 | > 0.9999 |
All the bivariate distributions studies here are symmetric about the origin. However, because the Spearman’s correlation is invariant to monotonic transformations of and , the results covered in our simulation studies can be readily extended to a wide range of asymmetric distributions, e.g. log-normal distributions.
4. Application
4.1. TCGA breast cancer data
As an illustration of our approach, we tested versus using The Cancer Genome Atlas (TCGA) breast cancer RNA sequencing (RNA-seq) data. The gene abundance was RSEM normalized (Li and Dewey, 2011). Fibroblast growth factor (FGF)2, FGF4, FGF7 and FGF20 are representative paracrine FGFs binding to heparan-sulfate proteoglycan and fibroblast growth factor receptors (FGFRs), whereas FGF19, FGF21 and FGF23 are endocrine FGFs binding to Klotho and FGFRs. FGFR1 is relatively frequently amplified and overexpressed in breast and lung cancer, and FGFE2 in gastric cancer. Moreover, FGF2 activates human dermal fibroblasts through transcriptional downregulation of the TP53 gene (Katoh, 2016). In this application, we examine whether the transcriptomic abundance of FGFR1 is correlated with that of TP53. To investigate the performance in small sample settings, we selected 18 samples from 17 mucinous carcinoma patients. The scatter plot of log-transformed TP53 and FGFR1 abundances is shown in Figure 3A. The marginal normality of data was examined by Shapiro-Wilk test and the bivariate normality was examined by Henze-Zikler test. The values of Shapiro-Wilk tests for log-transformed TP53 and FGFR1 abundances are 0.6077 and 0.0644, respectively. The value of Henze-Zikler test is 0.3478. Although there is no statistical significance, the marginal of FGFR1 abundance likely deviates from normal distribution.
Figure 3:
Scatter plot of log-transformed TP53 versus FGFR1 abundance for TCGA data (A), and age versus −log(PSA) for the PSA data (B).
The estimated Spearman’s correlation is . Table 3 shows the results of hypothesis testing. Only the result of studentized permutation test is non-significant at and suggests there is no evidence of positive correlation between TP53 and FGFR1. In fact the biology does not support a positive correlation either, since FGFR1 mediates negative regulation of TP53 by FGF2 at transcriptional level (Katoh, 2016). Indeed, if we include all samples () from TCGA breast cancer cohort, then all tests will fail to reject with -values over 0.5 except for Fisher-Yates test. Together with the results from the simulations, the result by studentized permutation test is clearly more reliable.
Table 3:
Results of testing versus for TCGA breast cancer data and PSA data.
| Tests | p value (TCGA) | p value (PSA) |
|---|---|---|
| t test | 0.047 | < 0.001 |
| Fisher’s Z | < 0.001 | < 0.001 |
| Fisher-Yates | < 0.001 | < 0.001 |
| Asymp Norm | 0.039 | < 0.001 |
| Permute | 0.033 | < 0.001 |
| Stu Permute | 0.081 | < 0.001 |
4.2. PSA data
The testing methods were also applied to a data set of age and baseline prostate-specific antigen (PSA) levels (Sweeney et al., 2015). The data consists of age and PSA levels of 480 subjects, of which 473 have complete paired observations. The sample Spearman’s correlation coefficient between age and PSA is . Since the alternative hypothesis of proposed test is , we applied a negative log transformation on PSA levels. Similar as the TCGA example, the marginal normality of data was examined by Shapiro-Wilk test, and the bivariate normality was examined by Henze-Zikler test. The values of Shapiro-Wilk tests for log-transformed age and PSA levels are 0.0208 and 0.0301, respectively. The value of Henze-Zikler test is < 0.0001. The results indicates the distribution is not bivariate normal. Figure 3B shows the scatter plot of age versus −log(PSA). Table 3 shows that all tests reject the and conclude there is a non-zero correlation between age and PSA. The is an example where all tests have consistent results when there is a true correlation. Although the normality tests are significant, such deviation may have been remedied by using the ranks in this specific example.
5. Discussion
Conventional tests of the Spearman’s correlation rely on normality assumption, including test, Fisher’s transformation, and naive permutation test, which fails to control Type I error rates when the assumption is violated. This was illustrated in our simulations studies (Section 3). Such defect cannot be remedied by transforming the marginal distributions such as by Fisher-Yates coefficient. Notably, the deviation from bivariate normality can result in a convergence of type I error rate to an arbitrary level when . This indicates that, under scenarios when two random variables are uncorrelated but dependent, the type I error will not be controlled at desired level no matter how large the sample size is. On the other hand, Serfling’s test based on the delta method guarantees that the type I error rate converges to as long as the fourth order moment is finite. However, it typically suffers an inflated type I error when sample size is under 50. It should be noted that although we only examined Spearman’s correlation, a similar phenomenon is expected for other rank-based correlation as well, such as Kendall’s . The tests based on large sample theories are also anticipated to have inflated type I error when the sample size is small.
We present a robust Spearman’s correlation permutation test based on studentized statistic for testing versus . The proposed approach is inspired by the work by DiCiccio and Romano (DiCiccio and Romano, 2017), which was developed for Pearson’s correlation. Through extensive simulation studies and real world application, we show the proposed test controls type I error even when sample size is as small as 10 and normality assumption is violated. Therefore, the test is valid in testing monotonic correlations under more general scenarios of underlying distributions and sample sizes. In addition, the studentized statistic can also be used for bootstrapping tests, so as to test for more general point null hypotheses (Hutson, 2019). In conclusion, the proposed studentized permutation test should be used as a routine for testing non-zero Spearman’s correlation coefficient.
Acknowledgements
This work was supported by Roswell Park Cancer Institute and National Cancer Institute (NCI) grant P30CA016056, NRG Oncology Statistical and Data Management Center grant U10CA180822 and IOTN Moonshot grant U24CA232979-01. The results shown here are in part based upon data generated by the TCGA Research Network: https://www.cancer.gov/tcga. The PSA data example is based on research using information obtained from www.projectdatasphere.org, which is maintained by Project Data Sphere, LLC. Neither Project Data Sphere, LLC nor the owner(s) of any information from the website have contributed to, approved or are in any way responsible for the contents of this publication.
References
- Arboretti R, Bonnini S, Corain L, and Salmaso L (2014). A permutation approach for ranking of multivariate populations. Journal of Multivariate Analysis, 132:39–57. [Google Scholar]
- Arboretti Giancristofaro R, Bonnini S, and Pesarin F (2009). A permutation approach for testing heterogeneity in two-sample categorical variables. Statistics and Computing, 19(2):209–216. [Google Scholar]
- Basso D and Salmaso L (2011). A permutation test for umbrella alternatives. Statistics and Computing, 21(1):45–54. [Google Scholar]
- DiCiccio CJ and Romano JP. (2017). Robust permutation tests for correlation and regression coefficients. Journal of the American Statistical Association, 112(519):1211–1220. [Google Scholar]
- Edgell SE and Noon SM (1984). Effect of violation of normality on the t test of the correlation coefficient. Psychological bulletin, 95(3):576. [Google Scholar]
- Fieller EC, Hartley HO, and Pearson ES (1957). Tests for rank correlation coefficients. i. Biometrika, 44(3/4):470–481. [Google Scholar]
- Fisher RA and Yates F (1938). Statistical tables: For biological, agricultural and medical research. Oliver and Boyd. [Google Scholar]
- Giancristofaro RA (2014). Permutation solutions for multivariate ranking and testing with applications. Communications in Statistics-Theory and Methods, 43(4):891–905. [Google Scholar]
- Giancristofaro RA and Bonnini S (2008). Moment-based multivariate permutation tests for ordinal categorical data. Journal of Non-parametric Statistics, 20(5):383–393. [Google Scholar]
- Hutson AD (2019). A robust pearson correlation test for a general point null using a surrogate bootstrap distribution. Plos one, 14(5):e0216287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hutson AD and Yu H (2021). A robust permutation test for the concordance correlation coefficient. Pharmaceutical Statistics, 20(4):696–709. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Institute, S. (2015). Base SAS 9.4 procedures guide. SAS Institute. [Google Scholar]
- Katoh M (2016). Fgfr inhibitors: Effects on cancer cells, tumor microenvironment and whole-body homeostasis. International journal of molecular medicine, 38(1):3–15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kendall M and Stuart A (1979). The advanced theory of statistics, 2: 240–80. London: Charles Griffin. [Google Scholar]
- Li B and Dewey CN (2011). Rsem: accurate transcript quantification from rna-seq data with or without a reference genome. BMC bioinformatics, 12(1):323. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Marozzi M (2014). Testing for concordance between several criteria. Journal of Statistical Computation and Simulation, 84(9):1843–1850. [Google Scholar]
- Moran P (1948). Rank correlation and product-moment correlation. Biometrika, 35(1/2):203–206. [PubMed] [Google Scholar]
- Pauly M and Smaga L (2020). Asymptotic permutation tests for coefficients of variation and standardised means in general one-way anova models. Statistical Methods in Medical Research, pages 2733–2748. [DOI] [PubMed] [Google Scholar]
- Pearson K (1896). Vii. mathematical contributions to the theory of evolution.—iii. regression, heredity, and panmixia. Philosophical Transactions of the Royal Society of London. Series A, containing papers of a mathematical or physical character, (187):253–318. [Google Scholar]
- Pesarin F and Salmaso L (2010a). Finite-sample consistency of combination-based permutation tests with application to repeated measures designs. Journal of Non-parametric Statistics, 22(5):669–684. [Google Scholar]
- Pesarin F and Salmaso L (2010b). The permutation testing approach: a review. Statistica, 70(4):481–509. [Google Scholar]
- Pesarin F and Salmaso L (2012). A review and some new results on permutation testing for multivariate problems. Statistics and Computing, 22(2):639–646. [Google Scholar]
- Pesarin F and Salmaso L (2013). On the weak consistency of permutation tests. Communications in Statistics-Simulation and Computation, 42(6):1368–1379. [Google Scholar]
- Salmaso L (2015). Combination-based permutation tests: Equipower property and power behavior in presence of correlation. Communications in Statistics-Theory and Methods, 44(24):5225–5239. [Google Scholar]
- Salmaso L and Pesarin F (2010). Permutation tests for complex data: theory, applications and software. John Wiley & Sons. [Google Scholar]
- Spearman C (1961). The proof and measurement of association between two things. [DOI] [PubMed] [Google Scholar]
- Stanton JM (2001). Galton, pearson, and the peas: A brief history of linear regression for statistics instructors. Journal of Statistics Education, 9(3). [Google Scholar]
- Sweeney CJ, Chen Y-H, Carducci M, Liu G, Jarrard DF, Eisenberger M, Wong Y-N, Hahn N, Kohli M, Cooney MM, et al. (2015). Chemohormonal therapy in metastatic hormone-sensitive prostate cancer. New England Journal of Medicine, 373(8):737–746. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Team, R. C. et al. (2013). R: A language and environment for statistical computing. [Google Scholar]



