Summary
We propose a method to test the correlation of two random fields when they are both spatially auto-correlated. In this scenario, the assumption of independence for the pair of observations in the standard test does not hold, and as a result we reject in many cases where there is no effect (the precision of the null distribution is overestimated). Our method recovers the null distribution taking into account the autocorrelation. It uses Monte-Carlo methods, and focuses on permuting, and then smoothing and scaling one of the variables to destroy the correlation with the other, while maintaining at the same time the initial autocorrelation. With this simulation model, any test based on the independence of two (or more) random fields can be constructed. This research was motivated by a project in biodiversity and conservation in the Biology Department at Stanford University.
Keywords: Geostatistics, Monte-Carlo methods, Resampling, Spatial autocorrelation, Spatial statistics, Variogram
1. Motivation
Assessing significance of the correlation coefficient is not straightforward if the values of the variables involved vary smoothly with location (throughout the paper smooth refers to spatial autocorrelation, and it may be the case that the variable changes abruptly over a short distance). With spatially autocorrelated data, nearby points may provide almost identical information. Hence there is a tension between sample size, resolution and the number of independent measurements; i.e. at some level, more data, meaning sampling the process at a higher resolution, does not mean more information. As a result, classical tests Fisher (1915) tend to incorrectly reject the null (larger type II errors). Some work has been done, particularly in the field of Geostatistics, to overcome this problem. For instance, Clifford, Richardson, and Hemon (1989) propose a method that estimates an effective sample size M < N to be used in such tests, in an attempt to capture the real uncertainty. The correlation coefficient is thus evaluated with a Student’s t distribution with M degrees of freedom (distribution with larger variance), which accounts for the loss of precision due to the underlying spatial component. The method, however, is developed for Gaussian random fields and in reality smoothed processes tend to be non-Gaussian.
We propose a Monte-Carlo method to test the correlation of two random fields that takes account of the spatial autocorrelation. By
randomly permuting the values of one of the fields across space we eliminate the dependence between them, and
smoothing and scaling the permuted field we approximately recover, with the help of the variogram, the spatial structure suppressed in (1).
In the same spirit, Allard, Brix, and Chadoeuf (2001) propose a method that is based on random local rotations, but applied to the characteristics of the spatial structure of point processes, where the intensity (rate parameter) is assumed to be constant at small scales and varies at large scales, the opposite situation we encounter with random fields, were the autocorrelation usually fades away as the distance increases.
By repeating steps 1) and 2) many times, we can obtain approximate realizations of the null distribution of interest. In fact, with this simulation model, it is possible to examine the null distribution of a larger variety of statistics. In particular any test based on the independence of two (or more) random fields can be constructed, the simplest example being the test for a single Pearson’s correlation coefficient between the two random fields. We will call it global correlation.
Other spatially local tests may be of interest — for example, which regions show strong correlation between the two fields — the question that our collaborators posed (McCauley et al., 2012). They mapped the locations of sites over the world using two criteria:
amount of species richness — biodiversity, and
travel time in days needed to reach the nearest city — remoteness.
Figures 1(a) and 1(b) represent both these fields, and we see that they are spatially very smooth. Are remoteness and biodiversity correlated with one another? i.e. are there more species in remote areas that are better insulated from human disturbance? To succinctly communicate the strength of these correlations, the authors were interested in reporting a p-value map for the areas where overlap between remoteness and biodiversity occurs. Initially they used Geographically Weighted Regression methods (Fotheringham, Brunsdon, and Charlton, 2002), a set of regression techniques that tackle spatially varying relationships. This book has captured considerable attention in the Geostatistics community. However, these methods focus on comparing coefficients for different spatial areas, and identifying the areas with stronger relationships, but with no assessment to whether the coefficients in the model are significant or not.
Given the map of correlations in Figure 1(c), where each value corresponds to the correlation between biodiversity and remoteness in a given neighborhood (we will call them local correlations), we will apply our method and produce a map of p-values, where each p-value assesses significance of the correlation in that particular location. See Appendix 2 for details on how to calculate the local correlations.
We organize the paper as follows. In section 2 we show the results of applying our method to the biodiversity dataset. Section 3 illustrates the limitations of the standard test under spatial autocorrelation. Section 4 describes in detail the algorithm proposed in this paper, and in Section 5 we study the behavior of the method by performing power and type I error analyses, and compares it to the approach in Clifford et al. (1989). We conclude and summarize our findings in the discussion.
2. Biodiversity Data
Protecting remote ecosystems is the future of global diversity, WWF ecologists divided the world into 16 unique regions (Biomes) based on land cover and climate. In this paper, and for simplicity, we will focus on the part of the biodiversity dataset that corresponds to the American region of Biome 1 (Tropical and subtropical moist broadleaf forests) to illustrate our methodology.
Biodiversity (X) is the result of estimating the number of species of plants, amphibians, birds and mammals in an area of 100 × 100 km and centered at location s. The estimates for the 4 groups is normalized to a maximum score of 10, with X being the average of those normalized counts. Remoteness (Y) combines a number of data sets that influence speed of travel: road networks, angle of slope, density of vegetation, river courses, etc. It takes values between 1 – 8 and indicates the travel time in days needed to reach the nearest city larger than 50, 000 inhabitants from location s, where 8 represents any travel time larger than 7 days. Our sample is denoted by (Xs, Ys) = [(Xs1, Ys1), …, (XsN, YsN)], s = (s1 … sN), where si ∈ ℝ2 are the longitude and latitude coordinates of observation i and N = 19, 926.
2.1 The empirical and smoothed variogram for Biodiversity
The theoretical variogram is a function describing the degree of spatial dependence of a random field Xs. It is defined as the variance of the difference between field values at two locations si and sj across realizations of the field: , see Cressie (1993). The empirical variogram for the sample Xs1, …, XsN is the collection of pairs of distances uij = ‖si − sj‖ between si and sj, and their corresponding variogram ordinates . The empirical variogram for biodiversity is plotted in Figure 2(a). Since γ is expected to be a smooth function of distance, it is common to smooth the empirical variogram to improve its properties as an estimator for γ. In our algorithm, we will use the smoothed variogram γ̂, which is obtained using a kernel smoother and defined in Appendix 1. The black curve in Figure 2(a) is the smoothed variogram for biodiversity with bandwidth h = 0.746, and truncated at distance u = 10, which corresponds to the 25% percentile of the distribution of pairs of distances. We truncate the variogram because the precision of the estimate is expected to decrease as the distance increases, since a decreasing number of pairs are involved.
When pairs of distances are small, the variance of Xsi − Xsj is small, indicating that observations close to each other are very similar. As pair distances increase, the autocorrelation dies off and the variogram grows; as farther apart observations are less similar. The increasing nature of the variogram as the distance increases is a common behavior that corroborates the spatial autocorrelation present in the variable biodiversity.
2.2 Applying our Method to the Biodiversity Dataset
In this section we apply to the biodiversity dataset the algorithm that we describe in Section 4. Given Xs and Ys, we will test significance of 1) the global correlation coefficient [H0 : ρXs, Ys = 0], and 2) the set of local correlations plotted in Figure 1(c) [H0 : ρXs, Ys(sj) = 0, ∀j], and calculated using a Gaussian kernel with neighborhood window truncated at bandwidth λ = 5.281 (see Appendix 2).
We will see that the algorithm returns as a result of randomizing and smoothing Xs B = 1000 times, with bandwidths Δ ∈ (0.005, 0.785). These are proxies for Xs with approximately the same autocorrelation but with the characteristic of being independent of Ys. The pairs will be used to test the situations 1) and 2) above.
We chose to randomize Xs (biodiversity), as we are free to pick the most convenient one, since the purpose is to break the dependence between both variables.
2.2.1 Testing the Global Correlation
The observed global correlation coefficient between biodiversity and remoteness is γ̂Xs, Ys = 0.224. We use r = (r1, …, rB) as an estimate of the sampling distribution of rXs, Ys under the null hypothesis of independence, where ; see Figure 2(b). Using this null distribution to test H0 : ρXs, Ys = 0, the p-value is (indicated with a black line in Figure 2(b)). Had we assumed that the sample pairs were independent, and used instead a Student’s t with N − 2 degrees of freedom, the p-value would have been effectively zero, since the variance of the Student’s t distribution is significantly smaller.
2.2.2 Testing Local Correlations
Each pair of random fields , can be used to calculate a new map of local correlations under the null hypothesis ( and Ys are constructed to be independent). Hence we can compute local p-values in exactly the same was as was done globally above. A sample distribution of the statistic rXs, Ys(sj) under the null is r(sj) = (r1(sj), …, rB(sj)), where . The p-value for testing H0 : ρXs, Ys(sj) = 0 at location sj is . The resulting p-value map is plotted in Figure 3, where we can identify the regions that are strongly correlated. For comparison, we also plot the map of p-values had we used the standard test.
3. Behavior of rXs, Ys under Spatial Autocorrelation
If (x1, y1), …, (xN, yN) is an independent and normally distributed sample, the distribution for the Pearson’s correlation coefficient under ρX,Y = 0 is
(1) |
The statistic is used to test for H0 : ρX,Y = 0, which follows a Student’s t distribution with N − 2 degrees of freedom, see Kenney and Keeping (1951).
In the previous section we have seen that even though the sample size of the biodiversity dataset was very large (N = 19, 296), the sample distribution for rXs, Ys was somewhat wide (Figure 2(b)).
This emphasizes the following point, which we demonstrate via a simulation. For a pair of spatially correlated random fields, the sample size or, more precisely, the resolution at which the fields are sampled, can play less of a role in the behavior of the distribution of rXs, Ys; rather it is the extent of the spatial autocorrelation that determines this distribution (see Walther (1997) for an equivalent problem with time series).
Let Ws be a stationary and isotropic Gaussian random field, s ∈ ℝ2, with Matérn auto-correlation function φ(u) = [2κ−1Γ(κ)]−1(u/ϕ)κKκ(u/ϕ) where ϕ > 0 is the scale, and the shape parameter κ > 0 determines the smoothness of the process, var(Ws) = σ2. For κ = 0.5, the Matérn autocorrelation function reduces to the exponential, and when κ → ∞ to the Gaussian.
Suppose Xsi is generated by a stationary process
(2) |
with Zi being mutually independent and identically distributed with zero mean and nugget variance τ2 (measurement error), see Diggle and Ribeiro (2007). The theoretical variogram of Xsi under stationarity is described in Appendix 1 and illustrated in Web Figure 4.
Figures 4(a) and 4(b) are Xs and Ys, two independent realizations of this process in the grid [0, 1] × [0, 1] with resolution N = 101 × 101 = 10, 201, and parameters κ = 0.5 and ϕ = 0.30. We have used the R package RandomFields to simulate these processes (R Core Team, 2013; Schlather et al., 2013).
In Web Figure 1, Figure (a) is the scatter plot of Xs and Ys, whereas Figure (d) is the scatter plot of two independent samples, each of them mutually independent (non-spatially correlated) and normally distributed. The correlation coefficient is much larger for Xs and Ys (r̂Xs, Ys = 0.3). Figure (b) and (c) are two new sets of independent simulated processes Xs and Ys, showing negative strong correlation and a correlation closer to zero respectively. Thus, as we have seen, a consequence of the spatial autocorrelation is a larger variance for the distribution of rXs, Ys. In fact, the larger κ, the larger the variance. Intuitively, consider the extreme case of two very smooth one-dimensional fields (i.e lines) with randomly chosen orientations (slopes); the correlation of sampled pairs will either be −1 and +1 depending on the randomly chosen orientation; a maximal variance situation. The same intuition applies to higher-dimensional fields.
The null distribution for rXs, Ys under ρXs, Ys = 0 can be estimated by generating the independent pairs . With this distribution, the probability of observing values as extreme or more than 0.3 is 0.16, with no evidence against H0.
Although Xs and Ys have been constructed to be independent of each other, if we use the Student’s t distribution to test H0 : ρXs, Ys = 0, the p-value for r̂Xs, Ys = 0.3 is 0.
4. The Algorithm
We propose a method that approximately recovers the null distribution of rXs, Ys, or any other statistic based on the independence of Xs and Ys. The following scheme summarizes the main steps of the algorithm.
Let Xs and Ys be a realization of two random fields. Repeat the following two steps B times:
Randomly permute the values of Xs over s, which we denote by Xπ(s); this means Xπ(s) and Ys are independent.
Smooth and scale Xπ(s) to produce X̂s, such that its smoothed variogram γ̂ approximately matches γ̂(Xs); i.e. the transformed variable X̂s has approximately the same autocorrelation structure as Xs.
The random fields have approximately the same autocorrelation structure as Xs, but are each independent of Ys. Hence, the pairs are the ingredients for the calculation of a null distribution. For instance, the distribution of rXs, Ys under H0 : ρXs, Ys = 0 can be approximated by r = (r1, …, rB), where , but we could proceed equivalently with any other statistic based on the independence of Xs and Ys.
Note that we do not pose restrictions on which random field to permute, but the algorithm assumes that Xs is stationary.
Step (2) of the algorithm is described in detail in the following section.
4.1 Step 2: Matching Variograms
This step focuses on recovering the intrinsic spatial structure of Xs that was eliminated with the random permutation. As we have seen, the null distribution of rXs, Ys is mainly determined by the amount of autocorrelation, and this step will determine how well we are able to recover it. Our approach is non-parametric, which implies that the variogram matching does not rely on model assumptions, such as choosing a parametric model for the variogram.
Formally, the problem reduces to choosing a variogram from the family
(3) |
that best approximates γ̂(Xs), the smoothed variogram of Xs.
Let Xπ(s) be the permuted random field obtained in step (1) above. Choose Δ to be a set of values for the proportion of neighbors to consider for the smoothing step.
Calculate the smoothed variogram γ̂(Xs) by smoothing the empirical variogram of Xs (see Appendix 1).
- For each δ ∈ Δ repeat:
- Construct the smoothed variable using a kernel smoother that fits a constant regression to Xπ(s) at each location sj (see Appendix 2).
- Calculate as in (1).
- Fit a linear regression between and γ̂(Xs), where (α̂δ, β̂δ) are the least-squares estimates.
Choose δ* ∈ Δ such that the sum of squares of the residuals of the fit in (2) (c) is minimized.
Transform by , where Z is a vector of mutually independent and identically distributed Zi’s with zero mean and unit variance.
By varying the tuning parameter or proportion of neighbors δ in step (2) we obtain a family of variograms with different shapes, choosing the one more similar to the variogram of the original variable Xs.
The linear transformation in Step (4) ensures that the scale and intercept of match those of γ̂(Xs), since the smoothing in step (2)(a) has changed the scale of Xs (the smoother , the smaller the variance), in addition to the intercept (nugget variance) of γ̂(Xs).
The variogram of the random field is a member of the family in (3), and has been constructed to match γ̂(Xs) in shape, scale, and intercept.
Note that is an estimate of σ2 in (A.2), |α̂δ*|is an estimate of τ2, and correspondingly is an estimate of σ2 + τ2.
For an illustration of the variogram matching see Figure 5. Different δs result in different shapes for , and the best match between γ̂(Xs) and (grey) is reached with δ = 0.085. The residual sum of squares of linearly regressing on γ̂(Xs) for different values of δ are plotted in Web Figure 2.
5. Simulation Results
In this section we study the performance of our method for testing H0 : ρXs, Ys = 0 by comparing the estimate of the null distribution for rXs, Ys provided by our algorithm with an estimate of the true null distribution.
In addition to that, we carry out power and type I error analyses (Section 5.2), and compare the results to Clifford’s approach.
Let Xs and Ys be two independent Gaussian random fields in the grid [0, 1] × [0, 1] with resolution N = 101 × 101 = 10, 201, following model (2) with σ2 = 1, μ = 0, and:
Gaussian autocorrelation function φ(u) = exp[−(u/ϕ)2] with scale parameter ϕ = 0.3.
exponential autocorrelation function φ(u) = exp(−u/ϕ) with ϕ = 0.3.
We apply our algorithm to one realization of the pair (Xs, Ys) in situations (1) and (2), with B = 1000 and bandwidths Δgauss = (0.1, 0.2, …, 0.9) and Δexp = (0.07, 0.080, …, 0.18) respectively.
The null distributions that the algorithm returns are plotted respectively in Figures 4(c) and 4(d), together with the true nulls obtained by simulating 1000 times the pairs from the models above. The results are slightly better for (1), but we manage to recover the null in both situations.
5.1 Clifford’s Effective Sample Size
We briey review Clifford’s procedure prior to using it as a comparator. Clifford et al. (1989) propose to use fM (r) in (1) as the distribution of reference under the null. The variance of fM(r) is . Their suggestion is to estimate the effective sample size M by , as a result of equating to , the variance of the sample correlation. Hence, an estimate for the null distribution of rXs, Ys is fM̂(r).
How to Estimate
In Clifford et al. (1989)’s Appendix they prove that to the first order, and under the assumption of normality, where SXsYs is the sample covariance, and are the sample variances of Xs and Ys, var(SXsYs) = trace(ΣξsΣηs), Σξs = PΣXsP, Σηs = PΣYsP, ΣXs and ΣYs are the covariance matrices of Xs and Ys respectively, and 1 is a vector of 1’s of dimension N.
An estimate for var(SXsYs) is obtained by imposing a stratified structure on ΣXs and ΣYs. More precisely, they assume that the set of all ordered pairs of elements of s can be divided into strata S0, S1, S2, …, and that the covariances within strata are constant. Then, , where Nk is the number of pairs in stratum Sk and is an autocovariance estimate for stratum Sk. They use the sample variogram to choose the number of strata.
5.2 Power Analysis and Type I Error Estimates
In this section we estimate the power and type I error of the test for our procedure, and compare it to Clifford’s approach.
Web Figure 3 summarizes different scenarios for the null and alternative distributions, as a function of the effect ρ we would like to detect and the amount of spatial autocorrelation ϕ. The power to detect a fixed effect ρ decreases as the spatial autocorrelation increases. Equivalently, for a fixed ϕ the power decreases with the effect size. These data have been simulated assuming we know the truth, and indicates how well we can do in each situation. The grid size g is set to 0.05 (N = 441), but identical results were obtained for grid size g = 0.01 (N = 10, 201).
The simulation experiment goes as follows. For each combination of autocorrelation in ϕ ∈ (0.05, 0.1, 0.3) and resolution in g ∈ (0.05, 0.01), and for the power calculations, for each size effect in ρ ∈ (0.2, 0.5, 0.8), we simulate 10 independent pairs from the usual model in the grid [0, 1] × [0, 1], with grid size g and Gaussian autocorrelation function with scale parameter ϕ. Then we do the following:
apply our algorithm to each pair with B = 1000, Δϕ=0.3 = (0.1, 0.2, …, 0.9), Δϕ=0.1 = (0.03, 0.034, …, 0.074), and Δϕ=0.05 = (0.013, 0.014, … 0.027) respectively. Let (r1, …, r10) be the resulting 10 nulls.
apply Clifford’s method to each pair by generating B independent and normally distributed random vectors Xi and Yi of dimension M̂. Let be the resulting 10 nulls, where and .
In addition to that, are 10 nulls under the truth, by simulating B independent pairs from the same model, where and .
The type I error of the test should be equal to the significance level α = 0.05. We use the nulls , (r1, …, r10) and to estimate the type I errors associated to both methods and the truth. We draw 100 samples from under H0 : ρXs, Ys = 0, and use respectively ,rj and to assess significance of , for k = 1, …, 100. Out of the 100 samples, the proportion of times the p-values are smaller than α = 0.05 is an estimate of the type I error. We repeat the same process for 100 replicates, obtaining 100 type I error estimates. We then repeat the process again for each null j = 1, …, 10 and average the 10 × 100 type I error estimates, which are shown in Table 1. Estimates of the standard errors are obtained by computing the standard deviation of the 10 × 100 estimates, and appear in brackets in the table.
Table 1.
Type I Error Estimates | ||||
---|---|---|---|---|
ϕ = 0.05 | ϕ = 0.1 | ϕ = 0.3 | ||
grid size = 0.05 | 0.052 (0.024) | 0.046 (0.022) | 0.048 (0.024) | true |
0.059 (0.029) | 0.056 (0.035) | 0.051 (0.041) | ours | |
0.055 (0.027) | 0.055 (0.029) | 0.093 (0.045) | Clifford | |
grid size = 0.01 | 0.053 (0.027) | 0.049 (0.023) | 0.051 (0.022) | true |
0.075 (0.030) | 0.051 (0.027) | 0.050 (0.045) | ours | |
0.055 (0.027) | 0.058 (0.026) | 0.086 (0.044) | Clifford |
The power of the test is calculated using a similar sampling scheme. We generate pairs as before, but then transform with (leading to a population correlation of ρ). The r̂k’s are thus obtained under the alternative hypotheses H1 : ρXs, Ys = ρ. We then follow the same procedure as was done for the type-I errors, namely steps (1) and (2) above, leading to the power estimates in Table 2.
Table 2.
Power Estimates | |||||
---|---|---|---|---|---|
ϕ = 0.05 | ϕ = 0.1 | ϕ = 0.3 | |||
grid size = 0.05 | ρ = 0.2 | 0.921 (0.026) | 0.418 (0.059) | 0.103 (0.034) | true |
0.926 (0.031) | 0.444 (0.087) | 0.089 (0.042) | ours | ||
0.926 (0.030) | 0.446 (0.077) | 0.153 (0.045) | Clifford | ||
ρ = 0.5 | 1 (0) | 0.997 (0.005) | 0.426 (0.055) | true | |
1 (0) | 0.997 (0.005) | 0.420 (0.089) | ours | ||
1 (0) | 0.997 (0.005) | 0.558 (0.100) | Clifford | ||
ρ = 0.8 | 1 (0) | 1 (0) | 0.951 (0.021) | true | |
1 (0) | 1 (0) | 0.879 (0.127) | ours | ||
1 (0) | 1 (0) | 0.950 (0.045) | Clifford | ||
grid size = 0.01 | ρ = 0.2 | 0.915 (0.033) | 0.405 (0.052) | 0.104 (0.030) | true |
0.935 (0.031) | 0.404 (0.073) | 0.124 (0.069) | ours | ||
0.927 (0.030) | 0.415 (0.064) | 0.169 (0.046) | Clifford | ||
ρ = 0.5 | 1 (0) | 0.996 (0.007) | 0.424 (0.059) | true | |
1 (0) | 0.996 (0.007) | 0.391 (0.154) | ours | ||
1 (0) | 0.996 (0.007) | 0.510 (0.111) | Clifford | ||
ρ = 0.8 | 1 (0) | 1 (0) | 0.937 (0.022) | true | |
1 (0) | 1 (0) | 0.949 (0.030) | ours | ||
1 (0) | 1 (0) | 0.972 (0.018) | Clifford |
We highlight the values in Table 1 that most deviate from α. Type I error estimates for our method are close to α, which is less of a case for Clifford’s method, except for ϕ = 0.05.
Note also in Table 2 that power results for our method are typically closer to the power obtained under the truth, and that better power results in Clifford’s case go along with larger type I error estimates, these values are highlighted in the table.
Finally, observe that there are no obvious differences between both resolutions, where the grid sizes g = 0.01 and g = 0.05 correspond to N = 101 × 101 = 10201 and N = 21 × 21 = 441 respectively. This is important, as it highlights the fact that the power of the analysis is mainly driven by the autocorrelation, and even if we increase significantly the resolution, the power stays the same as it is constrained by ϕ.
6. Discussion
This papers addresses the consequences of spatial autocorrelation on the distribution of statistics like correlation between random fields. We develop a nonparametric approach for sampling from the null-hypothesis of independence, that involves three steps:
pick one of the fields and estimate the spatial autocorrelation structure via its variogram
randomly permute the values in this field
apply a local smoothing to the permuted values, using a bandwidth and rescaling so that its resulting variogram matches the original in step (1).
Steps (2) and (3) are repeated B times to obtain B realizations from the null. These realizations can be used to measure many different kinds of independence, including measures of local dependence.
A referee of an earlier draft of this paper made a useful suggestion for an alternative and more direct approach. Their suggestion was to simulate a Gaussian random field according to a model of the variogram, with parameters estimated by weighted least square, maximum likelihood or composite likelihood, with non-Gaussian data transformed to normal scores prior to the estimation of the variogram. One of the advantages of our method, though, is that it does not rely on model assumptions, which turns it into a more flexible, non-parametric approach.
Allard et al. (2001) address a related problem in the context of pairs of point processes, where the focus is on local correlations between pairs of realizations of counts. They randomize one of the processes using local rotations, under the assumption that locally the process rates are constant. Our randomization scheme treats broader range correlations, so the locally constant assumption is not appropriate.
Another approach is proposed by Clifford et al. (1989), where they estimate an effective sample size that takes the autocorrelation into account. We compare this approach to our method through simulations, which show that our algorithm behaves well in practice, with type I error estimates close to α, and power estimates close to the maximum power for each combination of ϕ and ρ.
Since correlation may exist simultaneously at a number of different geographical scales, the method can be used to calculate a p-value for the global correlation, as well as p-values for the local correlations (p-values map), summarizing the strength of the relationship between both fields in a particular location.
One of the important consequences of autocorrelation is that increasing the sample size does not necessarily increase the power to find significance. There is no concept of sample size, since what we observe is one realization of the random field, and the amount of data that we have is the resolution. The information that the sample provides is limited by the spatial autocorrelation. Consequently, in practice it may be more important to focus on using methods that adjust for autocorrelation, than on collecting more data.
Supplementary Material
Acknowledgements
We thank Paul Switzer for some suggestions early on in this project. Trevor Hastie and Rahul Mazumder were partially supported by grant DMS-1007719 from the National Science Foundation (NSF), and grant RO1-EB001988-15 from the National Institutes of Health. Douglas McCauley was supported by NSF grants DEB-0909670 and GRFP-2006040852 as well funding from the Wood’s Institute for the environment. Julia Viladomat was supported by the Spanish Ministry of Education through the program ‘Programa Nacional de Movilidad de Recursos Humanos del Plan Nacional de I-D+i 2008–2011’.
Appendix 1
The smoothed variogram γ̂ for Xs is defined by smoothing the empirical variogram using a kernel smoother:
(A.1) |
It assigns weights that die off smoothly as distance to u0 increases, with wij = Kh(‖u0−uij‖), uij = ‖si−sj‖, and , the Gaussian kernel is scaled so that their quartiles are at ±0.25h, with h being the bandwidth (R function ksmooth). The variogram γ̂ is obtained by evaluating (A.1) at distances u = (u1, …, u100) uniformly chosen within the range of pair distances uij.
The theoretical variogram of a stationary random field Xsi = Wsi + Zi in (2) is:
(A.2) |
where φ(u) is the autocorrelation function of Wsi, typically a monotone decreasing function with φ(0) = 1 and φ(u) → 0 as u → ∞. Its most important feature is its behavior near u = 0, and how quickly it approaches zero when u increases, which reflects the physical extent of the spatial autocorrelation in the process. When φ(u) = 0 for u greater than some finite value, this value is known as the range of the variogram. The intercept τ2 corresponds to the nugget variance, the conditional variance of each measured value Xsi given the underlying signal value Wsi. The sill is the asymptote τ2 + σ2 and corresponds to the variance of the observation process Xsi. Web Figure 4 gives a schematic illustration.
Appendix 2
We smooth a given random field Xs by fitting a local constant regression at locations s1, …, sN using the R package locfit (Loader, 2013):
(A.3) |
where wsi = Kλs(‖s − si‖), is a Gaussian kernel, and λs controls the smoothness of the fit. For a fitting point s, λs is such that the neighborhood contains the k = ⌊Nδ⌋ nearest points to s in Euclidean distance, where δ ∈ (0, 1) is a tuning parameter that indicates the proportion of neighbors. Each f̂δ(s) is estimated with the k observations that fall within the ball Bλs(s) centered at s and of radius λs (the kernel truncates at one standard deviation). We use a varying λs because it reduces data sparsity problems by increasing the radius in regions with fewer observations.
The smoothed variable in Section 4.1 in step (2)(a) of the algorithm is the result of fitting the function (A.3) to Xπ(s).
The local correlations in Figure 1(c) are calculated using this same approach. The local correlation at location s is defined as:
(A.4) |
where and . We then compute (A.4) by breaking it down and separately evaluating the quantities ΣwsiXsi, ΣwsiYsi, ΣwsiXsiYsi, and using locfit as above.
Footnotes
Supplementary Materials
Web Figures referenced in Sections 3, 4.1, 5.2 and Appendix 1 are available with this article at the Biometrics website on Wiley Online Library.
The prototype R code is now available online at http://www.stanford.edu/~hastie/Papers/biodiversity/viladomat_code.R and at the Biometrics website in Wiley Online Library.
References
- Allard D, Brix A, Chadoeuf J. Testing Local Independence between Two Point Processes. Biometrics. 2001;57:508–517. doi: 10.1111/j.0006-341x.2001.00508.x. [DOI] [PubMed] [Google Scholar]
- Clifford P, Richardson S, Hemon D. Assessing the Significance of the Correlation Between Two Spatial Processes. Biometrics. 1989;45:123–134. [PubMed] [Google Scholar]
- Cressie NA. Statistics for Spatial Data. John Wiley; 1993. [Google Scholar]
- Diggle PJ, Ribeiro PJ. Model-based Geostatistics. Springer; 2007. [Google Scholar]
- Fisher RA. Frequency distribution of the values of the correlation coefficient in samples from an indefinitely large population. Biometrika. 1915;10:507–521. [Google Scholar]
- Fotheringham AS, Brunsdon C, Charlton M. Geographically weighted regression: the analysis of spatially varying relationships. John Wiley; 2002. [Google Scholar]
- Kenney JF, Keeping ES. Mathematics of Statistics (Part Two) New York: D. Van Nostrand Cmpany, Inc.; 1951. [Google Scholar]
- Loader C. locfit: Local Regression, Likelihood and Density Estimation. R package version 1.5–9.1. 2013 [Google Scholar]
- McCauley DJ, McInturff A, Nuñez TA, Young HS, Viladomat J, Mazumder R, et al. Nature’s last stand: Identifying the world's most remote and biodiverse ecosystems. In review. 2012 [Google Scholar]
- R Core Team. R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing; 2013. [Google Scholar]
- Schlather M, Menck P, Singleton R, Pfaff B, team RC. RandomFields: Simulation and Analysis of Random Fields. R package version 2.0.66. 2013 [Google Scholar]
- Walther G. Absence of Correlation between the Solar Neutrino Flux and the Sunspot Number. Physical Review Letters. 1997;79:4522–4524. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.