Abstract
Analysis of geospatial data has traditionally been model-based, with a mean model, customarily specified as a linear regression on the covariates, and a Gaussian process covariance model, encoding the spatial dependence. While nonlinear machine learning algorithms like neural networks are increasingly being used for spatial analysis, current approaches depart from the model-based setup and cannot explicitly incorporate spatial covariance. We propose NN-GLS, embedding neural networks directly within the traditional Gaussian process (GP) geostatistical model to accommodate non-linear mean functions while retaining all other advantages of GP, like explicit modeling of the spatial covariance and predicting at new locations via kriging. In NN-GLS, estimation of the neural network parameters for the non-linear mean of the Gaussian Process explicitly accounts for the spatial covariance through use of the generalized least squares (GLS) loss, thus extending the linear case. We show that NN-GLS admits a representation as a special type of graph neural network (GNN). This connection facilitates the use of standard neural network computational techniques for irregular geospatial data, enabling novel and scalable mini-batching, backpropagation, and kriging schemes. We provide methodology to obtain uncertainty bounds for estimation and predictions from NN-GLS. Theoretically, we show that NN-GLS will be consistent for irregularly observed spatially correlated data processes. We also provide a finite sample concentration rate, which quantifies the need to accurately model the spatial covariance in neural networks for dependent data. To our knowledge, these are the first large-sample results for any neural network algorithm for irregular spatial data. We demonstrate the methodology through numerous simulations and an application to air pollution modeling. We develop a software implementation of NN-GLS in the Python package geospaNN.
Keywords: geostatistics, Gaussian process, neural networks, graph neural networks, machine learning, kriging, consistency
1. Introduction
Geostatistics, the analysis of geocoded data, is traditionally based on stochastic process models, which offer a coherent way to model data at any finite collection of locations while ensuring the generalizability of inference to the entire region. Gaussian processes (GP), with a mean function capturing effects of covariates and a covariance function encoding the spatial dependence, is a staple for geostatistical analysis (Stein, 1999; Banerjee et al., 2014; Cressie and Wikle, 2015). The mean function of a Gaussian process is often modeled as a linear regression on the covariates. The growing popularity and accessibility of machine learning algorithms such as random forest (Breiman, 2001), gradient boosting (Freund et al., 1999), and neural networks (Goodfellow et al., 2016), capable of modeling complex non-linear relationships has heralded a paradigm shift. Practitioners are increasingly shunning models with parametric assumptions like linearity in favor of machine-learning approaches that can capture non-linearity and high-order interactions. In particular, deep neural networks have seen considerable recent adoption and adaptation for geospatial data (see Wikle and Zammit-Mangion, 2023, for a comprehensive review).
Many of the machine-learning based regression approaches assume independent observations, implicit in the choice of an additive loss (e.g., ordinary least squares or OLS loss) as the objective function used in estimating the algorithm parameters. Explicit encoding of spatial dependency via a covariance function, as is common in process-based geospatial models, is challenging within these algorithms. Current renditions of neural networks for spatial data circumvent this by using spatial co-ordinates or some transformations (like distances or basis functions) as additional covariates (Gray et al., 2022; Chen et al., 2024; Wang et al., 2019). These added-spatial-features neural networks incorporate all spatial information into the mean, thus assuming that the error terms are independent. Consequently, they leave the GP model framework, abandoning its many advantages. They can suffer from the curse of dimensionality on account of adding many basis functions and cannot provide a separate estimate of the covariate effect (see Section 2.2).
We propose a novel algorithm to estimate non-linear means with neural networks within the traditional Gaussian process models while explicitly accounting for the spatial dependence encoded in the GP covariance matrix. The core motivation comes from the extension of the ordinary least squares (OLS) to generalized least squares (GLS) for linear models with dependent errors. For correlated data, GLS is more efficient than OLS according to the Gauss-Markov theorem. For a neural network embedded as the mean of a GP, the GLS loss naturally arises as the negative log-likelihood. We thus refer to our algorithm as NN-GLS. We retain all advantages of the model-based framework, including separation of the covariate and spatial effects thereby allowing inference on both, parsimonious modeling of the spatial effect through the GP covariance function circumventing the need to create and curate spatial features, and seamlessly offering predictions at new locations via kriging. NN-GLS is compatible with any neural network architecture for the mean function and with any family of GP covariance functions.
We note that the philosophy of GLS has been adopted for non-linear spatial analysis before. Nandy et al. (2017) propose GAM-GLS, using GLS loss based on Gaussian Process covariance for estimating parameters of generalized additive models (GAM). GAM-GLS improves over GAM for non-linear additive function estimation by explicitly accounting for the spatial covariance. However, like GAM, GAM-GLS does not account for interactions among covariates. Recently, tree-based machine learning algorithms like Random Forests (RF-GLS, Saha et al., 2023) and boosted trees (GP-boost, Sigrist, 2022; Iranzad et al., 2022) has been extended to use GLS loss. Forest and tree estimators use a brute force search to iteratively grow the regression trees, requiring multiple evaluations of the GLS loss within each step. This severely reduces the scalability of these approaches. RF-GLS also requires pre-estimation of the spatial parameters, which are then kept fixed during the random forest estimation.
NN-GLS avoids both the issues of lack of scalability and prefixed spatial covariance parameters of existing GLS-based non-linear methods by offering a representation of the algorithm as a special type of graph neural network (GNN). We show that NN-GLS using any neural network architecture for the mean and a Nearest Neighbor Gaussian Process (NNGP, Datta et al., 2016) for the covariance is a GNN with two graph-convolution layers based on the nearest-neighbor graph where the graph-convolution weights are naturally derived as kriging weights. Leveraging this representation of the model as a GNN, we can exploit the various computing techniques used to expedite neural networks. This includes using mini-batching or stochastic gradients to run each iteration of the estimation on only a subset of the data and representing kriging predictions using GNN convolution and deconvolution. The spatial covariance parameters are now simply weights parameters in the network and are updated throughout the training. Finally, we provide a spatial bootstrap-based approach to construct interval estimates for the non-linear mean function.
We provide a comprehensive theoretical study of neural networks when the observations have spatially correlated errors arising from a stochastic process. Our theoretical contribution distinguishes itself from the current body of theoretical work on neural networks in two main ways. The existing asymptotic theory of neural networks (see Section S3 for a brief review) does not consider spatial dependence in either the error process generating the data (often being restricted to i.i.d. errors) or introduce any procedural modifications to the neural network algorithm to explicitly encode spatial covariance.
Our theory of NN-GLS accounts for spatial dependence in both the data generation and the estimation algorithm. We prove general results on the existence and the consistency of NN-GLS which subsumes special cases of interest including irregular spatial data designs, popular GP models like the Matérn Gaussian process, and using NNGP for the GLS loss. We also provide finite-sample error rates for the mean function estimation from NN-GLS. The error bound provides novel insights on the necessity of modeling the spatial covariance in neural networks, showing that the use of OLS loss in NN, ignoring spatial dependence, will lead to larger error rates. To our knowledge, these are the first theoretical results for neural networks for irregularly sampled, spatially dependent data.
The rest of the manuscript is organized as follows. In Section 2, we review process-based geostatistical models and existing spatial neural networks. Section 3 describes the idea of combining the two perspectives. Section 4 formally proposes the algorithm NN-GLS, depicts its connection with graph neural networks (GNN), and provides scalable estimation, prediction and inference algorithms. Section 5 presents the theoretical results. Section 6 and 7 respectively demonstrate the performance of NN-GLS in simulated and real data.
2. Preliminaries
2.1. Process-based geostatistical modeling
Consider spatial data collected at locations , , comprising a covariate vector and a response . Defining and the covariate matrix similarly, the spatial linear model is given by
| (1) |
where is a covariance matrix. A central objective in spatial analysis is to extrapolate inference beyond just the data locations to the entire continuous spatial domain. Stochastic processes are natural candidates for such domain-wide models. In particular, (1) can be viewed as a finite sample realization of a Gaussian process (GP)
| (2) |
where is a zero-mean Gaussian process modeling the spatial dependence via the covariance function such that . often, can be decomposed into a latent spatial GP and a non-spatial (random noise) process. This results in the variance decomposition where is the covariance matrix corresponding to the latent spatial GP and is the variance for the noise process. One can impose plausible modeling assumptions on the nature of spatial dependence like stationarity () or isotropy () to induce parsimony, thereby requiring a very-low dimensional parameter to specify the covariance function.
Nearest Neighbor Gaussian Processes (NNGP, Datta et al., 2016) are used to model the spatial error process when the number of spatial locations is large as evaluation of the traditional GP likelihood (1), requiring computations, becomes infeasible. NNGP provides a sparse approximation to the dense full GP precision matrix . This is constructed using a directed acyclic graph (DAG) based on pairwise distances among the data locations, such that each node (location) has almost directed (nearest) neighbors. Letting be the set of neighbors of location in the DAG, we have
| (3) |
Here is a strictly lower triangular matrix and is a diagonal matrix. NNGP precision matrices only require inversion of small matrices of size . As , evaluation of NNGP likelihood requires total time and storage.
2.2. Neural networks
Artificial neural networks (ANN) or, simply, neural networks (NN) are widely used to model non-parametric regression where . Mathematically, an -layer feed-forward neural networks or multi-layer perceptron (MLP) can be described as,
| (4) |
where for layer , represents the nodes, are the weight matrix, ’s are a linear combination of nodes with unknown weights ’s, and denotes the known non-linear activation (link) functions (e.g., sigmoid function, ReLU function) applied to each component of . The final layer is called the output layer and gives the modeled mean of the response, i.e., . For regression, the unknown weights are estimated using backpropagation based on the ordinary least squares (OLS) loss
| (5) |
Estimation is expedited by mini-batching where the data are split into smaller and disjoint mini-batches and at each iteration the loss (5) is approximated by restricting to one of the mini-batches, and cycling among the mini-batches over iterations.
Current extensions of neural networks for spatial data have mostly adopted the added-spatiai-features strategy, adding spatial features like the (geographical coordinates, spatial distance or basis functions) as extra covariates in the neural network (Wang et al., 2019; Chen et al., 2024). Formally, they model where denotes a set of spatial basis functions, and denotes a neural network on the joint space of the covariates and the basis functions . These methods thus depart the GP model framework, and do not explicitly model spatial covariance as all the spatial structure is put into the mean, implicitly assuming that there is no residual spatial dependence. They cannot provide a separate estimate of the covariate effect as it models jointly as a function of and . Also, the number of added basis functions carries a tradeoff, with more basis functions improving accuracy at the cost of increasing parameter dimensionality.
3. Neural networks for Gaussian process models
The two paradigms reviewed in Section 2 are complimentary in their scope. The popularity of the geospatial linear models (Section 2.1) is owed to their simplicity, interpretability, and parsimony – separating the covariate effect and the spatial effects, modeling the former through a linear mean, and the latter parsimoniously through the GP covariance, specified typically using only 2-3 interpretable parameters by encoding stationarity or isotropy. However, it relies on the strong assumption of a linear covariate effect. Neural networks can estimate arbitrary non-linear covariate effects. However, implicit in the usage of the OLS loss (5) for NN is the assumption that the data units are independent. This is violated for spatial data, where the error process is a dependent stochastic process as in (2). To our knowledge, there has been no previous work in directly incorporating spatial covariances in the neural network estimation process, with existing spatial neural networks mostly adopting the added-spatial-features strategy reviewed in Section 2.2.
We bridge the paradigms of traditional geospatial modeling and neural networks by embedding neural networks directly into the GP model, allowing the mean of the GP to be non-linear and modeling it with NN. We propose estimating the NN parameters using a novel GLS loss with the GP covariance matrix, which arises naturally from the log-likelihood of the GP model. We retain all advantages of GP based modeling framework, including separating the covariate and spatial effects into the mean and covariance structures, and obtaining predictions at new locations simply through kriging. In Section 4, we show that when using Nearest neighbor Gaussian processes for the spatial covariance, one can make a novel and principled connection between graph neural networks and geospatial models for irregular data, facilitating a scalable algorithm for estimation and prediction.
3.1. NN-GLS: Neural networks using GLS loss
We extend (2) to
| (6) |
Estimating in (6) using neural networks (4) needs to account for the spatial dependence modeled via the GP covariance. Using the OLS loss (5) ignores this covariance. We now extend NN to explicitly accommodate the spatial covariance in its estimation. Let . Then the data likelihood for the model (6) is
| (7) |
This is a non-linear generalization of the spatial linear model (1). For parameter estimation in linear models like (1) for dependent data, the OLS loss is replaced by a generalized least squares (GLS) loss using the covariance matrix , as it is more efficient according to the Gauss-Markov theorem. Similarly, to estimate using NN, we propose using the GLS loss
| (8) |
which accounts for the spatial dependency via the working precision matrix , which equals or, more practically, an estimate of it.
We refer to the neural network estimation using the GLS loss (8) as NN-GLS. Conceptually, generalizing NN to NN-GLS is well-principled as minimizing the GLS loss (8) with is equivalent to obtaining a maximum likelihood estimate of in (7). In practice, however, for spatial dependence modeled using GP, the GLS loss ushers in multiple computational issues for both mini-batching and backpropagation, techniques fundamental to the success of NN. As the GLS loss (8) is not additive over the data units, minibatching cannot be deployed as for the OLS loss, and back-propagation will involve computing an inverse of the dense matrix which requires storage and time for each iteration. These computing needs are infeasible for even moderate . Next, we develop an algorithm NN-GLS with a specific class of GLS loss that mitigates these issues and offers a pragmatic approach to using NN for GP models.
4. NN-GLS as Graph Neural Network
We offer a representation of NN-GLS as a special graph neural network (GNN). This connection allows use of OLS loss with transformed (graph-convoluted) data, leading to scalable mini-batching and backpropagation algorithms for NN-GLS. We propose choosing as the precision matrix from a Nearest Neighbor Gaussian Process (see Section 2.1), i.e., we optimize the objective function (8) with , as provided in (3). Basis functions derived from NNGP have been used as added-spatial-features in neural networks (Wang et al., 2019). This differs from our approach of using NNGP to directly model the spatial covariance, which is akin to the practice in spatial linear models.
A GLS loss can be viewed as an OLS loss for the decorrelated response , where is the Cholesky factor of . Hence, decorrelation is simply a linear operation. A convenience of choosing , the NNGP precision matrix, is that decorrelation becomes a convolution in the nearest neighbor DAG. To elucidate, note that defined in (3) denotes the kriging weights for predicting , based on its directed nearest neighbors using a GP with covariance . Similarly, in (3) is the corresponding nearest neighbor kriging variance. Letting denote the graph neighborhood for the node and defining weights
| (9) |
we can write . In this paper, index sets or are used in the subscript to subset a matrix or a vector. Thus, the decorrelated responses are simply convolution over the DAG used in NNGP with the graph convolution weights defined using kriging. Similarly, one can define the decorrelated output layer using the same graph convolution, where is the output layer of the neural network (see (4)).
The decorrelation step makes NN-GLS a special type of graph neural network (GNN) as depicted in Figure 1. In GNN, typically, the input observations are represented on a graph and the locality information is aggregated using convolution layers based on the graph structure (graph convolution). For NN-GLS, both the inputs and the responses are graph-valued objects as they both correspond to the locations , which are the nodes of the nearest neighbor DAG. First, the input is passed through the feed-forward NN (or multi-layer perceptron) to produce the respective output layer of . This is a within-node operation, and any architecture can be used (number of layers, number of nodes within each layer, sparsity of connections, choice of activation functions). Subsequently, the output layer from the MLP is passed through an additional graph-convolution to create the decorrelated output layer of using the weights (9). This layer is matched, using the OLS loss, to the decorrelated response layer of , created from the using the same graph convolution. So the objective function can be expressed as
| (10) |
Fig. 1.
NN-GLS as a graph neural network with two graph convolution layers
Thus fitting a GLS loss to an NN is simply fitting an OLS loss to a new NN with two additional decorrelation layers at the end of the MLP. From the form of in (9) and of and in (3), it is clear that can be calculated using matrices of dimension at −1 most , The NNGP precision matrix in (3) or the covariance matrix is never actually computed. Thus, the loss function (10) is evaluated without any large matrix computation.
To summarize, there are two ingredients of NN-GLS: a feed-forward NN or MLP (using any architecture) for intra-node operations; and a sparse DAG among the locations for incorporation of spatial correlation via inter-node graph convolutions. Information about the mutual distances among the irregular set of locations is naturally incorporated in the kriging-based convolution weights. We now discuss how this formulation of NN-GLS as a GNN with OLS loss helps leverage the traditional strategies to scale computing for NN.
4.1. Mini-batching
Leveraging the additivity of the OLS loss (10), we can write where being a partition of the data-locations each of size . The are uncorrelated and identically distributed (exactly under NNGP distribution, and approximately if the true distribution is a full GP), so the loss corresponding to the mini-batch are approximately i.i.d. for . Hence, parameter estimation via mini-batching can proceed like the i.i.d. case. The only additional computation we incur is during the graph convolution, as obtaining all for the mini-batch involves calculating for the neighbors of all units included in .
4.2. Back-propagation
Gradient descent or back-propagation steps for NN-GLS can be obtained without any large matrix inversions. We provide the back-propagation equations for a single-layer network with , where and are dimensional, being the number of hidden nodes, is the known activation (link) function. Here is the hidden layer created using weight-matrix and input vector . , are concatenated into vectors , , and , , are concatenated into matrices , , . With defined in (9), we introduce scalars and . Using the loss function (10) we derive the following customized back-propagation updates:
| (11) |
where are parameters for the graph convolution weights . Here is the learning rate, and is the mini-batch for the iteration and is the derivative of . Similar equations can be established for networks with more than one-layer, although the derivations will be tedious. Instead, for multi-layer networks, scalable gradient descent steps for NN-GLS can be conducted using off-the-shelf software. One can evaluate the mini-batch loss scalably, as the can be computed without any large matrix computation, and obtain scalable gradient descent updates using numerical, automatic, or symbolic differentiation.
Note that in this GNN representation, the convolution weights are parametrized by spatial covariance parameter (an example of is the one for Matérn covariance family (Definition S4.2) where ). Hence, the updates for can be just absorbed as a back-propagation step, as shown in (11). Alternatively, the spatial parameters can also be updated, given the current estimate of , by writing and to note the dependence on and maximizing the NNGP log-likelihood
| (12) |
We only use full optimization of (12) to update the spatial parameters , replacing its gradient descent update. The network parameters are always updated using gradient descent.
4.3. Kriging
Subsequent to estimating , NN-GLS leverages the GP model-based formulation to provide both point and interval predictions at a new location simply via kriging. Exploiting the GNN representation of NN-GLS with a NNGP working covariance matrix, kriging can be conducted entirely with the neural architecture, as we show below.
Given a new location and covariates , first is passed through the trained feed-forward part (MLP) of the GNN to obtain the output (see Figure 1). Next, the new location is added as a new node to the nearest-neighbor DAG. Let be its set of neighbors of on the DAG, and define the graph weights similar to (9). Then, using the graph convolution step, we obtain the decorrelated output . As the decorrelated output layer is the model for the decorrelated response layer in the GNN, we have . Finally, as is the graph-convoluted version of , we need to deconvolute over the DAG to obtain . This leads to the final prediction equation
| (13) |
It is easy to verify that the prediction (13) is exactly the same as the -nearest neighbor kriging predictor for the spatial non-linear model (7), i.e.,
| (14) |
Additionally, prediction variance is simply . As NN-GLS is embedded within the Gaussian Process framework, predictive distributions conditional on the data and parameters are Gaussian. So the prediction interval (PI) can be obtained as , where is the -quantile of a standard normal distribution.
Thus, the GNN connection for NN-GLS offers a simple and coherent way to obtain kriging predictions entirely within the neural architecture. We summarize the estimation and prediction from NN-GLS in Algorithm S1 in Section S1 of the Supplement. Section S2 presents a spatial bootstrap for obtaining pointwise confidence intervals of the mean.
5. Theory
We review some relevant theory of neural network regression in Section S3. Our theoretical contributions have two main differences from the existing theory. First, the overwhelming majority of the theoretical developments focus on the setting of i.i.d. data units. To our knowledge, asymptotic theory for neural networks under spatial dependence among observations from an irregular set of locations has not been developed.
Second, the existing theory has almost exclusively considered neural network methods like the vanilla NN (using the OLS loss) that do not explicitly accommodate spatial covariance among the data units. We provide general results on existence (Theorem 1) and asymptotic consistency (Theorem 2) of NN-GLS that explicitly encodes the spatial covariance via the GLS loss. In Propositions 1 and 2, we show that consistency of NN-GLS holds for spatial data generated from the popular Matérn GP under arbitrarily irregular designs. We also derive the finite-sample error rate of NN-GLS (Theorem 3) which shows how a poor choice of the working covariance matrix in the GLS loss will lead to large error rates, highlighting the necessity to accurately model spatial covariance in neural networks.
5.1. Notations and Assumptions
Let and denote the set of real numbers and natural numbers respectively. denotes the norm for vectors or matrices, . Given the covariates , for any function , we define the norm . Given a matrix denotes its eigenspace, , and . A sequence of numbers is if the sequence is bounded from above (goes to zero) as . Random variables (distributions) means and have the same distribution. We first specify the assumption of the data generation process.
Assumption 1 (Data generation process). The data is generated from a GP with a non-linear mean, i.e., , where is a continuous function, are fixed covariate vectors in a compact subset in , and the error process is a GP such that the maximum (minimum) eigenvalue of the covariance matrix is uniformly (in ) upper-bounded (lower-bounded) by .
Assumption 1 imposes minimal restrictions on the data generation process. The mean is allowed to be any continuous function. The restriction on the eigenvalues of the GP covariance matrix is tied to the spatial design. We show in Propositions 1 and 2 that this is satisfied for common GP covariance choices for any irregular set of locations in under the increasing domain asymptotics where locations are separated by a minimum distance.
We next state assumptions on the analysis model, i.e., the neural network family and the GLS working precision matrix. We consider a one-layer neural network class:
| (15) |
This formulation is equivalent to setting , as the sigmoid function and as the identity function in (4). It is easy to see that can control the approximation error, as one-layer NN are universal approximators. However, this class of functions can be too rich to control the estimation error. A common strategy to circumvent this is to construct a sequence of increasing function classes, also known as seive, to approximate , i.e.,
With careful tradeoff of the complexity of the function classes, it’s possible to control the estimation error (in terms of the covering number of ) using some suitable uniform-law-of-large-number (ULLN) while still being able to keep the approximation error in check. Following Shen et al. (2023) we consider the seive given below.
Assumption 2 (Function class). The mean function is modeled to be in the NN class
| (16) |
where as and satisfies the scaling
| (17) |
Here the activation function is a Lipschitz function on with range and Lipschitz constant . (For the sigmoid function, and ).
Hornik et al. (1989) has shown that is dense in the continuous function class which will control the approximation error. The estimation error will depend on the covering number for this class which can be controlled under the scaling rate (17).
Finally, to guarantee the regularity of the GLS loss (8) used for estimating the NN function , we require conditions on the working precision matrix . Instead of directly imposing these conditions on , we look at the discrepancy matrix
| (18) |
which measures the discrepancy between the true covariance matrix and the working covariance matrix . This is because we will see later in Theorem 3 that the finite sample error rates of NN-GLS will depend on through the spectral interval of . First, without loss of generality, we can assume . This is because the optimization in NN-GLS only depends on up to a scalar multiplier. The following assumption states that additionally the spectral interval of needs to be uniformly bounded in .
Assumption 3 (Spectral interval of the discrepancy matrix.). For all , all eigenvalues of lie in () for universal constants .
Uniform upper and lower bounds of are used for ensuring the continuity of the GLS loss function and the consistency of the loss function (Lemma S2) using empirical process results. We show in Propositions 1 and 2 how this Assumption is satisfied when is either the true GP precision matrix , its NNGP approximation. Of course, , which can be viewed as NNGP precision matrix with , trivially satisfies Assumption 3 if Assumption 1 holds as is simply .
5.2. Main results
We provide general results on the existence and consistency of neural network estimators minimizing a GLS loss for dependent data. The expected value of the GLS loss (8) is:
It is evident from above that naturally minimizes , while NN-GLS tries to minimize . We first show that such a minimizer exists in the seive class .
Theorem 1 (Existence of seive estimator). Given data generated from (6) under Assumption 1, and a working precision matrix satisfying Assumption 3, with the function classes defined in (16), there exists a seive estimator such that
| (19) |
All proofs are in Section S4 of Supplementary materials. The existence result ensures that a seive estimator in the class of neural networks that minimizes the GLS loss is well-defined. It is then natural to study its asymptotic consistency, as we do in the next result.
Theorem 2 (Consistency). Under Assumptions 1, 2, and 3, the NN-GLS estimate (19) minimizing the GLS loss in (8) is consistent in the sense .
To our knowledge, this is the first result on the consistency of neural networks for estimating the mean function of a spatially dependent process observed at an irregular set of locations. We do not impose any assumption on the true mean function beyond continuity and rely on very mild assumptions on the function class, and the covariance matrices of the data generation and analysis models. In Section 5.3 we show that these assumptions are satisfied for typical GP covariances and irregular spatial data designs. Also, note that this general result does not restrict the nature of the dependence to be spatial. Hence, while spatial applications are the focus of this manuscript, Theorem 2 can be used to establish consistency of neural networks for time-series, spatio-temporal, network, or other types of structured dependence. Section S6.7 presents a simulation study empirically demonstrating this consistency.
Theorem 2 generalizes the analogous consistency result of Shen et al. (2023) from i.i.d. data and OLS loss to dependent error processes and the use of a GLS loss. Consequently, the proof addresses challenges that do not arise in the i.i.d. case. The spatial dependency makes the standard Rademacher randomization fail and prevents using the standard Uniform law of large number (ULLN) result. We overcome this via construction of a different normed functional space equipped with a new Orlicz norm to adjust for data dependence and use of the GLS loss. This enables applying a ULLN for our dependent setting by showing that the empirical process is well-behaved with respect to this Orlicz norm.
5.3. Matérn Gaussian processes
We now establish consistency of NN-GLS for common GP covariance families, spatial data designs, and choice of working precision matrices. The main task in applying the general consistency result (Theorem 2) for these specific settings is verifying compliance to the regularity assumptions – i.e., the spectral bounds on the true Gaussian process covariance (Assumption 1) and on the working precision matrix (Assumptions 3).
We provide consistency results of NN-GLS for spatial data generated from the Matérn Gaussian process (see S4.2). This is the predominant choice of covariance family in geostatistical models due to the interpretability of its parameters with a marginal spatial variance , decay of spatial correlation , smoothness of the underlying process (Stein, 1999). Our first result considers data generated from a class of GP that contains the Matérn family and where the working precision matrix is the true GP precision matrix.
Proposition 1. Consider data generated from a spatial process at locations in , where is continuous, is a Gaussian process with covariance function , and is a covariance of a stationary spatial GP. Suppose the data locations are separated by a minimum distance , i.e., . Let denote the covariance matrix of , where . Then NN-GLS using is consistent, i.e., , if either (a) and for some , or (b) and is a Matérn covariance function with parameters .
The decay rate is satisfied by the Matérn family (Abramowitz and Stegun, 1948), so Proposition 1 proves the consistency of NN-GLS for Matérn GP, both with and without a nugget. The result holds for any irregular spatial design in being separated by a minimum distance. As the sample size grows, this is equivalent to considering the increasing domain paradigm that is commonly adopted as Matérn GP parameters are not identifiable if data are collected densely in a fixed spatial domain (Zhang, 2004).
Propositions 1 describes the case where true covariance structure is known. In that case, it’s possible to directly use the inverse of the covariance matrix as the working precision matrix in the GLS loss. However, this is often infeasible for multiple reasons. First, the true covariance parameters are usually unknown, and the working covariance matrix will typically use different (estimated) parameter values. Computationally, GLS loss using the full Matérn GP covariance matrix will require time and storage, which is not available even for moderate . The next proposition introduces a more pragmatic result proving the consistency of NN-GLS for data generated from Matérn GP but when using a working precision matrix derived from NNGP (as described in Section 4) and with parameter values different from the truth.
Proposition 2. Consider data generated as in Proposition 1 from a Matérn GP with parameters at locations separated by a minimum distance . Let be the NNGP precision matrix based on a Matérn covariance with parameters , i.e. , and using neighbor sets of maximum size with each location appearing in at most many neighbor sets. Then NN-GLS using is consistent, i.e., for any choice of , where is the Matérn covariance function with unit variance, unit spatial decay, and smoothness , and is its inverse function.
Proposition 2 provides consistency of NN-GLS for Matérn GP when using NNGP working precision matrices. This is the actual choice of used in our algorithm as it can be represented as a GNN, thereby facilitating a scalable implementation, as described in Section 4. The result allows the spatial parameter used in the working covariance to be different from the truth. The spatial decay parameter for the working precision matrix needs to be sufficiently large to ensure is not close to being singular, which will lead to numerical instability. We note that this is not a restriction on the data generation process, but just on the working precision matrix and can be chosen by the user. This is not enforced in practice, as these parameters are estimated as outlined in Section 4.2. In Section S5.1, we provide the theoretical intuition and empirical evidence that NN-GLS estimates the spatial parameters consistently. The restriction on each location appearing in at most a fixed number of neighbor sets of NNGP is usually satisfied in all but very pathological designs.
To our knowledge, Propositions 1 and 2 on the consistency of NN-GLS are the first examples of consistency of any machine-learning-based approach to estimating a nonlinear mean of a Matérn GP for irregular spatial designs. The only similar result in the literature is the consistency of RF-GLS, the GLS-based random forest approach of Saha et al. (2023). However, their result relies on a one-dimensional regular lattice design and restricts the true Matérn process smoothness to be half-integers. Our result is valid for any irregular spatial designs in the two-dimensional space – the most typical setting of spatial data collection. The result also holds for any true parameter values of the Matérn process.
5.4. Finite-sample error rates
We also obtain the finite-sample error rate for NN-GLS that informs about the importance of using the GLS loss in neural networks for spatial data.
Theorem 3 (Convergence rate). Let be the projection of in . Under Assumptions 1, 2, and 3, the NN-GLS estimate (19) satisfies
There are two main takeaways from this rate result. First, the rate for NN-GLS is, up to a scaling factor, the same as that for OLS neural networks in the i.i.d. case (Shen et al., 2023). We do not make any assumption on the class of the regression function beyond continuity and refer the readers to (Shen et al., 2023) for discussion on specific choices of classes of the true function , and where this rate is sharp up to logarithmic terms.
The second and novel insight is the scaling of the error rate by a factor — the ratio of the largest and smallest eigenvalues of the discrepancy matrix (18). In general, is close to 1 if the discrepancy matrix is closer to . Proposition S2 in Appendix S5.2 shows that the Kullback-Leibler distance between the identity matrix and the discrepancy matrix when using an NNGP working covariance matrix with -nearest neighbors is a decreasing function of . The best case scenario is when using the true covariance as the working covariance, i.e., yielding and . This corresponds to using NNGP with . At the other extreme, is the vanilla neural network using the OLS loss, corresponds to using a working covariance matrix of in NN-GLS. This corresponds to NNGP with and is thus the worst approximation according to Proposition S2. We verify this empirically, plotting in Figure S2 both the KLD and the spectral width of as a function of . We see a very large spectral width for , i.e., using OLS loss, implying a large error rate and inefficient estimation when using OLS loss in neural networks for spatial data. This is manifested in our results detailed later. The spectral width for NNGP approximation with is really narrow, yielding a near-optimal scaling factor close to 1. This demonstrates how modeling the spatial covariance in the neural network estimation via the GLS loss improves the error rates and guides the choice of the number of nearest neighbor in NN-GLS (see Section S5.3).
6. Simulation study
We conduct extensive simulation experiments to study the advantages of NN-GLS over existing methods in terms of both prediction and estimation. The data are simulated from the spatial GP model (6) with two choices for the non-linear mean function and (Friedman function, Friedman, 1991). Section S6.1 provides all parameter values and implementation details.
We consider a total of 10 methods that represent both linear and nonlinear, spatial and non-spatial, statistical and machine learning paradigms. We include 3 candidate neural network approaches for comparison to NN-GLS: NN without any spatial information (NN-nonspatial), NN using the spatial coordinates as two additional inputs (NN-latlon), and NN using spline basis as additional inputs (NN-splines). The NN-splines method is a rendition of Deepkriging (Chen et al., 2024). It uses the same basis functions as Deepkriging but to study the effect of additional covariates terms, we keep the numbers of layers and nodes the same for all the neural network based methods. Among these, only NN-GLS and NN-nonspatial offer both estimation and prediction, NN-latlon, and NN-splines only offer prediction and do not estimate of the covariate effect (see Section 2.2).
Other than neural networks, we also include 5 other popular non-linear methods, including Generalized Additive Models (GAM), Random Forests (RF), and their counterparts under spatial settings, GAM using the spatial coordinates as two additional inputs (GAM-latlon), GAM-GLS (Nandy et al., 2017), and RF-GLS (Saha et al., 2023). RF-GLS offers both estimation and spatial predictions. GAM-latlon is used for predictions, and the other three methods are only considered for estimation. We also include the spatial linear GP model, implemented using NNGP through the BRISC R-package (Saha and Datta, 2018). Section S6.1 provides details of all the methods, and Table S1 summarizes their estimation and prediction capabilities and scalability for large datasets.
A snapshot of the results from the different experiments is provided in Figure 2. We first evaluate the estimation performance of the methods that offer an estimate of the mean function — linear-spatial model, GAM, GAM-GLS, RF, RF-GLS, NN-nonspatial, and NN-GLS. We use the Mean Integrated Square Error (MISE) for the estimate . Figures S3(a) and S4(a) in Sections S6.2 and S6.3 respectively present the estimation results for all scenarios. A representative result is provided in Figure 2 (a) which presents the MISE for all methods when and for 3 choices of the nugget variance . NN-GLS consistently outperforms the other methods. The non-spatial neural network, which uses OLS loss, performs poorly for estimation even though it uses the same function class as NN-GLS for the mean. This shows the importance of explicitly accounting for the spatial covariance in the neural network estimation process. We further compare the two in Figure S5, which shows that NN-GLS has a more significant advantage over non-spatial neural networks when the spatial decay is small. This is expected as for small , there is a strong spatial correlation in the data, so the performance of NN-nonspatial suffers on account of ignoring this spatial information. The deterioration in the performance of NN-nonspatial over NN-GLS is lesser for large , as there is only a weak spatial correlation in the data.
Fig. 2.
(a): The estimation performance comparison with ; (b): (Section S6.4) Estimation performance comparison among GAM, GAM-GLS and NN-GLS against rho, i.e., the interaction strength; (c): The prediction performance comparison with (d): (Section S6.9) Prediction performance comparison among non-spatial NN, NN-GLS and NN-splines against sample size; (e): (Section S6.7) The consistency of estimation. (f): (Section S6.7) The running time for estimation.
In Figure 2(b) we present focussed comparisons between NN-GLS and methods from the GAM family. We consider variants of the Friedman function for the mean, controlling the weight of the interaction term with a parameter . Neither GAM nor GAM-GLS can model interaction, and this is reflected in their performance. When is small (weak interaction effect), their MISE are close to that from NN-GLS. However, when is large (strong interaction effect), the MISE of the GAM methods are considerably worse. This shows the advantage of the neural network family over GAM for non-linear regression in the presence of interaction terms (see Section S6.4 for details).
We compare prediction performances using the Relative Mean Squared Error (RMSE) on the test set, obtained by standardizing the MSE by the variance of the response so that the quantity can be compared across different experiments. Figures S3(b) and S4(b) in Sections S6.2 and S6.3 present the prediction results from all scenarios. When , i.e., the sine function (Figure S3(b)), the spatial linear model, unsurprisingly, offers the worst prediction performance by missing the strong non-linearity. NN-GLS is consistently the best, on account of using spatial information both in estimation and prediction, and the performances of other methods lie between NN-GLS and the linear model. In Figure 2(c), we present the more interesting scenario when . It’s surprising to notice that the linear-spatial model outperforms all methods except NN-GLS in terms of prediction performance. The reason is the partial-linearity of Friedman function and the power of the GP in the linear model, NN-GLS offers the lowest RMSE. In fact, all the GP-based approaches (spatial linear model, RF-GLS, NN-GLS) outperform the added-spatial-features approaches (GAM-latlon, NN-latlon, NN-splines). This demonstrates the limitation of the added-spatial-features approaches compared to NN-GLS which embeds NN in the GP model. The choice and the number of the added basis functions may need to be tuned carefully to optimize performance for specific applications. Also, the sample sizes may need to be much larger to fully unlock the non-parametric ability of the added-spatial-features approaches. We see this in Figure 2(d) which shows the prediction performance against sample size n under fixed-domain sampling. For large , the performance of NN-splines becomes similar to NN-GLS, although NN-GLS still has lower RMSE. NN-GLS circumvents basis functions by parsimoniously modeling the spatial dependence directly through the GP covariance, and performs well for both small and large sample sizes and fixed- and increasing-domain sampling paradigms (see Section S6.9 for all the scenarios).
Figures 2(e) and (f) present the estimation performance and running times for different methods up to sample sizes of 500, 000. The large sample runs demonstrate how the approximation error for NN-GLS goes to zero while for others like the GAM family and linear models stay away from zero as they cannot approximate the Friedman function. This empirically verifies the consistency result. NN-GLS also scales linearly in due to the innovations of Section 4 and is much faster than the other non-linear GLS approaches like RF-GLS and GAM-GLS. Details are in Section S6.7.
Section S6 of the Supplementary materials presents all the other simulation results. The confidence interval for and the prediction interval for are evaluated in Sections S6.5 and S6.6 respectively. Both types of intervals provided by NN-GLS provide near-nominal coverage as well as superior interval scores against the competing methods. To study the impact of sampling design, a denser sampling is considered in Section S6.8 keeping the domain fixed. This is to supplement the theoretical results, which are only for the increasing domain. In section S6.10 we investigate if the estimation of the spatial parameters has an impact on the performance of NN-GLS by comparing it to NN-GLS (oracle) which uses the true spatial parameters. We observe that NN-GLS’s performance is quite similar to that of NN-GLS (oracle) since it provides an accurate estimation of spatial parameters. NN-GLS also performs well for a higher dimensional mean function (of 15 covariates) (Section S6.11). We finally assess the robustness of NN-GLS to model misspecification, including misspecification of the GP covariance (Section S6.12), and complete misspecification of the spatial dependence (Section S6.13). For both estimation and prediction, NN-GLS performs the best or comparably with the best method for both cases of misspecification.
7. Real data example
The real data example considered here is the spatial analysis of PM2.5 (fine particulate matter) data in the continental United States. Following a similar analysis in Chen et al. (2024), we consider the daily PM2.5 data from 719 locations on June 18th, 2022 and regress PM2.5 on six meteorological variables — precipitation accumulation, air temperature, pressure, relative humidity, west wind (U-wind) speed, and north wind (V-wind) speed. Section S7.1 has more details about the data.
Figure 3 (a) demonstrates the PM2.5 distribution on the date (smoothed by inverse distance weighting interpolation), as well as the nationwide EPA monitor’s distribution (the orange spots). The spatial nature of PM2.5 is evident from the map. We consider the following 5 methods previously introduced in Section 6: GAM-latlon, RF-GLS, NN-latlon, NN-splines, and NN-GLS. To evaluate the prediction performance, we randomly take 83% (5/6th) of the data as a training set and the rest part as a testing set using block-random data splitting strategy, which is closer to a real-world scenario (see Appendix S7.2 for details), train the model and calculate the RMSE of the prediction on the testing set. The procedure is repeated for 100 times. The performance of each method is shown in Figure 3(b). We find that NN-GLS and RF-GLS have comparably the lowest average RMSE, while GAM-latlon, NN-latlon, and NN-splines are outperformed. This trend is consistent for other choices of dates and for different ways of data splitting, except for some cases where the spatial pattern is not clear and NN-splines may do similarly (see Section S7.3 of the Supplementary materials). Sensitivity to the choice of hold-out data is presented in Section S7.2. Section S7.4 evaluates the performance of prediction intervals on hold-out data. We see that prediction intervals from NN-GLS offers near nominal coverage. Model adequacy checks are performed in Section S7.5 and reveal that data from different days adhere to the modeling assumptions of Gaussianity and exponentially decaying spatial correlation to varying degrees. As NN-GLS performs well on all days, it gives confidence in the applicability of NN-GLS in a wide variety of scenarios.
Fig. 3.
PM2.5 data analysis.
NN-GLS also provides a direct estimate of the effect of the meteorological covariates on PM2.5, specified through the mean function . However, is six-dimensional in this application, precluding any convenient visualization of the function . Hence, we use partial dependency plots (PDP) – a common tool in the machine learning world for visualizing the marginal effect of one or two features on the response (see Section S7.6 for a detailed introduction). In Figure 4, we present the PDP of PM2.5 on temperature and wind. While the PM2.5 level is linearly affected by temperature, we see clear nonlinear effects in the wind, thereby demonstrating the need to move beyond the traditional linear spatial models and consider geostatistical models with non-linear means, like NN-GLS.
Fig. 4.
Partial dependency plots showing the marginal effects of temperature and west wind (U-wind).
8. Discussion
In this work, we showed how neural networks can be embedded directly within traditional geostatistical Gaussian process models, creating a hybrid machine learning statistical modeling approach that balances parsimony and flexibility. Compared to existing renditions of NN for spatial data which need to create and curate many spatial basis functions as additional features, NN-GLS simply encodes spatial information explicitly and parsimoniously through the GP covariance. It reaps all the benefits of the model-based framework including the separation of the non-spatial and spatial structures into the mean and the covariance, the use of GLS loss to account for spatial dependence while estimating the non-linear mean using neural networks, and the prediction at new locations via kriging.
We show that NN-GLS can be represented as a graph neural network. The resulting GLS loss is equivalent to adding a graph convolution layer to the output layer of an OLS-style neural network. This ensures the computational techniques used in standard NN optimization like mini-batching and backpropagation can be adapted for NN-GLS, resulting in a linear time algorithm. Also, kriging predictions can be obtained entirely using the GNN architecture using graph convolution and deconvolution.
Our connection of NN-GLS to GNN is of independent importance, as it demonstrates the connection of GNN to traditional geostatistics. Current adaptations of GNN to spatial data (Tonks et al., 2022; Fan et al., 2022) mostly use simple graph convolution choices like equal-weighted average of neighboring observations. For irregular spatial data, this disregards the unequal distances among the observations, which dictate the strength of spatial correlations in traditional geospatial models. Our representation result shows how graph-convolution weights arise naturally as nearest neighbor kriging weights which accounts for mutual distances among data locations. Future work will extend this GNN framework and consider more general graph deconvolution layers, and applicable to other types of irregular spatial data where graphical models are already in use, like multivariate spatial outcomes using inter-variable graphs (Dey et al., 2022) or areal data graphs (Datta et al., 2019).
We prove general theory on the existence and consistency of neural networks using GLS loss for spatial processes, including Matérn GP, observed over an irregular set of locations. We establish finite sample error rates of NN-GLS, which improves when the working covariance matrix used in the GLS loss is close to the true data covariance matrix. It proves that ignoring the spatial covariance by simply using the OLS loss leads to large error rates. To the best of our knowledge, we presented the first theoretical results for any neural network approach under spatial dependency.
There is a gap between the theory and the actual implementation of NN-GLS. The theory relies on a restricted class of neural networks, and does not consider steps like mini-batching and backpropagation used in practice. Even in a non-spatial context, this optimization error gap between the practice and theory of NN is yet to be bridged. Our theory is not complexity-adaptive as we do not assume any structure on the true mean function beyond continuity, It will be interesting to pursue adaptive rates for NN-GLS akin to recent works of Schmidt-Hieber (2020) and others in a non-spatial context (see Section S3). Extension to high-dimensional covariates as considered in Fan and Gu (2023) is also a future direction. Theoretically justified methodology for pointwise inference (interval estimates) for the mean function from neural networks remains open, even for i.i.d. case.
We primarily focussed on modeling the mean as a function of the covariates using a rich non-linear family, i.e., the neural network class, while using stationary covariances to model the spatial structure. However, non-stationarity can be easily accommodated in NN-GLS either by including a few basis functions in the mean or using the GLS loss with a non-stationary covariance function. For example, Zammit-Mangion et al. (2021) proposed rich classes of non-stationary covariance functions using transformations of the space modeled with neural networks.
Supplementary Material
Acknowledgement and Disclosures
This work is supported by National Institute of Environmental Health Sciences grant R01ES033739. The authors report there are no competing interests to declare.
Footnotes
Supplementary materials and Software
The supplement contains proofs, additional theoretical insights or methodological details, and results from more numerical experiments. A Python implementation of NN-GLS which is publicly available in the geospaNN package at https://pypi.org/project/geospaNN/.
References
- Abramowitz M and Stegun IA (1948), Handbook of mathematical functions with formulas, graphs, and mathematical tables, Vol. 55, US Government printing office. [Google Scholar]
- Banerjee S, Carlin BP and Gelfand AE (2014), Hierarchical modeling and analysis for spatial data, CRC press. [Google Scholar]
- Breiman L. (2001), ‘Random forests’, Machine learning 45(1), 5–32. [Google Scholar]
- Chen W, Li Y, Reich BJ and Sun Y (2024), ‘Deepkriging: Spatially dependent deep neural networks for spatial prediction’, Statistica Sinica 34, 291–311. [Google Scholar]
- Cressie N and Wikle CK (2015), Statistics for spatio-temporal data, John Wiley & Sons. [Google Scholar]
- Datta A, Banerjee S, Finley AO and Gelfand AE (2016), ‘Hierarchical nearest-neighbor Gaussian process models for large geostatistical datasets’, Journal of the American Statistical Association 111(514), 800–812. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Datta A, Banerjee S, Hodges JS and Gao L (2019), ‘Spatial disease mapping using directed acyclic graph auto-regressive (dagar) models’, Bayesian analysis 14(4), 1221. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dey D, Datta A and Banerjee S (2022), ‘Graphical gaussian process models for highly multivariate spatial data’, Biometrika 109(4), 993–1014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fan J, Bai J, Li Z, Ortiz-Bobea A and Gomes CP (2022), A gnn-rnn approach for harnessing geospatial and temporal information: application to crop yield prediction, in ‘Proceedings of the AAAI conference on artificial intelligence’, Vol. 36, pp. 11873–11881. [Google Scholar]
- Fan J and Gu Y (2023), ‘Factor augmented sparse throughput deep relu neural networks for high dimensional regression’, Journal of the American Statistical Association 0(0), 1–15. URL: 10.1080/01621459.2023.2271605 [DOI] [Google Scholar]
- Freund Y, Schapire R and Abe N (1999), ‘A short introduction to boosting’, Journal-Japanese Society For Artificial Intelligence 14(771−780), 1612. [Google Scholar]
- Friedman JH (1991), ‘Multivariate adaptive regression splines’, Annals of statistics 19(1), 1–67. [Google Scholar]
- Goodfellow I, Bengio Y and Courville A (2016), Deep learning, MIT press. [Google Scholar]
- Gray SD, Heaton MJ, Bolintineanu DS and Olson A (2022), ‘On the use of deep neural networks for large-scale spatial prediction.’, Journal of Data Science 20(4), 493–511. [Google Scholar]
- Hornik K, Stinchcombe M and White H (1989), ‘Multilayer feedforward networks are universal approximators’, Neural networks 2(5), 359–366. [Google Scholar]
- Iranzad R, Liu X, Chaovalitwongse WA, Hippe D, Wang S, Han J, Thammasorn P, Duan C, Zeng J and Bowen S (2022), ‘Gradient boosted trees for spatial data and its application to medical imaging data’, IISE Transactions on Healthcare Systems Engineering 12(3), 165–179. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nandy S, Lim CY and Maiti T (2017), ‘Additive model building for spatial regression’, Journal of the Royal Statistical Society Series B: Statistical Methodology 79(3), 779–800. [Google Scholar]
- Saha A, Basu S and Datta A (2023), ‘Random forests for spatially dependent data’, Journal of the American Statistical Association 118(541), 665–683. [Google Scholar]
- Saha A and Datta A (2018), ‘Brisc: bootstrap for rapid inference on spatial covariances’, Stat 7(1), e184. [Google Scholar]
- Schmidt-Hieber J. (2020), ‘Nonparametric regression using deep neural networks with ReLU activation function’, The Annals of Statistics 48(4), 1875–1897. URL: 10.1214/19-AOS1875 [DOI] [Google Scholar]
- Shen X, Jiang C, Sakhanenko L and Lu Q (2023), ‘Asymptotic properties of neural network sieve estimators’, Journal of Nonparametric Statistics 35(4), 839–868. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sigrist F. (2022), ‘Gaussian process boosting’, The Journal of Machine Learning Research 23(1), 10565–10610. [Google Scholar]
- Stein ML (1999), Interpolation of spatial data: some theory for kriging, Springer Science & Business Media. [Google Scholar]
- Tonks A, Harris T, Li B, Brown W and Smith R (2022), ‘Forecasting west nile virus with graph neural networks: Harnessing spatial dependence in irregularly sampled geospatial data’, arXiv preprint arXiv:2212.11367. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wang H, Guan Y and Reich B (2019), Nearest-neighbor neural networks for geostatistics, in ‘2019 international conference on data mining workshops (ICDMW)’, IEEE, pp. 196–205. [Google Scholar]
- Wikle CK and Zammit-Mangion A (2023), ‘Statistical deep learning for spatial and spatiotemporal data’, Annual Review of Statistics and Its Application 10, 247–270. [Google Scholar]
- Zammit-Mangion A, Ng TLJ, Vu Q and Filippone M (2021), ‘Deep compositional spatial models’, Journal of the American Statistical Association pp. 1–22.35757777 [Google Scholar]
- Zhang H. (2004), ‘Inconsistent estimation and asymptotically equal interpolations in model-based geostatistics’, Journal of the American Statistical Association 99(465), 250–261. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.




