Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2021 Feb 19;21(4):1460. doi: 10.3390/s21041460

Sequential Sampling and Estimation of Approximately Bandlimited Graph Signals

Sijie Lin 1, Ke Xu 2, Hui Feng 1,*, Bo Hu 1
Editor: Vladimir Stankovic
PMCID: PMC7922557  PMID: 33669801

Abstract

Graph signal sampling has been widely studied in recent years, but the accurate signal models required by most of the existing sampling methods are usually unavailable prior to any observations made in a practical environment. In this paper, a sequential sampling and estimation algorithm is proposed for approximately bandlimited graph signals, in the absence of prior knowledge concerning signal properties. We approach the problem from a Bayesian perspective in which we formulate the signal prior by a multivariate Gaussian distribution with unknown hyperparameters. To overcome the interconnected problems associated with the parameter estimation, in the proposed algorithm, hyperparameter estimation and sample selection are performed in an alternating way. At each step, the unknown hyperparameters are updated by an expectation maximization procedure based on historical observations, and then the next node in the sampling operation is chosen by uncertainty sampling with the latest hyperparameters. We prove that under some specific conditions, signal estimation in the proposed algorithm is consistent. Subsequent validation of the approach through simulations shows that the proposed procedure yields performances which are significantly better than existing state-of-the-art approaches notwithstanding the additional attribute of robustness in the presence of a broad range of signal attributes.

Keywords: graph signal, sequential sampling, consistent estimation

1. Introduction

Graph signal is a powerful tool to represent and analyze data with irregular interconnections by defining signal values on graph vertices and assigning edge weights according to correlation or similarity [1]. In the past decade, graph signal processing theory has developed rapidly, extending classical signal processing techniques such as Fourier analysis [1,2], filtering [3,4,5], sampling and interpolation [6,7,8] to graph signal setting. Related concepts and methods have found wide application in sensor networks [9,10], brain analysis [11], image processing [12], three-dimensional (3D) point cloud processing [13], and machine learning [14,15,16].

Sampling for (lossless) reconstruction or (minimum error) estimation is a fundamental problem in graph signal processing. For example, a sample survey in a social network needs to carefully select interviewees so as to better predict the attitudes of all users. In a sensor network, it is important to optimize sensor placement so that more information can be collected by a restricted number of sensors due to economic constraint [9]. In addition, any graph-based active semisupervised learning task may also be interpreted as a graph signal sampling and estimation problem [14,15].

There have been multiple works on graph signal sampling. The authors of [6,7] try to extend the Nyquist sampling theory to bandlimited graph signals. More works are seeking optimal sampling sets for graph signal estimation, where the graph signals are usually assumed to be bandlimited [17,18,19] or approximately bandlimited [17,20] in the frequency domain, or smooth in the vertex domain [17,21]. Sampling strategies mainly include topology-based methods that compute a score for each vertex [18,21], and design-of-experiment (DOE) approaches that optimize certain scalarization form of some target matrix [19,22]. Most of the current sampling methods rely on accurate bandwidth [6,18,19] or prior distribution [22,23] of the graph signal.

It is natural in many applications to assume that the graph signal is somehow bandlimited or smooth. For instance, people with strong social connection are likely to hold similar opinions and temperature sensors located close to each other usually get similar measurements. However, in practice it is hard to know the exact bandwidth or prior distribution of the graph signal before any observation. In such cases, those sampling strategies based on accurate signal model become not applicable.

A possible solution is to sample and estimate in a sequential way. Unknown signal properties can be estimated from previous observations, and the subsequent sampling node can be selected under the latest signal model. Such a sequential framework can overcome the difficulty of incomplete system model and take advantage of model-based sample selection and signal estimation [20].

In this work, we investigate the sequential sampling and estimation of approximately bandlimited (ABL) graph signals, whose bandwidth and energy level are both not known in advance. With the aid of a Butterworth low-pass filter, the ABL graph signal is assumed to follow a multivariate Gaussian distribution with unknown hyperparameters. The variance of observation noise is also considered unknown.

In the proposed algorithm, hyperparameter estimation and sample selection are performed in an alternating way. At each step, expectation maximization (EM) [24] is first used to update the maximum marginal likelihood (MML) estimation of unknown hyperparameters based on historical observations. Substituting these latest estimated values into the signal prior and noise distribution, a Bayesian estimation and prediction procedure can then be performed. According to the uncertainty sampling (US) criterion [25], the node with the largest predictive variance is selected as the next node to sample.

In fact, estimating prior distribution from data, namely, empirical Bayes (EB) method, has long been studied [26]. It has been proved that under certain assumptions, the EB posterior distribution of parameter is consistent at its true value [27]. Meanwhile, an alternating approach analogous to the proposed one has been used in sequential DOE for nonlinear least squares (LS) regression. There, parameter estimation is proved to be consistent, and the design is asymptotically optimal [28].

In this paper, we prove that under specific conditions, signal estimation in the proposed algorithm is consistent at the true value. That is, as the number of samples goes to infinity, the estimated graph signal can get arbitrarily close to the true value. The finite-time performance of the proposed algorithm is validated by simulation results. Our algorithm is able to adjust to ABL graph signals with different properties, and it performs better in estimation error than existing methods.

The rest of the paper is organized as follows. We detail the signal prior and observation model in Section 2, and develop a sequential sampling and estimation algorithm in Section 3. The consistency of the proposed algorithm is proved under specific conditions in Section 4, and its finite-time performance is validated by experiments in Section 5. We conclude the paper in Section 6.

Throughout the paper, we use normal-weight italic lowercase letters (e.g., x) to denote scalar variables, bold italic lowercase letters (e.g., x) to denote vectors, bold italic uppercase letters (e.g., X) to denote matrices, and calligraphic uppercase letters (e.g., X) to denote sets.

2. System Model

2.1. Preliminaries

We consider a simple connected weighted undirected graph G=(V,E,W), where V is the set of vertices indexed by {1,,N}; E is the set of edges between vertex pairs, e.g., (i,j)E denotes an edge between vertex i and vertex j; and W is the weighted adjacency matrix, with wij the weight of edge between vertex i and vertex j if they are connected, or 0 otherwise. A graph signal that takes a real value on each vertex of the graph can be represented as fRN, where fi is the signal value on vertex i.

The Laplacian matrix is defined as L=DW, where D=(d1,,dN) with di=j:(i,j)Ewij is the degree matrix. Being symmetric and positive semi-definite, L has a spectral decomposition L=VΛVT, where Λ is a diagonal matrix of the eigenvalues 0=λ1<λ2λN, and V contains the corresponding orthonormal eigenvectors {vk}k=1N as columns. These eigenvectors act as graph Fourier basis, and the associated eigenvalues are regarded as graph frequencies [1]. The frequency-domain coefficients f^ of a graph signal f can be calculated via the graph Fourier transform (GFT) as f^=VTf. The graph signal f can be expressed as f=Vf^, which is known as the inverse GFT (IGFT).

2.2. Signal Prior

In this work, the graph signal f is assumed to be drawn from the following distribution:

p(f;α,γ)=N(f|0,(αIN+γL2n)1), (1)

where IN denotes the N×N identity matrix, and n,α,γ>0 are hyperparameters, among which α,γ are assumed unknown in prior. The meanings of the signal model and hyperparameters are explained in the frequency domain as follows.

By GFT, the spectral coefficients {f^k}k=1N follow independent zero-mean Gaussian distributions with variances conforming to a Butterworth low-pass filter ([29], Section 7.3):

p(f^k)=Nf^k|0,g21+(λk/λc)2n, (2)

where n is the order, λc=α/γ2n is the cut-off frequency, and g=1/α is the amplitude gain of the filter. Under such a prior, we say the graph signal is approximately bandlimited in a probabilistic sense, where λc plays the role of bandwidth, n controls the strictness of banlimitedness, and g determines the energy level of the graph signal.

Specially, when n=1/2 and δ=α/γ is small, the signal prior Equation (1) becomes

p(f)expγ2fT(L+δIN)f, (3)

which is widely used in graph signal sampling and estimation [9,23]. The restriction on total variation fTLf=(i,j)Ewij(fifj)2 can lead to smooth signals over the graph.

If n, the out-of-band coefficients become deterministic zeros. The graph signal becomes strictly bandlimited, represented by

f=VKf^K, (4)

where K=max{k|λkλc}, f^K collects the first K GFT coefficients that may be non-zero, and VK contains the first K GFT basis.

In conclusion, our signal prior Equation (1) is flexible to describe approximately bandlimited graph signals with different bandwidth, strictness of bandlimitedness, and energy level. It can also cover the most commonly used smoothness prior [9,23] and bandlimitedness prior [6,7,18,19]. Additionally, this model is simple with three hyperparameters, and can be represented in the vertex domain without eigendecomposition.

Here, hyperparameters α,γ or equivalently λc,g in the signal prior are assumed unknown, that is, we do not need to have prior knowledge about the bandwidth or energy level of the graph signal. Only a value for n is required to control the strictness of bandlimitedness of the estimated signal. Compared to those methods for only strictly bandlimited graph signals with known bandwidth or smooth graph signals with known prior distribution, our signal prior is more general.

2.3. Observation Model

To overcome the difficulty caused by unknown hyperparameters, a sequential sampling process is considered, where one node is sampled at each step, so that unknown hyperparameters can be estimated from previous observations, and the next node to sample can be selected using the latest hyperparameters.

Suppose that vertex st is sampled at step t, and the observed signal value is yt. The observation model can be expressed as

yt=ψtTf+wt, (5)

where ψtT=estT is the sampling vector with its st-th element equal to 1 and others equal to 0, and wt is an additive zero-mean Gaussian observation noise with unknown precision β,

p(wt;β)=N(wt|0,β1). (6)

In hyperparameter, signal estimation, and sample selection at step t, the historical observations from step 1 to step t are considered together. Let

y1:t=y1yt,Ψ1:t=ψ1TψtT=es1TestT,w1:t=w1wt. (7)

The noise values w1,,wt involved in different samples (on either different vertices or the same vertex) are assumed to be independent and identically distributed (i.i.d.). According to Equations (5) and (6),

y1:t=Ψ1:tf+w1:t, (8)
p(w1:t;β)=N(w1:t|0,β1It). (9)

2.4. Problem Formulation

Under the signal and observation model described in Section 2.2 and Section 2.3, at each step t in the sequential sampling and estimation process, the two core problems are (1) how to estimate the unknown hyperparameters α,γ,β as well as the graph signal f based on the historical observations Ψ1:t,y1:t

αt,γt,βt,ftΨ1:t,y1:t, (10)

where αt,γt,βt are the estimated values of hyperparameters at step t, and ft denotes the estimated graph signal at step t, and (2) how to select the next node st+1 to be sampled at step t+1 with the latest estimated values of hyperparameters

st+1Ψ1:t,y1:t,αt,γt,βt. (11)

Our ultimate goal is to minimize signal estimation error within given budget, or from another angle, reach certain estimation accuracy with least samples.

3. Algorithm

3.1. Hyperparameter and Signal Estimation

We first focus on unknown hyperparameter estimation from observation data. Here, maximum marginal likelihood (MML) estimation of unknown hyperparameters is adopted:

αt,γt,βt=argmaxα,γ,βp(y1:t|Ψ1:t;α,γ,β)=argmaxα,γ,βRNp(y1:t|Ψ1:t,f;β)p(f;α,γ)df. (12)

Then, the signal posterior given estimated hyperparameters is Gaussian:

p(f|Ψ1:t,y1:t;αt,γt,βt)=N(f|μt,Ct), (13)

with mean vector and covariance matrix ([30], Section 10.6)

μt=βtCtΨ1:tTy1:t, (14)
Ct=(αtIN+γtL2n+βtΨ1:tTΨ1:t)1. (15)

The minimum mean square error (MMSE) or maximum a posteriori (MAP) estimation of the graph signal is ft=μt.

However, direct optimization of Equation (12) is intractable. We view f as a hidden variable, and introduce expectation maximization (EM) ([24], Section 9.4) in our algorithm to estimate α,γ,β, and thereby obtain the posterior distribution of f. To be concise, the subscripts t and 1:t are omitted in the introduction of our EM algorithm.

Consider the l-th EM iteration. In E step, hyperparameters α,γ,β are fixed at αl1,γl1, βl1, and the signal posterior p(f|Ψ,y;αl1,γl1,βl1) with mean μl1 and covariance Cl1 is computed as in Equations (13)–(15).

In M step, α,γ,β are updated according to

αl,γl,βl=argmaxα,γ,βEf|Ψ,y;αl1,γl1,βl1[lnp(y,f|Ψ;α,γ,β)]=argmaxα,γ,βEf|Ψ,y;αl1,γl1,βl1[lnp(y|Ψ,f;β)+lnp(f;α,γ)], (16)

where Ef|Ψ,y;αl1,γl1,βl1[·] means the expectation is taken with respect to p(f|Ψ,y;αl1, γl1,βl1) computed at E step, the signal prior p(f;α,γ) is given in Equation (1), and the signal likelihood p(y|Ψ,f;β)=N(y|Ψf,β1I) according to Equations (8) and (9).

For brevity, denote the expectation operator in Equation (16) as El1[·], and the expectation value as El1, which is also the objective value of the optimization problem. By direct calculations, we have

El1=12ln|αIN+γL2n|α2El1[fTf]γ2El1[fTL2nf]+M2lnββ2El1[(yΨf)T(yΨf)]+const, (17)

where M is the sample size; “const” denotes terms that are independent of α,γ,β; and

El1[fTf]=El1[(fμl1)T(fμl1)+2μl1Tfμl1Tμl1]=(Cl1)+μl1Tμl1, (18)
El1[fTL2nf]=El1[(fμl1)TL2n(fμl1)+2μl1TL2nfμl1TL2nμl1]=(L2nCl1)+μl1TL2nμl1, (19)
El1[(yΨf)T(yΨf)]=El1[(yΨμl1)T(yΨμl1)2(yΨμl1)TΨ(fμl1)+(fμl1)TΨTΨ(fμl1)]=(yΨμl1)T(yΨμl1)+(ΨTΨCl1). (20)

Note that El1 is concave with respect to α,γ,β, see Appendix A. The maximization problem in Equation (16) can be solved by any standard tool for convex optimization. Here, the first order conditions of the optimization problem are analyzed to give some insights into the hyperparameter estimation, and then a possible efficient search method based on these conditions is provided in Appendix B.

Take the partial derivatives of El1 with respect to α,γ,β, and set them to zero:

El1α=12((αIN+γL2n)1)12El1[fTf]=0, (21)
El1γ=12((αIN+γL2n)1L2n)12El1[fTL2nf]=0, (22)
El1β=M2β12El1[(yΨf)T(yΨf)]=0. (23)

Note that tr((αIN+γL2n)1), tr((αIN+γL2n)1L2n) and Mβ can be seen as the prior expectations of fTf, fTL2nf and wTw with respect to p(f;α,γ) and p(w;β). The M step actually looks for a group of hyperparameters that makes these prior expectations equal to their posterior ones. In this sense, the estimated hyperparameters agree with the observations.

Repeat the E and M steps as above until convergence, and we will obtain an MML estimation of unknown hyperparameters α,γ,β, together with a posterior of the graph signal f. This process is summarized in Algorithm 1. Although the estimated hyperparameters are not guaranteed to be globally optimal by MML, we find the performance good enough in experiments.

Algorithm 1 Hyperparameter and signal estimation by EM.
  • 1:

    Initialize α0,γ0,β0.

  • 2:

    forl=1,2,do

  • 3:

     Compute μl1,Cl1 as in Equations (14) and (15).

  • 4:

     Update αl,γl,βl according to Equation (16).

  • 5:

    end for

3.2. Sample Selection

Having been able to estimate the unknown hyperparameters from previous observations by EM, at each decision step, the subsequent sampling node can be selected using the latest estimated values of hyperparameters.

According to the uncertainty sampling (US) criterion in active learning ([25], Chapter 2), we should scan through all the nodes, and pick the one whose observed signal value we are most uncertain about as the next node to sample. Here, predictive variance is regarded as a measurement of uncertainty.

The predictive distribution of an observation y on vertex s with sampling vector ψT=esT given historical observations Ψ1:t,y1:t is

p(y|Ψ1:t,y1:t,ψT;αt,γt,βt)=N(y|esTμt,esTCtes+(βt)1), (24)

where μt and Ct are the posterior mean and covariance of the graph signal f, respectively, given Ψ1:t and y1:t. The predictive variance consists of two parts: estimative variance esTCtes and noise variance (βt)1, among which the latter is equal for all vertices. The next sampling node st+1 can be decided by

st+1=argmaxsVesTCtes. (25)

3.3. Sequential Sampling and Estimation

Finally, the complete procedure of the proposed sequential sampling and estimation algorithm for approximately bandlimited graph signals is given in Algorithm 2. At each step, first unknown hyperparameters are re-estimated by EM as in Section 3.1, and then the next node to sample is selected by US as in Section 3.2. The total number of sampling steps T is equal to the sampling budget.

Algorithm 2 Sequential sampling and estimation of approximately bandlimited graph signals.
  • 1:

    Choose the first sampling node s1 arbitrarily.

  • 2:

    fort=1,2,do

  • 3:

     Sample the vertex st and obtain an observation yt.

  • 4:

     Update hyperparameters αt,γt,βt and signal posterior μt,Ct by EM as in Algorithm 1.

  • 5:

     Select the next node to sample st+1 by US as in Equation (25).

  • 6:

    end for

Our sequential sampling strategy takes into account not only graph topology, but also previous observations, by both estimating hyperparameters based on historical data and selecting sampling nodes based on signal posterior. Making full use of historical observations and deciding which nodes to sample online, the proposed algorithm is able to cope with the situation where signal and noise distributions are not completely available in advance, and efficiently select samples to estimate the underlying graph signal.

The computational complexity of a decision step in the proposed algorithm is O(ntEMntbs¯N3), where ntEM is the number of EM iterations, ntbs¯ is the average iterations of binary search in M step, and N is the graph size. The computational cost is acceptable in view of the performance improvement, especially in time-insensitive scenarios with large observation cost. The proposed method is able to efficiently select samples to give accurate estimation of signals thanks to the sequential framework with EM hyperparameter update, and the signal estimation is consistent.

4. Asymptotic Analysis

In the previous section, we have developed a sequential sampling and estimation algorithm for approximately bandlimited graph signals with an incomplete system model, where hyperparameter estimation and sample selection are performed in an alternating way. The performance of the proposed algorithm will be validated via simulation in the next section, where we will see its efficient sample selection and accurate signal estimation with limited sampling budget and little prior knowledge. In this section, we are to emphasize that the proposed algorithm improves limited-budget performance without sacrifice of consistency, which is not trivial for a sequential decision process.

Intuitively, as the number of observations grows, hyperparameters better fitting the true model will be picked out, really important nodes will be selected to sample, and finally an excellent estimation of the graph signal can be obtained. Here, we try to provide some theoretical support for this intuition. The asymptotic performance of the proposed algorithm is analyzed, and we state and prove the consistency of signal estimation in our method.

Before concentrating on the asymptotic performance of signal estimation, we first investigate the asymptotic behavior of sample selection in the proposed algorithm.

Lemma 1.

Let mi,t denote the number of times vertex i is sampled up to step t. If limt1ttr(αtIN+γtL2n)Nβt=0,

δ>0,lim inftminiVmi,tt>δ.

That is, as t, all the nodes of the graph will be sampled again and again, each maintaining a sampling ratio greater than δ, including the lowest one. A proof of Lemma 1 is given in Appendix C. Then, two direct corollaries follow.

Corollary 1.

iV, δ>0, lim inftmi,tt>δ.

Corollary 2.

iV, limtmi,t=.

Based on them, we are now to state our main theorem on the consistency of signal estimation in the proposed algorithm.

Theorem 2.

The signal posterior p(f|Ψ1:t,y1:t;αt,γt,βt) is consistent at the true value f, if EM converges at MML hyperparameters satisfying the condition in Lemma 1.

The consistency of signal posterior here means, as t, the posterior distribution will become a point mass at the true value of the graph signal. A proof of Theorem 2 is provided in Appendix D. Then, we have the following corollary.

Corollary 3.

The MMSE/MAP estimation ft=μt is consistent at f under the conditions in Theorem 2.

It ensures that under specific conditions, as the sample size grows infinitely large, the MMSE/MAP estimation can get arbitrarily close to the true signal value, i.e., the graph signal can be estimated with arbitrary accuracy. This guarantees the asymptotic performance of the proposed algorithm in signal estimation.

The condition limt1ttr(αtIN+γtL2n)Nβt=0 is satisfied when αtβt=o(t), γtβt=o(t), or more sufficiently, if lim suptαt<, lim suptγt<, lim inftβt>0, which ensures that the estimated signal energy and bandwidth do not decrease to zero, and the estimated noise variance keeps finite. We did not observe any violation of these conditions in plenty of experiments, where the true graph signals do not have all same values, and all observations are finite.

5. Experiments

In this section, the finite-time performance of the proposed algorithm is validated via a series of simulation experiments, on both synthetic and real-world data.

The performance of the proposed algorithm is compared to the following methods:

  • RndEM: Select sampling nodes randomly and uniformly, and estimate hyperparameters and the signal using the proposed EM algorithm.

  • OrcUS: Suppose that the hyperparameters are known in advance (the “oracle”). Select sampling nodes by US, and estimate the signal by MMSE/MAP.

  • Anis2016: A heuristic sampling algorithm proposed in Anis et al. [7] to maximize the cut-off frequency such that the sampling set is a uniqueness set for the bandlimited subspace. The graph signal is recovered in the bandlimited subspace by least squares. The estimation order k of cut-off frequency is set to k=4.

  • Perraudin2018: A nonuniform random sampling method proposed in Perraudin et al. [21] with probability relevant to local uncertainty. Inpainting is done by minimizing total variation. The kernel in local uncertainty is g^(λ)=exp(τλ) where τ=8.

  • Sakiyama2019: In this method by Sakiyama et al. [31], the graph signal is recovered by a linear combination of localization operators at the sampled nodes. The sampling set is designed such that the energy of operator at each node is large and meanwhile the overlapping areas are small. The kernel in localization operator is g^(λ)=exp(τλ) where τ=0.5. The Chebyshev polynomial approximation order is P=12, and the signal recovery order is k=12.

RndEM and OrcUS are references to verify the effectiveness of the proposed algorithm. Anis2016, Perraudin2018, Sakiyama2019 are three state-of-the-art methods that do not directly rely on signal model. For Anis2016, the benefit of further increasing k is not noticeable in our experiments, but the computations will become less numerically stable. For Perraudin2018 and Sakiyama2019, we use the same kernel type as in the original papers, and the parameters are chosen by experiments to adjust to our settings.

The normalized l2-norm estimation error

e=ff2f2 (26)

is taken as a performance index, where f is the estimated signal, and f is the true value.

5.1. Simulations on Synthetic Data

Three graphs of representative types that have widespread applications in the real world are used for testing in our experiments:

  • G1: A random geometric graph with N=100 vertices randomly placed in a 1-by-1 square, and edge weights assigned via a thresholded Gaussian kernel
    wij=expdij22σ2,ifdij<r0,otherwise (27)
    where dij is the Euclid distance between vertex i and vertex j, r=0.2 and σ=0.1.
  • G2: A small world graph with N=100 nodes, generated by the Watts–Strogatz model [32] with mean node degree 6 and rewiring probability 0.2.

  • G3: A scale-free graph with N=100 nodes, generated by the Barabási–Albert model [33,34] with 1 initial node, each newly added node connected to 1 existing node, and connecting probabilities proportional to the degrees of existing nodes.

The graph topologies as well as their eigenvalue distributions are visualized in Figure 1. Due to limited space, we will mainly display and analyze our experimental results on G1 in detail, whereas example results on G2 and G3 are shown at the end to demonstrate the topology adaptability.

Figure 1.

Figure 1

The three graphs used in our simulations, and the histograms of their eigenvalue distributions. (a) G1, a random geometric graph and (d) its eigenvalue distribution. (b) G2, a Watts–Strogatz small world graph and (e) its eigenvalue distribution. (c) G3, a Barabási–Albert scale-free graph and (f) its eigenvalue distribution.

Graph signals are generated from the ABL prior distribution Equation (1). Moreover, observation noise follows i.i.d. zero-mean Gaussian distribution Equation (6). A variety of hyperparameters are considered, in order to cover ABL graph signals with different bandwidth, strictness of bandlimitedness and energy level, and scenarios with different signal-noise ratio (SNR). SNR here considering random ABL graph signal is defined as

SNRr=10lgk=1N1/(α+γλk2n)N/β. (28)

For each setting, 100 graph signals are generated to evaluate the average performance of each method. The average estimation errors under different sample sizes are computed and plotted.

First of all, we compare the performance of the proposed algorithm with RndEM and OrcUS on G1. Hyperparameters are set to n=2, λc=2, g=1, and SNRr=15dB. As shown in Figure 2, although in the initial stage, the estimation error of the proposed algorithm is similar to RndEM, as the sample size grows, our performance gradually approaches that of OrcUS. This result is as expected, as hyperparameter estimation gets better with growing sample size, which can lead to more efficient sample selection and more accurate signal estimation. Our superiority to RndEM (EM estimation without US sampling) demonstrates the effectiveness of choosing sampling nodes by US. The performance gap between our method and OrcUS (US sampling with known hyperparameters) tends to disappear when samples are abundant, also supporting the validity of hyperparameter estimation by EM.

Figure 2.

Figure 2

Estimation error with growing sample size of the proposed algorithm versus RndEM and OrcUS on G1. Hyperparameters are n=2, λc=2, g=1, and SNRr=15dB.

Subsequently, the proposed algorithm is compared to three existing sampling and estimation methods—Anis2016, Perraudin2018, and Sakiyama2019—in scenarios without much prior knowledge about signal and noise model. A series of experiments with different hyperparameters are done on G1 to test and compare their performance in all kinds of settings. In the first experiment, we fix g=1, SNRr=15dB, and repeat the simulation with n=1,2, λc=2,3, corresponding to ABL graph signals with different bandwidth and strictness of bandlimitedness. Simulation results are depicted in Figure 3. From the graphics we can see, the proposed algorithm can obtain better or competitive estimation performance compared to state-of-the-art methods on ABL graph signals with various properties. For Perraudin2018 and Sakiyama2019, the proposed method significantly outperforms them after a small number of initial samples in all the four settings. As for Anis2016, when n is small or λc is large, which means the signal is far from strictly bandlimited or occupies large bandwidth, we need far fewer samples than Anis2016 to recover the signal with tolerable error. This is because our method estimates the graph signal at all frequencies from the beginning thanks to the Bayesian framework, and our signal prior with hyperparameters estimated from data can better fit ABL graph signals with different properties. Therefore, the proposed algorithm is more robust against different signal properties and is able to obtain better estimation results in most cases.

Figure 3.

Figure 3

Estimation errors of the proposed algorithm versus Anis2016, Perraudin2018, and Sakiyama2019 on G1 with different signal bandwidth and strictness of bandlimitedness, i.e., different n and λc. We fix g=1 and SNRr=15 dB. (a) n=1,λc=2. (b) n=1,λc=3. (c) n=2,λc=2. (d) n=2,λc=3.

We continue to investigate our performance for ABL graph signals with different energy level as well as in scenarios with different signal–noise ratio. We fix n=2, λc=2. When changing g from 0.01 to 100, SNRr is fixed at 15 dB. When changing SNRr from 5 dB to 25 dB, g is fixed at 1. Sampling size is set to M=60. According to Figure 4a, the proposed algorithm gets better performance for graph signals with lower energy. For one thing, even though the estimated variance in the signal prior also becomes larger when the signal energy is larger, the zero-mean Gaussian prior still prefers signals with lower energy. For another, as the variance gets larger, the signal prior becomes flatter, and thus weaker, in the sense that it has less preference among different signals. Therefore, more observations are needed to reach the same estimation accuracy. Fortunately, this problem can be overcome by adaptive rescaling, dividing the observed signal values by a sufficiently large number, as long as we know the rough order of magnitudes of the signal energy. As for SNR, from Figure 4b, although lower SNR may result in worse performance of the proposed algorithm, our method keeps outperforming the others when samples are adequate.

Figure 4.

Figure 4

Estimation errors of the proposed algorithm versus Anis2016, Perraudin2018, and Sakiyama2019 on G1 with different signal energy level and signal–noise ratio, i.e., different g and SNRr. We fix n=2, λc=2. The sampling size is set to M=60. (a) g is changed from 0.01 to 100, while SNRr is fixed at 15 dB. (b) SNRr is changed from 5 dB to 25 dB, while g is fixed at 1.

Then, example simulation results on G2 and G3 are given in Figure 5. n=2, g=1, SNRr=15 dB. λc=3 for G2, and λc=1 for G3. The proposed algorithm also reaches higher estimation accuracy than the other methods with enough sampling budget. This demonstrates the robustness of the proposed algorithm against different graph topologies.

Figure 5.

Figure 5

Estimation errors of the proposed algorithm versus Anis2016, Perraudin2018, and Sakiyama2019 on G2 and G3. (a) Results on G2 with n=2, λc=3, g=1, SNRr=15 dB. (b) Results on G3 with n=2, λc=1, g=1, SNRr=15 dB.

From the simulation results above, without much prior knowledge, the proposed algorithm can adjust to ABL graph signals with different bandwidth, strictness of bandlimitedness, energy level, and SNR. In most cases, exceeding a small number of initial samples, the proposed algorithm is able to estimate the graph signals with higher accuracy than the existing methods.

Besides the validation of the limited-budget performance, we also verify the consistency of the proposed algorithm by a simulation experiment. We generate an ABL graph signal on G1 with n=2, λc=2, and g=1, and the SNR is set to 15dB. To observe the asymptotic performance, the maximum sample size is set to 105, which is 103 times the graph size. We repeat the proposed algorithm on the graph signal for 50 times to see the mean as well as the variance of the estimation errors.

As shown in Figure 6, the mean normalized l2-norm error tends to decrease continuously as the sample size grows, and has decreased to about 5×103 when the sample size reaches 105. The standard deviation also tends to decline with growing sample size, which is around 3×104 with 105 samples. Other ABL graph signals also have similar results. This agrees with the consistency of the proposed algorithm, which has been stated and theoretically proved in Section 4.

Figure 6.

Figure 6

Result of the consistency verification experiment. The graph signal is on G1 generated with n=2, λc=2, g=1, and the observation SNR is 15dB. The curve denotes the mean of the normalized l2-norm error, and the region corresponds to ± 1 standard deviation around the mean. In the overview figure, the y-axis is in logarithmic scale to better display the decrease of the mean estimation error, whereas in the partial enlargement, it is linear to present the decline of the standard deviation.

5.2. A Real-World Example of Temperature Sensor Network

As an application example, we apply the proposed algorithm to the air temperatures measured by weather stations in China [35]. This example shows a potential application of the proposed algorithm in sensor selection.

We construct a K-nearest neighbor (K-NN) graph from the N=381 weather stations with K=5, where each node corresponds to a weather station, and the great-circle distances between stations are considered in K-NN search. Edge weights are assigned via a Gaussian kernel:

wij=expdij22σ2for(i,j)E, (29)

where dij is the great-circle distance between station i and station j in degree, and σ=2.

Then, the air temperatures measured by these weather stations can be seen as a graph signal. We consider two such signals in our experiment: f1 at 0:00 on 1 January 2020, which is a winter night, and f2 at 12:00 on 1 July 2020, which is a summer noon. The two graph signals and their spectra are visualized in Figure 7. As expected, the signals are smooth in the vertex domain and approximately bandlimited in the frequency domain.

Figure 7.

Figure 7

Two graph signals of air temperatures measured by 381 weather stations in China, and their spectra. (a) f1 at 0:00 on 1 January 2020, and (b) its spectrum. (c) f2 at 12:00 on 1 July 2020, and (d) its spectrum.

We apply the proposed sequential sampling and estimation algorithm to these graph signals. Additive i.i.d. zero-mean Gaussian observation noise following Equation (6) is added, where SNRd=10lgfTfN/β is 15 dB. The known hyperparameter n in the signal prior Equation (1) is set to 1/2, which means a smoothness prior Equation (3) is considered. The observed signal values are divided by 100 as preprocessing, and accordingly the estimated graph signal is multiplied by 100 at output. The performance of the proposed algorithm is compared to Anis2016, Perraudin2018 and Sakiyama2019. As our first sampling node is chosen at random, Perraudin2018 is a random sampling method, and the observation noises also introduce randomness, we test each method on the same signal for 20 times, and compare their average estimation errors.

As shown in Figure 8, although the proposed algorithm needs some initial samples to “warm up”, after that the estimation error decreases rapidly and surpass Anis2016, Perraudin2018, and Sakiyama2019 in very short time, and we keep leading as the sample size further grows. This is because our sequential method with unknown hyperparameters requires a certain number of samples to have a relatively proper estimation of the signal and noise distributions. Then, based on the latest signal bandwidth, energy level, and noise precision estimated from data, we are able to choose the subsequent sampling nodes more efficiently and estimate the graph signal more accurately than other methods that take no account of such signal and noise properties.

Figure 8.

Figure 8

Estimation errors of the two temperature graph signals by the proposed algorithm versus Anis2016, Perraudin2018, and Sakiyama2019. (a) Results on f1. (b) Results on f2.

6. Conclusions

In this paper, a sequential sampling and estimation algorithm is proposed for approximately bandlimited graph signals without complete prior knowledge about signal and noise properties. In the proposed algorithm, hyperparameter estimation by EM based on historical observations and sample selection by US with latest estimated hyperparameters are done in an alternating way. We prove in theory the consistency of signal estimation in the proposed algorithm under specific conditions. Finite-time performance of the proposed algorithm is validated by a series of simulation results. The proposed method provides a novel and practical sequential framework for graph signal sampling with unknown hyperparameters in the system model, and the experiment on temperature data shows our potential application in sensor selection on sensor networks with smooth measurements over the topologies. Other potential applications include opinion prediction in social networks assuming that people with strong social connection are likely to hold similar opinions, and active semisupervised learning tasks on similarity graphs.

Appendix A. Concavity of Objective Function in M Step

The objective function in Equation (16) is concave with respect to α,γ,β, as the Hessian matrix

H=2El1α22El1αγ2El1αβ2El1γα2El1γ22El1γβ2El1βα2El1βγ2El1β2=12k=1N1(α+γλk2n)212k=1Nλk2n(α+γλk2n)2012k=1Nλk2n(α+γλk2n)212k=1Nλk4n(α+γλk2n)2000N2β2=12k=1N1(α+γλk2n)21λk2n01λk2n0+00000000N2β2 (A1)

is negative semi-definite for all α,γ,β.

Appendix B. Search Method for Optimization Problem in M Step

To maximize El1 to solve Equation (16), take the partial derivatives with respect to α,γ,β and set them to zero as in Equations (21)–(23).

From Equation (23), the M-step update of β has closed form

βl=MEl1[(yΨf)T(yΨf)], (A2)

which is independent of the update of α,γ.

As for α,γ, as the approximately bandlimited prior Equation (1) requires α,γ>0, if Equations (21) and (22) hold for some α,γ>0, this (α,γ) is the solution for (αl,γl); otherwise, (αl,γl) should be determined using gradient information based on concavity.

Figure A1.

Figure A1

Two cases of M-step update of α,γ. (a) Case I when Δ>0: The solution point (αl,γl) locates on PQ between P and Q. (b) Case II when Δ0: The endpoint Q is the solution for (αl,γl).

Combining Equations (21) and (22) to have αEl1α+γEl1γ=0, we get

αEl1[fTf]+γEl1[fTL2nf]=N, (A3)

which constrains (α,γ) on a straight line in the α-γ plane. Only the segment PQ in the first quadrant is considered as shown in Figure A1 to ensure α,γ>0. The endpoint P has coordinates 0,NEl1[fTL2nf], and the endpoint Q is at NEl1[fTf],0.

Define the gradient of El1 with respect to α,γ as g=El1α,El1γT. Any point A on PQ satisfies g|AOA. The gradient approaching P

g|(α,γ)P=+,0T (A4)

is always rightward as illustrated in Figure A1. Moreover, the gradient at Q

g|Q=0,12ΔT,whereΔ=(L2n)NEl1[fTf]El1[fTL2nf], (A5)

is either upward or downward, depending on the sign of Δ.

If Δ>0, the gradient at Q is upward as in Figure A1a. Let d be the unit vector of direction PQ. Consider the directional derivative

El1d=g,dEl1αEl1[fTL2nf]El1γEl1[fTf]. (A6)

El1d|(α,γ)P>0 and El1d|Q<0 are different in sign. By continuity of El1d, there exists a zero-crossing Z of El1d on PQ between P and Q. Moreover, by concavity of El1, El1d is monotonous along PQ. The zero-crossing Z can be found using, for example, binary search. Notice that g|Z satisfies both g|ZPQ and g|ZOZ, so g|Z=0, i.e., Equations (21) and (22) hold at Z. Therefore, Z is the solution point for (αl,γl).

Otherwise, if Δ0, the gradient at Q is zero or downward as in Figure A1b. By concavity of El1, the endpoint Q is the optimal point in the first quadrant. γl is set to 0, and α is updated by

αl=NEl1[fTf]. (A7)

In this case, p(f;αl,γl)=N(f|0,αl1IN), signal values on different nodes become i.i.d., and only the total energy of the graph signal is restricted.

Appendix C. Proof of Lemma 1

Proof. 

Let {m(r),t}r=1N be the order statistics of {mi,t}iV satisfying m(1),tm(N),t, and denote the sorted vertex indices by {i(r),t}r=1N. Consider ηi,t=mi,tt, and η(r),t=m(r),tt. Define

R=minr|δ>0,lim inftη(r),t>δ1. (A8)

Note that lim inftη(r),t always exists as ηi,t is bounded in [0,1]. Moreover, the set is not empty as η(N),t=maxiVηi,t1N. So 0RN1. Suppose that R1. Let ρt=η(R),t. We are to show that lim inftρtρ_ for some ρ_>0, which contradicts the definition of R. Then we will arrive at an assertion R=0, and lemma 1 is proved.

Define Pt=1tCt=1tAt+βtiVηi,teieiT, where At=αtIN+γtL2n. The sampling strategy Equation (25) can also be rewritten as

st+1=argmaxiVeiTPt1ei. (A9)

Define It={i(r),t|rR+1}, and Itc={i(r),t|rR}. By definition, there exists δ>0 and T<, for all t>T and iIt, it holds ηi,t>δ. As At is positive definite,

t>T,iIt,eiTPt1eieiTβtjVηj,tejejTei=1βtηi,t<1βtδ, (A10)

where (·) denotes the pseudo-inverse of a matrix.

Consider iItc. Define

Qt(1)=βt(1ρt)iIteieiT, (A11)
Qt(2)=1tAt+βtρtiVeieiT. (A12)

As PtQt(1)+Qt(2)=1tAt+βtiIteieiT+ρtiItceieiT,

maxiItceiTPt1eimaxiItceiT(Qt(1)+Qt(2))1ei=maxiItc,zRN2eiTzzT(Qt(1)+Qt(2))zmaxiItc,zN(Qt(1))2eiTzzTQt(2)z, (A13)

where N(Qt(1)) is the null space of Qt(1). Denote the maximum value in the last line, also a lower bound of maxiItceiTPt1ei, as Lt, and the maximizers as it, zt. On one hand, when i is set to it, we have

Lt=maxzN(Qt(1))2eitTzzTQt(2)z. (A14)

By Lagrange multiplier method,

zt=(Qt(2))1(INQt(1)(Qt(1)(Qt(2))1Qt(1))Qt(1)(Qt(2))1)eit, (A15)
Lt=2eitTzt(zt)TQt(2)zt=eitTzt=(zt)TQt(2)zt. (A16)

On the other hand, when z is fixed at zt, we know

eitTzt=maxiItceiTzt. (A17)

We further assert that

eitTzt=maxiItceiTzt. (A18)

Otherwise, there exists jItc such that ejTzt>eitTzt. This will lead to a contradiction 2ejT(zt)(zt)TQt(2)(zt)>Lt. Therefore,

(eitTzt)2=maxiItc(eiTzt)2=maxiItc(zt)TeieiTzt1R(zt)TiItceieiTzt=1R(zt)TiItceieiT+iIteieiTzt=1Rzt2, (A19)

where the last but one line involves ztN(Qt(1)). Therefore,

Lt=eitTzt1Rzt. (A20)

At the same time, according to the Courant–Fischer theorem, along with the positive definiteness of Qt(2),

Lt=(zt)TQt(2)ztλmax(Qt(2))zt2<(Qt(2))zt2, (A21)

where λmax(Qt(2)) denotes the maximum eigenvalue of Qt(2), and (Qt(2)) is the trace. Equations (A20) and (A21) together they give

zt>1R(Qt(2))=1R1t(At)+Nβtρt. (A22)

Substituting it back to (A20),

maxiItceiTPt1eiLt1Rzt>1R1t(At)+Nβtρt. (A23)

Let ρ=δNR, and ρt=ρ1t(At)Nβt. According to the sampling strategy Equation (A9), from previous results Equations (A10) and (A23), for t>T, ρt<ρt implies st+1Itc. Define ρ˜t=1RiItcηi,t.

ρt+1ρ˜t+1=Rtρ˜t+1R(t+1)=ρ˜t+1Rρ˜tR(t+1)>ρ˜t+(NR)δR(t+1)ρtR+(NR)δR(t+1). (A24)

By induction, this lower bound of ρt+τ increases with τ

ρt+τ>ρtR+(NR)δRk=1τ1t+k (A25)

until ρt+τρt+τ. Suppose that ρt1ρt1 but ρt<ρt at some t>T. Then,

ρtt1tρt1t1tρt1. (A26)

If limt1t(At)Nβt=0, then limtρt=ρ. We can obtain lim inftρtlimt(t1)ρt1Rt=ρR=δNR2>0, which contradicts the definition.

Therefore, the only possibility is R=0, and Lemma 1 is proved. □

The authors of Pronzato [28] have proved a similar statement in sequential DOE for nonlinear LS regression. In contrast, the proposed algorithm is an uncertainty sampling process for a Bayesian estimation problem, which makes our proposition and proof different in some way from theirs.

Appendix D. Proof of Theorem 2

Proof. 

Let Uϵ={fRN|iV,|fifi|<ϵ} be a “cubic” neighborhood around the true signal f, and its complement Uϵc=RNUϵ. We are to prove that P(Uϵc|Ψ1:t,y1:t;αt,γt, βt) converges to 0 with high probability for arbitrary ϵ>0.

Here, for brevity, γt is absorbed into αt. We can get a fractional form of the target

P(Uϵc|Ψ1:t,y1:t;αt,βt)=NtDt=Uϵcp(y1:t|Ψ1:t,f;βt)p(y1:t|Ψ1:t,f;β)p(f;αt)dfRNp(y1:t|Ψ1:t,f;βt)p(y1:t|Ψ1:t,f;β)p(f;αt)df, (A27)

where f is the true signal, and β is the true noise precision.

First look at the numerator. According to Corollary 2, as t, mi,t. We have

p(y1:t|Ψ1:t,f;βt)p(y1:t|Ψ1:t,f;β)=(a)expi=1Nmi,t1mi,tj=1mi,tlnp(yi,j|fi;β)p(yi,j|fi;βt)(b)expi=1Nmi,tKLp(yi|fi;β)||p(yi|fi;βt), (A28)

where yi,j denotes the j-th observation of vertex i; KL(p||q) is the Kullback–Leibler divergence between two distributions; (a) takes into account the independence of different observations in y conditioned on Ψ, f, and β; and (b) obeys the Law of Large Numbers.

For any fUϵc, there exists kV, such that |fkfk|>ϵ. Regardless of the value of βt,

KLp(yk|fk;β)||p(yk|fk;βt)=12βt(fkfk)2+βtβlnβtβ1>12ln(1+βϵ2). (A29)

For ik, KLp(yi|fi;β)||p(yi|fi;βt)0 by definition. Moreover, from Corollary 1, there exists δ>0, such that mk,tt>δ for all large t. Thus, for ϵ1=12δln(1+βϵ2)>0,

p(y1:t|Ψ1:t,f;βt)p(y1:t|Ψ1:t,f;β)<exp12mk,tln(1+βϵ2)<exp(tϵ1) (A30)

for all fUϵ for sufficiently large t with high probability.

Therefore, when t becomes large enough, no matter what values αt have, the numerator

Nt=Uϵcp(y1:t|Ψ1:t,f;βt)p(y1:t|Ψ1:t,f;β)p(f;αt)df<exp(tϵ1) (A31)

with high probability.

Then, come to the denominator. As αt, βt are estimated by MML,

DtD˜t(α,β)=RNp(y1:t|Ψ1:t,f;β)p(y1:t|Ψ1:t,f;β)p(f;α)df, (A32)

where α is allowed to be chosen arbitrarily.

Consider arbitrary ϵ2>0. Analogous to above, as t and mi,t, we have

exp(tϵ2)p(y1:t|Ψ1:t,f;β)p(y1:t|Ψ1:t,f;β)=expi=1Nmi,tϵ21mi,tj=1mi,tlnp(yi,j|fi;β)p(yi,j|fi;β)expi=1Nmi,tϵ2KLp(yi|fi;β)||p(yi|fi;β). (A33)

Define Xϵ2/2=fRN|iV,KLp(yi|fi;β)||p(yi|fi;β)<ϵ22. For all fXϵ2/2, when t grows large enough,

exp(tϵ2)p(y1:t|Ψ1:t,f;β)p(y1:t|Ψ1:t,f;β)>exptϵ22 (A34)

with high probability.

Note that the term inside integration is positive. We shrink the range of integration and obtain

exp(tϵ2)D˜t(α,β)>exp(tϵ2)Xϵ2/2p(y1:t|Ψ1:t,f;β)p(y1:t|Ψ1:t,f;β)p(f;α)df>exptϵ22Xϵ2/2p(f;α)df. (A35)

We can always choose an α such that Xϵ2/2p(f;α)df>0 since p(f;α) is Gaussian. Then, we get exp(tϵ2)D˜t(α,β) as t. Therefore, the denominator

DtD˜t(α,β)>exp(tϵ2) (A36)

for all large t with high probability, where ϵ2 can be any positive number.

Pick 0<ϵ2<ϵ1. For ϵ3=ϵ1ϵ2>0,

P(Uϵc|Ψ1:t,y1:t;αt,βt)=NtDt<exp(tϵ3) (A37)

for all sufficiently large t with high probability.

We finally reach a conclusion that P(Uϵc|Ψ1:t,y1:t;αt,βt) converges to 0 with high probability as t for any positive ϵ. □

The proposed algorithm is more complex than classical empirical Bayes method discussed in Petrone et al. [27], because in our setting, hyperparameters exist not only in signal prior but also in observation model, and i.i.d. complete observations are substituted by partial ones selected by strategy. That is why a more complicated proof of consistent signal estimation in our algorithm is provided here.

Author Contributions

Conceptualization, S.L., H.F., and B.H.; methodology, S.L. and H.F.; software, S.L.; validation, S.L., K.X., and H.F.; formal analysis, S.L. and H.F.; investigation, S.L. and H.F.; resources, H.F. and B.H.; data curation, S.L.; writing—original draft preparation, S.L.; writing—review and editing, K.X. and H.F.; visualization, S.L.; supervision, H.F. and B.H.; project administration, H.F., B.H.; funding acquisition, H.F. and B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Shanghai Municipal Natural Science Foundation (No. 19ZR1404700), Fudan University-CIOMP Joint Fund (FC2019-003), and 2020 Okawa Foundation Research Grant.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Shuman D.I., Narang S.K., Frossard P., Ortega A., Vandergheynst P. The emerging field of signal processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains. IEEE Signal Process. Mag. 2013;30:83–98. doi: 10.1109/MSP.2012.2235192. [DOI] [Google Scholar]
  • 2.Sandryhaila A., Moura J.M.F. Discrete signal processing on graphs. IEEE Trans. Signal Process. 2013;61:1644–1656. doi: 10.1109/TSP.2013.2238935. [DOI] [Google Scholar]
  • 3.Shuman D.I., Vandergheynst P., Frossard P. Chebyshev polynomial approximation for distributed signal processing; Proceedings of the 2011 International Conference on Distributed Computing in Sensor Systems and Workshops (DCOSS); Barcelona, Spain. 27–29 June 2011. [Google Scholar]
  • 4.Shi X., Feng H., Zhai M., Yang T., Hu B. Infinite impulse response graph filters in wireless sensor networks. IEEE Signal Process. Lett. 2015;22:1113–1117. doi: 10.1109/LSP.2014.2387204. [DOI] [Google Scholar]
  • 5.Loukas A., Simonetto A., Leus G. Distributed autoregressive moving average graph filters. IEEE Signal Process. Lett. 2015;22:1931–1935. doi: 10.1109/LSP.2015.2448655. [DOI] [Google Scholar]
  • 6.Chen S., Varma R., Sandryhaila A., Kovačević J. Discrete signal processing on graphs: Sampling theory. IEEE Trans. Signal Process. 2015;63:6510–6523. doi: 10.1109/TSP.2015.2469645. [DOI] [Google Scholar]
  • 7.Anis A., Gadde A., Ortega A. Efficient sampling set selection for bandlimited graph signals using graph spectral proxies. IEEE Trans. Signal Process. 2016;64:3775–3789. doi: 10.1109/TSP.2016.2546233. [DOI] [Google Scholar]
  • 8.Narang S.K., Gadde A., Ortega A. Signal processing techniques for interpolation in graph structured data; Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); Vancouver, BC, Canada. 26–31 May 2013; pp. 5445–5449. [Google Scholar]
  • 9.Sakiyama A., Tanaka Y., Tanaka T., Ortega A. Efficient sensor position selection using graph signal sampling theory; Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); Shanghai, China. 20–25 March 2016; pp. 6225–6229. [Google Scholar]
  • 10.Jabłoński I. Graph signal processing in applications to sensor networks, smart grids, and smart cities. IEEE Sens. J. 2017;17:7659–7666. doi: 10.1109/JSEN.2017.2733767. [DOI] [Google Scholar]
  • 11.Huang W., Bolton T.A.W., Medaglia J.D., Bassett D.S., Ribeiro A., Ville D.V.D. A graph signal processing perspective on functional brain imaging. Proc. IEEE. 2018;106:868–885. doi: 10.1109/JPROC.2018.2798928. [DOI] [Google Scholar]
  • 12.Cheung G., Magli E., Tanaka Y., Ng M.K. Graph spectral image processing. Proc. IEEE. 2018;106:907–930. doi: 10.1109/JPROC.2018.2799702. [DOI] [Google Scholar]
  • 13.Dinesh C., Cheung G., Bajić I.V. Point cloud denoising via feature graph Laplacian regularization. IEEE Trans. Image Process. 2020;29:4143–4158. doi: 10.1109/TIP.2020.2969052. [DOI] [PubMed] [Google Scholar]
  • 14.Gadde A., Anis A., Ortega A. Active semi-supervised learning using sampling theory for graph signals; Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; New York, NY, USA. 24–27 August 2014; pp. 492–501. [Google Scholar]
  • 15.Anis A., El Gamal A., Avestimehr A.S., Ortega A. A sampling theory perspective of graph-based semi-supervised learning. IEEE Trans. Inf. Theory. 2019;65:2322–2342. doi: 10.1109/TIT.2018.2879897. [DOI] [Google Scholar]
  • 16.Kipf T.N., Welling M. Semi-supervised classification with graph convolutional networks; Proceedings of the 5th International Conference on Learning Representations (ICLR), Palais des Congrès Neptune; Toulon, France. 24–26 April 2017. [Google Scholar]
  • 17.Chen S., Varma R., Singh A., Kovačević J. Signal recovery on graphs: Fundamental limits of sampling strategies. IEEE Trans. Signal Inf. Process. Netw. 2016;2:539–554. doi: 10.1109/TSIPN.2016.2614903. [DOI] [Google Scholar]
  • 18.Puy G., Tremblay N., Gribonval R., Vandergheynst P. Random sampling of bandlimited signals on graphs. Appl. Comput. Harmon. Anal. 2018;44:446–475. doi: 10.1016/j.acha.2016.05.005. [DOI] [Google Scholar]
  • 19.Xie X., Feng H., Jia J., Hu B. Design of sampling set for bandlimited graph signal estimation; Proceedings of the 2017 IEEE Global Conference on Signal and Information Processing (GlobalSIP); Montreal, QC, Canada. 14–16 November 2017; pp. 653–657. [Google Scholar]
  • 20.Lin S., Xie X., Feng H., Hu B. Active sampling for approximately bandlimited graph signals; Proceedings of the 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); Brighton, UK. 12–17 May 2019; pp. 5441–5445. [Google Scholar]
  • 21.Perraudin N., Ricaud B., Shuman D.I., Vandergheynst P. Global and local uncertainty principles for signals on graphs. APSIPA Trans. Signal Inf. Process. 2018;7:e3. doi: 10.1017/ATSIP.2018.2. [DOI] [Google Scholar]
  • 22.Xie X., Yu J., Feng H., Hu B. Bayesian design of sampling set for bandlimited graph signals; Proceedings of the 2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP); Ottawa, ON, Canada. 11–14 November 2019. [Google Scholar]
  • 23.Gadde A., Ortega A. A probabilistic interpretation of sampling theory of graph signals; Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); South Brisbane, Australia. 19–24 April 2015; pp. 3257–3261. [Google Scholar]
  • 24.Bishop C.M. Pattern Recognition and Machine Learning. Springer; New York, NY, USA: 2006. [Google Scholar]
  • 25.Settles B. Active Learning. Morgan & Claypool; San Rafael, CA, USA: 2012. [Google Scholar]
  • 26.Bernardo J.M., Smith A.F.M. Bayesian Theory. John Wiley & Sons; Hoboken, NJ, USA: 1994. [Google Scholar]
  • 27.Petrone S., Rousseau J., Scricciolo C. Bayes and empirical Bayes: Do they merge? Biometrika. 2014;101:285–302. doi: 10.1093/biomet/ast067. [DOI] [Google Scholar]
  • 28.Pronzato L. One-step ahead adaptive D-optimal design on a finite design space is asymptotically optimal. Metrika. 2010;71:219–238. doi: 10.1007/s00184-008-0227-y. [DOI] [Google Scholar]
  • 29.Oppenheim A.V., Schafer R.W. Discrete-Time Signal Processing. 3rd ed. Prentice Hall; Englewood Cliffs, NJ, USA: 2009. [Google Scholar]
  • 30.Kay S.M. Fundamentals of Statistical Signal Processing, Volume I: Estimation Theory. Prentice Hall; Englewood Cliffs, NJ, USA: 1993. [Google Scholar]
  • 31.Sakiyama A., Tanaka Y., Tanaka T., Ortega A. Eigendecomposition-free sampling set selection for graph signals. IEEE Trans. Signal Process. 2019;67:2679–2692. doi: 10.1109/TSP.2019.2908129. [DOI] [Google Scholar]
  • 32.Watts D.J., Strogatz S.H. Collective dynamics of ‘small-world’ networks. Nature. 1998;393:440–442. doi: 10.1038/30918. [DOI] [PubMed] [Google Scholar]
  • 33.Barabási A.L., Albert R. Emergence of scaling in random networks. Science. 1999;286:509–512. doi: 10.1126/science.286.5439.509. [DOI] [PubMed] [Google Scholar]
  • 34.Perraudin N., Paratte J., Shuman D., Martin L., Kalofolias V., Vandergheynst P., Hammond D.K. GSPBOX: A toolbox for signal processing on graphs. arXiv. 20161408.5781v2 [cs.IT] [Google Scholar]
  • 35.Integrated Surface Database (ISD) by National Climatic Data Center (NCDC) of the USA. [(accessed on 11 December 2020)]; Available online: https://www.ncdc.noaa.gov/isd.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES