Skip to main content
Springer logoLink to Springer
. 2015 Nov 7;26(6):1187–1211. doi: 10.1007/s11222-015-9604-3

Stability of noisy Metropolis–Hastings

F J Medina-Aguayo 1,, A Lee 1, G O Roberts 1
PMCID: PMC6991990  PMID: 32055107

Abstract

Pseudo-marginal Markov chain Monte Carlo methods for sampling from intractable distributions have gained recent interest and have been theoretically studied in considerable depth. Their main appeal is that they are exact, in the sense that they target marginally the correct invariant distribution. However, the pseudo-marginal Markov chain can exhibit poor mixing and slow convergence towards its target. As an alternative, a subtly different Markov chain can be simulated, where better mixing is possible but the exactness property is sacrificed. This is the noisy algorithm, initially conceptualised as Monte Carlo within Metropolis, which has also been studied but to a lesser extent. The present article provides a further characterisation of the noisy algorithm, with a focus on fundamental stability properties like positive recurrence and geometric ergodicity. Sufficient conditions for inheriting geometric ergodicity from a standard Metropolis–Hastings chain are given, as well as convergence of the invariant distribution towards the true target distribution.

Electronic supplementary material

The online version of this article (doi:10.1007/s11222-015-9604-3) contains supplementary material, which is available to authorized users.

Keywords: Markov chain Monte Carlo, Pseudo-marginal Monte Carlo , Monte Carlo within Metropolis, Intractable likelihoods, Geometric ergodicity

Introduction

Intractable target densities and the pseudo-marginal algorithm

Suppose our aim is to simulate from an intractable probability distribution π for some random variable X, which takes values in a measurable space X,B(X). In addition, let π have a density π(x) with respect to some reference measure μ(dx), e.g. the counting or the Lebesgue measure. By intractable we mean that an analytical expression for the density π(x) is not available and so implementation of a Markov chain Monte Carlo (MCMC) method targeting π is not straightforward.

One possible solution to this problem is to target a different distribution on the extended space X×W,B(X)×B(W), which admits π as marginal distribution. The pseudo-marginal algorithm (Beaumont 2003; Andrieu and Roberts 2009) falls into this category since it is a Metropolis–Hastings (MH) algorithm targeting a distribution π¯N, associated to the random vector (XW) defined on the product space X×W,B(X)×B(W) where WR0+:=[0,). It is given by

π¯N(dx,dw):=π(dx)Qx,N(dw)w, 1

where Qx,N(x,N)X×N+ is a family of probability distributions on W,B(W) satisfying for each (x,N)X×N

EQx,NWx,N1,forWx,NQx,N(·). 2

Throughout this article, we restrict our attention to the case where for each xX, Wx,N is Qx,N-a.s. strictly positive, for reasons that will become clear.

The random variables Wx,Nx,N are commonly referred as the weights. Formalising this algorithm using (1) and (2) was introduced by Andrieu and Vihola (2015), and “exactness” follows immediately: π¯ admits π as a marginal. Given a proposal kernel q:X×B(X)[0,1], the respective proposal of the pseudo-marginal is given by

q¯N(x,w;dy,du):=q(x,dy)Qy,N(du),

and, consequently, the acceptance probability can be expressed as

α¯N(x,w;y,u):=min1,π(dy)uq(y,dx)π(dx)wq(x,dy). 3

The pseudo-marginal algorithm defines a time-homogeneous Markov chain, with transition kernel P¯N on the measurable space X×W,B(X)×B(W). A single draw from P¯N(x,w;·,·) is presented in Algorithm 1.graphic file with name 11222_2015_9604_Figa_HTML.jpg

Due to its exactness and straightforward implementation in many settings, the pseudo-marginal has gained recent interest and has been theoretically studied in some depth, see e.g. Andrieu and Roberts (2009), Andrieu and Vihola (2014, 2015), Doucet et al. (2015), Girolami et al. (2013), Maire et al. (2014) and Sherlock et al. (2015). These studies typically compare the pseudo-marginal Markov chain with a “marginal” Markov chain, arising in the case where all the weights are almost surely equal to 1, and (3) is then the standard Metropolis–Hastings acceptance probability associated with the target density π and the proposal q.

Examples of pseudo-marginal algorithms

A common source of intractability for π occurs when a latent variable Z on (Z,B(Z)) is used to model observed data, as in hidden Markov models (HMMs) or mixture models. Although the density π(x) cannot be computed, it can be approximated via importance sampling, using an appropriate auxiliary distribution, say νx. Here, appropriate means πxνx, where πx denotes the conditional distribution of Z given X=x. Therefore, for this setting, the weights are given by

Wx,N=1Nk=1NπxZx(k)νxZx(k),whereZx(k)k{1,,N}i.i.d.νx(·),

which motivates the following generic form when using averages of unbiased estimators

Wx,N=1Nk=1NWx(k),whereWx(k)ki.i.d.Qx(·),EQxWx(k)1. 4

It is clear that (4) describes only a special case of (2). Nevertheless, we will pay special attention to the former throughout the article. For similar settings to (4) see Andrieu and Roberts (2009).

Since (2) is more general, it allows Wx,N to be any random variable with expectation 1. Sequential Monte Carlo (SMC) methods involve the simulation of a system of some number of particles, and provide unbiased estimates of likelihoods associated with HMMs (see Del Moral 2004, Proposition 7.4.1 or Pitt et al. 2012) irrespective of the size of the particle system. Consider the model given by Fig. 1. The random variables Xtt=0T form a time-homogeneous Markov chain with transition fθ(·|xt-1) that depends on a set of parameters θ. The observed random variables Ytt=1T are conditionally independent given the unobserved Xtt=1T and are distributed according to gθ(·|xt), which also may depend on θ. The likelihood function for θ is given by

l(θ;y1,,yT):=Efθt=1Tgθ(yt|Xt),

where Efθ denotes expectation w.r.t. the θ-dependent law of {Xt}t=1T, and we assume for simplicity that the initial value X0=x0 is known. If we denote by l^N(θ;y1,,yT) the unbiased SMC estimator of l(θ;y1,,yT) based on N particles, we can then define

Wθ,N:=l^N(θ;y1,,yT)l(θ;y1,,yT),

and (2) is satisfied but (4) is not. The resulting pseudo-marginal algorithm is developed and discussed in detail in Andrieu et al. (2010), where it and related algorithms are referred to as particle MCMC methods.

Fig. 1.

Fig. 1

Hidden Markov model

The noisy algorithm

Although the pseudo-marginal has the desirable property of exactness, it can suffer from “sticky” behaviour, exhibiting poor mixing and slow convergence towards the target distribution (Andrieu and Roberts 2009; Lee and Łatuszyński 2014). The cause for this is well-known to be related with the value of the ratio between Wy,N and Wx,N at a particular iteration. Heuristically, when the value of the current weight (w in (3)) is large, proposed moves can have a low probability of acceptance. As a consequence, the resulting chain can get “stuck” and may not move after a considerable number of iterations.

In order to overcome this issue, a subtly different algorithm is performed in some practical problems (see, e.g., McKinley et al. 2014). The basic idea is to refresh, independently from the past, the value of the current weight at every iteration. The ratio of the weights between Wy,N and Wx,N still plays an important role in this alternative algorithm, but here refreshing Wx,N at every iteration can improve mixing and the rate of convergence.

This alternative algorithm is commonly known as Monte Carlo within Metropolis (MCWM), as in O’Neill et al. (2000), Beaumont (2003) or Andrieu and Roberts (2009), since typically the weights are Monte Carlo estimates as in (4). From this point onwards it will be referred as the noisy MH algorithm or simply the noisy algorithm to emphasize that our main assumption is (2). Due to independence from previous iterations while sampling Wx,N and Wy,N, the noisy algorithm also defines a time-homogeneous Markov chain with transition kernel P~N, but on the measurable space (X,B(X)). A single draw from P~N(x,·) is presented in Algorithm 2, and it is clear that we restrict our attention to strictly positive weights because the algorithm is not well-defined when both Wy,N and Wx,N are equal to 0.graphic file with name 11222_2015_9604_Figb_HTML.jpg

Even though these algorithms differ only slightly, the related chains have very different properties. In Algorithm 2, the value w is generated at every iteration whereas in Algorithm 1, it is treated as an input. As a consequence, Algorithm 1 produces a chain on X×W,B(X)×B(W) contrasting with a chain from Algorithm 2 taking values on X,B(X). However, the noisy chain is not invariant under π and it is not reversible in general. Moreover, it may not even have an invariant distribution as shown by some examples in Sect. 2.

From O’Neill et al. (2000) and Fernández-Villaverde and Rubio-Ramírez (2007), it is evident that the implementation of the noisy algorithm goes back even before the appearance of the pseudo-marginal, the latter initially conceptualised as Grouped Independence Metropolis–Hastings (GIMH) in Beaumont (2003). Theoretical properties, however, of the noisy algorithm have mainly been studied in tandem with the pseudo-marginal by Beaumont (2003), Andrieu and Roberts (2009) and more recently by Alquier et al. (2014).

The noisy chain generated by Algorithm 2 can be seen as a perturbed version of an idealised Markov chain where the weights Wx,Nx,N are all equal to one. Perturbed Markov chains have been investigated in , e.g., Roberts et al. (1998), Breyer et al. (2001), Shardlow and Stuart (2000), Mitrophanov (2005), Ferré et al. (2013). More recently Pillai and Smith (2014) and Rudolf and Schweizer (2015) study such chains using the notion of Wasserstein distance. We focus on total variation distance, a particular case of the Wasserstein distance. The relationship between our work and these latter papers is pointed out in subsequent remarks.

Objectives of the article

The objectives of this article can be illustrated using a simple example. Let N(·|μ,σ2) denote a univariate Gaussian distribution with mean μ and variance σ2 and π(·)=N(·|0,1) be a standard normal distribution. Let the weights Wx,N be as in (4) with

Qx(·)=logN·|-12σ2,σ2andσ2:=5,

where logN(·|μ,σ2) denotes a log-normal distribution of parameters μ and σ2. In addition, let the proposal q be random walk given by q(x,·)=N·|x,4. For this example, Fig. 2 shows the estimated densities using the noisy chain for different values of N. It appears that the noisy chain has an invariant distribution, and as N increases it seems to approach the desired target π. Our objectives here are to answer the following types of questions about the noisy algorithm in general:

  1. Does an invariant distribution exist, at least for N large enough?

  2. Does the noisy Markov chain behave like the marginal chain for sufficiently large N?

  3. Does the invariant distribution, if it exists, converge to π as N increases?

We will see that the answer to the first two questions is negative in general. However, all three questions can be answered positively when the marginal chain is geometrically ergodic and the distributions of the weights satisfy additional assumptions.

Fig. 2.

Fig. 2

Estimated densities using the noisy chain with 100, 000 iterations for N=10 (left), N=100 (central) and N=1000 (right)

Marginal chains and geometric ergodicity

In order to formalise our analysis, let P denote the Markov transition kernel of a standard MH chain on X,B(X), targeting π with proposal q. We will refer to this chain and this algorithm using the term marginal (as in Andrieu and Roberts 2009; Andrieu and Vihola 2015), which is the idealised version for which the noisy chain and corresponding algorithm are simple approximations. Therefore

P(x,dy):=α(x,y)q(x,dy)+δx(dy)ρ(x),

where α is the MH acceptance probability and ρ is the rejection probability, given by

α(x,y):=min1,π(dy)q(y,dx)π(dx)q(x,dy)andρ(x):=1-Xα(x,y)q(x,dy). 5

Similarly, for the transition kernel P~N of the noisy chain, moves are proposed according to q but are accepted using α¯N (as in (3)) instead of α, once values for Wx,N and Wy,N are sampled. In order to distinguish the acceptance probabilities between the noisy and the pseudo-marginal processes, despite being the same after sampling values for the weights, define

α~N(x,y):=EQx,NQy,Nα¯N(x,Wx,N;y,Wy,N). 6

Here α~N is the expectation of a randomised acceptance probability, which permits defining the transition kernel of the noisy chain by

P~N(x,dy):=α~N(x,y)q(x,dy)+δx(dy)ρ~N(x),

where ρ~N is the noisy rejection probability given by

ρ~N(x):=1-Xα~N(x,y)q(x,dy). 7

As briefly noted before, the noisy kernel P~N is just a perturbed version of P involving a ratio of weights in the noisy acceptance probability α~N. When such weights are identically one, i.e. Qx,N({1})=1, the noisy chain reduces to the marginal chain, whereas the pseudo-marginal becomes the marginal chain with an extra component always equal to 1.

So far, the terms slow convergence and “sticky” behaviour have been used in a relative vague sense. A powerful characterisation of the behaviour of a Markov chain is provided by geometric ergodicity, defined below. Geometrically ergodic Markov chains have a limiting invariant probability distribution, which they converge towards geometrically fast in total variation (Meyn and Tweedie 2009). For any Markov kernel K:X×B(X)[0,1], let Kn be the n-step transition kernel, which is given by

Kn(x,·):=XKn-1(x,dz)K(z,·),forn2.

Definition 1.1

(Geometric ergodicity) A φ-irreducible and aperiodic Markov chain Φ:=(Φi)i0 on a measurable space X,B(X), with transition kernel P and invariant distribution π, is geometrically ergodic if there exists a finite function V1 and constants τ<1, R< such that

Pn(x,·)-π(·)TVRV(x)τn,forxX. 8

Here, ·TV denotes the total variation norm given by

μTV=12sup|g|1|μ(dy)g(y)|=supAB(X)μ(A),

where μ is any signed measure.

Geometric ergodicity does not necessarily provide fast convergence in an absolute sense. For instance, consider cases where τ, or R, from Definition 1.1 are extremely close to one, or very large respectively. Then the decay of the total variation distance, though geometric, is not particularly fast (see Roberts and Rosenthal 2004 for some examples).

Nevertheless, geometric ergodicity is a useful tool when analysing non-reversible Markov chains as will become apparent in the noisy chain case. Moreover, in practice one is often interested in estimating Eπf(X) for some function f:XR, which is done by using ergodic averages of the form

en,m(f):=1ni=m+1m+nfΦi,form,n0.

In this case, geometric ergodicity is a desirable property since it can guarantee the existence of a central limit theorem (CLT) for en,m(f), see Chan and Geyer (1994) and Roberts and Rosenthal (1997, 2004) for a more general review. Also, its importance is related with the construction of consistent estimators of the corresponding asymptotic variance in the CLT, as in Flegal and Jones (2010).

As noted in Andrieu and Roberts (2009), if the weights Wx,Nx,N are not essentially bounded then the pseudo-marginal chain cannot be geometrically ergodic; in such cases the “stickiness” may be more evident. In addition, under mild assumptions (in particular, that P¯N has a left spectral gap), from Andrieu and Vihola (2015, Proposition 10) and Lee and Łatuszyński (2014), a sufficient but not necessary condition ensuring the pseudo-marginal inherits geometric ergodicity from the marginal, is that the weights are uniformly bounded. This certainly imposes a tight restriction in many practical problems.

The analyses in Andrieu and Roberts (2009) and Alquier et al. (2014) mainly study the noisy algorithm in the case where the marginal Markov chain is uniformly ergodic, i.e. when it satisfies (8) with supxXV(x)<. However, there are many Metropolis–Hastings Markov chains for statistical estimation that cannot be uniformly ergodic, e.g. random walk Metropolis chains when π is not compactly supported. Our focus is therefore on inheritance of geometric ergodicity by the noisy chain, complementing existing results for the pseudo-marginal chain.

Outline of the paper

In Sect. 2, some simple examples are presented for which the noisy chain is positive recurrent, so it has an invariant probability distribution. This is perhaps the weakest stability property that one would expect a Monte Carlo Markov chain to have. However, other fairly surprising examples are presented for which the noisy Markov chain is transient even though the marginal and pseudo-marginal chains are geometrically ergodic. Section 3 is dedicated to inheritance of geometric ergodicity from the marginal chain, where two different sets of sufficient conditions are given and are further analysed in the context of arithmetic averages given by (4). Once geometric ergodicity is attained, it guarantees the existence of an invariant distribution π~N for the noisy chain. Under the same sets of conditions, we show in Sect. 4 that π~N and π can be made arbitrarily close in total variation as N increases. Moreover, explicit rates of convergence are possible to obtain in principle, when the weights arise from an arithmetic average setting as in (4).

Motivating examples

Homogeneous weights with a random walk proposal

Assume a log-concave target distribution π on the positive integers, whose density with respect to the counting measure is given by

π(m)exp-h(m)1mN+,

where h:N+R is a convex function. In addition, let the proposal distribution be a symmetric random walk on the integers, i.e.

q(m,{m+1})=12=q(m,{m-1}),formZ. 9

From Mengersen and Tweedie (1996), it can be seen that the marginal chain is geometrically ergodic.

Now, assume the distribution of the weights Wm,Nm,N is homogeneous with respect to the state space, meaning

Wm,N=WNQN(·),for allmN+. 10

In addition, assume WN>0QN-a.s., then for m2

P~N(m,m-1)=12EQNQNmin1,exp{h(m)}exp{h(m-1)}×WN(1)WN(2)andP~N(m,m+1)=12EQNQNmin1,exp{h(m)}exp{h(m+1)}×WN(1)WN(2),whereWN(k)k{1,2}i.i.d.QN(·).

For this particular class of weights and using the fact that h is convex, the noisy chain is geometrically ergodic, implying the existence of an invariant probability distribution.

Proposition 2.1

Consider a log-concave target density on the positive integers and a proposal density as in (9). In addition, let the distribution of the weights be homogeneous as in (10). Then, the chain generated by the noisy kernel P~N is geometrically ergodic.

It is worth noting that the distribution of the weights, though homogeneous with respect to the state space, can be taken arbitrarily, as long as the weights are positive. Homogeneity ensures that the distribution of the ratio of such weights is not concentrated near 0, due to its symmetry around one, i.e. for z>0

PQNQNWN(1)WN(2)z=PQNQNWN(1)WN(2)1z,whereWN(k)k{1,2}i.i.d.QN(·).

In contrast, when the support of the distribution QN is unbounded, the corresponding pseudo-marginal chain cannot be geometrically ergodic.

Particle MCMC

More complex examples arise when using particle MCMC methods, for which noisy versions can also be performed. They may prove to be useful in some inference problems. Consider again the hidden Markov model given by Fig. 1. As before, set X0=x0 and let

θ=x0,a,σX2,σY2,fθ·|Xt-1=N·|aXt-1,σX2andgθ·|Xt=N·|Xt,σY2.

Therefore, once a prior distribution for θ is specified, p(·) say, the aim is to conduct Bayesian inference on the posterior distribution

π(θ|y1,,yT)p(θ)l(θ;y1,,yT).

In this particular setting, the posterior distribution is tractable. This will allows us to compare the results obtained from the exact and noisy versions, both relying on the SMC estimator l^N(θ;y1,,yT) of the likelihood. Using a uniform prior for the parameters and a random walk proposal, Fig. 3 shows the run and autocorrelation function (acf) for the autoregressive parameter a of the marginal chain. Similarly, Fig. 4 shows the corresponding run and acf for both the pseudo-marginal and the noisy chain when N=250. Plots for the other parameters and different values of N can be found in Online Appendix 2. It is noticeable how the pseudo-marginal gets “stuck”, resulting in a lower acceptance than the marginal and noisy chains. In addition, the acf of the noisy chain seems to decay faster than that of the pseudo-marginal chain.

Fig. 3.

Fig. 3

Last 20,000 iterations of the marginal algorithm for the autoregressive parameter a (top). Estimated autocorrelation function of the corresponding marginal chain (bottom). The mean acceptance probability was 0.256

Fig. 4.

Fig. 4

Last 20,000 iterations of the pseudo-marginal (top left) and noisy (bottom left) algorithms, for the autoregressive parameter a when N=250. Estimated autocorrelation functions of the corresponding pseudo-marginal (top right) and noisy (bottom right) chains. The mean acceptance probabilities were 0.104 for the pseudo-marginal and 0.283 for the noisy chain

Finally, Figs. 5 and 6 show the estimated posterior densities for the parameters when N=250 and N=750, respectively. There, the trade-off between the pseudo-marginal and the noisy algorithm is noticeable. For lower values of N, the pseudo-marginal will require more iterations due to the slow mixing, whereas the noisy converges faster towards an unknown noisy invariant distribution. By increasing N, the mixing in the pseudo-marginal improves and the noisy invariant approaches the true posterior. Plots for other values of N can also be found in Online Appendix 2.

Fig. 5.

Fig. 5

Estimated densities using the marginal, pseudo-marginal and noisy chains for the four parameters when N=250. Vertical lines indicate the real values

Fig. 6.

Fig. 6

Estimated densities using the marginal, pseudo-marginal and noisy chains for the four parameters, when N=750. Vertical lines indicate the real values

Transient noisy chain with homogeneous weights

In contrast with example in Sect. 2.1, this one shows that the noisy algorithm can produce a transient chain even in simple settings. Let π be a geometric distribution on the positive integers, whose density with respect to the counting measure is given by

π(m)=12m1mN+. 11

In addition, assume the proposal distribution is a simple random walk on the integers, i.e.

q(m,{m+1})=θ=1-q(m,{m-1}),formZ. 12

where θ(0,1). Under these assumptions, the marginal chain is geometrically ergodic, see Proposition 5.1 in Appendix 1.

Consider N=1 and as in Sect. 2.1, let the distribution of weights be homogeneous and given by

W=(b-ε)Ber(s)+ε,forb>1andε(0,1), 13

where Ber(s) denotes a Bernoulli random variable of parameter s(0,1). There exists a relationship between s, b and ε that guarantees the expectation of the weights is identically one. The following proposition, proven in Appendix 1 by taking θ>1/2, shows that the resulting noisy chain can be transient for certain values of b, ϵ and θ.

Proposition 2.2

Consider a geometric target density as in (11) and a proposal density as in (12). In addition, let the weights when N=1 be given by (13). Then, for some b, ε and θ the chain generated by the noisy kernel P~N=1 is transient.

In contrast, since the weights are uniformly bounded by b, the pseudo-marginal chain inherits geometric ergodicity for any θ, b and ϵ. The left plot in Fig. 7 shows an example. We will discuss the behaviour of this example as N increases in Sect. 3.4 .

Fig. 7.

Fig. 7

Runs of the marginal, pseudo-marginal and noisy chains. Left plot shows example in Sect. 2.3, where θ=0.75, ε=2-3 and b=2εθ1-θ. Central and right plots show example in Sect. 2.4, where θ=0.5 and θ=0.25 respectively, with εm=m-(3-m(mod3)) and b=3+1-θθ3

Transient noisy chain with non-homogeneous weights

One could argue that the transient behaviour of the previous example is related to the large value of θ in the proposal distribution. However, as shown here, for any value of θ(0,1) one can construct weights satisfying (2) for which the noisy chain is transient. With the same assumptions as in the example in Sect. 2.3, except that now the distribution of weights is not homogeneous but given by

Wm,1=(b-εm)Ber(sm)+εm,forb>1andεm=m-(3-(m(mod3))), 14

the noisy chain will be transient for b large enough. The proof can be found in Appendix 1.

Proposition 2.3

Consider a geometric target density as in (11) and a proposal density as in (12). In addition, let the weights when N=1 be given by (14). Then, for any θ(0,1) there exists some b>1 such that the chain generated by the noisy kernel P~N=1 is transient.

The reason for this becomes apparent when looking at the behaviour of the ratios of weights. Even though εm0 as m, the non-monotonic behaviour of the sequence implies

εm-1εmOm2Om-1ifm(mod3)=0,m(mod3){1,2},

and

εm+1εmOm-2Omifm(mod3)=2,m(mod3){0,1}.

Hence, the ratio of the weights can become arbitrarily large or arbitrarily close to zero with a non-negligible probability. This allows the algorithm to accept moves to the right more often, if m is large enough. Once again, the pseudo-marginal chain inherits the geometrically ergodic property from the marginal. See the central and right plots of Fig. 7 for two examples using different proposals. Again, we will come back to this example in Sect. 3.4, where we look at the behaviour of the associated noisy chain as N increases.

Inheritance of ergodic properties

The inheritance of various ergodic properties of the marginal chain by pseudo-marginal Markov chains has been established using techniques that are powerful but suitable only for reversible Markov chains (see, e.g. Andrieu and Vihola 2015). Since the noisy Markov chains treated here can be non-reversible, a suitable tool for establishing geometric ergodicity is the use of Foster–Lyapunov functions, via geometric drift towards a small set.

Definition 3.1

(Small set) Let P be the transition kernel of a Markov chain Φ. A subset CX is small if there exists a positive integer n0, ε>0 and a probability measure ν(·) on X,B(X) such that the following minorisation condition holds

Pn0(x,·)εν(·),forxC. 15

The following theorem, which is immediate from combining Roberts and Rosenthal (1997, Proposition 2.1) and Meyn and Tweedie (2009, Theorem 15.0.1), establishes the equivalence between geometric ergodicity and a geometric drift condition. For any kernel K:X×B(X)[0,1], let

KV(x):=XK(x,dz)V(z).

Theorem 3.1

Suppose that Φ is a ϕ-irreducible and aperiodic Markov chain with transition kernel P and invariant distribution π. Then, the following statements are equivalent:

  • (i)
    There exists a small set C, constants λ<1 and b<, and a function V1 finite for some x0X satisfying the geometric drift condition
    PV(x)λV(x)+b1xC,forxX. 16
  • (ii)

    The chain is π-a.e. geometrically ergodic, meaning that for π-a.e. xX it satisfies (8) for some V1 (which can be taken as in (i)) and constants τ<1, R<.

From this point onwards, it is assumed that the marginal and noisy chains are ϕ-irreducible and aperiodic. In addition, for many of the following results, it is required that

(P1)

The marginal chain is geometrically ergodic, implying its kernel P satisfies the geometric drift condition in (16) for some constants λ<1 and b<, some function V1 and a small set CX.

Conditions involving a negative moment

From the examples of the previous section, it is clear that the weights play a fundamental role in the behaviour of the noisy chain. The following theorem states that the noisy chain will inherit geometric ergodicity from the marginal under some conditions on the weights involving a strengthened version of the Law of Large Numbers and convergence of negative moments.

(W1)
For any δ>0, the weights Wx,Nx,N satisfy
limNsupxXPQx,N|Wx,N-1|δ=0.
(W2)
The weights Wx,Nx,N satisfy
limNsupxXEQx,NWx,N-1=1.

Theorem 3.2

Assume (P1), (W1) and (W2). Then, there exists N0N+ such that for all NN0, the noisy chain with transition kernel P~N is geometrically ergodic.

The above result is obtained by controlling the dissimilarity of the marginal and noisy kernels. This is done by looking at the corresponding rejection and acceptance probabilities. The proofs of the following lemmas appear in Appendix 1.

Lemma 3.1

For any δ>0

PQz,NQx,NWz,NWx,N1-δ2supxXPQx,N|Wx,N-1|δ2.

Lemma 3.2

Let ρ(x) and ρ~N(x) be the rejection probabilities as defined in (5) and (7) respectively. Then, for any δ>0

ρ~N(x)-ρ(x)δ+2supxXPQx,N|Wx,N-1|δ2.

Lemma 3.3

Let α(x,y) and α~N(x,y) be the acceptance probabilities as defined in (5) and (6) respectively. Then,

α~N(x,y)α(x,y)EQx,NWx,N-1.

Notice that (W1) and (W2) allow control on the bounds in the above lemmas. While Lemma 3.2 provides a bound for the difference of the rejection probabilities, Lemma 3.3 gives one for the ratio of the acceptance probabilities. The proof of Theorem 3.2 is now presented.

Proof of Theorem 3.2

Since the marginal chain P is geometrically ergodic, it satisfies the geometric drift condition in (16) for some λ<1, b<, some function V1 and a small set CX. Now, using the above lemmas

P~NV(x)-PV(x)=Xq(x,dz)α~N(x,z)-α(x,z)V(z)+V(x)ρ~N(x)-ρ(x)supxXEWx,N-1-1PV(x)+δ+2supxXP|Wx,N-1|δ2V(x).

By (W1) and (W2), for any ε, δ>0 there exists N0N+ such that

supxXP|Wx,N-1|δ2<ε4andsupxXEWx,N-1-1<ε,

whenever NN0, implying

P~NV(x)PV(x)+εPV(x)+δ+ε2V(x)λ+ελ+δ+ε2V(x)+b1+ε1xC.

Taking δ=ε2 and ε0,1-λ1+λ, the noisy chain P~N also satisfies a geometric drift condition for the same function V and small set C, completing the proof.

Remark 3.1

In fact, (W1) and (W2) together guarantee for any δ>0

δα~N(x,y)-α(x,y)α(x,y)δ,

which is the crucial assumption in Pillai and Smith (2014, Lemma 3.6) for obtaining a similar drift condition.

Conditions on the proposal distribution

In this subsection a different bound for the acceptance probabilities is provided, which allows dropping assumption (W2) but imposes a different one on the proposal q instead.

(P1)
(P1) holds and for the same drift function V in (P1) there exists K< such that the proposal kernel q satisfies
qV(x)KV(x),forxX.

Theorem 3.3

Assume (P1*) and (W1). Then, there exists N0N+ such that for all NN0, the noisy chain with transition kernel P~N is geometrically ergodic.

In order to prove Theorem 3.3 the following lemma is required. Its proof can be found in Appendix 1. In contrast with Lemma 3.3, this lemma provides a bound for the additive difference of the noisy and marginal acceptance probabilities.

Lemma 3.4

Let α(x,y) and α~N(x,y) be the acceptance probabilities as defined in (5) and (6), respectively. Then, for any η>0

α~N(x,y)-α(x,y)η+2supxXPQx,N|Wx,N-1|η21+η.

Proof of Theorem 3.3

Using Lemmas 3.2 and 3.4 with η=δ

P~NV(x)-PV(x)=Xq(x,dz)α~N(x,z)-α(x,z)V(z)+V(x)ρ~N(x)-ρ(x)δ+2supxXP|Wx,N-1|δ21+δqV(x)+δ+2supxXP|Wx,N-1|δ2V(x)δ+2supxXP|Wx,N-1|δ21+δ×qV(x)+V(x).

By (W1), there exists N0N+ such that

supxXP|Wx,N-1|δ21+δ<ε4,

whenever NN0. This implies

P~NV(x)PV(x)+δ+ε2qV(x)+V(x),

and using (P1*)

P~NV(x)λ+δ+ε2K+1V(x)+b1xC.

Taking δ=ε2 and ε0,1-λ1+K, the noisy chain P~N also satisfies a geometric drift condition for the same function V and small set C, completing the proof.

Remark 3.2

By itself, (W1) implies for any δ>0

|α~N(x,y)-α(x,y)|δ,

but it needs to be paired with (P1*) to obtain the desired result. These assumptions are comparable to those in Pillai and Smith (2014, Lemma 3.6), taking f constant therein. Additionally, (W1) and (P1*) imply the required conditions on E and λ in Rudolf and Schweizer (2015, Corollary 31), where a similar result is proved in terms of V-uniform ergodicity.

In general, assumption (P1*) may be difficult to verify as one must identify a particular function V, but it is easily satisfied when restricting to log-Lipschitz targets and when using a random walk proposal of the form

q(x,dy)=q(y-x)dy, 17

where · denotes the usual Euclidean distance. To see this the following assumption is required, which is a particular case of (P1) and is satisfied under some extra technical conditions (see, e.g., Roberts and Tweedie 1996).

(P1)
XRd. The target π is log-Lipschitz, meaning that for some L>0
|logπ(z)-logπ(x)|Lz-x.
(P1) holds taking the drift function V=π-s, for any s(0,1). The proposal q is a random walk as in (17) satisfying
Rdexpauq(u)du<,
for some a>0.

See Appendix 1 for a proof of the following proposition.

Proposition 3.1

Assume (P1) and (W1). Then, (P1*) holds.

Conditions for arithmetic averages

In the particular setting where the weights are given by (4), sufficient conditions on these can be obtained to ensure geometric ergodicity is inherited by the noisy chain. For the simple case where the weights are homogeneous with respect to the state space (W1) is automatically satisfied. In order to attain (W2), the existence of a negative moment for a single weight is required. See Appendix 1 for a proof of the following result.

Proposition 3.2

Assume weights as in (4). If  EQxWx-1< then

limNEQx,NWx,N-1=1. 18

For homogeneous weights, (18) implies (W2). When the weights are not homogeneous, stronger conditions are needed for (W1) and (W2) to be satisfied. An appropriate first assumption is that the weights are uniformly integrable.

(W3)
The weights Wxx satisfy
limKsupxXEQxWx1{Wx>K}=0.

The second condition imposes an additional assumption on the distribution of the weights Wxx near 0.

(W4)
There exists γ(0,1) and constants M<, β>0 such that for w(0,γ) the weights Wxx satisfy
supxXPQxWxwMwβ.

These new conditions ensure (W1) and (W2) are satisfied.

Proposition 3.3

For weights as in (4),

  • (i)

    (W3) implies (W1);

  • (ii)

    (W1) and (W4) imply (W2).

The following corollary is obtained as an immediate consequence of the above proposition, Theorems 3.2 and 3.3.

Corollary 3.1

Let the weights be as in (4). Assume (W3) and either

  • (i)

    (P1) and (W4);

  • (ii)

    (P1*).

Then, there exists N0N+ such that for all NN0, the noisy chain with transition kernel P~N is geometrically ergodic.

The proof of Proposition 3.3 follows the statement of Lemma 3.5, whose proof can be found in Appendix 1. This lemma allows us to characterise the distribution of Wx,N near 0 assuming (W4) and also provides conditions for the existence and convergence of negative moments.

Lemma 3.5

Let γ(0,1) and p>0.

  • (i)
    Suppose Z is a positive random variable, and assume that for z(0,γ)
    PZzMzα,whereα>p,M<.
    Then,
    EZ-p1γp+pMγα-pα-p.
  • (ii)
    Suppose Zii=1N is a collection of positive and independent random variables, and assume that for each i1,,N and z(0,γ)
    PZizMizαi,whereαi>0,Mi<.
    Then, for z(0,γ)
    Pi=1NZizi=1NMizi=1Nαi.
  • (iii)
    Let the weights be as in (4). If for some N0N+
    EQx,N0Wx,N0-p<,
    then for any NN0
    EQx,N+1Wx,N+1-pEQx,NWx,N-p.
  • (iv)
    Assume (W1) and let g:R+R be a function that is continuous at 1 and bounded on the interval [γ,). Then
    limNsupxXEQx,N|gWx,N-g1|1Wx,Nγ=0.

Proof of Proposition 3.3

Part (i) is a consequence of Chandra (1989, Theorem 1). Assuming (W3), it implies

limNsupxXE|Wx,N-1|=0.

By Markov’s inequality

E|Wx,N-1|δP|Wx,N-1|δ,

and the result follows.

To prove (ii), assume (W4) and by part (ii) of Lemma 3.5, for w(0,γ)

PNWx,NwMNwNβ.

Take p>1 and define N0:=pβ+1, then using part (i) of Lemma 3.5 if NN0

supxXEWx,N-pNγp+pNMNγNβ-pNβ-p.

Hence, by Hölder’s inequality

E|Wx,N-1-1|1Wx,N(0,γ)EWx,N-11Wx,N(0,γ)EWx,N-p1pPWx,N<γp-1p,

and applying part (iii) of Lemma 3.5, for NN0

E|Wx,N-1-1|1Wx,N(0,γ)EWx,N0-p1pPWx,N<γp-1p.

Therefore,

supxXE|Wx,N-1-1|1Wx,N(0,γ)supxXEWx,N0-p1psupxXPWx,N<γp-1p.

Since γ<1 and by (W1)

limNsupxXPWx,N<γ=0,

implying

limNsupxXE|Wx,N-1-1|1Wx,N(0,γ)=0. 19

Now, for fixed γ(0,1) the function g(x)=x-1 is bounded and continuous on [γ,), implying by part (iv) of Lemma 3.5

limNsupxXE|Wx,N-1-1|1Wx,N[γ,)=0. 20

Finally, using (19) and (20)

limNsupxXE|Wx,N-1-1|=0,

and by the triangle inequality

supxXE|Wx,N-1-1|supxXEWx,N-1-1,

the result follows.

Remarks on results

Equipped with these results, we return to the examples in Sects. 2.3 and 2.4. Even though the noisy chain can be transient in these examples, the behaviour is quite different when considering weights that are arithmetic averages of the form in (4). Since in both examples the weights are uniformly bounded by the constant b, they immediately satisfy (W1). Additionally, by Proposition 3.2, condition (W2) is satisfied for the example in Sect. 2.3. This is not the case for example in Sect. 2.4, but condition (P1*) is satisfied by taking V=π-12. Therefore, applying Theorems 3.2 and 3.3 to examples in Sects. 2.3 and 2.4 respectively, as N increases the corresponding chains will go from being transient to geometrically ergodic.

Despite conditions (W1) and (W2) guaranteeing the inheritance of geometric ergodicity for the noisy chain, they are not necessary. Consider a modification of the example in Sect. 2.3, where the weights are given by

Wm,1=(bm-εm)Ber(sm)+εm,wherebm>1andεm(0,1]for allm1.

Again, there exists a relationship between the variables bm, εm and sm for ensuring the expectation of the weights is equal to one. Let BinN,s denote a binomial distribution of parameters NN+ and s(0,1). Then, in the arithmetic average context, Wm,N becomes

Wm,N=bm-εmNBinN,sm+εm,wherebm>1andεm(0,1]for allm1. 21

For particular choices of the sequences bmmN+ and εmmN+, the resulting noisy chain can be geometrically ergodic for all N1, even though neither (W1) nor (W2) hold.

Proposition 3.4

Consider a geometric target density as in (11) and a proposal density as in (12). In addition, let the weights be as in (21) with bm, εm0 as m and

limmεm-1εm=l,wherelR++.

Then, the chain generated by the noisy kernel P~N is geometrically ergodic for any NN+.

Finally, in many of the previous examples, increasing the value of N seems to improve the ergodic properties of the noisy chain. However, the geometric ergodicity property is not always inherited, no matter how large N is taken. The following proposition shows an example rather similar to Proposition 3.4, but in which the ratio εm-1εm does not converge as m.

Proposition 3.5

Consider a geometric target density as in (11) and a proposal density as in (12). In addition, let the weights be as in (21) with bm=m and

εm=m-(3-(m(mod3))).

Then, the chain generated by the noisy kernel P~N is transient for any NN+.

Convergence of the noisy invariant distribution

So far the only concern has been whether the noisy chain inherits the geometric ergodicity property from the marginal chain. As an immediate consequence, geometric ergodicity guarantees the existence of an invariant probability distribution π~N for P~N, provided N is large enough. In addition, using the same conditions from Sect. 3, we can characterise and in some cases quantify the convergence in total variation of π~N towards the desired target π, as N.

Convergence in total variation

The following definition, taken from Roberts et al. (1998), characterises a class of kernels satisfying a geometric drift condition as in (16) for the same V, C, λ and b.

Definition 4.1

(Simultaneous geometric ergodicity) A class of Markov chain kernels PkkK is simultaneously geometrically ergodic if there exists a class of probability measures νkkK, a measurable set CX, a real valued measurable function V1, a positive integer n0 and positive constants ε, λ, b such that for each kK:

  • (i)

    C is small for Pk, with Pkn0(x,·)ενk(·) for all xC;

  • (ii)

    the chain Pk satisfies the geometric drift condition in (16) with drift function V, small set C and constants λ and b.

Provided N is large, the noisy kernels {P~N+k}k0 together with the marginal P will be simultaneous geometrically ergodic. This will allow the use of coupling arguments for ensuring π~N and π get arbitrarily close in total variation. The main additional assumption is

(P2)
For some ε>0, some probability measure ν(·) on X,B(X) and some subset CX, the marginal acceptance probability α and the proposal kernel q satisfy
α(x,y)q(x,dy)εν(dy),forxC.

Remark 4.1

(P2) ensures the marginal chain satisfies the minorisation condition in (15), purely attained by the sub-kernel α(x,y)q(x,dy). This occurs under fairly mild assumptions (see, e.g., Roberts and Tweedie 1996, Theorem 2.2).

Theorem 4.1

Assume (P1), (P2), (W1) and (W2). Alternatively, assume (P1*), (P2) and (W1). Then,

  • (i)

    there exists N0N+ such that the class of kernels P,P~N0,P~N0+1, is simultaneously geometrically ergodic;

  • (ii)

    for all xX, limNP~N(x,·)-P(x,·)TV=0;

  • (iii)

    limNπ~N(·)-π(·)TV=0.

Part (iii) of the above theorem is mainly a consequence of Roberts et al. (1998, Theorem 9) when parts (i) and (ii) hold. Indeed, by the triangle inequality,

π~N(·)-π(·)TVP~Nn(x,·)-π~N(·)TV+Pn(x,·)-π(·)TV+P~Nn(x,·)-Pn(x,·)TV. 22

Provided NN0, the first two terms in (22) can be made arbitrarily small by increasing n. In addition, due to the simultaneous geometrically ergodic property, the first term in (22) is uniformly controlled regardless the value of N. Finally, using an inductive argument, part (ii) implies that for all xX and all nN+

limNP~Nn(x,·)-Pn(x,·)TV=0.

Proof of Theorem 4.1

From the proofs of Theorems 3.2 and 3.3, there exists N1N+ such that the class of kernels P,P~N1,P~N1+1, satisfies condition (ii) in Definition 4.1 for the same function V, small set C and constants λN1,bN1. Respecting (i), for any δ(0,1)

P~N(x,A)Aα~N(x,z)q(x,dz)AEmin1,Wz,NWx,Nα(x,z)q(x,dz)(1-δ)×A1-PWz,NWx,N1-δα(x,z)q(x,dz).

Then, by Lemma 3.1

P~N(x,A)(1-δ)1-2supxXP|Wx,N-1|δ2×Aα(x,z)q(x,dz).

By (W1), there exists N2N+ such that for NN2

supxXP|Wx,N-1|δ2δ2,

giving

P~N(x,A)(1-δ)2Aα(x,z)q(x,dz).

Due to (P2),

P~N(x,A)(1-δ)2εν(A),forxC.

Finally, take N0=maxN1,N2 implying (i).

To prove (ii) apply Lemma 3.2 and Lemma 3.4 to get

supAB(X)P~N(x,A)-P(x,A)η+2supxXPQx,N|Wx,N-1|η21+η×supAB(X)q(x,A)+ρN~(x)-ρ(x)supAB(X)1xAη+2supxXPQx,N|Wx,N-1|η21+η+δ+2supxXPQx,N|Wx,N-1|δ2 23

Finally, taking N and by (W1)

limNsupAB(X)P~N(x,A)-P(x,A)η+δ.

The result follows since η and δ can be taken arbitrarily small.

For (iii), see Theorem 9 in Roberts et al. (1998) for a detailed proof.

Remark 4.2

A Wasserstein distance variant of part (iii) in Theorem 4.1 has been proved in Rudolf and Schweizer (2015, Corollary 28), in which control of the difference between α~N and α is still required and can be obtained using (W1).

Rate of convergence

Let (Φ~nN)n0 denote the noisy chain and Φnn0 the marginal chain, which move according to the kernels P~N and P, respectively and define cx:=1-P~N(x,·)-P(x,·)TV. Using notions of maximal coupling for random variables defined on a Polish space (see Lindvall 2002 and Thorisson 2013), there exists a probability measure νx(·) such that

P(x,·)cxνx(·)andP~N(x,·)cxνx(·).

Let c:=infxXcx, define a coupling in the following way

  • If Φ~n-1N=Φn-1=y, with probability c draw Φnνy(·) and set Φ~nN=Φn. Otherwise, draw independently ΦnR(y,·) and Φ~nNR~N(y,·), where
    R(y,·):=1-c-1P(y,·)-cνy(·)andR~N(y,·):=1-c-1P~N(y,·)-cνy(·).
  • If Φ~n-1NΦn-1, draw independently ΦnP(y,·) and Φ~nNP~N(y,·).

Since

PΦ~nNΦn|Φ~0N=Φ0=xPΦ~nNΦn|Φ~n-1N=Φn-1,Φ~0N=Φ0=x+PΦ~n-1NΦn-1|Φ~0N=Φ0=x1-c+PΦ~n-1NΦn-1|Φ~0N=Φ0=x,

and noting

PΦ~1NΦ1|Φ~0N=Φ0=xsupxXP~N(x,·)-P(x,·)TV=1-c,

an induction argument can be applied to obtain

P[Φ~nNΦn|Φ~0N=Φ0=x]nsupxXP~N(x,·)-P(x,·)TV.

Therefore, using the coupling inequality, the third term in (22) can be bounded by

P~Nn(x,·)-Pn(x,·)TVPΦ~nNΦn|Φ~0N=Φ0=xnsupxXP~N(x,·)-P(x,·)TV. 24

On the other hand, using the simultaneous geometric ergodicity of the kernels and provided N is large enough, the noisy and marginal kernels will each satisfy a geometric drift condition as in (16) with a common drift function V1, small set C and constants λ,b. Therefore, by Theorem 3.1, there exist R>0, and τ<1 such that

P~Nn(x,·)-π~N(·)TVRV(x)τnandPn(x,·)-π(·)TVRV(x)τn. 25

Explicit values for R and τ are in principle possible, as done in Rosenthal (1995) and Meyn and Tweedie (1994). For simplicity assume infxXV(x)=1, then combining (24) and (25) in (22), for all nN+

π~N(·)-π(·)TV2Rτn+nsupxXP~N(x,·)-P(x,·)TV. 26

So, if an analytic expression in terms of N is available for the second term on the right hand side of (26), it will be possible to obtain an explicit rate of convergence for π~N and π.

Theorem 4.2

Assume (P1), (P2), (W1) and (W2). Alternatively, assume (P1*), (P2) and (W1). In addition, suppose

supxXP~N(x,·)-P(x,·)TV1r(N),

where r:N+R+ and limNr(N)=+. Then, there exists D>0 and N0N+ such that for all NN0,

π~N(·)-π(·)TVDlogr(N)r(N). 27

Proof

Let R>0,τ(0,1) and r>0. Pick r large enough, such that

log2Rrlogτ-11,

then the convex function f:[1,)R+ where

f(s)=2Rτs+sr,

is minimised at

s=log2Rrlogτ-1logτ-1.

Restricting the domain of f to the positive integers and due to convexity, it is then minimised at either

n1=sorn2=s.

In any case

minf(n1),f(n2)f(s+1),=1r1+τ+log2Rrlogτ-1logτ-1.

Finally take N large enough such that

log2Rr(N)logτ-11,

and from (26)

π~N(·)-π(·)TVminf(n1),f(n2)1r(N)1+τ+log2Rr(N)logτ-1logτ-1=Ologr(N)r(N),

obtaining the result.

Remark 4.3

A general result bounding the total variation between the law of a Markov chain and a perturbed version is presented in Rudolf and Schweizer (2015, Theorem 21). This is done using the connection between the V-norm distance and the Wasserstein distance introduced in Hairer and Mattingly (2011). With such a result, and considering the same assumptions in Theorem 4.2, one could in principle obtain an explicit value for D in (27).

Moreover, when the weights are expressed in terms of arithmetic averages as in (4), an explicit expression for r(N) can be obtained whenever there exists a uniformly bounded moment. This is a slightly stronger assumption than (W3).

(W5)
There exists k>0, such that the weights Wxx satisfy
supxXEQxWx1+k<.

Proposition 4.1

Assume (P1), (P2), (W4) and (W5). Alternatively, assume (P1*), (P2) and (W5). Then, there exists Dk>0 and N0N+ such that for all NN0,

π~N(·)-π(·)TVDklogNN1-22+k.

If in addition (W5) holds for all k>0, then for any ε(0,1) there will exist Dε>0 and N0N+ such that for all NN0,

π~N(·)-π(·)TVDεlogNN1-ε.

Discussion

In this article, fundamental stability properties of the noisy algorithm have been explored. The noisy Markov kernels considered are perturbed Metropolis–Hastings kernels defined by a collection of state-dependent distributions for non-negative weights all with expectation 1. The general results do not assume a specific form for these weights, which can be simple arithmetic averages or more complex random variables. The former may arise when unbiased importance sampling estimates of a target density are used, while the latter may arise when such densities are estimated unbiasedly using a particle filter.

Two different sets of sufficient conditions were provided under which the noisy chain inherits geometric ergodicity from the marginal chain. The first pair of conditions, (W1) and (W2), involve a stronger version of the Law of Large Numbers for the weights and uniform convergence of the first negative moment, respectively. For the second set, (W1) is still required but (W2) can be replaced with (P1*), which imposes a condition on the proposal distribution. These conditions also imply simultaneous geometric ergodicity of a sequence of noisy Markov kernels together with the marginal Markov kernel, which then ensures that the noisy invariant π~N converges to π in total variation as N increases. Moreover, an explicit bound for the rate of convergence between π~N and π is possible whenever an explicit bound (that is uniform in x) is available for the convergence between P~N(x,·) and P(x,·).

When weights are arithmetic averages as in (4), specific conditions were given for inheriting geometric ergodicity from the corresponding marginal chain. The uniform integrability condition in (W3) ensures that (W1) is satisfied, whereas (W4) is essential for satisfying (W2). Regarding the noisy invariant distribution π~N, (W5), which is slightly stronger than (W3), leads to an explicit bound on the rate of convergence of this distribution to π.

The noisy algorithm remains undefined when the weights have positive probability of being zero. If both weights were zero one could accept the move, reject the move or keep sampling new weights until one of them is not zero. Each of these lead to different behaviour.

As seen in the examples of Sect. 3.4, the behaviour of the ratio of the weights (at least in the tails of the target) plays an important role in the ergodic properties of the noisy chain. In this context, it seems plausible to obtain geometric noisy chains, even when the marginal is not, if the ratio of the weights decays sufficiently fast to zero in the tails. Another interesting possibility, that may lead to future research, is to relax the condition on the expectation of the weights to be identically one.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Acknowledgments

All authors would like to thank the EPSRC-funded Centre for Research in Statistical methodology (EP/D002060/1). The first author was also supported by Consejo Nacional de Ciencia y Tecnología. The third author’s research was also supported by the EPSRC Programme Grant i-like (EP/K014463/1).

Appendix 1: Proofs

On state-dependent random walks

The following proposition for state-dependent Markov chains on the positive integers will be useful for addressing some proofs. See Norris (1999) for a proof of parts (i) and (ii), for part (iii) see Callaert and Keilson (1973), which is proved within the birth-death process context.

Proposition 5.1

Suppose we have a random walk Φ on N+ with transition kernel P. Define for m1

pm:=P(m,m+1)andqm:=P(m,m-1),

with q1=0,p1(0,1] and pm,qm>0,pm+qm1 for all m2. The resulting chain is:

  • (i)
    recurrent if and only if
    m=2i=2mqipi;
  • (ii)
    positive recurrent if and only if
    m=2i=2mpi-1qi<;
  • (iii)
    geometrically ergodic if
    limmpm<limmqm.
Remark 5.1

Notice that (iii) is not an if and only if statement and that it implies (ii). Additionally, if the chain is not state-dependent, (ii) implies (iii).

Section 2

Proof of Proposition 2.1

Since h is convex

h(m)-h(m-1)h(m-1)andh(m)-h(m+1)-h(m),

implying

P~N(m,m-1)P~N(m,m+1)Emin1,exp{h(m-1)}WN(1)WN(2)Emin1,exp{-h(m)}WN(1)WN(2).

Define Z:=WN(1)WN(2), and since π(m)0 it is true that

log(k):=limmh(m)>0, 28

hence

limmP~N(m,m-1)P~N(m,m+1)Emin1,kZEmin1,k-1Z. 29

If k=+, it is clear that the limit in (29) diverges, consequently the noisy chain is geometrically ergodic according to Proposition 5.1. If k<, the noisy chain will be geometrically ergodic if

Emin1,kZ>Emin1,k-1Z,

which can be translated to

kEZ1{Zk-1}+PZ>k-1>k-1EZ1{Z<k}+PZk,

or equivalently to

kPk-1<Z<k+k2-1EZ1Zk-1>EZ1k-1<Z<k. 30

Now consider two cases, first if Pk-1<Z<k>0 then it is clear that

Ek-Z1k-1<Z<k>0,

which satisfies (30). Finally, if Pk-1<Z<k=0 then

PZk-1=12=PZk,

implying from (28)

k2-1EZ1Zk-1>0,

and leading to (30).

Proof of Proposition 2.2

For simplicity the subscript N is dropped. In this case,

Qm,1=Q=L(b-ε)Ber(s)+ε,

and the condition EQW=1 implies

s=1-εb-ε. 31

Let θ11+2ε,1 and set

b=ε2θ1-θ, 32

this implies α¯(m,w;m-1,u)1 and

α¯(m,w;m+1,u)=1-θ2θ11-θ2θ2ifu=wu=b,w=εu=ε,w=b.

Therefore, for m2,

α~(m,m-1)=1andα~(m,m+1)=1-θ2θs2+1-s2+1+1-θ2θ2s1-s.

Consequently, P~(m,m-1)=1-θ and

P~(m,m+1)=θ1-θ2θs2+1-s2+1+1-θ2θ2s1-s>θs(1-s).

From Proposition 5.1, if

P~(m,m+1)>P~(m,m-1),

then the noisy chain will be transient. For this to happen, it is enough to pick θ and s such that

θs(1-s)-(1-θ)0.

Let s=ε, then from (31) and (32)

θ=(1-ε+ε2)1-ε+3ε2=1-2ε21-ε+3ε2, 33

and if ε2-3 then

θs(1-s)-(1-θ)=ε1-ε+3ε2(1-ε+ε2)(1-ε)-2εε1-ε+3ε2(1-ε)2-2ε=ε1-ε+3ε2(2-ε)2-30.

Hence, for ε0,2-3 and setting s=ε, θ as in (33) and b as in (32), the resulting noisy chain is transient.

Proof of Proposition 2.3

For simplicity the subscript N is dropped. In this case,

Qm,1=Qm=L(b-εm)Ber(sm)+εm,

and the condition EQmWm=1 implies

sm=1-εmb-εm.

Then, for m large enough

α~(m,m-1)=Eα¯(m,Wm;m-1,Wm-1)=min1,2θ1-θsm-1sm+sm-1(1-sm)+(1-sm-1)(1-sm)1m(mod3)=0+Om-1andα~(m,m-1)=Eα¯(m,Wm;m-1,Wm-1)=min1,1-θ2θsmsm+1+(1-sm)sm+1+(1-sm)(1-sm+1)1m(mod3)2+Om-1.

Define

cm:=P~(m,m-1)P~(m,m+1)=(1-θ)α~(m,m-1)θα~(m,m+1).

Since sm1b as m,

c0,:=limkc3k=1-θθ×min1,2θ1-θ-11b2+1b+1-1b2min1,1-θ2θ-11b2+1b+1-1b21-θθ11-1b=1-θθbb-1=:l0,c1,:=limkc3kc3k+1=c0,1-θθ×min1,2θ1-θ-11b2+1bmin1,1-θ2θ-11b2+1b+1-1b2l01-θθ1b1-1b=1-θθ2b(b-1)2=:l1

and

limkc3kc3k+1c3k+2=c1,1-θθmin1,2θ1-θ-11b2+1bmin1,1-θ2θ-11b2+1bl11-θθ1b1b1-1b=1-θθ3b2b-13=:l2.

Therefore, for any δ>0 there exists k0N+, such that whenever kk0+1

K:=j=k0k-1c3jc3j+1c3j+2<(l2+δ)k-k0,

implying

Kc3k<(l2+δ)k-k0(l0+δ),Kc3kc3k+1<(l2+δ)k-k0(l1+δ)andKc3kc3k+1c3k+2<(l2+δ)k-k0(l2+δ).

Hence, for i{0,1,2} and some C>0

j=23k+icj<C(l2+δ)k.

Let am:=j=2mcj, then a sufficient condition for the series m=2am to converge, implying a transient chain according to Proposition 5.1, is l2<1. This is the case for b3+1-θθ3, since

1-l2=1-1-θθ3b2b-13=b2b-13b-13b2-1-θθ3=b2b-13b-3+3b-1b2-1-θθ3>b2b-13b-3-1-θθ30.

Hence, the resulting noisy chain is transient if b3+1-θθ3, for any θ(0,1).

Section 3

Proof of Lemma 3.1

For any δ>0

PWz,NWx,N1-δPWx,N1+δ2+PWz,N1-δ2P|Wx,N-1|δ2+P|Wz,N-1|δ22supxXPQx,N|Wx,N-1|δ2.

Proof of Lemma 3.2

Using the inequality

min1,abmin1,amin1,b,fora,b0,

and applying Markov’s inequality with δ>0,

ρ~N(x)=1-Xq(x,dz)α~N(x,z)1-Xq(x,dz)α(x,z)Emin1,Wz,NWx,N1-1-δ×Xq(x,dz)α(x,z)Pmin1,Wz,NWx,N>1-δ=1-1-δXq(x,dz)α(x,z)+1-δ×Xq(x,dz)α(x,z)PWz,NWx,N1-δ1-1-δ1-ρ(x)+Xq(x,dz)α(x,z)PWz,NWx,N1-δ.

Finally, using Lemma 3.1

ρ~N(x)ρ(x)+δ1-ρ(x)+2supxXP|Wx,N-1|δ21-ρ(x)ρ(x)+δ+2supxXP|Wx,N-1|δ2.

Proof of Lemma 3.3

For the first claim apply Jensen’s inequality and the fact that

min1,abmin1,ab,fora0andb1,

hence

α~N(x,z)min1,π(z)q(z,x)π(x)q(x,z)EWz,NWx,Nα(x,z)EWx,N-1EWz,N.

Proof of Lemma 3.4

Using the inequality

min1,abmin1,ab,fora0andb1,α~N(x,z)=Eα¯N(x,Wx,N;z,Wz,N)1Wz,NWx,N<1+η+Eα¯N(x,Wx,N;z,Wz,N)1Wz,NWx,N1+ηα(x,z)1+ηPWz,NWx,N<1+η+PWz,NWx,N1+ηα(x,z)+η+PWz,NWx,N1+η,

Notice that

PWz,NWx,N1+η=PWx,NWz,N11+η,

then applying Lemma 3.1 taking δ=η1+η.

α~N(x,z)α(x,z)+η+2supxXP|Wx,N-1|η21+η.

Proof of Proposition 3.1

Taking V=π-s, where 0<s<min1,aL,

qV(x)V(x)=XV(z)V(x)q(x,dz)=Xπ(x)π(z)sq(x,dz)Rdexpaz-xq(z-x)dz.

Finally, using the transformation u=z-x,

qV(x)V(x)Rdexpauq(u)du,

which implies (P1*).

Proof of Proposition 3.2

By properties of the arithmetic and harmonic means

1Nk=1N1Wx(k)-Nk=1NWx(k)0,

which implies, by Jensen’s inequality,

E1Nk=1N1Wx(k)-Nk=1NWx(k)EWx-1-1.

Then, using Fatou’s lemma and the law of large numbers

EWx-1-1lim supNE1Nk=1N1Wx(k)-Wx,N-1lim infNE1Nk=1N1Wx(k)-Wx,N-1Elim infN1Nk=1N1Wx(k)-lim supNWx,N-1EWx-1-1,

hence

limNE1Nk=1N1Wx(k)-Wx,N-1=EWx-1-1. 34

Finally, since

E1Nk=1N1Wx(k)-Wx,N-1=EWx-1-EWx,N-1,

the expression in (34) becomes

limNEWx,N-1=1.

Proof of Lemma 3.5

The proof of (i) is motivated by Piegorsch and Casella (1985, Theorem 2.1) and Khuri and Casella (2002, Theorem 3), however the existence of a density function is not assumed here. Since Z-p is positive,

EZ-p=R+PZ-pzdz1γp+γ-p,PZ-pzdz=1γp+0,γpu-p-1PZudu1γp+pMγα-pα-p.

For part (ii), since the random variables Zi are positive, then for any z>0

Pi=1NZiz=Pi=1NZiz,maxi{1,,N}Ziz.

Therefore, for z(0,γ)

Pi=1NZizPmaxi{1,,N}Ziz=i=1NPZizi=1NMizi=1Nαi.

Part (iii) can be seen as a consequence of Wx,N and Wx,N+1 being convex ordered and g(x)=x-p being a convex function for x>0 and p0, (see, e.g., Andrieu and Vihola 2014). We provide a self-contained proof by defining for j{1,,N+1}

Sx,N(j):=1Nk=1,kjN+1Wx(k),

and we have

Wx,N+1=1N+1j=1N+1Sx,N(j)

and since the arithmetic mean is greater than or equal to the geometric mean

Wx,N+1j=1N+1Sx,N(j)1N+1.

This implies for p>0

EWx,N+1-pEj=1N+1Sx,N(j)-pN+1j=1N+1ESx,N(j)-p1N+1=ESx,N(1)-p=EWx,N-p,

where Hölder’s inequality has been used and the fact that the random variables Sx,N(j):j1,,N+1 are identically distributed according to Qx,N.

For part (iv), let Mγ=supy[γ,)|g(y)| and due to continuity at y=1, for any ε>0 there exists a δ>0 such that

E|gWx,N-g(1)|1Wx,N[γ,)2MγPγWx,N1-δ+2MγP1+δWx,N+E|gWx,N-g(1)|1Wx,N1-δ,1+δ2MγP|Wx,N-1|δ+εP|Wx,N-1|<δ.

Therefore, for fixed ε and by (W1)

limNsupxXE|gWx,N-g(1)|1Wx,N[γ,)2MγlimNsupxXP|Wx,N-1|δ+εε,

obtaining the result since ε can be picked arbitrarily small.

Proof of Proposition 3.4

First notice that if l< then l1. To see this, define

am:=εm-1εm,

then for fixed δ>0, there exists MN such that for mM

am<l+δ.

Then, for mM

ε1εm=j=2maj<l+δm-Mε1εM,

and because εm0, it is clear that l+δm as m. Therefore, l+δ>1 and since δ can be taken arbitrarily small, it is true that l1.

Now, for weights as in (21) and using a simple random walk proposal, the noisy acceptance probability can be expressed as

α~N(m,m-1)=j=0Nk=0Nmin1,2θ1-θbm-1j+N-jεm-1bmk+N-kεmNj×Nksm-1jsmk1-sm-1N-j1-smN-k 35

and

α~N(m,m+1)=j=0Nk=0Nmin1,1-θ2θbm+1j+N-jεm+1bmk+N-kεmNj×Nksm+1jsmk1-sm+1N-j1-smN-k. 36

Since bm, then sm0 as m; therefore, any term in (35) and (36), for which j+k0, tends to zero as m. Hence,

α~N(m,m-1)=min1,2θ1-θ×εm-1εm×1-sm-1N1-smN+o(1)

and

α~N(m,m+1)=min1,1-θ2θ×εm+1εm×1-sm+1N1-smN+o(1),

implying,

limmP~Nm,m-1P~Nm,m+1=1-θlimmmin1,2θ1-θ×εm-1εmθlimmmin1,1-θ2θ×εm+1εm. 37

If l=+, (37) tends to +, whereas if l<

limmP~Nm,m-1P~Nm,m+1=2lmin1-θ,2θlmin2θl,1-θ2.

In any case, this implies

limmP~Nm,m-12limmP~Nm,m+1,

and since

limmP~Nm,m-1=min1-θ,2θl>0,

the noisy chain is geometrically ergodic according to Proposition Proposition 5.1.

Proof of Proposition 3.5

Noting that

εm-1εmOm2Om-1ifm(mod3)=0,m(mod3){1,2},andεm+1εmOm-2Omifm(mod3)=2,m(mod3){0,1},

expressions in (35) and (36) become

α~N(m,m-1)=1-sm-1N1-smN1m(mod3)=0+O(m-1)

and

α~N(m,m+1)=1-sm+1N1-smN1m(mod3)=0,1+Om-1.

Therefore,

P~Nm,m-1/P~Nm,m+1=1-θθ1-sm-1N+O(m-1)1-sm+1N+O(m-1)1m(mod3)=0+Om-11m(mod3)=1+O(1)1m(mod3)=2,

implying there exists CR+ such that for j=0,2

limkP~N3k+j,3k+j-1P~N3k+j,3k+j+1C,

and

limkP~N3k+1,3kP~N3k+1,3k+2=0.

Then, for fixed δ>0 there exists k0N+ such that whenever kk0

P~N3k+j,3k+j-1P~N3k+j,3k+j+1<C+δ,forj=0,2

and

P~N3k+1,3kP~N3k+1,3k+2<δ.

Let

cm:=P~N(m,m-1)P~N(m,m+1),

then for kk0+1

j=23k+1cj=j=1kc3j-1c3jc3j+1C+δ2δk-k0j=1k0c3j-1c3jc3j+1.

Take δ small enough, such that C+δ2δ<1, hence

k=1j=23k+1cj=k=1k0j=23k+1cj+k=k0j=23k+1cjk=1k0j=23k+1cj+j=1k0c3j-1c3jc3j+1×k=k0C+δ2δk-k0=k=1k0j=23k+1cj+j=1k0c3j-1c3jc3j+11-C+δ2δ<.

Similarly, it can be proved that

k=0j=23k+2cj<andk=1j=23kcj<,

thus

m=2j=2mcj<,

implying the noisy chain is transient according to Proposition 5.1.

Section 4

Proof of Proposition 4.1

From (23) and taking δ<12, η=δ1-δ

supxXP~N(x,·)-P(x,·)TV3δ+4supxXP|Wx,N-1|δ2.

Using Markov’s inequality

supxXP~N(x,·)-P(x,·)TV3δ+4supxXP|Wx,N-1|1+kδ21+k3δ+23+kδ1+ksupxXE|Wx,N-1|1+k3δ+23+kδ1+kNksupxXE|Wx-1|1+k.

Now, let

Ck=supxXE|Wx-1|1+k,

then the convex function f:R+R+ where

f(s)=3s+23+kCks1+kNk,

is minimised at

s=(1+k)23+kCk3Nk12+k=ON-k2+k.

Then,

supxXP~N(x,·)-P(x,·)TVf(s)=ON-k2+k+ON-k+k(1+k)2+k=ON-k2+k.

Applying Theorem 4.2 by taking

r(N)N1-22+k

and noting logN1-22+klog(N), the result is obtained.

For the second claim, for a given ε(0,1) take kε2ε-1-1 and apply the first part.

Footnotes

An erratum to this article is available at http://dx.doi.org/10.1007/s11222-017-9755-5.

References

  1. Alquier, P., Friel, N., Everitt, R., Boland, A.: Noisy Monte Carlo: convergence of Markov chains with approximate transition kernels. Stat. Comput., 1–19 (2014). doi:10.1007/s11222-014-9521-x
  2. Andrieu, C., Roberts, G.O.: The pseudo-marginal approach for efficient Monte Carlo computations. Ann. Stati. 37(2), 697–725 (2009). http://www.jstor.org/stable/30243645
  3. Andrieu, C., Vihola, M.: Establishing some order amongst exact approximations of MCMCs (2014). arXiv preprint. arXiv:14046909
  4. Andrieu C, Vihola M. Convergence properties of pseudo-marginal Markov chain Monte Carlo algorithms. Ann. Appl. Probab. 2015;25(2):1030–1077. doi: 10.1214/14-AAP1022. [DOI] [Google Scholar]
  5. Andrieu, C., Doucet, A., Holenstein, R.: Particle Markov chain Monte Carlo methods. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 72(3), 269–342 (2010). http://www.jstor.org/stable/40802151
  6. Beaumont, M.A.: Estimation of population growth or decline in genetically monitored populations. Genetics 164(3), 1139–1160 (2003). http://www.genetics.org/content/164/3/1139.abstract. http://www.genetics.org/content/164/3/1139.full.pdf+html [DOI] [PMC free article] [PubMed]
  7. Breyer, L., Roberts, G.O., Rosenthal, J.S.: A note on geometric ergodicity and floating-point roundoff error. Stat. Probab. Lett. 53(2), 123–127 (2001). doi:10.1016/S0167-7152(01)00054-2. http://www.sciencedirect.com/science/article/pii/S0167715201000542
  8. Callaert, H., Keilson, J.: On exponential ergodicity and spectral structure for birth-death processes, II. Stoch. Process. Appl. 1(3), 217–235 (1973). doi:10.1016/0304-4149(73)90001-X. http://www.sciencedirect.com/science/article/pii/030441497390001X
  9. Chan, K.S., Geyer, C.J.: Discussion: Markov chains for exploring posterior distributions. Ann. Stat. 22(4), 1747–1758 (1994). http://www.jstor.org/stable/2242481
  10. Chandra, T.K.: Uniform Integrability in the Cesàro sense and the weak law of large numbers. Sankhyā Ser. A (1961–2002) 51(3), 309–317 (1989). http://www.jstor.org/stable/25050754
  11. Del Moral, P.: Feynman–Kac formulae: genealogical and interacting particle systems with applications. In: Probability and Its Applications. Springer, New York (2004). http://books.google.co.uk/books?id=8LypfuG8ZLYC
  12. Doucet, A., Pitt, M.K., Deligiannidis, G., Kohn, R.: Efficient implementation of Markov chain Monte Carlo when using an unbiased likelihood estimator. Biometrika (2015). doi:10.1093/biomet/asu075. http://biomet.oxfordjournals.org/content/early/2015/03/07/biomet.asu075.abstract, http://biomet.oxfordjournals.org/content/early/2015/03/07/biomet.asu075.full.pdf+html
  13. Fernández-Villaverde, J., Rubio-Ramírez, J.F.: Estimating macroeconomic models: a likelihood approach. Rev. Econ. Stud. 74(4), 1059–1087 (2007). doi:10.1111/j.1467-937X.2007.00437.x. http://restud.oxfordjournals.org/content/74/4/1059.abstract, http://restud.oxfordjournals.org/content/74/4/1059.full.pdf+html
  14. Ferré D, Hervé L, Ledoux J. Regular perturbation of V-geometrically ergodic Markov chains. J. Appl. Probab. 2013;50(1):184–194. doi: 10.1017/S002190020001319X. [DOI] [Google Scholar]
  15. Flegal, J.M., Jones, G.L.: Batch means and spectral variance estimators in Markov chain Monte Carlo. Ann. Stat. 38(2), 1034–1070 (2010). http://www.jstor.org/stable/25662268
  16. Girolami, M., Lyne, A.M., Strathmann, H., Simpson, D., Atchade, Y.: Playing Russian roulette with intractable likelihoods (2013). arXiv preprint. arXiv:13064032
  17. Hairer, M., Mattingly, J.C.: Yet another look at Harris’ ergodic theorem for Markov chains. In: Seminar on Stochastic Analysis, Random Fields and Applications VI, Springer, Basel, pp. 109–117 (2011)
  18. Khuri, A., Casella, G.: The existence of the first negative moment revisited. Am. Stat. 56(1), 44–47 (2002). http://www.jstor.org/stable/3087326
  19. Lee, A., Łatuszyński, K.: Variance bounding and geometric ergodicity of Markov chain Monte Carlo kernels for approximate Bayesian computation. Biometrika (2014). doi:10.1093/biomet/asu027. http://biomet.oxfordjournals.org/content/early/2014/08/05/biomet.asu027.abstract, http://biomet.oxfordjournals.org/content/early/2014/08/05/biomet.asu027.full.pdf+html
  20. Lindvall, T.: Lectures on the Coupling Method. Dover Books on Mathematics Series. Dover Publications, Incorporated (2002). http://books.google.co.uk/books?id=GUwyU1ypd1wC
  21. Maire F, Douc R, Olsson J. Comparison of asymptotic variances of inhomogeneous Markov chains with application to Markov chain Monte Carlo methods. Ann. Stat. 2014;42(4):1483–1510. doi: 10.1214/14-AOS1209. [DOI] [Google Scholar]
  22. McKinley, T.J., Ross, J.V., Deardon, R., Cook, A.R.: Simulation-based Bayesian inference for epidemic models. Comput. Stat. Data Anal. 71(0), 434 – 447 (2014). doi:10.1016/j.csda.2012.12.012. http://www.sciencedirect.com/science/article/pii/S016794731200446X
  23. Mengersen, K.L., Tweedie, R.L.: Rates of convergence of the Hastings and Metropolis algorithms. Ann. Stat. 24(1), 101–121 (1996). http://www.jstor.org/stable/2242610
  24. Meyn, S.P., Tweedie, R.L.: Computable bounds for geometric convergence rates of Markov chains. Ann. Appl. Probab. 4(4), 981–1011 (1994). http://www.jstor.org/stable/2245077
  25. Meyn S, Tweedie RL. Markov Chains and Stochastic Stability. 2. New York: Cambridge University Press; 2009. [Google Scholar]
  26. Mitrophanov AY. Sensitivity and convergence of uniformly ergodic Markov chains. J. Appl. Probab. 2005;42(4):1003–1014. doi: 10.1017/S0021900200001066. [DOI] [Google Scholar]
  27. Norris, J.: Markov Chains. No. 2008 in Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press (1999). https://books.google.co.uk/books?id=qM65VRmOJZAC
  28. O’Neill, P.D., Balding, D.J., Becker, N.G., Eerola, M., Mollison, D.: Analyses of infectious disease data from household outbreaks by Markov chain Monte Carlo methods. J. R. Stat. Soc. Ser. C (Appl. Stat.) 49(4), 517–542 (2000). http://www.jstor.org/stable/2680786
  29. Piegorsch, W.W., Casella, G.: The existence of the first negative moment. Am. Stat. 39(1), 60–62 (1985). http://www.jstor.org/stable/2683910
  30. Pillai, N.S., Smith, A.: Ergodicity of approximate MCMC chains with applications to large data sets (2014). arXiv preprint arXiv:14050182
  31. Pitt, M.K., dos Santos Silva, R., Giordani, P., Kohn, R.: On some properties of Markov chain Monte Carlo simulation methods based on the particle filter. J. Econom. 171(2), 134–151 (2012). doi:10.1016/j.jeconom.2012.06.004. http://www.sciencedirect.com/science/article/pii/S0304407612001510
  32. Roberts, G., Rosenthal, J.: Geometric ergodicity and hybrid Markov chains. Electron. Commun. Probab. 2(2), 13–25 (1997). doi:10.1214/ECP.v2-981. http://ecp.ejpecp.org/article/view/981
  33. Roberts GO, Rosenthal JS. General state space Markov chains and MCMC algorithms. Probab. Surv. 2004;1:20–71. doi: 10.1214/154957804100000024. [DOI] [Google Scholar]
  34. Roberts, G.O., Tweedie, R.L.: Geometric convergence and central limit theorems for multidimensional Hastings and Metropolis algorithms. Biometrika 83(1), 95–110 (1996). http://www.jstor.org/stable/2337435
  35. Roberts, G.O., Rosenthal, J.S., Schwartz, P.O.: Convergence properties of perturbed Markov chains. J. Appl. Probab. 35(1), 1–11 (1998). http://www.jstor.org/stable/3215541
  36. Rosenthal, J.S.: Minorization conditions and convergence rates for Markov chain Monte Carlo. J. Am. Stat. Assoc. 90(430), 558–566 (1995). http://www.jstor.org/stable/2291067
  37. Rudolf, D., Schweizer, N.: Perturbation theory for Markov chains via Wasserstein distance (2015). arXiv preprint. arXiv:150304123
  38. Shardlow T, Stuart AM. A perturbation theory for ergodic Markov chains and application to numerical approximations. SIAM J. Numer. Anal. 2000;37(4):1120–1137. doi: 10.1137/S0036142998337235. [DOI] [Google Scholar]
  39. Sherlock C, Thiery AH, Roberts GO, Rosenthal JS. On the efficiency of pseudo-marginal random walk Metropolis algorithms. Ann. Stat. 2015;43(1):238–275. doi: 10.1214/14-AOS1278. [DOI] [Google Scholar]
  40. Thorisson, H.: Coupling, stationarity, and regeneration. In: Probability and Its Applications. Springer, New York (2013). http://books.google.co.uk/books?id=187hnQEACAAJ

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Statistics and Computing are provided here courtesy of Springer

RESOURCES