Skip to main content
Taylor & Francis Open Select logoLink to Taylor & Francis Open Select
. 2019 Jan 9;2019(4):308–335. doi: 10.1080/03461238.2018.1560357

Approximation methods for piecewise deterministic Markov processes and their costs

Peter Kritzer a, Gunther Leobacher b, Michaela Szölgyenyi c,d, Stefan Thonhauser e,CONTACT
PMCID: PMC6474733  PMID: 31058276

ABSTRACT

In this paper, we analyse piecewise deterministic Markov processes (PDMPs), as introduced in Davis (1984). Many models in insurance mathematics can be formulated in terms of the general concept of PDMPs. There one is interested in computing certain quantities of interest such as the probability of ruin or the value of an insurance company. Instead of explicitly solving the related integro-(partial) differential equation (an approach which can only be used in few special cases), we adapt the problem in a manner that allows us to apply deterministic numerical integration algorithms such as quasi-Monte Carlo rules; this is in contrast to applying random integration algorithms such as Monte Carlo. To this end, we reformulate a general cost functional as a fixed point of a particular integral operator, which allows for iterative approximation of the functional. Furthermore, we introduce a smoothing technique which is applied to the integrands involved, in order to use error bounds for deterministic cubature rules. We prove a convergence result for our PDMPs approximation, which is of independent interest as it justifies phase-type approximations on the process level. We illustrate the smoothing technique for a risk-theoretic example, and compare deterministic and Monte Carlo integration.

KEYWORDS: Risk theory, piecewise deterministic Markov process, quasi-Monte Carlo methods, phase-type approximations, dividend maximisation

MATHEMATICS SUBJECT CLASSIFICATION (2010): 60J25, 91G60, 65D32

1. Introduction

Many models in risk theory can be formulated as piecewise deterministic Markov processes (PDMPs) – a general class of finite-variation sample path Markov processes introduced by Davis (1984). This applies, among others, to the classical Cramér–Lundberg model, the renewal risk models, and multi-portfolio models recently introduced by Albrecher & Lautscham (2015). Moreover, PDMPs are sufficiently general to allow for non-constant model parameters, i.e. quantities such as the hazard rate or the premium rate may be state dependent. Examples of PDMPs and their control in the field of insurance mathematics are, e.g. Dassios & Embrechts (1989), Embrechts & Schmidli (1994), Schäl (1998), Rolski (1999), Cai et al. (2009), Leobacher & Ngare (2016), and Eichler et al. (2017).

The general theory of PDMPs is well developed, see for example the monographs by Davis (1993), Jacobsen (2006), or Bäuerle & Rieder (2011) for general results on PDMPs and their optimal control. More specialised contributions to the control theory of PDMPs can be found in Davis (1993), Lenhart & Liaot (1985), Costa & Davis (1989), Dempster & Ye (1992), Almudevar (2001), Forwick et al. (2004), Bäuerle & Rieder (2010), Costa & Dufour (2013), or Davis & Farid (1999) for viscosity solutions of associated Hamilton–Jacobi–Bellman equations, and Colaneri (2017) for a general comparison principle for solutions to control problems for PDMPs.

For the numerical treatment of (control) problems for PDMPs, however, only problem-specific solutions have been provided. A standard approach is to link expected values representing a quantity of interest in the problem to the solution of an associated integro-(partial) differential equation, see, e.g. Asmussen & Albrecher (2010). In only very few cases is it possible to derive an explicit solution to this integro-(partial) differential equation. Requiring an explicit solution typically restricts the complexity of the model significantly. One possibility is to solve the integro-(partial) differential equation numerically. This carries all the intricacies and difficulties of a combined numerical method for differential and integral equations. Alternatively one can apply crude Monte Carlo methods, see, e.g. Riedler (2013). Those methods, while robust, are limited in speed by the Monte Carlo convergence rate. Another – highly sophisticated – approach uses quantisation of the jump distribution, see de Saporta et al. (2016).

In this article we concentrate on particularly easy to implement methods similar to Monte Carlo. The aim is to adapt the problem in a way that also allows for deterministic numerical integration algorithms such as quasi-Monte Carlo (QMC). QMC has been applied successfully to problems in risk theory, see Tichy (1984), Coulibaly & Lefèvre (2008), Siegl & Tichy (2000), Albrecher & Kainhofer (2002), and Preischl et al. (2018). It should be noted that the finiteness of the total variation needed for the convergence estimate (Albrecher & Kainhofer 2002, Theorem 1) has not been proven.

We would like to highlight two features of our approach. Inspired by Albrecher & Kainhofer (2002), we reformulate a general cost functional as a fixed point of a particular integral operator, which allows for iterative approximation of the functional. In terms of numerical integration this means that we get a high-dimensional integration problem of fixed dimension, where the dimension is a multiple of the number of iterations. Having a fixed dimension is required for the application of standard QMC or other deterministic cubature rules.

The application of QMC requires some degree of regularity of the integrand. Only in rare cases these will be satisfied automatically. The examples from risk theory considered here lead to non-smooth integrands. For these situations, we introduce a smoothing technique which, in its simplest case, leads to C2 integrands. From the earlier considerations, we obtain deterministic error bounds for those. We prove convergence in distribution of the ‘smoothed processes’ to the original ones, which implies convergence of the corresponding expected values for every initial value of the process. In Section 2.1 we even obtain uniform convergence with respect to the initial value in a particular setup from risk theory.

Our convergence result has an additional benefit for a typical situation in risk-theoretic modelling. In the literature on the analysis of ruin probabilities, or more generally, on Gerber–Shiu functions, the assumption of a claim size distribution of mixed exponential or phase-type form is quite common. Apart from the possibility to obtain explicit expressions for quantities of interest in such setups, this modelling approach is motivated by the fact that the class of phase-type distributions is dense in the class of distributions with support on [0,), see Rolski (1999, Theorem 8.2.3). Under mild assumptions on the claim size distribution we want to approximate, our convergence result applies and justifies the phase-type approximation procedure even on the process level. Furthermore, efficient and easy to implement numerical methods for the computation of important targets such as Gerber-Shiu functions and expected discounted future dividend payments of an insurance company are of particular importance when models become more general and hence also more complicated. This makes our contribution valuable from both the analytical and the numerical point of view.

We would like to emphasise that the methods presented here per se do not provide solutions to optimal control problems, which is the main application of PDMPs in risk theory. However, the integration algorithms as introduced here can be used in a policy iteration procedure for calculating costs associated with a fixed policy.

The paper is structured as follows. In Section 2 we recall the definition of a PDMP and provide some risk-theoretic examples. In Section 3 we derive the fixed point approach for valuation of a cost functional of a PDMP. Section 4 reviews deterministic numerical integration of possibly multivariate Ck functions. Subsequently, Section 5 is devoted to the aforementioned smoothing procedure, and presents a stability result. Section 6 contains an application of the smoothing to one of the risk-theoretic examples and a comparative study of deterministic and Monte Carlo integration for this example.

2. Piecewise deterministic Markov processes

In this section we first define PDMPs. Then we give a couple of examples of practical interest.

A PDMP is a continuous-time stochastic process with (possibly random) jumps, which follows a deterministic flow, e.g. the solution of an ordinary differential equation (ODE), between jump times. We will not give the most general definition of PDMPs here, but instead refer to the monograph by Davis (1993). For a subset A of Rd we denote by A,A¯, and A its interior, closure, and boundary, respectively. We write B(A) for the Borel σ-algebra on A.

Definition 2.1

Let ARd. A function ϕ:A×RRd is called a flow on A, if

  • ϕ is continuous,

  • ϕ(x,0)=x for all xA;

  • for all xA and all s,tR it holds that if ϕ(x,t)A and ϕ(ϕ(x,t),s)A then ϕ(x,t+s)=ϕ(ϕ(x,t),s).

For fixed xA, let ϕ1(x,A)={tR:ϕ(x,t)A}. Then the function ϕ(x,):ϕ1(x,A)A is called a trajectory of the flow.

If ϕ is a flow on A, then we write ϕA={xA:ϵ(0,) such that t(0,ϵ):ϕ(x,t)A} and ϕ+A={xA:ϵ(0,) such that t(0,ϵ):ϕ(x,t)A}.

Thus ϕA consists of the points on the boundary of A from which the trajectory moves into A immediately, and ϕ+A consists of the points on the boundary of A to which a trajectory moves from A without passing other points on the boundary in-between. Furthermore, we write ϕ1A:=ϕAϕ+A.

Remark 2.2

The classical example of a flow arises through ODEs. Let g:RdRd be Lipschitz continuous. By the classical Picard–Lindelöf theorem on existence and uniqueness of solutions of ODEs we have that for every xR there exists a continuously differentiable function κ:RRd such that κ(0)=x and κ(s)=g(κ(s)) for all sR. For tR we define ϕ(x,t)=κ(t). The function ϕ defines a flow on Rd. If ARd, then the restriction of ϕ to A×R is a flow on A.

Definition 2.3

Let K be a finite set and let d:KN be a function which satisfies that, for every kK, EkRd(k) and ϕk is a flow on Ek with Ek=Ekϕk1Ek.

  1. The state space(E,E) of a PDMP is the measurable space defined by E=kK({k}×Ek) and E=σ({{k}×B:kK,BB(Ek)}).

  2. The flow of a PDMP is defined by ϕ={ϕk}kK.

  3. The active boundary of the PDMP is defined by Γ=k=1Kϕk+Ek. Furthermore, we define a σ-algebra on EΓ by E=σ({{k}×B:kK,BB(Ekϕk+Ek)}).

  4. The jump intensityλ of a PDMP is defined by a family of functions λ={λk}kK with λk:Ek[0,) measurable and bounded for all kK.

  5. The jump kernelQ of a PDMP is a function Q:E×(EΓ)[0,1] such that Q(A,) is E-B([0,1]) measurable for every AE, and Q(,x) is a probability measure on (E,E) for every xE with Q({x},x)=0.

We call the triple (ϕ,λ,Q) the local characteristics of a PDMP.

Given a state space (E,E) and local characteristics (ϕ,λ,Q) of a PDMP we define the function t:E[0,] by

t(k,y)=inf{t>0:ϕk(y,t)ϕk+Ek} if t>0:ϕk(y,t)ϕk+Ek,otherwise.

Definition 2.4

Let (E,E) be a state space and let (ϕ,λ,Q) be local characteristics of a PDMP, let xE, and let (Ω,F,P) be a probability space. A piecewise deterministic Markov process starting in x is a stochastic process X:[0,)×ΩE which satisfies the following. There exists a sequence of random variables (Tn)nN with Tn[0,] and TnTn+1 a.s. and limnTn= a.s. for all nN such that

  1. it holds P-a.s. that X0=x,

  2. for all nN, t[Tn,Tn+1), and for (k,y)E with XTn=(k,y) it holds P-a.s. that Xt=ϕk(y,tTn),

  3. for all s,t[0,) it holds P-a.s. that
    P(Tn+1Tn>t|Xs=(k,y)andTns<Tn+1)=e0tλk(ϕk(y,τ))dτif 0<t<t(k,y),0if tt(k,y),
  4. for all nN and all AE it holds P-a.s. that
    P(XTn+1A|XTn)=Q(A,XTn).

Theorem 2.5

Let (E,E) be a state space and let (ϕ,λ,Q) be local characteristics of a PDMP, let xE. There exist a probability space (Ω,F,Px) and a stochastic process X:[0,)×ΩE such that X is a PDMP starting in x with state space E and local characteristics (ϕ,λ,Q). Furthermore, X has the strong Markov property.

Proof.

The proof of Theorem 2.5 for a more general setup that also allows for the possibility of explosions and countable K can be found in Davis (1993, Section 2.25).

Figure 1 illustrates a path of a PDMP.

Figure 1.

Figure 1.

Illustration of a PDMP.

Let f:ER be a function. For all kK we denote by fk the function fk:EkR which satisfies for all xEk that fk(x)=f(k,x). It is not hard to see that f is measurable if and only if fk is measurable for every kK. We say that f is n-times continuously differentiable, if for every kK there exists an open set AkRd(k) with EkAk and an n-times continuously differentiable function fˆk:AkR such that fk=fˆk|Ek. We write Cn(E,Rm) for the space of n-times differentiable functions on E and Cbn(E,Rm) for the space of functions in Cn(E,Rm) for which all derivatives are bounded. Moreover, C0n(E,Rm) is the space of functions in Cbn(E,Rm) for which all derivatives vanish at infinity.

Further, for f:ER, a PDMP X, and t(0,) we write E(f(Xt)|X0=x)=:Ex(f(Xt)).

In the remainder of this section we provide some illustrative examples from risk theory. For other examples and applications in different fields we refer to Davis (1993), de Saporta et al. (2012), and Riedler (2013).

2.1. Examples

2.1.1. Classical Cramér–Lundberg model

Let X=(Xt)t0 be a stochastic process given by

Xt=x+ctSt,t0, (1)

where x,c0, N=(Nt)t0 is a homogeneous Poisson process with intensity λN>0, {Yi}iN is a family of positive i.i.d. random variables with distribution function FY, and St=i=1NtYi for all t0. A usual assumption in this kind of model is the independence of {Yi}iN and N. In risk theory the process X represents a standard model for the surplus of an insurance portfolio. A quantity of interest is the probability of X ever becoming negative, i.e. we are interested in ℙ(τ<∞) , where τ=inf{t0:Xt<0}. The model translates into a PDMP via

  • K={1,2},

  • E1=[0,), E2=(,0),

  • ϕ1(y,t)=y+ctyE1 and tR, ϕ2(y,t)=yyE2 and tR,

  • λ1(y)=λNyE1, λ2(y)=0yE2.

  • For B1B(E1), B2B(E2), and B=({1}×B1)({2}×B2),
    Q(B,(1,y))=P(YyB1)+P(YyB2)
    for yE1, and Q(B,(2,y))=P(YyB2),

where we have used the notation yB={yy:yB} for all yR and BB(R). For yE2, any definition for Q will do, since the jump intensity is 0 there, but the above definition is provided for definiteness.

2.1.2. Cramér–Lundberg model with dividend payments

A classical modification of the model from Section 2.1.1 is the introduction of a dividend barrier at level b>0. Then, once the surplus reaches the barrier, the incoming premium rate is immediately distributed as a dividend. Furthermore, if the process starts above b, the excess is distributed as a lump sum dividend, such that X0+=min{x,b}. A typical quantity of interest is the expected value of discounted future dividend payouts until ruin of the company, which is given by

V(x)=Ex0τeδtc1{Xt=b}dtif xb,xb+Eb0τeδtc1{Xt=b}dtif x>b, (2)

where δ>0 is a preference-based discount factor and τ=inf{t0:Xt<0}. The model translates into a PDMP via

  • K={1,2,3},

  • E1=[0,b), E2=(,0), E3={b},

  • ϕ1(y,t)=y+ctyE1 and tR, ϕ2(y,t)=yyE2 and tR, ϕ3(y,t)=yyE3 and tR,

  • λ1(y)=λNyE1, λ2(y)=0yE2, λ3(y)=λNyE3.

  • For BkB(Ek), 1k3, and B=({1}×B1)({2}×B2)({3}×B3),
    Q(B,(1,y))=P(YyB1)+P(YyB2)
    for yE1, Q(B,(2,y))=P(YyB2) for yE2, and
    Q(B,(3,y))=P(YyB1)+P(YyB2)
    for yE3. Finally, Q(B,(1,y))=1B3(y)(3,y) for yϕ11E1={b}.

Note that only initial values x(,b] translate to a viable initial value for the PDMP. However, this is sufficient for determining V(x) for all xR via (2).

2.1.3. Cramér–Lundberg model with time dependent dividend barrier

In Albrecher & Kainhofer (2002) the model from Section 2.1.2 is further extended to include a time dependent barrier b:[0,)|0,) of the form

b(t)=b0m+tα1/m,

where α,b0>0, m>1. The quantity of interest is again the expected value of discounted future dividend payments until the time of ruin, i.e.

V(x)=Ex0τeδt(cbt)1{Xt=bt}dt,

for xb0, where again τ=inf{t0:Xt<0} and δ>0 is a preference-based discount factor. The model translates into a PDMP via

  • K={1,2,3},

  • E1={(s,y)R2:0y<b(s)}, E2={(s,y)R2:y<0}, E3={(s,y)R2:y=b(s)},

  • ϕ1((s,y),t)=(s+t,y+ct)(s,y)E1 and tR, ϕ2((s,y),t)=(s+t,y)yE2 and tR, ϕ3((s,y),t)=(s+t,b(s+t))(s,y)E3 and tR,

  • λ1(y)=λNyE1, λ2(y)=0yE2, λ3(y)=λNyE3.

  • For BkB(Ek), 1k3, and B=({1}×B1)({2}×B2)({3}×B3),
    Q(B,(1,(s,y)))=P(Yy({s}×R)B1)+P(Yy({s}×R)B2)
    for (s,y)E1, Q(B,(2,(s,y)))=P(Yy({s}×R)B2) for (s,y)E2, and
    Q(B,(3,(s,y)))=P(Yy({s}×R)B1)+P(Yy({s}×R)B2)
    for (s,y)E3. Finally, Q(B,(1,(s,y)))=1B3((s,y))(3,(s,y)) for (s,y)ϕ11E1=E3.

2.1.4. Cramér–Lundberg model with loan

In Dassios & Embrechts (1989) the model from Section 2.1.2 is modified such that the insurance company is not ruined when the surplus hits zero, but has the possibility to take up a loan at an interest rate ρ>0. The time of ruin is given by τ=inf{t0:Xt<c/ρ}. The corresponding quantity of interest is

V(x)=Ex0τeδtc1{Xt=b}dt,

for xb, where δ>0 is a preference-based discount factor. The model translates into a PDMP via

  • K={1,2,3,4,5},

  • E1=[0,b), E2=((c/ρ),0), E3={b}, E4=(,(c/ρ)), E5={c/ρ},

  • ϕ1(y,t)=y+ctyE1 and tR, ϕ2(y,t)=yyE2 and tR, ϕ3 is the flow of the ODE z=c+ρz at (y,t)yE3 and tR, ϕ4(y,t)=yyE4 and tR, ϕ5(y,t)=yyE5 and tR,

  • λ1(y)=λNyE1, λ2(y)=λNyE2, λ3(y)=λNyE3, λ4(y)=0yE4, λ5(y)=0yE5.

  • For BkB(Ek), 1k5, and B=k=15({k}×Bk),
    Q(B,(1,y))=P(YyB1)+P(YyB2)+P(YyB4)
    for yE1, Q(B,(2,y))=P(YyB2)+P(YyB4) for yE2, and
    Q(B,(3,y))=P(YyB1)+P(YyB2)
    for yE3. Finally, Q(B,(1,y))=1B3(y)(3,y) for yϕ11E1={b}, and Q(B,(2,y))=1B2(y)(1,y) for yϕ21E2={0}.

2.1.5. Multidimensional Cramér–Lundberg model

In Albrecher & Lautscham (2015) a two-dimensional extension of the model in Section 2.1.2 is studied. The basis are independent surplus processes modelling two insurance portfolios Xt(j)=x(j)+c(j)tSt(j), j{1,2}, where c(1),c(2)0 and S(j) are compound Poisson processes with intensities λ(1),λ(2) and jump size distributions FY(1),FY(2). Furthermore, b(1),b(2)0 are barriers. As a new feature, the drift of the component at the barrier is added to the other component's drift, causing faster growth of the latter. Dividends are only paid when both surplus processes have reached their individual barriers. We show how the model translates into a PDMP, namely

E1={(x(1),x(2))R2:0x(1)<b(1),0x(2)<b(2)},E2={(x(1),x(2))R2:b(1)=x(1),0x(2)<b(2)},E3={(x(1),x(2))R2:0x(1)<b(1),b(2)=x(2)},E4={(x(1),x(2))R2:b(1)=x(1),b(2)=x(2)},E5=R2(E1E2E3E4).

The flow is given by

ϕ1(x,t)=x+c(1)c(2)t,ϕ2(x,t)=x+0c(1)+c(2)t,ϕ3(x,t)=x+c(1)+c(2)0t,

and ϕ4(x,t)=ϕ5(x,t)=x for all xR2, t0. It remains to describe the jump behaviour. We get deterministic ‘jumps’ at the active boundaries of E1,E2,E3 which do not manifest themselves as jumps of the process, i.e. Q(A,(1,x))=1A((2,x)) for (1,x)ϕ11(E1) and similar for the other active boundaries. Since each surplus process is a compound Poisson process with drift, jumps in the components occur due to realisations of independent identically distributed exponential random variables (independence implies that mutual jumps occur with probability zero). The two-dimensional process thus jumps at the minimum of the individual jump times. This means that we have a constant jump intensity λk=λ(1)+λ(2) for k=1,2,3,4, and λ5=0. If a jump occurs at time t0, it happens with probability λ(1)/(λ(1)+λ(2)) in the first surplus process with jump size distribution FY(1), and with probability λ(2)/(λ(1)+λ(2)) in the second surplus process with jump size distribution FY(2). It remains to describe the jump kernel for the jumps from xE. To this end define, for k1,k2{1,2,3,4} and BB(Ek2)B(R2), and (y(1),y(2))Ek1,

B(1)={z(1)R:(z(1),z(2))B,z(2)=y(2)},B(2)={z(2)R:(z(1),z(2))B,z(1)=y(1)}.

Furthermore,

Q({k2}×B,(k1,y(1),y(2)))=λ(1)λ(1)+λ(2)FY(1)(y(1)B(1))+λ(2)λ(1)+λ(2)FY(2)(y(2)B(2)).

A quantity of interest in this model is again the expected value of discounted future dividend payments until the time of ruin of one of the portfolios,

V(x(1),x(2))=Ex(1),x(2)0τeδt(c(1)+c(2))1E4(Xt(1),Xt(2))dt, (3)

for x(1)b(1), x(2)b(2), with τ=inf{t0:(Xt(1),Xt(2))E5}, and δ>0 being a preference-based discount factor.

3. Iterated integrals and a fixed point approach

In this section we derive a method for numerical approximation of the quantities of interest appearing in the models introduced in the previous section. We rewrite the quantity of interest as a sum of integrals with fixed dimension and an error term that goes to zero exponentially fast with increasing dimension of the integral. This allows for the use of deterministic integration rules. The starting point for the derivation of this integral representation is the observation that the quantity of interest is a fixed point of a certain integral operator associated to the PDMP.

Definition 3.1

Suppose there exists a set KcK such that for all kKc it holds that λk(x)=0, and ϕk(x,t)=x for all xEk and all tR. We call Ec:=kKcEk a cemetery of the PDMP.

Definition 3.2

Let a PDMP be given and let Ec be a cemetery of the PDMP. A running reward function:ER is a measurable function satisfying |Ec0. A terminal cost functionΨ:EcR is a measurable function satisfying Ψ|EEc0. The cost functionalV:ER corresponding to Ec,,Ψ is defined by

V(x)=Ex0τeδt(Xt)dt+eδτΨ(Xτ), (4)

where τ=inf{t0:XtEc}.

Let T1 be the first jump time. Equation (4) can be rewritten as follows,

V(x)=Ex[0T1eδt(ϕ(x,t))dt+T1τeδt(ϕ(XT1,tT1))dt+eδτΨ(Xτ)1{T1<τ}+0τeδt(ϕ(x,t))dt+eδτΨ(ϕ(x,τ))1{τ<T1}+0T1eδt(ϕ(x,t))dt+eδT1Ψ(XT1)1{T1=τ}].

Since X is a PDMP and hence a strong Markov process, this yields V=H+GV with H:ER, G:C2(E,R)R defined by

H(x)=Ex[0T1eδt(ϕ(x,t))dt1{T1<τ}+0τeδt(ϕ(x,t))dt+eδτΨ(ϕ(x,τ))1{τ<T1}+0T1eδt(ϕ(x,t))dt+eδT1Ψ(XT1)1{T1=τ}],GV(x)=Ex[eδT1V(XT1)1{T1<τ}]. (5)

Recall that for every t0 it holds that Px(T1>t)=exp(0tλ(ϕ(x,s))ds)=:1FW(t,x) and denote the corresponding density by fW. With this, the function H and the operator G admit representations as integrals,

H(x)=0t(x)fW(t,x)0teδs(ϕ(x,s))ds+eδtEcΨ(y)Q(dy,ϕ(x,t))dt+(1FW(t(x),x))0t(x)eδs(ϕ(x,s))ds+eδt(x)Ψ(ϕ(x,t(x))),GV(x)=0t(x)fW(t,x)eδtEV(y)Q(dy,ϕ(x,t))dt.

Note that H(x) corresponds to the expected discounted rewards collected before the first jump at time T1 when starting in x. GV(x) represents the expected discounted rewards from time T1 onwards conditional on the event {XT1Ec,X0=x}. Iterating the above steps nN times leads to

V(x)=GnV(x)+i=0n1GiH(x). (6)

Lemma 3.3

Let Ψ:EcR and :ER be bounded, for all kK assume that the functions λk are bounded by Cλ(0,), and for all xE let t(x)=. Then for all xE and for all nN it holds that |GnV(x)|CV(Cλ/(Cλ+δ))n and, in particular, it holds that limnGnV(x)=0 uniformly in xE.

Proof.

The boundedness of ℓ and Ψ implies that also V is bounded by CV=(/δ)+Ψ. Using the strong Markov property and Equation (5) we have by induction on n,

GnV(x)=ExeδT1Gn1V(XT1)1{T1<τ}=ExeδT1EXT1eδ(TnT1)V(XTn)1{Tn<τ}1{T1<τ}=ExEXT1eδTnV(XTn)1{Tn<τ}1{T1<τ}=ExeδTnV(XTn)1{τ>Tn}, (7)

where we used 1{Tn<τ}1{T1<τ}=1{Tn<τ} in the last equality. Recall that P(TnTn1>t|Tn1,XTn1)=exp(0tλ(ϕ(s,XTn1))ds)exp(tCλ). For every nN let ZnErlang(n,Cλ) be an Erlang-distributed random variable. Combining this with (7) we get that

GnV(x)CVExeδTnCVEeδZn=CVCλCλ+δn.

The latter expression converges to zero as n uniformly in xE.

Combining Lemma 3.3 with (6) results in the error estimate

V(x)i=0n1GiH(x)CVCλCλ+δn. (8)

Finally, we obtain the following representation,

Gi1H(x0)=t1=0t(x0)fW(t1,x0)eδt1x1Et2=0t(x1)fW(t2,x1)eδt2x2E×ti1=0t(xk2)fW(ti1,xi2)eδti1xi1EH(xi1)Q(dxi1,ϕ(xi2,ti1))dti1Q(dx1,ϕ(x0,t1))dt1=t1=0t(x0)x1Eti1=0t(xi2)xi1Ej=1i1fW(tj,xj1)eδtjH(xi1)Q(dxi1,ϕ(xi2,ti1))dti1Q(dx1,ϕ(x0,t1))dt1. (9)

In (9) we denote by {tj}j{1,,i1} the family of inter-jump times and by {xj}j{1,,i1} the family of post-jump locations.

Remark 3.4

Solving the integral Gi1H(x0) brings several advantages compared to a crude Monte Carlo approach. First, (9) is an integral with a fixed dimension. Hence, it can be approximated using deterministic integration rules like QMC, for which deterministic error bounds are available. Second, the bias of restricting oneself to a fixed number of jumps can be estimated uniformly in x=x0 using the bias estimate in Lemma 3.3. Third, rare events like surviving a large number of jumps are – in this formulation – not rare in the sense that it is unlikely to draw such a realisation, which has the effect of importance sampling.

4. Cubature rules for Cκ-functions

In order to obtain convergence estimates for numerical integration methods such as QMC methods or other cubature rules, we need more regularity of the integrands than they admit in many practical applications. For example, we may need to bound a certain norm of the Hessian matrix of the integrand. In Section 5, we will rewrite the problem introduced in Section 3 so that the integrand is a function f:[0,1]dR which satisfies fC2([0,1]d), or more generally fCκ([0,1]d) for some κN. We outline two different methods for treating such integrands f by cubature rules.

4.1. QMC methods

QMC methods are equal-weight cubature rules with M deterministically chosen integration nodes. Let the integrand f:[0,1]dR satisfy fC2([0,1]d). In order to obtain a convergence estimate for numerical integration of f using QMC, we require a so-called Koksma–Hlawka type inequality. The original Koksma–Hlawka inequality bounds the integration error of a QMC rule by the product of the variation of the integrand (in the sense of Hardy and Krause) and the so-called discrepancy of the integration node set (see, e.g. Niederreiter (1992, Chapter 2)). We remark, however, that we cannot easily apply the classical Koksma–Hlawka inequality in this paper, as we cannot rely on the integrands to have bounded variation in the sense of Hardy and Krause. Hence, we are going to resort to a variant of the Koksma–Hlawka inequality which was recently proven in Pausinger & Svane (2015). Let QM,d(f)=1/Mj=1Mf(xj) be a QMC rule using M integration nodes x1,,xM[0,1)d. Then by Pausinger & Svane (2015, Theorem 3.12) we have

[0,1]df(x)dxQM,d(f)supx[0,1]df(x)infx[0,1]df(x)+d16M(f)DiscI(x1,,xM), (10)

where M(f)=supx[0,1]dHess(f,x), Hess(f,x) is the Hessian matrix of f at x, denotes the usual operator norm, and where DiscI(x1,,xM) is the isotropic discrepancy of the integration node set,

DiscI(x1,,xM)=supC[0,1]dC convex1Mj=1M1{xjC}μd(C),

where μd denotes the Lebesgue measure on the Rd. Now let x1,,xM[0,1]d. In Niederreiter (1992, Chapter 2) it is shown that

DiscI(x1,,xM)8dDisc(x1,,xM)1/d,

where by Disc(x1,,xM) we denote the star discrepancy of x1,,xM, defined as

Disc(x1,,xM)=supa[0,1)d1Mj=1M1{xj[0,a)}μd([0,a)),

where [0,a) denotes [0,a1)××[0,ad) for a=(a1,,ad). It is well known that common point sequences that are employed in QMC methods, such as Sobol' sequences or Halton sequences, have a star discrepancy of order (logM)d/M (and it is known that this order can, if at all, only be improved with respect to the exponent of the log-term). Hence, by using, e.g. Sobol' points in a QMC method for numerically integrating a C2-function, we cannot expect an error that converges to zero faster than (logM)/M1/d.

As we shall see below, this order of magnitude can, with respect to the disadvantageous dependence on d, not be improved further for C2-functions. However, there is room for improvement if we make additional smoothness assumptions on the integrand.

4.2. Product rules

In Hinrichs et al. (2017) it is shown that, by using products of Gauss rules, one can obtain the following result. Let f:[0,1]dR be such that fCκ for some κN. Then, by using a product rule QG,M~,d of M~-point Gauss quadrature rules, one obtains

[0,1]df(x)dxQG,M~,d(f)cκdM~κfCκ,for M~κ+1, (11)

where cκ=(π/2)(e/(63))κ, and where

fCκ=maxβN0dβ1κDβ(f)L,

where Dβ denotes the (weak) partial derivative of order β for βN0d. A d-fold Gauss product rule as described above uses M=M~d points in total, and hence yields a convergence order of Mκ/d. It is known due to Bakhvalov (1959) that this convergence order is best possible. For the special case κ=2, we only obtain a relatively small improvement over the bound implied by (10). However, there is an additional advantage in the bound (11). By requiring that the function f satisfies additional smoothness assumptions, namely that fCκ for some κN which is possibly larger than 2, we obtain an improved convergence rate. Hence, we face a trade-off between imposing a higher degree of smoothness on the integrand f to obtain a higher accuracy in the quadrature rule, and the error we make by smoothing the integrand to that extent. It is therefore likely that the method needs to be fine-tuned on a case-by-case basis. In practice, product rules often cannot be applied, since, for example, for integrating a d=1024-variate integrand using only M~=2 integration nodes per coordinate requires a point set consisting of M=21024 integration nodes. To overcome the latter problem, it might be useful to apply the theory of weighted integration as introduced in Sloan & Woźniakowski (1998), possibly combined with truncation (see, e.g. Kritzer et al. (2016)) or multivariate decomposition methods (see, e.g. Kuo (2010)). A detailed analysis of these approaches applied to the present problem is left open for future research.

5. Smoothing of the integrand

The integrand in (9) is not necessarily a Cκ-function. Therefore, in this section we provide a technique for smoothing the integrand in order to apply convergence results for integration rules that are described in Section 4.

The piecewise construction of the process described in Definition 2.4 leads to the situation that Xt=ϕ(XTj1,tTj1) for t[Tj1,Tj) is a function of XTj1 and Tj1. In particular, all subsequent pre-jump locations depend on all previous post-jump locations and jump times, via ϕ and λ. Consequently, regularity of the integrand depends on regularity of the flow ϕ and the intensity function λ. The analysis in this section is restricted to the case where the flow originates from autonomous ODEs, i.e. for all kK there exist Lipschitz continuous functions gk:Rd(k)Rd(k) such that (/t)ϕk(y,t)=gk(ϕk(y,t)). General results from the literature on ODEs, see, e.g. Grigorian (2009), yield that the derivatives (/y)ϕk,(2/y2)ϕk,(/t)ϕk can be described by so-called associated first- and second-order variational equations for which one requires gk to be a C2-function.

For the density fW of the inter-jump times to be C2 we need that λC2(E,R). Also we need Cb2(E,R), and ΨCb2(E,R) since they appear in the integral defining H.

A serious problem with respect to smoothness arises if the PDMP model allows for jumps from the active boundary. Suppose (k,y)E and t(k,y)<. Then, conditional on Xt=(k,y), the time of the next jump is distributed as min(T,t(k,y)), where T has distribution function FT(t)=1exp(0tλk(ϕk(y,s))ds). But in general neither t(k,y) nor min(T,t(k,y)) will depend smoothly on y, even if λk has arbitrarily high regularity. We are not aware of a general remedy for this problem. However, for all PDMP models put forward in Section 2.1, the jumps from the active boundary do not constitute jumps of the original problem. In the following subsection we describe by example how PDMPs can be approximated by PDMPs that do not allow for jumps from the boundary.

Concerning the jump kernel Q, it is hard to state general sufficient regularity conditions. An exemplary favourable situation arises if the jump kernel Q admits a C2-density fY in the sense that Q(A,x)=AfY(x1,x)dx1 for all AE and all xE. In the one-dimensional examples from risk theory in Sections 2.1.12.1.4, this condition is satisfied and for the two-dimensional example in Section 2.1.5 we present a smoothing technique in Section 5.2.

5.1. Smoothing of the flow

Consider the example from Section 2.1.4 without dividend barrier. We can describe the problem alternatively with a state space consisting of three components:

  • K={1,2,3},

  • E1=((c/ρ),), E2=(,(c/ρ)), E3={(c/ρ)},

  • ϕ1 is determined by an autonomous ODE of the form g1:RR,
    g1(y)=c,if y(0,),c+ρy,if ycρ,0,0,if y,cρ, (12)
    for some c>0,ρ>0. The function ϕ2 is given by ϕ2(y,t)=yyE2 and tR, and ϕ3 by ϕ3(y,t)=yyE3 and tR,
  • λ1(y)=λNyE1, λ2(y)=0yE2, λ3(y)=0yE3.

  • For B=({1}×B1)({2}×B2)({3}×B3)E,
    Q(B,(1,y))=P(YyB1)+P(YyB2)+P(YyB3)(for yE1),Q(B,(2,y))=P(YyB2)(for yE2),Q(B,(3,y))=P(YyB2)+P(YyB3)(for yE3).

Here, g1 is not differentiable in 0. However, we may smoothen this discontinuity using a ‘smoothened Heaviside function’. Note that Γ=.

Definition 5.1

Let κN{0}. We call a function h:R[0,1] a Cκ-Heaviside function, if

  • h(y)=0 for y<−1,

  • h(y)=1 for y>1,

  • h is non-decreasing,

  • h(y)+h(y)=1,

  • h is κ-times continuously differentiable.

Lemma 5.2

Let κN{0}, and let f:RR be a piecewise Cκ-function with discontinuity in ξR, i.e. let there exist Cκ-functions f1,f2:RR with f=f1 on (,ξ) and f=f2 on (ξ,). Let h be a Cκ-Heaviside function. For every ϵ>0 define fϵ:RR by fϵ(y)=f1(y)h(yξ/ϵ)+f2(y)h(yξ/ϵ). Then,

  1. fϵCκ for every ϵ>0,

  2. fϵ|R(ϵ,ϵ)=f|R(ϵ,ϵ) for every ϵ>0,

  3. for all yR{ξ} it holds that limϵ0+fϵ(y)=f(y),

  4. for all δ>0 it holds that limϵ0+supyR(ξδ,ξ+δ)|fϵ(y)f(y)|=0.

Proof.

The elementary proof is left to the reader.

There are various possible choices for the smoothing: from the left fϵ(y)=f1(y)h(yξ+ϵ/ϵ)+f2(y)h(yξ+ϵ/ϵ) and from the right fϵ+(y)=f1(y)h(yξϵ/ϵ)+f2(y)h(yξϵ/ϵ). Figure 2 depicts these three possible smoothings for a function with a discontinuity in ξ=1. A concrete example for a function h that satisfies the above requirements is given by

h(y)=0if y<1,12+15y165y38+3y516if y[1,1],1if y>1. (13)

For every ϵ>0, a smoothed version of the function g1 defined in (12) is given by

g1ϵ(y)=(c+ρy)h(yϵ)+ch(yϵ).

We can finally formulate a PDMP corresponding to the new model, where the flow has been smoothened,

  • K={1,2,3},

  • E1=((c/ρ),), E2=(,(c/ρ)), E3={(c/ρ)},

  • (/t)ϕ1ϵ(y,t)=g1ϵ(ϕ1ϵ(y,t))yE1 and tR, ϕk(y,t)=yyEk and tR, k{2,3};

  • λ1(y)=λNyE1, λk(y)=0yEk, k{2,3};

  • for B=({1}×B1)({2}×B2)({3}×B3)E,
    Q(B,(1,y))=P(YyB1)+P(YyB2)+P(YyB3)(for yE1),Q(B,(2,y))=P(YyB2)(for yE2),Q(B,(3,y))=P(YyB2)+P(YyB3)(for yE3).

Figure 2.

Figure 2.

Illustration of smoothing a piecewise C2-function with a discontinuity in ξ=1.

Note that Γ=. Since the dividend barrier b is never reached, we also have to smoothen the reward function in a way that the region where dividends are paid can be reached, i.e. ϵ(y)=ch(yb+ϵ/ϵ). We will show convergence of the corresponding value functions in Section 6.

5.2. Smoothing of jump measures

We give an example for a class of jump measures that can be approximated by measures leading to C2-integrands in (9).

Let (E,E) be the state space of a PDMP and let (ϕ,λ,Q) be its local characteristics. Let the probability kernel Q satisfy the following assumption.

Assumption 5.3

We assume that

  1. there exists a positive integer n such that for every kK, and every yEk, there exist sets B1(k,y),,Bn(k,y) such that
    1. for every j{1,,n} there exists k1K such that Bj(k,y)Ek1,
    2. for every j{1,,n} it holds that {(y¯,z):y¯Ek,zBj((k,y¯))} is a connected C2-manifold,
  2. for every kK and every j{1,,n} the mapping from Ek to R, y¯Q(Bj((k,y¯),x) is C2,

  3. for all xE it holds that j=1nQ(Bj(x),x)=1,

  4. for every xE and every j{1,,n} there exists a C2-mapping Gj,x:[0,1]dim(Bj)Bj such that for all AE it holds that
    Q(ABj,x)=μdim(Bj)(Gj,x1(ABj))Q(Bj,x),
    where μm denotes the m-dimensional Lebesgue measure,
  5. for every kK and every j{1,,n} the mapping from Ek×[0,1]dim(Bj) to lKEl, (y,u)Gj,(k,y)(u) is C2.

Note that Assumption 5.3(1) implies that, for every xE, Bj(x) is a C2-manifold, and that for all x1=(k1,y1),x2=(k2,y2)E with k1=k2 we have dimBj(x1)=dimBj(x2).

Under Assumption 5.3 we have for xE and for fCb2(E,R) that

Ef(y)Q(dy,x)=j=1npj(x)[0,1]dim(Bj(x))f(Gj,x(u))du,

where pj(x)=Q(Bk,j,x) for all xE. For the integral in (9) this implies that we have iterated sums for each jump, which increases the complexity for large numbers of jumps. Instead, we may write the sum as an integral over [0,1],

Ef(y)Q(dy,x)=01j=1n1[qk,j1(x),qk,j(x))(u0)[0,1]dim(Bj(x))f(Gj,x(u))dudu0,

where q0(x)=0 and qj(x)=p1(x)++pj(x). However, with this ‘trick’ we have lost the property of the integrand being C2. So, using again our smoothened Heaviside function h:R[0,1], we can smoothen the indicator functions,

Ef(y)Qϵ(dy,x)=01j=1n(h(u0qj1(x)ϵ)+h(qj(x)u0ϵ))[0,1]dim(Bj(x))f(Gj,x(u))dudu0=01[0,1]dim(Bj(x))j=1n(h(u0qj1(x)ϵ)+h(qj(x)u0ϵ))f(Gj,x(u1,,udim(Bj(x))))dudu0.

This expression, considered as a function of x, is C2 as it is a composition of C2-functions.

Theorem 5.4

In the setup of this section we have for all fCb0(E,R) that

limϵ0Ef(y)Qϵ(dy,x)=Ef(y)Q(dy,x).

Proof.

It holds that

|Ef(y)(Qϵ(dy,x)Q(dy,x))|=|j=1n01(h(u0qj1(x)ϵ)+h(qj(x)u0ϵ)1[qj1(x),qj(x))(u0))du0×[0,1]dim(Bj(x))f(Gj,x(u))du|j=1n01|h(u0qj1(x)ϵ)+h(qj(x)u0ϵ)1[qj1(x),qj(x))(u0)|du0×[0,1]dim(Bj(x))|f(Gj,x(u))|du.

For our concrete example of h the first integral can be estimated by 58ϵ. Thus

|Ef(y)(Qϵ(dy,x)Q(dy,x))|5ϵn8f,

yielding the statement of the theorem.

Now, consider the example from Section 2.1.5. Here, a jump can be either a jump in x1-direction or a jump in x2-direction, i.e.

XTj=XTJ+(Y1,0)with probability λ1λ1+λ2,XTJ+(0,Y2)with probability λ2λ1+λ2.

In this case we can find functions G1,G2:[0,1][0,) such that Y1dG1(Θ1) and Y2dG2(Θ2) for uniform random variables Θ1,Θ2. Hence,

Ef(y)Q(dy,(x1,x2))01[0,1]2h(ϵ1(λ1λ1+λ2u))f(x1+G1(ϑ1),x2)+h(ϵ1(uλ1λ1+λ2))f(x1,x2+G2(ϑ2))dϑ1dϑ2du.

Remark 5.5

If we consider, say, i=100 in (9), then we get a very high number of terms to be summed in the integral. However, we always assume ε to be very small, in particular, we may assume that per jump at most two, and in most situations only one, of the terms h(ϵ1(uqj1(x)))+h(ϵ1(qj(x)u)) are nonzero.

5.3. Convergence

In this section we prove a general convergence result for approximated versions of PDMPs with smoothing as above. We will exploit results on Feller processes presented in Kallenberg (2002, Chapter 19) and Ethier & Kurtz (1986, Chapters 4.2 and 4.8). For the remainder of this section we make the following assumptions:

  1. t(x)= for all xE,

  2. λCb(E,R),

  3. for all fCb(E) the mapping xEf(x¯)Q(dx¯,x) is continuous.

With this, we can utilise the following theorem.

Theorem 5.6 Davis 1993, Theorem 27.6 —

If t(x)= for all xE and for all λCb(E,R), and if the mapping xEf(y)Q(dy,x) is continuous for all fCb(E,R), then the PDMP is a Feller process.

We give an example for a class of jump kernel which comprises the jump kernels of the one-dimensional examples in Section 2.1 and which satisfies (iii).

Example 5.7

Let EkR be an interval for every kK and let fY be a bounded density function on R. Furthermore, let, for every x=(k,y)E and every AE, Q(A,(k,y))=jK(yA)EjfY(y¯)dy¯. Then for every fCb(E,R) it holds that

Ef(x)Q(dx,(k,y1))Ef(x)Q(dx,(k,y2))=jKR1Ej(y1y¯)fj(y1y¯)fY(y¯)dy¯jKR1Ej(y2y¯)fj(y2y¯)fY(y¯)dy¯jKR1Ej(y1y¯)fj(y1y¯)fY(y¯)dy¯R1Ej(y2y¯)fj(y2y¯)fY(y¯)dy¯jKR|1Ej(y1y¯)fj(y1y¯)1Ej(y2y¯)fj(y2y¯)|fY(y¯)dy¯.

Since, by assumption, all fj are continuous and all Ej are intervals, it holds that |1Ej(y1y¯)fj(y1y¯)1Ej(y2y¯)fj(y2y¯)| is bounded by 2fj and goes to zero as y1y2 for almost all y¯.

Therefore, bounded convergence implies that the above sum converges to 0. From this the desired continuity follows.

The generator of X in the setup of the current section is given by

Af(x)=Xf(x)+λ(x)E(f(x¯)f(x))Q(dx¯,x),xE, (14)

where for x=(k,y)E we define Xf(x) by (Xf)k(y)=(/t)fk(ϕk(y,t))|t=0. Note that for fCb1(E,R) this means (Xf)(y)=g(y)f(y). So the domain D(A) of the generator consists of all functions in Cb(E,R) which are continuously differentiable along the trajectories of the flow on all components, cf. Ethier & Kurtz (1986, p. 8), and Cb1(E,R)D(A).

Definition 5.8 Kallenberg 2002, Chapter 19 —

Let A be a closed linear operator with domain of definition D(A). A core for A is a linear subspace DD(A) such that the restriction A|D has closure A.

Proposition 5.9 Kallenberg 2002, Proposition 19.9 —

If A is the generator of a Feller semigroup (Pt)t0, then any dense, (Pt)t0-invariant subspace DD(A) is a core for A.

Proposition 5.10

Under the assumptions made in this section, and for A being defined as in (14), it is true that Cb(E,R) is a core for A.

Proof.

We certainly have that Cb(E,R) is a dense subspace of Cb(E,R). Furthermore, the transition semigroup satisfies Pt:Cb(E,R)Cb(E,R) for all t[0,), see (Davis 1993, p.76), since the PDMP is Feller by Theorem 5.6.

We have to prove that Cb(E,R) is invariant under (Pt)t[0,). We show this by proving that, for all kN, PtCbk(E,R)Cbk(E,R). For k=0 this is just the Feller property. Since all derivatives are bounded in the sup-norm, differentiation and application of Pt commute, i.e. (k/xk)Ptf=Pt(k/xk)fCb(E,R) for all kN. Consequently, Cb(E,R) is a core for A.

Theorem 5.11 Kallenberg 2002, Theorem 19.25 —

Let X be a Feller process in E with semigroup (Pt)t0 and generator A with domain D(A), and for all nN let Xn be Feller processes in E with semigroups (Ptn)t0 and generators An with domains D(An). Let D be a core for A. Then the following statements are equivalent:

  1. for every fD there exists a sequence (fn)nN with fnD(An) for all nN and such that fnf and AnfnAf uniformly as n,

  2. for all t>0 we have PtnPt as n in the strong operator topology,

  3. for every fC0(E,R) and every t0(0,) it holds that PtnfPtf as n uniformly for t[0,t0],

  4. if X0ndX0 in E, then XndX in D([0,),E).

Remark 5.12

The notion of weak convergence of processes in Item (iv) needs an explanation. Here, D([0,),E) is the space of càdlàg functions, equipped with the Skorokhod topology, see Ethier & Kurtz (1986, p. 118). With this topology, D([0,),E) is a Borel subset of a Polish space and for a sequence (Xn)nN of D([0,),E)-valued random variables (i.e. processes in E with càdlàg paths), and a D([0,),E)-valued random variable X we have XndX if and only if limnE(F(Xn))=E(F(X)) for all bounded Skorokhod continuous functions D([0,),E)R, see Kurtz & Protter (1996, Section 6) or Ethier & Kurtz (1986, Chapter 3). We do not wish to go into the details of the notion of Skorokhod continuity. It suffices to mention that from Kurtz & Protter (1996, Section 8, Example 8.1) we know that for given continuous functions f1:E×[0,)Rd and f2:[0,)[0,), and fixed t[0,), the following functionals exhibit this property:

F1(ω)=f1(ω(t),t)(for ωD([0,),E)),F2(ω)=0tf2(ts)f1(ω(s),s)ds(for ωD([0,),E)).

Lemma 5.13

Let f:ER be continuous and bounded, then the functional

F3(ω)=0eδsf(ω(s))ds(for ωD([0,),E))

is Skorokhod continuous.

Proof.

Let σ denote the Skorokhod metric on D([0,),E). Let ϵ>0. There exists t>0 such that teδsfds<ϵ/4. By Skorokhod continuity of F2 there exists an η>0 such that for all ω1,ω2D([0,),E) it holds that, if σ(ω1,ω2)<η then |0teδsf(ω1(s))ds0teδsf(ω2(s))ds|<ϵ/2. Therefore,

|F3(ω1)F3(ω2)|=0eδsf(ω1(s))ds0eδsf(ω2(s))ds0teδsf(ω1(s))ds0teδsf(ω2(s))ds+2teδsfds<ϵ.

We remark that a function f:ER is continuous if and only if fk:EkR is continuous for all k. In particular, every indicator function of a component {k}×Ek is continuous.

We are in the position to show that cost functionals indeed commute with weak limits of PDMPs.

Lemma 5.14

Let X be a PDMP and (Xn)nN be a sequence of PDMPs on the same state space E and with the same cemetery Ec, and let :ER and Ψ:ER be a running reward function and a terminal cost function, respectively. Assume that both ℓ and Ψ are continuous and bounded. Assume further that X0n=x for all nN and X0=x, and XndX in D([0,),E).

Then

Ex0τeδt(Xtn)dt+eδτΨ(Xτn)Ex0τeδt(Xt)dt+eδτΨ(Xτ)

as n.

Proof.

Recall that 0 on Ec, and Ψ0 on EEc, so that 0eδs(ω(s))ds=0τeδs(ω(s))ds and 0δeδsΨ(ω(s))ds=τδeδsΨ(ω(s))ds. Thus by Lemma 5.13 the mappings ω0τeδs(ω(s))ds and ωτδeδsΨ(ω(s))ds are Skorokhod continuous.

Moreover, if ω is a path of the PDMPs, then it holds that ω(s)=ω(τ) for all sτ, such that τδeδsΨ(ω(s))ds=eδτΨ(ω(τ)). This completes the proof.

Also, finite time ruin probabilities, i.e. the probability of the PDMP reaching the cemetery before a given time horizon t, commute with weak limits, as we show next.

Lemma 5.15

Let X be a PDMP and (Xn)nN be a sequence of PDMPs on the same state space E and with the same cemetery Ec. Assume further that X0n=x for all nN and X0=x, and XndX in D([0,),E).

Then limnPx(XtnEc)=Px(XtEc) for all t0.

Proof.

Consider a functional of the same form as F1 in Remark 5.12, with f1=1Ec. Since the cemetery is the union of only entire ({k}×Ek), and is therefore a union of connected components of E, the indicator function of the cemetery is continuous. Therefore if we define ψ(x,t)=Px(τt)=Px(XtEc)=Ex(1Ec(Xt)) and ψn(x,t)=Px(τnt)=Px(XtnEc)=Ex(1Ec(Xtn)), nN, we have limnψn(x,t)=ψ(x,t) for all xE and for all t0.

The following theorem specifies conditions under which Theorem 5.11 is applicable in the PDMP setting.

Theorem 5.16

Let X be a Feller PDMP with local characteristics (ϕ,λ,Q) and let Xn,nN, be Feller PDMPs with local characteristics (ϕn,λn,Qn). Further, let the following assumptions hold:

  1. gng and λnλ as n, uniformly in xE,

  2. for all fCb(E,R),
    limnsupxEEf(y)Qn(dy,x)Ef(y)Q(dy,x)=0, (15)
  3. X0ndX0 in E.

Then XndX in D([0,),E).

Proof.

Let D(An), nN, and D(A) be the domains of the generators An, nN, and A, corresponding to Xn and X, respectively. For fnD(An) we have

Anfn(x)=Xnfn(x)+λn(x)E(fn(y)fn(x))Qn(x,dy),(Xnfn)(x)=(gn)(x)(fn)(x).

By Proposition 5.10, D=Cb(E,R) is a core for all generators involved. For every fD we set fn=f for all nN, such that trivially fnf as n. Next, observe that we have for all nN,

|Anf(x)Af(x)||gn(x)f(x)g(x)f(x)|+λn(x)E(f(y)f(x))Qn(dy,x)λ(x)E(f(y)f(x))Q(dy,x)=|(gn(x)g(x))f(x)|+λn(x)E(f(y)f(x))Qn(dy,x)λ(x)E(f(y)f(x))Q(dy,x)gngf+fλn(x)EQn(dy,x)λ(x)EQ(dy,x) (16)
+λn(x)Ef(y)Qn(dy,x)λ(x)Ef(y)Q(dy,x). (17)

Since Qn, nN, and Q are probability measures on (E,B(E)), and since, by assumption, gng and λnλ uniformly in xE, the terms in (16) converge to zero. The term in (17) can be estimated as follows,

λn(x)Ef(y)Qn(dy,x)λ(x)Ef(y)Q(dy,x)λnEf(y)Qn(dy,x)Ef(y)Q(dy,x)+Ef(y)Q(dy,x)λnλ.

The latter expression tends to zero, since for all xE it was assumed that (15) holds, and since λnλ uniformly in xE. Thus, Item (i) of Theorem 5.11 holds. This implies that Item (iv) of Theorem 5.11 holds. The latter is equivalent to the assertion of this theorem.

Remark 5.17

Note that in the Feller case we can move to another external state only due to a purely random jump, i.e. a jump determined by Qn for nN or Q. Therefore, if we assume uniform convergence of the local characteristics across all state components, and in particular also QnQ in the sense of (15), then the result of Theorem 5.16 still holds.

Since uniform convergence of the local characteristics and the assumption that t(x)= are essential in the proof of Theorem 5.16, we need an alternative argument for situations with an active boundary or for situations in which a smooth approximation fails. A prototypical univariate example for both cases is a drift of the form g(x)=c1{xb} for some bR. Here one faces either a discontinuity or a subdivision of R into two state components, i.e. R={xR:xb}{xR:x>b}, with a continuous drift on each component. For a specific example, we find a method for dealing with this particular situation in the next section.

6. Application to the Cramér–Lundberg model with loan

In this section we apply our smoothing technique to the example presented in Section 2.1.4 and calculate the quantity of interest using different numerical integration methods. In this setup, ϕ1 solves the ODE (/t)ϕ1(y,t)=g1(ϕ1(y,t))yE1 and tR, with

g1(y)=cif y(0,),c+ρyif ycρ,0,0if y,cρ.

In the setup of Section 2.1.4, the quantity of interest is the expected value of discounted future dividend payments until the time of ruin. The cemetery Ec is given by Ec=({2}×E2)({3}×E3), the running reward ℓ is given by 10,4c, and the terminal cost is Ψ(x)=0 for xEc. For xE, t0, let

L(t,x)=0teδs(ϕ(s,x))ds.

Since g is not differentiable in 0 and t(x)< for all x=(1,y) with yE1, we replace g by a smoothed version and we also modify ℓ accordingly. For ϵ>0 let

g1ϵ(y)=cif y(ϵ,bϵ],c(by)315ϵ(yb)+6(by)2+10ϵ2ϵ5if y(bϵ,b),c+ρyif ycρ,ϵ,c+ρ(y+3ϵ)(yϵ)316ϵ3if y[ϵ,ϵ],0if y,cρ[b,).

Observe that g1ϵC2(R), that limybg1ϵ(y)=0 and that g1ϵ0. For ϵ>0 define the PDMP Xϵ so that for all yR its flow ϕ1ϵ(,y) is the solution to the ODE ddtϕ1ϵ(t,y)=gϵ(ϕ1ϵ(t,y)) with ϕ1ϵ(0,x)=x. Apart from that all specifications are the same as for the original PDMP X. In addition, we replace 1 by

1ϵ(y)=chyb+ϵϵ,

where h can be chosen as in (13) and we define

Lϵ(t,x)=0teδsϵ(ϕϵ(s,x))ds.

We aim at computing Gi1H for the smoothed process, in order to observe how (9) simplifies in this example. By the definition of the cemetery, Gi1H(x)=0 for all x=(k,z)E with k{2,3}. For x=(1,z) with zE1, any jumps bigger than z+c/ρ lead to the cemetery, so we only need to integrate over jump sizes up to z+c/ρ. Thus, we get that

GV(x)=GV((1,z))=0fW(t,x)eδtEV(y)Q(dy,ϕ(x,t))dt=0fW(t,x)eδt0z+c/ρV((1,zy))dFY(y)dt.

Moreover, since λ is constant on E1 it holds for all x=(1,z) with zE1, t0 that fW(t,x)=λNeλNt, where λN is as in Section 2.1.1. For x=(1,z) with zE1 we get

Gi1H(x)=t1=0λNe(λN+δ)t1y1=0χ1+cρti1=0λNe(λN+δ)ti1yi1=0χ(i1)+cρti=0λNeλNtiLϵ(ti,χ(i1))dtidFY(yi1)dti1dFY(y1)dt1, (18)

where the functions χj,χj, j=1,2,,i1 solve χj=ϕ1ϵ(tj,χj1) and χj=χjyj.

Thus χj depends on t1,,tj and y1,,yj1, whereas χj depends on t1,,tj and y1,,yj. However, this depensdence has been suppressed in (18) for the sake of readability.

Assumption 6.1

The jump distribution admits a density fY=FY, with fYC02.

In what follows, suppose that Assumption 6.1 holds. A variable substitution tj=ln(vj) and yj=(χj+c/ρ)zj, where vj[0,1],zj[0,1], χˆj(v1,,vj,z1,,zj)=χj(t1,,tj,y1,,yj). We then put

ν(v1,,vi,z1,,zi1)=Lϵ(ln(vi),χˆi1(v1,,vi1,z1,,zi1)),

which leads to

Gi1H(x)=v1=01vi=01z1=01zi1=01λNij=1i1vjδ+λN1viλN1ν(v1,,vi,z1,,zi1)×j=1i1fYzjχˆj+cρχˆj+cρdzjj=1idvj. (19)

Due to the recursive structure of the functions χˆ1,χˆ2,,χˆi1, the Jacobi matrix of the substitution has lower triangular shape, such that its determinant is the product of the diagonal elements. For being able to reasonably apply (10) we need to bound the Hessian of the integrand. If for example the jump size distribution is the Gamma distribution with parameters α,β>0, i.e. dFY(y)=(yα1βαeβy)/Γ(α)dy, then this boils down to the condition β3 and δ+λ>3, which implies that the integrand is bounded in 0. In the original problem statement this corresponds to an additional integrability condition on the jump size distribution. Finally, for xE of the form x=(4,b) we have

Gi1H((4,b))=0λeλteδt0b+c/ρGi2H((1,by))dFY(y)dt.

Remark 6.2

In Section 5.3 the stability, with respect to the smoothing parameter ε of the considered functional of the process, is dealt with in a fairly general setting. Unfortunately, because of the discontinuity of the drift g in the present example, we cannot achieve uniform convergence of the smoothed drift around the barrier level b, whereas point-wise convergence is achieved.

Theorem 6.3

In the setup of this section, the following assertion holds true. There exists C>0 such that VϵVCϵ.

Proof.

Recall that

V(x)=Ex0τeδs(Xs)dsandVϵ(x)=Ex0τϵeδsϵ(Xsϵ)ds,

where τ=inf{t0:XtEc} and τϵ=inf{t0:XtϵEc}.

It is readily checked that supy(c/ρ,bϵ)|g1(y)g1ϵ(y)|3ϵρ/16 and that |g1(y1)g1(y2)|ρ|y1y2| for all y1,y2(c/ρ,bϵ). Hence we get from Kamke (1964, Theorem 8, p. 111) that

|ϕ1ϵ(t,y)ϕ1(t,y)|<3ϵ16eρt1

for all y(c/ρ,bϵ) and for all t[0,min{θbϵϵ,θ~bϵ}], where

θbϵϵ=inf{t0:ϕ1ϵ(t,y)bϵ},θ~bϵ=inf{t0:ϕ1(t,y)bϵ},θ~b=inf{t0:ϕ1(t,y)b}.

Since g1ϵ and g1 coincide on (c/ρ,bϵ)(ϵ,ϵ) and g1ϵg10, we can refine this estimate to get

|ϕ1ϵ(t,y)ϕ1(t,y)|<3ϵ16eρC(ϵ)1,

for all t[0,min{θbϵϵ,θ~bϵ}], where C(ϵ)[0,) is the time needed for the trajectory ϕ1(,y) to cross (ϵ,ϵ). Note that g1ϵg10 yields ϕ1ϵ(t,y)ϕ1(t,y) for all t[0,θ~bϵ], and hence θ~bϵθbϵϵ. For tmin{θbϵϵ,θ~bϵ}=θbϵϵ we have by construction that |ϕ1ϵ(t,y)ϕ1(t,y)|ϵ. In total we get

|ϕ1ϵ(t,y)ϕ1(t,y)|ϵmax1,316eρC(ϵ)1. (20)

Since limϵ0C(ϵ)0, it holds that |ϕ1ϵ(t,y)ϕ1(t,y)|ϵ for sufficiently small ϵ>0.

Recall that T1 is the time of the first jump of X conditional on X0=(1,y). Since the jump intensity is constant on E1, T1 is exponentially distributed with intensity λN. Hence, we can write

V((1,y))=E(1,y)L(T1,(1,y))+eδT1V(XT1)=0λNeλNsL(s,(1,y))+eδsEV(x1)Q(dx1,ϕ1(s,y))ds=0λNeλNsL(s,(1,y))ds+0λNe(λN+δ)sEV(x1)Q(dx1,ϕ1(s,y))ds,

and analogously for Vϵ. We write V((1,y))=V1(y) and Vϵ((1,y))=V1ϵ(y) for yE1. Therefore,

|V1(y)V1ϵ(y)|0λNeλNs|L(s,(1,y))Lϵ(s,(1,y))|ds+0λNe(λN+δ)sEV(x1)Q(dx1,ϕ1(s,y))EVϵ(x1)Q(dx1,ϕ1ϵ(s,y))ds0λNeλNs|L(s,(1,y))Lϵ(s,(1,y))|ds+0λNe(λN+δ)sEV(x1)Q(dx1,ϕ1(s,y))EV(x1)Q(dx1,ϕ1ϵ(s,y))ds+0λNe(λN+δ)sEV(x1)Q(dx1,ϕ1ϵ(s,y))EVϵ(x1)Q(dx1,ϕ1ϵ(s,y))ds. (21)

For x=(1,y) and t0 it holds that Lϵ(s,x)=0 for sθb2ϵϵ, and

Lϵ(s,x)=0seδrϵ(ϕϵ(r,x))dr=c0seδrh((ϕϵ(r,x)b+ϵ)/ϵ)drcθb2ϵϵseδrdr

for sθb2ϵϵ. On the other hand, we have that, for x=(1,y) and s0, L(s,x)=0 for sθ~b, and

L(t,x)=cθ~bseδrdr

for s>θ~b. Using ϕ1ϵ(s,y)ϕ1(s,y) for all s0, we get θ~bθb2ϵϵ, such that

|Lϵ(s,(1,y))L(s,(1,y))|=Lϵ(s,(1,y))L(s,(1,y))cθb2ϵϵθ~beδrdr

for all t0. Hence,

0λNeλNs|L(s,(1,y))Lϵ(s,(1,y))|dscθb2ϵϵθ~beδrdrc(θ~bθb2ϵϵ).

Now θ~bθb2ϵϵ(b(b2ϵϵC1(ϵ)))/c=ϵ(2+C1(ϵ))/c, where C1(ϵ)=max(1,316(eρC(ϵ)1)), see (20). With this, the first term in (21) can be estimated by

0λNeλNs|L(s,(1,y))Lϵ(s,(1,y))|dsϵ(2+C1(ϵ)). (22)

Next, observe that (we remind the reader that the states xE are denoted by x=(k,y), which is why in the following the terms y1,y2 are not to be confused with the integration variables yj used in and below (18)),

EV(x1)Q(dx1,(1,y1))EV(x1)Q(dx1,(1,y2))=0y1+c/ρV1(y1z)fY(z)dz0y2+c/ρV1(y2z)fY(z)dzmin(y1,y2)+c/ρmax(y1,y2)+c/ρV1(z)fY(z)dzV1fY|y1y2|.

Combining this with (20), we can estimate the second term in (21) by

0λNe(λN+δ)sEV(x1)Q(dx1,ϕ1(s,y))EVϵ(x1)Q(dx1,ϕ1ϵ(s,y))ds (23)
λNλN+δV1fYsups0|ϕ1(s,y)ϕ1ϵ(s,y)|λNλN+δV1fYϵmax×1,316eρC(ϵ)1. (24)

Furthermore, since

EV(x1)Q(dx1,(1,y2))EVϵ(x1)Q(dx1,(1,y2))V1V1ϵ,

the third term in (21) can be estimated as follows,

0λNe(λN+δ)sEV(x1)Q(dx1,ϕ1ϵ(s,y))EVϵ(x1)Q(dx1,ϕ1ϵ(s,y))dsλNλN+δV1V1ϵ. (25)

Taking the supremum over yE1 in (21) and using (22), (23), and (25) we obtain that

V1V1ϵCϵ+λNλN+δV1V1ϵ

for some constant C and for sufficiently small ε. Thus,

δλN+δV1V1ϵCϵ,

which completes the proof.

6.1. Numerical experiment

We now solve the example presented above numerically. We set the following parameter values. The initial value of the PDMP x0=0, the premium income rate c=5, the credit rate ρ=0.05, the intensity of the Poisson process λ=4, the jump size distribution is for all x[0,) given by FY(x)=1eαx with α=1, and the discount rate δ=0.02. With this, the optimal dividend threshold according to Dassios & Embrechts (1989) is b=3.24289. Furthermore, we set the smoothing parameter ϵ=0.01. For computing the flow it is enough to solve the corresponding ODE once and to store the solution for repeated use.

We implemented Monte Carlo (random), QMC with the Sobol' sequence (Sobol), and QMC with a scrambled version of the Halton sequence (scrambled Halton), where scrambling refers to a permutation of digits (see, e.g. Owen (2000)). The Sobol' point generator we used was taken from Frances Y. Kuo's homepage Kuo (n.d.) and is based on Joe & Kuo (2008).

The reference solution was calculated using Monte Carlo with M=5000210=5120000 sample paths and d=1024, meaning that the maximum number of jumps we allow for is 512. In our plots we show the results plotted over an increasing number of integration nodes M{502j:1j16}.

Figure 3 shows the estimated standard deviation (root mean square error) of the estimation, which is calculated by using 50 repetitions with randomly shifted versions of our integration nodes.

Figure 3.

Figure 3.

The estimated standard deviation of the estimation.

Funding Statement

Peter Kritzer P. Kritzer is supported by the Austrian Science Fund (FWF): Project F5506-N26, which is part of the Special Research Program ‘Quasi-Monte Carlo Methods: Theory and Applications’. P. Kritzer is partially supported by the National Science Foundation (NSF) [grant number DMS-1638521] to the Statistical and Applied Mathematical Sciences Institute. P. Kritzer, G. Leobacher, and M. Szölgyenyi gratefully acknowledge the partial support of the Erwin Schrödinger International Institute for Mathematics and Physics (ESI) in Vienna under the thematic programme ‘Tractability of High Dimensional Problems and Discrepancy’. G. Leobacher is supported by the Austrian Science Fund (FWF): Project F5508-N26, which is part of the Special Research Program ‘Quasi-Monte Carlo Methods: Theory and Applications’. M. Szölgyenyi is supported by the AXA Research Fund grant ‘Numerical Methods for Stochastic Differential Equations with Irregular Coefficients with Applications in Risk Theory and Mathematical Finance’ and supported by the Vienna Science and Technology Fund (WWTF): Project MA14-031.

Acknowledgments

The authors would like to thank an anonymous referee for useful comments on how to improve the presentation of the results. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation. Part of this article was written while G. Leobacher was affiliated with the Institute of Financial Mathematics and Applied Number Theory, Johannes Kepler University Linz, Altenbergerstraße 69, 4040 Linz, Austria. A part of this article was written while M. Szölgyenyi was affiliated with the Institute of Statistics and Mathematics, Vienna University of Economics and Business, Welthandelsplatz 1, 1020 Vienna, Austria.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

  1. Albrecher H. & Kainhofer R. (2002). Risk theory with a nonlinear dividend barrier. Computing 68(4), 289–311. doi: 10.1007/s00607-001-1447-4 [DOI] [Google Scholar]
  2. Albrecher H. & Lautscham V. (2015). Dividends and the time of ruin under barrier strategies with a capital-exchange agreement. Anales del Instituo de Actuarios Espanoles 21(3), 1–30. [Google Scholar]
  3. Almudevar A. (2001). A dynamic programming algorithm for the optimal control of piecewise deterministic Markov processes. SIAM Journal on Control and Optimization 40(2), 525–539. doi: 10.1137/S0363012999364474 [DOI] [Google Scholar]
  4. Asmussen S. & Albrecher H. (2010). Ruin probabilities, 2nd ed Advanced Series on Statistical Science and Applied Probability Hackensack, NJ: World Scientific. [Google Scholar]
  5. Bakhvalov N. S. (1959). On the approximate calculation of multiple integrals. Vestnik MGU, Series Mathematical, Mechanics & Astronomy Physical Chemistry 4, 3–18. In Russian. [Google Scholar]
  6. Bäuerle N. & Rieder U. (2010). Optimal control of piecewise deterministic Markov processes with finite time horizon. Modern Trends in Controlled Stochastic Processes: Theory and Applications 123, 143. [Google Scholar]
  7. Bäuerle N. & Rieder U. (2011). Markov decision processes with applications to finance. Heidelberg: Universitext, Springer. [Google Scholar]
  8. Cai J., Feng R. & Willmot G. E. (2009). On the expectation of total discounted operating costs up to default and its applications. Advances in Applied Probability 41(2), 495–522. doi: 10.1239/aap/1246886621 [DOI] [Google Scholar]
  9. Colaneri K., Eksi Z., Frey R. & Szölgyenyi M. (2017). Optimal liquidation under partial information with price impact. arXiv:1606.05079
  10. Costa O. L. & Davis M. H. A. (1989). Impulse control of piecewise-deterministic processes. Mathematics of Control, Signals, and Systems (MCSS) 2(3), 187–206. doi: 10.1007/BF02551384 [DOI] [Google Scholar]
  11. Costa O. L. & Dufour F. (2013). Continuous average control of piecewise deterministic Markov processes. New York: Springer. [Google Scholar]
  12. Coulibaly I. & Lefèvre C. (2008). On a simple quasi-Monte Carlo approach for classical ultimate ruin probabilities. Insurance: Mathematics and Economics 42(3), 935–942. [Google Scholar]
  13. Dassios A. & Embrechts P. (1989). Martingales and insurance risk. Communications in Statistics. Stochastic Models 5(2), 181–217. doi: 10.1080/15326348908807105 [DOI] [Google Scholar]
  14. Davis M. H. A. (1984). Piecewise-deterministic Markov processes: A general class of nondiffusion stochastic models. Journal of the Royal Statistical Society Series B 46(3), 353–388. With discussion. [Google Scholar]
  15. Davis M. H. A. (1993). Markov models and optimization. Monographs on Statistics and Applied Probability London: Chapman & Hall. [Google Scholar]
  16. Davis M. H. A. & Farid M. (1999). Piecewise-deterministic processes and viscosity solutions. In W. M. McEneaney, G. George Yin, and Q. Zhang, eds., Stochastic Analysis, Control, Optimization and Applications. Boston: Springer. P. 249–268.
  17. de Saporta B., Dufour F. & Zhang H. (2016). Numerical methods for simulation and optimization of piecewise deterministic markov processes. Mathematics and Statistics Series London: ISTE; Hoboken, NJ: John Wiley & Sons, Inc. [Google Scholar]
  18. de Saporta B., Dufour F., Zhang H. & Elegbede C. (2012). Optimal stopping for the predictive maintenance of a structure subject to corrosion. Journal of Risk and Reliability 226(2), 169–181. [Google Scholar]
  19. Dempster M. A. H. & Ye J. J. (1992). Necessary and sufficient optimality conditions for control of piecewise deterministic Markov processes. Stochastics: An International Journal of Probability and Stochastic Processes 40(3–4), 125–145. [Google Scholar]
  20. Eichler A., Leobacher G. & Szölgyenyi M. (2017). Utility indifference pricing of insurance catastrophe derivatives. European Actuarial Journal 7(2), 515–534. doi: 10.1007/s13385-017-0154-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Embrechts P. & Schmidli H. (1994). Ruin estimation for a general insurance risk model. Advances in Applied Probability 26(2), 404–422. doi: 10.2307/1427443 [DOI] [Google Scholar]
  22. Ethier S. N. & Kurtz T. G. (1986). Markov processes. Wiley Series in Probability and Mathematical Statistics: Probability and Mathematical Statistics New York: John Wiley & Sons, Inc. [Google Scholar]
  23. Forwick L., Schäl M. & Schmitz M. (2004). Piecewise deterministic Markov control processes with feedback controls and unbounded costs. Acta Applicandae Mathematica 82(3), 239–267. doi: 10.1023/B:ACAP.0000031200.76583.75 [DOI] [Google Scholar]
  24. Grigorian A. (2009). Ordinary differential equation. Lecture notes. https://www.math.uni-bielefeld.de/grigor/odelec2009.pdf
  25. Hinrichs A., Novak E., Ullrich M. & Woźniakowski H. (2017). Product rules are optimal for numerical integration in classical smoothness spaces. Journal of Complexity 38, 39–49. doi: 10.1016/j.jco.2016.09.001 [DOI] [Google Scholar]
  26. Jacobsen M. (2006). Point process theory and applications. Probability and its Applications Boston, MA: Birkhäuser Boston, Inc. [Google Scholar]
  27. Joe S. & Kuo F. Y. (2008). Constructing Sobol' sequences with better two-dimensional projections. SIAM Journal of Scientific Computation 30, 2635–2654. doi: 10.1137/070709359 [DOI] [Google Scholar]
  28. Kallenberg O. (2002). Foundations of modern probability, 2nd Probability and its Applications (New York) New York: Springer-Verlag. [Google Scholar]
  29. Kamke E. (1964). Differentialgleichungen. I. Gewöhnliche differentialgleichungen. 5th ed Leipzig: Akademische Verlagsgesellschaft. [Google Scholar]
  30. Kritzer P., Pillichshammer F. & Wasilkowski G. W. (2016). Very low truncation dimension for high dimensional integration under modest error demand. Journal of Complexity 35, 63–85. doi: 10.1016/j.jco.2016.02.002 [DOI] [Google Scholar]
  31. Kuo F. Y. (n.d.). F. Y. Kuo's homepage. http://web.maths.unsw.edu.au/∼fkuo/sobol/index.html. Last visited 14/12/2017.
  32. Kuo F. Y., Sloan I. H., Wasilkowski G. W. & Woźniakowski H. (2010). Liberating the dimension. Journal of Complexity 26, 422–454. doi: 10.1016/j.jco.2009.12.003 [DOI] [Google Scholar]
  33. Kurtz T. G. & Protter P. E. (1996). Weak convergence of stochastic integrals and differential equations I. In D. Talay and L. Tubaro, eds., Probabilistic Models for Nonlinear Partial Differential Equations. Berlin, Heidelberg: Springer.
  34. Lenhart S. & Liaot Y. (1985). Integro-differential equations associated with optimal stopping time of a piecewise-deterministic process. Stochastics: An International Journal of Probability and Stochastic Processes 15(3), 183–207. doi: 10.1080/17442508508833356 [DOI] [Google Scholar]
  35. Leobacher G. & Ngare P. (2016). Utility indifference pricing of derivatives written on industrial loss indexes. Journal of Computational and Applied Mathematics 300, 68–82. doi: 10.1016/j.cam.2015.11.028 [DOI] [Google Scholar]
  36. Niederreiter H. (1992). Random number generation and quasi-Monte Carlo methods. CBMS-NSF Regional Conference Series in Applied Mathematics, Philadelphia: SIAM.
  37. Owen A. B. (2000). Monte Carlo, quasi-Monte Carlo, and randomized quasi-Monte Carlo. In H. Niederreiter and J. Spanier, eds., Monte Carlo and Quasi- Monte Carlo Methods 1998. Springer. P. 86–97.
  38. Pausinger F. & Svane A. M. (2015). A Koksma–Hlawka inequality for general discrepancy systems. Journal of Complexity 31, 773–793. doi: 10.1016/j.jco.2015.06.002 [DOI] [Google Scholar]
  39. Preischl M., Thonhauser S. & Tichy R. F. (2018). Integral equations, quasi-monte carlo methods and risk modeling. In J. Dick, F. Y. Kuo and H. Woźniakowski, eds., Contemporary Computational Mathematics – A Celebration of the 80th Birthday of Ian Sloan, Vol. 1, 2. Cham: Springer. P. 1051–1074.
  40. Riedler M. G. (2013). Almost sure convergence of numerical approximations for piecewise deterministic Markov processes. Journal of Computational and Applied Mathematics 239, 50–71. doi: 10.1016/j.cam.2012.09.021 [DOI] [Google Scholar]
  41. Rolski T., Schmidli H., Schmidt V. & Teugels J. (1999). Stochastic processes for insurance and finance. Wiley Series in Probability and Statistics New York: John Wiley & Sons. [Google Scholar]
  42. Schäl M. (1998). On piecewise deterministic Markov control processes: Control of jumps and of risk processes in insurance. Insurance: Mathematics and Economics 22(1), 75–91. [Google Scholar]
  43. Siegl T. & Tichy R. F. (2000). Ruin theory with risk proportional to the free reserve and securitization. Insurance: Mathematics and Economics 26(1), 59–73. [Google Scholar]
  44. Sloan I. H. & Woźniakowski H. (1998). When are quasi-Monte Carlo algorithms efficient for high dimensional integrals? Journal of Complexity 14, 1–33. doi: 10.1006/jcom.1997.0463 [DOI] [Google Scholar]
  45. Tichy R. F. (1984). Über eine zahlentheoretische Methode zur numerischen Integration und zur Behandlung von Integralgleichungen. Osterreichische Akademie der Wissenschaften Mathematisch-Naturwissenschaftliche Klasse. Sitzungsberichte. Abteilung II 193(4–7), 329–358. [Google Scholar]

Articles from Scandinavian Actuarial Journal are provided here courtesy of Taylor & Francis

RESOURCES