Skip to main content
Elsevier - PMC COVID-19 Collection logoLink to Elsevier - PMC COVID-19 Collection
. 2021 Dec 16;344:108758. doi: 10.1016/j.mbs.2021.108758

Optimal control of the SIR model with constrained policy, with an application to COVID-19

Yujia Ding 1, Henry Schellhorn 1,
PMCID: PMC8675184  PMID: 34922976

Abstract

This article considers the optimal control of the SIR model with both transmission and treatment uncertainty. It follows the model presented in Gatto and Schellhorn (2021). We make four significant improvements on the latter paper. First, we prove the existence of a solution to the model. Second, our interpretation of the control is more realistic: while in Gatto and Schellhorn (2021) the control α is the proportion of the population that takes a basic dose of treatment, so that α>1 occurs only if some patients take more than a basic dose, in our paper, α is constrained between zero and one, and represents thus the proportion of the population undergoing treatment. Third, we provide a complete solution for the moderate infection regime (with constant treatment). Finally, we give a thorough interpretation of the control in the moderate infection regime, while Gatto and Schellhorn (2021) focused on the interpretation of the low infection regime. Finally, we compare the efficiency of our control to curb the COVID-19 epidemic to other types of control.

Keywords: SIR model, Population control, COVID-19, Stochastic optimal control, Epidemiology

1. Introduction

This article extends the analysis of the model presented in [1]. In that article the authors gave analytical expressions for the optimal proportion of infected to undergo treatment in a pandemic. Analytical approaches allow to better understand the form of the solution compared to numerical approaches, which are currently more prevalent in this domain. It was possible for the authors to find formulae because of the type of objective function they chose. Rather than minimize an expected value of the number of infected or the cumulated number of infected (under a constraint on the dispersion of the results), we chose to minimize the expected value of a convex increasing function of the number of infected at the horizon. The quadratic utility function is a special case of the latter, and we indeed show results for a whole family of functions, called isoelastic functions. Each function in this family is characterized by a single parameter, called the risk-aversion parameter, and varying this parameter as we shall see enables us to better understand how the optimal control depends smoothly on the level of aversion to risk, where risk is understood as the probabilistic uncertainty of the result. In contradistinction, the optimal control of deterministic epidemiological models is often of the bang–bang type (see for instance [2], [3]). Several authors, such as [3], [4], [5], [6] use Pontryagin’s maximum principle. Some authors (e.g., [7]) use dynamic programming. Laarabi et al. [8] use the same framework, but add delay. The literature on the control of stochastic epidemiologic models tends to be more sparse and more recent. A variety of models and numerical methods has been proposed: Markov chain [9], backward SDE solved by the 4-step scheme [10], simulated annealing [11], stochastic programming [12], genetic algorithms [13]. Wang et al. [14] consider time-varying parameters.

The contributions of this article are fourfold. First, we prove existence of a solution. Second, whereas in [1] the optimal control α has the interpretation of the proportion of the population that takes a basic dose of treatment, so that α>1 occurs only if a proportion of the population takes more than a basic dose of treatment. In the low infection regime part of our paper, α is constrained to be between zero and one, and represents thus the proportion of the population undergoing treatment. The latter interpretation is much more realistic, as it is uncommon to ration treatment. Third, we provide a complete solution for the moderate infection regime (with constant treatment). The final improvement is a thorough numerical analysis and sensitivity analysis of the moderate infection regime, while [1] focused exclusively on the interpretation of the control in the low infection regime. This enables us to discover some errors in the second-order term of the solution in [1], which we correct here. Finally, we compare the efficiency of our control to curb the COVID-19 pandemic to other types of control. Our optimal control is, as expected, superior to a full control (or no control), in terms of expected utility. It is clearly superior in the case of low infection, but the benefit is less pronounced in the case of moderate infection. There are two possible reasons for that. First, we included only few terms in the analytic series of the optimal control. Adding more terms would have yielded slightly better results. More importantly, the quality of the treatment available to COVID-19 patients in the US in 2020 was probably not sufficient to make a difference if the number of infected had climbed to more than 1%.

The structure of the article is as follows. In Section 2 we briefly introduce the model in [1], and provide a proof of existence of the solution. In Section 3, we show our results for the low infection regime. In Section 4, we extend and analyze the solution in the moderate infection regime. Section 5 shows our experimental results when applying our methodology to the COVID-19 in the US in 2020. We draw the conclusion in Section 6.

2. A stochastic SIR model with treatment uncertainty

Let S, I, R be the proportions of susceptible, infected, and out of infection (recovered, and dead), respectively. Let β be the transmission rate and μ be the death rate.

In the SIR model, the rate of decrease dSdt of the proportion of susceptible is equal to the constant transmission rate β time SI. As in [1], we add a term σSSIdB1dt, where dB1dt is white noise, in order to model the error in the transmission rate:

dSdt=βSI+σSSIdB1dt

Infected patients are either treated or not treated against the disease. In both cases they either recover or die. We denote by μ0 (μ1) the constant death rate without (with) treatment and by K(t) the recovery rate of the treatment. The optimal policy α is a progressively measurable process that represents the proportion of the infected population that receives treatment, thus α(t)[0,1]. This constraint is an important addition to the model in [1]. Depending whether the individual is treated or not, there are then four different ways for an infected individual to exit the pool of infected:

  • not treated and recover

  • not treated and die

  • treated and recover

  • treated and died.

Thus, the “out of infection rate” will be:

dR(t)dt=(1α(t))I(t)K0not treated and recover+(1α(t))I(t)μ0not treated and die+α(t)I(t)K1(t)treated and recover+α(t)I(t)μ1treated and dieα(t)I(t)σdB2dttreatment measurement error (1)

For simplicity, we assume that the Brownian motion driving transmission uncertainty (B1) is independent from the Brownian motion driving treatment uncertainty (B2). Usually μ0μ1 (people die faster without treatment than with treatment), but not necessarily. Most of the time K1(t)>K0 (treatment is better than no treatment), but not necessarily. We relax this requirement somewhat by requiring:

P(K0<K1(t)) is close to one (2)

In order to keep the treatment rate within bounds, we model it as an Ornstein–Uhlenbeck process:

dK1(t)=λk(k¯1K1(t))dt+σkdB2(t)

with the mean-reversion rate λk>0 and the long run value of the treatment rate k¯1. It is well-known that K1 is Gaussian, with variance equal to:

Var[K1(t)]=σk22λk(1e2λkt)

Thus, if mean-reversion is large compared to volatility σk, constraint (2) is satisfied. We simplify (1) by:

dR(t)dtI(t)=K0+μ0+α(t)(K0+K1(t)μ0+μ1)α(t)σdB2dt

Putting everything together, the dynamics of the infected is:

dI(t)dt=βS(t)I(t)dR(t)dtσSS(t)I(t)dB1dt

We try to minimize a measure of the infected over our horizon T. This article focuses on the solution of the Mayer problem, i.e., the objective is expressed as a measure of I(T). Another possible control would have been the time-integral of the number of infected over the horizon (the Lagrange problem). Both figures have their merit,1 and we leave for future research the control of (a measure) of the Lagrange and Bolza problems. Rather than trying to minimize the expected value of the infected, namely E[I(T)], we include in our objective the risk caused by the uncertainty of the model and its observations. Decision-makers are notoriously risk-averse. For this reason, Morgenstern and Von Neumann [15] introduced a class of utility functions U that bears their names. A decision-maker in epidemiology that is averse to risk will thus minimize the expected utility of the infected, namely E[U(I(T))] where U is increasing and convex. Alternately, one can maximize the negative thereof, i.e., maximize the expected value of a concave and decreasing function of I(T). The policy obtained in maximizing the expected value of a concave utility function can be showed, under certain conditions, to maximize the expected value of the outcome (here I) under a constraint on the dispersion of the outcome. Out of the universe of concave decreasing utility functions, we choose the power utility function:

U(I)=I1γ1γ

The coefficient γ is often called the risk-aversion parameter . When γ=0 the decision-maker is risk-neutral, meaning that the uncertainty does not have an influence on her decisions. It is straightforward to check that this power utility function is concave in I when γ<0, which we will assume. The more negative γ the more risk-averse is the decision-maker. Taking for instance γ=1, we see that the objective is to

maxE[I22]

which returns the same policy as:

minE[I22]

The importance of analytic formulations is that other figures of interest in this model, like the expected number of deaths from treatment can be analytically calculated, and depend on γ. Thus, a decision-maker can calibrate its risk-aversion parameter γ on other goals. Expected number of deaths is only one type of goal and economic factors that can be easily added. We define

τ=min{t>0|I(t)=0 or I(t)=1}

Our controlled SIR model is thus:

sup0α(t)1E[I(min(τ,T))1γ1γ]
dS(t)=βS(t)I(t)dt+σSS(t)I(t)dB1(t) (3)
dI(t)=βS(t)(K0+μ0)+α(t)(K0K1(t)+μ0μ1)I(t)dt+α(t)I(t)σdB2(t)σSS(t)I(t)dB1(t) (4)
dK(t)=λk(K¯K(t))dt+σkdB2(t) (5)

The relative sign of our volatilities σ and σk is important. We will assume without loss of generality that σ<0. The sign of σk is the sign of covariance between the measured value of today’s treatment rate and the change in value of the treatment rate between today and a future date. An example may help illustrate the difference. Suppose that over a week one performs daily measurements of the treatment recovery rate as well as daily forecasts of the evolution of the treatment recovery rate over the next day. The two quantities measured each day t are proportional to the same white noise B2(t+1 day)B2(t). One then calculates weekly estimates σˆ of σ and σˆk of σk over these 7 daily observations. Since we arbitrarily choose σ<0, a negative σˆk shows a correlation of +1 between the measurement (of today’s treatment rate) and the forecast. Fig. 1 is a depiction of our model.

Fig. 1.

Fig. 1

A stochastic SIR model.

Let τ be the time when the infection stops, i.e., the first time when either I(τ) or S(τ) or both are equal to zero. We naturally require the control to go down to zero at that time, for definiteness reasons. The next result considers an open-loop control.

Theorem 1

For a given progressively measurable control 0α1 , with and given initial values (S(0),I(0),R(0),K(0)) there exists a unique solution of (3) (4) (5) up to time τ .

The proof of Theorem 1, included in Appendix A, follows the proof of a theorem of Yamada and Watanabe (1971), as exposed in the book by Karatzas and Shreve [16, Prop. 2.13, Sec. 5.2]. We showed in a companion document that the probability that either I or S is equal to zero over a finite interval is zero, following the method of proof in [17], [18]. It is thus unlikely that a simple discretization of our model results in negative values.2 For notational simplicity we define the impact of treatment risk X:

X(t)=K0+μ0μ1K1(t)σ

as well as the long run impact of the treatment risk X¯:

X¯=K0+μ0μ1k¯1σ

We define λx=λk and σx=σk/σ. For simplicity we write μ=K0+μ0.

In the absence of bounds on α, Gatto and Schellhorn [1] show that, when a smooth optimal closed-loop control α exists, there exists a probability measure P~ 3 under which I(t)eμt is a martingale and I is the optimal state. Moreover, I has the explicit form:

I(t)=F(Z(t),X(t),S(t),t) (6)
(Z(T))1/γ=I(T) (7)

and F is twice continuously differentiable. In the sequel we will shorten this statement by saying that “I solves the martingale problem (6)”.

3. Results in the low infection regime

We assume S(t) close to one and σ=0. Thus the term:

r=βS(t)(K0+μ0)βK0μ0

is assumed constant. With this simplification, we give an analytical solution to the constrained problem, i.e., the case where 0α(t)1, a significant improvement over [1], who considered the unconstrained case.

We consider first the case where the treatment rate is constant, and then the case where it follows an Ornstein–Uhlenbeck process.

3.1. Constant treatment rate

Let b=βμ1k¯1. The problem is:

sup0α(t)1E[I(T)1γ1γ]dI(t)=(r+α(t)(br))I(t)dt+α(t)σI(t)dB2(t) (8)

Theorem 2

The following constant control is optimal:

α=min(1,max(0,k¯1K0σ2|γ|))

The proof is in Appendix B, and follows closely [19].

3.2. Treatment rate as Ornstein–Uhlenbeck process

The problem is

supE[I(T)1γ1γ]dI(t)=(r+α(t)σX(t))I(t)dt+α(t)σI(t)dB2(t)dX(t)=λx(X¯X(t))dtσxdB2(t) (9)

In the low infection regime our solution will depend on a kernel H0(Xt,τ) with τ=Tt, while in the moderate infection regime it will also depend on two other kernels H1(Xt,τ) and H2(Xt,τ) that are closely related. In order to unify notation we define the kernels. Define

H0(Xt,τ)=exp1γ(A1(τ,γ)Xt22+A2(τ,γ)Xt+A3(τ,γ)+(1γ)(μ+r)τ) (10)

and, for i>0

Hi(Xt,τ)=expiγ(A1(τ,γ/i)Xt22+A2(τ,γ/i)Xt+A3(τ,γ/i)) (11)

where

A1(τ,γ)=1γγ2(1exp(θ(γ)τ))2θ(γ)(b2(γ)+θ(γ))(1exp(θ(γ)τ)) (12)
A2(τ,γ)=4λxX¯b1(γ)1expθ(γ)τ/22θ(γ)2θ(γ)(θ(γ)+b2(γ))(1exp(θ(γ)τ)) (13)
A3(τ,γ)=0τσx22γ+λxX¯A22(s,γ)+σx22A1(s,γ)+(γ1)μds (14)
b1(γ)=1γγb2(γ)=2(γ1γσxλx)b3(γ)=σx2γ
θ(γ)=b22(γ)4b1(γ)b3(γ)

We provide an explicit formula for A3(τ,γ) in Appendix C.

Gatto and Schellhorn [1, Prop. 1] provide an explicit solution to the PDE that F in (6) satisfies, but with some typos in the expression of H0(Xt,τ), which we correct here.

Theorem 3 Proposition 1 in [1]

If σx<0 then I solves the martingale problem (6) , where

I(t)=(Z(t))1γH0(Xt,Tt)

where

dZ(t)Z(t)=(r+X2(t))dt+X(t)dB2(t)
Z(0)=(I(0)H0(X(0),T))γ

The corresponding control α(t) is given by:

α0(t)=X(t)γσσxγσ(A0,1(Tt)X(t)+A0,2(Tt)) (15)

This control has some very clear properties. It is decomposed into a myopic policy X(t)γσ and a hedging policy, namely the second term of (15). Recall that X is most often negative. Both myopic control and hedging policies are thus inversely proportional to the degree of risk aversion |γ| and to the volatility σ which corresponds to the contemporaneous transmission measurement error. The hedging policy gives protection against the risk of making decisions too soon. As expected, its magnitude decreases as time approaches the horizon T. This is a typical feature of the Mayer problem, which is usually attenuated in the Bolza problem.

4. Results in the moderate infection regime

We first handle the Ornstein–Uhlenbeck treatment rate case, which was presented in [1, Prop. 2].

4.1. Treatment rate as Ornstein–Uhlenbeck process

The problem is defined in Section 2. We rewrite here for convenience,

supE[I(min(τ,T))1γ1γ]dS(t)=βS(t)I(t)dt+σS(t)I(t)dB1(t)dI(t)=βS(t)μ+α(t)σX(t)I(t)dt+α(t)I(t)σdB2(t)σSS(t)I(t)dB1(t)dX(t)=λx(X¯X(t))dtσxdB2(t) (16)

We further define

M~(t,τ)=2θ(γ)e12(b2(γ/2)b2(γ)θ(γ))(τt)2θ(γ)(b2(γ)+θ(γ))1eθ(γ)(τt) (17)
mY(τ,x)=xM~(t,τ)+s=tτM~(s,τ)(λxX¯+σx2γA2(τs,γ))ds+A2(Tτ,γ)A1(Tτ,γ)VY(τ,x)=σx2tτM~2(s,τ)dsg(X,t)=τ=tTH2(X,τt)12β2σS2γ112VY(τ,X)A1(Tτ,γ)/γ×exp(2γA3(Tτ,γ)A22(Tτ,γ)γA1(Tτ,γ) (18)
+mY2(τ,X)A1(Tτ,γ)γ2VY(τ,X)A1(Tτ,γ))dτ

From this, we can calculate:

gX=τ=tTH2(X,τt)12β2σS2γ112VY(τ,X)A1(Tτ,γ)/γ×exp2γA3(Tτ,γ)A22(Tτ,γ)γA1(Tτ,γ)+mY2(τ,X)A1(Tτ,γ)γ2VY(τ,X)A1(Tτ,γ)×(A1(τt,γ/2)X(t)+A2(τt,γ/2)γ/2+2mY(τ,X)A1(Tτ,γ)γ2VY(τ,X)A1(Tτ,γ)M~(t,τ))dτ

Theorem 4

Let I(0)=ɛ . If σx<0 then I solves the martingale problem (6) , where:

I(t)=ɛZ1/γ(t)H1(X(t),Tt)+ɛ2Z2/γ(t)S(t)g(X(t),t)+O(ɛ3)

where Z(t) satisfies:

dZZ=(μ+X2+β2SIσS2)dtβSIσSdB1+XdB2
Z(0)=H1(X(0),T)+H12(X(0),T)4ɛS(0)g(X(0),0)(O(ɛ2)1)2ɛS(0)g(X(0),0)γ

The corresponding control α(t)=α0(t)+ɛα1(t)+O(ɛ2) where:

α0(t)=X(t)γσσxγσA1(Tt,γ)X(t)+A2(Tt,γ)
α1(t)=Z1/γ(t)S(t)H1(X(t),Tt)σ(g(X(t),t)X(t)γσxgX+σxg(X(t),t)γA1(Tt,γ)X(t)+A2(Tt,γ))

The proof is in Appendix D. We refer to [1] for a discussion of α0. In the moderate infection regime, the control α0 (which is the control in the low infection regime) dominates the corrections due to transmission of the disease. Our asymptotic expansion thus indicates how the optimal control should change as the pandemic progresses. The sign of α1 is determined by the signs of σ and

g(X(t),t)γ(X(t)+σxA1(Tt,γ)X(t)+A2(Tt,γ))σxgX (19)

More specifically, α1 is positive if σ and (19) are both positive or negative. α1 is negative if one of them is positive and the other one is negative.

It is obvious that the magnitude of both g(X(t),t) and gX decrease with time and are equal to zero when t=T. Therefore, the importance of α1 decreases as time increases.

To further discuss the sign of (19), we rewrite it by

|g(X(t),t)γ|1+|σxA1(Tt,γ)|X(t)+X¯|σxA2(Tt,γ)X¯|+|σx|gX

Thus, suppose gX, X(t), and X¯ are all positive, (19) is positive, and vice versa. In the following cases, we provide two simple cases that we can easily discuss the sign of α1:

  • if gXσ>0, μμ1>max(K1(t),k¯1), then α1 is positive.

  • if gXσ<0, μμ1<min(K1(t),k¯1), then α1 is negative.

In the following, we discuss the full expansion of the solution in Theorem 4. Consider equation (57) in [1]:

(t+L1+ɛL2)f=0

This time we use full asymptotic expansion:

f=f1+ɛf2+=i=1fiɛi1

and obtain:

0=(t+L1)f1+i=1((t+L1)fi+1+L2fi)ɛi

The terms of our asymptotic expansion are thus determined by:

(t+L1)f1=0 (20)
(t+L1)fi+1=L2fii=1,2, (21)

We use the Ansatz:

fi(Z(t),X(t),t)=Z(t)2i1/γS(t)2i11gi(X(t),t)i=1,2,

We have showed that g1=H1 and g2=g in the proof of Theorem 4. By the same process, we can also calculate the expressions for g3, g4, in the sequel.

4.2. Constant treatment rate

The problem is:

supE[I(T)1γ1γ]dS(t)=βS(t)I(t)dt+σSS(t)I(t)dB1(t)dI(t)=(r+α(t)(br))I(t)dt+α(t)σI(t)dB2(t)σSS(t)I(t)dB1(t) (22)

The HJB equation of this problem is obtained by simplifying (6), i.e., making the value function independent from X. Let τ=Tt, the solution kernels hi(τ) for i=1,2, are given by:

hi(τ)=exp(2i1γ(ai,12(brσ)2+ai,2)τ) (23)

where

ai,1=1γ/2i1γ/2i1 (24)
ai,2=(γ/2i11)μ (25)

Theorem 5

Let I(0)=ɛ , then I solves the martingale problem (6) , where:

I(t)=i=1Z(t)2i1/γS(t)2i11gi(t)ɛi

where g1(t)=h1(Tt) , gi(t) , i>1 can be obtained by (47) , and Z(t) satisfies:

dZZ=(μ+(brσ)2+β2SIσS2)dtβSIσSdB1+brσdB2
1=i=1Z(0)2i1/γS(0)2i11gi(0)ɛi1

The corresponding control α(t) is equal to α0(t)+ɛα1(t)+O(ɛ2) , where α0(t) and α1(t) are equal to

α0=brγσ2α1=Z1/γ(t)S(t)g2(t)h1(Tt)brγσ2

The proof is in Appendix E, where we also provide a formula for g3. Observe that

g2(t)=β22σS2h2(Tt)h12(Tt)(brσ)2(a1,1a2,1)+2(a1,2a2,2)=β22σS2h2(Tt)h12(Tt)γμ(brσ)2/γ

is always positive because the signs of h2(Tt)h12(Tt) and γμ(brσ)2/γ are the same. The signs of α0 and α1 are determined by the sign of rbσ2.

5. Application to COVID-19

We use the same weekly US COVID-19 data from June 7, 2020 to November 1, 2020 as in ([1] Sec. 5.1) and the parameters in Table 1 are estimated using the COVID-19 data set. In the following, we show both simulation plots and one scenario real data plots under low infection regime with constant treatment, low infection regime with OU treatment, moderate infection regime with constant treatment, moderate infection regime with OU treatment, respectively. We compare three types of treatment:

  • no control, i.e., α(t)=0.

  • full control, i.e., α(t)=1.

  • optimal control, i.e. the control from Theorems 2, 3, 4, 5.

The Github repository for implementing the models and generating the plots can be found at https://github.com/yujiading/optimal-control-sir-model.

Table 1.

Parameters.

Treatment parameter Symbol Value
Death rate/no treatment μ0 0.0575
Death rate μ1 0.0575
Recovery rate/no treatment K0 0.2559
Recovery rate at time 0 K1(0) 0.2559
Long run value of recovery rate k¯1 0.4612
Volatility of the measurement of today’s recovery rate σ 0.4418
Volatility of changes in the recovery rate σk −1.1647
Speed of mean-reversion of the recovery rate λk 0.7692
Transmission rate β 0.025
Proportion of infected at time 0 ɛ 0.01
Time step Δt 0.001
Volatility of the measurement of today’s susceptible rate σS 2.17

5.1. Simulations

In Figs. 2, 3, 4, and 5 we use the Euler scheme to simulate 1,000,000 scenarios using the parameters in Table 1 and calculate the expected values of I(T) and utility of I(T) for each regime. The risk-aversion parameters γ that are considered are between 0.5 and 5. In all the regimes, the expected utility of our control is higher than full control and no control, as expected. This effect is more pronounced in the low infection than in the moderate infection regime. When γ becomes more negative, the expected utility of no control is higher than that of full control. This is because when γ becomes more negative, the decision-maker becomes more risk averse and trades off expected value of I(T) against dispersion of I(T). In the low infection regime, the expected number of infected is lower with full control than with no control, as expected, and, with the optimal control, it depends on the level of risk-aversion, as expected. When treatment is risky the more risk-averse a decision maker, the less he or she is likely to invest in treatment.

Fig. 2.

Fig. 2

Expected infection and expected utility of optimal control, full control, and no control of low infection regime with constant treatment. Using 1,000,000 simulation scenarios.

Fig. 3.

Fig. 3

Expected infection and expected utility of optimal control, full control, and no control of low infection regime with OU treatment. Using 1,000,000 simulation scenarios.

Fig. 4.

Fig. 4

Expected infection and expected utility of optimal control, full control, and no control of moderate infection regime with constant treatment. Using 1,000,000 simulation scenarios.

Fig. 5.

Fig. 5

Expected infection and expected utility of optimal control, full control, and no control of moderate infection regime with OU treatment. Using 1,000,000 simulation scenarios.

5.2. One scenario real data

In Fig. 6, we use the one scenario COVID-19 data as introduced above to plot the infections with optimal control, full control, and no control for each of the regimes. We can see that as gamma varies, some plots show contrafactual results. In fact, it is possible for a particular scenario to result in a better or worse outcome. This does not contradict our theoretical results, as in stochastic control the goal is not to optimize a single scenario.

Fig. 6.

Fig. 6

Infection of optimal control, full control, and no control of low infection regime with constant treatment, low infection regime with OU treatment, moderate infection regime with constant treatment, and moderate infection regime with OU treatment. Using real COVID-19 data.

6. Conclusion

We showed that a stochastic optimal control approach enables to more efficiently use treatment in an pandemic such as COVID-19. On a theoretical level, we show that, in a first approximation, the control does not depend on the transmission of the pandemic, i.e., the control in the moderate infection regime resembles the control in the low infection regime. The influence of key parameters of the problems, namely risk-aversion and volatility, are clearly demonstrated in our formulas. On a practical level, we show that an optimal control would have been better than a full control, had it been available in the low infection regime which we experienced in summer 2020 in the US for COVID-19. However the relative poor efficiency of the treatment that we experienced then would have translated into a poor performance of any type of control, had the pandemic moved into a moderate infection regime. In that regime, the influence of the type of control would have turned out not to be significant.

In the Mayer problem, which we study here, the control depends significantly on the horizon. The study of the full Bolza problem remains to be done. Many other interesting problems remain to be solved. For instance, we showed optimality of the constrained control only in the constant, low infection regime case. Verification theorems need to be worked out in the multiple treatment case or the Ornstein–Uhlenbeck case. Optimal vaccination is another area where we believe a similar asymptotic approach can be used. Finally, Bertozzi et al. [20] use Hawkes processes to model COVID-19. The control of Hawkes processes remains a largely open problem that deserves attention, in particular for its application to epidemiology.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Footnotes

1

We thank an anonymous referee for this comment.

2

We note that we never encountered a negative value of I(t) in our simulations.

3

Equivalent to the original measure P, which is the measure under which B1 and B2 are Brownian motions.

Appendix F

Supplementary material related to this article can be found online at https://doi.org/10.1016/j.mbs.2021.108758.

Appendix A. Proof of Theorem 1

We follow the proof in [16, Prop. 2.13, Sec. 5.2]. They consider the one-dimensional case. Let h:[0,)[0,) be a strictly increasing function with h(0) and

(0,ɛ)h2(u)=,   ɛ>0 (26)

In our case, we take h(x)=x.  Because of (26), there exists a strictly decreasing sequence {an}(0,1] with a0 and limnan=0 such that anan1h2(u)du=n. For every n there exists continuous function ρn on R with support on (an,an1) so that

0ρn(x)2nh(x),   x>0

and an1anρn(u)du=1. Then the function

Ψn(x)=0|x|0yρn(u)dudy

is even and twice continuously differentiable, with |Ψn(x)|1 and limn Ψn(x) =|x|. Suppose there are two strong solutions (I(1),S(1)) and (I(2),S(2)),

d(I(1)I(2)E[I(1)I(2)])ασ(I(1)I(2))dB2=σSS(1)I(1)S(2)I(2)dB1=d(S(1)S(2)E[S(1)S(2)])

so that

(d(I(1)I(2)))2<σS2(S(1)I(1)S(2)I(2))dt+(ασ)2(I(1)I(2))2dt(d(S(1)S(2)))2<σS2(S(1)I(1)S(2)I(2))dtd((I(1)I(2))(S(1)S(2)))<σS2(S(1)I(1)S(2)I(2))dt

Thus, since |Ψn|<1,

E[dΨn(It(1)It(2))+dΨn(St(1)St(2))]=E[Ψn(It(1)It(2))(d(I(1)I(2)))+Ψn(St(1)St(2))(d(S(1)S(2)))]+12E[Ψn(It(1)It(2))(d(I(1)I(2)))2]+12E[Ψn(St(1)St(2))(d(S(1)S(2)))2]E[2|β||S(1)I(1)S(2)I(2)|dt+|D||I(1)I(2)|dt]+12E[Ψn(It(1)It(2))σS2(S(1)I(1)S(2)I(2))dt]+12E[Ψn(St(1)St(2))σS2(S(1)I(1)S(2)I(2))dt]+12E[Ψn(It(1)It(2))(ασ)2(I(1)I(2))2]dt

where

D=(K0+μ0)+α(K0K1+μ0μ1)

Observe that:

S(1)I(1)S(2)I(2)=S(1)(I(1)I(2))+I(2)(S(1)S(2))<|I(1)I(2)|+|S(1)S(2)|

Since Ψn<2/nh and h is positive,

Ψn(It(1)It(2))+Ψn(St(1)St(2))S(1)I(1)S(2)I(2)<2n1h(|I(1)I(2)|)+1h(|S(1)S(2)|)(|I(1)I(2)|+|S(1)S(2)|)<2n|I(1)I(2)|h(|I(1)I(2)|)+|S(1)S(2)|h(|S(1)S(2)|)

Taking h(x)=x results in

E[dΨn(It(1)It(2))+dΨn(St(1)St(2))]<(E[(2|β|+|D|)|I(1)I(2)|]+E[2|β||S(1)S(2)|]+2σS2n+(ασ)22E[|I(1)I(2)|])dt

Since limnΨn(x)=|x|,

E[|It(1)It(2)|+|St(1)St(2)|]<0tE[(2|β|+|Ds|)|Is(1)Is(2)|]+E[2|β||Ss(1)Ss(2)|]+(αsσ)22E[|Is(1)Is(2)|]ds

But,

E[(2|β|+|Ds|)|Is(1)Is(2)|]<E[(2|β|+|Ds|)2]E[|Is(1)Is(2)|2]

Since |Is(1)Is(2)|<1, E[(Is(1)Is(2))2]<1 and E[(Is(1)Is(2))2]<E[|Is(1)Is(2)|] thus

E[|It(1)It(2)|+|St(1)St(2)|]<0tE[(2|β|+|Ds|)2]+(ασ)22E[|Is(1)Is(2)|]+2|β|E[|Ss(1)Ss(2)|]ds<0tmaxE[(2|β|+|Ds|)2]+(ασ)22,2|β|×E[|Is(1)Is(2)|+|Ss(1)Ss(2)|]ds

and local uniqueness follows by Gronwall’s inequality.

Appendix B. Proof of Theorem 2

We refer to the problem treated by Gatto and Schellhorn [1] as the unconstrained problem. Indeed, in that problem α was not constrained. We refer to our problem as the constrained problem. We follow the method of proof in [19], referred to hereafter as CK. They introduce auxiliary problems, which are unconstrained. They show that there exists an auxiliary problem which solution can be used to construct the solution of the original constrained problem. We follow the numbering of the sections in CK in order to ease understanding.

CK Section 2. The model.

To ease the correspondence with the CK paper, we define br=K0+μ0μ1k¯1, θ(br)/σ, and

H(0)(t)=exp(rt)exp(0tθdB2(s)120tθ2ds)

Observe that E[0tθ2ds]<.

CK Section 3. Portfolio and consumption processes.

Define:

B2(0)(t)=B2(t)+0tθds

Denote by Ii,α the infected process subject to I(0)=i and control α. It is admissible if

0Ii,α(t)1    0tT

The set of admissible α is denoted A0(i). Note that (See (3.5) in CK)

H(0)(t)I(t)=i+0tH(0)(s)I(s)(α(s)σθ)dB2(s)

CK Section 4. Convex sets and their support functions.

The difference between CK and this paper is that our objective is to minimize. This means that the key relation between our auxiliary infected and infected is reversed compared to the first equation in CK. Indeed if αν solves the auxiliary problem and α the original problem, we must have:

Iνi,αν(t)Ii,α(t)

Define

δ(ν)=0ν<0νν>0 (27)

It is subadditive:

δ(λ+ν)δ(λ)+δ(ν) (28)

CK Section 5. Utility functions.

The main difference between our utility functions and the utility functions in financial economics is that our utility functions are decreasing for positive arguments. Recall indeed that our utility function is, for γ<0:

U(i)=i1γ1γ

Since

U(i)=iγ

We have limiU(i)= and limi0U(i)=0, again for γ<0. This is unlike CK and [21] who consider the case 0<γ<1 with utility of wealth U2(x)=x1γ1γ. In their case, limxU2(x)=0 and limx0U2(x)=.

We define I2 to be the inverse of U, with I2(y) on y0. By straightforward calculations:

I2(y)=(y)1/γ

We also define the Legendre–Fenchel dual

U~(y)=maxx>0[U(x)xy]=U(I2(y))yI2(y)

This function satisfies:

U~(y)=I2(y)y0

CK Section 6. The constrained and unconstrained optimization problems.

We define:

A(i)={αA0(i)|0α1}

The supremum of the unconstrained problem is denoted by V0, while the supremum of the constrained problem is denoted by V, namely:

V0(i)=supαA0(i)E[U(Ii,α(T))|I(0)=i]
V(i)=supαA(i)E[U(Ii,α(T))|I(0)=i]

CK Section 7. Solution of the unconstrained problem.

We note that the expectation

X0(y)E[H(0)(T)I2(yH(0)(T))]

is finite for every y(,0]. We define its inverse Y0:

Y0(X0(y))=y

The solution of the unconstrained problem is well-known, and equal to:

α(s)=θσγ=rbσ2|γ|

CK Section 8. Auxiliary unconstrained optimization problems.

Recall δ(ν) in (27). It is easily seen that:

ανδ(ν)=ανν<0(α1)νν>00

We introduce a new process I(ν) by:

dI(ν)(t)I(ν)(t)=(r+α(t)ν(t)δ(ν(t)))dt+α(t)σdB2(0)(t)

Likewise we introduce

θ(ν)=θ+ν/σ
B(ν)(t)=B2(t)+0tθ(ν)(s)ds
H(ν)(t)=exp(rt+0tδ(ν(s))ds)E(0tθ(ν)(s)dB2(s))
E(0tθ(ν)(s)dB2(s))exp(0tθ(ν)(s)dB2(s)120t(θ(ν)(s))2ds)

We denote by Aν(i) the class of α for which

Iνi,α(t)1

Since the solution of our dual problem will have α(t)ν(t)δ(ν(t))0, clearly A(i)Aν(i). We define:

Vν(i)=supπAν(i)E[U(Ii,α(T))]
Xν(y)E[H(ν)(T)I2(yH(ν)(T))]

We define a class of progressively measurable processes ν in R by:

D={ν;E0Tδ(ν(t))dt,Eν2(t)dt<,Xν(y)<,y(,0]}

Proposition 8.3. in CK shows that, if for some λ D the corresponding control αλ is optimal for the auxiliary optimization problem and if

δ(λ)+αλ(t)λ(t)=0

then αA(i) and is optimal for the constrained problem.

The solution of the unconstrained problem is:

α(s)=θ(ν)σγ=θ+ν/σσγ=rbνσ2|γ| (29)

CK Section 9. Contingent claims attainable by constrained portfolios.

We sketch the proof of Theorem 9.1 in CK, as the signs are different, and the structure of the control is slightly different.

CK 9.1 Theorem

Let B be a positive FT -measurable random variable and suppose there is a process λD such that, for all νD

E[H(ν)(T)B]E[H(λ)(T)B]i (30)

Then there exists a control αA(i) such that Ii,α=B .

Sketch of Proof

See CK p.782 for a definition of the stopping time τn. By (30) and subadditivity of δ (28):

0limɛ0sup1ɛE[(H(λ)(T)H(λ+ɛ(νλ))(T))B]=limɛ0sup1ɛE[H(λ)(T)B(1exp(0Tτnδ(λ(t)+ɛ(ν(t)λ(t)))δ(λ(t))dt×E(0Tτn(θ(λ)(t)+θ(λ+ɛ(νλ))(t))dB2(λ)(t))))]limɛ0supE[H(λ)(T)B(LT+NT)]

where

δ˘(ν)(λ(t))=δ(λ(t))ν=0δ(ν(t)λ(t))otherwise
LT=0Tτnδ˘(ν)(λ(t))dt
NT=0Tτnν(t)λ(t)σdB2(λ)(t)

By Ito’s lemma.

d[H(λ)(t)I(t)(Lt+Nt)]=I(t)H(λ)(t)d(Lt+Nt)+(Lt+Nt)H(λ)(t)I(t)α(t)σdB2(λ)(t)+I(t)H(λ)(t)α(t)(ν(t)λ(t))dt

which implies

H(λ)(T)I(T)(LT+NT)=0τnI(t)H(λ)(t)(ν(t)λ(t)σ+(Lt+Nt)σα(t))dB2(λ)(t)+0τnH(λ)(t)I(t)(α(t)(ν(t)λ(t))dt+dLt)

Therefore,

0E[H(λ)(T)B(LT+NT)]=E[0τnH(λ)(t)I(t)(α(t)(ν(t)λ(t))dt+dLt)]

It is easy to see that, for any ρD, take ν=λ+ρ:

δ(ρ(t))+α(t)ρ(t)0 (31)

and, taking ν(t)=0, we obtain:

δ(λ(t))+α(t)λ(t)0

which together with (31) for ρ=λ yields:

δ(λ(t))+α(t)λ(t)=0

CK Section 10. Equivalent optimality conditions.

The most important implication to prove is (D) (B) (A) in CK. It shows that the solution of the dual problem solves the auxiliary problem, and that, moreover, it is feasible and optimal for the original constrained problem. We make it more explicit here.

(Part of) CK 10.1 Theorem

Suppose that for every νD ,

E[U~(Yλ(i)H(λ)(T))]E[U~(Yλ(i)H(ν)(T))] (32)

then there exists a control αλ[0,1] that is optimal for the constrained problem Vλ(i)=E[U(Ii,αλ(T))] and such that

Vλ(i)=V(i)

Proof

E[U~(Yλ(i)H(λ)(T))]E[U~(Yλ(i)H(λ+ɛ(νλ))(T))]

Since U~(y)=I2(y),

0limɛ0sup1ɛE[U~(Yλ(i)H(λ+ɛ(νλ))(T))U~(Yλ(i)H(λ)(T))]=Yλ(i)limɛ0sup1ɛE[I2(Yλ(i)H(λ)(T))(H(λ)(T)H(λ+ɛ(νλ))(T))]

By Theorem 9.1 there exists a control αλAλ(i) such that:

Ii,αλ(T)=I2(Yλ(i)Hλ(T))

Clearly αλ is optimal for the constrained problem, and

δ(λ)+αλ(t)λ(t)=0

Thus by proposition 8.3, αλ is optimal for the constrained problem. □

CK Section 12. A dual problem.

Define:

Vˆ(y)=infνDE[U~(yHν(T))]

In our case,

U~(y)=maxx>0[x1γ1γxy]

Thus

y=U(x)=xγI2(y)=(y)1/γ

Let ρ=(1γ)/γ. Then:

U~(y)=(y)ρ/ρ

Typically, γ=1, so that:

U~(y)=y2/2

The main problem in condition (32) is to find the optimal process H(λ) (across all H(ν)) but it depends on y which depends on λ. Thus the dual must be fixed for a fixed but arbitrary real number y. The objective has the form

E[U~(yH(ν)(T))]=E[U(I2(yH(ν)(T)))yH(ν)(T)I2(yH(ν)(T))]

The right hand-side of the equation (see [22, p.134]) is the maximum of the function h(B,y)L(B,y) for all non-negative FT measurable B with E[H(ν)(T)B]i. Thus a minimization over all positive numbers y of h(B,y) would yield the optimal utility of the unconstrained problem. We could thus first minimize E[U~(yH(ν)(T))] in y, and then minimize over ν. However, the main idea is to first minimize over μ, and then minimize over y, hoping that the two can be interchanged.

CK 12.1 Proposition

Suppose that for any y there exists λy such that Vˆ(y)=E[U~(yH(λy)(T))] . Then there exist an αA(i) with i=Xλy(y) which is optimal for the primal problem, and we have:

Vˆ(y)=supi[V(i)iy]

Proof

Write λ for λYλ(i). Then

E[U~(Yλ(i)H(λ)(T))]E[U~(Yλ(i)H(ν)(T))]

and we conclude by CK Theorem 10.1. □

CK Section 15. Deterministic coefficients and feedback formulae.

Define:

Q(y,t)=E[U~(yH(ν)(T))|yH(ν)(t)=y]

Recall

dH(ν)H(ν)=(r+δ(ν))dt(θ+ν/σ)dB2

The HJB equation is:

minν12y2(θ+ν/σ)2Qyy+y(r+δ(ν))Qy+Qt=0
Q(T,y)=U~(y)=(y)ρρ

Again, with ρ=(1γ)/γ<0. We choose

Q(y,t)=1ρ(y)ρv(t)

Thus

12y2(θ+ν/σ)2Qyy+y(r+δ(ν))Qy=12(ρ+1)(y)ρv(t)(θ+ν/σ)2+(r+δ(ν))(y)ρv(t)

Dividing by (y)ρv(t), the problem becomes:

argminν1+ρ2(θ+ν/σ)2+δ(ν) (33)

Recall that if ν is positive, then δ(ν)=ν thus we solve (33) and obtain

ν=σ21+ρ+rb=σ2|γ|+rb

since 1+ρ=1/γ and γ is negative. If ν is negative, then δ(ν)=0, thus ν=rb.

From (29), the solution is

α(s)=rbνσ2|γ|=min(1,max(0,rbσ2|γ|))

Suppose μ0=μ1 and treatment is better than no treatment k¯1>K0. Thus rb=k¯1K0 is positive. Thus

α(s)=min(1,max(0,k¯1K0σ2|γ|))

Appendix C. Explicit formula of A3(τ,γ) in (14)

A3(τ,γ)=0τσx22γ+λxX¯A22(s,γ)+σx22A1(s,γ)+(γ1)μds=σx22γ+λxX¯(2λxX¯b2(γ)A2(τ,γ)θ3(γ)b3(γ)+2X¯2λx2θ3(γ)(A1(τ,γ)b3(γ)+8b12(γ)τlog2θ(γ)θ(γ)+b2(γ)1eθ(γ)τ2θ(γ)(b2(γ)θ(γ))b1(γ)b3(γ)+b2(γ)(θ(γ)2b1(γ)b3(γ))θ(γ)b32(γ)×logb2(γ)2b1(γ)b3(γ)θ(γ)2b2(γ)+4b1(γ)b3(γ)eθ(γ)τ/22θ(γ)b2(γ)+θ(γ)1eθ(γ)τ(b2(γ)+θ(γ))A1(τ,γ)2b1(γ)+4b12(γ)τ(b2(γ)+θ(γ))2))+σx221b3(γ)log2θ(γ)eθ(γ)τ2θ(γ)b2(γ)+θ(γ)1eθ(γ)τ2b1(γ)τb2(γ)+θ(γ)+(γ1)μτ

Appendix D. Proof of Theorem 4

We follow the proof of Proposition 2 in [1]. Recall the equations (58) (59) in [1] and following same notations:

(t+L1)f1=0 (34)
(t+L1)f2=L2f1 (35)

where L1,L2 are equations (53) (54) in [1].

Solution of (34).

We postulate that:

f1(Z,X,t)=Z1/γH1(X,Tt)

Substitution in (34) shows that H1 solves:

(t+Lγ)H1=0 (36)
H1(X,0)=1

where the operator Lγ is defined by:

LγH12σx22HX2+((γ1γσxλx)X+λxX¯)HX+(X2(121γ(1γ1))+μ(11γ))H

Using the Ansatz (11), we can rewrite the LHS of (36) into:

(C1(t)X2+C2(t)X+C3(t))H1/γ=0

Clearly all terms C1,C2,C3 must be identically zero. Thus:

dA1(t,γ)dt=σx2γA12(t,γ)+2(γ1γσxλx)A1(t,γ)+1γγ
dA2(t,γ)dt=σx2A1(t,γ)γA2(t,γ)+(γ1γσxλx)A2(t,γ)+λxX¯A1(t,γ)
dA3(t,γ)dt=σx22(A1(t,γ)+A22(t,γ)γ)+λxX¯A2(t,γ)μ(1γ)

which admit the solutions (12), (13), (14).

Solution of (35).

The second equation can be rewritten

(t+L1)f2=12β2γσS2Z2/γSH1(X,Tt)2 (37)

We try the Ansatz:

f2(Z(t),X(t),t)=Z(t)2/γS(t)g(X(t),t) (38)

Thus

(t+Lγ/2)g(X,t)=12β2σS2γH1(X,Tt)2
g(X,T)=0

We use Lemma 6 to obtain the g(X,t) in (18).

The optimal policy is:

α=1σFFZXZFXσx=α0+ɛα1+O(ɛ2)

where

α0=f1ZXZσf1f1Xσxσf1=X(t)γσσxγσA1(Tt,γ)X(t)+A2(Tt,γ)
α1=XZσf1f2Zf1Zf2f1σxσf1f2Xf1Xf2f1=Z1/γ(t)S(t)H1(X,Tt)σg(X(t),t)X(t)γσxgX+σxg(X(t),t)γA1(Tt,γ)X(t)+A2(Tt,γ)

Lemma 6

Let u(x,t)=12β2σS2γH1(x,Tt)2 . The solution to

g(x,t)t+Lγ/2g(x,t)=u(x,t) (39)
g(x,T)=0

is in (18) .

Sketch of Proof

Define m(x) and r(x) to be such that:

Lγ/2f(x,t)=12σx22f(x,t)x2+m(x)f(x,t)xr(x)f(x,t)
m(x)=(γ/21γ/2σxλx)x+λxX¯
r(x)=(x21γ(2γ1)+μ(12γ))

Let f(x,t) be the solution of:

f(x,t)t+12σx22f(x,t)x2+m(x)f(x,t)x=r(x)f(x,t) (40)

Defining:

dX(t)=m(X)dt+σxdW(t) (41)

we see that:

f(x,t)t+12σx22f(x,t)x2+m(x)f(x,t)x=E[df(X,t)|X(t)=x]/dt (42)

Thus (40) can be rewritten:

E[df(X(t),t)r(X(t))f(X(t),t)dt|X(t)]=0

Using the integrating factor exp(0tr(X(s))ds), we have:

E[d(exp(0tr(X(s))ds)f(X(t),t))|X(t)]=0

Under the boundary condition f(X(T),T)=1 the only possible solution is:

f(x,t;T)=E[exp(tTr(X(s))ds)|X(t)=x]

Define P(t,T)=f(X(t),t;T)=H2(X(t),Tt). Clearly:

dP(t,T)P(t,T)=r(X(t))dt+v(t,T)dW(t)

where:

v(t,T)=σxfxf

By Ito’s lemma, and for the exact same reason as (42):

g(x,t)t+12σx22g(x,t)x2+m(x)g(x,t)x=E[dg(X,t)|X(t)=x]/dt

The stochastic equivalent of (39) is:

E[dg(X(t),t)r(X(t))g(x,t)dt|X(t)]=E[u(X(t),t)dt|X(t)]

The solution is:

g(X(t),t)=τ=tTQ(t,τ)dτ

where:

Q(t,τ)=E[exp(tτr(X(s))ds)u(X(τ),τ)|X(t)]

Clearly, for some volatility σQ(t,τ)

dQ(t,τ)Q(t,τ)=r(X(t))dt+σQ(t,τ)dW(t)

We are now ready to define a change of numeraire. Let

dWτ=dWv(t,τ)dt

By Theorem 9.2.2 in [23], Q(t,τ)/P(t,τ) is a Pτ-martingale, i.e., 

Q(t,τ)=P(t,τ)Etτ[u(X(τ))]

where

dX(t)=m(X)dt+σxdW(t)=m(X)dt+σx(dWτ(t)+v(t,τ)dt)

From (11),

u(X(t),t)=12β2σS2γe2γA1(Tt,γ)2X(t)2+A2(Tt,γ)X(t)+A3(Tt,γ)

Let us now take:

P(t,T)=exp(2γ(A1(Tt,γ/2)2X2(t)+A2(Tt,γ/2)X(t)+A3(Tt,γ/2)))

Thus:

v(t,τ)=σxγ(A1(τt,γ/2)X(t)+A2(τt,γ/2))
dX(t)=[(γ21γ/2σxλx+σx2γA1(τt,γ/2))X(t)+λxX¯+σx2γA2(τt,γ/2)]dt+σxdWτ(t) (43)

Thus Etτ[u(X(τ))] when (43) holds can be calculated exactly the same way as E[u(X(τ))] when (41) holds. The structure is also affine, and there will be a solution of the form:

Etτ[u(X(τ),τ)]=12β2σS2γEtτe2γA1(Tτ,γ)2X2(τ)+A2(Tτ,γ)X(τ)+A3(Tτ,γ)

To summarize, since P(t,T)=H2(X(t),Tt)

g(x,t)=τ=tTP(t,τ)Etτ[u(X(τ),τ)]dτ=τ=tTH2(X,τt)12β2σS2γEtτe2γA1(Tτ,γ)2X2(τ)+A2(Tτ,γ)X(τ)+A3(Tτ,γ)dτ (44)

Let M~(t,τ) as in (17) and

Y(τ)=X(τ)+A2(Tτ,γ)A1(Tτ,γ)

Clearly:

Eτ[X(τ)|X(t)=x]=xM~(t,τ)+s=tτM~(s,τ)(λxX¯+σx2γA2(τs,γ))ds
Varτ[X(τ)|X(t)=x]=σx2tτM~2(s,τ)ds

Thus we can calculate:

mY(τ,x)=Eτ[Y(τ)|X(t)=x]=Eτ[X(τ)|X(t)=x]+A2(Tτ,γ)A1(Tτ,γ)
VY(τ,x)=Varτ[Y(τ)|X(t)=x]=σx2tτM~2(s,τ)ds

We can further develop:

Etτ[exp(2γ(A1(Tτ,γ)2X2(τ)+A2(Tτ,γ)X(τ)+A3(Tτ,γ)))]=Etτe2γA3(Tτ,γ)+1γA1(Tτ,γ)X(τ)+A2(Tτ,γ)A1(Tτ,γ)2A22(Tτ,γ)γA1(Tτ,γ)=Etτe2γA3(Tτ,γ)A22(Tτ,γ)γA1(Tτ,γ)eA1(Tτ,γ)γX(τ)+A2(Tτ,γ)A1(Tτ,γ)2=e2γA3(Tτ,γ)A22(Tτ,γ)γA1(Tτ,γ)12πVY(τ,x)eA1(Tτ,γ)γy2e(ymY(τ,x))22VY(τ,x)dy=e2γA3(Tτ,γ)A22(Tτ,γ)γA1(Tτ,γ)+mY2(τ,x)A1(Tτ,γ)γ2VY(τ,x)A1(Tτ,γ)112VY(τ,x)A1(Tτ,γ)/γ

providing γ<2A1(Tτ,γ)VY(τ,x). Thus Eq. (18) follows from (44). □

Appendix E. Proof of Theorem 5

When X(t) is a constant, equations (53) and (54) in [1] become

L1F=12Z2(brσ)22FZ2μZFZ+μF
L2F=12β2σS2ZSFFZ

Use the Ansatz f1(Z(t),t)=Z1/γ(t)h1(Tt) and insert in (20) shows that h1 solves:

t+Lγh1=0h1(0)=1 (45)

where the operator Lγ is defined by:

LγH((brσ)2(121γ(1γ1))+μ(11γ))H

Using the Ansatz (23), we can rewrite (45) into:

(C1(t)(brσ)2+C2(t))h1/γ=0

Clearly all terms C1,C2 must be identically zero. Thus

da1,1tdt=1γγda1,2tdt=μ(γ1)

which admit the solutions (24), (25) at i=1.

Now use fi(Z(t),t)=Z2i1/γ(t)S2i11(t)gi(t). We can rewrite (21) by

t+Lγ/2igi+1(t)=12β22i1σS2γgi2(t)gi+1(T)=0 (46)

Let u(t)=12β22i1σS2γgi2(t) and

ri=2iγ(12(brσ)2ai+1,1+ai+1,2)

Then

Lγ/2igi+1(t)=rigi+1(t)

and the stochastic equivalent of (46) is:

gi+1(t)t+rigi+1(t)=u(t)gi+1(T)=0

which admits

gi+1(t)=12β22i1σS2γ1hi+1(t)tTgi2(s)hi+1(s)ds (47)

We have showed that g1(t)=h1(Tt). Here we also provide the g2 and g3 in the following:

g2(t)=β22σS2h2(Tt)h12(Tt)(brσ)2(a1,1a2,1)+2(a1,2a2,2)
g3(t)=β616σS61(brσ)2(a1,1a2,1)+2(a1,2a2,2)21h3(t)h22(T)h3(T)h22(T)h3(t)h22(t)a3,1a2,12(brσ)2+a3,2a2,2+h14(T)h3(T)h14(T)h3(t)h14(t)a3,1a1,12(brσ)2+a3,2a1,22h2(T)h12(T)h3(T)h12(T)h2(T)h3(t)h12(t)h2(t)a3,1a2,1a1,122(brσ)2+a3,2a2,2a1,22

Suppose we use the first two expansions, the optimal policy is given by:

α=α0+ɛα1+O(ɛ2)

where

α0=f1ZbrσZσf1=brσγσ
α1=brσZσf1f2Zf1Zf2f1=Z1/γ(t)S(t)h1(Tt)σg2(t)brσγ

Appendix F. Supplementary data

The following is the Supplementary material related to this article.

MMC S1

Online supplement for Optimal Control of the SIR model with constrained policy, with an application to COVID-19.

mmc1.pdf (234.5KB, pdf)

References

  • 1.Gatto Nicole M., Schellhorn Henry. Optimal control of the SIR model in the presence of transmission and treatment uncertainty. Math. Biosci. 2021;333 doi: 10.1016/j.mbs.2021.108539. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Ledzewicz Urszula, Schättler Heinz. Conference Publications. American Institute of Mathematical Sciences; 2011. On optimal singular control for a general SIR model with vaccination and treatment; pp. 981–990. (Special). [DOI] [Google Scholar]
  • 3.Bolzoni Luca, Bonacini Elena, Soresina Cinzia, Groppi Maria. Time-optimal control strategies in SIR epidemic models. Math. Biosci. 2017;292:86–96. doi: 10.1016/j.mbs.2017.07.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Grigorieva E.V., Khailov E.N. Optimal vaccination, treatment, and preventive campaigns in regard to the SIR epidemic model. Math. Model. Nat. Phenom. 2014;9(4):105–121. doi: 10.1051/mmnp/20149407. [DOI] [Google Scholar]
  • 5.Gaff Holly, Schaefer Elsa. Optimal control applied to vaccination and treatment strategies for various epidemiological models. Math. Biosci. Eng. 2009;6(3):469–492. doi: 10.3934/mbe.2009.6.469. [DOI] [PubMed] [Google Scholar]
  • 6.Tchuenche J.M., Khamis S.A., Agusto F.B., Mpeshe S.C. Optimal control and sensitivity analysis of an influenza model with treatment and vaccination. Acta Biotheor. 2011;59(1):1–28. doi: 10.1007/s10441-010-9095-8. [DOI] [PubMed] [Google Scholar]
  • 7.Hethcote Herbert W., Waltman Paul. Optimal vaccination schedules in a deterministic epidemic model. Math. Biosci. 1973;18(3–4):365–381. doi: 10.1016/0025-5564(73)90011-4. [DOI] [Google Scholar]
  • 8.Laarabi Hassan, Abta Abdelhadi, Hattaf Khalid. Optimal control of a delayed SIRS epidemic model with vaccination and treatment. Acta Biotheoretica. 2015;63:87–97. doi: 10.1007/s10441-015-9244-1. [DOI] [PubMed] [Google Scholar]
  • 9.Ball Frank G., Lyne Owen D. Optimal vaccination policies for stochastic epidemics among a population of households. Math. Biosci. 2002;177–178:333–354. doi: 10.1016/S0025-5564(01)00095-5. [DOI] [PubMed] [Google Scholar]
  • 10.Ishikawa Masaaki. Proceeding of the ISCIE International Symposium on Stochastic Systems Theory and its Applications, Vol. 2012. 2012. Stochastic optimal control strategies of a SIR epidemic model with vaccination; pp. 57–62. [DOI] [Google Scholar]
  • 11.Pennisi Marzio, Catanuto Roberto, Pappalardo Francesco, Motta Santo. Optimal vaccination schedules using simulated annealing. Bioinformatics. 2008;24(15):1740–1742. doi: 10.1093/bioinformatics/btn260. [DOI] [PubMed] [Google Scholar]
  • 12.Tanner Matthew W., Sattenspiel Lisa, Ntaimo Lewis. Finding optimal vaccination strategies under parameter uncertainty using stochastic programming. Math. Biosci. 2008;215(2):144–151. doi: 10.1016/j.mbs.2008.07.006. [DOI] [PubMed] [Google Scholar]
  • 13.Patel Rajan, Longini Jr. Ira M., Halloran M. Elizabeth. Finding optimal vaccination strategies for pandemic influenza using genetic algorithms. J. Theoret. Biol. 2005;234(2):201–212. doi: 10.1016/j.jtbi.2004.11.032. [DOI] [PubMed] [Google Scholar]
  • 14.Wang Xinwei, Peng Haijun, Shi Boyang, Jiang Dianheng, Zhang Sheng, Chen Biaosong. Optimal vaccination strategy of a constrained time-varying SEIR epidemic model. Commun. Nonlinear Sci. Numer. Simul. 2019;67:37–48. doi: 10.1016/j.cnsns.2018.07.003. [DOI] [Google Scholar]
  • 15.Morgenstern Oskar, Von Neumann John. Princeton University Press; 1953. Theory of Games and Economic Behavior. [Google Scholar]
  • 16.Karatzas Ioannis, Shreve Steven. Springer; 2014. Brownian Motion and Stochastic Calculus. [DOI] [Google Scholar]
  • 17.Agarwal Ravi P., Badshah Qaisar, ur Rahman Ghaus, Islam Saeed. Optimal control & dynamical aspects of a stochastic pine wilt disease model. J. Franklin Inst. B. 2019;356(7):3991–4025. doi: 10.1016/j.jfranklin.2019.03.007. [DOI] [Google Scholar]
  • 18.Gray Alison, Greenhalgh David, Hu Liangjian, Mao Xuerong, Pan Jiafeng. A stochastic differential equation SIS epidemic model. SIAM J. Appl. Math. 2011;71(3):876–902. doi: 10.1137/10081856X. [DOI] [Google Scholar]
  • 19.Cvitanić Jakša, Karatzas Ioannis. Convex duality in constrained portfolio optimization. Ann. Appl. Probab. 1992;2:767–818. doi: 10.1214/aoap/1177005576. [DOI] [Google Scholar]
  • 20.Bertozzi Andrea L, Franco Elisa, Mohler George, Short Martin B, Sledge Daniel. The challenges of modeling and forecasting the spread of COVID-19. Proc. Natl. Acad. Sci. 2020;117(29):16732–16738. doi: 10.1073/pnas.2006520117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Wachter Jessica A. Portfolio and consumption decisions under mean-reverting returns: An exact solution for complete markets. J. Financ. Quant. Anal. 2002;37:63–91. doi: 10.2307/3594995. [DOI] [Google Scholar]
  • 22.Korn Ralf. World Scientific; 1997. Optimal Portfolios: Stochastic Models for Optimal Investment and Risk Management in Continuous Time. [DOI] [Google Scholar]
  • 23.Shreve Steven E. Springer-Verlag New York; 2004. Stochastic Calculus for Finance II: Continuous-Time Models. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

MMC S1

Online supplement for Optimal Control of the SIR model with constrained policy, with an application to COVID-19.

mmc1.pdf (234.5KB, pdf)

Articles from Mathematical Biosciences are provided here courtesy of Elsevier

RESOURCES