Skip to main content
Springer Nature - PMC COVID-19 Collection logoLink to Springer Nature - PMC COVID-19 Collection
. 2021 Oct 11;12(1):214–236. doi: 10.1007/s13235-021-00403-1

Dynamic Games of Social Distancing During an Epidemic: Analysis of Asymmetric Solutions

Ioannis Kordonis 1,, Athanasios-Rafail Lagos 1, George P Papavassilopoulos 1,2
PMCID: PMC8503885  PMID: 34659872

Abstract

Individual behaviors play an essential role in the dynamics of transmission of infectious diseases, including COVID-19. This paper studies a dynamic game model that describes the social distancing behaviors during an epidemic, assuming a continuum of players and individual infection dynamics. The evolution of the players’ infection states follows a variant of the well-known SIR dynamics. We assume that the players are not sure about their infection state, and thus, they choose their actions based on their individually perceived probabilities of being susceptible, infected, or removed. The cost of each player depends both on her infection state and on the contact with others. We prove the existence of a Nash equilibrium and characterize Nash equilibria using nonlinear complementarity problems. We then exploit some monotonicity properties of the optimal policies to obtain a reduced-order characterization for Nash equilibrium and reduce its computation to the solution of a low-dimensional optimization problem. It turns out that, even in the symmetric case, where all the players have the same parameters, players may have very different behaviors. We finally present some numerical studies that illustrate this interesting phenomenon and investigate the effects of several parameters, including the players’ vulnerability, the time horizon, and the maximum allowed actions, on the optimal policies and the players’ costs.

Keywords: COVID-19 pandemic, Games of social distancing, Epidemics modeling and control, Nash games, Nonlinear complementarity problems

Introduction

COVID-19 pandemic is one of the most important events of this era. Until early April 2021, it has caused more than 2.8 million deaths, an unprecedented economic depression, and affected most aspects of people’s lives in the larger part of the world. During the first phases of the pandemic, non-pharmaceutical interventions (primarily social distancing) have been one of the most efficient tools to control its spread [13]. Due to the slow roll-out of the vaccines, their uneven distribution, the emergence of SARS-CoV-2 variants, age limitations, and people’s resistance to vaccination, social distancing is likely to remain significant in large part of the globe for the near future.

Mathematical modeling of epidemics dates back to early twentieth century with the seminal works of Ross  [33] and Kermack and McKendrick [24]. A widely used modeling approach separates people in several compartments according to their infection state (e.g., susceptible, exposed, infected, recovered, etc.) and derives differential equations describing the evolution of the population of each compartment (for a review, see [1]).

However, the description of individual behaviors (practice of social distancing, use of face masks, vaccination, etc.) is essential for the understanding of the spread of epidemics. Game theory is thus a particularly relevant tool. A dynamic game model describing voluntary implementation of non-pharmaceutical interventions (NPIs) was presented in [32]. Several extensions were published, including the case where infection severity depends on the epidemic burden [6], and different formulations of the cost functions (linear vs. nonlinear, and finite horizon vs. discounted infinite horizon or stochastic horizon) [11, 15, 37]. Aurell et al. [4] study the design of incentives to achieve optimal social distancing, in a Stackelberg game framework. The works [2, 10, 12, 17, 22, 30, 31, 39] study different aspects of the coupled dynamics between individual behaviors and the spread of an epidemic, in the context of evolutionary game theory. Network game models appear in [3, 18, 27, 29], and the effects of altruism on the spread of epidemics are studied in [8, 14, 23, 26]. Another closely related stream of research is the study of the adoption of decentralized protection strategies in engineered and social networks [19, 21, 36, 38]. A review of game theoretic models for epidemics (including also topics other than social distancing, e.g., vaccination) is presented in [9, 20].

This paper presents a dynamic game model to describe the social distancing choices of individuals during an epidemic, assuming that the players are not certain about their infection state (susceptible (S), infected (I), or removed (R)). The probability that a player is at each health state evolves dynamically depending on the player’s distancing behavior, the others’ behavior, and the prevalence of the epidemic. We assume that the players care both about their health state and about maintaining their social contacts. The players may have different characteristics, e.g., vulnerable versus less vulnerable, or care differently about maintaining their social contacts.

We assume that the players are not sure about their infection state, and thus, they choose their actions based on their individually perceived probabilities of being susceptible, infected, or removed. In contrast with most of the literature, in the current work, players—even players with the same characteristics—are allowed to behave differently. We first characterize the optimal action of a player, given the others’ behavior, and show some monotonicity properties of optimal actions. We then prove the existence of a Nash equilibrium and characterize it in terms of a nonlinear complementarity problem.

Using the monotonicity of the optimal solution, we provide a simple reduced-order characterization of the Nash equilibrium in terms of a nonlinear programming problem. This formulation simplifies the computation of the equilibria drastically. Based on that result, we performed numerical studies, which verify that players with the same parameters may follow different strategies. This phenomenon seems realistic since people facing the same risks or belonging to the same age group often have different social distancing behaviors.

The model presented in this paper differs from most of the dynamic game models presented in the literature in the following ways:

  1. The players act without knowing their infection state. However, they know the probability of being susceptible, infected, or recovered, which depends on their previous actions and the prevalence of the epidemic.

  2. The current model allows for asymmetric solutions, i.e., players with the same characteristics may behave differently.

The rest of the paper is organized as follows. Section 2 presents the game theoretic model. In Sect. 3, we analyze the optimization problem of each player and prove some monotonicity properties. In Sect. 4, we prove the existence of the equilibrium and provide Nash equilibrium characterizations. Section 5 presents some numerical results. Finally, ‘Appendix’ contains the proof of the results of the main text.

The Model

This section presents the dynamic model for the epidemic spread and the social distancing game among the members of the society.

We assume that the infection state of each agent could be susceptible (S), infected (I), recovered (R), or dead (D). A susceptible person gets infected at a rate proportional to the number of infected people she meets with. An infected person either recovers or dies at constant rates, which depend on her vulnerability. An individual who has recovered from the infection is immune, i.e., she could not get infected again. The evolution of the infection state of an individual is shown in Fig. 1.

Fig. 1.

Fig. 1

The evolution of the infection state of each individual

We assume that there is a continuum of agents. This approximation is frequently used in game theoretic models dealing with a very large number of agents. The set of players is described by the measure space ([0,1),B,μ), where B is the Borel σ-algebra and μ the Lebesgue measure. That is, each player is indexed by an i[0,1).

Denote by Si(t),Ii(t),Ri(t),Di(t) the probability that player i[0,1) is susceptible, infected, removed, or dead at time t. The dynamics is given by:

S˙i=-ruiSiIfI˙i=ruiSiIf-αiIiR˙i=α¯iIiD˙i=(αi-α¯i)Ii, 1

where r,αi are positive constants, and ui(t) is the action of player i at time t. The quantity ui(t) describes player i’s socialization, which is proportional to the time she spends in public places. The quantity If, which denotes the density of infected people in public places, is given by:

If(t)=Ii(t)ui(t)μ(di). 2

For the actions of the players, we assume that there are positive constants um, uM, such that ui(t)[um,uM][0,1]. The constant um describes the minimum social contacts needed for an agent to survive, and uM is an upper bound posed by the government.

The cost function for player i is given by:

Ji=Gi(1-Si(T))-si0Tui(t)u~(t)dt-si0Tκui(t)dt, 3

where T is the time horizon. The parameter Gi>0 depends on the vulnerability of the player and indicates the disutility a player experiences if she gets infected. The quantity 1-Si(T) corresponds to the probability that player i gets infected before the end of the time horizon. (Note that in that case the infection state of the player at the end of the time horizon is I, R, or D.) The second term corresponds to the utility a player derives from the interaction with the other players, whose mean action is denoted by u~(t):

u~(t)=ui(t)μ(di). 4

The third term indicates the interest of a person to visit public places. The relative magnitude of this desire is modeled by a positive constant κ. Finally, constant si indicates the importance player i gives on the last two terms that correspond to going out and interacting with other people.

Considering the auxiliary variable u¯(t):

u¯(t)=κ+u~, 5

and computing S(T) by solving (1), the cost can be written equivalently as:

Ji=Gi1-Si(0)e-r0Tui(t)If(t)dt-si0Tui(t)u¯(t)dt. 6

Without loss of generality, assume that Ri(0)=Di(0)=0 for all i[0,1).

Assumption 1

(Finite number of types) There are M types of players. Particularly, there are M+1 values 0=i¯0<<i¯M=1 such that the functions Si(0),Ii(0),Gi,si,αi:[0,1)R are constant for i[i¯0,i¯1), i[i¯1,i¯2),, i[i¯M-1,i¯M). Denote by mj=μ([i¯j-1,i¯j)) the mass of the players of type j. Of course m1++mM=1.

Remark 1

The finite number of types assumption is very common in many applications dealing with a large number of agents. For example, in the current COVID-19 pandemic, people are grouped based on their age and/or underlying diseases to be prioritized for vaccination. Assumption 1, combined with some results of the following section, is convenient to describe the evolution of the states of a continuum of players using a finite number of differential equations.

Assumption 2

(Piecewise constant actions) The interval [0, T) can be divided in subintervals [tk,tk+1), with t0=0<t1<<tN=T, such that the actions of the players are constant in these intervals.

Remark 2

Assumption 2 indicates that people decide only a finite number of times (tk) and follow their decisions for a time interval [tk,tk+1). A reasonable length for that time interval could be 1 week.

The action of player i in the interval [tk,tk+1) is denoted by uki.

Assumption 3

(Measurability of the actions) The function uk·:[0,1)[um,uM] is measurable.

Under Assumptions 13, there is a unique solution to differential equations (1), with initial conditions S·(0),I·(0), and the integrals in (2), (4) are well defined (see Appendix A.1). We use the following notation:

u¯k=tktk+1u¯dt,andIkf=tktk+1If(t)dt.

For each player, we define an auxiliary cost, by dropping the fixed terms of (6) and dividing by si:

J~i(ui,u¯,If)=-biexp-rk=0N-1ukiIkf-k=0N-1ukiu¯k, 7

where bi=Si(0)Gi/si, and ui=[u0i,,uN-1i]T. Denote by U=[um,uM]N the set of possible actions for each player. Observe that ui minimizes Ji over the feasible set U if and only if it minimizes the auxiliary cost J~i. Thus, the optimization problem for player i is equivalent to:

minimizeuiUJ~i(ui,u¯,If). 8

Note that the cost of player i depends on the actions of the other players through the terms u¯ and If. Furthermore, the current actions of all the players affect the future values of u¯ and If through the SIR dynamics.

Assumption 4

For a player i of type j, denote bj=bi. Assume that the different types of players have different bj’s. Without loss of generality, assume that b1<b2<<bM.

Assumption 5

Each player i has access only to the probabilities Si and Ii and the aggregate quantities u¯ and If, but not the actual infection states.

Remark 3

This assumption is reasonable in cases where the test availability is very sparse, so the agents are not able to have a reliable feedback for their estimated health states.

In the rest of the paper, we suppose that Assumptions 15 are satisfied.

Analysis of the Optimization Problem of Each Player

In this section, we analyze the optimization problem for a representative player i, given u¯k and Ikf>0, for k=0,,N-1.

Let us first define the following composite optimization problem:

minimizeA-bie-A+f(A), 9

where:

f(A)=infuiU-k=0N-1ukiu¯k:k=0N-1ukiIkf=A/r. 10

The following proposition proves that (8) and (9) are equivalent and express their solution in a simple threshold form.

Proposition 1

  • (i)

    If ui is optimal for (8), then uiU~={um,uM}N.

  • (ii)

    Problems (8) and (9) are equivalent, in the sense that they have the same optimal values, and ui minimizes (8) if and only if there is an optimal A for (9) such that ui attains the minimum in (10).

  • (iii)

    Let Am=rumk=0N-1Ikf and AM=ruMk=0N-1Ikf. For A[Am,AM], the function f is continuous, non-increasing, convex, and piecewise affine. Furthermore, it has at most N affine pieces and f(A)=, for A[Am,AM].

  • (iv)

    There are at most N+1 vectors uiU that minimize (8).

  • (v)

    If ui is optimal for (8), then there is a λ such that u¯k/Ikfλ implies uki=um, and u¯k/Ikf>λ implies uki=uM.

Proof

The idea of the proof of Proposition 1 is to reduce the problem of minimizing (8) into the minimization of the sum of the concave function -bie-A with a piecewise affine function f(A). Then, the candidates for the minimum are only the corner points and of f(A) and the endpoints of the interval, where f is defined. The form of the optimal action ui comes from a Lagrange multiplier analysis of (10). For a detailed proof, see Appendix 2.

Remark 4

Part (v) of the proposition shows that if the density of infected people in public places If is high, or the average socialization u¯ is low, then it is optimal for a player to choose a small degree of socialization. The optimal action for each player depends on the ratio u¯k/Ikf. Particularly, there is a threshold λ such that the action of player i is um for values of the ratio below the threshold and uM for ratios above the threshold. Note that the threshold is different for each player.

Remark 5

The fact that the optimal value of a linear program is a convex function of the constraints constants is known in the literature (e.g., see [35] chapter 2). Thus, the convexity of the function f is already known from the literature.

Corollary 1

There is a simple way to solve the optimization problem (8) using the following steps:

  1. Compute Λ={u¯k/Ikf:k=0,,N-1}{0}.

  2. For all λΛ compute uλ with:
    ukλ=uMifu¯k/Ikf>λumifu¯k/Ikfλ,
    and Ji(uλ).
  3. Compare the values of Ji(uλ), for all λΛ, and choose the minimum.

We then prove some monotonicity properties for the optimal control.

Proposition 2

Assume that for two players i1 and i2, with parameters bi1 and bi2, the minimizers of (9) are A1 and A2, respectively, and ui1 and ui2 are the corresponding optimal actions. Then:

  • (i)

    If bi1<bi2, then A1A2.

  • (ii)

    If bi1<bi2, then uki2uki1, for k=0,,N-1.

  • (iii)

    If bi1=bi2, then either uki2uki1 for all k, or uki1uki2 for all k.

Proof

See Appendix 3.

Remark 6

Recall that bi=Si(0)Gi/si. Thus, Proposition 2(ii) expresses of the fact that if (a) a person is more vulnerable, i.e., she has large Gi, or (b) she derives less utility from the interaction with the others, i.e., she has smaller si, or (c) it is more likely that she is not yet infected, i.e., she has larger Si(0), then she interacts less with the others.

Remark 7

The optimal control law can be expressed in feedback form (see Appendix A.4.1).

Nash Equilibrium Existence and Characterization

Existence and NCP Characterization

In this section, we prove the existence of a Nash equilibrium and characterize it in terms of a nonlinear complementarity problem (NCP).

We consider the set U~={um,uM}N, defined in Proposition 1. Let v1,,v2N be the members of the set U~ and plj be the mass of players of type j following action vlU~. Let also pj=[p1j,,p2Nj] be the distribution of actions of the players of type j and π=[p1,,pM] be the distribution of the actions of all the players.

Denote by:

Δj=pjR2N:plj0,l=12Nplj=mj, 11

the set of possible distributions of actions of the players of type j and by Π=Δ1××ΔM the set of all possible distributions.

Finally, let F:ΠR2N·M be the vector function of auxiliary costs; that is, the component F(j-1)2N+l(π) is the auxiliary cost of the players of type j playing a strategy vl, as introduced in (7), when the distribution of actions is π. We denote Fj(π)=[F(j-1)2N+1(π),...,Fj2N(π)] the vector of the auxiliary costs of the players of type j playing vl, l=1,,2N.

Let us recall the notion of a Nash equilibrium for games with a continuum of players (e.g., [28]).

Definition 1

A distribution of actions πΠ is a Nash equilibrium if for all j=1,,M and l=1,,2N:

π(j-1)2N+l>0largminlF(j-1)2N+l(π) 12

Let δj(π) be the value of problem (8), i.e., the minimum value of the auxiliary cost of an agent of type j. This value depends on π, through the terms If and u¯. Define Φj(π)=Fj(π)-δj(π) and Φ(π)=[Φ1(π)...ΦM(π)]. We then characterize a Nash equilibrium in terms of a nonlinear complementarity problem (NCP):

0πΦ(π)0, 13

where πΦ(π) means that πTΦ(π)=0.

Proposition 3

  • (i)

    A distribution πΠ corresponds to a Nash equilibrium if and only if it satisfies the NCP (13).

  • (ii)
    A distribution πΠ corresponds to a Nash equilibrium if and only if it satisfies the variational inequality:
    (π-π)TF(π)0,for allπΠ 14
  • (iii)

    There exists a Nash equilibrium.

Proof

See Appendix A.5.

Remark 8

In principle, we can use algorithms for NCPs to find Nash equilibria. The problem is that the number of decision variables grows exponentially with the number of decision steps. Thus, we expect that such methods would be applicable only for small values of N.

Structure and Reduced-Order Characterization

In this section, we use the monotonicity of the optimal strategies, shown in Proposition 2, to derive a reduced-order characterization of the Nash equilibrium.

The actions on a Nash equilibrium have an interesting structure. Assume that π is a Nash equilibrium and:

V={vlU~:j:π(j-1)2N+l>0}U~, 15

is the set of actions used by a set of players with a positive mass. Let us define a partial ordering on U~. For v1,v2U~, we write v1v2 if vk1vk2 for all k=1,,N. Proposition 2(iii) implies that V is a totally ordered subset of U~ (chain).

Lemma 1

There are at most N! maximal chains in U~, each of which has length N+1. Thus, at a Nash equilibrium, there are at most N+1 different actions in V.

Proof

See Appendix A.6.

For each time step k, denote by ρk the fraction of players who play uM, that is, ρk=μ({i:uki=uM}). Given any vector ρ=[ρ1ρN][0,1]N, we will show that there is a unique πΠ, such that the corresponding actions satisfy the conclusion of Proposition 2(iii) and induce the fractions ρ. An example of the relationship between π and ρ is given in Fig. 2.

Fig. 2.

Fig. 2

In this example, N=5 and there are M=3 types of players, depicted with different colors. The mass of players below each solid line play uM, and the mass of players above the line play um. Example 1 computes π from ρ

Let us define the following sets:

Ik={i[0,1):uki=uM},Kk={k:ρkρk},k=0,,N-1.

Let k1,,kN be a reordering of 0,,N-1 such that ρk1ρk2ρkN. Consider also the set V~={v¯1,,v¯N+1} of N+1 actions v¯n with:

v¯kn=uMifkKknumotherwise,n=1,,N, 16

and v¯kN+1=um, for all k. Observe that v¯n+1v¯n. The following proposition shows that the set V, defined in (15) is subset of the set V~.

Before stating the proposition, let us give an example.

Example 1

Consider the fractions described in Fig. 2. There are three types of players with total mass m1,m2, and m3 with corresponding colors blue, pink, and yellow. In this example, we assume that the actions of each player i[0,1) are given by:

uki=1,ifi<ρk0,otherwise 17

The sets Ik are given by:

Ik=[0,ρk).

The sets Kk are given by:

K0={2},K1={2,3},K2={2,3,1},K3={2,3,1,4},K4={2,3,1,4,0}.

The actions v¯n are given by:

v¯0=[uM,uM,uM,uM,uM],v¯1=[uM,uM,um,uM,uM],v¯2=[uM,uM,um,um,uM],v¯3=[uM,um,um,um,uM],v¯4=[uM,um,um,um,um],v¯5=[um,um,um,um,um]. 18

The mass of the players of each type following each action is described in the following table:

Type 1 1 2 2 2 2 3 3
Mass ρ2 i¯1-ρ2 ρ3-m1 ρ1-ρ3 ρ4-ρ1 i¯2-ρ4 ρ0-i¯2 1-ρ0
Action v¯0 v¯1 v¯1 v¯2 v¯3 v¯4 v¯4 v¯5

The following proposition and its corollary present a method to compute π from ρ in the general case (i.e., without assuming a set of actions in a form similar to (17)).

Proposition 4

Assume that (uki)i[0,1),k=0,,N-1, with uiU~, be a set of actions satisfying the conclusions of Proposition 2. Then:

  • (i)

    For kk, either IkIk or IkIk.

  • (ii)

    If for some k,k, it holds ρk=ρk, then μ-almost surely all the players have the same action on k,k, i.e., μ({i:uki=uki})=1.

  • (iii)
    Up to subsets of measure zero, the following inclusions hold:
    graphic file with name 13235_2021_403_Equ56_HTML.gif
    where Inline graphic indicates that μ(Ikn\Ikn+1)=0. Furthermore, μ(Ik)=ρk.
  • (iv)

    For μ-almost all iIkn+1\Ikn, the action ui is given by v¯n+1, for μ-almost all iIk1, ui=v¯1, and for μ-almost all i[0,1)\IkN, uki=v¯N+1.

Proof

See Appendix A.7.

Corollary 2

The mass of players of type j with action v¯n is given by:

μ(i:iis of typej,ui=v¯n)=μ([i¯j-1,i¯j)[ρkn-1,ρkn)), 19

where we use the convention that ρk0=0, and ρkN+1=1. Thus:

π(j-1)2N+l=μ([i¯j-1,i¯j)[ρkn-1,ρkn))ifvl=v¯n0otherwise 20

Proof

The proof follows directly from Propositions 4 and 2(ii).

Remark 9

There are at most M+N+1 combinations of jl such that π(j-1)2N+l>0.

Let us denote by π~(ρ) the value of vector π computed by (20).

Example 2

The situation is the same as in Example 1, but without assuming that the actions are given by (17). Then, Corollary 2 shows that π~(ρ) is given by the table in Example 1.

Proposition 5

The fractions ρ0,,ρN-1 correspond to a Nash equilibrium if and only if:

H(ρ)=j=1Mn=1N+1μ([i¯j-1,i¯j)[ρkn-1,ρkn))(F¯j,v¯n(π~(ρ))-δj(π~(ρ)))=0, 21

where F¯j,v¯n(π) is the cost of action v¯n, for a player of type j. Furthermore, H(π) is continuous and nonnegative.

Proof

See Appendix A.8.

Remark 10

The computation of an equilibrium has been reduced to the calculation of the minimum of an N-dimensional function. We exploit this fact in the following section to proceed with the numerical studies.

Numerical Examples

In this section, we give some numerical examples of Nash equilibria computation. Section 5.1 presents an example with a single type of players and Sect. 5.2 an example with many types of players. Section 5.3 studies the effect of the maximum allowed action uM on the strategies and the costs of the players.1

Single Type of Players

In this subsection, we study the symmetric case, i.e., all the players have the same parameter bi. The parameters for the dynamics are r=0.4 and a=1/6 which correspond to an epidemic with basic reproduction number R0=2.4, where an infected person remains infectious for an average of 6 days. (These parameters are similar to [22] which analyzes COVID-19 epidemic.) We assume that um=0.4 and that there is a maximum action uM=0.75, set by the government. The discretization time intervals are 1 week, and the time horizon T is approximately 3 months (13 weeks). The initially infected players are I0=0.01. We chose this time horizon to model a wave of the epidemic, starting at a time point where 1% of the population is infected. We assume that κ=3.

We then compute the Nash equilibrium using a multi-start local search method for (21). Figure 3 shows the fraction ρ of players having action uM at each time step and the evolution of the total mass of infected players for the several values of b. We observe that, for small values of b, which correspond to less vulnerable or very sociable agents, the players do not engage in voluntary social distancing. For intermediate values of b, the players engage voluntary social distancing, especially when there is a large epidemic prevalence. For large values of b, there is an initial reaction of the players which reduces the number of infected people. Then, the actions of the players return to intermediate levels and keep the number of infected people moderate. In all the cases, voluntary social distancing ‘flattens the curve’ of infected people mass, in the sense that it reduces the pick number of infected people and leaves more susceptible persons in the end on the time horizon.

Fig. 3.

Fig. 3

a The fractions ρk, for k=1,,13, for different values of b. b The evolution of the number of infected people under the computed Nash equilibrium

We then present some results for the case where bi=200. Figure 4 illustrates the evolution of Si(t) and Ii(t), for the players having different strategies. We observe that the trajectories of Si’s do not intersect. What is probably interesting is that the trajectories of Ii may intersect. This indicates that, toward the end of the time horizon, it is probable for a person who was less cautious, i.e., she used higher values of ui, to have a lower probability of being infectious.

Fig. 4.

Fig. 4

The evolution of the probabilities Si and Ii, for players following different strategies, for b=200. Different colors are used to illustrate the evolution of the probabilities for players using different strategies

Many Types of Players

We then compute the Nash equilibrium for the case of multiple types of players. We assume that there are six types of players with vulnerability parameters G1=100,G2=200,G3=400,G4=800,G5=1600,G6=3200. The sociability parameter si is equal to 1, for all the players. The masses of these types are m1=0.5 and m2==m6=0.1. The initial condition is for all the players I0=0.0001, and the time horizon is 52 weeks (approximately a year). Here, we assume that the maximum action is uM=0.8. The rest of the parameters are as in the previous subsection.

Figure 5 presents the fractions ρ and the evolution of the probability of each category of players to be susceptible and infected. Let us note that the analysis of Sect. 4.2 simplifies a lot the analysis. Particularly, the set Π has (252-1)68.3·1093 dimensions, while Problem (21) only 52.

Fig. 5.

Fig. 5

a The fractions ρk of players having an action uM. Note that the fractions correspond to players of all types. Since si=1, for all the types, it holds b1<<b6, and thus, the more vulnerable players cannot have an action higher than the less vulnerable ones. b The probability of a class of people to be susceptible and infected. The colored lines correspond to the probabilities of being susceptible Si(t) and infected Ii(t), for the several strategies of the players. The bold black line represents the mass of susceptible and infected persons (Color figure online)

Effect of uM

We then analyze the case where the types of the players are as in Sect. 5.2, and the initial condition is I0=0.005 for all the players, for various values of uM. The time horizon is 13 weeks.

Figure 6a illustrates the equilibrium fractions ρk, for the various values of uM. We observe that as uM increases, the fractions ρk decrease. Figure 6b shows the evolution of the mass of infected players, for the different values of the maximum action uM. We observe that as uM increases, the mass of infected players decreases. Figure 6c presents the cost of the several types of players, for the different value of the maximum action uM. We observe that players with low vulnerability (G=100) prefer always a larger value of uM, which corresponds to less stringent restrictions. For vulnerable players (e.g., G=3200), the cost is an increasing function of uM; that is, they prefer more stringent restrictions. For intermediate values of G, the players prefer intermediate values of uM. The mean cost in this example is minimized for uM=0.6.

Fig. 6.

Fig. 6

a The fractions ρk, for the several values of the maximum action uM. b The mass of infected people as a function of time, for the different values of the maximum action uM. c The cost for the several classes of players, for the different values of the maximum action uM. The bold black line represents the mean cost of all the players

Conclusion

This paper studied a dynamic game of social distancing during an epidemic, giving an emphasis on the analysis of asymmetric solutions. We proved the existence of a Nash equilibrium and derived some monotonicity properties of the agents’ strategies. The monotonicity result was then used to derive a reduced-order characterization of the Nash equilibrium, simplifying its computation significantly. Through numerical experiments, we show that both the agents’ strategies and the evolution of the epidemic depend strongly on the agents’ parameters (vulnerability, sociality) and the epidemic’s initial spread. Furthermore, we observed that agents with the same parameters could have different behaviors, leading to rich, high-dimensional dynamics. We also observe that more stringent constraints on the maximum action (set by the government) benefit the more vulnerable players at the expense of the less vulnerable. Furthermore, there is a certain value for the maximum action constant that minimizes the average cost of the players.

There are several directions for future work. First, we can study more general epidemics models than the SIR. Second, we can investigate different information patterns, including the cases where the agents receive regular or random information about their health state. Finally, we can compare the behaviors computed analytically with real-world data.

A Appendix: Proof of the Results of the Main Text

A.1 Existence of Solution to (1)

Note that the first two equations in (1) do not depend on Ri and Di. Thus, it suffices to show that the first two equations of (1) have a unique solution.

For any i[0,1), if [Si(t),Ii(t)] solve the differential equations (1), with initial condition (Si(0),Ii(0))[0,1]2, then (Si(t),Ii(t)) remain in [0,1]2. Thus, we consider the solution of the differential equations:

S˙i=satB(-ruiSiIf(t))I˙i=satB(ruiSiIf(t)-αiIi), 22

where satB(z)=max(min(z,B),-B), and B=ruM2+maxjαj.

Consider the Banach space X=L1([0,1),R2), and let x0=(S·(0),I·(0)):[0,1)R2. Then, under Assumptions 1,3, it holds x0X. For each interval [tk,tk+1), the differential equations (22) with the corresponding initial conditions can be written as:

x˙=fk(x),x(tk)=x0k, 23

where for x:i[Si,Ii]T, the value of fk(x)X is given by:

fk(x):i[satB(-rukiSiMkx),satB(rukiSiMkx-αiIi)]T,

where Mk:XR is a linear bounded operator with Mkx=Iiukiμ(di). For the initial condition, it holds x00=x0, and x0k=x(tk) is computed from the solution of (23) on the interval [tk-1,tk), for k1. For all k, fk is Lipschitz and thus there is a unique solution to (23) (e.g., Theorem 7.3 of [7]). Furthermore, both I·(t) and u·(t) are measurable and bounded. Thus, the integrals in (4), (2) are well defined.

Note that from Assumption 1, we only used the fact that S·(0),I·(0):[0,1)R are measurable and not the piecewise constant property.

Proof of Proposition 1

  • (i)

    Since rIkf>0 and bi>0, the cost (7) is strictly concave, with respect to uki. Thus, the minimum with respect to uki is either um or uM.

  • (ii)
    Since U is compact and J~ is continuous, there is an optimal solution for (8). Denote by ui, this solution. Further, denote by V1=J~i(ui,) and V2=infA{-bie-A+f(A)} the values of problems (8) and (9), respectively. Then, for A=k=0N-1ruki,Ikf, we have
    V2-bie-A+f(A)-biexp-rk=0N-1uki,Ikf-k=0N-1uki,u¯k=J~i(ui,)=V1,
    where the first inequality is due to the definition of V2 and the second inequality is due to the definition of f(A). To contradict, assume that V2<V1. Then, there is some A and an ε>0 such that:
    -bie-A+f(A)<V1-2ε. 24
    Thus, there is a u~i such that A=k=0N-1ru~kiIkf and k=0N-1u~kiu¯k<f(A)+ε. Combining with (24), we get J~i(u~)<V1-ε, which contradicts the definition of V1. For ui, minimizing (8), the problem (9) is minimized for A=k=0N-1ruki,Ikf and ui, attains the minimum in (10). To see this, observe that otherwise we would have V2<V1. On the other hand, assume that A minimizes (9) and ui attains the minimum in (10). Then, V2=-be-A+f(A)=Ji(ui). Furthermore, since V2=V1, it holds Ji(ui)=V1, and thus, ui minimizes Ji.
  • (iii)

    The set uiU:k=0N-1ukiIkf=A/r is non-empty if and only if A[Am,AM]. Thus, the f(A) is finite if and only if A[Am,AM].

    For A[Am,AM], there exists an optimal solution ui that attains the minimum in (10). Since (10) is a feasible linear programming problem, there is a Lagrange multiplier λ (e.g., Proposition 5.2.1 of [5]), and ui minimizes the Lagrangian:
    L(ui,λ)=-k=0N-1u¯kuki+λk=0N-1Ikfuki-λA/r. 25
    Thus, uki=um, if u¯k/Ikf<λ and uki=uM, if u¯k/Ikf>λ. To compute f(A), we reorder k, using a new index k, such that u¯k/Ikf is non-increasing. Let:
    kA=maxk¯A:k=0k¯A-1uMIkf+k=k¯AN-1umIkfA/r.
    Then:
    ΣkA+uk¯AiIk¯Af=A/r.
    where ΣkA=k=0k¯A-1uMIkf+k=k¯A+1N-1umIkf. Thus:
    f(A)=-k=0kA-1u¯kuM-k=kA+1N-1u¯kum-u¯kAIkAf(A/r-ΣkA),
    for A/r[ΣkA+umIkAf,ΣkA+uMIkAf]. It holds:
    ΣkA+uMIkAf=ΣkA+1+umIkA+1f.
    Therefore, f is continuous and piecewise affine. Furthermore, since u¯k/Ikf is non-increasing with respect to k, the slope of f is non-decreasing, i.e., it is convex. Thus, f is differentiable in all (Am,AM) except of the points A=Σk+uMIkf with u¯k/Ikf>u¯k+1/Ik+1f. The linear parts of f are at most N.
  • (iv)
    Since -bie-A is strictly concave in A, there are at most N+1 possible minima of -bie-A+f(A), which correspond to the points of non-differentiability of f in (Am,AM) and the points Am and AM. Observe that for A=Am or A=AM, there is a unique ui minimizing (10). We then show that for all the non-differentiability points A of f, there is a unique ui minimizing (10). If A is a non-differentiability point, there is a k0 such that A/r=k=0k0IkfuM+k=k0+1N-1Ikfum and u¯k0/Ik0f>u¯k0+1/Ik0+1f. We then show that the unique minimizer in (10) is given by uki=uM for kk0 and uki=um for k>k0. Indeed, ui is feasible and if uui is another feasible point, it holds:
    k=0k0(uM-uk)Ikf+k=k0+1N-1(um-uk)Ikf=0.
    Multiplying by u¯k0/Ik0f, we get:
    k=0k0(uM-uk)Ikfu¯k0Ik0f+k=k0+1N-1(um-uk)Ikfu¯k0Ik0f=0.
    Then, using that uM-uk0, um-uk0, and that for kk0, it holds u¯k/Ikfu¯k0/Ik0f and for k>k0, it holds u¯k/Ikf<u¯k0/Ik0f, we have:
    -k=0N-1uku¯k--k=0N-1ukiu¯k=k=0k0(uM-uk)Ikfu¯kIkf+k=k0+1N-1(um-uk)Ikfu¯kIkf0,
    and the inequality is strict if for some k>k0, ukum. Therefore, ui is optimal and if u is also optimal, then it should satisfy uk=um for all k>k0. Combining this with the fact that k=0N-1ukIkf=A/r and Ikf>0, we get u=ui.
  • (v)
    We have shown that if ui is optimal, then there is a k0 such that uki=uM for kk0 and uki=um for k>k0. Then, using the original index k, the optimal control can be expressed as:
    uki=uMifu¯k/Ikfλumifu¯k/Ikf<λ,
    where λ=u¯k0/Ik0f.

A.3 Proof of Proposition 2

  • (i)
    Since A1 is optimal for bi1 and A2 is optimal for bi2, it holds:
    -bi1e-A1+f(A1)-bi1e-A2+f(A2),-bi2e-A2+f(A2)-bi2e-A1+f(A1).
    Adding these equations and reordering, we get:
    (bi2-bi1)e-A2(bi2-bi1)e-A1.
    And since bi2>bi1, we get A1A2
  • (ii)
    Using (v) of Proposition 1, and A1A2, we get:
    A1/r=k=0N-1uki1Ikf=k=0N-1um+(uM-um)hλ1(u¯k/Ikf)Ikfk=0N-1um+(uM-um)hλ2(u¯k/Ikf)Ikf=k=0N-1uki2Ikf=A2/r
    where hλ(·) is the Heaviside function, i.e., hλ(x)=1 if xλ and hλ(x)=0 otherwise. Therefore, λ1λ2.
  • (iii)
    Assume that for k1k2, uk1i1=uk2i2=um and uk1i2=uk2i1=uM. Then, using (v) of Proposition 1 we have:
    λ2<u¯k1Ik1fλ1,λ1u¯k2Ik2f<λ2,
    which is a contradiction.

A.4 Optimal Control and Equilibria in Feedback Form

A.4.1 Optimal Control in Feedback Form

In this section, we express the optimal control law in feedback form using dynamic programming. Let Vk(Ski,si,Gi) be the optimal cost-to-go from time k of a player with parameters si,Gi:

Vk(Ski,si,Gi)=minuiGi-GiSkiexp-rk=kN-1ukiIkf-sik=kN-1ukiu¯k, 26

where Ski=Si(tk)=Si(0)exp-rk=0k-1ukiIkf. The optimal cost-to-go can be expressed as:

Vk(Ski,si,Gi)=Gi+siV~k(Ski,Gi/si).

We call V~k the ‘auxiliary cost-to-go.’

Proposition 6
  • (i)

    The auxiliary cost-to-go V~k(Ski,Gi/si) is non-increasing and concave in Ski.

  • (ii)

    The optimal cost-to-go Vk(Ski,si,Gi) is non-decreasing in Gi, non-increasing in si, and non-increasing and concave in Ski.

  • (iii)

    The optimal control law can be expressed in threshold form. That is, there are constants l0,,lN-1 such that the optimal control law for each player satisfies: uki=um if SkiGi/si>lk and uki=uM if SkiGi/si<lk. For SkiGi/si=lk, both uki=um and uki=uM are optimal.

Proof
  • (i)
    The auxiliary cost-to-go can be written as:
    V~k(Ski,si/Gi)=minui-GiSkisiexp-rk=kN-1ukiIkf-k=kN-1ukiu¯k. 27
    Then, applying the principle of optimality we get:
    V~k(Ski,si/Gi)=minuki-ukiu¯k+V~k+1(Skiexp[-rukiIkf],si/Gi). 28
    We use induction. For k=N, the auxiliary cost-to-go is given by V~k(SNi,si/Gi)=-GiSNi/si. That is, V~k(SNi,si/Gi) is non-increasing and concave in SNi. Assume that it holds for k=k0. Then, for each fixed uk0i, the quantity:
    -uk0iu¯k0+V~k0(Skiexp[-ruk0iIk0f],si/Gi),
    is non-increasing and concave in Sk0i. Using (28), i.e., minimizing with respect to uk0i, we conclude to the desired result.
  • (ii)

    The monotonicity properties of Vk are a direct consequence of its definition (26). The concavity of the optimal value follows easily from (i).

  • (iii)

    From Proposition 1(i), we know that the optimal uki is in the set, {um,uM}. Using Proposition 2 to the subproblem starting at time step k, and applying the principle of optimality, we get the desired result. The fact that if SkiGi/si=lk, then both uki=um and uki=uM are optimal, is a consequence of the continuity of V~k.

Remark 11

Propositions 6(iii) and 1(v) both express the optimal actions for a player i. The primary difference is that Proposition 1(v) uses dynamic programming, and thus, the policies obtained can better handle uncertainties in the individual dynamics of player i.

Equilibrium in Feedback Strategies

Consider a Nash equilibrium π that induces the fractions ρk, the mean actions u¯k, and the mass of infected people in public places Ikf, for k=0,,N-1. Consider also the auxiliary cost-to-go function V~k defined in (27), and the corresponding thresholds lk. Then, based on the strategies in Proposition 28(iii) we will describe a Nash equilibrium in feedback strategies. Consider the set of strategies:

uki=umifSkiGi/si>lkuMifSkiGi/si<lkuMifSkiGi/si=lk,andi<ρkumifSkiGi/si=lk,andiρk 29

.

Proposition 7

The set of strategies (29) is a Nash equilibrium.

Proof

Observe that the set of strategies (29) generates the same actions with the Nash equilibrium π.

A.5 Proof of Proposition 3

  • (i)

    Assume that a πΠ satisfies (13) and fix a j{1,,M}. For any l such that π(j-1)2N+l>0, it holds Φ(j-1)2N+l(π)=0, that is F(j-1)2N+l(π)=δj(π)=minlF(j-1)2N+l(π). Thus, πΠ is a Nash equilibrium.

    Conversely, assume that πΠ is a Nash equilibrium and fix a j{1,,M}. There is an l such that π(j-1)2N+l>0. Since π is a Nash equilibrium, it holds F(j-1)2N+l(π)=δj(π) and for all other l, it holds F(j-1)2N+l(π)F(j-1)2N+l(π)=δj(π), which implies (13).

  • (ii)
    Assume that π is a Nash equilibrium and πΠ. Then, l=12Nπ(j-1)2N+l=l=12Nπ(j-1)2N+l=mj. Since π is a Nash equilibrium, it holds:
    l=12N(π(j-1)2N+l-π(j-1)2N+l)TF(j-1)2N+l(π)0.
    Thus, (14) holds.

    Conversely, assume that (14) holds, for some πΠ. If π is not a Nash equilibrium, then there is a jl such that π(j-1)2N+l>0 and F(j-1)2N+l>δj(π). Then, if l is such that F(j-1)2N+l=δj(π), taking π=π+π(j-1)2N+le(j-1)2N+l-π(j-1)2N+le(j-1)2N+l, we get (π-π)TF(π)<0, which is a contradiction.

  • (iii)

    With a slight abuse of notation, we write Ikf(π),u¯k(π) to describe the quantities Ikf,u¯k when the distribution of actions is π and J~j(vl,π) to describe the auxiliary cost of a player of type j who plays action vj when the distribution of the actions is π.

Lemma 2

The quantities Ikf(π),u¯k(π),J~j(vl,π), are continuous on π.

Proof

The state of the system evolves according to the set of M2N+1+1 differential equations:

S˙j,vl=-rvklSj,vlIf,I˙j,vl=rvklSj,vlIf-αjIj,vl,z˙=If,

where j=1,,M, l=1,,2N, k:t[tk,tk+1), and:

If=j=1Ml=12Nπ(j-1)2N+lIj,vlvkl.

The initial conditions are Sj,vl(0)=Sj(0), Ij,vl(0)=Ij(0), (Assumption 1) and z(0)=0.

The right-hand side of the differential equations depends continuously on π through the term If. Furthermore, Sj,vl(t),Ij,vl(t) remain in [0, 1] for all j,vl. Thus, the state space of the system remains in [0,1]M·2N×R and the right-hand side of the differential equation is Lipschitz. Thus, Theorem 3.4 of [25] applies and Sj,vl(t), Ij,vj(t) and z(t) depend continuously on π. Thus, Ikf=z(tk+1)-z(tk) depends continuously on π. Furthermore, u¯k is continuous (linear) on π. Finally, the auxiliary cost Jj(vl,π), due to its form (7), depends continuously π.

To complete the proof, observe that F(π) is continuous and Π is compact and convex. Thus, the existence is a consequence of Corollary 2.2.5 of [16].

Remark 12

An alternative would be to use Theorem 1 of [34] or Theorem 1 of [28], combined with Lemma 2 to prove the existence of a mixed Nash equilibrium and then use Assumption 1, to construct a pure strategy equilibrium. However, the reduction to an NCP is useful computationally.

A.6 Proof of Lemma 1

Every maximal chain begins with the least element [um,,um] and ends at the greatest element [uM,,uM]. Every two consecutive elements of a maximal chain vl, vl+1 differ at exactly one point; otherwise, there exists a vector v: vlvvl+1 and thus the chain is not maximal.

Thus, beginning from [um,,um] and changing at each step one point from um to uM, we get a sequence of N+1-ordered vectors. So, every maximal chain has length N+1.

Then, we prove that the number of such chains is N! using induction.

For N=2, it is easy to verify that we have two chains of 3 elements.

For N=n, we have n! maximal chains of n+1 elements. Then, for N=n+1 we consider one of the previous chains v1v2vN and at each of its elements, we add an extra bit: v~i=[vi,βi]. We observe that if βi=uM, then for all j>i it should hold βj=uM, in order for the new vectors to remain ordered under .

Denote by ic the point that βi change from um to uM. For each choice of ic:βj=um,j<icandβj=uM,j>ic, we take two ordered vectors v~1ic=[vic,um] and v~2ic=[vic,uM] in the new chain, so we have two βic. Thus, we have N+1 possible choices for the β=[βi]{um,uM}N+1. This way we observe that from each chain in ({um,uM}N,), we can construct N+1 chains in ({um,uM}N+1,).

Remark 13

The fact that V has at most N+1 elements is also a consequence of Corollary 1.

A.7 Proof of Proposition 4

  • (i)

    To contradict, assume that IkIk and IkIk. Then, there is a pair of players i1,i2 such that uki1=uki2=uM and uki2=uki1=um, which contradicts Proposition 2(iii).

  • (ii)
    Without loss of generality, assume that IkIk. Then:
    μ(Ik)=ρk=ρk=μ(Ik)=μ(Ik)+μ(Ik\Ik).
    Thus, μ(Ik\Ik)=0. Combining with IkIk and the definition of Ik,Ik, we get μ({i:uki=uki})=1.
  • (iii)

    The equality μ(Ik)=ρk is immediate from the definition of ρk. Consider a pair Ikn,Ikn+1. There are two cases, ρkn<ρkn+1 and ρkn=ρkn+1. In the first case, we cannot have Ikn+1Ikn. Thus, from (i) we have IknIkn+1. If ρkn=ρkn+1, then Inline graphic from part (ii).

    The inclusion KknKkn+1 is immediate from the definition.

  • (iv)

    Let iIkn+1\Ikn. Then, since iIkn ukni=um for nn. On the other hand, μ-almost all iIkn+1 satisfy iIkn, for n>n. Thus, for μ-almost all iIkn+1\Ikn, the action ui is given by (16). The proof is similar for iIk1, and i[0,1)\IkN.

A.8 Proof of Proposition 5

If ρ corresponds to a Nash equilibrium, then combining (13) and (20) we conclude that H(ρ)=0. Conversely, since all the terms of (21) are nonnegative, H(ρ)=0 implies that if μ([i¯j-1,i¯j)[ρkn-1,ρkn))>0, then F(j-1)2N+n(π~(ρ))=δj(π~(ρ)). Combining this with (20), we conclude that if for some jl, π(j-1)2N+l>0, then F(j-1)2N+l(π)=δj(π), where π=π~(ρ). That is, π is a Nash equilibrium.

From (20), we observe that π(ρ) is continuous with respect to ρ, since μ(·) is the Lebesgue measure. Moreover, (21) can be expressed as:

H(ρ)=π(ρ)TΦ(π(ρ)).

The fact that H(ρ) is nonnegative is a result of (13). Furthermore, from Lemma 2, F(j-1)2N+n(π)=J~j(vn,π) is continuous with respect to π. Additionally, δj(π), which is the minimum of F(j-1)2N+l(π)=J~j(vl,π) for all vl, is continuous with respect to π as the minimum of continuous functions. So, Φ(π)=F(π)-δ(π) is continuous with respect to π and H(ρ) is continuous with respect to ρ as composition of continuous functions.

Footnotes

1

Data availability: The datasets generated during and analyzed during the current study are available from the corresponding author on reasonable request.

This article is part of the topical collection “Modeling and Control of Epidemics” edited by Quanyan Zhu, Elena Gubar and Eitan Altman.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Ioannis Kordonis, Email: jkordonis1920@yahoo.com.

Athanasios-Rafail Lagos, Email: lagosth@mail.ntua.gr.

George P. Papavassilopoulos, Email: yorgos@netmode.ntua.gr

References

  • 1.Allen LJ, Brauer F, Van den Driessche P, Wu J (2008) Mathematical epidemiology, vol 1945, Springer
  • 2.Amaral MA, de Oliveira MM, Javarone MA (2020) An epidemiological model with voluntary quarantine strategies governed by evolutionary game dynamics. arXiv preprint arXiv:2008.05979 [DOI] [PMC free article] [PubMed]
  • 3.Amini H, Minca A (2020) Epidemic spreading and equilibrium social distancing in heterogeneous networks [DOI] [PMC free article] [PubMed]
  • 4.Aurell A, Carmona R, Dayanikli G, Lauriere M (2020) Optimal incentives to mitigate epidemics: a Stackelberg mean field game approach. arXiv preprint arXiv:2011.03105
  • 5.Bertsekas DP (1997) Nonlinear programming. J Oper Res Soc 48(3):334–334
  • 6.Bhattacharyya S, Reluga T. Game dynamic model of social distancing while cost of infection varies with epidemic burden. IMA J Appl Math. 2019;84(1):23–43. doi: 10.1093/imamat/hxy047. [DOI] [Google Scholar]
  • 7.Brezis H (2010) Functional analysis. In: Sobolev spaces and partial differential equations, Springer
  • 8.Brown PNN, Collins B, Hill C, Barboza G, Hines L (2020) Individual altruism cannot overcome congestion effects in a global pandemic game. arXiv preprint arXiv:2103.14538
  • 9.Chang SL, Piraveenan M, Pattison P, Prokopenko M (2020) Game theoretic modelling of infectious disease dynamics and intervention methods: a review,” Journal of biological dynamics, vol. 14, no. 1, pp. 57–89 [DOI] [PubMed]
  • 10.Chen FH (2009) “Modeling the effect of information quality on risk behavior change and the transmission of infectious diseases,” Mathematical biosciences, vol. 217, no. 2, pp. 125–133 [DOI] [PubMed]
  • 11.Cho S (2020) Mean-field game analysis of SIR model with social distancing. arXiv preprint arXiv:2005.06758
  • 12.d’Onofrio A, Manfredi P (2009) Information-related changes in contact patterns may trigger oscillations in the endemic prevalence of infectious diseases. J Theor Biol 256(3), 473–478 [DOI] [PubMed]
  • 13.ECDC (2020) Guidelines for the implementation of non-pharmaceutical interventions against COVID-19
  • 14.Eksin C, Shamma JS, Weitz JS. Disease dynamics in a stochastic network game: a little empathy goes a long way in averting outbreaks. Sci Rep. 2017;7(1):1–13. doi: 10.1038/srep44122. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Elie R, Hubert E, Turinici G. Contact rate epidemic control of COVID-19: an equilibrium view. Math Model Nat Phenomena. 2020;15:35. doi: 10.1051/mmnp/2020022. [DOI] [Google Scholar]
  • 16.Facchinei F, Pang J-S (2007) Finite-dimensional variational inequalities and complementarity problems, Springer
  • 17.Funk S, Salathé M, Jansen VA (2010) “Modelling the influence of human behaviour on the spread of infectious diseases: a review,” Journal of the Royal Society Interface, vol. 7, no. 50, pp. 1247–1256 [DOI] [PMC free article] [PubMed]
  • 18.Hota AR, Sneh T, Gupta K (2020) Impacts of game-theoretic activation on epidemic spread over dynamical networks. arXiv preprint arXiv:2011.00445
  • 19.Hota AR, Sundaram S. Game-theoretic vaccination against networked SIS epidemics and impacts of human decision-making. IEEE Trans Control Netw Syst. 2019;6(4):1461–1472. doi: 10.1109/TCNS.2019.2897904. [DOI] [Google Scholar]
  • 20.Huang Y, Zhu Q (2021) Game-theoretic frameworks for epidemic spreading and human decision making: a review. arXiv preprint arXiv:2106.00214 [DOI] [PMC free article] [PubMed]
  • 21.Huang Y, Zhu Q. A differential game approach to decentralized virus-resistant weight adaptation policy over complex networks. IEEE Trans Control Netw Syst. 2019;7(2):944–955. doi: 10.1109/TCNS.2019.2931862. [DOI] [Google Scholar]
  • 22.Kabir KA, Tanimoto J. Evolutionary game theory modelling to represent the behavioural dynamics of economic shutdowns and shield immunity in the COVID-19 pandemic. R Soc Open Sci. 2021;7(9):201095. doi: 10.1098/rsos.201095. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Karlsson C-J, Rowlett J. Decisions and disease: a mechanism for the evolution of cooperation. Sci Rep. 2020;10(1):1–9. doi: 10.1038/s41598-020-69546-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Kermack WO, McKendrick AG (1927) A contribution to the mathematical theory of epidemics. In: Proceedings of the royal society of London, series A, containing papers of a mathematical and physical character, vol 115, no 772, pp 700–721
  • 25.Khalil HK, Grizzle JW (2002) Nonlinear systems, vol 3, Prentice Hall, Upper Saddle River
  • 26.Kordonis I, Lagos A-R, Papavassilopoulos GP (2020) Nash social distancing games with equity constraints: how inequality aversion affects the spread of epidemics. arXiv preprint arXiv:2009.00146
  • 27.Lagos A-R, Kordonis I, Papavassilopoulos G (2020) Games of social distancing during an epidemic: local vs statistical information. arXiv preprint arXiv:2007.05185 [DOI] [PMC free article] [PubMed]
  • 28.Mas-Colell A. On a theorem of Schmeidler. J Math Econ. 1984;13(3):201–206. doi: 10.1016/0304-4068(84)90029-6. [DOI] [Google Scholar]
  • 29.Paarporn K, Eksin C, Weitz JS, Shamma JS. Networked SIS epidemics with awareness. IEEE Trans Comput Soc Syst. 2017;4(3):93–103. doi: 10.1109/TCSS.2017.2719585. [DOI] [Google Scholar]
  • 30.Poletti P, Ajelli M, Merler S (2012) “Risk perception and effectiveness of uncoordinated behavioral responses in an emerging epidemic,” Mathematical Biosciences, vol. 238, no. 2, pp. 80–89 [DOI] [PubMed]
  • 31.Poletti P, Caprile B, Ajelli M, Pugliese A, Merler S (2009) “Spontaneous behavioural changes in response to epidemics,” Journal of theoretical biology, vol. 260, no. 1, pp. 31–40 [DOI] [PubMed]
  • 32.Reluga TC. Game theory of social distancing in response to an epidemic. PLoS Comput Biol. 2010;6(5):e1000793. doi: 10.1371/journal.pcbi.1000793. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Ross R (1916) An application of the theory of probabilities to the study of a priori pathometry. In: Proceedings of the royal society of London, series A, containing papers of a mathematical and physical character, vol 92, no 638, pp 204–230
  • 34.Schmeidler D (1973) “Equilibrium points of nonatomic games,” Journal of statistical Physics, vol. 7, no. 4, pp. 295–300
  • 35.Shapiro A, Dentcheva D, Ruszczyński A. Lectures on stochastic programming: modeling and theory. Philipedia: SIAM; 2014. [Google Scholar]
  • 36.Theodorakopoulos G, Le Boudec J-Y, Baras JS (2012) “Selfish response to epidemic propagation,” IEEE Transactions on Automatic Control, vol. 58, no. 2, pp. 363–376
  • 37.Toxvaerd F (2020) Equilibrium social distancing, Cambridge working papers in economics
  • 38.Trajanovski S, Hayel Y, Altman E, Wang H, Van Mieghem P. Decentralized protection strategies against SIS epidemics in networks. IEEE Trans Control Netw Syst. 2015;2(4):406–419. doi: 10.1109/TCNS.2015.2426755. [DOI] [Google Scholar]
  • 39.Ye M, Zino L, Rizzo A, Cao M (2020) Modelling epidemic dynamics under collective decision making. arXiv preprint arXiv:2008.01971

Articles from Dynamic Games and Applications are provided here courtesy of Nature Publishing Group

RESOURCES