Skip to main content
PLOS One logoLink to PLOS One
. 2019 Jan 15;14(1):e0209819. doi: 10.1371/journal.pone.0209819

Intervention on default contagion under partial information in a financial network

Yang Xu 1,*
Editor: Fenghua Wen2
PMCID: PMC6333338  PMID: 30645587

Abstract

We study the optimal interventions of a regulator (a central bank or government) on the illiquidity default contagion process in a large, heterogeneous, unsecured interbank lending market. The regulator has only partial information on the interbank connections and aims to minimize the fraction of final defaults with minimal interventions. We derive the analytical results of the asymptotic optimal intervention policy and the asymptotic magnitude of default contagion in terms of the network characteristics. We extend the results of Amini, Cont and Minca’s work to incorporate interventions and adopt the dynamics of Amini, Minca and Sulem’s model to build heterogeneous networks with degree sequences and initial equity levels drawn from arbitrary distributions. Our results generate insights that the optimal intervention policy is “monotonic” in terms of the intervention cost, the closeness to invulnerability and connectivity. The regulator should prioritize interventions on banks that are systematically important or close to invulnerability. Moreover, the regulator should keep intervening on a bank once having intervened on it. Our simulation results show a good agreement with the theoretical results.

Introduction

Introduction

The systemic risk in a financial network has been drawing more and more interest of regulators and researchers, especially after the Asian financial crisis in the late 1990’s and the more recent economic recession during 2007-2009. Financial institutions (hereafter, banks) are connected to form an interbank network to allow liquidity reallocation between the banks in that banks with liquidity surpluses can lend to banks with liquidity deficits. However, the interbank network may also introduce aggregate liquidity shortage and default contagion. Liquidity shock such as a run will cause some banks to default, leading to losses of their creditor banks through interbank connections, which may in turn result in losses of their creditors. In the financial crisis of 2007-2009 the interbank market dysfunctioned because the market participants perceived heightened counterparty risk and liquidity risk [1], and the severe reduction in transaction volume was a major contributing factor to the collapse of many banks. When the interbank market is stressed or freezes, the central bank as the lender of last resort has to provide extensive short term liquidity support. For example, the Federal Reserve in the US established many facilities, including the traditional discount window, the Term Auction Facility (TAF), Primary Dealer Credit Facility (PDCF), and Term Securities Lending Facility (TSLF). The central bank or government also recapitalize the banks and provide risk capital in the form of a bailout. Naturally, we ask the following questions: Should the regulator (the central bank or government) intervene if the bankruptcy of one or several banks has occurred or is imminent? What is the optimal intervention policy of the regulator based on the measurable features of the network and the banks, such as the degrees of connectivity and the levels of capitalization? How much improvement can the optimal strategy achieve regarding the fraction of market protected from defaults?

To answer these questions we study the uncollateralized interbank funding market where the majority of interbank loans are overnight. The connections are constantly changing so the regulator may not know exactly all the connections over time. After some liquidity incident, e.g. runs on a few banks, some banks default, which initiates the default contagion process. During the process, the regulator has to intervene on the defaulted banks to prevent the contagion from spreading. Due to the system panics, no banks want to lend new loans to other banks but meanwhile they are still obliged to pay back their current loans which are due in the time frame of the model. So it is reasonable to fix the in and out degrees of banks and assume all the connections do not change any more after the inception of default contagion. Theoretically, the regulator would be able to find out the connections between the banks by communicating with the banks; however, the contagion process happens so fast that the regulator may not have ample time to find out the connections between the banks. In other words, the regulator will have to intervene while all the connections between the banks are unknown. So the regulator has only partial information because in the beginning the regulator only knows the initially defaulted banks and the magnitude of connections of each bank but the connections between the banks are unknown. Every time a default occurs the regulator learns the connections of the defaulted bank (i.e. the banks that are affected by the default), represented by the connections being revealed after the default.

We set up a probability space under which the financial network is generated by a uniform matching of the in and out degrees (a configuration model). A directed link in the network represents one unit of loan. In the following we may use “bank” and “node” interchangeably. Before an external shock to the system, each node has a positive equity level (the difference of total assets and liabilities), which indicates the number of defaulted loans due to the default of its debtors a node can withstand before it defaults. In other words, it is the “distance to default”. After an external shock, some nodes in the system default initially and we set their equity levels to zero. We adopt the dynamics of the model in [2, 3]. When a node defaults, it defaults on all of its loans. We assume a zero recovery rate of the loan, i.e. the creditor receives zero value from the loan, which is the most realistic assumption for short term default as suggested in [4]. We assume there is a time span between a node’s default and the time its creditor records the loan as a loss (by writing down the loan from its balance sheet). We model this time span by independently and identically distributed exponential random variables. After the affected node records the defaulted loan, it may request the regulator for interventions. If the regulator decides to intervene by replacing the defaulted loan or by infusing one unit of equity, the equity level of the affected node will stay the same, otherwise its equity level will decrease by one. Once the equity level reaches zero, the node defaults. We assume that the once a bank has defaulted it cannot become liquid again within the time horizon of the model because it is very unlikely for a bank that has declared default to gain enough capital in the short term considered in the model.

We emphasize that one essential feature of the model is partial information. Because it takes time and effort for the regulator to find out the exact connections in the interbank market and the default contagion process may progress very rapidly, the regulator may have to intervene even before it is able to figure out the connections in order to take early actions and save the maximum amount of banks from bankruptcy. During the contagion process, the regulator only knows the default set (the set of defaulted nodes) with some out-links revealed, but meanwhile other out-links remain hidden until the affected nodes record the defaulted loans. More generally, unlike the complete information assumption in other theoretical models, the partial information setting aligns with the reality better, as pointed out by [5]: “Interbank exposure data are never publicly available, and in many countries nonexistent even for central regulators”.

Our methodology is illustrated in Fig 1. The regulator’s goal is to minimize the number of final defaulted nodes with the minimum number of interventions. Thus we obtain a stochastic control problem minμnobjn(Grn,μn) where the objective function depends on the graph Grn and the intervention sequence μn, shown in (1). We aim to solve it for the optimal intervention sequence μn* and thus obtain the optimal objective function value objn(Grn,μn*), shown in (2). However, solving the problem with the usual dynamic programming approach will incur intractability problem because of the fast expansion of the state space as pointed out in [3], especially for a heterogeneous network. We take an alternative approach based on the fact that under some regularity conditions, the objective function converges as n → ∞. We solve the asymptotic optimal control problem minμ obj(Gr, μ) in (3) where obj, Gr and μ are the limit forms of objn, Grn and μn, respectively, and obtain the optimal intervention μ* and the objective function obj(Gr, μ*) in (4) which allow us to construct the optimal intervention sequence μn* for a finite n through μ* and approximate objn(Grn,μn*) with obj(Gr, μ*). Our results of the numerical experiments validate the approximation for networks with sizes close to the real financial networks.

Fig 1.

Fig 1

The methodology: Approximation of the finite network with the infinite network.

Relations to previous literature

Our work is closely related to the current literature on the role of the central bank as the lender of last resort, including providing liquidity by a loan, recapitalizing the banks and bank bailout. These studies differ in their perspectives and focuses in their models. The influencing Diamond & Dybvig model [6] about the market panics and bank runs has two Nash equilibria: depositors withdraw only for their real expenditure needs or a bank run. [7] extend the Diamond & Dybvig model to a financial network of four banks to study the default contagion. The interbank network is formed to allocate liquidity among the banks to satisfy regional liquidity demands. In a complete market where the banks exchange deposits or in a disconnected market no contagion occurs while in an incomplete market where the banks do not exchange deposits with all other banks, high connectivity may entail contagion. The role of the central bank is thus to complete the market. [8] introduce a similar model but assume the depositors are uncertain about where they have to consume. In their model the central bank acts as a “crisis manager”: when a bank is to be liquidated, the central bank has to organize the bypass of the defaulting bank in the payment network and provide liquidity to the banks that depend on the defaulting bank. [9] consider that the fire sales of the banking assets occur when a large number of banks default and investors outside the banking sector who are inefficient users may end up purchasing the liquidated assets. To avoid the allocation inefficiency, the regulator may bail out the banks directly or provides liquidity to surviving banks to purchase defaulting banks. [10] argue that the government should bail out banks in distress because it can provide liquidity more efficiently than private investors. [11] consider three forms of regulator interventions: buying equity, purchasing assets and providing debt guarantee to alleviate debt overhang in a financial market but the regulator has limited information and resources. All the works discuss the optimal interventions based on equilibrium analysis. In contrary, our model emphasizes the aspect of the interbank market that it is a complex network and focus on how the regulator should make intervention decisions under the network dynamics. Moreover, some papers study the related problem about the banks in the interbank network bailing each other out, such as [12] and [13].

In addition to the theoretical studies, empirical studies abound. [14] analyze the data on the interbank transactions derived from the main euro area payment system and find that the European Central Bank took the role of the overnight unsecured interbank market in liquidity provision to the banks during the global financial crisis in 2008-2010. [15] analyze the daily transaction data and find that in 2008 counterparty risk plays a more important role than liquidity hoarding in reducing liquidity and increasing the cost of finance in the federal funds market in the US. By analyzing supervisory data of Germany, [16] find that regulatory interventions decrease liquidity creation while capital support does not affect it. [17] discuss the relations between liquidity regulations and the lender of last resort practice and argue that they are complementary rather than conflicting tools.

Our work is also related to the strand of literature on systemic risk and default contagion in the financial networks without considering regulator interventions. These works focus on understanding the dependence of the default contagion on various features of the financial network and the banks within it, including the degrees of connectivity, the equity levels and so on. Similarly, there are mainly two types of literature: empirical and theoretical. The empirical studies conduct statistical analyses on the interbank markets using data on interbank lending as far as they are available and provide an overview of the structural characteristics of the interbank network in different countries ([4, 18, 19]). The theoretical studies model the financial network with network models but differ in their assumptions about the network structure and approaches: some focus on “stylized” networks whose structures are hypothetical ([7, 20, 21, 22]) while others rely on simulations ([4, 23]). Among them, [24] and [25] propose random network models that allow more realistic and heterogeneous structures. [26] survey theoretical works on contagion and systemic risk in financial networks and categorize them according to different topics including network connectivity, bank heterogeneity, uncertainty in financial markets, and portfolio composition of the banks.

In regards to section Introduction, this paper is closely related to [24] and [2, 3]. [24] study the magnitude of default contagion in a heterogeneous network with given degree sequence and arbitrary distribution of weights and derive the analytical expressions of the asymptotic fractions of defaults in terms of the network characteristics. Our work incorporates interventions into a model proven to be equivalent to theirs. Thus if there are no interventions, the asymptotic fraction of final defaults will be the same as in [24]. [2] consider a stylized core-periphery financial network as an intermediary to provide liquidity to fund projects in outside economy but it may also incur contagion when the banks hoard liquidity. The regulator intervenes by providing loans to defaulting banks. [3] consider a similar core-periphery model where the regulator intervenes by injecting equity. We adopt the dynamics of their model that constructs the default set under interventions through a configuration model because the configuration model can be adapted to the contagion process [27]. But we differ from [2, 3] in two important ways. First, our model is a more general heterogeneous random network with degree sequences and initial equity levels drawn from arbitrary distributions. More importantly, [2, 3] focus on the benefits and costs of the connectivity in the presence of the regulator and draw conclusions mainly from numerical studies, while we focus on the optimal intervention policy and its relations to the network characteristics. Mathematically we have successfully addressed two major mathematical difficulties arising from considering interventions on a general complex network: Interventions introduce discontinuity into the asymptotic process thus the main supporting theorem used in [24] is no longer applicable directly; moreover, the high dimensional optimal control problem we obtain later is well known among control theorists difficult to solve, especially analytically. We give analytical formulations for the asymptotic optimal intervention policy as well as the asymptotic number of interventions and final defaulted banks. The asymptotic results provide a good approximation to real financial networks, which are heterogeneous and have several hundred to a thousand of banks thanks to the fast convergence behavior of our results.

Contributions

The main contribution is that we have proposed a new approach to determine the optimal intervention strategy on contagions in a large and heterogeneous financial network-to be specific, we derive the asymptotic optimal intervention strategy as the size of the network tends to infinity and then show that it is a good approximation of the optimal intervention strategy for a real financial network. This new approach has the advantage of avoiding the stylized model in the equilibrium analysis or the intractability of the dynamic programming approach in previous literature and enabling us to obtain analytical results. In light of this, we derive rigorous asymptotic results of the optimal strategy for the regulator and the fraction of final defaulted banks under the dynamics of the default contagion process in a heterogeneous network with a degree sequence and initial equity levels that can be drawn from arbitrary distributions. The analytical expressions are presented in terms of measurable features of the network. For convergence of the results, we assume the network is sparse as in [24] which is supported by the empirical studies of real financial networks ([28, 29]).

The key insights of our findings of the optimal strategies are summarized in the following. We should only consider intervening on a bank when it records the loss of a defaulted loan and is very close to default. The optimal intervention policy depends strongly on the intervention cost. The smaller the intervention cost, the more interventions are implemented. Moreover, the optimal intervention policy is “monotonic” with respect to the measurable features of the network. We should not intervene on the banks with out degrees in a certain range regardless of their other features; for those banks worth interventions, the larger the sum of initial equity and accumulative interventions received, the earlier we should begin intervention on them; the time to start intervention on a node is also “monotonic” in its in and out degrees. Interestingly, once we begin intervening on a node, we keep intervening on it every time it records the loss of a defaulted loan. By comparing the fractions of final defaults under no interventions and the optimal intervention policy, we are able to quantify the improvement made by interventions in terms of the network features. This gives guidance for the maximum impact the regulator can have to offset the effects of default contagion.

The paper is organized as follows. We set up the model and introduce the stochastic control problem (SCP) in section Model description and dynamics. In section The asymptotic control problem we formulate the asymptotic control problem Eq (9) that gives the limit for the objective function of the SCP as the size of the network goes to infinity and present the necessary conditions for the optimal intervention policy, which lead to the main theorems. In section Numerical experiments we show the results of the numerical experiments to validate the approximation of Eq (4) by Eq (9). We present in Appendix A: Proofs all the proofs and in Appendix B: Wormald’s theorem and Appendix C: Extended pontryagin maximum principle the two theorems used in the proofs as well as a list of notations in Appendix D: Preliminary list of notations.

Model description and dynamics

Basic setup

We consider default contagion in an unsecured interbank lending market under short term illiquidity risk based on the model of [2]. Due to the system panics, no banks want to lend new loans to other banks but meanwhile they are still obliged to pay back their current loans which are due in the time frame of the model, so we fix the in and out degrees of banks in the network, denoted as (d(v), d+(v))v∈[n], where [n] = {1, …, n} the set of nodes. Let m = ∑v∈[n] d(v) = ∑v∈[n] d+(v).

Then we model the financial network with prescribed degree sequence (d(v), d+(v))v∈[n] as an unweighted directed network ([n],En), where En denotes the set of links. A directed link (v,w)En represents v borrows a unit of loan from w, i.e. v is obliged to repay w one unit of loan. We allow multiple loans to exist between two nodes and also self links representing the internal loans between different departments of the bank.

Now we set up a probability space (Gn,m,P) where Gn,m is the set of networks on n nodes with at most m directed links. So the random financial network with m directed links lives in this probability space and under P the law of the random link set En is determined as follows. We start with n unconnected nodes and assign node v d(v) in-half-links and d+(v) out-half-links. An in-half-link represents an offer of a loan and an out-half-link a demand for a loan. Then the m in-half-links and m out-half-links are matched uniformly so that the borrowers and lenders are determined. The resulting random network is called the configuration model.

The uniform matching of the in and out-half-links allows us to construct the random network sequentially: at every step we can choose any out half-link by any rule and choose the in-half-link uniformly over all unconnected in-half-links to form a directed link. This is because conditional on any subset of connected links, the unconnected links also follow the uniform distribution. Moreover, the conditional law of unconnected links only depends on the number of connected links, not the matching history. Additionally we can restrict the matching to choosing only the out-half-links from the defaulted nodes so that we can model the development of the default set with their revealed out-links.

Remark 1. As a result of the uniform connection of in and out-half-links, a node gets selected with the probability proportional to the number of its unconnected in-half-links. The rationale is that when the regulator searches for the lender of a defaulting loan, a bank that lends out more loans to other banks is more likely to be the lender and be affected by the defaulting loan.

Then we endow a node v ∈ [n] with its initial equity level e(v)N0{0,1,2,} which represents the number of defaulted loans v can tolerate until v defaults, so it is the “distance to default”. Next after the system receives some external shock, some nodes default and the system begins to evolve. Define time 0 as the time right after the shock. Let (Gk)0km be the filtration for the probability space (Gn,m,P) which models the arrival of new information, i.e. the revealed link at each step. Because this implies that the the remaining equity of the selected node will decrease by one, (Gk)0km also models the default contagion at the same time. Note in the following the network with the set of revealed links evolves in the space Gn,m as the result of the contagion process.

Initial condition

Define the set of initially defaulted nodes D0{v[n]:e(v)0} and the σ-algebra representing the information available initially G0σ((d-(v),d+(v),e(v))v[n]). Given the degree pairs for all nodes (d(v), d+(v))v∈[n], envision for a node v ∈ [n], there are d(v) in-half-links each representing a loan another node is obliged to pay v and d+(v) out-half-links each representing a loan v is obliged to pay another node. Let ckv be the sum of initial equity and accumulative number of interventions on node v and lkv be the number of revealed in-links of node v at step k, so c0v=e(v), l0v=0. An example of the initial condition of a four node network is illustrated in Fig 2 with D0={1,2}.

Fig 2. Financial network before default contagion occurs.

Fig 2

(d(v), d+(v)) are the degrees and e(v) is the initial equity of node v. The nodes in the initial default set D0={1,2} are marked in blue.

Dynamics

We adopt the dynamics from [2]. At the kth step for k ∈ [1, m], if the out-links of nodes in Dk-1 have not all been revealed, then the new link is revealed following the rule: an out-half-link of any node in Dk-1 is picked by any rule and then connected uniformly to another unconnected in-half-link. Let (Vk, Wk) be a pair of random variables denoting the link from node Vk to node Wk being revealed at step k. We call Wk is selected at step k. Assume (Vk, Wk) = (v, w), then the uniformity in connecting the half-links leads to the probability of w being selected conditional on Gk-1 as

P(Wk=wGk-1)=numberofwsunrevealedin-linksatk-1totalnumberofunrevealedin-linksatk-1=d-(w)-lk-1wm-(k-1). (1)

So a node is selected with the probability proportional to the number of its unrevealed (unconnected) in-half-links. After a directed link (v, w) is revealed, then proceed with the following steps:

  • Update Gk=σ(Gk-1{(v,w)}).

  • Update the number of revealed out-links: lkw=lk-1w+1 and lkη=lk-1η for ηw.

  • Determine the intervention μk ∈ {0, 1} Gk measurable at step k for the selected node w. Note that ckηlkη indicates that the node η has defaulted by step k. Because we do not intervene on defaulted node, μk = 0 if ck-1wlk-1w.

  • Update ckw=ck-1w+μkw, otherwise ckη=ck-1η for ηw.

  • Update the default set Dk. If ckwlkw and wDk-1, then Dk=Dk-1{w}, otherwise Dk=Dk-1.

If all out-links from Dk have been revealed, the process ends and let the process end time be Tn = k, otherwise repeat the process. Define DTn as the number of defaulted nodes by the process end time Tn. In Fig 3 we show the first step of the dynamics for the network in Fig 2. The link (1, 4) is revealed (connected) and node 4 is selected with probability 26. Because the node 4 is liquid, the regulator needs to decide whether to intervene. Node 4 remains liquid if it receives one intervention, or it will default and be included in the default set.

Fig 3. Dynamics at step one.

Fig 3

Node 4 is selected with probability 26 and a link between 1 and 4 is revealed. If one intervention is applied, node 4 remains liquid, otherwise its equity level will decrease by one and node 4 defaults.

Remark 2. The notion of partial information is reflected by the fact that there are always unrevealed links before the process ends and the regulator cannot make decisions depending on the knowledge of it. The regulator can only be certain that if the remaining equity is nonpositive (ckηlkη) then the node η has defaulted at step k. On the other hand, every out-half-link from the default set represents a defaulted loan which is possible to impact every node in the network at a later time. So all the current liquid vulnerable nodes are subject to default at a later time and the regulator should take this into account when making intervention decisions at each step.

Let Rk be the accumulative number of interventions by step k and Dk=|Dk| be the number of defaults at step k. The regulator aims to minimize the number of defaulted nodes by Tn with the minimum amount of interventions, so we define the objective function as a linear combination of the (scaled) number of interventions and defaults by the end of the process Tn as

Jn=E(KRTnn+DTnnG0), (2)

where K > 0 is the relative “cost” of an intervention. Further by the definition of cTnv and noting that a node defaults at last if cTnvlTnv, i.e. the number of defaulted loans exceeds the total of the initial equity level and the number of interventions received by Tn, we can express RTn and DTn as

RTn=v[n](cTnv-e(v)),DTn=v[n]1(cTnvlTnv). (3)

Now we define the stochastic optimal control problem as

minμUJn, (4)

where μ = (μk)1≤km, μk ∈ {0, 1} and U contains all (Gk)0km adapted process μ.

The asymptotic control problem

Assumptions and definitions

We assume that a bank cannot become liquid again once it has defaulted, thus we cannot save defaulted banks. This assumption is reasonable in the setting of default contagion in a stressed network and a short time window. Nor do we intervene on invulnerable nodes, because they never default but intervening on them will only prevent us from saving the banks that are very close to default especially when the interventions are costly.

In the model description we only intervene on the node that is selected at each step. Now we show that even if the regulator intervenes on multiple nodes and applies more than one unit of credit every time, it will not be better.

Proposition 1. For the stochastic control problem Eq (4), we only need to consider intervening on a node that, when selected, has only one unit of equity remaining.

We see proposition 1 implies that it is never optimal to intervene on a node if it is not selected or has more than one unit of equity remaining when selected. Let (i, j, c, l) be the state of a node, meaning it has the in and out degree (i, j), sum of the initial equity and the number of interventions c and l revealed in-links. Note that by definition li. We characterize nodes with states because nodes with the same state have the same probability of being selected at each step and are statistically the same in influencing other nodes. Note in particular:

  1. c = 0 denotes that the node has defaulted initially.

  2. cl denotes the remaining equity or “distance to default”, i.e. the number of times of being selected before a node defaults without interventions. Thus cl means the node has defaulted.

  3. Because li by definition, i < c implies that a node is invulnerable, i.e. even all loans lent out to the counterparties are written down from the balance sheet, the node still has positive remaining equity. On the contrary, 0 < ci denotes the node has the possibility to default, i.e. vulnerable.

  4. In the beginning of the contagion process, all nodes have l = 0, i.e. are in states of the form (i, j, c, 0).

Then we define the state of the system at each step. Note that the number of nodes that have defaulted initially (c = 0) or invulnerable (i < c) in the beginning will not change throughout the process, so we only need to keep track of the nodes that are initially vulnerable (0 < ci) and currently liquid and if needed, we can always calculate the number of defaulted nodes at any time in the process. Further note that the possible states throughout the process for nodes that are vulnerable in the beginning and liquid at a later step are

Γ{(i,j,c,l):0i,0j,0l<ciorc=i+1,l=i}. (5)

Note particularly the state (i, j, i + 1, i) is the result that a node in state (i, j, i, i − 1) is selected and receives one intervention and thus becomes invulnerable.

Definition 1. (State variable Sk) Let Ski,j,c,l denote the number of nodes that are vulnerable initially and are in state (i, j, c, l) at step k, for k = 0, …, m and Sk(Ski,j,c,l)(i,j,c,l)Γ be the state of the system. Note in the following we may use α to represent (i, j, c, l)∈Γ and write Skα instead of Ski,j,c,l to simplify the notation.

Recall m = m(n) is the number of the total in (or out) degree of the network, which is also the maximum steps of the process. Throughout this paper we follow the convention that the superscript (usually a multi-index) denotes the state and the subscript denotes the time (discrete or continuous), e.g. sτi,j,c,l, uτi,j,c,c-1, sti,j,c,l, wti,j,c,l and uti,j,c,c-1 in the following. Then we define the empirical probability of in, out degrees and initial equity levels.

Definition 2. (Empirical probability) Define the empirical probability of the triplet (in degree, out degree, initial equity level) as

Pn(i,j,c)=1n|{v[n]d-(v)=i,d+(v)=j,e(v)=c}|. (6)

Note that c0Pn(i,j,c)=1n{v[n]d-(v)=i,d+(v)=j} represents the empirical probability of the in and out degree pair (i, j).

Previously we use Wk to denote the selected node at step k. Now with a little abuse of notation, let Wk denote the state of the selected node at step k, k = 1, …, m, so Wk ∈ Γ+ ≔ {(i, j, c, l):0 ≤ i, 0 ≤ j, 0 ≤ c, 0 ≤ li}. We consider a Markovian control policy Gn=(g1(n)(S0,W1),,gm(n)(Sm-1,Wm)) where gk+1(n):N0|Γ|×Γ+{0,1} specifies the intervention at step k + 1 on the selected node which has state Wk+1 given the state Sk and the superscript (n) shows the dependence on n.

Letting Pn = (Pn(i, j, c))i,j,0≤ci, we rewrite the terms Jn as JGn(Pn), RTn as RTn(Gn,Pn) and DTn as DTn(Gn,Pn) in Eq 4 based on Gn and Pn, so

RTn(Gn,Pn)=k=1Tngk(n)(Sk-1,Wk),DTn(Gn,Pn)=ni,jPn(i,j,0)+ni,j,1ciPn(i,j,c)-(i,j,c,l)ΓSTni,j,c,l=ni,j,0ciPn(i,j,c)-(i,j,c,l)ΓSTni,j,c,l. (7)

Note that the first equality for DTn(Gn,Pn) holds because the nodes that default at the end of the process consist of two parts: the nodes that have defaulted initially i.e. ni,j Pn(i, j, 0) and those nodes that are vulnerable initially and default during the process i.e. ni,j,1ciPn(i,j,c)-(i,j,c,l)ΓSTni,j,c,l.

Assumption 1. Consider a sequence ([n],En) of random networks, indexed by the size of the network n. For each nN, (d(v))v∈[n], (d+(v))v∈[n] are sequences of nonnegative integers with ∑v∈[n] d(v) = ∑v∈[n] d+(v) and such that for some probability distribution p on N03 independent of n with λ ≔ ∑i,j,c ip(i, j, c) = ∑i,j,c jp(i, j, c)<∞, the following holds

  1. Pn(i, j, c) → p(i, j, c) ∀ i, j, c ≥ 0 as n → ∞.

  2. v∈[n][(d(v))2+ (d+(v))2] = O(n).

Note that the second assumption implies by uniform integrability that m(n)nλ as n → ∞ and recall that m(n) ≔ ∑v∈[n] d(v) = ∑v∈[n] d+(v). Since km(n), for large n it holds that knm(n)nλ+1. Assumption 1 essentially implies the network is sparse which is justified in many empirical study literature on the structure of financial networks [24].

Remark 3. We previously defined that the vector Pn only includes Pn(i, j, c) in the range 0 ≤ ci, i.e. the fractions of initially defaulted and vulnerable nodes. Accordingly define p ≔ (p(i, j, c))i,j,0≤ci, i.e. the vector p only includes p(i, j, c) in the range 0 ≤ ci.

Next we present our assumptions on the control functions gk(n).

Assumption 2. Define Φ ≔ {(i, j, c, c − 1): 0 ≤ i, 0 ≤ j, 1 ≤ ci}. Let Gn=(g1(n),,gm(n)) be the a control policy (a sequence of control functions) for the contagion process on a network of size n where n is large enough such that m(n)nλ+1. Assume that

gk+1(n)(s,w)={ukni,j,c,c-1ifw=(i,j,c,c-1)Φ0otherwise, (8)

for 0 ≤ km − 1. uτi,j,c,c-1=ui,j,c,c-1(τ) where ui, j, c, c−1: [0, λ + 1] → {0, 1} is a piecewise constant function on [0, λ + 1], i.e. there is a partition of the interval into a finite set of intervals such that ui, j, c, c−1 is constant 0 or 1 on each interval. Let u = (uβ)β∈Φ and Π contain all piecewise constant vector function u on [0, λ + 1].

Note that Φ includes possible states having the distance to default equal to one and Φ ⊂ #x0393;. Further gk+1(n)(s,w)=0 for w ∉ Φ follows from proposition 1. In the following we may use β to represent (i, j, c, c − 1) ∈ Φ and write uτβ instead of uτi,j,c,c-1 to simplify the notation.

Remark 4. By this assumption the function u is independent of the state but only a function of time. This implies that the control function gk+1(n)(s,w) depends on the scaled time kn and the state of the currently selected node w but not on the state s. We will show it suffices to consider such control policy Gn later after proposition 3 because given a function u, we can predict a deterministic process to which the scaled stochastic contagion process converges in probability at any time as the size of the network n → ∞. Moreover, this type of control policies is the one that can be solved in the optimal control problem Eq (39) we will introduce later.

In summary, assumption 1 assumes the convergence of the empirical probabilities of the in and out degrees and the initial equity. On the other hand, proposition 3 indicates that the control functions depend on the scaled time and the state of the currently selected node. These two assumptions allow us to define the following asymptotic control problem by ensuring that the limits in the objective function are well defined.

Definition 3. For a sequence of networks with Pn and Gn satisfying assumption 1 and assumption 2, respectively, define the asymptotic control problem as

minuΠlimnJGn(Pn)=minuΠlimnKERTn(Gn,Pn)n+EDTn(Gn,Pn)n. (9)

In the following we will show the limits in Eq (9) are well defined by applying Wormald’s theorem [30].

Dynamics of the default contagion process with interventions

Recall that Rk is the accumulative number of interventions up to step k, so

R0=0,Rk==1kg(S-1,W)==1kβΦ1(W=β)u-1nβ. (10)

We shall show that (Sk, Rk)k=0, …, m is a controlled Markov chain given a control policy Gn. In Fig 4 we illustrate for the same (i, j) pair the state space as well as their transition relations between the states.

Fig 4. The state space for the same (i, j) pair, 0 ≤ i, 0 ≤ j and their transition relations.

Fig 4

To describe the transition probabilities, assume the state of the selected node at step k + 1 is Wk+1 = (i, j, c, l), for k = 0, …, m − 1, there are three possibilities:

The selected node has defaulted, i.e. cl or the node is invulnerable, i.e. c > i, then Sk+1 = Sk, Rk+1 = Rk.

  1. The selected node is vulnerable but has the “distance to default” greater than one, i.e. cl ≥ 2, then the node w is selected with probability (i-l)Ski,j,c,lm-k and
    Sk+1i,j,c,l=Ski,j,c,l-1,Sk+1i,j,c,l+1=Ski,j,c,l+1+1,Rk+1=Rk, (11)
    while other entries of Sk+1 are the same as Sk.
  2. The selected node has the “distance to default” of one, i.e. cl = 1, then the node is selected with probability (i-c+1)Ski,j,c,c-1m-k and by assumption 2,
    Sk+1i,j,c,c-1=Ski,j,c,c-1-1,Sk+1i,j,c+1,c=Ski,j,c+1,c+gk+1(n)(Sk,(i,j,c,c-1))=Ski,j,c+1,c+ukni,j,c,c-1,Rk+1=Rk+ukni,j,c,c-1, (12)
    while other entries of Sk+1 are the same as Sk.

Let (Fk)k=0,,m be the natural filtration of Sk, ΔSkα=Sk+1α-Skα, α ∈ Γ and ΔRk = Rk+1Rk, it follows that

E[ΔSki,j,c,0|Fk]=-iSki,j,c,0m-kfor1ci,E[ΔSki,j,c,l|Fk]=(i-l+1)Ski,j,c,l-1m-k-(i-l)Ski,j,c,lm-kfor3ci,1lc-2,E[ΔSki,j,c,c-1|Fk]=(i-c+2)Ski,j,c-1,c-2m-kukni,j,c-1,c-2+(i-c+2)Ski,j,c,c-2m-k-(i-c+1)Ski,j,c,c-1m-kfor2ci,E[ΔSki,j,i+1,i|Fk]=Ski,j,i,i-1m-kukni,j,i,i-1,E[ΔRk|Fk]=(i,j,c,c-1)Φ(i-c+1)Ski,j,c,c-1m-kukni,j,c,c-1. (13)

Convergence of the default contagion process with interventions

Based on the dynamics of the contagion process under interventions described in the previous section, we will show that the state variable Sk, the accumulative interventions Rk, the number of defaults Dk and the number of unrevealed out-links from the default set Dk- (defined later) after being scaled by n all converge to a deterministic process which depends on the solution of the system of ODEs we will present now. Then we are able to show that the stochastic control problem Eq (4) converges to the asymptotic control problem Eq (9).

Definition 4. (ODEs of sτ) Given a set of piecewise constant function u = (uβ)β∈Φ on [0, λ], i.e. uΠ, define the system of ordinary differential equations (ODEs) of sτ=(sτα)αΓ as

dsτi,j,c,0dτ=-isτi,j,c,0λ-τfor1ci,dsτi,j,c,ldτ=(i-l+1)sτi,j,c,l-1λ-τ-(i-l)sτi,j,c,lλ-τfor3ci,1lc-2,dsτi,j,c,c-1dτ=(i-c+2)sτi,j,c-1,c-2λ-τuτi,j,c-1,c-2+(i-c+2)sτi,j,c,c-2λ-τ-(i-c+1)sτi,j,c,c-1λ-τfor2ci,dsτi,j,i+1,idτ=sτi,j,i,i-1λ-τuτi,j,i,i-1. (14)

The ODEs can be expressed in the form dsτdτ=h(τ,sτ;uτ) where h = (hα)α∈#x0393;.

For what is needed below we analyze the solutions of the ODEs in definition 4 for a subinterval of [0, λ] on which uτ is a constant vector function.

Proposition 2. Let sτ=(sτα)αΓ satisfy the system of ordinary differential equations in definition 4 with the initial conditions sτ1=s1(s1α)αΓ and assume uτ is a constant vector function uτ = b ≔ (bβ)β∈Φ in the interval [τ1, τ2) ⊆ [0, λ) where bβ ∈ {0, 1} is a constant, then the solution sτ on [τ1, τ2) is

sτi,j,c,l=(λτλτ1)ilr=0ls1i,j,c,r(irlr)(1λτλτ1)lrfor2ci,0lc2, (15)
sτi,j,c,c1=(λτλτ1)ic+1r=0c1q=r+1ck=qc1bi,j,k,k1s1i,j,q,r(irc1r)(1λτλτ1)c1rfor1ci, (16)
sτi,j,i+1,i=s1i,j,i+1,i+r=0i1q=r+1ik=qibi,j,k,k1s1i,j,q,r(1λτλτ1)ir, (17)

where k=cc-1bi,j,k,k-11. As a direct result, if we take the initial condition sτ1i,j,c,l=p(i,j,c)1(l=0) for (i, j, c, l)∈Γ at τ1 = 0, it follows that

sτi,j,c,l=p(i,j,c)(il)(1-τλ)i-l(τλ)lfor2ci,0lc-2. (18)

Remark 5. We discuss some properties of sτ. Observe that the ODEs are “separable” in that sτi,j,c,l only depends on the entries of sτ and p with the same (i, j). Fix an (i, j) pair, define Γi,j ≔ {(c, l):0 ≤ l < ci or c = i + 1, l = i}. If uτ is a constant vector of 1’s on [τ1, τ2) ⊆ [0, λ), then we can show after some algebra that for any τ ∈ [τ1, τ2),

(c,l)Γi,jsτi,j,c,l=(c,l)Γi,jsτ1i,j,c,l. (19)

If there exists some c0 such that 1 ≤ c0i, uτi,j,c0,c0-1=0, then (c,l)Γi,jsτi,j,c,l<(c,l)Γi,jsτ1i,j,c,l. Since the initial condition is s0i,j,c,l=p(i,j,c)1(l=0) for 1 ≤ ci, it follows that

(c,l)Γi,jsτi,j,c,l1cip(i,j,c). (20)

In the following part our goal is to approximate Rkn and Dkn as n → ∞ given a function u. However, the number of variables depends on n, so we need to bound the terms associated with large in or out degrees. Fix ϵ > 0 and by assumption 1 we have that

λ=i,j,cip(i,j,c)=i,j,cjp(i,j,c)<, (21)

then there exists an integer Mϵ such that

iMϵj,cip(i,j,c)+jMϵi,cjp(i,j,c)<ϵ, (22)

so letting ij = max{i, j}, we have

ijMϵ,cjp(i,j,c)=iMϵj<Mϵcjp(i,j,c)+iMϵjMϵcjp(i,j,c)+i<MϵjMϵcjp(i,j,c)iMϵj<Mϵcip(i,j,c)+iMϵjMϵcjp(i,j,c)+i<MϵjMϵcjp(i,j,c)<ϵ. (23)

We can prove similarly that there exists an integer Lϵ such that ∑ijLϵ,c ip(i, j, c) < ϵ, but without loss of generality we write Mϵ instead of Lϵ in what follows. Moreover, by assumption 1, as n → ∞,

i,j,ciPn(i,j,c)=i,j,cjPn(i,j,c)λ<, (24)

so for n large enough, we can show that

ijMϵ,cjPn(i,j,c)<ϵ,ijMϵ,ciPn(i,j,c)<ϵ. (25)

So we define the integer Mϵ formally.

Definition 5. Given any ϵ > 0, define Mϵ as the integer such that

ijMϵ,cip(i,j,c)<ϵ,ijMϵ,cjp(i,j,c)<ϵ. (26)

Accordingly, define

Γϵ{(i,j,c,l):ij<Mϵ,0l<ciorc=i+1,l=i},Φϵ{(i,j,c,c-1):ij<Mϵ,0ci},λ^λ-ϵ, (27)

where ab = max{a, b}.

Next we show that the scaled state variable Sk and Rk converges in probability to the solution of the ODEs in definition 4 given the function u. The difficulty in proving proposition 3 arises from the fact that the right sides of the ODEs for sτ in definition 4 are discontinuous due to interventions so the auxiliary Wormald’s theorem in Appendix B: Wormald’s theorem is not applicable and needs to be adapted.

Proposition 3. Consider a sequence of networks with initial conditions (Pn)n ≥ 1 satisfying assumption 1 and let (Gn)n≥1 be the sequence of control policies for the contagion process on the sequence of networks and (Gn)n≥1 satisfy assumption 2 with the function u = (uβ)β∈Φϵ, then

sup0knλ^Skαn-sknα=O(n-14),sup0knλ^R˜kn-r˜kn=O(n-14), (28)

with probability 1-O(n14exp(-n14)) and α ∈ Γϵ, where sτ=(sτα)αΓ is the solution for the ODEs in definition 4 with the initial conditions s0i,j,c,l=p(i,j,c)1(l=0) and

R˜0=0,R˜k==1kβΦϵ1(W=β)u-1nβ, (29)

and

r˜τ=0τ(i,j,c,c-1)Φϵ(i-c+1)sti,j,c,c-1λ-tuti,j,c,c-1dt. (30)

From proposition 3 we see that given (Pn)n≥1 and (Gn)n≥1 satisfying assumption 1 and assumption 2, respectively, the scaled stochastic process Skn converges to the deterministic process skn for any k in [0,nλ^]. This justifies assumption 2 on the control policy because given a control policy Gn depending on the function u, we can predict with high probability the scaled stochastic contagion process at any time k.

Next we discuss the convergence of the scaled number of defaults Dkn and the process end time Tnn. Note in definition 6, definition 7 and proposition 4 it is not required that ij < Mϵ

Definition 6. Define Dk- the number of unrevealed out links from the default set at step k.

Recall that Dk denotes the number of defaulted nodes at step k which consist of two parts: the nodes that have defaulted initially ni,j Pn(i, j, 0) and those that are vulnerable initially and default by step k, i.e. ni,j,1ciPn(i,j,c)-(i,j,c,l)ΓSki,j,c,l, thus

Dk=ni,jPn(i,j,0)+ni,j,1ciPn(i,j,c)-(i,j,c,l)ΓSki,j,c,l=ni,j,0ciPn(i,j,c)-(i,j,c,l)ΓSki,j,c,l. (31)

Similarly, among all defaulted nodes at step k the nodes with out degree j consist of two parts: the nodes that have defaulted initially ni Pn(i, j, 0) and those nodes that are vulnerable initially and default by step k, ni,1ciPn(i,j,c)-i,0l<ciorc=i+1,l=iSki,j,c,l, thus

Dk-=jj(niPn(i,j,0)+ni,1ciPn(i,j,c)-i,0l<ciorc=i+1,l=iSki,j,c,l)-k=ni,j,0cijPn(i,j,c)-(i,j,c,l)ΓjSki,j,c,l-k. (32)

Correspondingly we make the following definitions to approximate Dkn and Dk-n as n → ∞.

Definition 7. Define

dτ=i,j,0cip(i,j,c)-(i,j,c,l)Γsτi,j,c,l,dτ-=i,j,0cijp(i,j,c)-(i,j,c,l)Γjsτi,j,c,l-τ. (33)

Proposition 4. Based on definition 6 and definition 7, it follows that

sup0knλ^|Dk-n-dkn-|op(1)+2ϵ,sup0knλ^|Dkn-dkn|op(1)+2ϵ. (34)

To summarize the results we have so far, we have shown in proposition 3 and proposition 4 that the state variable Sk, the accumulative interventions Rk, the number of defaults Dk and the number of unrevealed out-links from the default set Dk- after being scaled by n all converge to a deterministic process which depends on the solution of ODEs in definition 4. This convergence applies to any k before nλ^. By definition 6, Tn=min{0km:Dk-=0}. Additionally define τf=inf{0τλ:dτ-=0}. Next we show that when Tnn converges in probability to τf, then RTnn and DTnn also converge in probability to the corresponding deterministic variables, rτf and dτf, which in light of the boundedness of RTnn and DTnn further implies convergence in expectations, thus the limits in Eq (9) are well defined.

Proposition 5. Consider a sequence of networks with initial conditions (Pn)n≥1 satisfying assumption 1 and let (Gn)n≥1 be the sequence of control policies for the contagion processes on the sequence of networks and (Gn)n≥1 satisfy assumption 2 with the function u. If τf = λ, or τf < λ and ddτdτf-<0, it follows that as n → ∞,

RTn(Gn,Pn)nprτf(u,p),DTn(Gn,Pn)npdτf(u,p). (35)

where

rτf=0τf(i,j,c,c-1)Φ(i-c+1)sti,j,c,c-1λ-tuti,j,c,c-1dt. (36)

Further it follows that as n → ∞,

ERTn(Gn,Pn)nrτf(u,p),EDTn(Gn,Pn)ndτf(u,p). (37)

Under the conditions in proposition 5, the asymptotic control problem Eq (9) becomes

minuΠK·rτf(u,p)+dτf(u,p). (38)

In the following let uτ=(uτβ)βΦ and u = (uτ)τ ∈ [0, λ].

Substituting the expressions of rτf(u,p) and dτf(u,p) in Eqs (36) and (33) respectively into Eq (38) and putting together the system of ordinary differential equations of sτ=(sτα)αΓ, i.e. ddτsτ=h(τ,sτ;uτ) in definition 4 as well as the condition that determines τf, i.e. dτf-=0, we attain the following deterministic optimal control problem.

minu,τfK·rτf(u,p)+dτf(u,p)stddτsτ=h(τ,sτ;uτ)s0i,j,c,l=p(i,j,c)1(l=0)dτf-=0uτβ{0,1}βΦτf[0,λ), (39)

where ddτsτ=h(τ,sτ;uτ) is defined as in definition 4 and

rτf(u,p)=0τfi,j,1ci(i-c+1)sτi,j,c,c-1λ-τuτi,j,c,c-1dτ,dτf(u,p)=i,j,0cip(i,j,c)-(i,j,c,l)Γsτfi,j,c,l,dτf-=i,j,0cijp(i,j,c)-(i,j,c,l)Γjsτfi,j,c,l-τf. (40)

Some difficulties arise because Eq (39) is an infinite dimensional optimal control problem. In light of assumption 1, it suffices to solve a finite dimensional problem to approximate the objective function of the infinite dimensional problem. First we define the finite dimensional optimal control problem.

Definition 8. (FOCP) For ϵ > 0, recall Mϵ as in definition 5. Define the finite dimensional optimal control problem (FOCP) as Eq (39) with the indexes (i, j) restricted to ij < Mϵ.

Remark 6. The restriction of (i, j) to ij < Mϵ indicates that we use only p(i, j, c), ij < Mϵ, 0 ≤ ci in the calculation. It is equivalent to setting p(i, j, c) = 0 for ijMϵ, 0 ≤ ci while keeping p(i, j, c) for ij < Mϵ, 0 ≤ ci unchanged, which implies asymptotically nodes with ijMϵ are all invulnerable. By the solution of the ODE in proposition 2, it implies that sτα=0, for α ∈ Γ∖Γϵ. Note we use tilde sign with the variables to indicate the indexes (i, j) are in the range ij < Mϵ, for example, r˜τf, d˜τf and d˜τf-.

We have the following lemma regarding the objective functions of the infinite and finite dimensional optimal control problems.

Lemma 1. Let ζ(u,τf,p)Krτf(u,p)+dτf(u,p) be the objective function for the infinite dimensional Eq (39) and ζ˜(u,τf,p)Kr˜τf(u,p)+d˜τf(u,p) for (FOCP). Let (u*,τf*) and (u˜,τ˜f) be the optimal solutions for the infinite dimensional Eq (39) and (FOCP), respectively, then for the same p we have that

|ζ˜(u˜,τ˜f,p)-ζ(u*,τf*,p)|<(K+1)ϵ. (41)

By lemma 1 we only need to solve the finite dimensional optimal control problem (FOCP) in definition 8. Because ϵ can be arbitrarily small, we can approximate the objective function of the infinite dimensional problem to any precision. Given p for (FOCP), the Pontryagin’s maximum principle provides the necessary conditions for the optimal control u˜ and end time τ˜f. We can obtain the optimal asymptotic number of interventions r˜τ˜f and fraction of final defaults d˜τ˜f, which lead to the main results of our work. In the next section we focus on solving (FOCP) and suppress the tilde sign for the variables for notational convenience.

Necessary conditions for the optimal control problem

In the following we solve the finite dimensional optimal control problem (FOCP) in definition 8. Throughout this section we understand that the degrees are in the bounded range ij < Mϵ unless specified otherwise. We also suppress the tilde sign for notational convenience. Let t = t(τ) ≔ − ln(λ − τ), t0t(0) = −ln λ and tf ≔ −ln(λ − τf). Note t(τ) is a strictly increasing function of τ and so is the inverse function τ = τ(t). We remark that we assume in the following that τ < λ which implies tf < ∞, but later we can see that the solutions of sτ, uτ and wτ do allow τ = λ. Then we can reformulate the optimal control problem Eq (39) into an autonomous one, i.e. the differential equations of the system dynamics do not depend on time explicitly. Let ut=(utβ)βΦ and u=(ut)tt0 (note previously u = (uτ)τ∈[0,λ]. Additionally, we allow an arbitrary starting time t0 here).

(42)minu,tfK·rtf(u,p)+dtf(u,p)stddtst=h(st;ut)st0i,j,c,l=p(i,j,c)1(l=0)dtf=0utβ{0,1},βΦϵtf[0,),(43)

where ddtst=h(st;ut) denotes the system of differential equations

dsti,j,c,0dt=-isti,j,c,0for1ci,dsti,j,c,ldt=(i-l+1)sti,j,c,l-1-(i-l)sti,j,c,lfor3ci,1lc-2,dsti,j,c,c-1dt=(i-c+2)sti,j,c-1,c-2uti,j,c-1,c-2+(i-c+2)sti,j,c,c-2-(i-c+1)sti,j,c,c-1for2ci,dsti,j,i+1,idt=sti,j,i,i-1uti,j,i,i-1, (44)

and

rtf(u,p)=t0tfi,j,1ci(i-c+1)sti,j,c,c-1uti,j,c,c-1dt, (45)
dtf(u,p)=i,j,0cip(i,j,c)-(i,j,c,l)Γϵstfi,j,c,l, (46)
dtf-=i,j,0cijp(i,j,c)-(i,j,c,l)Γϵjstfi,j,c,l-τf=i,j,0cijp(i,j,c)-(i,j,c,l)Γϵjstfi,j,c,l-λ(1-et0-tf). (47)

Note that Eq (47) follows from 1λ=et0 and thus τf=λ-e-tf=λ(1-1λe-tf)=λ(1-et0-tf).

To determine the necessary conditions for the optimal terminal time tf* and optimal control ut* in Eq (42), we need the Extended Pontryagin Maximum Principle in Appendix C: Extended pontryagin maximum principle. Then we put together the objective function Eq (43) and the necessary conditions to construct the optimization problem Eq (86) we will introduce later.

Applying the Extended Pontryagin Maximum Principle to the optimal control problem Eq (42) yields the following necessary conditions of optimality. Note in order to simplify notations, we suppress the apostrophe “*” in the following. In other words, we use st, ut, wt, v, tf instead of st*,ut*,wt*,v*,tf* to denote the optimal values.

Proposition 6. (Necessary conditions of optimality) Let (st, ut)t∈[t0, tf] be the optimal state and control trajectory of Eq (42) where tf is the optimal terminal time. Define the Hamiltonian function

H(st,ut,wt)=i,j,1ciwti,j,c,0(-isti,j,c,0)+i,j,2ci,1lc-1wti,j,c,l[(i-l+1)sti,j,c,l-1-(i-l)sti,j,c,l]+i,j,2ci+1(K+wti,j,c,c-1)(i-c+2)sti,j,c-1,c-2uti,j,c-1,c-2, (48)

then there exists a piecewise continuously differentiable vector function wt=(wtα)αΓC^1[t0,)Γ and a scalar vR such that the following conditions hold:

  1. The optimal control ut satisfies that ∀t ∈ [t0, tf], 1 ≤ ci, if sti,j,c,c-1>0,
    uti,j,c,c-1={0ifwti,j,c+1,c>-K1ifwti,j,c+1,c<-K0or1ifwti,j,c+1,c=-K, (49)
    if sti,j,c,c-1=0,
    uti,j,c,c-1=0or1. (50)
  2. For 2 ≤ ci, 0 ≤ lc − 2,
    ddtwti,j,c,l=(i-l)(wti,j,c,l-wti,j,c,l+1), (51)
    and for 1 ≤ ci,
    ddtwti,j,c,c-1=(i-c+1)(wti,j,c,c-1-(K+wti,j,c+1,c)uti,j,c,c-1), (52)
    and
    ddtwti,j,i+1,i=0. (53)

    We denote the set of ordinary differential equations for wt as ddtwt=h(wt;ut).

  3. H(st,ut,wt) is a constant for t ∈ [t0, tf].

  4. Transversal conditions
    wtfi,j,c,l=vj-1(i,j,c,l)Γϵ, (54)
    H(stf,utf,wtf)=-ve-tf, (55)
    dtf-=i,j,0cijp(i,j,c)-(i,j,c,l)Γϵjstfi,j,c,l-λ(1-et0-tf)=0. (56)

Remark 7. (Singular control policy) Observe that if the coefficient of uti,j,c,c-1 in the Hamiltonian function H(st,ut,wt) Eq (48), i.e. (i-c+1)(K+wti,j,c+1,c)sti,j,c,c-1 vanishes, uti,j,c,c-1=0 or 1 both satisfy conditions of the Extended Pontryagin Maximum Principle in Appendix C: Extended pontryagin maximum principle i.e. minimizing H(st,ut,wt). In other words, the Extended Pontryagin Maximum Principle cannot determine the optimal control uti,j,c,c-1 in this case. Moreover, since ic + 1 > 0, so if (K+wti,j,c+1,c)sti,j,c,c-1=0 can be sustained over some interval (θ1, θ2) ⊂ [t0, tf], then uti,j,c,c-1 can be 0 or 1 at any time on (θ1, θ2) and switch arbitrarily often between 0 and 1. In the terminology of optimal control theory, the control ut is “singular” on (θ1, θ2) and the corresponding portion of the state trajectory st on (θ1, θ2) is called a singular arc. Further note that (K+wti,j,c+1,c)sti,j,c,c-1=0, t ∈ (θ1, θ2) implies two cases: sti,j,c,c-1=0 or sti,j,c,c-1>0, wti,j,c+1,c=-K, t ∈ (θ1, θ2). We can show that in the first case any feasible control function uti,j,c,c-1 will not affect other entries of st and the second case only occurs when c = i and (θ1, θ2) = (t0, tf).

Solutions of the necessary conditions

Throughout this section we understand that the degrees are in the bounded range ij < Mϵ unless specified otherwise. A well known fact in the control theorists community is that solving for the optimal (st, ut, wt) from the necessary conditions presented in proposition 6 is difficult, especially analytically. In what follows, we solve for the optimal (st, ut, wt) in three steps.

First, solve the the two-point boundary value problem (TPBVP) consisting of the differential equations for st in Eq (42) and wt in condition (2) of proposition 6 where for st the boundary values are given at t = t0 and for wt at t = tf as follows.

ddtst=h(st;ut),st0i,j,c,l=p(i,j,c)1(l=0)(i,j,c,l)Γϵ,ddtwt=h(wt;ut),wtfi,j,c,l=vj1(i,j,c,l)Γϵ, (57)

and additionally the optimal control policy ut satisfies Eq (50) in condition 1 of proposition 6. We solve for (st, ut, wt) in terms of the auxiliary variables (v, tf, ts) and in the following we call these expressions in terms of (v, tf, ts) as the solutions of (st, ut, wt).

Second, because the optimal (st, ut, wt) satisfies the two equations in the necessary conditions of proposition 6, i.e. the Hamiltonian function Eq (55) at t = tf and the terminal condition Eq (56), while minimizing the objective function Eq (43), we have the following optimization problem for (st, ut, wt):

(58)minst,ut,wtK·ittf(u,p)+dtf(u,p)(59)stH(stf,utf,wtf)=vetf(60)dtf=0,(61)

where

rtf(u,p)=t0tfi,j,1ci(i-c+1)sti,j,c,c-1uti,j,c,c-1dt,dtf(u,p)=i,j,0cip(i,j,c)-(i,j,c,l)Γϵstfi,j,c,l,dtf-=i,j,0cijp(i,j,c)-(i,j,c,l)Γϵjstfi,j,c,l-λ(1-et0-tf). (62)

After substituting (st, ut, wt) expressed in terms of (v, tf, ts) into the optimization problem (58), we are able to solve the optimal (v, tf, ts) based on which we can calculate the optimal (st, ut, wt).

Third, we change the variables from (v, tf, ts) to (v, y, z) for further simplification and attain the optimization problem Eq (86) later.

Now we carry out the first step. The main difficulty of solving problem Eq (57) comes from the fact that wt and st depend on ut which depends on wt and st recursively through Eq (50). To disentangle the recursive dependence, the idea is to analyze the properties of st based on which we can either derive the properties of wt or the explicit solutions of wt depending on signs of vj − 1 + K. Then by Eq (50) we attain the optimal control policy ut which leads to the solution of st.

It turns out that we only need wtf as well as ut and st in (58). From problem Eq (57) we know that wtfβ=vj-1 for β ∈ Γϵ. For ut and st, we give out their solutions in the following directly due to the limited space of the paper. The solutions of ut and st can be verified by substituting into problem Eq (57).

Lemma 2. The optimal control policy ut in terms of the variables (v, tf, ts) is given as below.

For 1 ≤ ci except c = i and vj − 1 = −K, ∀t ∈ [t0, tf],

uti,j,c,c-1=1(tti,j,c), (63)

where

ti,j,c={tfifvj-1-Ktf+ln(1+K+vj-1(i-c)K)ifvj-1<-Kand1c<i+K+vj-1K(1-et0-tf)t0otherwise, (64)

If vj − 1 = −K, ∀t ∈ [t0, tf],

uti,j,i,i-1=1(tts)forsomets[t0,tf]. (65)

The following is the solution for st.

Lemma 3. Letting p(i, j, i + 1) = 0, under the optimal control policy in lemma 2, we have the following solutions of st for the two-point boundary value problem (TPBVP).

  1. For 2 ≤ ci, 0 ≤ lc − 2 and c = 1, l = 0,
    sti,j,c,l=p(i,j,c)(il)(et0-t)i-l(1-et0-t)l. (66)
  2. If vj − 1 < −K, consider t ∈ [ti, j, h, ti, j, h−1), for some 1 ≤ hi where
    ti,j,h={tfifh=0tf+ln(1+K+vj-1(i-h)K)if1h<i+K+vj-1K(1-et0-tf)t0otherwise. (67)
    If 1 ≤ c < h,
    sti,j,c,c-1=p(i,j,c)(ic-1)(et0-t)i-c+1(1-et0-t)c-1. (68)
    If hci + 1,
    sti,j,c,c-1=(ic-1)(et0-t)i-c+1m=hcp(i,j,m)n=0m-1(c-1n)(1-et0-ti,j,m)n(et0-ti,j,m-et0-t)c-1-n. (69)
  3. If vj − 1 > −K, for 1 ≤ ci + 1, t ∈ [t0, tf],
    sti,j,c,c-1=p(i,j,c)(ic-1)(et0-t)i-c+1(1-et0-t)c-1. (70)
  4. If vj − 1 = −K, for 1 ≤ ci, t ∈ [t0, tf],
    sti,j,c,c-1=p(i,j,c)(ic-1)(et0-t)i-c+1(1-et0-t)c-1, (71)
    and
    sti,j,i+1,i=p(i,j,i)1{tst)[(1-et0-t)i-(1-et0-ts)i], (72)
    where ts ∈ [t0, tf].

Since Eqs (56) and (55) require the state variable value particularly at t = tf, we can apply lemma 3 at t = tf to attain stf. Next we proceed to the second step, i.e. we substitute (st,ut,wtf) in terms of (v, tf, ts) into the optimization program (58) which leads to the following results.

Proposition 7. Based on the solutions of optimal st in lemma 3 (particularly stf), ut in lemma 2 and wtfα=vj-1, ∀α ∈ Γϵ, letting

ti,j,c={tfifK+vj-10orc=0tf+ln(1+K+vj-1(i-c)K)ifK+vj-1<0and1c<i+K+vj-1K(1-et0-tf),t0otherwise, (73)

the Hamiltonian equation Eq (60) at t = tf is equivalent to

jmax{-K,vj-1}iic=1ip(i,j,c)m=ci(i-1m-1)(et0-tf)i-m+1n=0c-1(m-1n)(1-et0-ti,j,c)n(et0-ti,j,c-et0-tf)m-1-n=vλet0-tf. (74)

The terminal condition Eq (61) is equivalent to

ijj{c=0ip(i,j,c)n=ci(in)(1-et0-ti,j,c)n(et0-ti,j,c)i-n-1(vj-1=-K)p(i,j,i)[(1-et0-tf)i-(1-et0-ts)i]}=λ(1-et0-tf). (75)

And the objective function Eq (59) becomes

K·rtf(u,p)+dtf(u,p)=Kij{c=1ip(i,j,c)m=cin=0c-1(m-c+1)(im)(mn)(et0-tf)i-m(1-et0-ti,j,c)n(et0-ti,j,c-et0-tf)m-n+1(vj-1=-K)p(i,j,i)[(1-et0-tf)i-(1-et0-ts)i]}+ji{c=0ip(i,j,c)n=ci(in)(1-et0-ti,j,c)n(et0-ti,j,c)i-n-1(vj-1=-K)p(i,j,i)[(1-et0-tf)i-(1-et0-ts)i]}. (76)

For the third step, we further simplify the expressions in proposition 7. Define

y1-et0-tf=τfλ,z1-et0-ts,xi,j,c,c-11-et0-ti,j,c={yifK+vj-10orc=01-(1-y)(i-c)K(i-c+1)K+vj-1ifK+vj-1<0and1c<i+K+vj-1Ky0otherwise, (77)

where the first equality follows from t0 = −ln λ and tf = −ln(λ − τf). Because t0tstf and the function 1 − et0t is increasing in t, 0 ≤ zy ≤ 1. As a result, we can substitute the new variables (y, v, z) into the objective function Eq (59), the Hamiltonian equation Eq (60) and the terminal condition Eq (61). Moreover, we add the definition of xi,j,c,c−1 and 0 ≤ zy ≤ 1. Then we obtain a new optimization problem defined as Eq (86) in section Main results. After solving Eq (86) for (y*, v*, z*), we are able to calculate ut* and st* (or uτ* and sτ* after changing the time index) in order to present theorem 2 and theorem 3.

Main results

Contagion process with no interventions

We first present the theorem when no interventions are provided in the contagion process. For ϵ > 0, recall Mϵ is defined as in definition 5 and note that all indexes (i, j) are in the range ij < Mϵ in what follows.

Definition 9. (I function) Define a function I: [0, 1] → [0, 1] as

I(y)1λij<Mϵjc=0ip(i,j,c)P(Bin(i,y)c) (78)

where Bin(i, y) denotes a binomial random variable with i trials and the probability of occurrence y, so P(Bin(i,y)c)=m=ci(im)ym(1-y)i-m. I(y) is constructed to represent the asymptotic scaled total out degree from the default set at scaled time y and attains its form Eq (78) from the solution of a set of differential equations.

Since I(0)=1λij<Mjp(i,j,0)0, and from the definition of λ,

I(1)=1λij<Mϵjc=0ip(i,j,c)1, (79)

and I(y) is continuous and increasing, it has at least one fixed point in [0, 1]. Further define

J(y)ij<Mϵc=0ip(i,j,c)P(Bin(i,y)c). (80)

Theorem 1. (Extends from theorem 3.8 of [24]) Consider a sequence of networks with initial conditions (Pn)n≥1 satisfying assumption 1 where p = (p(i, j, c))i,j,0≤ci such that p(i, j, c) = 0 for ijMϵ, 0 ≤ ci and no interventions are implemented, let y* ∈ [0, 1] be the smallest fixed point of I.

  1. If y* = 1, then asymptotically almost all nodes default during the contagion process, i.e.
    DTnnp1. (81)
  2. If y* < 1 and it is a stable fixed point, i.e. I′(y*) < 1, then asymptotically the fraction of final defaulted nodes
    DTnnpJ(y*). (82)
    Particularly, if I(0) = 0 and I′(0) < 1, then
    DTnnpij<Mϵp(i,j,0). (83)

Remark 8. Theorem 1 states that the stopping time of the default contagion process is fully governed by the smallest fixed point of the function I(y) and the asymptotic fraction of defaulted nodes at the end of the process can be 1, 0 or a fractional number, representing almost all nodes default, almost no nodes default or a partial number of nodes default, respectively.

Contagion process with interventions

We present the theorems for the contagion process with interventions as the result of solving the finite dimensional optimal control problem (FOCP) in definition 8. For ϵ > 0, recall Mϵ is defined as in definition 5. By lemma 1, the optimal objective function value of (FOCP) is within (K + 1)ϵ distance to the optimal objective function value of the infinite dimensional Eq (39). That is why we solved (FOCP) and present the results below with (i, j) in the range ij < Mϵ in the following. From remark 6, for a given vector p = (p(i, j, c))0≤i,0≤j,0≤ci, the restriction of (i, j) to ij < Mϵ indicates that we use only p(i, j, c), ij < Mϵ, 0 ≤ ci in the calculation. It is equivalent to setting p(i, j, c) = 0, ijMϵ, 0 ≤ ci while keeping p(i, j, c), ij < Mϵ, 0 ≤ ci unchanged, which implies asymptotically nodes with ijMϵ are all invulnerable (Note all p(i, j, c) need not to sum up to one).

First we define the optimization problem Eq (86) based on which we present theorem 2 and theorem 3.

Definition 10. (I˜ and J˜ function) Let x = (xβ)β∈Φϵ where xβ = xβ(y, v) and p = (p(i, j, c))i,j,0≤ci. We define the functions I˜(y,v,z) and J˜(y,v,z) as

I˜(y,v,z)=1λij<Mϵj[c=0ip(i,j,c)P(Bin(i,xi,j,c,c-1)c)-1(vj-1=-K)p(i,j,i)(yi-zi)], (84)
J˜(y,v,z)=ij<Mϵ[c=0ip(i,j,c)P(Bin(i,xi,j,c,c-1)c)-1(vj-1=-K)p(i,j,i)(yi-zi)]. (85)

Note we may write I˜(y;v,z) and J˜(y;v,z) to indicate that we treat y as the variable and v, z as the fixed parameters. To present the main results, we define the following optimization problem.

Definition 11. Define the following optimization problem.

(86)miny,v,zK·r˜(y,v,z)+J˜(y,v,z)(87)st(1y)H˜(y,v)=λv(1y)(88)I˜(y,v,z)=y(89)xi,j,c,c1={yifK+vj10orc=01(1y)(ic)K(ic+1)K+vj1ifK+vj1<0and1c<i+K+vj1Ky0otherwise(i,j,c,c1)Φϵ(90)0zy1y,v,z,(91)

where

r˜(y,v,z)=ij<Mϵ{c=1ip(i,j,c)m=cin=0c-1(m-c+1)P(Multin(i,xi,j,c,c-1,y-xi,j,c,c-1,1-y)=(n,m-n,i-m))+1(vj-1=-K)p(i,j,i)(yi-zi)},H˜(y,v)=ij<Mϵmax{-K,vj-1}ic=1ip(i,j,c)[P(Bin(i-1,y)c-1)-P(Bin(i-1,xi,j,c,c-1)c)], (92)

where Bin(i, y) denotes a binomial random variable in i trials with the probability of occurrence y, so P(Bin(i,y)c)=m=ci(im)ym(1-y)i-m and Multin(i, x, y, 1 − xy) = (a, b, iab) denotes a multinomial distribution in i trials with the probabilities of occurrence of each of three types being x, y and 1 − xy, and turns out to have a, b and iab occurrences of each type, respectively, so P(Multin(i,x,y,1-x-y)=(a,b,i-a-b))=(ia,b,i-a-b)xayb(1-x-y)i-a-b.

Remark 9. A feasible solution (y, v, z) has its own meanings for the optimal control problem Eq (39) on the deterministic process (sτ)τ∈[0,λ]: y=τfλ is the scaled end time of the process; v is an intervention indicator in that we should intervene on nodes with out degree j satisfying vj − 1 ≤ −K and v also determines the starting time of the intervention; z specifies the starting time of the intervention for nodes in state (i, j, i, i − 1) when vj − 1 = −K. Moreover, the auxiliary variables (xi,j,c,c−1)(i,j,c,c−1)∈Φϵ specifies the starting time of interventions for nodes with different (i, j) values.

Then we are ready to present the next main theorem about the optimal control policy.

Theorem 2. For the asymptotic control problem Eq (9), then

  1. Consider a sequence of networks with initial conditions (Pn)n≥1 satisfying assumption 1 where p = (p(i, j, c))i,j,0≤ci such that p(i, j, c) = 0 for ijMϵ, 0 ≤ ci.

  2. Let (Gn)n≥1 be the sequence of control policies for the contagion process on the sequence of networks and (Gn)n≥1 satisfy assumption 2.

  3. Let (y*, v*, z*) be the optimal solution for the optimization problem Eq (86).

If y* = 1, or y* ∈ [0, 1) and it is a stable fixed point of I˜(y;v*,z*), i.e. ddyI˜(y*;v*,z*)<1, the optimal control policy Gn*=(g1(n)*,,gm(n)*) satisfies that for 0 ≤ km − 1,

gk+1(n)*(s,w)={1(knλ(x*)i,j,c,c-1)ifw=(i,j,c,c-1)Φϵexceptc=iandv*j-1=-K1(knλz*)ifw=(i,j,i,i-1)andv*j-1=-K0otherwise, (93)

where

(x*)i,j,c,c-1={y*ifK+v*j-10orc=01-(1-y*)(i-c)K(i-c+1)K+v*j-1ifK+v*j-1<0and1c<i+K+v*j-1Ky*0otherwise, (94)

for (i, j, c, c − 1)∈Φϵ.

The next theorem states conclusions for the asymptotic fraction of final defaulted nodes under the optimal policy satisfying theorem 2.

Theorem 3. For the asymptotic control problem Eq 9, then

  1. Consider a sequence of networks with initial conditions (Pn)n≥1 satisfying assumption 1 where p = (p(i, j, c))i,j,0≤ci such that p(i, j, c) = 0 for ijMϵ, 0 ≤ ci.

  2. Let (Gn)n≥1 be the sequence of control policies for the contagion process on the sequence of networks and (Gn)n≥1 satisfy assumption 2.

  3. Let (y*, v*, z*) be the optimal solution for the optimization problem Eq (86).

Then under the optimal control policy Gn* as in theorem 2, we have the following conclusions for the asymptotic fraction of final defaulted nodes.

  1. If y* = 1, then asymptotically almost all nodes default during the default contagion process, i.e.
    DTnnp1. (95)
  2. If y* ∈ [0, 1) and it is a stable fixed point of I˜(y;v*,z*), i.e. ddyI˜(y*;v*,z*)<1, then asymptotically the fraction of final defaulted nodes
    DTnnpJ˜(y*,v*,z*), (96)
    where xi,j,c,c−1 in J˜ is
    xi,j,c,c-1={y*ifK+v*j-10orc=01-(1-y*)(i-c)K(i-c+1)K+v*j-1ifK+v*j-1<0and1c<i+K+v*j-1Ky*0otherwise, (97)
    for (i, j, c, c − 1) ∈ Φϵ. Particularly, if y* = 0 and ddyI˜(0;v*,z*)<1, then
    DTnnpij<Mϵp(i,j,0), (98)
    i.e. the final defaulted nodes only consist of the initially defaulted nodes.

In theorem 3 the first case indicates that the network is highly vulnerable and interventions are costly, then the regulator rather lets the whole network default without implementing any interventions, while in the second case interventions are less expensive or the contagion effect is not as high, it is better for the regulator to implement interventions to counteract the contagion process.

Discussion and summary

The key to solve Eq (86) depends on solving the two equations Eqs (88) and (89). First we claim that the optimal v* for Eq (86) must be nonpositive.

Lemma 4. For Eq (86), the optimal v* ≤ 0.

Here we give an algorithm to solve Eq (86) numerically.

Algorithm 1. Solving Eq (86) numerically.

  1. Assume v = 0,
    1. if K = 1, then solve Eqs (88) and (89) by e.g. Newton’s method, for y and z;
    2. if K ≠ 1, then Eq (86) does not depend on z, so solve for y and let zy arbitrary.
  2. Assume v < 0,
    1. if K = 1, then Eq (86) does not depend on z, so solve for y and v such that v < 0 and let zy arbitrary;
    2. if K ≠ 1,
    3. let y = z and solve Eqs (88) and (89) for y and v such that 0 ≤ y ≤ 1 and v < 0;
    4. if additionally K > 1, let v=1-Kj for j > 0 and solve Eqs (88) and (89) for y and z such that 0 ≤ zy ≤ 1 for each j ∈ {1, …, Mϵ}.
  3. Choose among all the solutions above (if any) the one that minimizes the objective function Eq (87).

Recall that a node is in state (i, j, c, l) if it has in and out degree pair (i, j), the sum of initial equity and accumulative interventions c (called total buffer) and l revealed in-links. Similar to [24], we call the in-links to a node that has only one unit of equity remaining (“distance to default” equal to one) as “contagious” links. So a node in state (i, j, 1, 0) has i contagious links and a node in state (i, j, 2, 1) has i − 1 contagious links and so on, as shown in Fig 5 the states associated with contagious links are colored in blue. The insights from the optimal interventions policy are summarized as follows.

Fig 5. Optimal intervention policy summary where the states indicating one unit of equity remaining are colored in blue.

Fig 5

  1. It is never optimal to intervene on a node if it is not selected or has at least two units of remaining equity when selected. Thus the optimal control policy described in theorem 2 only specifies the optimal intervention decision on a node that, when selected, has one unit of equity remaining, i.e. l = c − 1. In other words, the use of interventions is to counteract the effects of contagious links.

  2. The optimal control policy depends strongly on K, the relative cost of interventions. At a higher K value, interventions are costly and the regulator rather lets the contagion to evolve without any interventions. At a lower K value, the regulator implements more and more interventions, even a “complete” intervention strategy, that is, intervening on every selected node having the “distance to default” of one since the beginning of the process.

  3. The optimal control policy is “monotonic” concerning the number of out degree of a node. There exists a cutoff value of the out degree such that it is only optimal to intervene on a node with out degree larger than this cutoff value and not otherwise, regardless of its in degree, total buffer and revealed in-links. For nodes with out degree equal to the cutoff value, we have the singular control case that only those in state (i, j, i, i − 1) needs interventions and the starting time of interventions is determined by the variable z from the optimization problem Eq (86).

  4. For nodes that are candidates to receive interventions, the starting time of interventions (depends on the variable xi,j,c,c−1) is “monotonic” in terms of the total buffer. The higher the total buffer is, the earlier we should begin to intervene as illustrated in Fig 5 that xi,j,c,c−1 is decreasing in c. Moreover, the starting time is also “monotonic” in terms of the in and out degree. For the same out degree, the smaller the in degree is, the earlier the intervention should begin. If v < 0, the larger the out degree is, the earlier the intervention should begin. The economic meaning is that we should focus on systematically important nodes as well as nodes that are close to invulnerability and thus easier to save. In other words, we achieve our objective by protecting the nodes that would incur large impact on the network after defaulting and by bringing easy-to-save nodes into invulnerable states.

  5. Once we have begun intervening on a node we should keep implementing interventions on it every time it is selected. In other words, we do not allow nodes that have received interventions to default. This is an interesting result. In the partial information setting, [3] consider the interventions on banks that record more write-downs later and default as the “wasted government money” and mention the tradeoff associated with intervention: potentially wasted money versus less capital depletion. But our finding is that there is no wasted money under the optimal policy and the tradeoff is the high intervention cost versus less magnitude of defaults.

Indeed, following the optimal policy, we are able to achieve earlier termination time of the contagion process and smaller fraction of final defaulted nodes. We can quantify the improvement by comparing I˜ and J˜ in theorem 3 with I and J defined in theorem 1, respectively. Note in the following we suppress the apostrophe “*” and the indexes (i, j) are in the range ij < Mϵ.

  1. I˜(y,v,z) plays the same role as I(y) in theorem 1, which represents the asymptotic scaled total out degree from the default set at scaled time y. Since
    I(y)-I˜(y,v,z)=1λij<Mϵj{c=0ip(i,j,c)[P(Bin(i,y)c)-P(Bin(i,xi,j,c,c-1)c)]+1(vj-1=-K)p(i,j,i)(yi-zi)}, (99)
    and note that
    xi,j,c,c-1yfor(i,j,c,c-1)Φϵ, (100)
    thus
    P(Bin(i,y)c)-P(Bin(i,xi,j,c,c-1)c)0. (101)

    Then for the same initial conditions p = (p(i, j, c))ij<Mϵ,0≤ci, the smallest fixed point of I˜(y;v*,z*) is no greater than that of I(y), which implies that the default contagion process under optimal interventions terminates no later than under no interventions.

  2. Similarly J˜(y,v,z) plays the same role as J(y) in theorem 1, which represents the asymptotic fraction of final defaulted nodes under the optimal control policy. The difference
    J(y)-J˜(y,v,z)=ij<Mϵ{c=0ip(i,j,c)[P(Bin(i,y)c)-P(Bin(i,xi,j,c,c-1)c)]+1(vj-1=-K)p(i,j,i)(yi-zi)}0 (102)
    quantifies the fraction of nodes that are prevented from default because of the optimal control policy.

Numerical experiments

Introduction

While the theoretical framework described before allows heterogeneous networks with degree sequences and initial equity levels drawn from arbitrary distributions, we present a relatively simplified case in numerical experiments for illustration purpose, in which the degree sequences and initial equity levels satisfy specific distributions. Consider a sequence of networks with the number of nodes n growing to infinity, whose in and out degrees are between 1 and 10, and each node’s in degree equal to its own out degree, i.e. d(v) = d+(v), v ∈ [n], respectively, so we call either the in or out degree as the degree of the node. This allows us to combine two indexes i and j into one index i, so the state of a node becomes (i, c, l) and the empirical probability Pn and the limiting probability p of the degree and initial equity become Pn(i, c) and p(i, c) respectively. Additionally we assume the initial equity levels between 1 and 10. In sum, we consider the degree and initial equity level pair in the set Γ{(i,c)N02:1i10,0c10}.

Next we decide on the limiting probability p. Note that Γ′ contains three initial types of nodes: defaulted (with c = 0), vulnerable (with ci) and invulnerable (with c > i). In this numerical experiment, we set the total fraction of initial defaults as ξ and assume the fraction of initial defaults is the same across all degrees, i.e. p(i,0)=ξ10 for i ∈ [1, 10]. For the initially liquid nodes, the joint probability of the degree and initial equity conditional on being liquid is constructed through a binormal copula with correlation ρ and two marginal probabilities. The marginal probabilities of the degree and initial equity are assumed to follow the Zipf’s law, i.e.

P(deg=i)=i-(1+a1)l=110l-(1+a1),P(initialequity=c)=c-(1+a2)l=110l-(1+a2), (103)

where a1, a2 > 0. The Zipf’s law is a form of the power law with Pareto tails, which is observed for the distribution of the degrees and equity levels of the financial networks in many empirical studies, see e.g. [19, 29].

In a network of size n with the joint probability Pn(i, c) of the degree and initial equity, a contagion process under interventions occurs as described in section Dynamics. Recall that we only need to consider intervening on a node that, when selected, has only one unit of equity left, i.e. a node with “distance to default” equal to one. Here we consider two types of intervention policies, the optimal policy and the alternative policy: intervening on nodes with degree between 8 and 10 and “distance to default” equal to one from the beginning of the process. The alternative policy implies interventions on nodes with high degrees and close to default, representing the usual policy employed by the central bank or government in a real financial crisis.

Our objective is to verify the convergence in probability of RTnn and DTnn as well as the convergence of the scaled termination time Tnm as stated in proposition 5. Moreover, we shall study the convergence rate of the standard deviation and IQR (interquartile range) to examine if the asymptotic variables provide good approximations under realistic n values.

Under the optimal policy in the form given in theorem 2, the limits for RTnn, DTnn and Tnm as n → ∞ are r˜(y*,v*,z*), J˜(y*,v*,z*) and y*, respectively in Eq (86) where (y*, v*, z*) is the optimal solution. On the other hand, the alternative policy is that for 0 ≤ km − 1,

gk+1(n),alt(s,w)={1(knλxalti,c,c-1)ifw=(i,c,c-1)Φ0otherwise, (104)

where for (i, c, c − 1)∈Φ′,

xalti,c,c-1={0ifi{8,9,10}yotherwise, (105)

and y is the solution of 1λi=010ic=0ip(i,c)P(Bin(i,xalti,c,c-1)c)=y. Then the limits for RTnn, DTnn and Tnm as n → ∞ can be calculated as:

RTnnpi=010c=1ip(i,c)m=cin=0c-1(m-c+1)P(Multin(i,xalti,c,c-1,y-xalti,c,c-1,1-y)=(n,m-n,i-m)),DTnnpi=010c=0ip(i,c)P(Bin(i,xalti,c,c-1)c),Tnmpy, (106)

where P(Bin(i,y)c)=m=ci(im)ym(1-y)i-m and P(Multin(i,x,y,1-x-y)=(a,b,i-a-b))=(ia,b,i-a-b)xayb(1-x-y)i-a-b.

Simulation

The set up

We have the following setup.

  1. A sequence of six networks with increasing number of nodes n ∈ {54, 64, …, 104} and there are 100 runs for each network under either intervention policy.

  2. To determine the asymptotic fraction p(⋅, ⋅) of the degree and initial equity pair (i, c) where (i, c) ∈ Γ′, we set the following parameters.

    1. The fraction of initial defaults ξ = 0.5, indicating half of the nodes have defaulted. As stated before, we assume in this numerical experiment that the fraction of initial defaults is the same across all degrees, thus p(i,0)=ξ10 for i ∈ [1, 10].

    2. The probability of the degree and initial equity for liquid nodes p(i, e), i ∈ {1, …, 10}, e ∈ {1, …, 10} is determined by a binormal coupula with the exponents of the marginal probabilities of the degree and initial equity (a1, a2) = (0.8, 0.7) and the correlation coefficient ρ = 0.9. Note that a smaller a1 indicates larger fraction of nodes with higher degrees, thus higher connectivity and a smaller a2 indicates larger fraction of nodes with higher initial equities, and ρ implies how likely that higher degree nodes have higher initial equities.

  3. After determining the asymptotic fraction p(⋅, ⋅), we construct a sequence of empirical fractions Pn(⋅, ⋅) for each network that converge to p(⋅, ⋅) by
    Pn(i,c)=[np(i,c)]n(i,c)Γ, (107)
    where [⋅] is the round function. In other words, the number of nodes with degree i and initial equity c are [np(i, c)] for a network of n nodes.
  4. We consider two intervention policies described as before.

  5. The relative cost for the interventions K = 0.5.

In the following we suppress Tn in the subscripts and write Rn, Dn and Tm in stead of RTnn, DTnn and Tnm respectively.

Theoretical results

The theoretical limits of Rn, Dn and Tm under the optimal and alternative policies are summarized in Table 1.

Table 1. Theoretical limits of Rn, Dn and Tm under the optimal and alternative policies.
Policies Rn Dn KRn+Dn Tm
Optimal 0.306 0.503 0.657 0.728
Alternative 0.019 0.821 0.830 0.866

We first verify that the objective function KRn+Dn in Eq (86) is indeed less under the optimal policy. Moreover, compared with the alternative policy, the optimal policy intervens more but results in smaller fraction of final defaulted nodes and ends the contagion process earlier.

Simulation results

We show the plots for Rn, Dn and Tm under the optimal and alternative policies as shown in Figs 611.

Fig 6. The boxplot and log-log plot of standard deviation and IQR for R/n under optimal policy.

Fig 6

Fig 11. The boxplot and log-log plot of standard deviation and IQR for T/m under alternative policy.

Fig 11

Fig 7. The boxplot and log-log plot of standard deviation and IQR for D/n under optimal policy.

Fig 7

Fig 8. The boxplot and log-log plot of standard deviation and IQR for T/m under optimal policy.

Fig 8

Fig 9. The boxplot and log-log plot of standard deviation and IQR for R/n under alternative policy.

Fig 9

Fig 10. The boxplot and log-log plot of standard deviation and IQR for D/n under alternative policy.

Fig 10

  1. Under either policy and for each variable, there are four plots in each figure. The first two plots are two boxplots. The above boxplot visualizes five summary statistics (min, mean−standard deviation, mean, mean+ standard deviation, max) while the bottom boxplot uses another set of summary statistics (1st quartile−1.5IQR, 1st quartile, median, 3rd quartile, 3rd quartile+ 1.5IQR) and the data outside the range are treated as outliers, where IQR stands for interquartile range, i.e. the difference between the third and the first quartiles.

  2. The blue dashed horizontal line in each plot indicates the theoretical limits of Rn, Dn and Tm based on p(⋅, ⋅) and the red solid line in each box indicates the theoretical limits of those values calculated with Pn(⋅, ⋅) for each n. We calculate the theoretical values in both ways because for small n, Pn(⋅, ⋅) determined by Eq (107) has a relatively large rounding error and thus deviates a bit from p(⋅, ⋅) but calculating using Pn(⋅, ⋅) instead of p(⋅, ⋅) can effectively remove the deviations in the inputs to the model. However, given p(⋅, ⋅), Pn(⋅, ⋅) is different for different n values, thus the theoretical values of a variable calculated with Pn(⋅, ⋅) are also different for different n’s.

  3. The black dots in the boxplots indicates the results of 100 runs and they are jittered by a random amount left and right to avoid overplotting. From the black dots we can see the distributions of the results. Note that the black dots in the above and bottom boxplots show the same results for each n. They look different because they are jittered left and right by a different random amount.

  4. The last two plots in every figure shows the log-log plot of the standard deviation and IQR of Rn, Dn and Tm against n and a fitted straight line with the slope.

From the simulation results, we make the following conclusions.

  1. From the boxplots of Rn, Dn and Tm under both intervention policies, we observe that the mean or median converge to the calculated theoretical value with shrinking standard deviation or IQR. Because the theoretical value is a constant given the joint probability of degree and initial equity p(⋅, ⋅), convergence of mean to the theoretical value with variance converging to zero is equivalent to convergence in probability, this observation provides evidence for the convergences in probability of Rn, Dn and Tm to their theoretical values, thus proving proposition 5 and theorem 3.

  2. Be comparing the blue dashed line and the red solid line we see that the mean or median is closer to the red solid line, i.e. the theoretical value calculated with Pn(⋅, ⋅) instead of p(⋅, ⋅). This reflects the rounding error caused by Eq (107) in the inputs into the calculation. By using the more accurate fraction we observe that the closeness of the mean or median to the theoretical value does not vary in n although the results of different runs are more and more concentrated around the mean or median as n grows.

  3. The log-log plots of the standard deviation and IQR of each variable with the fitted straight lines further show that both of them decrease with power law tails, i.e. in the form of z = Cxa where C is a constant and a > 0 is the exponent. The absolute value of the slope of the straight line serves as the exponent. It is interesting to observe that the exponents for the standard deviation and IQR of different variables are in the range 0.4 ∼ 0.5 under both intervention policies. This implies that the dispersions of all variables converge to zero at roughly the same rate under both policies.

Conclusion

From the simulation part we can make the following conclusions.

  1. The convergences of Rn, Dn and Tm to their theoretical values (stated in proposition 5) are supported by the simulation results. It is worth noting that the closeness of the mean or median to the theoretical value does not vary for different n after the rounding error in the initial fractions are removed, but the dispersion of the variable shrinks as n grows.

  2. The dispersion of each variable decreases following a power law. The exponents are close to each other under both intervention policies and for all variables, indicating a uniform convergence rate of the dispersions of all the variables under both policies.

Appendix A: Proofs

Proof of proposition 1

We give a proof in words similar to the proof of proposition 3.4 in [2] for a different objective function of optimizing the value of the financial system at the end of the process under some budget constraint. We observe that the objective function Jn depends on the set of defaulted nodes only through its cardinality. Any node will affect the states of other nodes only after it defaults because the set of unrevealed out links of the default set determining the contagion process grows only after a node defaults. And it is possible for a default to occur only when a node has one unit of equity (distance to default equal to one) at the time of being selected. Before that time, the equity only decreases by one every time it is selected. Moreover, there is always a chance to intervene on a node before it defaults. However, if we intervene on a node that is not selected at the current step or has more than one units of remaining equity when selected, it is possible that the node may not be selected in the following steps before the process ends in which case we implemented redundant interventions without reducing the number of defaults.

Proof of proposition 2

Proof. Assume uτ = b ≔ (bβ)β∈Φ a constant vector for τ ∈ [τ1, τ2) ⊆ [0, λ), 0 ≤ τ1 < τ2 < ∞ and bβ ∈ {0, 1}. Note that the ODEs are “separable” in that sτi,j,c,l only depends on the entries of sτ with the same (i, j), so we can only focus on the system of ODEs with the same (i, j). For the same (i, j), define Γi,j ≔ {(c, l):0 ≤ l < ci or c = i + 1, l = i} and suppress (i, j) in the superscripts in definition 4, then we obtain the system of ODEs for τ ∈ [τ1, τ2) with the initial condition sτ1=s1(s1c,l)(c,l)Γi,j.

Letting t = −ln(λ − τ), t1 = −ln(λ − τ1) and t2 = −ln(λ − τ2), we have an autonomous system of ODEs for t ∈ [t1, t2) that

dstc,0dt=-istc,0for1ci,dstc,ldt=(i-l+1)stc,l-1-(i-l)stc,lfor3ci,1lc-2,dstc,c-1dt=(i-c+2)stc-1,c-2bc-1,c-2+(i-c+2)stc,c-2-(i-c+1)stc,c-1for2ci,dsti+1,idt=sti,i-1bi,i-1, (108)

with the initial condition st1=s1.

By induction, we can prove the solution st on [t1, t2) is

stc,l=e(i-l)(t1-t)r=0ls1c,r(i-rl-r)(1-et1-t)l-rfor2ci,0lc-2,stc,c-1=e(i-c+1)(t1-t)r=0c-1q=r+1ck=qc-1bk,k-1s1q,r(i-rc-1-r)(1-et1-t)c-1-rfor1ci,sti+1,i=s1i+1,i+r=0i-1q=r+1ik=qibk,k-1s1q,r(1-et1-t)i-r, (109)

where k=cc-1bk,k-11.

By changing the variable t to τ by t = −ln(λ − τ), Eqs (15), (16) and (17) follow. Let the initial condition be sτ1i,j,c,l=p(i,j,c)1(l=0) for (i, j, c, l) ∈ Γ at τ1 = 0, then Eq (18) follows from Eq (15).

Proof of proposition 3

Proof. For the following proof we need to adapt the Wormald’s theorem in Appendix B: Wormald’s theorem. For notational convenience we suppress the tilde sign for R˜, r˜. Since mnλ as n → ∞, for the given ϵ and λ^=λ-, we can find n0N, such that 0<λ^<mn<λ+0.1 for nn0. Let z = (zα)α∈Γϵ and

U={(τ,z,r)R|Γϵ|+2:-ϵ<τ<λ^,-ϵ<zα<1.1,-ϵ<r<λ+0.1}, (110)

then U contains the closure of

{(0,z,0):P(S0α=zαn,αΓϵ,R0=0)0forsomen}. (111)

Define the stopping time TU=min{1km:(kn,Skn,Rkn)U}.

By definition 1 and definition of Rk, 0Skαn, α ∈ Γϵ and 0 ≤ Rk ≤ (λ + 0.1)n hold ∀k ≥ 0 and nn0. Recall that Sk=(Skα)αΓ and Skn=(Skαn)αΓ. The following conditions hold:

  1. For 0 ≤ k < TU and α ∈ Γϵ,
    |Sk+1α-Skα|1,|Rk+1-Rk|1, (112)

    i.e. ρ1 = 1.

  2. There exists ρ2 = O(n−1) such that for 0 ≤ k < TU and α ∈ Γϵ,
    |E(Sk+1α-SkαFk)-hα(kn,Skn)|ρ2,|E(Rk+1-RkFk)-h0(kn,Skn)|ρ2, (113)
    where h = (hα)α∈Γϵ and h0 are
    hi,j,c,l(t,z)={-izi,j,c,0λ-tif1ci,l=0(i-l+1)zi,j,c,l-1λ-t-(i-l)zi,j,c,lλ-tif3ci,1lc-2(i-c+2)zi,j,c-1,c-2λ-tuti,j,c-1,c-2+(i-c+2)zi,j,c,c-2λ-t-(i-c+1)zi,j,c,c-1λ-tif2cizi,j,i,i-1λ-tuti,j,i,i-1ifc=i,l=i-1,h0(t,z)=(i,j,c,c-1)Φϵ(i-c+1)zi,j,c,c-1λ-tuti,j,c,c-1. (114)
    In particular, Eq (113) follows from
    |(i,j,c,c-1)Φϵ(i-c+1)Ski,j,c,c-1m-kukni,j,c,c-1-(i,j,c,c-1)Φϵ(i-c+1)Ski,j,c,c-1nλ-knukni,j,c,c-1|(i,j,c,c-1)Φϵ|(i-c+1)Ski,j,c,c-1nmn-knukni,j,c,c-1-(i-c+1)Ski,j,c,c-1nλ-knukni,j,c,c-1|=O(n-1). (115)

However, for β ∈ Φϵ, hβ and h0 are not Lipschitz continuous because uτβ can have step changes on [0, λ). So we need to adapt the proof. Note that uτβ is piecewise constant {0, 1} valued function thus hβ(τ, s) and h0(τ, s) are Lipschitz continuous in each interval where u˜=(uβ)βΦ is a constant vector and then we can apply the Wormald’s theorem in Appendix B: Wormald’s theorem. In the following define sτ(τ′, x) as the solution of the ODEs,

ddτsτ=h(τ,sτ), (116)

with initial condition at τ′, sτ = x ≔ (xα)α∈Γϵ.

In what follows define the points where any component of u˜τ has a step change. τlinf{τ>τl-1:uτβhasastepchangeforsomeβΦ}λ^ for l ≥ 1 and τ0 ≔ 0. Also let kl = ⌈l⌉, where ⌈⋅⌉ is the ceiling function. As a result, kl − 1 < lkl. Recall the initial condition s0=(s0α)αΓ with s0i,j,c,l=p(i,j,c)1(l=0). Because every uβ for β ∈ Φϵ has only a finite number of step changes on [0, λ) and Φϵ is a finite set, there are in total a finite number of step changes for all the component functions of u˜ on [0, λ).

Then by the Wormald’s theorem, let ρ=n-14, it follows that

sup0kk1-1Skαn-sknα(0,s0)=O(n-14) (117)

with probability 1-O(n14exp(-n14)), ∀α ∈ Γϵ. Note that we will write “with probability 1-O(n14exp(-n14))” as whp hereinafter.

In particular, we have that

Sk1-1αn-sk1-1nα(0,s0)=O(n-14)whp. (118)

Additionally by the Wormald’s theorem again we have that

supk1kk2-1Skαn-sknα(k1n,Sk1n)=O(n-14)whp. (119)

Note that

|Sk1αn-Sk1-1αn|1nαΓϵ, (120)

and by the Lipschitz continuity of sτα(0,s0) on (0,τ1-),

sk1-1nα(0,s0)-sτ1α(0,s0)=O(n-1). (121)

So by Eqs (118), (120) and (121), we have

|Sk1αn-sτ1α(0,s0)||Sk1αn-Sk1-1αn|+|Sk1-1αn-sk1-1nα(0,s0)|+|sk1-1nα(0,s0)-sτ1α(0,s0)|=n-1+O(n-14)+O(n-1)whp. (122)

Thus we have that

Sk1n-sτ1(0,s0)=O(n-14)+O(n-1)whp. (123)

where ‖η‖ is the norm for ηR|Γ|. We do not specify the norm because all norms in Rl are equivalent, lN. From proposition 2 we see that the partial derivatives of sτα(τ,x) with respect to the initial time τ′ and every entry of x are continuous in τ′ and every entry of x respectively, and are bounded for any τ in a subinterval of [0,λ^) on which u˜ is a constant vector function, i.e.

sτα(τ,x)(τ,x)M1< (124)

where M1 is a constant. Recall that |k1n-τ1|<n-1, so by Eqs (123) and (124), it follows from the fundamentals of calculus (e.g. theorem 9.19 and 9.21 in [31]) that

sτα(k1n,Sk1n)-sτα(τ1,sτ1(0,s0))=sτα(k1n,Sk1n)-sτα(0,s0)=O(n-14)+O(n-1)whp, (125)

for τ ∈ (τ1, τ2). So it follows from Eq (119) that ∀α ∈ Γϵ,

supk1kk2-1Skαn-sknα(0,s0)=O(n-14)whp. (126)

Similarly for Rk, define rτ(τ′, x, y) as the solution of

ddτrτ=h0(τ,sτ), (127)

with the initial condition at τ′, (sτ, rτ) = (x, y). Applying the Wormald’s theorem for Rk and rτ gives that,

sup0kk1-1Rkn-rkn(0,s0,0)=O(n-14)whp,supk1kk2-1Rkn-rkn(k1n,Sk1n,Rk1n)=O(n-14)whp. (128)

In particular,

Rk1-1n-rk1-1n(0,s0,0)=O(n-14)whp, (129)

Further note that

|Rk1n-Rk1-1n|1nαΓϵ, (130)

and by the Lipschitz continuity of rτ(0, s0, 0) on (0,τ1-),

rk1-1n(0,s0,0)-rτ1(0,s0,0)=O(n-1). (131)

So by Eqs (129), (130) and (131) we have

|Rk1n-rτ1(0,s0,0)||Rk1n-Rk1-1n|+|Rk1-1n-rk1-1n(0,s0,0)|+|rk1-1n(0,s0,0)-rτ1(0,s0,0)|=n-1+O(n-14)+O(n-1)whp. (132)

Here we apply the fact we shall prove later that the partial derivatives of rτ(τ′, x, y) with respect to the initial time τ′ and every entry of x and y are continuous in τ′, every entry of x and y respectively, and are bounded for any τ in a subinterval of [0,λ^) on which u˜ is a constant vector function, i.e.

rτ(τ,x,y)(τ,x,y)M2< (133)

for some constant M2. Recall that |k1n-τ1|<n-1, so by Eqs (123), (132) and (133), it follows from the fundamentals of calculus that

rτ(k1n,Sk1n,Rk1n)-rτ(τ1,sτ1(0,s0),rτ1(0,s0,0))=rτ(k1n,Sk1n,Rk1n)-rτ(0,s0,0)=O(n-14)+O(n-1)whp, (134)

for τ ∈ (τ1, τ2). So it follows from Eq (128) that

supk1kk2-1Rkn-rkn(0,s0,0)=O(n-14)whp. (135)

We can repeat the above procedure every time any uτβ has a step change, β ∈ Φϵ and there are only a finite number of step changes in [0, λ). Because sτα1 and rτ ≤ λ, d((sτ,rτ),U)0.1Cn-14, for a sufficiently large constant C. Thus the supremum of τ that (sτ, rτ) can be extended to the boundary of U is λ^, i.e. in Eq (177) of the Wormald’s theorem in Appendix B: Wormald’s theorem,

σ=sup{τ0:d((sτ,rτ),U)Cn-14}=λ^. (136)

So it follows that

sup0knλ^Skαn-sknα(0,s0)=O(n-14)whp,sup0knλ^Rkn-rkn(0,s0,0)=O(n-14)whp. (137)

At last we prove the claim that the partial derivatives of rτ(τ′, x, y) with respect to the initial time τ′ and every entry of x and y are all continuous and bounded as in Eq (133) for any τ in a subinterval of [0,λ^) on which u˜ is a constant vector function b = (bβ)β∈Φϵ. Note first that rτ with initial condition s¯=(sτ,rτ) at τ = τ′ in a subinterval of [0,λ^) on which u˜=b satisfies that

rτ=rτ+ττ(i,j,c,c-1)Φϵ(i-c+1)bi,j,c,c-1λ-ysyi,j,c,c-1(τ,sτ)dy. (138)

We shall prove the boundedness by showing the boundedness of rτs¯ and rττ, seperately. First we take the derivatives of rτ with respect to the initial condition s¯ and obtain

rτs¯=elast+ττ(i,j,c,c-1)Φϵ(i-c+1)bi,j,c,c-1λ-ysyi,j,c,c-1(τ,sτ)s¯dy, (139)

where elast is a vector of zeros except an entry of one at the last. The continuity of every entry in rτs¯ is obvious. For boundedness,

rτs¯1+ττ(i,j,c,c-1)Φϵ(i-c+1)bi,j,c,c-1λ-ysyi,j,c,c-1(τ,sτ)s¯dy. (140)

By Eq (124), syi,j,c,c-1s¯<M1, so rτs¯ is bounded. Next we take the derivative of rτ with respect to the initial time τ′ by the Leibniz integral rule and obtain that

rττ=-(i,j,c,c-1)Φϵ(i-c+1)bi,j,c,c-1λ-τsτi,j,c,c-1(τ,sτ)+ττ(i,j,c,c-1)Φϵ(i-c+1)bi,j,c,c-1λ-ysyi,j,c,c-1(τ,sτ)τdy, (141)

where sτi,j,c,c-1(τ,sτ)=sτi,j,c,c-1. The continuity of rττ follows. By Eq (124), syi,j,c,c-1(τ,sτ)τ is bounded, so rττ is bounded. We have proved Eq (133).

Proof of proposition 4

Proof. For some [τ1, τ2) ⊆ [0, λ) on which uτ is a constant vector function, by remark 5 we have for every fixed (i, j) pair and Γi,j = {(c, l):0 ≤ l < ci or c = i + 1, l = i} that

(c,l)Γi,jsτi,j,c,l1cip(i,j,c), (142)

and thus it follows from Eq (23) that

0ijMϵ0cijp(i,j,c)-(i,j,c,l)Γ\Γϵjsτi,j,c,lijMϵ0cijp(i,j,c)<ϵ. (143)

Similarly because by the definition of Ski,j,c,l for 1 ≤ km, for fixed (i, j) pair, (i, j, c, l) ∈ Γ,

0c,lSki,j,c,ln1ciPn(i,j,c), (144)

thus it follows from Eq (25) that

0ijMϵ0cijPn(i,j,c)-(i,j,c,l)Γ\ΓϵjSki,j,c,lnijMϵ0cijPn(i,j,c)<ϵ. (145)

For any k where 0kλ^, by proposition 3, it follows that

|Dk-n-dkn-|<|ij<Mϵ,0cijPn(i,j,c)-(i,j,c,l)ΓϵjSki,j,c,ln-(ij<Mϵ,0cijp(i,j,c)-(i,j,c,l)Γϵjskni,j,c,l)|+2ϵ=|ij<Mϵ,0cij(Pn(i,j,c)-p(i,j,c))-(i,j,c,l)Γϵj(Ski,j,c,ln-skni,j,c,l)|+2ϵij<Mϵ,0cij|Pn(i,j,c)-p(i,j,c)|+(i,j,c,l)Γϵj|Ski,j,c,ln-skni,j,c,l|+2ϵMϵ|Γϵ|(o(1)+op(1))+2ϵ=op(1)+2ϵ, (146)

and similarly,

|Dkn-dkn|<|ij<Mϵ,0ciPn(i,j,c)-(i,j,c,l)ΓϵSki,j,c,ln-(ij<Mϵ,0cip(i,j,c)-(i,j,c,l)Γϵskni,j,c,l)|+2ϵ=|ij<Mϵ,0ci(Pn(i,j,c)-p(i,j,c))-(i,j,c,l)Γϵ(Ski,j,c,ln-skni,j,c,l)|+2ϵij<Mϵ,0ci|Pn(i,j,c)-p(i,j,c)|+(i,j,c,l)Γϵ|Ski,j,c,ln-skni,j,c,l|+2ϵ|Γϵ|(o(1)+op(1))+2ϵ=op(1)+2ϵ. (147)

Proof of proposition 5

Proof. By Eq (25) for n large enough and 1 ≤ km, we have

1n=1kijMϵ1ci1(W=(i,j,c,c-1))u-1ni,j,c,c-11nijMϵ1ciinPn(i,j,c)ijMϵ,ciPn(i,j,c)<ϵ. (148)

The first inequality holds because the number of times nodes with states in the range ijMϵ, 1 ≤ ci are selected during the process is bounded above by their total in degree. Similarly by Eq (23), for τλ^,

0τijMϵ,1ci(i-c+1)sti,j,c,c-1λ-tuti,j,c,c-1dt0τijMϵ,1ciip(i,j,c)λ-tdtϵ0τ1λ-tdt=ϵlnλλ-τϵlnλϵ=O(ϵ). (149)

For any k where 0knλ^, by proposition 3 it follows that

|Rkn-rkn||R˜kn+1n=1kijMϵ1ci1(W=(i,j,c,c-1))u-1ni,j,c,c-1-(r˜kn+0knijMϵ,1ci(i-c+1)sti,j,c,c-1λ-tuti,j,c,c-1dt)||R˜kn-r˜kn|+O(ϵ)op(1)+O(ϵ), (150)

thus we have that

sup0knλ^|Rkn-rkn|=op(1)+O(ϵ). (151)

If τf = λ, it implies that dτ->0 for τ(0,λ^), then it follows from proposition 4 that Tnn=λ^+O(ϵ)+op(1). Then because at each step there is at most one more node defaulting, DTnn=Dnλ^n+O(ϵ)+op(1) and from proposition 4 again, D[nλ^]n=dλ^+O(ϵ)+op(1). ⌊⋅⌋ denotes the floor function. Further by the continuity of dτ on [0, λ], DTnn=dλ+O(ϵ)+op(1). Similarly, by Eq (151) and the continuity of rτ on [0, λ], we have that RTnn=rλ+O(ϵ)+op(1).

If τf < λ and ddτdτf-<0, by definition 4, sτ is continuous and thus by Eq (33) dτ- is also continuous. So there exists some τ′ > 0 such that dτ-<0 for τ ∈ (τf, τf + τ′) by the continuity of dτ-. Since ϵ is arbitrary, let ϵ be small enough such that infτ(τf,τf+τ)dτ-<-2ϵ and τ^ be the first time dτ- reaches the minimum. Because dτ^-<-2ϵ, then by proposition 4 Dnτ^-n<0 with high probability, so it holds that Tnn=τf+O(ϵ)+op(1). Again by the continuity of dτ and rτ on [0, λ], proposition 4 and Eq (151), DTnn=dτf+O(ϵ)+op(1) and RTnn=rτf+O(ϵ)+op(1).

In both cases we conclude that Eq (35) holds by tending ϵ → 0.

To prove Eq (37), since RTnm(λ+0.1)n for large n and DTnn, RTn(Gn,Pn)n and DTn(Gn,Pn)n are bounded and thus uniformly integrable. For a sequence of uniformly integrable random variables, convergence in probability implies convergence in expectation. Therefore Eq (37) holds.

Proof of lemma 1

Proof. Solve (FOCP) for the optimal (u˜,τ˜f). Note that u˜β=0 for β ∈ Φ∖Φϵ. If there exists some p(i, j, c)>0, ijMϵ, 0 ≤ ci, then by remark 5, at τ˜f by summing over (i, j) pairs satisfying ijMϵ we can show that

ijMϵ,0cip(i,j,c)-(i,j,c,l)Γ\Γϵsτ˜fi,j,c,l0, (152)

and by the definition of τ˜f that

dτ˜f-=ijMϵ,0cijp(i,j,c)-(i,j,c,l)Γ\Γϵjsτ˜fi,j,c,l+ij<Mϵ,0cijp(i,j,c)-(i,j,c,l)Γϵjsτ˜fi,j,c,l-τ˜f=ijMϵ,0cijp(i,j,c)-(i,j,c,l)Γ\Γϵjsτ˜fi,j,c,l0. (153)

Now we construct a function u as uτ=u˜τ, ττ˜f and uτβ=1, τ˜f<τ, β ∈ Φ. Note that under u, there are always interventions after τ˜f, thus by remark 5, for a fixed (i, j) pair with the set Γi,j = {(c, l):0 ≤ l < ci or c = i + 1, l = i}, (c,l)Γi,jsτi,j,c,l will not change, i.e. (c,l)Γi,jsτi,j,c,l=(c,l)Γi,jsτ˜fi,j,c,l for τ>τ˜f. Let τf be the solution of dτf-=0 under u, then it follows that

dτf-d˜τ˜f=i,j,0cip(i,j,c)-(i,j,c,l)Γsτfi,j,c,l-(ij<Mϵ,0cip(i,j,c)-(i,j,c,l)Γϵsτ˜fi,j,c,l)=ijMϵ,0cip(i,j,c)-((i,j,c,l)Γsτ˜fi,j,c,l-(i,j,c,l)Γϵsτ˜fi,j,c,l)=ijMϵ,0cip(i,j,c)-(i,j,c,l)Γ\Γϵsτ˜fi,j,c,lijMϵ,0cip(i,j,c)<ϵ, (154)

and similarly

rτf-r˜τ˜f=ijMϵ,0cijp(i,j,c)-(i,j,c,l)Γ\Γϵjsτ˜fi,j,c,lijMϵ,0cijp(i,j,c)<ϵ. (155)

By the definition of ζ and ζ˜,

ζ(u,τf,p)ζ˜(u˜,τ˜f,p)+(K+1)ϵ. (156)

Recall (u*,τf*) is the optimal solution for the infinite dimensional Eq (39). By remark 6, (FOCP) assumes that the high degree nodes are invulnerable and because the (u˜,τ˜f) is the optimal solution for (FOCP), it provides the lower bound for the optimal objective function of the infinite dimensional Eq (39), i.e.

ζ˜(u˜,τ˜f,p)ζ(u*,τf*,p). (157)

Let the objective function be ζ(u, τf, p) under u, then by the optimality of (u*,τf*), we have that

ζ(u*,τf*,p)ζ(u,τf,p). (158)

In sum, we have that

ζ˜(u˜,τ˜f,p)ζ(u*,τf*,p)ζ(u,τf,p)ζ˜(u˜,τ˜f,p)+(K+1)ϵ. (159)

Thus the conclusion follows.

Proof of proposition 6

Proof. We apply the Extended Pontryagin Maximum Principle (EPMP) in Appendix C: Extended pontryagin maximum principle. First we present the correspondence of a notation A in EPMP and B in our application in the form AB.

tt,t0t0,tftf,(xti)i{1,,nx}(stα)αΓϵ,(uti)i{1,,nu}(utβ)βΦϵ,U{0,1},λ˚w˚,(λti)i{1,,nx}(wtα)αΓϵ,(t,xt,ut)Ki,j,1ci(i-c+1)sti,j,c,c-1uti,j,c,c-1,ϕ(tf,xtf)dtf,ψk(tf,xtf)=0,k=1,,nψdtf-=0. (160)

Let (w˚,wt) be the adjoint variables then w˚=1, since otherwise the necessary conditions of optimality becomes independent of the cost functional in Eq (42). The Hamiltonian function Eq (48) is a direct result of Eq (183). Note that nψ = 1 and

ψ(t,s)=i,j,0cijp(i,j,c)-(i,j,c,l)Γϵjsi,j,c,l-λ(1-et0-t). (161)

Taking partial derivative yields sψ(tf,stf)=(j,j,,j) which has rank 1.

Since the Hamiltonian function is affine in the control variable ut, by condition (1) of EPMP, we attain that, for 1 ≤ ci,

uti,j,c,c-1={0if(K+wti,j,c+1,c)sti,j,c,c-1>01if(K+wti,j,c+1,c)sti,j,c,c-1<00or1if(K+wti,j,c+1,c)sti,j,c,c-1=0. (162)

By distinguishing the two cases sti,j,c,c-1>0 and sti,j,c,c-1=0, we have the equivalent form in Eq (50).

Taking partial derivative of H with regard to s yields the differential equations of wt in condition (2). Note that H is autonomous, then according to condition (3) of EPMP, H(st,ut,wt) is a constant for t ∈ [t0, tf], which is condition (3).

Then define

Ψ(t,s)i,j,0cip(i,j,c)-(i,j,c,l)Γϵsi,j,c,l+v(i,j,0cijp(i,j,c)-(i,j,c,l)Γϵjstfi,j,c,l-λ(1-et0-tf)) (163)

and taking partial derivatives with respect to s and t respectively by condition (4) of EPMP together with the terminal condition Eq (47) leads to condition (4).

Proof of theorem 1

Proof. For the contagion process without intervention, we relate our model to the auxiliary model used in the proof of theorem 3.8 in [24].

Recall that in Dynamics we are given a set of nodes [n] and the degree sequence (d(v), d+(v))v∈[n] as well as the initial equity levels (e(v))v∈[n] and the network is constructed sequentially by matching any out half-link from the default set to a uniformly chosen unconnected in half-link at every step. For each node v we assign each in half-link a number ranging in {1,…,d(v)}. Let ∑v be the set of all permutations of the in half-links of node v ∈ [n], then a permutation π ∈ ∑v specifies the order in which the in half-links are connected.

Because every in half-link of v represents one unit of loan, v will default after e(v) of its in half-links have been connected (or e(v) of its in links have been revealed) for every permutation πΣv. So the default threshold θ(v, π) for node v if the order in which the in half-links are connected is specified by π is θ(v, π) = e(v), ∀πΣv. Further our assumption 1 is equivalent to the assumption 4.1 and 4.2 in [24]. Moreover, under no intervention, the random graph generated in Dynamics conforms to the model defined in definition 5.4 in [24] with in and out degree sequences (d(v), d+(v))v∈[n] and default thresholds (e(v))v∈[n] So by theorem 3.8 in [24] we achieve the conclusions of theorem 1.

Proof of theorem 2

Proof. To simplify the notations we suppress the apostrophe “*”. In lemma 2 we have presented the optimal control policy (ut)t∈[t0,tf] in terms of t,t0,tf,ts,ti,j,c. Recall that in Eq (77) we have the following relations,

y=1-et0-tf,z=1-et0-ts,xi,j,c,c-1=1-et0-ti,j,c={yifK+vj-10orc=01-(1-y)(i-c)K(i-c+1)K+vj-1ifK+vj-1<0and1c<i+K+vj-1Ky0otherwise, (164)

as well as t = −ln(λ − τ), t0 = −ln λ, so we can change the variable from t to τ. Particularly we apply mapping f(t) = 1 − et0t which is strictly increasing in t, then we have

f(t)=τλ,f(t0)=0,f(ti,j,c)=xi,j,c,c-1,f(ts)=z,f(tf)=y. (165)

We replace each variable t, t0, tf, ts, ti,j,c in lemma 2 with its corresponding variable in Eq (165) resulting in the expressions for uτi,j,c,c-1. At last by assumption 2 on the relations between the control policy Gn=(g1(n),,gm(n)) and the function u, we have the conclusion in theorem 2.

Proof of theorem 3

Proof. In proposition 7 we have obtained the expressions for dtf- and dtf with ij < Mϵ in terms of (v, tf, ts), after change of variables to (v, y, z) with y=τfλ we have the following expressions for dτf- and dτf with their relations to I˜(y;v,z) and J˜(y;v,z) in definition 10.

dτf-=ij<Mϵj[c=0ip(i,j,c)P(Bin(i,xi,j,c,c-1)c)-1(vj-1=-K)p(i,j,i)((τfλ)i-zi)]-τf=λ(I˜(τfλ;v,z)-τfλ),dτf=ij<Mϵ[c=0ip(i,j,c)P(Bin(i,xi,j,c,c-1)c)-1(vj-1=-K)p(i,j,i)((τfλ)i-zi)]=J˜(τfλ;v,z). (166)

Suppose (y*, v*, z*) is an optimal solution for the optimization problem Eq (86) and note that y* is the smallest fixed point of I˜(y;v*,z*) and y*=τf*λ.

  1. If y* = 1, then τf*=λ. By the definition of dτf-, it can only occur when ij<Mjc=0ip(i,j,c)=λ and z*=τf*λ=1, thus we have dτf*=dλ=1, then by proposition 5,
    DTnnp1, (167)
    which proves (1) of theorem 3.
  2. If y* < 1 and I˜(y*;v*,z*)<1, then τf*<λ and ddτdτf*-=I˜(τf*λ;v*,z*)-1<0. Again it follows from proposition 5,
    DTnnpdτf*=J˜(y*;v*,z*). (168)
    which proves (2) of theorem 3. This concludes the proof of theorem 3.

It is important to note that the two cases in theorem 3 corresponds to τf*=λ, and τf*<λ, ddτdτf*-<0, respectively. By proposition 5 they guarantees that the limits of ERTn(Gn,Pn)n and EDTn(Gn,Pn)n in Eq 9 as n → ∞ are well defined, which are rτf and dτf, respectively.

Proof of lemma 4

Proof. In the following we suppress “*”. Note first that if v > 0, xi,j,c,c−1 is increasing in j. This implies that xi,j1,c,c−1 < xi,j2,c,c−1 for the two states in Φϵ, (i, j1, c, c − 1) and (i, j2, c, c − 1) where j1 < j2. By theorem 2, this further implies that at some step k such that nλxi,j1,c,c−1knλxi,j2,c,c−1, we should intervene on a node in state (i, j1, c, c − 1) when it is selected at k but not on a node in state (i, j2, c, c − 1). But this control policy is not optimal because both nodes are the same except the out degree and the node in state (i, j2, c, c − 1) is systematically more important.

Appendix B: Wormald’s theorem

The following is from [30]. Let a ≥ 2 be a fixed integer and ((Ytl)1la)t0 denote a sequence of real valued random variables indexed by n with its natural filtration (Ft)t0. Assume that there is a constant C0 > 0 such that YtlC0n for ∀n, t ≥ 0 and 1 ≤ la. Let fl:Ra+1R be functions and URa+1 be some bounded connected open set containing the closure of

{(0,z1,,za):P(Y0l=zln,1la)0forsomen}. (169)

Define the stopping time TU=inf{t1:(tn,Yt1n,,Ytan)U}. Assume the following three conditions are satisfied:

  1. (Boundedness) For some function ρ1 = ρ1(n)≥1 and ∀t < TU and 1 ≤ la,
    |Yt+1l-Ytl|ρ1. (170)
  2. (Trend) For some function ρ2 = ρ2(n) = o(1) and ∀t < TU and 1 ≤ la,
    |E(Yt+1l-YtlFt)-fl(tn,Yt1n,,Ytan)|ρ2. (171)
  3. (Lipschitz continuity) The functions (fl)1≤la are continuous and satisfies a Lipschitz condition on
    U{(t,z1,,za):t0} (172)
    with the same Lipschitz constant for each l.

Then the following holds:

  1. For (0,z^1,,z^a)U the system of differential equations
    dzlds=fl(s,z1,,za),1la (173)
    has a unique solution in U for zl:RR passing through
    z0l=z^l,1la (174)
    and which extends to points arbitrarily close to the boundary of U.
  2. Let ρ > ρ2 and ρ = o(1). For a sufficiently large constant C, with probability 1-O(ρ1ρexp(-nρ3ρ13)), it holds that
    sup0tnσ(Ytln-ztnl)=O(ρ) (175)
    where zsl is the solution in (1) with
    z0l=Y0ln (176)
    and
    σ=σ(n)=sup{s0:d(((zsl)1la),U)Cρ}, (177)
    where d(u, v) = max1≤ij|uivi| for u=(u1,,uj)Rj and v=(v1,,vj)Rj.

Appendix C: Extended pontryagin maximum principle

The following is from [32]. Consider the optimal control problem to minimize the cost functional including a terminal term

J(u,tf)t0tf(t,xt,ut)dt+ϕ(tf,xtf), (178)

with fixed initial time t0 and free terminal time tf, subject to the dynamical system

x˙t=f(t,xt,ut);xt0=x0, (179)

where the vector function xC1^[t0,T]nx represents the state variables characterizing the behavior of the system at any time instant t, and some general terminal constraints

ψk(tf,xtf)=0,k=1,,nψ. (180)

The admissible controls shall be taken in the class of piecewise continuous functions

uU[t0,T]{uC^[t0,T]nu:utUfort0ttf}, (181)

with tf ∈ [t0, T], where T > t0 and the nonempty, possibly closed and nonconvex set U denotes the control region.

Suppose ℓ and f are continuous and have continuous first partial derivatives with respect to (t, x, u) on [t0,T]×Rnx×Rnu, and also ϕ and ψ(ψk)k=1,,nψ are continuous and have continuous first partial derivatives with respect to (t, x) on [t0,T]×Rnx. Suppose that the terminal constraints Eq (180) satisfy the constraint qualification

rank(ψx(tf*,xtf**))=nψ (182)

where ψx(tf*,xtf**) denotes the Jacobian matrix of the partial derivatives of components of ψ with respect to x evaluated at (tf*,xtf**). Define the Hamiltonian function

H(t,x,u,λ˚,λ)=λ˚(t,x,u)+λTf(t,x,u). (183)

Let (u*,tf*)C^[t0,T]nu×[t0,T) denote a minimizer for the problem, and x*C1^[t0,T] the optimal state, then there exists a nx dimensional piecewise continuously differentiable vector function λt* and λ˚*{0,1} ((λ˚*,λt*) are called adjoint variables) and a Lagrange multiplier vector v*Rnψ such that (λ˚*,λt*)0 for every t[t0,tf*] and the following conditions hold:

  1. The function H(t,xt*,w,λ˚*,λt*) attains its minimum on U at w=ut* for every t[t0,tf*], i.e.
    H(t,xt*,w,λ˚*,λt*)H(t,xt*,ut*,λ˚*,λt*),wU. (184)
  2. (xt*,ut*,λ˚*,λt*) verifies the equations
    ddtxt*=f(t,xt*,ut*),ddtλt*=-xH(t,xt*,ut*,λ˚*,λt*) (185)
    at each instant t of continuity of u* and λ˚*{0,1}.
  3. H(t,xt*,ut*,λ˚*,λt*)=H(tf*,xtf**,utf**,λ˚*,λtf**)-ttf*tH(τ,xτ*,uτ*,λ˚,λτ*)dτ. Therefore, if tH=0, i.e. H is autonomous, then H is a constant over time.

  4. (Transversal condition) Define Ψ(t,x)λ˚*ϕ(t,x)+v*Tψ(t,x), then
    λtf**=xΨ(tf*,xtf**),H(tf*,xtf**,utf**,λ˚tf**,λtf**)=-tΨ(tf*,xtf**) (186)
    together with the terminal condition Eq (180) at t=tf*, i.e. ψk(tf*,xtf**)=0 for k = 1, …, nψ.
  5. The optimal control u* may or may not be continuous; in the latter case we have a corner point. In particular, the conditions that must hold at any corner point θ[t0,tf*] are
    xθ-*=xθ+*,λθ-*=λθ+*,H(θ-,xθ*,uθ-*,λ˚*,λθ*)=H(θ+,xθ*,uθ+*,λ˚*,λθ*). (187)

Proof. See theorem 3.33 and theorem 3.34 in [32].

Appendix D: Preliminary list of notations

Dk: the default set at step k;

Dk=Dk;

DTn: the number of defaulted nodes by the end of the process Tn;

Dk-: the number of unrevealed out links from the default set at step k;

En: the set of links in a random network;

Grn: a graph on n nodes;

Gn,m: the set of networks on n nodes with at most m directed links;

(Gk)0km: the filtration on the default contagion process;

Gn=(g1(n),,gm(n)): a control policy for a network of size n;

I(y),J(y), I˜(y,v,z), J˜(y,v,z): special functions defined for theorem 1, theorem 2 and theorem 3;

K: the “cost” of an intervention relative to a defaulted node;

Mϵ: an integer defined based on ϵ;

N0{0,1,2,};

N{1,2,}, the set of positive integers;

P: probability measure on the set Gn,m;

Pn(i, j, c): the empirical probability of the degrees and initial equity levels and Pn = (Pn(i, j, c))i,j,0≤ci;

Qk: the set of hidden out links from the default set at step k;

Rk: the accumulative number of interventions by step k;

RTn: the accumulative number of interventions by the end of the process Tn;

Ski,j,c,l: the number of nodes that are vulnerable initially and in state (i, j, c, l) at step k and Sk(Skα)αΓ;

Tn: the contagion process end time;

U: the set of all (Gk)0km adapted process μ;

(Vk, Wk): a pair of random variables denoting the link from node Vk to node Wk revealed at step k; By abuse of notation, Wk denotes the state of the node selected at k when there is no confusion;

b ≔ (bβ)β∈Φ: a vector of {0, 1} constants;

ckv: the sum of initial equity and accumulative interventions of node v at step k;

d(v),d+(v): the in and out degree of a node v;

dτ-: the asymptotic number of unrevealed out links of the default set at time τ;

dτ: the asymptotic fraction of defaults at time τ;

e(v): the initial equity of a node;

gk(n): the control function at step k for a network of size n;

h: the set of ODEs of sτ and st;

h′: the set of ODEs of wt;

lkv: the number of revealed in links of node v at step k;

m = m(n): the total in or out degree of a network of size n, maybe index variable as well;

n: the number of nodes, may be index variable as well;

[n] = {1, …, n};

uτi,j,c,c-1: a piecewise constant function; uτ=(uτβ)βΦ;

rτ: the asymptotic scaled number of interventions by time τ;

v: the Lagrange multiplier;

(vk, wk): the link from node vk to node wk revealed at step k;

wti,j,c,l: the adjoint variable;

ϵ: a positive real;

λ: the asymptotic mean of in (out) degree;

λ^λ-;

μk: intervention at step k and μ = (μk)1≤km;

μn: in Introduction denotes an intervention sequence for a network of n nodes;

τf: the end time in of the asymptotic process;

‖‖: any norm of Rl for some lN;

⌈⋅⌉: the ceiling function;

[⋅]: the round function;

˜: the corresponding variable resulted by the constraint ij < Mϵ, e.g. R˜k, u˜, τ˜f, r˜τ, d˜τ;

Bin(i, y): a binomial random variable in i trials with the probability of occurrence y;

Multin(i, x, y, 1 − xy) = (a, b, iab): a multinomial distribution in i trials with the probabilities of occurrence of each of three types being x, y and 1 − xy, and turns out to have a, b and iab occurrences of each type.

Important sets:

Φ{(i,j,c,c-1):0i,0j,1ci},Φϵ{(i,j,c,c-1):ij<Mϵ,1ci}.Γ+{(i,j,c,l):0i,0j,0c,0li},Γ{(i,j,c,l):0i,0j,0l<ciorc=i+1,l=i},Γϵ{(i,j,c,l):ij<Mϵ,0l<ciorc=i+1,l=i},Γi,j{(c,l):0l<ciorc=i+1,l=i}. (188)

Supporting information

S1 Dataset. Dataset generated from the numerical experiments and based on which Figs 611 are presented.

The spreadsheet contains two tables which includes respectively, under the optimal and alternative intervention policies the scaled number of interventions RTnn, the scaled number of defaults DTnn, the scaled process end time Tnm and the objective function value for the number of nodes n = 54, 64, …, 104 and there are 100 runs for each n. The dataset is publicly available via https://figshare.com/articles/Simulation_result/7477562 with DOI 10.6084/m9.figshare.7477562.

(XLSX)

Acknowledgments

The author thanks Lisa R Goldberg, Ilan Adler, Mariana Olvera-Cravioto and Alexander D Shkolnik for their valuable comments and suggestions.

Data Availability

All relevant data are within the paper and its Supporting Information files.

Funding Statement

The author received no specific funding for this work.

References

  • 1. Vollmer U. In: Backhaus J, editor. Lender of Last Resort. New York, NY: Springer; 2014. p. 1–6. [Google Scholar]
  • 2. Amini H, Minca A, Sulem A. Control of interbank contagion under partial information. Journal of Financial Mathematics. 2015;6(1):1195–1219. 10.1137/140981538 [DOI] [Google Scholar]
  • 3. Amini H, Minca A, Sulem A. Optimal equity infusions in interbank networks. Journal of Financial Stability. 2017;31:1–17. 10.1016/j.jfs.2017.05.008 [DOI] [Google Scholar]
  • 4. Cont R, Moussa A, Santos EBe. Network Structure and Systemic Risk in Banking Systems. SSRN eLibrary. 2010; p. 327–368. [Google Scholar]
  • 5.Hurd TR. Contagion! The Spread of Systemic Risk in Financial Networks; 2015.
  • 6. Diamond DW, Dybvig PH. Bank Runs, Deposit Insurance, and Liquidity. Journal of Political Economy. 1983;91(3):401–419. 10.1086/261155 [DOI] [Google Scholar]
  • 7. Allen F, Gale D. Financial contagion. Journal of Political Economy. 2000;1(108):1–33. 10.1086/262109 [DOI] [Google Scholar]
  • 8. Freixas X, Parigi B, Rochet JC. Systemic Risk, Interbank Relations and Liquidity Provision by the Central Bank. Journal of Money, Credit and Banking. 2000;3(3):611–638. 10.2307/2601198 [DOI] [Google Scholar]
  • 9. Acharya VV, Yorulmazer T. Cash-in-the-market pricing and optimal resolution of bank failures. Review of Financial Studies. 2008;21(6):2705–2742. 10.1093/rfs/hhm078 [DOI] [Google Scholar]
  • 10. Gorton G, Huang L. Liquidity, efficiency, and bank bailouts. American Economic Review. 2004;94(3):455–483. 10.1257/0002828041464650 [DOI] [Google Scholar]
  • 11. Philippon T, Schnabl P. Efficient Recapitalization. Journal of Finance. 2013;68(1):1–42. 10.1111/j.1540-6261.2012.01793.x [DOI] [Google Scholar]
  • 12. Rogers LCG, Veraart LAM. Failure and Rescue in an Interbank Network. Management Science. 2013;59(4):882–898. 10.1287/mnsc.1120.1569 [DOI] [Google Scholar]
  • 13. Leitner Y. Financial networks: Contagion, commitment, and private sector bailouts. Journal of Finance. 2005;60:2925–2953. 10.1111/j.1540-6261.2005.00821.x [DOI] [Google Scholar]
  • 14. Garcia-de Andoain C, Heider F, Hoerova M, Manganelli S. Lending-of-last-resort is as lending-of-last-resort does: Central bank liquidity provision and interbank market functioning in the euro area. Journal of Financial Intermediation. 2016;28:32–47. 10.1016/j.jfi.2016.01.003 [DOI] [Google Scholar]
  • 15. Afonso G, Kovner A, Schoar A. Stressed, not frozen: The federal funds market in the financial crisis. Journal of Finance. 2011;66(4):1109–1139. 10.1111/j.1540-6261.2011.01670.x [DOI] [Google Scholar]
  • 16. Berger AN, Bouwman CHS, Kick T, Schaeck K. Bank liquidity creation following regulatory interventions and capital support. Journal of Financial Intermediation. 2016;26:115–141. 10.1016/j.jfi.2016.01.001 [DOI] [Google Scholar]
  • 17.Carlson MA, Duygan-Bump B, Nelson WR. Why Do We Need Both Liquidity Regulations and a Lender of Last Resort? A Perspective from Federal Reserve Lending During the 2007-09 U.S. Financial Crisis; 2015. Available from: http://www.ssrn.com/abstract=2573767.
  • 18. Furfine CH. Interbank Exposures: Quantifying the Risk of Contagion. Journal of Money, Credit and Banking. 2003;35(1):pp. 111–128. 10.1353/mcb.2003.0004 [DOI] [Google Scholar]
  • 19. Boss M, Elsinger H. An empirical analysis of the network structure of the Austrian interbank market. Financial stability Report. 2004;(7):77–87. [Google Scholar]
  • 20. May RM, Arinaminpathy N. Systemic risk: the dynamics of model banking systems. Journal of the Royal Society, Interface. 2010;7(46):823–838. 10.1098/rsif.2009.0359 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Haldane AG, May RM. Systemic risk in banking ecosystems. Nature. 2011;469:351–355 10.1038/nature09659 [DOI] [PubMed] [Google Scholar]
  • 22. Gai P, Haldane A, Kapadia S. Complexity, concentration and contagion. Journal of Monetary Economics. 2011;58:453–470 10.1016/j.jmoneco.2011.05.005 [DOI] [Google Scholar]
  • 23. Nier E, Yang J, Yorulmazer T, Alentorn A. Network models and financial stability. Journal of Economic Dynamics and Control. 2007;31(6):2033–2060 10.1016/j.jedc.2007.01.014 [DOI] [Google Scholar]
  • 24. Amini H, Cont R, Minca A. Resilience to contagion in financial networks. Mathematical Finance. 2013;00(0):1–37. [Google Scholar]
  • 25. Eisenberg L, Noe TH. Systemic risk in financial systems. Management Science. 2001;47(2):236–249 10.1287/mnsc.47.2.236.9835 [DOI] [Google Scholar]
  • 26. Chinazzi M, Fagiolo G. Systemic Risk, Contagion, and Financial Networks: A Survey. SSRN Electronic Journal. 2013. 10.2139/ssrn.2243504 [DOI] [Google Scholar]
  • 27.van der Hofstad R. Random Graphs and Complex Networks; 2014.
  • 28. Furfine CH. The Microstructure of the Federal Funds Market. Financial Markets, Institutions and Instruments. 1999;8(5):24–44. 10.1111/1468-0416.00031 [DOI] [Google Scholar]
  • 29. Bech ML, Atalay E. The topology of the federal funds market. Physica A: Statistical Mechanics and its Applications. 2010;389(22):5223–5246. 10.1016/j.physa.2010.05.058 [DOI] [Google Scholar]
  • 30.Wormald NC. The Differential Equation Method for Random Network Processes and Greedy Algorithms. In: Lectures on Approximation and Randomized Algorithms; 1999. p. 73–155.
  • 31.Rudin W. Principles of Mathematical Analysis. 3rd ed. McGraw-Hill Education; 1976. Available from: http://onlinelibrary.wiley.com/doi/10.1002/cbdv.200490137/abstract.
  • 32.Chachuat BC. Nonlinear and Dynamic Optimization: From Theory to Practice—IC-32: Spring Term 2009. EPFL; 2009.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

S1 Dataset. Dataset generated from the numerical experiments and based on which Figs 611 are presented.

The spreadsheet contains two tables which includes respectively, under the optimal and alternative intervention policies the scaled number of interventions RTnn, the scaled number of defaults DTnn, the scaled process end time Tnm and the objective function value for the number of nodes n = 54, 64, …, 104 and there are 100 runs for each n. The dataset is publicly available via https://figshare.com/articles/Simulation_result/7477562 with DOI 10.6084/m9.figshare.7477562.

(XLSX)

Data Availability Statement

All relevant data are within the paper and its Supporting Information files.


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES