Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2014 Jan 6;9(1):e84563. doi: 10.1371/journal.pone.0084563

Cascading Failures in Spatially-Embedded Random Networks

Andrea Asztalos 1,2,3,¤, Sameet Sreenivasan 1,2,3,*, Boleslaw K Szymanski 1,2, Gyorgy Korniss 1,3
Editor: Alain Barrat4
PMCID: PMC3882255  PMID: 24400101

Abstract

Cascading failures constitute an important vulnerability of interconnected systems. Here we focus on the study of such failures on networks in which the connectivity of nodes is constrained by geographical distance. Specifically, we use random geometric graphs as representative examples of such spatial networks, and study the properties of cascading failures on them in the presence of distributed flow. The key finding of this study is that the process of cascading failures is non-self-averaging on spatial networks, and thus, aggregate inferences made from analyzing an ensemble of such networks lead to incorrect conclusions when applied to a single network, no matter how large the network is. We demonstrate that this lack of self-averaging disappears with the introduction of a small fraction of long-range links into the network. We simulate the well studied preemptive node removal strategy for cascade mitigation and show that it is largely ineffective in the case of spatial networks. We introduce an altruistic strategy designed to limit the loss of network nodes in the event of a cascade triggering failure and show that it performs better than the preemptive strategy. Finally, we consider a real-world spatial network viz. a European power transmission network and validate that our findings from the study of random geometric graphs are also borne out by simulations of cascading failures on the empirical network.

Introduction

Cascading failures represent a particular vulnerability of networked systems, and their effects have been experienced and documented in several domains such as infrastructure networks [1], financial systems [2], and biological systems [3]. An important feature of real-world networks that has not been incorporated into most studies on cascading failures, with some notable exceptions [4][6], is the fact that they are subject to spatial constraints. In other words, in most real-world networks, which node a given node connects to, or interacts with, is largely determined by the geographic distance between the two nodes. This rather severe constraint has important consequences on the network's behavior, and gives rise to significant differences in the scaling behavior of quantities of interest when compared to spatially unconstrained networks [7].

In the context of cascading failures and strategies for their mitigation, studying the effect of spatial constraints is crucial to providing fundamental insights that are practically applicable. A specific context within which studies of cascading failures have proliferated is that of electrical power transmission systems [4], [5], [8][12]. However, understanding such failures in the more general context of flow bearing networks is just as important, especially when confronted with the imminent rise of technologies like the Internet of Things [13], which essentially consists of everyday physical objects equipped with sensors to communicate with users or other objects within their range.

Motivated by these reasons, we study a model of load-based cascading failures on networks on a particular class of spatially constrained networks – the Random Geometric Graph (RGG) [14], [15] – carrying distributed flows and compare its behavior to that of unconstrained network classes. Closely related earlier and recent works, employing resistor networks, investigated transport efficiency, flow optimization, and vulnerability in complex networks [16][20], and the emergence of traffic gridlocks and congestion in road networks [21], [22].

To validate the insights obtained from these spatially-embedded model networks (RGGs), we also study the same load-based cascading failure process on a real-world network with spatial constraints – the European power transmission network maintained and operated by the Union for the Co-ordination of Transmission of Electricity (UCTE). We find several revealing features that arise from the presence of spatial constraints, the most noticeable being a lack of self-averaging on such networks. This is in stark contrast to the results for unconstrained random networks, and thus points to the potential pitfalls of ignoring spatial constraints when studying cascade mitigation strategies.

Methods

Here, we briefly describe the distributed flow model and cascade model that we utilize in this study. For clarity, we note that we use the term ‘node’ and ‘vertex’ interchangeably in the rest of the paper.

1. Distributed flow

We assume the flow on the network to be both directed and distributed. Specifically, each unit of flow is associated with a source and a sink, and takes advantage of all possible paths between the source and the sink. We adopt a simple model of such flow, by modeling the network as a random resistor network with unit conductances along the links [19], [20]. As each node and edge plays a role in forwarding the current from the source to the target node, each of them experiences a load. For one source-target pair and for unit current flowing between them, the load on an arbitrary edge Inline graphic is the current along that edge: Inline graphic; analogously, the load on an arbitrary node Inline graphic is the net current flowing through that node: Inline graphic. These two loads are related by the expression.

graphic file with name pone.0084563.e027.jpg (1)

For all our studies presented here, we assume that unit current flows between Inline graphic source/target pairs simultaneously. Specifically, we assume that all nodes are simultaneously sources and unit current flows into the network at each source. For each source node, a target is chosen randomly and uniformly from the remaining Inline graphic nodes. Consequently, the load is defined as the superposition of all currents flowing through an arbitrary edge/node, which is identical to the edge/node current-flow betweenness [20], [23], [24]:

graphic file with name pone.0084563.e030.jpg (2)

Currents Inline graphic along the edges due to one source/target pair are obtained by writing down Kirchhoff's law for each node Inline graphic in the network and solving the system of linear equations:

graphic file with name pone.0084563.e033.jpg (3)

where we assumed that Inline graphic units of current flow into the network at a source Inline graphic and leave at a target Inline graphic, and where Inline graphic denotes the adjacency matrix of the network. Rewritten in terms of the weighted network Laplacian Inline graphic, where Inline graphic denotes the degree of node Inline graphic, the system (3) transforms into the matrix equation Inline graphic, where Inline graphic is the unknown column voltage vector, while Inline graphic is the net current flowing into the network at node Inline graphic, which is zero for all nodes with the exception of the source and target nodes. As the network Laplacian Inline graphic is singular, we use spectral decomposition [18], [25] to find the pseudo-inverse Laplacian Inline graphic, defined in the space orthogonal to the zero mode. For example, by choosing the reference potential to be the mean voltage [17], Inline graphic, where Inline graphic, one obtains:

graphic file with name pone.0084563.e049.jpg (4)

for each node Inline graphic. Thus, for Inline graphic units of current and for a given source/target pair, the current flowing through edge Inline graphic is:

graphic file with name pone.0084563.e053.jpg (5)

This relation shows that current along an arbitrary edge is uniquely determined by the network topology. Therefore, this is a fully deterministic model of flow and the only source of randomness in the problem arises in the specific instantiation of the network from its parent ensemble.

Electrical flows when applied to explicitly modeling the power grid have commonly used a DC power flow model [4], [5], [11], [12], [26] wherein links also possess a reactance in addition to resistance. However, as pointed out in [4], the equations for this DC model of power flow bear a close resemblance to that of a simple electrical circuit with the current playing the analogous role of power. Further, Scala et al. [11] have demonstrated that inferences made using a DC power flow model, can still be useful despite neglecting the true AC nature of the power transmission network [27]. We emphasize that although the empirical network on which we validate our results is an electrical grid, our studies are aimed at understanding fundamental aspects of cascades on spatial networks carrying distributed flow, and not towards designing strategies specifically tailored for electrical power transmission systems.

2. Cascade model

We model a cascading failure on a network carrying distributed flow following the seminal model of Motter and Lai [28]. We assume that each node is equipped with a load handling capacity that is proportional to the steady-state load on it when the network is intact. Specifically, the capacity of a node Inline graphic is Inline graphic where Inline graphic plays the role of a tolerance parameter, and Inline graphic is the load on the node for the intact network. If a node on the network fails, i.e. is absent or removed from the network, then the flow undergoes a redistribution, and consequently, so do the loads on the surviving nodes. If the new load on any surviving node exceeds its capacity, i.e. if Inline graphic, then that node also fails which leads to a further redistribution and possibly further failures. This process constitutes the model of a cascading failure that we utilize here.

3. Network models

We briefly outline the network models used in this paper and the methods employed for generating associated ensembles.

Random geometric graphs

A Random Geometric Graph (RGG) of size Inline graphic in 2D is constructed by placing Inline graphic nodes randomly in the unit square with open boundary conditions, and connecting any pair of nodes if the Euclidean distance between them is less than a distance Inline graphic, the connection radius [14], [15]. The average degree of the graph Inline graphic can be controlled by varying Inline graphic since Inline graphic.

Erdös-Rényi graphs

An Erdös-Rényi (ER) graph [29] of size Inline graphic is constructed by connecting every pair of nodes with probability Inline graphic. The average degree of the network can be controlled through Inline graphic since Inline graphic.

Scale-free networks

Scale-free (SF) networks [30] of size Inline graphic and degree-exponent Inline graphic are constructed by first generating a degree sequence by sampling the prescribed power-law distribution Inline graphic that yields a desired average-degree Inline graphic. The network is then constructed using this degree sequence following the Configuration Model [31].

Rewired random geometric graphs

To better understand the role of spatial constraints in the observed characteristics of cascades on spatial networks, we generated rewired RGGs as follows. Starting with the original spatial network, we rewire an arbitrarily chosen end of each link to a randomly chosen node in the network with probability Inline graphic. During this process, we ensure that no self-loops or multiple edges are generated, by rejecting any rewiring step that leads to these undesired features.

4. Empirical network

As a realistic testbed on which to validate our results, we use the UCTE European power transmission network from the year 2002 [32][34], which we will henceforth refer to simply as the UCTE network. This network is spread over a geographic area that comprises 18 countries, and consists of Inline graphic buses which constitute the nodes for our purposes. The average degree of the network is Inline graphic.

Results

1. Load landscapes in RGGs

We begin by analyzing the vertex load distributions in RGGs and comparing them to those in ER graphs with the same average degree, the latter playing the role of a null-model where spatial constraints are absent. Both RGGs and ER graphs are characterized by homogeneous (Poissonian) degree distributions [35]. In addition, RGGs exhibit a high clustering coefficient [15], resulting from the spatial dependance of the connectivity and the transitivity of spatial relationships. Path lengths on RGGs scale with the network size Inline graphic in contrast to the Inline graphic scaling found in ER graphs. Given these differences, we expect that the vertex load distribution for RGGs would also differ significantly from that of ER graphs. Indeed, as shown in Figs. 1A and 1B respectively, the vertex load distribution for RGGs has an exponentially decaying tail with a decay constant Inline graphic, while the distribution for ER graphs is best-fitted by a Gaussian distribution (parameters in caption). For identical average degrees, the mean vertex-load in RGGs, (Inline graphic), is almost six times as large as that for ER graphs. Figure 1C shows the average vertex-load conditioned on the vertex-degree, as a function of the degree. Again, in contrast to the case of ER graphs, the plot for RGGs does not display an unambiguously positive correlation of vertex-load with degree over the entire degree range. The vertex-loads are strongly correlated with degrees up until a value close to the average degree, after which they show a subtle decline. A visualization of the network (Fig. 1D) makes it clear that the nodes with the highest loads do not have degrees anywhere as high as the largest degree in the network.

Figure 1. Vertex load profiles in RGGs and ER networks.

Figure 1

Calculated on (A) random geometric graphs and (B) Erdős-Rényi random grahs, composed of Inline graphic nodes with Inline graphic and averaged over Inline graphic network realizations. (C) Positive correlations are shown in the case of ER graphs, while these correlations seem to disappear in RGGs for degree classes higher than the average degree of the network ensemble. Data were averaged over more than Inline graphic network realizations for networks of Inline graphic and Inline graphic. The fluctuating tail of the red curve originates from the lack of sufficient number of samples in the specific degree classes. The error bars correspond to one standard deviations. (D) A single network realization (Inline graphic, Inline graphic) showing the vertex loads. Note, that the node with the highest connections (blue arrows indicate the 3 highest degree nodes) does not carry the highest load in the network (loads are color coded, and node sizes are proportional to loads).

For a network where connections are spatially constrained, we intuitively expect that a high load on a node is indicative of a high load in its neighborhood. To substantiate this, we investigate the spatial correlation between vertex loads on the network. Specifically, we measure the correlation between vertex loads as a function of the distance separating them, by systematically obtaining all pairs of nodes Inline graphic separated by a distance that lies within Inline graphic, and computing the Pearson correlation coefficient between these pairs:

graphic file with name pone.0084563.e129.jpg (6)

Figure 2A shows the Pearson correlation coefficient between loads at a distance Inline graphic apart from each other. Inline graphic evenly spaced values of Inline graphic were considered within the complete range Inline graphic, with Inline graphic. The resulting picture shows that loads are positively correlated for nodes within a distance Inline graphic where Inline graphic is the connection radius, while just beyond this value the correlation sharply drops and continues to decrease monotonically thereafter, reaching slightly negative values at very large separations. The picture obtained for networks with different average degrees is qualitatively similar, and does not change significantly for rewired RGGs generated using small values of the rewiring parameter. It is worth mentioning that although the spatial correlation captured by the Pearson correlation coefficient indicates vertex loads being correlated only within a short distance, it does not preclude the existence of lower dimensional correlated structures – such as a 1D backbone formed by vertices with high loads [36] -within the network. To conclude this study of the load profiles, we analyze the extreme value scaling of the load distribution with network size Inline graphic, a quantity of significance in determining the effective throughput of the network [37]. As shown in Fig. 2B, the maximum vertex load on RGGs scales as a power law with Inline graphic, with an exponent of Inline graphic. This is a much faster growth in comparison to the scaling,Inline graphic found for ER graphs. Rewiring the links of the RGG with increasing probability Inline graphic, gradually but systematically lowers the loads, and their scaling. (The scaling exponents found for Inline graphic and Inline graphic are Inline graphic and Inline graphic respectively.).

Figure 2. Correlations of vertex loads as a function of distance between vertices and extreme value characteristics of loads.

Figure 2

(A) Load and distance correlations in RGGs and rewired RGGs with rewiring probability Inline graphic (see text). Pearson correlation coefficient as function of distance Inline graphic measured between two arbitrary nodes. Data were averaged over Inline graphic network realizations for networks having Inline graphic nodes. (B) System size dependence of the maximum vertex load in networks with Inline graphic. Data were averaged over Inline graphic network realizations. The parameter Inline graphic corresponds to the rewiring probability for links in the RGG.

2. Cascades of overload failures

Next, we simulate cascading failures on a network triggered either by random or targeted removals of nodes, and quantify the resilience of the network to such failures. The model used (see Methods) is identical to that used in earlier studies [19], [28], [38], and is parametrized by a single tolerance parameter Inline graphic which quantifies the excess load bearing capacity of a node. Following the notation introduced in [28], the resilience of a network is quantified in terms of the fractional size of the surviving largest connected component after the cascade ends: Inline graphic, where Inline graphic is the number of nodes belonging to the largest network component after the cascade and Inline graphic is the undamaged (connected) network size.

Figure 3A shows the probability that a cascade ensues after an initial node removal. As seen, irrespective of the tolerance parameter, cascades triggered by the removal of the node with the highest load in the network leave behind the largest damage when compared with those resulting from removal of the highest degree node or a random node. Figures 3B,C show the fractional size of the surviving giant component Inline graphic as a function of the tolerance parameter in the presence and the absence of a cascade. Once again, the damage done is the worst for the case where the initial node removed is the one with the maximal load, even in the case where no cascade is triggered, suggesting that vertices with the highest loads are those responsible for bridging together distinct connected components and ensuring the structural integrity of the network. Finally, Fig. 3D compares the damage done due to cascading failures on RGGs with the damage in SF and ER networks, all having the same average degree. Clearly, while increasing excess capacity does lead to an increase, on average, of the surviving giant component, the growth is profoundly slower for RGGs than for the spatially unconstrained networks. Henceforth, as we further investigate more detailed characteristics of cascades, we restrict our studies to cascades triggered by the removal of the vertex with the highest load, since the damage done to the network is the most severe in this case.

Figure 3. Cascades triggered by targeted and random removals.

Figure 3

(A) Probability that a single node removal will trigger a cascade as function of the tolerance parameter. (B) The ratio Inline graphic of the size of the largest surviving network component to the initial network size, as function of the tolerance parameter Inline graphic, when the initial failure triggers a cascade. (C) Similar to (B), except for the case where the initial failure does not trigger a cascade. In all (A), (B) and (C) subplots the red curve corresponds to the case when the triggered node is the node with the highest load, the blue curve to the case when the triggered node is the most connected (highest degree) node in the network and the green curve shows the case when the triggered node was chosen randomly. Network parameters are Inline graphic, Inline graphic, while the data was averaged over Inline graphic network realizations. Error bars correspond to the standard error of the mean. (D) Inline graphic as a function of tolerance parameter, unconditioned on whether or not a cascade was triggered for RGGs, SF networks and ER networks. For each point, the data was averaged over Inline graphic network realizations.

As shown above, increased capacity allocation results in a monotonic increase in the average surviving giant component size, where the averaging is done over an ensemble of network realizations. If such a monotonic increase was also obtained for individual network instances, then increased capacity allocation, although only weakly effective, would at least be a justifiable preventative measure against cascades. Figures 4 A, B show the size of the surviving giant component Inline graphic as a function of the tolerance parameter Inline graphic for three individual instances of RGGs, for different respective average degrees. As is clearly seen, the variation in Inline graphic is far from monotonic for a single network instance, and differs significantly across instances. Thus, the trend observed by averaging a macroscopic quantity, Inline graphic, over an ensemble of RGG networks (as was the case in Fig. 3) provides little indication of the true behavior of the same quantity for an individual network instance. This behavior persists even if the network size is increased (not shown). Such lack of self-averaging has been observed previously in fragmentation processes on lattices, to which cascades bear a close resemblance [39]. In contrast, results of cascades on single instances of ER and SF networks, shown in Figs. 4B,C, are consistent with those obtained by averaging Inline graphic over respective network ensembles.

Figure 4. Cascades on single network realizations.

Figure 4

Simulations were performed on networks of size Inline graphic. Fractional size of surviving giant component as a function of Inline graphic for (A),(B) RGGs, (C),(D) ER networks and (E),(F) SF networks.

Presumably, this lack of self-averaging is a feature that results from the embedding of the network in two-dimensions (with no shortcuts). To conclusively validate this argument, we study how the presence of a few spatially unconstrained links affects the surviving giant component size, since the addition of such links has the effect of increasing the underlying dimensionality of the space in which the network is embedded. Specifically, for each link, we rewire with probability Inline graphic one end of the link with a randomly chosen node in the network, without allowing self-loops or multiple edges to form. Similar constructions have been used before in [40][42]. By varying Inline graphic between Inline graphic and Inline graphic, we can interpolate between RGGs and ER graphs, as is confirmed by the results shown in Fig. 5, where both the degree-conditioned average load and the average load undergo a smooth crossover from the results expected for RGGs to those expected for ER graphs. Figure 6 shows that even with as low as Inline graphic of the links of the RGG rewired, the non-monotonicity in Inline graphic versus Inline graphic completely disappears. The interval of Inline graphic within which the crossover takes place contains values larger than the theoretical estimate of Inline graphic [43] at which the small-world crossover occurs, likely a finite-size-effect due to the small system sizes considered here. Thus, from a theoretical network-design point of view, the incorporation of a few long-range links would be a simple step in stabilizing flows and managing cascades, since it results in a more predictable relationship between surviving-component size and excess capacity. However, in practical situations the cost of adding such long-range links could be prohibitive, and therefore may not constitute a feasible solution for controlling the grid.

Figure 5. Effect of rewiring in RGGs.

Figure 5

(A) Transition in network structure from an RGG towards an ER network through the process of rewiring. Multiple rewired versions (different values of Inline graphic) are shown together with the two extreme cases. (B) Average vertex load in RGG, ER and rewired versions of RGG as function of the fraction of rewired links Inline graphic. Network parameters are: Inline graphic, Inline graphic. Data were averaged over Inline graphic network realizations.

Figure 6. Effect of rewiring on cascades in RGGs.

Figure 6

Cascades were triggered by the removal of the highest load. Simulations were performed on networks of size Inline graphic with Inline graphic. As Inline graphic is increased the lack of self-averaging that manifests itself in the form of non-monotonicities in the curves for Inline graphic versus Inline graphic, disappears.

We also studied how length dependent link-conductances affected our results. Specifically, we assumed that Inline graphic for a link connecting nodes Inline graphic and Inline graphic where Inline graphic denotes the Euclidean distance between them, and performed simulations to study the dependence of the surviving giant component size Inline graphic as a function of the tolerance parameter Inline graphic (analogous to results in Fig. 4 A,B), and to investigate the effect of rewiring links to create a few long-range connections in the network (similar to the results in Fig. 6). For both cases, we found no significant quantitative difference in the results for the case where conductances were length-dependent. In particular, the non-self-averaging nature of cascades manifested itself in exactly the same manner as is demonstrated in Figs. S1 and S2.

As a next step in understanding the nature of spatial cascades, we measure the spatial correlations between nodes that fail in successive stages of the cascade. Here, a single stage refers to a round of calculating vertex loads, and removing those nodes whose load exceeds their respective capacity. Figure. 7A shows the average location of failing nodes per stage of the cascade, relative to the node that triggers the cascade. The most significant feature observed here, as well as in the distribution of distance (from the cascade-triggering node) for failing nodes in each cascade stage (Fig. 7B) is the separation between the most likely locations for nodes removed in the first and second stages. In subsequent stages, the distribution of the location of failing nodes gets progressively more uniform. In general, as seen from our simulations, cascades last for only a few stages (the longest found in the systems studied here was Inline graphic stages) with most of the damage occurring by the second stage, and then declining rapidly. The stage-wise distributions in Fig. 7B were obtained by aggregating all nodes removed in a particular stage and belonging to a particular distance bin over Inline graphic distinct cascades, and normalizing them by the total number of nodes removed over the distinct cascades. Thus, declining contribution of later stages is due to a combination of two factors: the reduction in the number of nodes removed during later stages, and the decrease in the probability of the cascade surviving up to that stage. The all-stage distribution was generated in a similar fashion as the stage-wise distribution, but disregarding the stages associated with the nodes.

Figure 7. Location of overloaded nodes.

Figure 7

(A) Position (distance and angle) of failed nodes relative to the initially removed one, here the highest load in the network. Different colors correspond to different iterations of the cascade: blue squares (1st), red squares (2nd), green squares (3rd), light blue triangles (4th), black squares (5th), magenta circles (6th), orange circles (7th), light green squares (8th), yellow triangles (9th). Network parameters are the same as in Fig. 1, while each data point is the averaged location of nodes removed in a given stage over Inline graphic independent network cascades. (B) The distribution of the distance Inline graphic from the cascade-triggering node for nodes that fail in the course of a cascade (See text for details).

Finally in this section, we study the effect of average degree of RGGs on their resilience to cascading failures. Figure 8 compares the fractional size of the largest connected component as a function of Inline graphic for networks with average degree Inline graphic and Inline graphic. Surprisingly, the damage caused by cascading failures is far more severe for the more well connected of the two network ensembles. Although, for other dynamical processes such as epidemic spreading and diffusion it is intuitively obvious that more connections lead to more spread, here we would expect that the presence of more paths between any source-sink pair on a denser network would lead to more effective load balancing, and therefore weaker cascading failures. However, while increasing the average degree does cause loads for each node to be lower and more balanced initially, the excess capacity allocation in proportion to these lower and more uniform loads, makes the network ill-equipped to handle variations in load resulting from the initial node removal. As a result, cascades cause more damage for a denser RGG than a sparser one. In contrast, as is well known, denser RGGs are structurally more resilient to random (non-cascading) failures occurring in the network, since the giant component undergoes a transition in size at Inline graphic [44], [45] where Inline graphic is the critical fraction of randomly removed nodes from a RGG, and Inline graphic is an embedding-dimension dependent constant taking the value Inline graphic [15] for two dimensions.

Figure 8. The effect of average degree upon cascading failures.

Figure 8

Fraction of the largest surviving network component following cascading failures Inline graphic triggered by the removal of a single, randomly chosen node as function of Inline graphic tolerance parameter. The two curves correspond to two ensembles of random geometric graphs, one with Inline graphic (maroon) and one with Inline graphic (green). Data were obtained for RGGs of size Inline graphic, averaged over more than Inline graphic network realizations. The error bars correspond to the standard error of the mean.

3. Cascade mitigation strategies

Next, we study two cascade mitigation strategies and evaluate their effectiveness. We begin by analyzing the preemptive node removal strategy proposed by Motter [38]. Intuitively, this method aims to utilize node removal in such a way that the two competing objectives of reducing the load on the network, and keeping the network connected, are balanced. Specifically, the method involves removing a fraction Inline graphic of the lowest load nodes after the initial node failure. This method was motivated by studies on scale-free networks where the load distribution is heavy-tailed implying that a significant fraction of nodes despite contributing to the total load on the network by acting as sources of current/packets, only frugally participate in the carrying of loads generated by other source-sinks pairs, due to their low betweenness. The load distribution for RGGs however, is comparatively much narrower, and we would therefore expect that preemptive node removal would not yield significant success. The results of investigating the efficacy of preemptive node removal as a cascade mitigation strategy are presented in Fig. 9. As shown in Fig. 9A, the probability of a cascade occurring decreases (with increasing Inline graphic) until it reaches a minimum, and beyond which, it increases again. A similar profile is also observed for the ensemble averaged values of the fractional size of the giant surviving component Inline graphic as a function of Inline graphic. Both plots show however, that even at the optimal Inline graphic, and for as large as Inline graphic additional capacity (i.e. Inline graphic), the gains obtained are weak. Furthermore, as a consequence of the lack of self-averaging, individual network instances show profiles that are highly variable and showing little resemblance to the ensemble averaged results. Three such examples for individual network instances are shown in Fig. 9C. Finally, we study how the throughput in the giant surviving component after a cascade, Inline graphic, compares to the throughput on the original network, Inline graphic. The throughput captures the maximum current that can be injected per source without the network becoming congested. For Inline graphic units of current injected at every source, the network is uncongested if for every node Inline graphic, the inequality, Inline graphic holds. Consequently, for the intact network (indicated by subscript Inline graphic), the throughput is Inline graphic. The throughput can similarly be calculated for the surviving component after a cascade. As shown in Fig. 9D, the throughput after the cascade Inline graphic is larger than the initial throughput for Inline graphic. The increase in throughput is expected since the size of the network is smaller after a cascade, leading to a reduction in loads (due to the Inline graphic dependance in the definition of loads, see Eq. 2) and thereby an increase in the quantity Inline graphic. For the case where Inline graphic, although the ensemble average of the ratio Inline graphic is smaller than one (Inline graphic), in most individual instances the ratio is exactly one. In these cases, the throughput after the cascade is determined by a node whose connectivity before and after the cascade is Inline graphic. Such a dangling end has initial load equal to Inline graphic which remains unchanged after the cascade as well i.e. the reduction in the number of sources and sinks in the system has no effect on its load, unlike for other nodes which have higher connectivity. Therefore the value of Inline graphic after the cascade for such a node often ends up being the highest among all nodes, and by definition results in the throughput after the cascade being identical to that before the cascade i.e. Inline graphic. When Inline graphic, such dangling ends are removed as part of the preemptive node removal process, and all surviving nodes end up experiencing a reduction in load due to the reduced size of the surviving giant component. As a result, the final throughput is higher than the initial throughput, resulting in Inline graphic being greater than one.

Figure 9. Preemptive node removal in RGGs.

Figure 9

(A) Probability that a cascade occurs after removal of the node with highest load, despite a fraction Inline graphic of nodes being preemptively removed immediately after the initial trigger. (B) Fractional size of the largest surviving network component Inline graphic as a function of preemptively removed fraction Inline graphic, when there is a cascade. (C) Fractional size of the largest surviving network component Inline graphic as a function of preemptively removed fraction Inline graphic for a single network instance for different values of the tolerance parameter Inline graphic. (D) The ratio of the throughput (defined in text) of the surviving giant component and the throughput of the original network as a function of the altruist node fraction. The red circle corresponds to the case when there no nodes are preemptively removed. Network parameters are: Inline graphic, Inline graphic, and Inline graphic for results shown in (A),(B) and (D).

In view of the observation that the pre-cascade vertex load distribution in the RGG is not highly skewed, we propose a cascade mitigation strategy where rather than reducing the total load on the network by the making a fraction of nodes “absent” from the network as we did for the preemptive strategy, we assign a random fraction Inline graphic of nodes to be altruists who cease to act as sources in the event of a node failure, but continue conducting flow between other source-sink pairs. Figure 10A shows the drop in the probability of a cascade as a function of the fraction Inline graphic of altruistic nodes. Clearly, the drop is significant in comparison to that achieved by the preemptive node removal strategy. We also show the results of a third strategy which involves all surviving nodes reducing the net current they inject into the network (per sink) to a fraction Inline graphic of its original value. We show the results (in red) for the two values of Inline graphic, and note that the probability of a cascade is approximately the same as that obtained when only a fraction Inline graphic of nodes are fully altruistic (i.e. inject no current into the network). Figures 10B and C show comparative plots of the size of the surviving giant component obtained for each of these strategies, conditioned on whether a cascade occurs or not. In both cases, the altruistic strategy, as well as the overall current reduction strategy, show a significant improvement over the preemptive node removal strategy. Understandably, this improvement comes at the cost of the overall throughput in the network. Figure 10D shows the effective throughput in the surviving component Inline graphic normalized by the initial throughput Inline graphic of the intact network, as a function of the altruist fraction Inline graphic. For a principled comparison of the throughputs before and after the cascade, we define the effective throughput of the surviving giant component as the current per source on the intact network that would yield the same total current as that flowing through the surviving giant component after the cascade. Mathematically, when the number of altruist nodes in the surviving component is Inline graphic, this effective throughput is written as:

graphic file with name pone.0084563.e221.jpg (7)

As seen in Fig. 10D, the ratio Inline graphic decreases as the altruist fraction is increased, thus indicating that the increased surviving fraction comes at the expense of the throughput of the network.

Figure 10. Increasing the resilience of the network by introducing altruist nodes.

Figure 10

(A) Probability that a cascade is triggered for an altruist/preemptively removed fraction Inline graphic. The orange squares indicated the probability of cascade when no nodes (other than the initial cascade-triggering node) are removed, but when the current per source is reduced by Inline graphic (upper square) or Inline graphic (lower square) immediately after the initial node removal. (B) The fractional size of the surviving giant component Inline graphic when a cascade is triggered, as a function of the altruist/preemptively-removed node fraction. Also shown are the results when the current per source is reduced by Inline graphic (upper square) or Inline graphic (lower square) immediately after the initial node removal, which coincide with the Inline graphic and Inline graphic results respectively for altruistic node removal. (C) Similar to (B), but for the cases where a cascade is not triggered. (D) The ratio of the effective throughput (defined in text) of the surviving giant component and the throughput of the original network as a function of the altruist node fraction. The red circle corresponds to the case when there are no altruist nodes. Network parameters for all these plots are: Inline graphic, Inline graphic, and Inline graphic for results shown in (A),(B) and (D).

4. Cascade model on an empirical spatial network: The UCTE network

Thus far, our studies have been confined to a stylized model of a spatial network, viz. the RGG. We now study the outcomes of the same cascading failure model on the UCTE network, several aspects of which, have been studied elsewhere [5], [32], [46]. The network consists of Inline graphic transmission stations, with an average degree Inline graphic, spanning 18 European countries in 2002. The network is disassortative with an assortativity coefficient of Inline graphic, and with a higher average clustering coefficient than an ER graph (Inline graphic). Figure 11 shows several other properties of this network. The load appears to be positively correlated with the degree (Fig. 11A), while the degree and load distributions span a relatively narrow range (Fig. 11B,C respectively), as observed also for RGGs. It is worth noting however, that the variance of loads is significant even for small degree values, which makes it difficult to straightforwardly assess the load bearing responsibility of a node purely from its degree.

Figure 11. Characteristics of the UCTE network.

Figure 11

The network consists of Inline graphic nodes with an average degree of Inline graphic. (A) Scatter of loads as a function of node degree Inline graphic (black squares) and the average load (red squares) as a function of node degree Inline graphic. (B) The load distribution on the intact UCTE network. (C) The degree distribution of the UCTE network. (D) A visualization of the UCTE network with loads indicated using both node size and color.

Figures 12 A,B show the cascades triggered on the UCTE network by the removal of a a single edge and a single node, respectively. The non-monotonicity observed in Inline graphic versus Inline graphic for the model spatial networks is also observed here, thus reinforcing the non-self-averaging nature of spatially constrained networks. In the case of node-removal triggered cascades, removal of the highest-load node results in the worst overall damage, as was also the case for RGGs.

Figure 12. Cascades on the UCTE network.

Figure 12

(A) Cascades triggered by the removal of a single node where the node was chosen using three different criteria i.e. randomly, highest load or highest degree. (B) Cascades triggered by the removal of a single edge where the edge was either chosen randomly or was the one with the highest load. Data obtained for cascade triggered by the random removal of a single node (edge) were averaged over Inline graphic different scenarios.

The visualization panels presented in Fig. 13 provide some intuition on the cause of the observed non-monotonicity in Inline graphic as the tolerance parameter is increased. Figure 13 A shows the landscape of loads on the network before the initiation of a cascade where the size of the node is directly proportional to the load on the node. Figure 13 B shows, the state of the network with tolerance parameter Inline graphic after a cascade initiated by the removal of the highest load, has terminated. Figure 13 C shows a similar picture for the case where the tolerance parameter is higher, (Inline graphic), but where the eventual damage is greater (i.e. Inline graphic is smaller than the value obtained for Fig. 13 B). In this last panel, the network consists of several nodes, indicated in red, that had been removed in the course of the cascade depicted in Fig. 13 B, but are now intact as a consequence of the increased tolerance. However, counter-intuitively, the survival of these nodes result in wider load imbalances, resulting in a larger overall number of failures and a smaller surviving giant component. Thus, to some degree, the nodes shown in red, behave like fuses which if removed in the course of a cascade, end up saving a larger part of the network from failure. Dynamic visualizations of the progression of the cascades resulting in the final states shown in Figs. 13B,C are provided in Movies S1 and S2, respectively. A feature that becomes apparent in these dynamic visualizations is the non-local nature of the progression of the cascade. As pointed out in [4] such non-local progression is commonly observed in real cascade situations, and is a feature which can be reproduced by a more realistic DC power flow model, but not by simpler epidemic or percolation based models. Thus it is worth noting that the model presented in this work, despite being simpler than the DC power flow model used in [4], can nevertheless capture a distinctive attribute of real cascade progression.

Figure 13. UCTE network snapshots before and after cascades.

Figure 13

(A) The intact network with node sizes in proportion to their respective steady-state loads. (B) The network and the loads after a highest-load-removal-triggered cascade has terminated, with the tolerance parameter Inline graphic. (C) The network and the loads after a highest-load-removal-triggered cascade has terminated, with the tolerance parameter Inline graphic. The red nodes here indicate nodes that were removed in the cascade leading to (B), but survived in the cascade leading to (C).

Next, we compare the two cascade mitigation strategies, viz. preemptive node removal and assignment of altruistic nodes, for cascades initiated by highest load removal on the UCTE network. As Fig. 14 A and B show, the altruistic strategy generally results in a larger surviving giant component after the cascade, than in the case when preemptive node removal is employed. It is also worth noting that non-monotonicities due to the lack of self-averaging in the cascade process, manifest themselves in these plots as well.

Figure 14. Cascade mitigation on the UCTE network.

Figure 14

Comparison between the preemptive and altruistic node removal strategies on the UCTE network with tolerance parameter (A) Inline graphic and (B) Inline graphic.

We conclude with an investigation of whether, in the case of multiple initial failures, the failures being spatially localized has any effect on the severity of the cascade. Figure 15A shows for a given value of the tolerance parameter Inline graphic, the size of the surviving giant component Inline graphic as a function of the number of nodes removed, for concentrated and random failures on an RGG. Random failures are only marginally more destructive than concentrated ones, which is understandable in light of how the different cascade stages resulting from just a single node's removal can cover a wide spatial spread, as seen in Fig. 7. We arrive at a similar conclusion for the case of concentrated and randomly located failures within the UCTE network from the results shown in Fig. 15B. Dynamic visualizations of the progression of spatially localized and distributed cascades on the UCTE network for the same number of initially removed nodes are provided in Movies S3 and S4, respectively.

Figure 15. Cascades triggered by concentrated versus randomly distributed removals.

Figure 15

(A) Fractional surviving giant component size after a cascade as a function of number of initial nodes removed in concentrated and random removals for RGGs with Inline graphic and Inline graphic. (B) Fractional surviving giant component size after a cascade as a function of number of initial nodes removed in concentrated and random removals for the UCTE network. Each data point is an average over more than Inline graphic realizations.

Discussion

In summary, we have attempted a thorough analysis of the characteristics of cascading failures and strategies for their mitigation on spatially constrained networks, including a model of such networks viz. the random geometric graph, as well as a real-world power transmission network. The key finding worth emphasizing from these studies is the inherent lack of self-averaging for cascade processes on spatial networks. In other words, conclusions gleaned from aggregate statistics on an ensemble of such networks, yield information of little value pertaining to a single network instance. For example, in contrast to the observation for an ensemble of RGGs, for a single network instance, increasing the excess load bearing capacity does not necessarily reduce the severity of the cascade in a monotonic fashion. Thus a straightforward measure for cascade prevention could yield counter-intuitive results. We demonstrate that increasing the effective dimensionality of the system i.e. easing the effect of the spatial constraints by introducing rewired long-range links eliminates these non-intuitive features. A standard cascade mitigation strategy, extensively studied in the past, of preemptively removing a fraction of underperforming nodes does not effectively reduce the severity of cascades on spatially constrained networks, due to the fairly narrow initial range of loads in spatial networks. Instead, the strategy of introducing a fraction of altruistic nodes appears to be a more effective alternative. This holds true both for the model networks as well as for the empirical network. Finally, we also find that cascades resulting from spatially concentrated node failures do not appear to be significantly less destructive than ones that are distributed over the network. Thus, our results paint a complex picture for how failure cascades induced by load redistribution on spatial networks carrying distributed flow propagate through the network. In short, for spatial networks, details specific to a network instance play a very important role in determining strategies to increase the resilience of the network against cascading failures, and methods based on aggregate observations from a network ensemble will present substantial pitfalls.

Note: Data on the UCTE network [32] that we used in this work was obtained from the website http://www.see.ed.ac.uk/~jbialek/Europe_load_flow which is currently non-functional. A processed version of the original data (the UCTE network structure) can be obtained by emailing the corresponding author (SS).

Supporting Information

Figure S1

Cascade realizations on a single RGG of size N  = 1300 where conductances on links are inversely proportional to their lengths. The behavior of the surviving giant component size Inline graphic as a function of the tolerance parameter Inline graphic (three individual realizations are shown) is practically indistinguishable from that found in the case where conductances on all links are identical, shown in Fig. 4A in the main text. All remaining parameters (besides conductances) and simulation details are identical to that in Fig. 4A.

(EPS)

Figure S2

Effect of rewiring links in an RGG with link-length dependent conductances. As the rewiring probability Inline graphic is increased, the non-monotonicities in Inline graphic as a function of tolerance parameter Inline graphic gradually disappear, similarly to the case where link conductances are independent of their length (see Fig. 5). Simulations were performed with Inline graphic and Inline graphic.

(EPS)

Movie S1

Progression of the cascade initiated by the removal of the node with the highest load on the UCTE network ( N  = 1254) with tolerance parameter α  = 0.45. Node sizes are proportional to the load on them. The single orange node at the beginning of the movie indicates the node with the largest node which is removed to trigger a cascade. The overloaded nodes in subsequent stages are shown in orange before they are removed. The total number of nodes removed in the cascade is 167, and the number of nodes in the surviving giant component is 465.

(MP4)

Movie S2

Progression of the cascade initiated by the removal of the node with the highest load on the UCTE network with tolerance parameter α  = 0.45. Although the tolerance parameter is greater than in the case of Movie S1, a greater number of nodes, Inline graphic, fail in the cascade, and the resulting giant component is also smaller, with Inline graphic nodes. The nodes shown in gray indicate those nodes which failed in course of the cascade occurring for Inline graphic (shown in Movie S1), but survived when Inline graphic was increased to Inline graphic. The survival of these nodes potentially plays a role in making the cascade more severe. All other color and node size conventions are identical to those in Movie S1.

(MP4)

Movie S3

Progression of a cascade initiated by a spatially localized removal of 9 nodes. Color and node size conventions are as explained in caption for Movie S1. The tolerance parameter used here is Inline graphic. The number of nodes removed in the course of the cascade is Inline graphic, and the number of nodes in the surviving giant component is Inline graphic.

(MP4)

Movie S4

Progression of a cascade initiated by distributed (random) removal of 9 nodes. Color and node size conventions are as explained in caption for Movie S1. The tolerance parameter used here is Inline graphic. The number of nodes removed in the course of the cascade is Inline graphic (same as for Movie S3), and the number of nodes in the surviving giant component is Inline graphic.

(MP4)

Funding Statement

This work was supported by DTRA Award No. HDTRA1-09-1-0049, by the Army Research Laboratory under Cooperative Agreement Number W911NF-09-2-0053 and by NSF Grant No. DMR-1246958. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Northeast blackout of 2003. Available: https://en.wikipedia.org/wiki/Northeast_blackout_of_2003. Accessed 2013 Jun 17.
  • 2. Sachs JD (2009) Rethink the global money supply. Scientific American 300: 34. [PubMed] [Google Scholar]
  • 3. Anrather CIJ (2011) The immunology of stroke: from mechanisms to translation. Nature Medicine 17: 796–808. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Bernstein A, Bienstock D, Hay D, Uzunoglu M, Zussman G (2011) Power grid vulnerability to geographically correlated failures – analysis and control implications. Technical Report Technical Report No. 2011-05-06, Nov. 2011, Columbia University, Electrical Engineering. arXiv:1206.1099 [cs.SY].
  • 5.Verma T, Ellens W, Kooij R (2013) E Context-independent centrality measures underestimate the vulnerability of power grids. arXiv:1304.5402 [physics.soc-ph].
  • 6. Li W, Bashan A, Buldyrev SV, Stanley HE, Havlin S (2012) Cascading failures in interdependent lattice networks: The critical role of the length of dependency links. Phys Rev Lett 108: 228702. [DOI] [PubMed] [Google Scholar]
  • 7. Barthélemy M (2011) Spatial networks. Physics Reports 499: 1–101. [Google Scholar]
  • 8. Kinney R, Crucitti P, Albert R, Latora V (2005) Modeling cascading failures in the north american power grid. The European Physical Journal B – Condensed Matter and Complex Systems 46: 101–107. [Google Scholar]
  • 9. Dobson I, Carreras BA, Lynch VE, Newman DE (2007) Complex systems analysis of series of blackouts: Cascading failure, critical points, and self-organization. Chaos 17: 026103. [DOI] [PubMed] [Google Scholar]
  • 10. Dorfler F, Chertkov M, Bullo F (2013) Synchronization in complex oscillator networks and smart grids. Proceedings of the National Academy of Sciences 110: 2005–2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Scala A, Pahwa S, Scoglio C (2012) Cascade failures from distributed generation in power grids. arXiv:1209.3733v2.
  • 12.Rahnamay-Naeini M, Wang Z, Ghani N, Mammoli A, Hayat MM (2013) Stochastic Analysis of Cascading-Failure Dynamics in Power Grids. Under review in IEEE Transaction on Power Systems.
  • 13.Zaslavsky A, Perera C, Georgakopoulos D (2013) Sensing as a service and big data. arXiv:1301.0159 [cs.CY].
  • 14.Penrose M (2003) Random Geometric Graphs, volume 5 of Oxford Studies in Probability. Oxford University Press.
  • 15. Dall J, Christensen M (2002) Random geometric graphs. Phys Rev E 66: 016121. [DOI] [PubMed] [Google Scholar]
  • 16. López E, Buldyrev SV, Havlin S, Stanley HE (2005) Anomalous Transport in Scale-Free Networks. Phys Rev Lett 94: 248701. [DOI] [PubMed] [Google Scholar]
  • 17. Korniss G, Hastings MB, Bassler KE, Berryman MJ, Kozma B, et al. (2006) Scaling in small-world resistor networks. Phys Lett A 350: 324–330. [Google Scholar]
  • 18. Korniss G (2007) Synchronization in weighted uncorrelated complex networks in a noisy environment: Optimization and connections with transport efficiency. Phys Rev E 75: 051121. [DOI] [PubMed] [Google Scholar]
  • 19. Asztalos A, Sreenivasan S, Szymanski BK, Korniss G (2012) Distributed flow optimization and cascading effects in weighted complex networks. The European Physical Journal B 85: 288. [Google Scholar]
  • 20.Korniss G, Huang R, Sreenivasan S, Szymanski BK (2012) Optimizing synchronization, flow and robustness in weighted complex networks. in Handbook of Optimization in Complex Networks, edited by Thai MT, Pardalos P, Springer Optimization and Its Applications Vol. 58, Part 1. New York: Springer. 61–96.
  • 21. Mendes GA, da Silva LR, Herrmann HJ (2012) Traffic gridlock on complex networks. Physica A 391: 362–370. [Google Scholar]
  • 22. Colak S, Schneider CM, Wang P, González MC (2013) On the role of spatial dynamics and topology on network ows. New Journal of Physics 15: 113037. [Google Scholar]
  • 23. Newman MEJ (2005) A measure of betweenness centrality based on random walks. Social Networks 27: 39–54. [Google Scholar]
  • 24. Brandes U, Fleischer D (2005) Centrality measures based on current ow. In: Diekert V, Durand B, editors, Lecture Notes in Computer Science, Springer, NY, volume 3404: 533–544. [Google Scholar]
  • 25. Hernandez V, Roman JE, Vidal V (2005) SLEPc: A scalable and exible toolkit for the solution of eigenvalue problems. ACM Trans Math Software 31: 351–362. [Google Scholar]
  • 26. Zimmerman R, Murillo-Sánchez C, Thomas R (2011) Matpower: Steady-state operations, planning, and analysis tools for power systems research and education. Power Systems, IEEE Transactions on 26: 12–19. [Google Scholar]
  • 27. Simonsen I, Buzna L, Peters K, Bornholdt S, Helbing D (2008) Transient dynamics increasing network vulnerability to cascading failures. Phys Rev Lett 100: 218701. [DOI] [PubMed] [Google Scholar]
  • 28. Motter AE, Lai YC (2002) Cascade-based attacks on complex networks. Phys Rev E 66: 065102 (R).. [DOI] [PubMed] [Google Scholar]
  • 29.Bollobás B (1985) Random Graphs. Orlando, FL: Academic Press.
  • 30. Barabási A-L, Albert R (1999) Emergence of scaling in random networks. Science 286: 509–512. [DOI] [PubMed] [Google Scholar]
  • 31. Molloy M, Reed B (1995) A critical point for random graphs with a given degree sequence. Random Struct Algorithms 6: 161–179. [Google Scholar]
  • 32. Zhou Q, Bialek JW (2005) Approximate model of European interconnected system as a benchmark system to study effects of cross-border trades. IEEE Transactions on Power Systems 20: 782–788. [Google Scholar]
  • 33.Zhou Q, Bialek JW (2013) Updated and validated power ow model of the main continental European transmission network. Draft of a paper accepted for IEEE PowerTech 2013.
  • 34.Updated UCTE data (2013) Available: http://www.powerworld.com/bialek Accessed 2013 Nov 15.
  • 35. Herrmann C, Barthélemy M, Provero P (2003) Connectivity distribution of spatial networks. Phys Rev E 68: 026128. [DOI] [PubMed] [Google Scholar]
  • 36. Ercsey-Ravasz M, Toroczkai Z (2010) Centrality scaling in large networks. Phys Rev Lett 105: 038701. [DOI] [PubMed] [Google Scholar]
  • 37. Sreenivasan S, Cohen R, López E, Toroczkai Z, Stanley HE (2007) Structural bottlenecks for communication in networks. Phys Rev E 75: 036105. [DOI] [PubMed] [Google Scholar]
  • 38. Motter AE (2004) Cascade control and defense in complex networks. Phys Rev Lett 93: 098701. [DOI] [PubMed] [Google Scholar]
  • 39. Krapivsky PL, Grosse I, Ben-Naim E (2000) Scale invariance and lack of self-averaging in fragmentation. Phys Rev E 61: R993–R996. [DOI] [PubMed] [Google Scholar]
  • 40. Watts DJ, Strogatz SH (1998) Collective dynamics of ‘small-world’ networks. Nature 393: 440. [DOI] [PubMed] [Google Scholar]
  • 41. Diaz-Guilera A, Gomez-Gardenes J, Moreno Y, Nekovee M (2009) Synchronization in random geometric graphs. Int J Bifurcation and Chaos 19: 687–693. [Google Scholar]
  • 42. Lu Q, Korniss G, Szymanski BK (2008) Naming Games in Two-Dimensional and Small-World- Connected Random Geometric Networks. Phys Rev E 77: 016111. [DOI] [PubMed] [Google Scholar]
  • 43. Newman MEJ, Watts DJ (1999) Scaling and percolation in the small-world network model. Phys Rev E 60: 7332–7342. [DOI] [PubMed] [Google Scholar]
  • 44. Díaz J, Petit J, Serna M (2001) Faulty random geometric networks. Parallel Processing Letters 10: 2000. [Google Scholar]
  • 45. Kong Z, Yeh E (2010) Resilience to degree-dependent and cascading node failures in random geometric networks. Information Theory, IEEE Transactions on 56: 5533–5546. [Google Scholar]
  • 46. Rosas-Casals M, Valverde S, Solé RV (2007) Topological vulnerability of the European power grid under errors and attacks. International Journal of Bifurcation and Chaos 17: 2465–2475. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Figure S1

Cascade realizations on a single RGG of size N  = 1300 where conductances on links are inversely proportional to their lengths. The behavior of the surviving giant component size Inline graphic as a function of the tolerance parameter Inline graphic (three individual realizations are shown) is practically indistinguishable from that found in the case where conductances on all links are identical, shown in Fig. 4A in the main text. All remaining parameters (besides conductances) and simulation details are identical to that in Fig. 4A.

(EPS)

Figure S2

Effect of rewiring links in an RGG with link-length dependent conductances. As the rewiring probability Inline graphic is increased, the non-monotonicities in Inline graphic as a function of tolerance parameter Inline graphic gradually disappear, similarly to the case where link conductances are independent of their length (see Fig. 5). Simulations were performed with Inline graphic and Inline graphic.

(EPS)

Movie S1

Progression of the cascade initiated by the removal of the node with the highest load on the UCTE network ( N  = 1254) with tolerance parameter α  = 0.45. Node sizes are proportional to the load on them. The single orange node at the beginning of the movie indicates the node with the largest node which is removed to trigger a cascade. The overloaded nodes in subsequent stages are shown in orange before they are removed. The total number of nodes removed in the cascade is 167, and the number of nodes in the surviving giant component is 465.

(MP4)

Movie S2

Progression of the cascade initiated by the removal of the node with the highest load on the UCTE network with tolerance parameter α  = 0.45. Although the tolerance parameter is greater than in the case of Movie S1, a greater number of nodes, Inline graphic, fail in the cascade, and the resulting giant component is also smaller, with Inline graphic nodes. The nodes shown in gray indicate those nodes which failed in course of the cascade occurring for Inline graphic (shown in Movie S1), but survived when Inline graphic was increased to Inline graphic. The survival of these nodes potentially plays a role in making the cascade more severe. All other color and node size conventions are identical to those in Movie S1.

(MP4)

Movie S3

Progression of a cascade initiated by a spatially localized removal of 9 nodes. Color and node size conventions are as explained in caption for Movie S1. The tolerance parameter used here is Inline graphic. The number of nodes removed in the course of the cascade is Inline graphic, and the number of nodes in the surviving giant component is Inline graphic.

(MP4)

Movie S4

Progression of a cascade initiated by distributed (random) removal of 9 nodes. Color and node size conventions are as explained in caption for Movie S1. The tolerance parameter used here is Inline graphic. The number of nodes removed in the course of the cascade is Inline graphic (same as for Movie S3), and the number of nodes in the surviving giant component is Inline graphic.

(MP4)


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES