Skip to main content
Proceedings of the National Academy of Sciences of the United States of America logoLink to Proceedings of the National Academy of Sciences of the United States of America
. 2012 Mar 1;109(12):4395-4400. doi: 10.1073/pnas.1105757109

Collaboration in social networks

Luca Dall’Asta a,b,c,1, Matteo Marsili b, Paolo Pin d
PMCID: PMC3311355  PMID: 22383559

Abstract

The very notion of social network implies that linked individuals interact repeatedly with each other. This notion allows them not only to learn successful strategies and adapt to them, but also to condition their own behavior on the behavior of others, in a strategic forward looking manner. Game theory of repeated games shows that these circumstances are conducive to the emergence of collaboration in simple games of two players. We investigate the extension of this concept to the case where players are engaged in a local contribution game and show that rationality and credibility of threats identify a class of Nash equilibria—that we call “collaborative equilibria”—that have a precise interpretation in terms of subgraphs of the social network. For large network games, the number of such equilibria is exponentially large in the number of players. When incentives to defect are small, equilibria are supported by local structures whereas when incentives exceed a threshold they acquire a nonlocal nature, which requires a “critical mass” of more than a given fraction of the players to collaborate. Therefore, when incentives are high, an individual deviation typically causes the collapse of collaboration across the whole system. At the same time, higher incentives to defect typically support equilibria with a higher density of collaborators. The resulting picture conforms with several results in sociology and in the experimental literature on game theory, such as the prevalence of collaboration in denser groups and in the structural hubs of sparse networks.

Keywords: constraint-satisfaction problems, message-passing algorithms, prisoner's dilemma, social networks


The social network influences and constrains in nontrivial ways the behavior of individuals (1) but also contributes to aspects generically referred to as social capital (24), which favor the emergence of coordinated actions or collaboration.

The theoretical investigation about the emergence of collaboration and the fate of repeated actions on networks has, up to now, mostly focused on adaptive learning, imitation, and evolutionary game theory. The corresponding theoretical models (69) seem to suggest that sustaining collaboration is easier on sparse networks and spatial structures than on dense groups. On the contrary, there is no clear experimental evidence of such an effect. Results of recent experiments on voluntary contribution suggest instead that the establishment and maintenance of collaboration is easier on well connected groups than on sparse graphs (1012), in some ways confirming the intuition behind the classical Coleman’s closure argument (13). The picture becomes more complex as we move to analyze individual positions in the network: Experiments performed on star-shaped groups suggest that central nodes are fostered to collaborate more than peripherals (12), corroborating another well known idea; i.e., Burt’s argument on the importance of structural holes (14).

However, the network is not only the channel for the achievement of payoffs or the exchange of information, which are the key ingredients of learning and imitative behavior. The network also supports the establishment of trust, norms, contracts, and other continuative collaborative relations in which the temporal dimension is crucial. In fact, a key aspect of a social network is that individuals connected by a link interact repeatedly with each other, calling into play forward looking strategic behavior typical of the theory of repeated games. This feature provides a formal framework for concepts such as threats, punishment, and credibility, which have been so enlightening on the emergence of collaboration in simple setups (e.g., the prisoner’s dilemma) (15).

Here we investigate the extension of the theory of repeated games to network games. We focus on simple local contribution games which provide a perfect framework to highlight the main conceptual issues, and are closely linked to experimental works on voluntary contribution (1012, 16). Under very general assumption on the structure of the network, we prove that, in a multiplayer setting, rationality is consistent with conditional collaboration that is (i) player-specific and (ii) reciprocal. Interestingly, the fact that collaboration can be sustained through a reciprocal relation of control and punishment is a largely verified experimental fact (16).

We satisfy these requirements with a refinement of Nash equilibria, that we call Collaborative Equilibria, with a precise graph-theoretical interpretation. This refinement turns the problem of finding and characterizing Nash equilibria into that of finding specific subgraphs of the social network, which can be addressed by graph theoretical methods (17, 18). These techniques provide considerable insight on the nature of possible collaborative structures which can be supported, depending on incentives and on the topology of the social network. In particular, we find that when the contribution cost is low, collaboration can be sustained by local commitments (corresponding to dimers or loops on the network). But when costs are high, collaboration requires nonlocal structures which span a finite fraction of the system. This structure has clear implications in terms of systemic fragility, because when costs are high a single individual defection may bring to the collapse of the whole collaborative network.

Our analysis of Collaborative Equilibria on ensembles of random graphs unveils a generic picture which, as will be discussed in the closing section, has many correspondences with well known results in sociology and in experimental economics.

A Repeated Game of Local Contribution

We focus on a class of local contribution games, recently popularized in refs. 16, 19, which have the peculiarity that only the neighbors of a contributor and not the contributor herself benefit of the positive effects of a contribution. In this regard, we can interpret these local contribution games as a simple extension of the celebrated prisoners’ dilemma to the case where there are more than two players, where the range of possible interactions between individuals is limited only to local ones by a fixed network structure, and where agents play the same strategy against each neighbor.

The Model.

Players occupy the nodes N of an undirected social network and they interact with their neighbors in a local contribution game. Specifically, each player iN has the option of either contributing (si = 1) or not (si = 0) to a local public good that has effect only on the neighbors of i in Ni. Contributing is costly, which means that players who contribute incur a cost Xi > 0. At the same time, each player receives a unitary payoff from all agents jNi who contribute in her neighborhood. The payoff function of agent i is then given by

graphic file with name pnas.1105757109eq31.jpg [1]

where s-i stands for the vector of choices of all other players, except i. It is clear that, in the single stage game, a positive si is only a cost for player i, and hence defection (i.e., si = 0) is a dominant strategy for all Xi > 0 : πi(0,s-i) > πi(1,s-i). If the game is played just once, the unique Nash equilibrium is one where si = 0 for all iN. When players are engaged in repeated plays of the game, with the same opponents, a much richer set of outcomes is possible. Indeed, a continuum set of outcomes can be sustained as Nash equilibria (15).

Let the game above be played repeatedly at discrete times t = 0,1,2,3,… and let Inline graphic be the option taken by agent i at time t. A strategy now becomes an action plan Inline graphic, which specifies the behavior of agent i at all times. In particular, Inline graphic may be any function of the opponents’ behavior Inline graphic in the previous stage games t < t. The payoff function is generalized to an intertemporal utility function

graphic file with name pnas.1105757109eq32.jpg [2]

where δ∈(0,1] is the factor by which agents discount future payoffs with respect to present ones. For δ → 0 the game reverts to the single stage game with the unique noncollaborative equilibrium. The limit δ → 1 of (temporally) very far-sighted players is instead much more interesting, because players can choose from a huge (a priori infinite) space of possible strategies.

Trigger Strategies.

To attack this problem, we start considering the simpler and well known situation of a two-player game, where enforceable payoffs can be attained as Nash equilibria by reducing the set of possible strategies to a particular subset, known as trigger strategies (20). Trigger strategies encode the idea of punishment: agent i collaborates a priori, but if her opponent j misbehaves (i.e., Inline graphic for some t≥0), she will punish her with defective behavior for the infinite future (i.e., Inline graphic for all t > t). If the opponent is “threatened” in this way, her best reply, if δ is large enough and Xi < 1, is to collaborate, or equivalently to use the same trigger strategy. This strategy in fact guarantees the collaboration payoff 1 - Xi > 0 in each round, as opposed to the zero payoff she would get if both defected. A crucial issue is that threats of punishment must be credible: a rational opponent will not consider credible a trigger strategy that inflicts a payoff loss to the player herself.

How does this insight carry over to multiplayer settings, where players interact with their neighbors on the network? It is instructive to consider the case of three players, without loss of generality call them 1, 2, and 3, all connected together in a closed triangle§,, with Xi < 1. Imagine that players 1 and 2 are both using trigger strategies, making their collaboration conditional on the collaboration of the others. How should player 3 behave? If she takes the threat of other players seriously, then she will also collaborate. But should she consider their threat credible? No, because if player 3 were to defect, and both 1 and 2 also turn to defection, they would get a payoff of zero, whereas if they were to continue collaborating they would get a payoff of 1 - Xi > 0 in each stage game. Hence the threat of players 1 and 2 is not credible and player 3 better defects, free—riding with a payoff of 2. On the contrary, if we assume 1 < Xi < 2 ∀i then full collaboration is individually rational and can be sustained in the repeated game as a Nash equilibrium outcome by means of credible threats. This example points out that generalizing trigger strategies entails making the strategies conditional on the behavior of a subset of the players: for Xi < 1, player 1 will punish player 2 but not player 3, because only one collaborator is needed in the neighborhood to make collaboration the best choice. In other words, control by punishment is player-specific.

Also, the same arrangement of players shows that punishment should be reciprocal. Imagine a situation where player 1 punishes 2, 2 punishes 3, and 3 punishes 1. Should 3 abide to this arrangement? No. Again if she decides to defect, she may argue that it is not credible that 2 will punish her, because again 2 would end up with a lower payoff (zero) than what she would get by continuing collaborating with 1.

Collaborative Equilibria.

The three conditions that punishment must be: (i) credible, (ii) player-specific, and (iii) reciprocal identify a notion of Nash equilibrium sustained by trigger strategies on general graphs. We shall first consider the case of trees (Proposition 1) and then generalize to general graphs (Proposition 2). Central to our analysis is the notion of nearsightedness; i.e., we focus on Nash equilibria where players can only punish their neighbors on the graph.

Let σ = 0 be the “always defect” strategy and τ(Δ) be a trigger strategy, conditional on the behavior of agents j∈Δ⊆N. An agent playing τ(Δ) will collaborate as long as all agents in Δ also collaborate. Let CN be the subset of agents who play trigger strategies, whereas agents iC always defect (σ = 0). Individuals only control their neighbors; i.e., if iC plays τi), then ΔiNi. It is convenient to introduce also the set of direct punishers of iC, which is Γi = {ji∈Δj}⊆CNi. Then let ci = |CNi| be the number of collaborators in i’s neighborhood and γi = |Γi| be the number of punishers of i.

We start with a result on networks without loops; i.e., trees.

Proposition 1

(Collaborative Equilibria in Trees) Assume that the network has no loops. The arrangement where, for all iC, Xi ≤ γi < Xi + 1, is a Nash equilibrium of the repeated game where Γi = Δi (i.e., j∈Γi if and only if i∈Γj), for a sufficiently large discount factor δ.

We refer the interested reader to SI Text for a detailed proof. We shall call Nash equilibria of this type Collaborative Equilibria (CE). In a CE, the defection of a player will generate a cascade of defections due to punishment. When player i considers whether she should punish defection of jNi, these indirect effects need to be considered. However, on trees, the defection of a neighbor jNi of player i, will only have consequences on the subtree downstream of j, and will have no direct effects on other neighbors of i. The proof of Proposition 1 relies on this fact.

The above proposition proves only sufficiency, as there could be other trigger strategies equilibria; e.g., the one in which there is no nearsightedness and in equilibrium any collaborator would immediately defect as a result of any defection. Interestingly, CE turn out to correspond, on trees, to subgame perfect Nash equilibria of a similar game with the same payoffs (22, 23). In this related game, the assumptions in Proposition 1 are also necessary, which provides even more support for the notion of CE as a refinement of repeated games on networks. We refer to the SI Text for a detailed treatment of this issue.

On general networks, the indirect effects of cascades of punishment need to be taken into account. Consider a strategy profile as in Proposition 1, where for all iC, Xi ≤ γi < Xi + 1 and Γi = Δi. Let Dji be the neighbors of i, who are not in Γi and who would defect as an indirect consequence of the defection of jC, if i does not defect. Let

graphic file with name pnas.1105757109eq33.jpg

be the set of indirect punishers of i; i.e., all those neighbors not in Γi that would defect as an indirect consequence of the defection of i herself. Then we have the following result.

Proposition 2.

(Collaborative Equilibria) Consider the arrangement where, for all iC, Xi ≤ γi < Xi + 1 and Γi = Δi. If moreover, for all iC and j∈Γi, we have that Dji = Dii, then this arrangement is a Nash equilibrium of the repeated game, for a sufficiently large discount factor δ.

We refer the interested reader to SI Text for a detailed proof. A direct consequence of Proposition 2 is that it turns the characterization of Nash equilibria into a graph theoretical problem. Whenever the conditions of Propositions 2 applies, every collaborative equilibrium of the local contribution game identifies a subgraph Inline graphic of the social network Inline graphic, with CN, ΓiNi, and |Γi| being the smallest integer larger than Xi. Every subgraph Inline graphic with these properties supports a Nash equilibrium where players iC collaborate Inline graphic and players not in C defect Inline graphic. Every player iC belongs to one and only one component in Inline graphic.

The technical condition Dji = DiiiC and j∈Γi, seems more restrictive than it actually is. It is clearly satisfied on trees, where Dji = Dii = ∅, and is expected to be verified on large random graphs, which typically have a tree-like structure (18), when C is the union of connected components of few sites (as in the example below with γi = 1∀i). It is also true on the polar case of the complete graph, where again Dii = Dji = ∅. Indeed, in a general graph, imagine two collaborating neighbors of i, call them j and k, and suppose that kDii but kDji: this can happen only if j and k have no other paths connecting them in the subgraph Inline graphic apart from the one passing through i. Interestingly, in contrast with the intuition from Proposition 1, the emergence of loops and cliques makes the statement of Proposition 2 very likely to hold in dense random networks. At the same time, on sparse graphs, the network is tree-like and it is often the case that the subgraph Inline graphic has no loops. Then, in this case, Proposition 1 applies.

For this reason, an efficient strategy to find Collaborative Equilibria is to first solve the graph theoretical problem of identifying subgraphs Inline graphic that satisfy the conditions of Proposition 1 and then to check that the technical condition in Proposition 2 is satisfied. This is the strategy we will follow in the sequel to investigate the structure of collaborative equilibria and their computational complexity.

The Complexity of Collaboration on Networks

How many collaborative equilibria exist on a given social network and how does this number depend on the number of players, the payoffs, and the network structure? How hard is it to compute an equilibrium? How does the equilibrium “respond” to local perturbations? The mapping of collaborative equilibria to collections of possibly disconnected subgraphs of a given socio-economic network allows us to address a number of questions. Some of these questions can be answered in very general terms: For example, if 0 < Xi < 1 ∀i, the subgraphs which support collaborative equilibria are collections of disjoint “dimers” (see Fig. 1 left). The number of dimer covers is generically expected to increase exponentially with the size of the network. Local deviations have only local effects, as the defection of one player affects at most the behavior of the other player on the same dimer. If 1 < Xi < 2 collaborative equilibria coincide with configurations of loops on the social network and larger Xi entail more complex structures (see Fig. 1 right). Correspondingly, the answer to the questions above becomes nontrivial. Also, the problem of finding subgraphs of a given network can be computationally very hard**, but it is typically very easy for dimers and loops, or for specific classes of graphs, such as planar graphs. Note that Proposition 2 applies for dimers and loops because Dji = ∅ for all i,j in the first case, whereas Dji = Dii on loops.

Fig. 1.

Fig. 1.

Example of collaborative equilibria for N = 40 players on a regular random graph (|Ni| = K = 4) with 0 < Xi < 1,∀i (left) and 2 < Xi < 3,∀i (right). Defectors are in black whereas collaborators are in yellow. Thick links are those between players who collaborate conditionally on each other (i∈Γj and j∈Γi).

Collaborative Equilibria on Random Graphs.

Quantitative predictions are possible for networks which are drawn from ensembles of random graphs. For large network sizes (|N| → ∞) several properties attain a limit which is independent of the particular realization of the social network. We refer to these as typical properties; i.e., properties which are expected to hold with high probability for large networks. The local tree-like structure of large graphs with a finite degree makes the characterization of typical behavior tractable by means of message passing techniques (18), such as Belief Propagation (BP), which are exact on trees and are approximately correct also for finite, moderately clustered graphs. The theoretical treatment is formally the same as in refs. 2527, which have considered similar graph-theoretic problems††. BP allows one to derive results on ensembles of graphs in the limit |N| → ∞ (18) but it also provides efficient heuristic algorithms to find collaborative equilibria on given graph instances (28, 29). The messages which are exchanged in the BP algorithm are the probabilities Inline graphic that player i collaborates and punishes j, in the collaborative equilibrium‡‡. These probabilities, are updated through the equation (26)

graphic file with name pnas.1105757109eq34.jpg [3]

where, for any integer q and subset VN,

graphic file with name pnas.1105757109eq35.jpg [4]

and the indicator function Inline graphic restricts the sum only to subsets of q elements. In words, the numerator in the right hand side (r.h.s.) of Eq. 3, which is based on Eq. 4, asserts that i should control j if there are γi - 1 other players k ≠ j who control i. The denominator expresses the fact that, in a collaborative equilibrium, three situations are possible: (i) iC, and hence i need not be controlled by any neighbor, (ii) i∈Γj and iC, or (iii) i∉Γj and iC; i.e., i is already controlled by γi neighbors and does not need to control j. The parameter ϵ∈(-∞,+∞) is a statistical weight for collaborators, which is introduced to bias the distribution over equilibria towards those with a higher (ϵ < 0) or lower (ϵ > 0) density of collaborators. The probability P{iC} that a player i collaborates or the probability P{i∈Γj} that players i,jC conditionally collaborate can be expressed in terms of the solution {μij} of the set of Eqs. 3 and 4 as follows

graphic file with name pnas.1105757109eq36.jpg [5]
graphic file with name pnas.1105757109eq37.jpg [6]

In practice, Eqs. 3 and 4 can be iterated on a specific graph, substituting the value μij with the value of the function on the r.h.s., until a fixed point is reached, which is guaranteed to occur on tree-like structures (18). The fixed point, however, is not unique. Indeed, for any collaborative equilibrium Inline graphic, binary messages Inline graphic are a solution of the BP equations. These “pure strategy” fixed points coexist with an internal solution§§. As in similar problems (27), the BP iteration converges to the internal solution, and makes possible a series of estimates on the statistical properties of the problem. In particular, we can compute the number Inline graphic of collaborative equilibria as a function of the density of collaborators Inline graphic. For large population size |N|, the number of equilibria is, to leading order, exponentially large in |N|, and one can define the entropy function Inline graphic. Similarly, one can study the entropy of collaborative equilibria as a function of other parameters of the problem (e.g., costs of contribution) or topological property (e.g., average degree). Moreover, simple adaptations of the message-passing algorithm; e.g., by iteratively fixing some variables [BP-decimation (28)] or by introducing self-consistent biases on the messages [BP-reinforcement (29)], can be used to converge towards specific pure-strategy equilibria. In this respect, varying ε allows for searching collaborative equilibria with a given average density of collaborators ρ.

On the (ensemble of) subgraphs Inline graphic identified in this manner, the validity of the technical condition in Proposition 2 has to be assessed. It is possible to do it in polynomial time on single graphs or with probabilistic arguments on ensembles.

Results.

A clear picture of the properties of collaborative equilibria can be obtained considering the conceptually simple situation of regular subgraphs of random regular graphs. It corresponds to assume a uniform cost; i.e., Xi = X ∀i, hence γi = γ = ⌈X⌉, ∀i. The condition in Proposition 2 applies for Xi = X < 3, where Inline graphic is a collection of loops and dimers. It also applies for X≥3 with probability close to one, as the probability that a random regular graph of degree γ≥3 can be split in two disjoint component by removing one node, is negligible. Examples of collaborative equilibria obtained by means of the BP-decimation algorithm on a random regular graph of degree |Ni| = K = 4 are reported in Fig. 1. The figure suggests that the density of collaborators ρ increases with the cost Xi. This conclusion is also supported by the behavior of the entropy s(ρ), shown in Fig. 2. The entropy attains a maximum at a value ρtyp of the density of collaborators, which implies that, for large |N|, almost all subgraphs span a fraction ρ ≃ ρtyp of the nodes of the network. Fig. 2 shows that ρtyp increases with Xi, suggesting that higher costs enforce higher densities of collaborators. This is an apparently counterintuitive result that has however a simple explanation in the context of the model. An agent collaborates only as long as there are enough neighbors around her that also collaborate, and that would punish her defecting if she defects: the higher the cost of collaboration, the more profitable it would be to defect, but hence also the more collaborating neighbors are needed to balance this incentive.

Fig. 2.

Fig. 2.

Entropy s(ρ) as a function of density ρ of collaborators for γ = ⌈X⌉ = 1,2, and 3 on regular random graph with |Ni| = K = 4. Results are derived analytically for |N| → ∞ as e.g., in ref. 27.

A remarkable feature of Fig. 2 is that while for γ = 1 and 2 equilibria exist for all densities ρ of collaborators (i.e., s(ρ) > 0,∀ρ∈(0,1]), for γ = 3 equilibria only exist for ρρc ≈ 0.8. This behavior reflects the fact that for γ = 1, collaborative equilibria are collections of “dimers” of collaborators who reciprocally control each others in pairs. For γ = 2, equilibria are collections of loops, whose typical size in a sparse random graph is Inline graphic. For large graphs, collaborative equilibria with γi = 1 or 2 can be constructed adding and removing dimers or loops, adjusting the density ρ in a continuous manner in (0,1]. For γ = 3, instead, collaboration requires the formation of a regular subgraph of degree 3, which is only possible if more than a fraction ρc of the nodes are involved. Hence collaborative equilibria do not exist for ρ < ρc, and when they emerge (for ρ > ρc), they are exponentially many and they span a large fraction of the network. The existence of a threshold has clear implications for the systemic stability of collaborative equilibria: while a local deviation for γ < 3 only entails local rearrangements, for γ≥3 it is likely to cause the collapse of the whole collaborative structure. Hence collaborative equilibria acquire a nonlocal character whereby the fate of agreements in a neighborhood depend on what happens in distant regions of the social network. Such systemic fragility has its roots in the organization of the space of equilibria. Collaborative equilibria are formally obtained as solutions of constraint satisfaction problems (CSP) that belong to the class of locked CSP, recently introduced in ref. 27. In these CSP, if one modifies one variable in a solution, the rearrangement required to find another solution propagates across the network possibly affecting a sizable part of it. Here, for γ < 3, the shift from a given equilibrium to another one involves the rearrangement of at most Inline graphic nodes; i.e., a negligible fraction of the system. On the contrary, for γ≥3 the minimal distance between two equilibria is proportional to the number |N| of nodes.

We have extended our analysis to heterogeneous networks; i.e., Erdös-Rényi and scale free random graphs with fixed costs Xi = X [see also (26)] and with costs of collaboration Xi = x|Ni| that are proportional to the number of neighbors each player interacts with. For these, the verification of the technical condition in Proposition 2 requires a more demanding analysis. We shall work under the assumption that this condition is verified, which means that our estimate of the number of equilibria can only be considered as an upper bound¶¶.

All the instances which have been analyzed confirmed the following set of generic features: (i) on average the typical fraction ρtyp of collaborators increases with costs; (ii) the absence of equilibria at small densities ρ for large costs; and (iii) their fragility with respect to (w.r.t.) small perturbations and nonlocal character, for sufficiently large costs. When Xi = x|Ni|, if the marginal cost x is small, equilibria are mainly formed by dimers and loops and can be found for any density ρ. When x is large, nontrivial collaborative equilibria only exist for sufficiently large density of collaborators ρ, reproducing the “critical mass” effect observed in regular random graphs. Interestingly, when the marginal cost exceeds a graph-dependent threshold xc (xc ≃ 0.79 for Erdös-Rényi random graphs with average degree equal to 4), the number of equilibria vanishes in the full range of ρ (i.e., s(ρ) < 0), suggesting that the only possible equilibria are the all-defect or the fully collaborative (for x = 1) ones.

In addition, we also found that (iv) increasing the average degree promotes collaboration on average, because denser graphs admit typical equilibria of larger density ρtyp of collaborators, as shown in Fig. 3 (left) for Erdös-Rényi random graphs. The monotonic behavior of s(ρtyp) with graph density (see the inset) is an evidence of the fact that in denser graphs there are much more possible ways of arranging a collaborating subgraph. Finally, we found that (v) within a collaborative equilibrium, the more neighbors a player has, the more she is likely to collaborate (see Fig. 3 right).

Fig. 3.

Fig. 3.

(Left) Density ρtyp of collaborators in a typical equilibrium in Erdös-Rényi graphs as function of the average degree Inline graphic for x = 0.1 (black circles) and 0.5 (red squares). The inset shows the corresponding typical entropies smax = s(ρtyp). (Right) Probability PC(k) (red squares) that a node of degree |Ni| = k collaborates in a typical collaborative equilibrium (ρtyp) and low marginal cost (x = 0.05) on an uncorrelated scale-free random graph of size |N| = 5·103 with degree distribution P(k) ∝ k-2.5 for 2 ≤ k ≤ 70 (blue circles).

Social networks are far from being uncorrelated. Therefore it is important to consider the role played by degree correlations. To this end, we have generated assortative/disassortative networks starting from uncorrelated ones by means of a link-exchange Monte Carlo algorithm proposed in ref. 30. The corresponding curves s(ρ) computed for a moderate contribution cost x = 0.1 are displayed in Fig. 4. Our results suggest*** that (vi) positive degree—correlation favors collaboration in that, the number of collaborative equilibria s(ρ) and the typical fraction of collaborators ρtyp increase with degree correlation. Remarkably, in strongly disassortative networks, collaboration can be suppressed altogether for large ρ (see Fig. 4).

Fig. 4.

Fig. 4.

Entropy s(ρ) as a function of density ρ of collaborators on correlated scale-free random graphs. We first considered an uncorrelated network (J = 0 in Eq. 9 of ref. 30) of |N| = 5·103 nodes, with degree distribution P(k) ∝ k-2.5 for 2 ≤ k ≤ 70 (black line and solid black diamond). Then we generated assortative/disassortative networks from this network, by means of a link-exchange Monte Carlo algorithm proposed in ref. 30, with positive (J = 0.1, triangles) and negative (J = -1, squares) degree-correlation. The cost is Xi = x|Ni| with x = 0.1.

Discussion

Experiments suggest that, apart from a very small fraction of innate altruists, most individuals are self-interested, hence they rationally condition their own collaboration to that of their peers (31). However, in order to sustain conditional collaboration over time in a social group, credible punishment is necessary. In a network setting, we have that it is sufficient for agents to condition their behavior only on the behavior of their neighbors. Conditional cooperation, in order to be credible, generally implies retaliation of defection only for a subset of neighbors. In other words, punishment is player-specific. This condition implies that, even in symmetric situations (e.g., complete network with Xi = X for all iN), the equilibrium can be asymmetric, with heterogeneity of behaviors arising as an emergent property. Collaborative equilibria define a subgraph of pairwise interactions that reminds endogenous network formation games, in which stable coalitions depend on the existence of reciprocal pairwise interactions (32, 33).

In brief, our main result is that credibility and locality of threats identify a class of Nash equilibrium refinements, that we call “collaborative equilibria,” that can be described in terms of a purely graph theoretical problem. This result shows that the contribution cost (incentive to defect) has a major effect on the structure of the equilibria. When it is small, agents exert a low control on the neighbors and collaboration can be easily sustained in a repeated game. Increasing the cost/incentive, the system develops strong long-range correlations and collaboration may require a critical mass. Here, the absence of equilibria with low density of collaborators evokes Oliver and Marwell’s “critical mass theory,” which is one of the most celebrated theories of collective actions (34). The theory suggests that simultaneous coordination of a sufficiently large subset of individuals could be one of the possible solutions of the free-riding problem in situations in which individual contribution is extremely disadvantageous. Here we show that the existence of a critical mass emerges naturally from simple interaction, without the need of artificially introducing a threshold in the behavior of the agents. In this case, collaboration is very fragile and, even if there are exponentially many equilibria, learning to coordinate on one of them can be very difficult.

An important conclusion from our analysis, that partially contradicts results from evolutionary game theory and learning dynamics based on imitation (9), is that network’s sparseness and degree fluctuations do not necessarily favor the emergence of collaboration. In fact, the set of constraints imposed by control and punishment relations can be much more difficult to satisfy in sparse graphs than in dense groups. This result is somehow in agreement with experimental observation (10, 12), that measured systematically higher contribution levels in cliques and dense groups than in circular or linear arrangements.

The analysis of collaborative equilibria on different topologies suggests that the global properties of typical collaborative equilibria do not change considerably on different sparse networks. However, the local properties inside a network can be very different. It is evident in heterogeneous networks, where high-degree nodes have a considerably larger probability to collaborate than low degree ones. The result is in agreement with predictions of evolutionary game theory (9) and with experimental results obtained comparing the contribution levels of central and peripheral nodes in star-like graphs (12). It also reminds the idea, firstly proposed by Haag and Lagunoff (35), of the existence of an uncollaborative fringe of agents connected to a collaborative core that can tolerate them. This picture is particularly true when considering heterogeneous networks with assortative mixing. On the contrary, in networks with negative degree correlations collaboration can be problematic, because of the mismatch between the conditions on high-degree nodes and the neighboring low-degree ones.

We conclude with a note on a recent experiment (36) on the absence of social contagion with respect to collaborative behavior. Social contagion refers to the idea (19) that some personal behaviors (as the inclination to collaborate) may be transmitted via social networks. Suri and Watts (36) find that increasing the number of collaborating neighbors does not directly imply a larger probability to collaborate (and vice versa). Our model provides a theoretical foundation for such an observation. In fact, strategic conditional collaboration requires a precise number of collaborating neighbors and when the level of collaboration in the neighborhood is too high the temptation to free-ride takes over. On the other hand, for sufficiently large costs, a collaborative state can be destroyed because of a single deviation, that may induce a cascade effect over the network.

In summary, we have put forward a unique framework to study the emergence of collaboration on networks, that is completely based on assumptions drawn from the experimental observations and on a rational forward-looking strategic behavior. Our results provide a theoretical explanation to a series of important empirical facts, and could motivate other experiments on repeated voluntary contribution games in networked systems. They provide insights in the analysis of strictly strategic problems, such as coalition formation, contractual agreements, and negotiation.

Supplementary Material

Supporting Information

Acknowledgments.

L.D. acknowledges the support from the European Community grant STAMINA 265496.

Footnotes

The authors declare no conflict of interest.

*This Direct Submission article had a prearranged editor.

This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1105757109/-/DCSupplemental.

We will explicitly refer to this endogenous profitable mutual exchange as collaboration, and not as cooperation, to stress the fact that it emerges from individual opportunistic incentives and not from group-based profit maximization, as in the cooperative games already introduced in (5).

We adopt the notation Inline graphic to specify the social network, where N is the set of nodes and the ith element of Inline graphic, denoted Ni, is the subset of the neighbors of i (iNi. The social network is undirected: iNj implies jNi).

§From now on, for the sake of simplicity and without loss of generality, we focus on the limit δ → 1 where the utility is dominated by the asymptotic behavior of players. Each statement derived for δ → 1 will be true for a δ < 1, which is large enough.

This example is discussed also in the SI Text.

This notion was already implicitly present in several multiplayer generalization of the Prisoner’s Dilemma introduced to study the free-riding problem in groups of individuals (21).

**For instance, in the case of random regular subgraphs; i.e., γi = γ ∀i, the problem is known to be NP-complete on general graphs (24).

††A finer approximation, going beyond the so-called replica-symmetric ansatz, is possible (17). This approximation allows one to explore possible organization of the space of solutions which are more complex than the one described by a single Gibbs measure, as in the BP approach, but it goes beyond the scope of the present paper and will not be pursued further.

‡‡In the subgraph problem, this is the probability that the link (i,j) is part of the subgraph, in the modified graph in which the link from i to j is removed. On trees, this operation disconnects the graph in subtrees.

§§The probabilities P{iC} and P{i∈Γj} should be interpreted in a frequentist sense, as the fraction of CE in which iC and i∈Γj. Their interpretation as mixed strategies can be misleading.

¶¶Again we expect that for small and for large Xi the condition should hold with high probability. Scale free random graphs with Xi = x|Ni| is the most delicate case, where removing highly connected nodes might disconnect parts of the subgraph Inline graphic.

***Again, these results have to be taken with caution as they are derived assuming the validity of the technical assumption in Proposition 2.

References

  • 1.Jackson MO. Social and Economic Networks. New York: Princeton; 2008. [Google Scholar]
  • 2.Homans CG. Social Behavior as Exchange. Am J Soc. 1958;62:597–606. [Google Scholar]
  • 3.Axelrod R. The Evolution of collaboration. New York: Basic Books; 1984. [Google Scholar]
  • 4.Sobel J. Can we trust social capital? J Econ Lit. 2002;40:139–154. [Google Scholar]
  • 5.von Neumann J, Morgenstern O. Theory of Game Econ Behav. New York: John Wiley and Sons; 1944. [Google Scholar]
  • 6.Ellison G. Collaboration in the prisoner’s dilemma with anonymous random matching. Rev Econ Stud. 1994;61:567–588. [Google Scholar]
  • 7.Eshel I, Samuelson L, Shaked A. Altruists, egoists, and hooligans in a local interaction model. Am Econ Rev. 1998;88:157–179. [Google Scholar]
  • 8.Nowak MA, May RM. Evolutionary games and spatial chaos. Nature. 1992;359:826–829. [Google Scholar]
  • 9.Santos FC, Pacheco JM, Lenaerts T. Evolutionary dynamics of social dilemmas in structured heterogeneous populations. Proc Natl Acad Sci USA. 2006;103:3490–3494. doi: 10.1073/pnas.0508201103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Kirchkamp O, Nagel R. Naive learning and collaboration in network experiments. Game Econ Behav. 2007;58:269–292. [Google Scholar]
  • 11.Cassar A. Coordination and collaboration in local, random and small world networks: experimental evidence. Game Econ Behav. 2007;58:209–230. [Google Scholar]
  • 12.Fatas E, Meléndez-Jiménez MA, Solaz H. An experimental analysis of team production in networks. Exp Econ. 2010;13:399–411. [Google Scholar]
  • 13.Coleman JS. Social capital in the creation of human capital. Am J Sociol. 1988;94:S95–S120. [Google Scholar]
  • 14.Burt RS. Structural Holes: The Social Structure of Competition. Cambridge: Harvard University Press; 1992. [Google Scholar]
  • 15.Fudenberg D, Maskin E. The folk theorem in repeated games with discounting or with incomplete information. Econometrica. 1986;54:533–554. [Google Scholar]
  • 16.Fehr E, Gachter S. Altruistic punishment in humans. Nature. 2002;415:137–140. doi: 10.1038/415137a. [DOI] [PubMed] [Google Scholar]
  • 17.Mézard M, Parisi G, Zecchina R. Analytic and algorithmic solution of random satisfiability problems. Science. 2002;297:812–815. doi: 10.1126/science.1073287. [DOI] [PubMed] [Google Scholar]
  • 18.Mézard M, Montanari A. Information, Physics, and Computation. Oxford: Oxford University Press; 2009. [Google Scholar]
  • 19.Fowler JH, Christakis NA. Collaborative behavior cascades in human social networks. Proc Nat Acad Sci USA. 2010;107:5334–5338. doi: 10.1073/pnas.0913149107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Rubinstein A. Finite automata play the repeated prisoner’s dilemma. J Econ Theory. 1986;39:83–96. [Google Scholar]
  • 21.Glance NS, Huberman BA. The outbreak of collaboration. J Math Sociol. 1993;17:281–302. [Google Scholar]
  • 22.Benoit J, Krishna V. Renegotiation in finitely repeated games. Econometrica. 1993;61:303–323. [Google Scholar]
  • 23.Kurzban R, Houser D. Individual differences in collaboration in circular public goods game. Eur J Personality. 2001;15:37–52. [Google Scholar]
  • 24.Garey MR, Johnson DS. Computers and Intractability. New York: Freeman; 1979. [Google Scholar]
  • 25.Marinari E, Semerjian G. On the number of circuits in random graphs. J Stat Mech. 2006:P06019. [Google Scholar]
  • 26.Pretti M, Weigt M. Sudden emergence of q-regular subgraphs in random graphs. Europhys Lett. 2006;75:8–14. [Google Scholar]
  • 27.Zdeborová L, Mézard M. Constraint satisfaction problems with isolated solutions are hard. J Stat Mech. 2008:P12004. [Google Scholar]
  • 28.Braunstein A, Mézard M, Zecchina R. Survey propagation: an algorithm for satisfiability. Random Struct Algor. 2005;27:201–226. [Google Scholar]
  • 29.Braunstein A, Zecchina R. Learning by message passing in networks of discrete synapses. Phys Rev Lett. 2006;96:030201. doi: 10.1103/PhysRevLett.96.030201. [DOI] [PubMed] [Google Scholar]
  • 30.Noh JD. Percolation transition in networks with degree-degree correlation. Phys Rev E. 2007;76:026116. doi: 10.1103/PhysRevE.76.026116. [DOI] [PubMed] [Google Scholar]
  • 31.Fischbacher U, Gächter S, Fehr E. Are people conditionally collaborative? Evidence from a public goods experiment. Econ Lett. 2001;71:397–404. [Google Scholar]
  • 32.Jackson MO, Wolinsky A. A Strategic model of social and economic networks. J Econ Theory. 1996;71:44–74. [Google Scholar]
  • 33.Vega-Redondo F. Building up social capital in a changing world. J Econ Dyn Control. 2006;30:2305–2338. [Google Scholar]
  • 34.Marwell G, Oliver P. The Critical Mass in Collective Action: A Micro-Social Theory. Cambridge University Press; 1993. [Google Scholar]
  • 35.Haag M, Lagunoff R. Social norms, local interaction and neighborhood planning. Int Econ Rev. 2006;47:265–296. [Google Scholar]
  • 36.Suri S, Watts D. Cooperation and contagion in web-based, networked public goods experiments. PLoS ONE. 2011;6:e16836. doi: 10.1371/journal.pone.0016836. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supporting Information

Articles from Proceedings of the National Academy of Sciences of the United States of America are provided here courtesy of National Academy of Sciences

RESOURCES