Significance
Many of the world’s most pressing problems, like the prevention of climate change, have the form of a large-scale social dilemma with numerous involved players. Previous results in evolutionary game theory suggest that multiplayer dilemmas make it particularly difficult to achieve mutual cooperation because of the lack of individual control in large groups. Herein, we extend the theory of zero-determinant strategies to multiplayer games to describe which strategies maintain cooperation. Moreover, we propose two simple models of alliances in multiplayer dilemmas. The effect of these alliances is determined by their size, the strategy of the allies, and the properties of the social dilemma. When a single individual’s strategic options are limited, forming an alliance can result in a drastic leverage.
Keywords: evolutionary game theory, alliances, public goods game, volunteer's dilemma, cooperation
Abstract
Direct reciprocity and conditional cooperation are important mechanisms to prevent free riding in social dilemmas. However, in large groups, these mechanisms may become ineffective because they require single individuals to have a substantial influence on their peers. However, the recent discovery of zero-determinant strategies in the iterated prisoner’s dilemma suggests that we may have underestimated the degree of control that a single player can exert. Here, we develop a theory for zero-determinant strategies for iterated multiplayer social dilemmas, with any number of involved players. We distinguish several particularly interesting subclasses of strategies: fair strategies ensure that the own payoff matches the average payoff of the group; extortionate strategies allow a player to perform above average; and generous strategies let a player perform below average. We use this theory to describe strategies that sustain cooperation, including generalized variants of Tit-for-Tat and Win-Stay Lose-Shift. Moreover, we explore two models that show how individuals can further enhance their strategic options by coordinating their play with others. Our results highlight the importance of individual control and coordination to succeed in large groups.
Cooperation among self-interested individuals is generally difficult to achieve (1–3), but typically the free rider problem is aggravated even further when groups become large (4–9). In small communities, cooperation can often be stabilized by forms of direct and indirect reciprocity (10–17). For large groups, however, it has been suggested that these mechanisms may turn out to be ineffective, as it becomes more difficult to keep track of the reputation of others and because the individual influence on others diminishes (4–8). To prevent the tragedy of the commons and to compensate for the lack of individual control, many successful communities have thus established central institutions that enforce mutual cooperation (18–22).
However, a recent discovery suggests that we may have underestimated the amount of control that single players can exert in repeated games. For the repeated prisoner’s dilemma, Press and Dyson (23) have shown the existence of zero-determinant strategies (or ZD strategies), which allow a player to unilaterally enforce a linear relationship between the own payoff and the coplayer’s payoff, irrespective of the coplayer’s actual strategy. The class of zero-determinant strategies is surprisingly rich: for example, a player who wants to ensure that the own payoff will always match the coplayer’s payoff can do so by applying a fair ZD strategy, like Tit-for-Tat. On the other hand, a player who wants to outperform the respective opponent can do so by slightly tweaking the Tit-for-Tat strategy to the own advantage, thereby giving rise to extortionate ZD strategies. The discovery of such strategies has prompted several theoretical studies, exploring how different ZD strategies evolve under various evolutionary conditions (24–30).
ZD strategies are not confined to the repeated prisoner’s dilemma. Recently published studies have shown that ZD strategies also exist in other repeated two player games (29) or in repeated public goods games (31). Herein, we will show that such strategies exist for all symmetric social dilemmas, with an arbitrary number of participants. We use this theory to describe which ZD strategies can be used to enforce fair outcomes or to prevent free riders from taking over. Our results, however, are not restricted to the space of ZD strategies. By extending the techniques introduced by Press and Dyson (23) and Akin (27), we also derive exact conditions when generalized versions of Grim, Tit-for-Tat, and Win-Stay Lose-Shift allow for stable cooperation. In this way, we find that most of the theoretical solutions for the repeated prisoner’s dilemma can be directly transferred to repeated dilemmas with an arbitrary number of involved players.
In addition, we also propose two models to explore how individuals can further enhance their strategic options by coordinating their play with others. To this end, we extend the notion of ZD strategies for single players to subgroups of players (to which we refer as ZD alliances). We analyze two models of ZD alliances, depending on the degree of coordination between the players. When players form a strategy alliance, they only agree on the set of alliance members, and on a common strategy that each alliance member independently applies during the repeated game. When players form a synchronized alliance, on the other hand, they agree to act as a single entity, with all alliance members playing the same action in a given round. We show that the strategic power of ZD alliances depends on the size of the alliance, the applied strategy of the allies, and on the properties of the underlying social dilemma. Surprisingly, the degree of coordination only plays a role as alliances become large (in which case a synchronized alliance has more strategic options than a strategy alliance).
To obtain these results, we consider a repeated social dilemma between n players. In each round of the game, players can decide whether to cooperate (C) or to defect (D). A player’s payoff depends on the player’s own decision and on the decisions of all other group members (Fig. 1A): in a group in which j of the other group members cooperate, a cooperator receives the payoff , whereas a defector obtains . We assume that payoffs satisfy the following three properties that are characteristic for social dilemmas (corresponding to the individual-centered interpretation of altruism in ref. 32): (i) irrespective of the own strategy, players prefer the other group members to cooperate ( and for all j); (ii) within any mixed group, defectors obtain strictly higher payoffs than cooperators ( for all j); and (iii) mutual cooperation is favored over mutual defection . To illustrate our results, we will discuss two particular examples of multiplayer games (Fig. 1B). In the first example, the public goods game (33), cooperators contribute an amount to a common pool, knowing that total contributions are multiplied by r (with ) and evenly shared among all group members. Thus, a cooperator’s payoff is , whereas defectors yield . In the second example, the volunteer’s dilemma (34), at least one group member has to volunteer to bear a cost in order for all group members to derive a benefit . Therefore, cooperators obtain (irrespective of j), whereas defectors yield if and . Both examples (and many more, such as the collective risk dilemma) (7, 8, 35) are simple instances of multiplayer social dilemmas.
Fig. 1.
Illustration of the model assumptions for repeated social dilemmas. (A) We consider symmetric n−player social dilemmas in which each player can either cooperate or defect. The player’s payoff depends on its own decision and on the number of other group members who decide to cooperate. (B) We will discuss two particular examples: the public goods game (in which payoffs are proportional to the number of cooperators) and the volunteer’s dilemma (as the most simple example of a nonlinear social dilemma). (C) In addition to individual strategies, we will also explore how subjects can enhance their strategic options by coordinating their play with other group members. We refer to the members of such a ZD alliance as allies, and we call group members that are not part of the ZD alliance outsiders. Outsiders are not restricted to any particular strategy. Some or all of the outsiders may even form their own alliance.
We assume that the social dilemma is repeated, such that individuals can react to their coplayers’ past actions (for simplicity, we will focus here on the case of an infinitely repeated game). As usual, payoffs for the repeated game are defined as the average payoff that players obtain over all rounds. In general, strategies for such repeated games can become arbitrarily complex, as subjects may condition their behavior on past events and on the round number in nontrivial ways. Nevertheless, as in pairwise games, ZD strategies turn out to be surprisingly simple.
Results
Memory-One Strategies and Akin’s Lemma.
ZD strategies are memory-one strategies (23, 36); they only condition their behavior on the outcome of the previous round. Memory-one strategies can be written as a vector . The entries denote the probability to cooperate in the next round, given that the player previously played and that j of the coplayers cooperated (in the SI Text, we present an extension in which players additionally take into account who of the coplayers cooperated). A simple example of a memory-one strategy is the strategy Repeat, , which simply reiterates the own move of the previous round, and . In addition, memory-one strategies need to specify a cooperation probability for the first round. However, our results will often be independent of the initial play, and in that case we will drop .
Let us consider a repeated game in which a focal player with memory-one strategy p interacts with arbitrary coplayers (who are not restricted to any particular strategy). Let denote the probability that the outcome of round t is . Let be the vector of these probabilities. A limit distribution v is a limit point for of the sequence . The entries of such a limit distribution correspond to the fraction of rounds in which the focal player finds herself in state over the course of the game.
There is a surprisingly powerful relationship between a focal player’s memory-one strategy and the resulting limit distribution of the iterated game. To show this relationship, let be the probability that the focal player cooperates in round t. By definition of we can write . Similarly, we can express the probability that the focal player cooperates in the next round as . It follows that . Summing up over all rounds from 1 to t, and dividing by t, yields , which has absolute value at most . By taking the limit we can conclude that
| [1] |
This relation between a player’s memory-one strategy and the resulting limit distribution will prove to be extremely useful. Because the importance of Eq. 1 has been first highlighted by Akin (27) in the context of the pairwise prisoner’s dilemma, we will refer to it as Akin’s lemma. We note that Akin’s lemma is remarkably general, because it neither makes any assumptions on the specific game being played, nor does it make any restrictions on the strategies applied by the remaining group members.
Zero-Determinant Strategies in Multiplayer Social Dilemmas.
As an application of Akin’s lemma, we will show in the following that single players can gain an unexpected amount of control over the resulting payoffs in a multiplayer social dilemma. To this end, we first need to introduce some further notation. For a focal player i, let us write the possible payoffs in a given round as a vector , with and . Similarly, let us write the average payoffs of i’s coplayers as , where the entries are given by and . Finally, let 1 denote the -dimensional vector with all entries being one. Using this notation, we can write player i’s payoff in the repeated game as , and the average payoff of i’s coplayers as . Moreover, by definition of v as a limit distribution, it follows that . After these preparations, let us assume player i applies the memory-one strategy
| [2] |
with α, β, and γ being parameters that can be chosen by player i (with the only restriction that ). Due to Akin’s lemma, we can conclude that such a player enforces the relationship
| [3] |
Player i’s strategy thus guarantees that the resulting payoffs of the repeated game obey a linear relationship, irrespective of how the other group members play. Moreover, by appropriately choosing the parameters α, β, and γ, the player has direct control on the form of this payoff relation. As in Press and Dyson (23), who were first to discover such strategies for the prisoner’s dilemma, we refer to the memory-one strategies in Eq. 2 as zero-determinant strategies or ZD strategies.
For our purpose, it will be convenient to proceed with a slightly different representation of ZD strategies. Using the parameter transformation , , and , ZD strategies take the form
| [4] |
and the enforced payoff relationship according to Eq. 3 becomes
| [5] |
We refer to l as the baseline payoff of the ZD strategy and to s as the strategy’s slope. Both parameters allow an intuitive interpretation: when all players adopt the same ZD strategy p such that , it follows from Eq. 5 that each player yields the payoff l. The value of s determines how the mean payoff of the other group members varies with . The parameter ϕ does not have a direct effect on Eq. 5; however, the magnitude of ϕ determines how fast payoffs converge to this linear payoff relationship as the repeated game proceeds (37).
The parameters and ϕ of a ZD strategy cannot be chosen arbitrarily, because the entries are probabilities that need to satisfy . In general, the admissible parameters depend on the specific social dilemma being played. In SI Text, we show that exactly those relations 5 can be enforced for which either (in which case the parameter l in the definition of ZD strategies becomes irrelevant) or for which l and satisfy
| [6] |
It follows that feasible baseline payoffs are bounded by the payoffs for mutual cooperation and mutual defection, , and that the slope needs to satisfy . With s sufficiently close to 1, any baseline payoff between and can be achieved. Moreover, because the conditions in Eq. 6 become increasingly restrictive as the group size n increases, larger groups make it more difficult for players to enforce specific payoff relationships.
Important Examples of ZD Strategies.
In the following, we discuss some examples of ZD strategies. At first, let us consider a player who sets the slope to . By Eq. 5, such a player enforces the payoff relation , such that i’s payoff matches the average payoff of the other group members. We call such ZD strategies fair. As shown in Fig. 2A, fair strategies do not ensure that all group members get the same payoff; due to our definition of social dilemmas, unconditional defectors always outperform unconditional cooperators, no matter whether the group also contains fair players. Instead, fair players can only ensure that they do not take any unilateral advantage of their peers. Our characterization 6 implies that all social dilemmas permit a player to be fair, irrespective of the group size. As an example, consider the strategy proportional Tit-for-Tat (pTFT), for which the probability to cooperate is simply given by the fraction of cooperators among the coplayers in the previous round
| [7] |
For pairwise games, this definition of pTFT simplifies to Tit-for-Tat, which is a fair ZD strategy (23). However, also for the public goods game and for the volunteer’s dilemma, pTFT is a ZD strategy, because it can be obtained from Eq. 4 by setting and , with c being the cost of cooperation.
Fig. 2.

Characteristic dynamics of payoffs over the course of the game for three different ZD strategies. Each panel depicts the payoff of the focal player (blue) and the average payoff of the other group members (red) by thick lines. Additionally, the individual payoffs of the other group members are shown as thin red lines. (A) A fair player ensures that the own payoff matches the mean payoff of the other group members. However, fair strategies cannot ensure that all group members yield the same payoff. (B) For games in which mutual defection leads to the lowest group payoff, extortionate players ensure that their payoffs are above average. (C) In games in which mutual cooperation is the social optimum, generous players let their coplayers gain higher payoffs. The three graphs depict the case of a public goods game with , , and group size . For the strategies of the other group members, we used random memory-one strategies, where the cooperation probabilities were independently drawn from a uniform distribution. For the strategies of the focal player, we used (A) pTFT, (B) with , and (C) with .
As another interesting subclass of ZD strategies, let us consider strategies that choose the mutual defection payoff as baseline payoff, , and that enforce a positive slope . The enforced payoff relation 5 becomes , implying that on average the other group members only get a fraction s of any surplus over the mutual defection payoff. Moreover, as the slope s is positive, the payoffs and are positively related. As a consequence, the collective best reply for the remaining group members is to maximize i’s payoffs by cooperating in every round. In analogy to Press and Dyson (23), we call such ZD strategies extortionate, and we call the quantity the extortion factor. For games in which , Eq. 5 shows that the extortion factor can be written as . Large extortion factors thus signal a substantial inequality in favor of player i. Extortionate strategies are particularly powerful in social dilemmas in which mutual defection leads to the lowest group payoff (as in the public goods game and in the volunteer’s dilemma). In that case, they enforce the relation ; on average, player i performs at least as well as the other group members (as also depicted in Fig. 2B). As an example, let us consider a public goods game and a ZD strategy with , , for which Eq. 4 implies
| [8] |
independent of the player’s own move . In the limit , approaches the fair strategy pTFT. As s decreases from 1, the cooperation probabilities of are increasingly biased to the own advantage. Extortionate strategies exist for all social dilemmas (this follows from condition [6] by setting and choosing an s close to 1). However, larger groups make extortion more difficult. For example, in public goods games with , players cannot be arbitrarily extortionate any longer as [6] implies that there is an upper bound on χ (SI Text).
As the benevolent counterpart to extortioners, Stewart and Plotkin described a set of generous strategies for the iterated prisoner’s dilemma (24, 28). Generous players set the baseline payoff to the mutual cooperation payoff while still enforcing a positive slope . These parameter choices result in the payoff relation . In particular, for games in which mutual cooperation is the optimal outcome for the group (as in the public goods game and in the prisoner’s dilemma but not in the volunteer’s dilemma), the payoff of a generous player satisfies (Fig. 2C). For the example of a public goods game, we obtain a generous ZD strategy by setting and , such that
| [9] |
For , approaches the fair strategy pTFT, whereas lower values of s make more cooperative. Again, such generous strategies exist for all social dilemmas, but the extent to which players can be generous depends on the particular social dilemma and on the size of the group.
As a last interesting class of ZD strategies, let us consider players who choose . By Eq. 5, such players enforce the payoff relation , meaning that they have unilateral control over the mean payoff of the other group members (for the prisoner’s dilemma, such equalizer strategies were first discovered in ref. 38). However, unlike extortionate and generous strategies, equalizer strategies typically cease to exist once the group size exceeds a critical threshold. For the example of a public goods game this threshold is given by . For larger groups, single players cannot determine the mean payoff of their peers any longer.
Stable Cooperation in Multiplayer Social Dilemmas.
Let us next explore which ZD strategies give rise to a Nash equilibrium with stable cooperation. In SI Text, we prove that such ZD strategies need to have two properties: they need to be generous (by setting and ), but they must not be too generous [the slope needs to satisfy ]. In particular, whereas in the repeated prisoner’s dilemma any generous strategy with is a Nash equilibrium (27, 28), larger group sizes make it increasingly difficult to uphold cooperation. In the limit of infinitely large groups, it follows that s needs to approach 1, suggesting that ZD strategies need to become fair. For the public goods game, this implies that stable cooperation can always be achieved when players cooperate in the first round and adopt proportional Tit-for-Tat thereafter. Interestingly, this strategy has received little attention in the previous literature. Instead, researchers have focused on other generalized versions of Tit-for-Tat, which cooperate if at least k coplayers cooperated in the previous round (4, 39, 40). Such memory-one strategies take the form if and if . Unlike pTFT, these threshold strategies neither enforce a linear relation between payoffs, nor do they induce fair outcomes, suggesting that pTFT may be the more natural generalization of Tit-for-Tat in large-scale social dilemmas.
In addition to the stable ZD strategies, Akin’s lemma also allows us to characterize all pure memory-one strategies that sustain mutual cooperation. In SI Text, we show that any such strategy p needs to satisfy the following four conditions
| [10] |
with no restrictions being imposed on the other entries . The first condition ensures that individuals continue to play C after mutual cooperation; the second condition guarantees that any unilateral deviation is punished; and the last two conditions describe whether players are allowed to revert to cooperation after rounds with almost uniform defection. Surprisingly, only these last two conditions depend on the specific payoffs of the social dilemma. As an application, condition 10 imply that the threshold variants of Tit-for-Tat discussed above are only a Nash equilibrium if they use the most stringent threshold: . Such unforgiving strategies, however, have the disadvantage that they are often susceptible to errors: already a small probability that players fail to cooperate may cause a complete breakdown of cooperation (41). Instead, the stochastic simulations by Hauert and Schuster (5) showed that successful strategies tend to cooperate after mutual cooperation and after mutual defection [i.e., and for all other states ]. We refer to such a behavior as WSLS, because for pairwise dilemmas it corresponds to the Win-Stay, Lose-Shift strategy described by ref. 36. Because of condition [10], WSLS is a Nash equilibrium if and only if the social dilemma satisfies . For the example of a public goods game, this condition simplifies to , which is always fulfilled for . For social dilemmas that meet this condition, WSLS provides a stable route to cooperation that is robust to errors.
Zero-Determinant Alliances.
In agreement with most of the theoretical literature on repeated social dilemmas, our previous analysis is based on the assumption that individuals act independently. As a result, we observed that a player’s strategic options typically diminish with group size. As a countermeasure, subjects may try to gain strategic power by coordinating their strategies with others. In the following, we thus extend our theory of ZD strategies for single individuals to subgroups of players. We refer to these subgroups as ZD alliances. Because the strategic power of ZD alliances is likely to depend on the exact mode of coordination between the allies, we consider two different models: when subjects form a strategy alliance, they only agree on the set of alliance members and on a common ZD strategy that each ally independently applies. During the actual game, there is no further communication between the allies. Strategy alliances can thus be seen as a boundary case of coordinated play, which requires a minimum amount of coordination. Alternatively, we also analyze synchronized alliances, in which all allies synchronize their actions in each round (i.e., the allies cooperate collectively, or they defect collectively). In effect, such a synchronized alliance thus behaves like a new entity that has a higher leverage than each player individually. Synchronized alliances thus may be considered as a boundary case of coordinated play that requires substantial coordination.
To model strategy alliances, let us consider a group of allies, with . We assume that all allies make a binding agreement that they will play according to the same ZD strategy p during the repeated game. Because the ZD strategy needs to allow allies to differentiate between the actions of the other allies and the outsiders, we need to consider a more general state space than before. The state space now takes the form . The first entry S corresponds to the focal player’s own play in the previous round, gives the number of cooperators among the other allies, and is the number of cooperators among the outsiders. A memory-one strategy p again needs to specify a cooperation probability for each of the possible states. Using this state space, we can define ZD strategies for a player i in a strategy alliance as
| [11] |
The vector contains the average payoff of the other allies for each possible state, and is the corresponding vector for the outsiders. The weights and are additional parameters that determine the relative importance of outsiders and other allies, being subject to the constraint . In the special case of a single player forming an alliance, , this guarantees that the two definitions of ZD strategies 4 and 11 are equivalent.
Similarly to the case of single individuals, we can apply Akin’s lemma to show that strategy alliances enforce a linear relationship between their own mean payoff and the mean payoff of the outsiders (for details, see SI Text)
| [12] |
where the slope of the alliance is given by A strategy alliance can enforce exactly those payoff relationships 12 for which either or for which l and satisfy the conditions
| [13] |
Interestingly, to reach this strategic power, an alliance needs to put a higher weight on the within-alliance payoffs (i.e., needs to exceed ; SI Text), such that the allies are stronger affected by what the other allies do, as opposed to the actions of the outsiders. For single player alliances, , condition 13 again simplifies to the previous condition 6. However, as the alliance size increases, condition 13 becomes easier to satisfy. Larger alliances can therefore enforce more extreme payoff relationships. For the example of a public goods game, we noted that single players cannot be arbitrarily extortionate when . Alliances, on the other hand, only need to be sufficiently large, . Once an alliance has this critical mass, there are no bounds to extortion.
In a similar way, we can also analyze the strategic possibilities of a synchronized alliance. Because synchronized alliances act as a single entity, they transform the symmetric social dilemma between n independent players to an asymmetric game between independent players. From the perspective of the alliance, the state space now takes the form , where is the common action of all allies and where is the number of cooperators among the outsiders. ZD strategies for the synchronized alliance can be defined analogously to ZD strategies for single players
| [14] |
with being the payoff vector for the allies and being the payoff vector of the outsiders. For a single player alliance, , this again reproduces the definition of ZD strategies in 4. By applying Akin’s lemma to Eq. 14, we conclude that synchronized alliances enforce , which is the same as relationship 12 for strategy alliances. Surprisingly, we even find that for reasonable alliance sizes, , strategy alliances and synchronized alliances have the same set of enforceable parameters l and , as given by Eq. 13 (see SI Text for details). Thus, for the two models of ZD alliances considered here, the exact mode of coordination is irrelevant for the alliance’s strategic power unless the alliance has reached a substantial size.
Table 1 gives an overview of our findings on ZD strategies and ZD alliances in multiplayer social dilemmas. It shows that, although generally, ZD strategies exist for all group sizes, the power of single players to enforce particular outcomes typically diminishes or disappears in large groups. Forming ZD alliances then allows players to increase their strategic scope. The impact of a given ZD alliance, however, depends on the specific social dilemma: although ZD alliances can become arbitrarily powerful in public goods games, their strategic options remain limited in the volunteer’s dilemma.
Table 1.
Strategic power of different ZD strategies for three different social dilemmas
| Strategy class | Typical property | Prisoner’s dilemma | Public goods game | Volunteer’s dilemma |
| Fair strategies | Always exist | Always exist | Always exist | |
| Extortionate strategies | Always exist | In large groups, single players cannot be arbitrarily extortionate, but sufficiently large ZD alliances can be arbitrarily extortionate | Even large ZD alliances cannot be arbitrarily extortionate | |
| Generous strategies | Always exist | In large groups, single players cannot be arbitrarily generous, but sufficiently large ZD alliances can be arbitrarily generous | Do not ensure that own payoff is below average | |
| Equalizers | Always exist | May not be feasible for single players, but is always feasible for sufficiently large ZD alliances | Only feasible if size of ZD alliance is , can only enforce |
Analogously to the case of individual players, ZD alliances are fair when they set ; they are extortionate when and ; they are generous for and ; and they are equalizers when . For each of the three considered social dilemmas, we explore whether a given ZD strategy is feasible by examining the respective conditions in Eq. 13. In the repeated prisoner’s dilemma, single players can exert all strategic behaviors (23, 28, 29). Other social dilemmas either require players to form alliances to gain sufficient control (as in the public goods game), or they only allow for limited forms of control (as in the volunteer’s dilemma). These results hold both for strategy alliances and for synchronized alliances.
Discussion
When Press and Dyson (23) discovered the new class of ZD strategies for the repeated prisoner’s dilemma, this came as a big surprise (24, 25): after more than five decades of research, it seemed unlikely that any major property of the prisoner’s dilemma has been overlooked. For repeated multiplayer dilemmas the situation is different. Although various Folk theorems guarantee that cooperation is also feasible in large groups (42, 43), there has been considerably less theoretical research on the evolution of cooperation in repeated multiplayer dilemmas (4, 5, 39, 40). This lack of research may be due to the higher complexity: the mathematics of repeated n-player dilemmas seems to be more intricate, and numerical investigations are impeded because the time to compute payoffs increases exponentially in the number of players (5). Nevertheless, we showed here that many of the results for the repeated prisoner’s dilemma can be directly transferred to general social dilemmas, with an arbitrary number of involved subjects. The foundation for this progress is a new framework, provided by Akin’s lemma and the theory of Press and Dyson.
Using this framework, we extended the theory of repeated multiplayer dilemmas into three directions. The first and most immediate direction is our finding that ZD strategies exist in all social dilemmas. These strategies allow players to unilaterally dictate linear payoff relations, irrespective of the specific social dilemma being played, irrespective of the group size, and irrespective of the counter measures taken by the other group members. In particular, we showed that any social dilemma allows players to be fair, extortionate, or generous. Each of these strategy classes has its own particular strengths: extortionate strategies give a player a relative advantage compared with the other group members; fair strategies help to avoid further inequality within a group; and generous strategies allow players to revert to mutual cooperation when a coplayer defected by accident. At the same time, ZD strategies are remarkably simple. For example, to be fair in a public goods game, players only need to apply a rule called proportional Tit-for-Tat: if j of the other group members cooperated in the previous round, then cooperate with probability in the following round. Extortionate and generous strategies can be obtained in a similar way, by slightly modifying pTFT to the own advantage or to the advantage of the others.
As the second direction, we explored which ZD strategies and which pure memory-one strategies can be used to sustain cooperation in multiplayer dilemmas. Among ZD strategies, such strategies need to be generous (such that players never try to outperform their peers) (27, 28), but at the same time they must not be too generous. The right degree of generosity depends on the size of the group but not on the specific social dilemma being played. As a rule of thumb, we obtain that in larger groups, subjects are required to show less generosity.
As the last direction, we extended the concept of zero-determinant strategies from single players to subgroups of players, to which we refer to as ZD alliances. Depending on the degree of coordination, we explored two forms of ZD alliances: members of a strategy alliance only agree on using a common ZD strategy during the game, but they do not coordinate each of their decisions; members of a synchronized alliance, on the other hand, act as a single entity—they either all cooperate or they all defect in a given round. The effect of such ZD alliances depends on the size of the alliance, the applied strategy, and the properties of the underlying social dilemma. In general, we find that by coordinating their play with others, subjects can increase their strategic options considerably. The exact mode of coordination, however, only turns out to play a minor role: As long as the size of the ZD alliance is below half the group size, strategy alliances and synchronized alliances have the same strategic power. In addition to their static properties, ZD strategies for the prisoner’s dilemma also have a remarkable dynamic component (23, 44): when a player commits himself to an extortionate ZD strategy, then adapting coplayers learn to cooperate over time. Numerical simulations in the SI show an analogous result for multiplayer dilemmas: when ZD alliances apply strategies with a positive slope, they can trigger a positive group dynamics among the outsiders. The magnitude of this dynamic effect again depends on the size of the alliance, and on the applied strategy of the allies.
Here, we focused on ZD strategies; but the toolbox that we apply (in particular Akin’s lemma) is more general. As an example, we identified all pure memory-one strategies that allow for stable cooperation, including the champion of the repeated prisoner’s dilemma, Win-Stay Lose-Shift (36, 45). We expect that there will be further applications of Akin’s lemma to come. Such applications may include, for instance, a characterization of all Nash equilibria among the stochastic memory-one strategies or an analysis of how alliances are formed and whether evolutionary forces favor particular alliances over others (46, 47).
Overall, our results reveal how single players in multiplayer games can increase their control by choosing the right strategies and how they can increase their strategic options by joining forces with others.
Supplementary Material
Acknowledgments
C.H. gratefully acknowledges generous funding by the Schrödinger scholarship of the Austrian Science Fund (J3475).
Footnotes
The authors declare no conflict of interest.
This article is a PNAS Direct Submission. J.B.P. is a guest editor invited by the Editorial Board.
This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1407887111/-/DCSupplemental.
References
- 1.Hardin G. The tragedy of the commons. The population problem has no technical solution; it requires a fundamental extension in morality. Science. 1968;162(3859):1243–1248. [PubMed] [Google Scholar]
- 2.Olson M. The Logic of Collective Action: Public Goods and the Theory of Groups. Revised Ed Harvard Univ Press; Cambridge, MA: 1971. [Google Scholar]
- 3.Nowak MA. Five rules for the evolution of cooperation. Science. 2006;314(5805):1560–1563. doi: 10.1126/science.1133755. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Boyd R, Richerson PJ. The evolution of reciprocity in sizable groups. J Theor Biol. 1988;132(3):337–356. doi: 10.1016/s0022-5193(88)80219-4. [DOI] [PubMed] [Google Scholar]
- 5.Hauert C, Schuster HG. Effects of increasing the number of players and memory size in the iterated prisoner’s dilemma: A numerical approach. Proc Biol Sci. 1997;264(1381):513–519. [Google Scholar]
- 6.Suzuki S, Akiyama E. Evolution of indirect reciprocity in groups of various sizes and comparison with direct reciprocity. J Theor Biol. 2007;245(3):539–552. doi: 10.1016/j.jtbi.2006.11.002. [DOI] [PubMed] [Google Scholar]
- 7.Santos FC, Pacheco JM. Risk of collective failure provides an escape from the tragedy of the commons. Proc Natl Acad Sci USA. 2011;108(26):10421–10425. doi: 10.1073/pnas.1015648108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Abou Chakra M, Traulsen A. Evolutionary dynamics of strategic behavior in a collective-risk dilemma. PLOS Comput Biol. 2012;8(8):e1002652. doi: 10.1371/journal.pcbi.1002652. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Yang W, et al. Nonlinear effects of group size on collective action and resource outcomes. Proc Natl Acad Sci USA. 2013;110(27):10916–10921. doi: 10.1073/pnas.1301733110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Trivers RL. The evolution of reciprocal altruism. Q Rev Biol. 1971;46(1):35–57. [Google Scholar]
- 11.Axelrod R. The Evolution of Cooperation. Basic Books; New York: 1984. [Google Scholar]
- 12.Alexander R. The Biology of Moral Systems. Aldine de Gruyter; New York: 1987. [Google Scholar]
- 13.Wedekind C, Milinski M. Cooperation through image scoring in humans. Science. 2000;288(5467):850–852. doi: 10.1126/science.288.5467.850. [DOI] [PubMed] [Google Scholar]
- 14.Henrich J, et al. Cooperation, reciprocity and punishment in fifteen small scale societies. Am Econ Rev. 2001;91(2):73–78. [Google Scholar]
- 15.Nowak MA, Sigmund K. Evolution of indirect reciprocity. Nature. 2005;437(7063):1291–1298. doi: 10.1038/nature04131. [DOI] [PubMed] [Google Scholar]
- 16.van Veelen M, García J, Rand DG, Nowak MA. Direct reciprocity in structured populations. Proc Natl Acad Sci USA. 2012;109(25):9929–9934. doi: 10.1073/pnas.1206694109. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Zagorsky BM, Reiter JG, Chatterjee K, Nowak MA. Forgiver triumphs in alternating Prisoner’s Dilemma. PLoS ONE. 2013;8(12):e80814. doi: 10.1371/journal.pone.0080814. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Yamagishi T. The provision of a sanctioning system as a public good. J Pers Soc Psychol. 1986;51(3):110–116. [Google Scholar]
- 19.Ostrom E. Governing the Commons: The Evolution of Institutions for Collective Action. Cambridge Univ Press; Cambridge, UK: 1990. [Google Scholar]
- 20.Sigmund K, De Silva H, Traulsen A, Hauert C. Social learning promotes institutions for governing the commons. Nature. 2010;466(7308):861–863. doi: 10.1038/nature09203. [DOI] [PubMed] [Google Scholar]
- 21.Guala F. Reciprocity: Weak or strong? What punishment experiments do (and do not) demonstrate. Behav Brain Sci. 2012;35(1):1–15. doi: 10.1017/S0140525X11000069. [DOI] [PubMed] [Google Scholar]
- 22.Hilbe C, Traulsen A, Röhl T, Milinski M. Democratic decisions establish stable authorities that overcome the paradox of second-order punishment. Proc Natl Acad Sci USA. 2014;111(2):752–756. doi: 10.1073/pnas.1315273111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Press WH, Dyson FJ. Iterated Prisoner’s Dilemma contains strategies that dominate any evolutionary opponent. Proc Natl Acad Sci USA. 2012;109(26):10409–10413. doi: 10.1073/pnas.1206569109. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Stewart AJ, Plotkin JB. Extortion and cooperation in the prisoner’s dilemma. Proc Natl Acad Sci USA. 2012;109(26):10134–10135. doi: 10.1073/pnas.1208087109. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Ball Physicists suggest selfishness can pay. Nature. 2012 doi: 10.1038/nature.2012.11254. [DOI] [Google Scholar]
- 26.Hilbe C, Nowak MA, Sigmund K. The evolution of extortion in iterated prisoner’s dilemma games. Proc Natl Acad Sci USA. 2013;110(17):6913–6918. doi: 10.1073/pnas.1214834110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Akin E. 2013 Stable cooperative solutions for the iterated prisoner’s dilemma. arXiv:1211.0969v2. [Google Scholar]
- 28.Stewart AJ, Plotkin JB. From extortion to generosity, evolution in the Iterated Prisoner’s Dilemma. Proc Natl Acad Sci USA. 2013;110(38):15348–15353. doi: 10.1073/pnas.1306246110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Hilbe C, Nowak MA, Traulsen A. Adaptive dynamics of extortion and compliance. PLoS ONE. 2013;8(11):e77886. doi: 10.1371/journal.pone.0077886. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Szolnoki A, Perc M. Evolution of extortion in structured populations. Phys Rev E Stat Nonlin Soft Matter Phys. 2014;89(2):022804. doi: 10.1103/PhysRevE.89.022804. [DOI] [PubMed] [Google Scholar]
- 31.Pan L, Hao D, Rong Z, Zhou T. 2014 Zero-determinant strategies in the iterated public goods game. arXiv:1402.3542v1. [Google Scholar]
- 32.Kerr B, Godfrey-Smith P, Feldman MW. What is altruism? Trends Ecol Evol. 2004;19(3):135–140. doi: 10.1016/j.tree.2003.10.004. [DOI] [PubMed] [Google Scholar]
- 33.Ledyard JO. Public goods: A survey of experimental research. In: Kagel JH, Roth AE, editors. The Handbook of Experimental Economics. Princeton Univ Press; Princeton: 1995. pp. 111–194. [Google Scholar]
- 34.Diekmann A. Volunteer’s dilemma. J Conflict Resolut. 1985;29(4):605–610. [Google Scholar]
- 35.Milinski M, Sommerfeld RD, Krambeck H-J, Reed FA, Marotzke J. The collective-risk social dilemma and the prevention of simulated dangerous climate change. Proc Natl Acad Sci USA. 2008;105(7):2291–2294. doi: 10.1073/pnas.0709546105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Nowak M, Sigmund K. A strategy of win-stay, lose-shift that outperforms tit-for-tat in the Prisoner’s Dilemma game. Nature. 1993;364(6432):56–58. doi: 10.1038/364056a0. [DOI] [PubMed] [Google Scholar]
- 37.Hilbe C, Röhl T, Milinski M. Extortion subdues human players but is finally punished in the prisoner’s dilemma. Nat Commun. 2014;5:3976. doi: 10.1038/ncomms4976. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Boerlijst MC, Nowak MA, Sigmund K. Equal pay for all prisoners. Am Math Mon. 1997;104(4):303–307. [Google Scholar]
- 39.Kurokawa S, Ihara Y. Emergence of cooperation in public goods games. Proc Biol Sci. 2009;276(1660):1379–1384. doi: 10.1098/rspb.2008.1546. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Van Segbroeck S, Pacheco JM, Lenaerts T, Santos FC. Emergence of fairness in repeated group interactions. Phys Rev Lett. 2012;108(15):158104. doi: 10.1103/PhysRevLett.108.158104. [DOI] [PubMed] [Google Scholar]
- 41.Sigmund K. The Calculus of Selfishness. Princeton Univ Press; Princeton: 2010. [Google Scholar]
- 42.Friedman J. A non-cooperative equilibrium for supergames. Rev Econ Stud. 1971;38(1):1–12. [Google Scholar]
- 43.Fudenberg D, Tirole J. Game Theory. 6th Ed MIT Press; Cambridge, MA: 1998. [Google Scholar]
- 44.Chen J, Zinger A. The robustness of zero-determinant strategies in Iterated Prisoner’s Dilemma games. J Theor Biol. 2014;357:46–54. doi: 10.1016/j.jtbi.2014.05.004. [DOI] [PubMed] [Google Scholar]
- 45.Kraines DP, Kraines VY. Pavlov and the prisoner’s dilemma. Theory Decis. 1989;26(1):47–79. [Google Scholar]
- 46.Peleg B, Sudhölter P. Introduction to the Theory of Cooperative Games. Springer; Berlin: 2003. [Google Scholar]
- 47.Mesterton-Gibbons M, Gavrilets S, Gravner J, Akçay E. Models of coalition or alliance formation. J Theor Biol. 2011;274(1):187–204. doi: 10.1016/j.jtbi.2010.12.031. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.

