Abstract
Recent theory shows that extortioners taking advantage of the zero-determinant (ZD) strategy can unilaterally claim an unfair share of the payoffs in the Iterated Prisoner’s Dilemma. It is thus suggested that against a fixed extortioner, any adapting coplayer should be subdued with full cooperation as their best response. In contrast, recent experiments demonstrate that human players often choose not to accede to extortion out of concern for fairness, actually causing extortioners to suffer more loss than themselves. In light of this, here we reveal fair-minded strategies that are unbending to extortion such that any payoff-maximizing extortioner ultimately will concede in their own interest by offering a fair split in head-to-head matches. We find and characterize multiple general classes of such unbending strategies, including generous ZD strategies and Win-Stay, Lose-Shift (WSLS) as particular examples. When against fixed unbending players, extortioners are forced with consequentially increasing losses whenever intending to demand a more unfair share. Our analysis also pivots to the importance of payoff structure in determining the superiority of ZD strategies and in particular their extortion ability. We show that an extortionate ZD player can be even outperformed by, for example, WSLS, if the total payoff of unilateral cooperation is smaller than that of mutual defection. Unbending strategies can be used to outlearn evolutionary extortioners and catalyze the evolution of Tit-for-Tat-like strategies out of ZD players. Our work has implications for promoting fairness and resisting extortion so as to uphold a just and cooperative society.
Keywords: evolutionary game theory, repeated games, cooperation, reciprocity, fairness
Significance Statement.
Extortioners witting of the zero-determinant strategy can gain the upper hand in Iterated Prisoner’s Dilemma games by unilaterally enforcing an unfair linear relation between their own payoff and that of their coplayer. Therefore, theory predicts that acceding to extortion is the best response for any adapting coplayer. Recent empirical evidence, however, shows that human players seldom yield to extortion out of concern for fairness and are willing to discipline extortioners by refusing to fully cooperate. To shed light on such fair-minded responses, here we find and characterize general classes of unbending strategies such that the best response of any payoff-maximizing extortioner against a fixed unbending player is to offer a fair split, thereby guaranteeing equal pay for both parties.
Introduction
The Prisoner’s Dilemma (PD) has been considered a central paradigm for understanding a wide variety of cooperation problems (1). In this game, two players decide whether to cooperate (C) or defect (D). If both players choose to cooperate, they receive the same reward for mutual cooperation, R, and if they both defect, they receive the same punishment for mutual defection, P. However, if one cooperates but the other defects, the defector receives the temptation to defect, T, whereas the cooperator receives the sucker’s payoff, S. These payoff values satisfy (2).
Departure from one-shot games, the dynamics of the Iterated Prisoner’s Dilemma (IPD) can be analyzed by examining the four possible outcomes that arise when two players simultaneously play the game at each time step: , , , and . The former two outcomes result in equal payoffs for both players, while the latter two create a payoff inequality, with one player receiving a higher payoff than the other. In repeated interactions, it is possible for both players to have equal long-term average payoffs, or for one player to receive a higher payoff than the other.
To shed light on a range of cooperative or exploitative strategies in IPD games (3), prior studies have extensively investigated various behavioral choices and responses that can be characterized by prescribed intentions or preferences to be fair, cooperative, reciprocal, generous, and forgiving (or the opposite). These concepts involving reciprocal fairness and cooperation (also known as direct reciprocity, put it simply, “I will if you will”) can be investigated within the framework of IPD games.
Various strategies can be employed in an IPD game, with some being more cooperative and fair-minded than others. For instance, a fair-minded reciprocator would reciprocate cooperation at least as often as their coplayer does, rather than seeking an advantage over them. Among the common IPD strategies, Tit-for-Tat (TFT) and its variants, such as generous TFT (GTFT), are cooperative and fair-minded in nature (4). TFT-like players do not defect initially unless their coplayers had defected once or more. On the other hand, adaptive learning strategies (5), such as Win-Stay, Lose-Shift (WSLS), are more robust to noise and error than TFT (6). WSLS deterministically keeps the current strategy if the resulting payoff is above a fixed aspiration level, or switches otherwise.
An “equalizer” is capable of unilaterally setting any coplayer’s payoff level to the same arbitrary level within the range of (7). Even more capable of bilateral payoff control is the zero-determinant (ZD) strategy, discovered by Press and Dyson (8). A ZD player can unilaterally set a linear relation between the payoff of themselves and that of the coplayer, regardless of the strategy of the coplayer. In recent years, the discovery of ZD strategies has generated renewed interest in studying IPD games in light of Press and Dyson’s finding (9–18).
Of particular interest is the existence of a continuous spectrum of ZD strategies that vary in their level of generosity, ranging from extortionate ZD to generous ZD (19). Undoubtedly, witting of ZD strategies enables players to gain the upper hand in IPD games, even allowing an implicit form of extortion (8). Self-serving extortioners can leverage ZD strategies to their advantage to the fullest extent, aiming to dominate any coplayer preemptively. However, it is shown that two extortioners, both equipped with the knowledge of extortionate ZD, will neutralize each other in their interactions and lead to their own demise, both receiving P (8). The lack of mutual cooperation among extortionate ZD players may prevent them from being favored by natural selection in an evolutionary population dynamics setting, particularly in larger populations (14). However, ZD players can still be successful in small populations, and even more so when they either adapt to be more generous towards others (19) or establish reconciliation and cooperation among themselves (20).
Prior work almost invariably considers ZD fixed, while their coplayer tries to adapt to ZD’s unilateral control. In reality, extortion can be met with resistance; unbending individuals are willing to push back any attempt to extort out of concern for fairness (10, 15, 21). Indeed, recent experiments demonstrate that fixed computer ZD players are able to outcompete their human counterparts but at a huge cost in a way that human players are significantly less cooperative (15). In a variety of experimental scenarios involving incentives for extortionate human players to receive additional bonuses based on their competitiveness (10), unbending players may give up their disciplinary efforts against extortioners, ultimately losing to them. Nevertheless, these unbending players can still sabotage the extortioners’ success through occasional defections, causing a decline in the extortioners’ payoffs compared to other control conditions (10). Thus, the success of ZD’s extortion attempt can be undermined and become less effective in reaching the fullest possible extent. Moreover, ZD players need to prescribe their strategies in a sophisticated way that explicitly depends on the underlying payoff matrix in the first place. It remains unknown how potential variations in the payoff matrix, which can arise from the uncertainty of evolving game environment (22), will impact the pairwise dominance of ZD strategies and in particular their extortion ability, since not all PD games are qualitatively the same (23).
These considerations lead us to reveal the previously unforeseen Achilles’ heel of ZD strategies, specifically in one-on-one encounters. Namely, there exist simple strategies (including TFT-like strategies and WSLS as particular examples) that are unbending to extortion and can cause an unfair demand to backfire on extortioners. When against a fixed unbending player, the best response of any payoff-maximizing extortioner, characterized by their prescribed smallest possible level of generosity, is to offer a fair split, thereby guaranteeing equal payoffs for both parties.
Moreover, we show that in interactions of more adversarial nature (24), characterized by the payoff structure condition , ZD’s dominance is drastically impaired, and extortioners tempting to dominate the coplayer are likely to become victims of their own success. The fixed unbending strategies, discovered in the present study, are able to not only force greedy ZD coplayers to be fair in their own interest but also more importantly, steer adapting coplayers (including those ZD coplayers) towards fairness and cooperation in adaptive learning settings. Our work provides useful insights into understanding the important role played by unbending strategies as an enforcer and stabilizer of fairness and cooperation in dyadic interactions, of relevance and interest for studying direct reciprocity.
Results
We begin with studying the effectiveness of ZD strategies in payoff control and extortion and how it depends on their prescribed parameter choices and the underlying payoff matrix. Doing so will provide a new perspective on understanding specific conditions required for intended extortion to be successful or lack thereof. These critical considerations ultimately lead us to reveal unbending strategies that are able to outlearn ZD players and foster fairness and cooperation in pairwise interactions (see Figs. S1–S15 and Tables S1–S7 in the Online Supplementary Material).
Following common practice (6), we denote memory-one IPD strategies by , where , for , is the conditional probability to cooperate, respectively, after experiencing one of the four possible outcomes each round , that is, written from the perspective of a focal player X (the first letter represents X’s last move, and the second letter for the coplayer Y’s). Suppose that player X uses a ZD strategy and the coplayer Y uses an arbitrary strategy , and let denote the average payoff of player X and that of player Y. A general yet intuitive parameterization of memory-one ZD strategies are based on three control parameters :
(1) |
including the extortion factor , the baseline payoff , and the normalization factor ϕ that ensures to be a proper probability vector. A complete discussion of the admissible ranges of these parameters can be found in the Online Supplementary Material.
Regardless of the strategy used by Y, X unilaterally enforces a linear relative payoff relation of the form (8, 19):
(2) |
which represents a straight line in the parametric plot of with the slope (the reciprocal of the extortion factor χ) (Fig. 1a and b). In this plane, the baseline payoff determines the intercept to the line of equal payoffs, , and also dictates the level of generosity (19).
Fig. 1.
Pairwise dominance and extortion ability of ZD strategies. The baseline payoff O used by the ZD player X is regarded as extortionate P (least level of generosity), generous R (maximum level of generosity), and in between (intermediate level of generosity). The optimality of extortionate ZD strategies (with ) nontrivially depends on both the strategy of their coplayer and the payoff structure. When playing against a certain type of coplayers (which we call unbending strategies), extortioners can maximize their prospective payoffs only if aiming for an equal split by letting the extortion factor . Moreover, when , extortioners can even be outperformed. In panels a) and b), we show the scatter plot of payoff pairs of ZD players against random coplayers uniformly drawn from all possible memory-one strategies in a) and in b) . Shown in c) and d) is the probability that a ZD player X actually gets better payoff than their coplayer Y () who uses a random strategy uniformly drawn from memory-one strategies , with respect to varying their baseline payoff . The parameter O controls the level of generosity of a ZD player but also impacts their chance to outperform their coplayers (“extortion ability”). Increasing O above P makes ZD less likely to be able to ensure the dominance over their coplayers. Noticeably, the payoff structure plays an even more pronounced role than does the parameter O: c) for the curvature is concave downward and ZD is able to maintain dominance for most of the time even using intermediate values, d) whereas concave upward for and ZD is more likely to lose dominance for any . In line with a) and b), extortion with always leads to superior payoff than the coplayers a) when , but not necessarily true for b) . Parameters: a–d) ZD player X’s extortion factor , and ϕ is uniformly distributed and truncated at the admissible upper bound, a,c) , , , , b,d) , , , .
The payoff control as given in Eq. 2 enables an implicit form of extortion where ZD player X can prescribe their strategies in a way that they reciprocate cooperation less frequently than their coplayer Y (8, 14). For ZD players, the way to attempt such dominance and extortion is to deliberately choose their parameters O and χ in advance, which will in turn determine admissible values of ϕ. The chosen values of O and χ can be observed directly from pairwise payoff plots (Fig. 1a and b), and together with the underlying payoff matrix, they jointly determine the upper bound of admissible ϕ values. For example, a widely used parameterization of this ZD class, which is called extortionate ZD strategy (8), ensures that holds with . The admissible range of ϕ for extortionate ZD is given by
(3) |
Notably, the parameter ϕ has an upper bound that explicitly depends on the sign of . We emphasize that this previously overlooked payoff structure condition, whether holds or not, surprisingly strikes out as an important condition for determining the optimality of ZD strategies and their extortion ability (see Fig. S16 in the Online Supplementary Material). As shown in Fig. 1c and d, as long as a ZD player uses the minimal and , they secure the most favorable position to dominate and get higher payoffs than their opponent as compared to other O values, regardless of the underlying payoff matrix (Fig. 1c and d). Despite such a contextual difference of ZD’s extortion ability owing to the change in the underlying payoff structure, we still refer to this class of “extortionate ZD” as extortioner as in Ref. (14), for the sake of consistency. First, the effect of varying their control parameter O on their resulting level of generosity and extortion ability remains qualitatively consistent across IPD games of drastically different nature. Second, for ZD players with , no matter what types of IPD games they are engaged in, the chosen value of P characterizes the least level of generosity, and thus preemptively sets their extortion ability at maximum, even though these so-called extortioners will not always succeed in securing advantage as intended, particularly when (cf. Fig. 1c and d).
On the other hand, given the uncertainty of vastly possible strategies the coplayer could use against ZD players, it is worthwhile to quantify the robustness of the dominance and the performance of ZD strategies with particular respect to varying their baseline payoff O. It is likely that ZD players choose deviating from P for plausible reasons like the trembling hand (25) or “blurred minds” (4), and as a consequence, they will respond with nonzero cooperation (i.e. ) after entering mutual defection state with their coplayer. Even so, any ZD player using still has extortion ability to some extent unless they use the generous ZD with that ensures their average payoffs are never above the coplayer’s (19) (Fig. 1). It is thus reasonable to consider the extortion ability of ZD strategies as a continuous spectrum—“the likelihood of getting better payoffs than any kind of opponent”—instead of a binary character (either always or not at all). In doing so, we are able to quantify and compare the extortion ability of ZD strategies and how it depends on their control parameters , and more remarkably, on the underlying payoff structure specified by the sign of (Figs. 1 and 2).
Fig. 2.
Impacts of control parameters on the average payoff of a ZD player when playing against a fixed coplayer. The ZD player X’s payoff, , is shown as a function of the normalization factor ϕ and the extortion factor χ, along with contour lines projected on the -plane: a–c) for and d–f) for . The ZD player X’s payoff is either monotonic or remains constant with respect to ϕ while it can exhibit nonmonotonic behavior with respect to χ. Despite being able to enforce a linear payoff relationship , ZD player X that unilaterally uses a larger extortion factor χ does not necessarily lead to further payoff gains as demonstrated in a). As long as “winning isn’t everything” (actual payoff performance is concerned in wide-ranging scenarios), ZD can subtly tune their control parameters to optimize their own payoff performance against a fixed coplayer. Parameters: a–f): X uses the most formidable ZD strategy with (also known as extortionate ZD), a–c) , , , , the upper bound of , d–f) , , , , the upper bound of , coplayer Y’s strategy , , .
We further note that ϕ is a hidden parameter, which has received little attention in prior studies. However, we find that albeit the normalization factor ϕ has no impact on the linear payoff relation, it can nontrivially affect the average payoff values that a ZD player will receive (Fig. 2). Mathematically, ZD’s average payoff is given by the ratio of the determinants of two matrices, giving rise to a rational function (8). We can show that is a monotonic function of ϕ (Figs. S17 and S18 in the Online Supplementary Material) but can have strict nonmonotonicity with respect to χ, exhibiting as a one-humped function of χ (see Online Supplementary Material for derivation details). Fig. 2 plots an extortionate ZD’s average payoff (with the baseline payoff ) against a fixed coplayer Y using a specific strategy as a function of the parameter space . This result further demonstrates that ZD can unilaterally fine tune their control parameters, in particular the previously overlooked parameter ϕ to their own advantage (which would be boundary values of its admissible interval, either infinitely small or the upper bound).
Only if is an extortionate ZD unbeatable, ensuring no less payoffs than their opponent (the worst scenario is a tie, e.g. against TFT as shown in Fig. 2c and f). In this case, making the extortion factor χ excessively larger surely can help ZD impose a greater relative advantage over their opponent, but their actual average payoff can be seriously comprised (Fig. 2a). Even worse, when , can drop below P and due to we have (Fig. 2e). In accordance with Fig. 1, the payoff structure can completely change the impact of varying ϕ and χ on ZD’s performance (cf. Fig. 2a and d, cf. 2b and e). This is one of the novel insights stemming from the present study, complementing the prior finding that ZD strategies are disfavored in population dynamics settings (14, 19, 20). Altogether, these results are key to improving our understanding of previously unforeseen limitations of ZD strategies in head-to-head matches in IPD games.
When an individual is knowingly confronted with extortion and especially has known the limitations of ZD strategies (Figs. 1 and 2), should this player be subdued or otherwise unbending? Prior work demonstrates that if an individual accedes by fully cooperating with an extortioner who fixes their strategies, both their payoffs are maximized (Fig. 1a and b). Conversely, here we ask whether there exist unbending players who choose to fix their strategies such that extortioners could maximize their payoffs only if they try to be fair by letting . Otherwise, extortioners would have experienced a decline in their average payoffs if they ever demanded an unequal share by increasing χ.
Motivated by these, we further explore unbending strategies that are able to force adapting ZD strategies, among these least generous ones with and hence equipped with the greatest level of extortion ability, to offer a fair split by letting in their own interest and guarantee equal pay for both sides. Considering that any ZD player can always modulate their hidden parameter ϕ to extreme values to favor their gains in the interactions (Fig. 2), we suppose unbending strategies, without loss of generality, will need to (i) neutralize the parameter ϕ in the first place such that both of their average payoffs are independent of ϕ, , and (ii) guarantee that the derivative of with respect to χ is strictly negative, .
These required properties of unbending strategies lead us to search and identify general classes of strategy candidates that can counteract the adversary imposed by extortioners, provided that they can trigger the backfire of being extortionate. To put it simply, when confronted with a fixed unbending player, any extortionate ZD player is disciplined with payoff reductions in the way that a higher degree of extortion leads to a smaller average payoff. Here, we simplify the interaction process by assuming targeted interactions between an unbending player and an extortionate ZD coplayer, without requiring the recognition and assessment of the possibility of coplayer’s extortion as discussed in Ref. (26). However, we later relax this assumption during our study of steering learning dynamics of adaptive players to consider more general strategies beyond extortionate ZD.
Thus, a potential candidate of unbending strategies outlearning any extortionate ZD coplayer needs to mitigate the impact of χ and ϕ, which are unilaterally controlled by the extortioner. To this end, we find that four classes of unbending strategies that can make their average payoffs independent of ϕ (as detailed in the Online Supplementary Material):
Class A: | and | Class B: | |
Class C: | Class D: |
Here in class D, , which is exactly the same linear relation satisfied by any ZD strategy. Class B only exists when and the maximum payoffs for both sides can receive is R (which is an equal-pay outcome). Class C contains “willing” on the boundary (27), against which an extortioner can only maximize their own payoffs by being fair (), thereby ensuring equal payoffs () with unbending “willing” for , and (Table S12 in the Online Supplementary Material). The complete analysis and discussion of these two classes B and C can be found in the Online Supplementary Material.
Furthermore, the entire strategy space comprised of all admissible unbending strategies can be characterized by requiring the derivative (Fig. 3). Again, the sign of determines the geometry of the strategy space satisfying unbending properties (see Fig. 3a and b for class A, Fig. 3c and d for class D). Of particular interest, the memory-one particle swarm optimization (PSO) Gambler , an optimized strategy using PSO algorithms in IPD games with the conventional payoff values (13), belongs to class A of unbending strategies (Fig. 3a), and WSLS is an unbending strategy only if (Fig. 3b).
Fig. 3.
Revealing strategies that are unbending to extortioners in IPD games. Shown is the strategy space of unbending players that are able to cause the monotonic decrease of an extortionate ZD (parameterized with , namely, the least generous type regardless of the sign ) player’s payoff with respect to the extortion factor χ. Extortioners can demand an even more unfair share by unilaterally raising their extortion factor χ. However, an unexpected drop in their prospective payoffs if intentionally being more extortionate is likely to compel self-interested extortioners who want to maximize their payoffs to be fair. In this sense, unbending strategies can be used to steer their coplayers from extortion to fairness. The strategy space of unbending players depends on the sign of , and we show two general classes of interest (see Online Supplementary Material for the complete classification): one class has the form with combinations of and shown in a) and b), and another class has the form where . This latter class in fact contains all ZD strategies that enforce a linear payoff relation with . Particular examples of unbending strategies include a) the memory-one PSO Gambler which is optimized by using particle swarm algorithms, b) WSLS, and c, d) all ZD strategies with . The dashed lines show the altered boundary of unbending strategies against the ZD player X using as opposed to ; the region of unbending strategies for class A () remains unchanged as shown in panel (b). The shaded area in b) shows the region where ZD player X, even though using minimal , can be outperformed by unbending strategies (see Table S9 in the Online Supplementary Material for details). This result is in line with Fig. 1, which shows that the payoff structure drastically hinders a ZD player’s ability to extort and dominate their coplayers, let alone those unbending ones. Parameters: a, b) , , , , ; c, d) , , , , .
Interestingly and coincidentally, we find that all ZD strategies with and are unbending to extortionate ZD (Fig. 3c and d). It is worth noting that these planes specifying the boundary of class D have particular meanings. As shown in Fig. 3c, the shaded triangle ADE represents the set of extortionate ZD strategies with and , and the shaded area by the four-sided polygon BCDE represents the set of equalizer strategies, and all unbending strategies in class D are in between these two planes and bounded by the unit cube. Besides, the triangle ACD represents the set of generous ZD strategies with , and the triangle ABD represents the set of ZD strategies with . For (Fig. 3d), the strategy space degenerates into the region between the triangle ABD (extortionate ZD) and triangle BCD (equalizer). Hence, we conclude that class D contains all ZD strategies with .
We also have extended our search of fixed unbending strategies with respect to an even broader class of ZD strategies just with positive (namely, using the baseline payoff and still having extortion ability to some degree as shown in Fig. 1c and d), such that ZD’s payoff is independent of the normalization factor ϕ and monotonically decreases with their extortion factor χ. As shown in Fig. 3 (highlighted with dashed lines) and the Online Supplementary Material, our classification of unbending strategies (especially nontrivial classes A and D) remains largely robust with respect to this important extension. Unexpectedly, we also find a set of nonlinear memory-one (non-ZD) strategies, when having the knowledge of the ZD coplayer’s baseline payoff O, will always be able to ensure equal payoffs O for both (see Fig. S21 and Table S27 in the Online Supplementary Material for details).
To provide further intuition about why extortion against unbending players fails to yield better average payoffs, we consider the case where an extortioner X with plays against a fixed generous ZD player Y with which in fact belongs to class D of unbending strategies. Both of their resulting payoffs are independent of their ϕ values, and the extortioner X has an average payoff given by
(4) |
We see that is monotonically decreasing with χ, as the derivative for , , and (Fig. 4a).
Fig. 4.
Intuition for how unfair demand can backfire on extortioners. As shown in panel a), despite being able to enforce payoff control against generous ZD (with ), the prospective payoff of extortionate ZD monotonically decreases with their extortion factor χ. The unfair extortion backfires on ZD which intended to demand a higher proportion but ended up with less payoff than what they would have obtained if being fairer otherwise. Even more, extortioners are outperformed by WSLS, as shown in panel b) when . The payoffs of the extortioner and WSLS, and , are both less than P, but intended extortion inflicts the unprecedented opposite outcomes: extortioner suffers more, whereas WSLS gains more. Targeted at extortioners, unbending strategies can be used to foster fairness in IPD games. The inset plots in a) and b) show the zoomed-in view of the scatter plot of payoffs, in a fashion similar to Fig. 1a and b, with their arrows to indicate their directions of change starting from fair, equal split as ZD increases χ above one. Parameters: a) , , , , b) , , , .
Geometrically visualizing this specific example, the generous ZD player Y enforces a linear payoff relation as , whereas the extortioner X enforces , and the resulting payoff pair lies in the intersection of these two straight lines. If the extortioner X increases the extortion factor χ, the intersection point will move down along the line of (if the generous ZD player Y remains unchanged). Therefore, the more unfair demand towards a fixed generous ZD player, the less payoff extortion yields. This previously unforeseen “backfire” is self-inflicted by the attempt to extort. For a self-interested individual who cares about how much they get, not just about monopolizing control of relative payoff, it does not pay to extort a generous ZD coplayer, and unfair demand backfires on extortioners who would have received the maximum R if trying to be fair by setting (Fig. 4a).
We now turn to explain the intuition behind the payoff structure of IPD games that can impact the dominance (optimality) of ZD strategies. It is well known that the condition is needed for mutual cooperation to fare better than alternating C and D pairs in the IPD. Yet another condition comes into sight if one ponders the condition under which the average payoff of any IPD strategy cannot be worse than P, the payoff for ending up with the deadlock of mutual defection. IPD is typically studied using the conventional values , , , , satisfying , and thus it ensures the average payoff of any IPD strategy cannot be less than P. Extortionate ZD players attain payoff control and extortion as desired in this scenario using the conventional payoff values (Fig. 1a and c), but the tide will turn against extortioners if the payoff structure satisfies . In this latter case, the average payoff of extortionate ZD strategies can be lower than P when facing off certain IPD strategies (Fig. 1b and d).
As ZD strategies are explicitly dependent on the underlying payoff matrix whose elements are , we discover that the particular payoff structure, which is governed by the sign of , can fundamentally change the dominance of extortionate ZD strategies (Fig. 1, also see Tables S8–S16 in the Online Supplementary Material). For example, when an extortionate ZD player is pitted against WSLS with , the stationary distribution of pairwise outcomes is, up to a positive normalization factor, given by
(5) |
Therefore, in order to gain an advantage, extortion ZD must ensure . However, this condition cannot always be satisfied when (the shaded region in Fig. 3b, see Tables S9 and S13 in the Online Supplementary Material for details). On the contrary, the extortionate ZD player in fact reciprocates unilateral cooperation more frequently than WSLS if holds, which is equivalent to requiring . Under this payoff structure condition, WSLS outperforms any extortionate ZD player (Fig. 4b); the more greedy extortion, the more ZD loses. Noteworthy, there is absolutely no mutual cooperation between WSLS and extortionate ZD players. Extortionate ZD does not fully cooperate after a mutual cooperation move, and thus ZD and WSLS will eventually end up with mutual defection from which ZD will never respond with cooperation while WSLS will always respond with cooperation; they will never be back to mutual cooperation. As a consequence, in the long run, no mutual cooperation between them can be established at all.
To further understand fixed unbending player’s unprecedented steering role in enforcing fairness and cooperation, we consider adaptive learning dynamics of a focal player X using a much broader space of strategies, rather than being limited to extortionate ZD, in a donation game which is a simplified PD (14, 19) (see Tables S17 and S18 in the Online Supplementary Material). Under this donation game satisfying the “equal gains from switching” (i.e. ), the memory-one reactive strategies is actually a subset of ZD strategies (14) (also see Fig. S19 in the Online Supplementary Material). The shaded triangle in Fig. 5 indicates all such ZD strategies with positive : BA represents extortionate ZD with ; point A is TFT with and ; point B is equalizer with and ; point C is GTFT with and ; BC represents the class of “equalizer” strategies. We find that if the benefit-to-cost ratio (“golden ratio”), the cooperative edge is guaranteed to have the maximum average payoff value for player X when interacting with any fixed unbending player from class A (the region highlighted in Fig. 3a). Depending on the specific strategy of the unbending player Y from class A (see Fig. S20 and Table S23 for more details in the Online Supplementary Material), there could exist bistable learning outcomes of X’s final strategies: X can converge to the all defection corner or otherwise to the cooperative edge (Fig. 5a), but there is a subset of class A of unbending strategies that ensures the global convergence to the cooperative edge (Fig. 5b, and Fig. S20, Tables S19–S22 in the Online Supplementary Material).
Fig. 5.
Steering learning dynamics towards fairness and cooperation with unbending strategies. Shown are the stream plots (vector fields) from the adaptive learning dynamics of a self-interested focal player X who uses a general reactive strategy against a fixed unbending coplayer Y: a, b) from class A and c, d) from class D. Class A of unbending strategies are able to steer their coplayer X ultimately to behave like GTFT. Panel a) demonstrates that, depending on the specific unbending strategy player Y uses, the direction of change of player X’s can exhibit bistability (separated by the dashed line in (a)), which further depends on the initial state of X’s strategy. Panel b) shows that there exists a subset of class A that is able to direct the change of always towards full cooperation. In a, b), on the edge , the direction of change of is neutral, and the line segment of AC indicates the set of compliers; on the line segment of the edge , indicated by BA, which represents a subset of extortionate ZD players (extortioners), the learning dynamics of the ZD player X always converges to TFT . Panels c) and d) show that class D of unbending strategies are able to steer adaptive learning dynamics of X globally to the cooperative edge on which the direction of change of c) either is increasing if Y uses a strategy from class D that is an intermediate ZD with ) d) or remains neutral if Y uses a strategy from class D that happens to be the generous ZD with . In both a) and b), the global maximum for X’s payoff is reached at the edge if , and in c) only at does the global payoff maximum for X occur whereas so does the entire cooperative edge in d). The shaded triangle ABC in each panel indicates the part of reactive strategies which belong to the subset of general ZD strategies with positive χ. The color of the curves and dots corresponds to the payoff values of X in situ, as specified by the given colorbar. The PD game is parameterized using a donation game , , , with and .
The steered learning dynamics under the influence of the coplayer Y from class D of unbending strategies is shown in Fig. 5c and d (also see Table S24 in the Online Supplementary Material). Generally speaking, the learning dynamics of a focal player X against class D of unbending strategies (in other words, ZD strategies with a higher level of generosity than player X) adds useful insights by complementing previous results in Ref. (8) that focuses on adapting coplayer Y against an extortionate ZD player X. Here, we show that the final strategy of player X converges to the cooperative edge, reaching full cooperation if against an unbending strategy from class D with (still more generous than any extortionate ZD strategies on the edge BA with ) (Fig. 5c) or remains neutral on the cooperative edge once reaching there when against an unbending strategy from class D with (generous ZD) (Fig. 5d).
In the Online Supplementary Material, we study the corresponding learning dynamics of a general ZD player within the parameter space (see Tables S25 and S26 in the Online Supplementary Material) and confirm qualitatively similar results as reported here. In particular, when against a fixed unbending player, an adapting extortioner with the intended extortion factor χ unexpectedly suffers greater payoff reductions than their counterpart who chooses not to accede unless offering a fair split. For this reason, any evolutionary extortioner who aspires to maximize their own payoff will be compelled from extortion to fairness by adjusting their χ values. Since there is no interference by the parameter ϕ as , such reactive learning dynamics of extortioners is governed solely by the evolution of χ towards payoff optimization:
(6) |
where the properly chosen timescale parameter τ guarantees that the state of Markov chains of gameplay reaches equilibrium faster than the learning dynamics (see Online Supplementary Material).
Thus, a self-interested extortioner tends to adjust χ as small as possible and ultimately behaves like TFT by letting , thereby guaranteeing equal payoffs for both parties (see the change of direction on the edge of in Tables S25 and S26 in the Online Supplementary Material). In evolving populations, natural selection favors generosity over extortion (19), and in head-to-head matches as demonstrated here, players with the knowledge of unbending strategies can outlearn extortioners and foster fairness and reciprocity in dyadic interactions.
Noteworthy, in the aforementioned steering learning dynamics, we have focused on an adaptive payoff-maximizing player against a fixed unbending coplayer in various scenarios that directly complement the original study by Press and Dyson, where they assume an evolutionary adaptive player against a fixed extortionate ZD coplayer (8). In previous experiments with human subjects, it was observed that while some players refuse to comply with unfair demands to discourage extortionate behavior, some players who consistently refuse to be extorted may eventually give up on punishing extortionists (10, 21). This situation can arise when the extortionists are known to be preprogrammed computer agents (21) or when they are incentivized to win an advantage (10), making it challenging to discipline them effectively. To address this tug-of-war situation in the adaptive dynamics of behavior response, we introduce a relative time scale ω that governs the time evolution of the behavioral change of an unbending player as compared to their coplayer. Specifically, the coadaptive dynamics between a ZD player X () and an unbending player Y from class A (starting from a prescribed ) under the previous donation games can be described by the following system of differential equations:
(7) |
As ω approaches 0, the dynamics revert back to the original scenario we studied (Fig. 5), where an unbending player is fixed in their behavior. Conversely, when ω approaches 1, the dynamics converge to the scenario studied by Press and Dyson, which features a fixed ZD player. For intermediate values of ω, an interesting arms race emerges between the two adaptive players, which is similar to the Red Queen dynamics. Namely, adapting quickly means being responsive to instant payoff improvement, but this may lead to unexpected shifts in the game where one player turns out to have more influence than the other in the long run. Our theoretical findings are supported by an analysis of the extended coadaptive dynamics between a ZD player and their unbending coplayer from class A or class D. Our results show that unbending players, even when their learning rates are high, can enforce fairness and cooperation in pairwise interactions as long as they retain some degree of unbending characteristics, as demonstrated in Fig. S22 in the Online Supplementary Material.
Taken together, these results suggest that unbending strategies cannot just outlearn self-interested extortionate ZD and force them to be fair and cooperative (that is, both parties eventually get equal payoffs out of mutual cooperation, Figs. S22a and b, S22d and e in the Online Supplementary Material), but also steer the evolution of TFT-like strategies out of any focal player using a much broader strategy space (represented by the unit square in Fig. 5) including but not necessarily limited to extortionate ZD strategies (namely, the edge BA in Fig. 5).
So far, we have focused on characterizing properties of unbending strategies and demonstrating their steering role in enforcing fairness and cooperation in pairwise interactions that only involve two parties. It is equally, if not less, worthy of studying the evolutionary dynamics of unbending strategies in stochastic population dynamics together with a set of other prescribed IPD strategies. As shown in Fig. 6, we demonstrate the evolutionary advantage of unbending strategies and their stability in stochastic dynamics of invasion and fixation under the limit of rare mutations (28). We see that unbending strategies, including class A (i.e. the PSO Gambler) and class D (i.e. generous ZD and TFT), are favored by natural selection; their abundance is greater than the population average, which holds from weak selection () to strong selection (). Noticeably, the abundance of extortionate ZD almost vanishes under strong selection (see Fig. 6a). In pairwise competition dynamics (arising from the limit of rare mutations such that the system has at most two IPD strategies simultaneously present in the population), the PSO Gambler, generous ZD, and TFT all have an evolutionary advantage over extortionate ZD (see Fig. 6b). Namely, the fixation probability of an unbending strategy is greater than that of an extortionate ZD (see Online Supplementary Material for details).
Fig. 6.
Evolutionary dynamics of unbending strategies in finite populations. Shown are a) the stationary abundance of IPD strategies (as indicated on the x-axis) under the limit of rare mutations and for different selection strengths β and b) the evolutionary pathways in pairwise competition dynamics (the direction of the arrows indicates dominance where the IPD strategy at the end is favored over that at the start; double arrows indicate neutral evolution). We consider a set of prescribed IPD strategies, including members belonging to unbending strategies from class A (the PSO Gambler) and from class D (generous ZD with ). TFT is an extreme boundary case of unbending strategies from class D (cf. point A in Fig. 3c). We use the Moran process for evolutionary updating and study the long-term mutation-selection equilibrium. Parameters: population size , mutation rate , selection strength , 0.1, 1, payoff values: , , , .
Nevertheless, we note that the general picture depicted here could qualitatively change for extremely small population size N (see Online Supplementary Material, and Fig. S23). In fact, extortionate ZD can be favored over unbending strategies (UB), such as the PSO Gambler, if . To see this, let us express the average payoff matrix for their game interactions as follows:
(8) |
In a population of size N, UB is favored over ZD if and only if
(9) |
This condition holds for any selection strength and for any mutation rate (29). As for the PSO Gambler vs an extortionate ZD (with ) using conventional payoff values, we have , , (ZD enforcing a linear relation, ), and . For , ZD completely dominates UB as . Only for population size (i.e. ) is it possible for natural selection to favor UB over ZD.
Moreover, the presence of noise can have an impact on the evolutionary performance of IPD strategies, as demonstrated by the susceptibility of TFT against noise (4). In the Online Supplementary Material, we quantify how the level of implementation errors impacts the ability to foster mutual cooperation among unbending strategies themselves (see Online Supplementary Material for further details, Fig. S24). Like WSLS, unbending strategies, such as the PSO Gambler, are robust against noise; their mutual cooperation level is impacted only as . Altogether, our results demonstrate that previously unforeseen unbending properties actually exist in some common IPD strategies and that they can be leveraged to foster fairness and cooperation not only in pairwise interactions but also in population dynamics settings.
Discussions
It is thought that an evolutionary (adapting) player should be subdued to a fixed extortionate ZD player by fully cooperating as the best response (8). In contrast, recent experimental evidence suggests that human players often choose not to accede to extortion out of concern for fairness (10, 15). Inspired by this empirical finding, here we show that there exist general classes of unbending strategies such that the best response of any payoff-maximizing extortioner against a fixed unbending player is to be fair, thereby ensuring equal pay for both parties. From this perspective, the witting of unbending strategies has effectively turned the opponent’s choices of whether or not to adopt extortionate ZD strategy into an Ultimatum game (30): to demand unfair division via unilaterally setting a large χ value, or to guarantee fair share by letting . In the former, the extortion effort is sabotaged by unbending, and both sides will be hurt, whereas in the latter, both sides will get an equal split of the payoffs. Our results demonstrate that unbending strategies can be used to rein in seemingly formidable extortionate ZD players, whose fair offer ultimately can be cultivated in their own interest.
In light of unbending strategies, there is no guarantee that extortioners will be able to subdue their opponents with certainty and get their own way as desired. Extortion cannot be successful unless their coplayers give up resistance in the first place. The extortion ZD exerts on the coplayer can backfire on themselves. For example, an extortionate ZD player will not be able to rein in TFT-like players (a limiting case belonging to class D of unbending strategies) who are fair-minded but willing to punish defection by responding with defection. They will end up in a tie both receiving P (Fig. 2). Even if an extortionate ZD player X does end up with dominance over the coplayer Y, namely, , a higher ratio of relative payoff surplus, , does not necessarily translate to higher actual payoffs (Fig. 2a). Increasing χ appears to put ZD in a more advantageous position, but such unfair demand would be pushed back by unbending players such as generous ZD and TFT-like players and hence does not always yield higher actual payoffs (Fig. 4a). As recently demonstrated in experiments involving human players against fixed machine extortioners (15), human players respond to more extortionate ZD players with much lower cooperation levels, which are in essence passive punishment measures to counter ZD’s intended extortion.
Unbending behavior is related to the concept of fairness, which has been extensively studied in economics and game theory, particularly through the Ultimatum game (31–34). In such games, individuals may refuse to make concessions they perceive as unfair, even if it would be rational to do so. Third-party mechanisms such as punishment (35) and reputation effects (30) are often necessary to enforce cooperative behavior or fairness. However, in repeated games like the PD, individuals can employ a variety of strategies to balance their expectations of fair play and reinforce niceness, which can lead to reciprocal fairness and cooperation. The unbending strategies we discovered may be part of the repertoires used to counteract the ZD’s extortion and foster mutual cooperation and cultivate fairness, given the ubiquity of unbending properties in some common IPD strategies (Fig. 3).
Our work highlights the importance of payoff structure in determining the optimality of ZD strategies in IPD games (Fig. 1). In particular, if the condition holds, which means the total payoff, , from alternating C and D pairs of two players is worse than that of their mutual defection, , extortionate ZD players can be outperformed (Fig. 1b and d). This surprising finding is an important new insight that stems from the present study. Moreover, the sign of qualitatively determines the admissible strategy space of unbending players that can cause the backfire on extortioners (Fig. 3). Noticeably, only if is WSLS an unbending strategy, and in this case, WSLS dominates any extortionate ZD strategy (Fig. 4b).
The payoff condition implies a more adversarial nature in pairwise interactions than the conventional IPD games where typically holds (24). Intuitively, this means that the best response for a pair of individuals alternating between and is always to switch to mutual defection (cf. Fig. 1a and b). Unbending strategies (those highlighted in the shaded area in Fig. 3b) can outcompete seemingly invincible extortionate ZD players who would have the greatest potential to dominate by using the minimal and (Fig. 1). As aforementioned, since ZD’s strategy is parameterized explicitly by the underlying payoff matrix, variations in the payoff structure can have a previously unforeseen effect that will turn the tables on ZD (Fig. 4b): an extortionate ZD may become a victim of their own success in IPD games satisfying and more broadly, in social dilemmas of more adversarial nature as discussed in Ref. (24).
In the presence of errors and noises (36, 37), complex strategies informed by longer memory of past moves are likely at an advantage against simple memory-one strategies. Beyond pairwise interactions, higher order ones in multiperson games (38), such as the public goods game, as well as asymmetric interactions (39), are also of relevance in studying reciprocity in these generalized situations. Extensions incorporating these considerations are meaningful, but incur computational and theoretical challenges in search of robust optimal strategies. Nevertheless, the recent breakthrough in reinforcement (deep) learning of zero-sum games (40), like the Go (41), can lend some insight into the study of nonzero-sum games where learning agents, despite being self-serving, can mutually foster cooperation for the greater good under certain conditions (42). Thus, the classic framework of IPD still has the potential to be used as a primary testbed for synergistically combining artificial intelligence (AI) and game theory in future work (13, 43, 44), all with an eye towards helping us to enhance global cooperation in many challenging issues confronting our common humanity (45).
While the theoretical and modeling insights of unbending strategies are enlightening, we would like to briefly discuss the limitations of their practical success. First, when individuals encounter unknown coplayers with limited prior information, accurately detecting and assessing potential extortion during repeated interactions may prove challenging due to cognitive constraints. This presents an obstacle to effectively countering extortion using unbending strategies in real time. Second, even when unbending players are made fully aware of the extortion, they may give up their resistance efforts all too quickly, as it could be more beneficial for them to do so, according to recent empirical findings that show extortion prevails under incentive (10) or power (46, 47) asymmetries (see Ref. (48) for a mini review). Moving forward, we hope the present study will help stimulate future studies, both empirical and theoretical, to assess the efficacy of unbending strategies in more realistic scenarios, such as those involving unknown coplayers or short-term incentives that encourage self-interest.
In summary, we have found and characterized general classes of unbending strategies that are fair-minded and can outlearn extortioners in their head-to-head encounters. When an extortionate ZD player attempts to demand an unfair greater share from an unbending player who instead uses a fixed strategy, the unbending player is able to restrain the extortioner from profiting more. The intent to extort an unbending player has unprecedented consequences: extortioners would fare worse than if being fairer, and they can even be outperformed by, for example, WSLS, if the payoff matrix satisfies . Such previously unforeseen backfires caused by unbending players can steer reactive learning dynamics of extortionate ZD players from extortion to fairness. Our work offers novel insights into fostering fairness and suppressing extortion for a more equitable and just society.
Model and methods
Model and analytical approach
We use the same analytical approach invented by Press and Dyson (8) to calculate the expected payoffs of any two given players that are head-to-head in the IPD games. In this work, we focus on revealing strategies that are unbending to extortionate ZD players using explicit closed-form solutions (see details in the Online Supplementary Material). The ZD strategies are usually parameterized by three important parameters, the extortion factor χ, the normalization factor ϕ, plus an additional baseline payoff which controls the level of generosity (8, 19). Tuning the parameter ϕ of extortionate ZD strategies with and does not affect the linear payoff relation , but will impact the dependence of their own average payoffs on the extortion factor χ in a nontrivial way (see Fig. 2). Therefore, we restrict our search for unbending strategies that can neutralize the impact of this parameter ϕ, that is, we find specific classes of strategies that are able to render the independence of their payoffs on ϕ. Further, we narrow down the search of unbending strategies that can cause the “backfire” of extortion, namely, the expected payoffs of extortionate ZD strategies against a fixed unbending player are monotonically decreasing with χ. Ultimately, these considerations lead us to discover multiple general classes of unbending strategies, against which attempt to extort and dominate, if any, does not pay off at all for ZD players using (including but not limited to extortionate ZD). In some cases, extortionate ZD strategies can even be outperformed by unbending coplayers, if the payoff matrix satisfies (Fig. 3b). We also investigate how fixed unbending players can steer the learning dynamics of their adapting coplayers who use a much broader range of memory-one strategies beyond the class of extortionate ZD towards fairness and cooperation. We detail our comprehensive analysis in the Online Supplementary Material (see Figs. S1–S24 and Tables S1–S21).
Supplementary Material
Acknowledgments
We would like to thank three anonymous referees for their constructive comments which helped to improve our work. X.C. gratefully acknowledges the generous faculty startup fund support by BUPT. F.F. is supported by the Bill & Melinda Gates Foundation (award no. OPP1217336), the NIH COBRE Program (grant no. 1P20GM130454), a Neukom CompX Faculty Grant, the Dartmouth Faculty Startup Fund, and the Walter & Constance Burke Research Initiation Award.
Contributor Information
Xingru Chen, School of Science, Beijing University of Posts and Telecommunications, Beijing 100876, China; Department of Mathematics, Dartmouth College, Hanover, 03755 NH, USA.
Feng Fu, Department of Mathematics, Dartmouth College, Hanover, 03755 NH, USA; Department of Biomedical Data Science, Geisel School of Medicine at Dartmouth, Lebanon, 03756 NH, USA.
Supplementary material
Supplementary material is available at PNAS Nexus online.
Funding
This work is supported in part by funds from the Bill & Melinda Gates Foundation (award no. OPP1217336) and the Neukom CompX Faculty Grant.
Author contributions
X.C. and F.F. conceived the study, conducted the research, analyzed the results, and wrote and reviewed the manuscript.
Previous presentation
These results were not previously presented.
Preprints
A preprint of this article is published at arXiv:2201.04198.
Data Availability
Our Python Jupyter notebooks that can be used to reproduce results reported in this work are available at GitHub: https://github.com/fufeng/unbending.
References
- 1. Axelrod R, Hamilton WD. 1981. The evolution of cooperation. Science. 211(4489):1390–1396. [DOI] [PubMed] [Google Scholar]
- 2. Rapoport A, Chammah AM, Orwant CJ. 1965. Prisoner’s Dilemma: a study in conflict and cooperation. Vol. 165. Ann Arbor: University of Michigan Press. [Google Scholar]
- 3. Hilbe C, Chatterjee K, Nowak MA. 2018. Partners and rivals in direct reciprocity. Nat Hum Behav. 2(7):469–477. [DOI] [PubMed] [Google Scholar]
- 4. Nowak MA, Sigmund K. 1992. Tit for tat in heterogeneous populations. Nature. 355(6357):250–253. [Google Scholar]
- 5. Macy MW, Flache A. 2002. Learning dynamics in social dilemmas. Proc Natl Acad Sci USA. 99(Suppl 3):7229–7236. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Nowak M, Sigmund K. 1993. A strategy of win-stay, lose-shift that outperforms tit-for-tat in the Prisoner’s Dilemma game. Nature. 364(6432):56–58. [DOI] [PubMed] [Google Scholar]
- 7. Boerlijst MC, Nowak MA, Sigmund K. 1997. Equal pay for all prisoners. Am Math Mon. 104(4):303–305. [Google Scholar]
- 8. Press WH, Dyson FJ. 2012. Iterated Prisoner’s Dilemma contains strategies that dominate any evolutionary opponent. Proc Natl Acad Sci USA. 109(26):10409–10413. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Akin E. 2015. What you gotta know to play good in the Iterated Prisoner’s Dilemma. Games. 6(3):175–190. [Google Scholar]
- 10. Becks L, Milinski M. 2019. Extortion strategies resist disciplining when higher competitiveness is rewarded with extra gain. Nat Commun. 10(1):1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Chen J, Zinger A. 2014. The robustness of zero-determinant strategies in Iterated Prisoner’s Dilemma games. J Theor Biol. 357:46–54. [DOI] [PubMed] [Google Scholar]
- 12. Hao D, Rong Z, Zhou T. 2015. Extortion under uncertainty: zero-determinant strategies in noisy games. Phys Rev E. 91(5):052803. [DOI] [PubMed] [Google Scholar]
- 13. Harper M, Campbell O. 2017. Reinforcement learning produces dominant strategies for the Iterated Prisoner’s Dilemma. PLoS ONE. 12(12):e0188046. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Hilbe C, Nowak MA, Sigmund K. 2013. Evolution of extortion in Iterated Prisoner’s Dilemma games. Proc Natl Acad Sci USA. 110(17):6913–6918. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Hilbe C, Röhl T, Milinski M. 2014. Extortion subdues human players but is finally punished in the Prisoner’s Dilemma. Nat Commun. 5(1):1–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Ichinose G, Masuda N. 2018. Zero-determinant strategies in finitely repeated games. J Theor Biol. 438:61–77. [DOI] [PubMed] [Google Scholar]
- 17. McAvoy A, Hauert C. 2016. Autocratic strategies for iterated games with arbitrary action spaces. Proc Natl Acad Sci USA. 113(13):3573–3578. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Szolnoki A, Perc M. 2014. Defection and extortion as unexpected catalysts of unconditional cooperation in structured populations. Sci Rep. 4(1):1–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Stewart AJ, Plotkin JB. 2013. From extortion to generosity, evolution in the Iterated Prisoner’s Dilemma. Proc Natl Acad Sci USA. 110(38):15348–15353. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Adami C, Hintze A. 2013. Evolutionary instability of zero-determinant strategies demonstrates that winning is not everything. Nat Commun. 4(1):1–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Wang Z, Zhou Y, Lien JW, Zheng J, Xu B. 2016. Extortion can outperform generosity in the Iterated Prisoner’s Dilemma. Nat Commun. 7(1):11125. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Weitz JS, Eksin C, Paarporn K, Brown SP, Ratcliff WC. 2016. An oscillating tragedy of the commons in replicator dynamics with game-environment feedback. Proc Natl Acad Sci USA. 113(47):E7518–E7525. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Nowak M, Sigmund K. 1990. The evolution of stochastic strategies in the Prisoner’s Dilemma. Acta Appl Math. 20(3):247–265. [Google Scholar]
- 24. D’Orsogna MR, Perc M. 2015. Statistical physics of crime: a review. Phys Life Rev. 12:1–21. [DOI] [PubMed] [Google Scholar]
- 25. Bielefeld RS. 1988. Reexamination of the perfectness concept for equilibrium points in extensive games. In: Models of strategic rationality. Dordrecht: Springer. p. 1–31. [Google Scholar]
- 26. Knight VA, Harper M, Glynatsi NE, Gillard J. 2019. Recognising and evaluating the effectiveness of extortion in the Iterated Prisoner’s Dilemma, arXiv, arXiv:1904.00973, preprint: not peer reviewed. [DOI] [PMC free article] [PubMed]
- 27. Van den Berg P, Weissing FJ. 2015. The importance of mechanisms for the evolution of cooperation. Proc R Soc B: Biol Sci. 282(1813):20151382. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Fudenberg D, Imhof LA. 2006. Imitation processes with small mutations. J Econ Theory. 131(1):251–262. [Google Scholar]
- 29. Antal T, Nowak MA, Traulsen A. 2009. Strategy abundance in games for arbitrary mutation rates. J Theor Biol. 257(2):340–344. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Nowak MA, Page KM, Sigmund K. 2000. Fairness versus reason in the ultimatum game. Science. 289(5485):1773–1775. [DOI] [PubMed] [Google Scholar]
- 31. Fehr E, Schmidt KM. 1999. A theory of fairness, competition, and cooperation. Q J Econ. 114(3):817–868. [Google Scholar]
- 32. Kahneman D, Knetsch JL, Thaler RH. 1986. Fairness and the assumptions of economics. J Bus. 59:S285–S300. [Google Scholar]
- 33. Rabin M. 1993. Incorporating fairness into game theory and economics. Am Econ Rev. 83:1281–1302. [Google Scholar]
- 34. Suleiman R. 1996. Expectations and fairness in a modified ultimatum game. J Econ Psychol. 17(5):531–554. [Google Scholar]
- 35. Fehr E, Gächter S. 2002. Altruistic punishment in humans. Nature. 415(6868):137–140. [DOI] [PubMed] [Google Scholar]
- 36. Axelrod R. 2012. Launching “the evolution of cooperation”. J Theor Biol. 299:21–24. [DOI] [PubMed] [Google Scholar]
- 37. Fudenberg D, Maskin E. 1990. Evolution and cooperation in noisy repeated games. Am Econ Rev. 80(2):274–279. [Google Scholar]
- 38. Hilbe C, Wu B, Traulsen A, Nowak MA. 2014. Cooperation and control in multiplayer social dilemmas. Proc Natl Acad Sci USA. 111(46):16425–16430. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Su Q, Allen B, Plotkin JB. 2022. Evolution of cooperation with asymmetric social interactions. Proc Natl Acad Sci USA. 119(1):e2113468118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Balduzzi D, et al. 2019. Open-ended learning in symmetric zero-sum games. In: International Conference on Machine Learning. PMLR. p. 434–443.
- 41. Silver D, et al. 2017. Mastering the game of go without human knowledge. Nature. 550(7676):354–359. [DOI] [PubMed] [Google Scholar]
- 42. Barfuss W, Donges JF, Vasconcelos VV, Kurths J, Levin SA. 2020. Caring for the future can turn tragedy into comedy for long-term collective action under risk of collapse. Proc Natl Acad Sci USA. 117(23):12915–12922. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. McAvoy A, Mori Y, Plotkin JB. 2022. Selfish optimization and collective learning in populations. Physica D. 439:133426. [Google Scholar]
- 44. Noordman CR, Vreeswijk GAW. 2019. Evolving novelty strategies for the Iterated Prisoner’s Dilemma in deceptive tournaments. Theor Comput Sci. 785:1–16. [Google Scholar]
- 45. Dafoe A, et al. 2021. Cooperative AI: machines must learn to find common ground. Nature. 593(7857):33–36. [DOI] [PubMed] [Google Scholar]
- 46. Hilbe C, Hagel K, Milinski M. 2016. Asymmetric power boosts extortion in an economic experiment. PLoS ONE. 11(10):e0163867. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47. Milinski M, Hilbe C, Semmann D, Sommerfeld R, Marotzke J. 2016. Humans choose representatives who enforce cooperation in social dilemmas through extortion. Nat Commun. 7(1):10915. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Milinski M. 2022. Extortion—a voracious prosocial strategy. Curr Opin Psychol. 44:196–201. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Our Python Jupyter notebooks that can be used to reproduce results reported in this work are available at GitHub: https://github.com/fufeng/unbending.