Skip to main content
Proceedings of the Royal Society B: Biological Sciences logoLink to Proceedings of the Royal Society B: Biological Sciences
. 2010 Aug 18;278(1704):371–377. doi: 10.1098/rspb.2010.1275

The evolution of punishment through reputation

Miguel dos Santos 1,2,*,, Daniel J Rankin 2,3,, Claus Wedekind 1
PMCID: PMC3013410  PMID: 20719773

Abstract

Punishment of non-cooperators has been observed to promote cooperation. Such punishment is an evolutionary puzzle because it is costly to the punisher while beneficial to others, for example, through increased social cohesion. Recent studies have concluded that punishing strategies usually pay less than some non-punishing strategies. These findings suggest that punishment could not have directly evolved to promote cooperation. However, while it is well established that reputation plays a key role in human cooperation, the simple threat from a reputation of being a punisher may not have been sufficiently explored yet in order to explain the evolution of costly punishment. Here, we first show analytically that punishment can lead to long-term benefits if it influences one's reputation and thereby makes the punisher more likely to receive help in future interactions. Then, in computer simulations, we incorporate up to 40 more complex strategies that use different kinds of reputations (e.g. from generous actions), or strategies that not only include punitive behaviours directed towards defectors but also towards cooperators for example. Our findings demonstrate that punishment can directly evolve through a simple reputation system. We conclude that reputation is crucial for the evolution of punishment by making a punisher more likely to receive help in future interactions, and that experiments investigating the beneficial effects of punishment in humans should include reputation as an explicit feature.

Keywords: game theory, punishment, cooperation, humans, experimental game

1. Introduction

Cooperation is often enhanced if non-cooperators can be punished [15], but this simple fact cannot yet explain the evolution of punishment, especially not if punishment inflicts immediate costs to the punisher. Indeed, in settings where individuals interact repeatedly with the same partner [68], or when third-party punishment is possible (i.e. punishing players for being unkind to others [9,10]), punishers usually finish with lower pay-offs than non-punishers (but see [5]). It has therefore been concluded that punishment is mostly maladaptive within the respective games, and that it may have evolved for other reasons than for promoting cooperation [610]. However, the maladaptive argument is not very satisfying [11] and, given the widespread prevalence of punishment, a significant direct evolutionary advantage of punishment is still likely, for example, in the context of reputation games [4].

Although it is well established that reputation can play a key role in social interactions (e.g. in the evolution of human cooperation [1221], the possible advantage from a reputation of being a punisher has not been sufficiently explored as an explanation for the evolution of punishment (if I punish you because you have defected against me, others may later not defect against me). Some theoretical studies have suggested that, under certain circumstances, natural selection favours strategies that take the likelihood of being punished into account [2227], for example, when information about a neighbours' behaviour is available [25], or in situations where individuals cooperate according to the average punishment strategy played by all social partners [26]. Thus, punishment could act as an indirect threat to observers. However, the majority of empirical studies so far have investigated punishment in anonymous settings where players did not have information on the others' punitive behaviours, or in settings where this information was confounded with information about the others' cooperative behaviours [2,3,6,8,10,2830].

Here, we explore the evolution of punishment in a helping game. In this game, a donor can either help, or refuse to help a receiver who then can punish in return [26,27,31]. Such punitive actions influence an individual's ‘punishment score’, which reflects how much an individual punished previously and can then be used by others to discriminate between punishers and non-punishers. Such a reputation system is analogous to the image scoring proposed for the evolution of indirect reciprocity [1416,18,21,32,33]. In a simple analytical model, we investigate the evolutionary stability of a strategy that discriminates between punishers and non-punishers. We then use computer simulations to extend our model and to test the robustness of the major outcome of the model.

2. Methods and results

(a). Analytical model

We first build an analytical model. Individuals interact randomly and can choose to either help or not help their social partners. Defectors, who never help, are denoted by frequency xj, cooperators, who always help, are denoted by frequency yj and discriminators, who only help individuals that punished in their last encounter, are denoted by frequency zj. The subscript j ∈ {N, P} denotes whether individuals are punishers (j = P), who punish upon not receiving help, or non-punishers (j = N), who never punish. We assume a simple form of scoring, where players have either a positive punishment score, or a negative punishment score. The parameter qij (where i ∈ {x, y, z} and j ∈ {N, P}) is the probability that an individual punished in their last encounter (and therefore has a positive punishment score) while 1 − qij is the probability that an individual did not punish in their last encounter (and therefore has a negative punishment score). For non-punishers, this probability is always zero. For punishers, it is contingent on them having experienced defection in their last encounter. Following replicator dynamics, and assuming an infinite population, the probability that a given individual punished at time τ + 1 can be calculated by the following recursion relations:

(a).

We assume that behavioural dynamics occur on a very fast time scale relative to evolutionary dynamics, and therefore the above recurrence relations will equilibrate very quickly to give the equilibrium punishment scores for each state, which are

(a).

Since each punishment score is identical (as it merely depends on the relative frequency of defectors and discriminators in the population), we will write Inline graphic We assume that cooperation imposes a cost c on an actor, and confers a benefit b on a receiver. Punishment imposes a cost s on the punisher, while inflicting a cost e on the individual being punished. The pay-off of each of the six strategies is denoted by gij (given in appendix A). The fitness of a given strategy is given by Inline graphic, where Inline graphic is the average pay-off in the population, such that

(a).

The dynamics of a given strategy are therefore given by kij(t + 1) = kij(t)wij, where kij is the frequency of the strategy in question. Table 1 shows a list of the symbols used in our model.

Table 1.

List of symbols.

symbol definition
xj defectors (with j ∈ {N, P})
yj cooperators (with j ∈ {N, P})
zj discriminators (with j ∈ {N, P})
iN non-punishing i (with i ∈ {x, y, z})
iP punishing i (with i ∈ {x, y, z})
gij pay-off of strategy ij
Inline graphic average pay-off in the population
wij fitness of strategy ij
c cost of helping
b benefit of receiving help
s cost of punishing
e cost of being punished
q* probability that an individual punished a defection in its last encounter
n population sizea
m number of interactions per individuala
μ mutation ratea
ɛ error ratea
ka punishment scorea (with a ∈ {1, 2, 3})
Is image scorea

aOnly used in the simulations.

We analyse the condition under which punishing discriminators (with a frequency zP) cannot be invaded by any other strategy, and is therefore an evolutionarily stable strategy (ESS). The condition under which punishing discriminators, zP, will be an ESS with respect to non-punishing defectors (i.e. xN → 0 and zP → 1) is if zP have a greater pay-off than xN (i.e. gzP > gxN), which is fulfilled if

(a).

under which they are also an ESS with respect to non-punishing cooperators (i.e. when yN → 0 and zP → 1). This condition is always respected since b > c and e > s. Punishing discriminators will also be an ESS with respect to punishing defectors (i.e. xP → 0 and zP → 1) if being punished is less costly than helping (e > c), and to non-punishing discriminators (i.e. zN → 0 and zP → 1) if the benefit of helping is greater than the cost of punishing (b > s). If the cost of helping is greater than the cost of being punished (c > e), punishing cooperators will be able to invade, and if this condition is held, punishing cooperators are an ESS with respect to all strategies (except non-punishing cooperators, to which they are neutral). This allows for non-punishing cooperators to invade through drift. If this occurs, any of the other four strategies (i.e. xN, xP, zN and zP) will be able to invade. However, punishment is frequency dependent, and, during an invasion of a population of non-punishing cooperators, punishing discriminators will be able to outcompete non-punishing defectors (i.e. gzP > gxN), which will occur if

(a).

which is independent of the frequency of non-punishing cooperators (yN). Thus, as punishing discriminators become more common, they are increasingly favoured over defectors. But how can punishment invade a population of non-punishing defectors (xN)? When common, xN will be an ESS with respect to all strategies except zN, who can invade through drift. Then, as soon as they are common in the population and that the benefit of help is greater than the cost of punishing (b > s), a single punisher can outweigh the costs of punishing by receiving help and zP can invade.

(b). Computer simulations

Our analytical model considers a simplified case, with large populations and a limited number of strategies. To test the robustness of our results and to put them into the wider context of cooperative strategies, we built an individual-based model. We modelled a population of finite size n. In each generation, pairs of players are randomly chosen to interact in the following manner: one player (the donor) has to decide whether to help the other one (the receiver) or not. Helping incurs a cost c to the donor and a benefit b to the receiver (where b > c). No help results in zero pay-off for both individuals. After the donor's decision, the receiver has the possibility to pay a cost s in order to punish the donor. Punishment reduces the donor's total pay-off by e (where e > s). In each generation, a player has on average m interactions, each of which could be in the donor or receiver role. The fitness of a player is given by its total number of points at the end of its m interactions. Individuals then leave offspring in proportion to their fitness. Mutations occur during reproduction and both the helping and the punishment strategies mutate independently with a probability μ, in which case they are replaced at random by another helping/punishment strategy, respectively. This potentially creates new combinations of punishment and helping strategies into our simulations. We used n = 500, m = 30 and μ = 0.02 in all our simulations. It is well established that for a reputation system to be efficient, a relatively high number of interactions per individual is required [14,32,33], hence our choice of m = 30 interactions. As a consequence, we chose a sufficiently large number of individuals in the population to avoid direct reciprocity effects, that is, a given player meeting the same partner in reversed roles (i.e. probability <0.03). In the electronic supplementary material, we show that decreasing the mutation probability μ can often hinder the emergence of punishment and cooperation for low benefits b and low punishment ratios (s/e; electronic supplementary material, figure S1).

Our mechanism of punishment scoring is analogous to the image scoring of [14] but applied to punitive actions: individuals have a punishment score that starts at 0 and can reach −5 or +5. We implemented four different punishment/image scores, which we denote k1, k2, k3 and Is.

  • — k1punishment of defection. Each time an individual punishes defection, his k1 increases by 1 unit, whereas each time he does not punish defection, his k1 decreases by 1 unit.

  • — k2punishment of either defection/cooperation. Each time an individual punishes (defection or cooperation), his k2 increases by 1 unit, whereas each time he does not punish, his k2 decreases by 1 unit.

  • — k3punishment of cooperation. Each time an individual punishes cooperation, his k3 increases by 1 unit, whereas each time he does not punish cooperation, his k3 decreases by 1 unit.

  • — Iscooperation. Each time an individual cooperates, his Is increases by 1 unit, whereas each time he does not cooperate, his Is decreases by 1 unit (i.e. image score).

These punishment scores are used to define the different strategies. Table 2 shows our 10 different helping strategies specifying how a player acts in the donor role, and table 3 shows the four different punishment strategies specifying how a player acts in the receiver role. In summary, we have 10 different helping strategies that could be combined with four different punishing strategies; i.e. there were in total 40 behavioural strategies possible.

Table 2.

The 10 strategies in the simulations that specify how a player acts in the donor role.

strategy name strategy notation help if receiver … helping rule
defectors xj never help
cooperators yj always help
discriminators zj punished defections k1 ≥ 0
discriminators' z'j did not punish defections k1 < 0
helpers to punishers uj punished k2 ≥ 0
helpers to punishers' u'j did not punish k2 < 0
helpers to C-punishers vj punished cooperative actions k3 ≥ 0
helpers to C-punishers' v'j did not punish cooperative actions k3 < 0
image scorers aj cooperated Is ≥ 0
image scorers' a'j defected Is < 0

Table 3.

The four strategies in the simulations that specify how a player acts in the receiver role. (Non-punishers never punish.)

strategy name strategy notation punish if donor …
unconditional punishers iA cooperated or defected
non-punishers iN
punishers of defection iP defected
punishers of cooperation iC cooperated

Figure 1 gives a typical example of how punishing strategies emerged in our simulations and how populations arrived at a mix of cooperative strategies. The mean frequencies across 20 replicates and for generations 18 000–20 000 were: punishing discriminators = 0.58; punishing cooperators = 0.12, non-punishing discriminators = 0.06, mean rate of cooperation = 0.98, mean rate of punishment = 0.01 (standard errors always <0.01). Note that low levels of punishment are sufficient to prevent defectors from invading (figure 1b). Punishing discriminators (zP) consistently evolved as the dominant strategy in a wide range of different parameters (figure 2), provided that the cost of being punished was greater than the cost of cooperating (e > c), and that the cost of punishing was smaller than the benefit of cooperation (s < b; figure 2). Increasing the benefit b of cooperation also had the consequence of increasing the frequency of punishing cooperators (yP) and punishing image scorers (aP; figure 2). Although some drift effects were possible (e.g. non-punishing defectors can be invaded by other non-cooperative strategies or punishing cooperators can invade punishing discriminators; figure 2), variation brought by mutations maintained a certain selective pressure for discrimination, such that unconditional cooperators could never dominate, i.e. the cycling between strategies seen in the analytical model was avoided, and cooperation was more stable (figure 1b). Decreasing the mutation rate μ made it harder for punishers to invade for low benefits b of cooperation (electronic supplementary material, figure S1af), and also decreased the selection pressure for discrimination (electronic supplementary material, figure S1di).

Figure 1.

Figure 1.

Punishment scoring leads to cooperation in computer simulations. (a) The six most frequent strategies (out of 40 possible ones) in a typical simulation for a finite population initially constituted of non-punishing defectors (red). Non-punishing discriminators (brown) appear and pave the way for punishing discriminators (green) to invade and dominate all other strategies. The other frequent strategies are defectors punishing cooperation (yellow), punishing cooperators (blue) and punishing image scorers (purple). (b) Frequency of cooperative moves (blue line) and punitive moves (red line). Parameters values are: c = 1, b = 2, s = 1, and e = 4.

Figure 2.

Figure 2.

Evolution of punishment at cost of helping c = 1 and benefit b = 2 (ad), b = 4 (eh), and b = 10 (il). The average frequencies of the most successful strategies non-punishing defectors (xN), punishing cooperators (yP), punishing discriminators (zP) and punishing image scorers (aP) are calculated across 20 replicates for generations 18 000–20 000 for different costs s to the punisher and costs e of being punished (standard errors always <0.1). Punishing discriminators (zP) predominate in the simulations for a large set of parameters. The category ‘others’ is the sum of the remaining 36 strategies, some of which mainly arise through drift, for example, defectors punishing cooperation if non-punishing defectors (xN) are most frequent (e.g. (d)).

As in other indirect reciprocity models, our punishment scoring system depends on the ability of a player to correctly assess the punitive reputation of others. Thus, we went on to test whether the inclusion of errors had an impact on our results. We introduced errors in the perception of an individual's score. After each interaction, the donor's cooperation score Is was replaced randomly with a probability ɛ by either Is 1, Is + 0 or Is + 1, and the receiver's punishment score ki was replaced randomly with a probability ɛ by either ki − 1, ki + 0 or ki + 1. As a consequence, the number of wrong actions caused by an incorrect score perception was greater at the beginning of each generation and decreased as the game went along (e.g. if my image score is 4 and I cooperate, replacing it by 3 instead of 5 will not result in a wrong action). This realistic assumption reflected the fact that more mistakes are made when players do not know their partners at the beginning of the interactions. As shown in the electronic supplementary material, the inclusion of errors often hinders the emergence of punishment and cooperation when the benefit b of cooperation is low (electronic supplementary material, figure S2ac). With greater b, however, punishment still emerged, but increasing the error rate reduced the efficiency of punishing discriminators (zP) and punishing image scorers (aP) relative to punishing cooperators (yP) who do not use reputation for their actions (electronic supplementary material, figure S2di).

3. Discussion

Punishment that serves to prevent an individual from repeating a damaging action towards the punisher or that serves to prevent future defection towards the punisher seems to be very common in humans and some animals [1]. Recent studies have concluded, however, that punishment may have evolved for something else other than for promoting cooperation, because significant benefits to punishers could typically not be found in the context of pure cooperation games [4,611,34]. Third-party punishment in an indirect reciprocity setting, for example, rarely favours punishment [9,10]. Indeed, it has been suggested that benefits to groups rather than to individuals could explain the evolution of punishment [35], even though punishment is expected to provide either direct or indirect fitness benefits to the individual in order to evolve [26,31,36]. For example, punishment can be favoured by indirect reciprocity when it discourages future aggression by observers [37]. Our models show that even if punishing defectors is immediately costly, it acts to discourage future refusals to help from observers to such a degree that the immediate costs of punishment are outweighed by the additional donations it evokes over the long run. Hence, punishment evolved in our simulations entirely through a punitive reputation, i.e. without punishment directed towards non-punishers [4,22] or the need of spatial constraints [25,35,38].

The cognitive abilities of humans may allow reputation to be used not only for assessing the cooperativeness of individuals within a social group, but also with regard to the readiness of group members to punish defectors. In our simulations, both types of reputation could be used, and individuals merely using the other's reputation of being generous (i.e. image scoring [14]) do not fare better than punishing discriminators. The latter strategy has a higher pay-off than punishing cooperators, a strategy that could be interpreted as strong reciprocity [35,38]. Strong reciprocity can be evolutionarily stable when common, but when punitive actions are observable by others, punishing cooperators cannot invade punishing discriminators unless the cost of helping is greater than the cost of being punished (c > e, as shown in our analytical model above). It is likely that situations where individuals actions were observable, and thus formed public information, occurred during much of human evolutionary history [39].

In experimental public goods games, punishment is typically perceived as conferring benefits on a social group and being an act of cooperation in itself [3,40]. Indeed, humans seem to be more likely to punish if they are observed by others [41], suggesting that they care about the reputation effects that arise from punishing. It has been shown that refusing low offers (a sort of punishment) in the ultimatum game when there are observers made the ‘punishers’ more likely to receive higher offers in later interactions [42]. Thus, it is still unknown whether the proximate incentives to cooperate more often with punishers would come from the fear of being punished (and thus defecting when there is no threat of punishment would be opportunism [27]) or from the wish to reward punishers for their pro-social behaviour [3,40]. However, the latter hypothesis raises a second-order dilemma as rewarding punishers is also costly.

We focused on strategies using indirect information on the punishment reputation of others. Strategies using their own experience of direct punishment (i.e. whether you punished me) compared with those using indirect information (i.e. whether you punished others) may perform differently depending on the number of interactions as well as the group size, as it is the case with direct and indirect reciprocity [33]. Similarly, the weight of information attached to reputation gained from punishment of defectors and cooperators was the same in our model, but both types of information could be weighted differently [21]. We also assumed no retaliation from punished individuals in our models. While more realistic, the option of retaliation seems to lower the cooperation level in public goods games [43], but it is still unknown what would be the influence on cooperation if retaliative actions also impact one's reputation. Another assumption of our model was that all individuals have the same capacity to punish. This is probably unrealistic for many animals when punishment could be used to establish and maintain dominance relationships, in which case dominant individuals would receive more cooperation [1,44,45].

Our model bears similarity to the model of Hilbe & Sigmund [27] (hereafter H&S), where a game was played in which reputation was implemented as a probability to know about the others' behaviour and individuals had the option to reward cooperators. For instance, they found that punishing discriminators (denoted [OC, P] in H&S) can invade a population initially constituted of non-punishing defectors (denoted [ALLD, N] in H&S) if the probability to know whether the co-player punishes or not is greater than s(b + s), denoted (μ > γ) (b + γ) in H&S. Moreover, the possibility for receivers to reward cooperation seems to foster the evolution of cooperation (and then punishment) when the probability to know the co-player's reputation is small [27]. We included 40 different strategies in our simulations including the possibility to always punish, or to punish cooperation, an often missing feature of many models [22,2527,31], and still find that punishing discriminators can bring cooperation and prevent defectors from invading.

Our study highlights the importance of reputation in driving the evolution of punishment. By allowing reputation to be based on either the punishment of defectors, cooperators or both, we have shown that punishing defectors and always cooperating with punishers emerges as a dominant strategy. Our results are also robust to the other strategies, such as image scoring. We conclude that reputation is the key to the evolution of punishment, and that simple reputation games can explain the high prevalence of punishment in humans. The combination of reputation and punishment acts as a strong mechanism promoting the evolution of cooperation.

Acknowledgements

We thank R. Bergmüller, R. Bshary, M. Chapuisat, C. Clavien, C. El Mouden, A. Gardner, C. Hauert, M. Hochberg, L. Keller, C. Metzger, S. Nusslé, A. Ross-Gillespie and the anonymous referee for discussion or comments on a previous version of the manuscript, and the Swiss National Science Foundation (grants to C.W. and D.J.R.) for funding.

Appendix A

Our analytical model above consists of six strategies, namely xN, xP, yN, yP, zN and zP. The pay-offs for these respective strategies can be written as:

graphic file with name rspb20101275-e6.jpg

References


Articles from Proceedings of the Royal Society B: Biological Sciences are provided here courtesy of The Royal Society

RESOURCES