Abstract
Although positive incentives for cooperators and/or negative incentives for free-riders in social dilemmas play an important role in maintaining cooperation, there is still the outstanding issue of who should pay the cost of incentives. The second-order free-rider problem, in which players who do not provide the incentives dominate in a game, is a well-known academic challenge. In order to meet this challenge, we devise and analyze a meta-incentive game that integrates positive incentives (rewards) and negative incentives (punishments) with second-order incentives, which are incentives for other players’ incentives. The critical assumption of our model is that players who tend to provide incentives to other players for their cooperative or non-cooperative behavior also tend to provide incentives to their incentive behaviors. In this paper, we solve the replicator dynamics for a simple version of the game and analytically categorize the game types into four groups. We find that the second-order free-rider problem is completely resolved without any third-order or higher (meta) incentive under the assumption. To do so, a second-order costly incentive, which is given individually (peer-to-peer) after playing donation games, is needed. The paper concludes that (1) second-order incentives for first-order reward are necessary for cooperative regimes, (2) a system without first-order rewards cannot maintain a cooperative regime, (3) a system with first-order rewards and no incentives for rewards is the worst because it never reaches cooperation, and (4) a system with rewards for incentives is more likely to be a cooperative regime than a system with punishments for incentives when the cost-effect ratio of incentives is sufficiently large. This solution is general and strong in the sense that the game does not need any centralized institution or proactive system for incentives.
Author Summary
Although social dilemmas can be resolved if punishing non-cooperators or rewarding cooperators works, such rewards and punishments, i.e., external incentives, entail certain expenses. As a result, a cooperative player who shirks his or her duty to provide an incentive to other players will emerge, and he or she will be more advantageous than an incentive-provider. In fact, the problem of excluding such cooperative incentive-non-providers, or second-order free-riders, is a well-known academic challenge. In order to meet this challenge, we devise and analyze a meta-incentive game that integrates positive incentives (rewards) and negative incentives (punishments) with second-order incentives, which are incentives for other players’ incentives. In this paper, we solve the replicator dynamics for a simple version of the game and analytically categorize the game types into four groups. We show that second-order incentives for first-order reward are necessary for cooperative regimes. This solution is general and strong in the sense that the game does not need any centralized institution or proactive system for incentives.
Introduction
Even though society is based on cooperation, achieving cooperation in social dilemmas is still a big challenge. The free-rider problem, for example, hinders cooperation. Many studies have addressed this problem, and the methods proposed for solving it include giving players sufficient ability to remember their direct [1] or indirect experiences [2]. The idea is that this additional information generates cooperation through direct or indirect reciprocity. Other research has attempted to solve the problem by assigning tags [3], reputations [4], spatial structures [5] or networks [6] to players in the game. A third approach is to give players choices other than simply whether or not to contribute. A drawback of this approach, however, is that there may be a loner, i.e., a player who does not participate in the game [7, 8], or a joker, i.e., a destroyer who damages the public good [9]. Some researchers have devised another sort of game that promotes cooperation by giving players incentives explicitly. Important incentives for the evolution of cooperation are rewards and punishments as they tend to capture strong views of human nature [10, 11]. The approach we took integrates positive incentives such as rewarding and negative ones such as punishing into a system for promoting cooperation.
A meta-analysis of reviews on reward and punishment systems using a common framework [12] revealed that this approach is still controversial. Here, we tackle this issue by focusing on three perspectives: (1) the contrast between an individual incentive system and a centralized institutional one, (2) an incentive-integrated punishment and reward system, and (3) incentives on a meta-level.
First, many studies have focused on either individually-dealt-with punishments or rewards. While some studies [13–16] have shown that a costly punishment can effectively achieve cooperation, others [17–21] have shown just the opposite. Some researchers claim that the findings of peer-punishment studies may not be broadly applicable to modern human societies, because rewards and punishments are typically carried out by rules-bound institutions [22, 23] rather than by individuals. The reason might be a difficulty of establishing and maintaining such a peer-punishment system because it does not involve a nomocracy, i.e., a proactive or ex-ante commitment.
Second, should a free-rider be punished and/or should a contributor be rewarded? Experimental studies [24, 25] have indicated that rewards and punishments induce similar levels of cooperation when the incentive is very large. Economists have used experimental games to study the effects of positive and negative incentives (i.e., rewards and punishments) on the propensity to collaborate [26]. A theoretical study [27] showed that a punishment is more effective than a reward because these incentives have an asymmetric relationship with one another. In other words, punishments are not needed once cooperation is established, and thus, cooperators do not need to pay the costs of punishment. On the other hand, attaining rewards requires participants to pay costs by following a cooperative strategy. The threat of a strong punishment can achieve cooperation at a very low cost [28]. For an intermediate level of incentive, however, although punishments can induce greater cooperation than rewards [25], they cannot do so consistently [29]. Moreover, cooperation easily breaks down if both forms of incentive are removed [30]. Compared with the numerous studies on punishments, there have been relatively few on rewards [12, 24, 25, 29]. For example, an experimental study [31] explored the situation in which unkind newcomers are strictly exploited and found that indirect rewards are effective in such situations.
Third, in the approach we took, the focus is on meta-level incentives. When the incentive for cooperation is either a punishment or reward, there is still the second-order free-rider problem. The effort made to maintain a cooperative society is a cost that must be defrayed by someone. Generally, a player who contributes to a game but never defrays the cost for providing incentives is more evolutionarily adaptive than one who contributes to the game and does pay the cost. This means that eventually no one defrays the cost of maintaining the incentive system. The second-order free-rider problem can come down to the problem of costly incentives [32]. One solution is to implement a second-order incentive system.
The pioneering work on meta-level incentives was performed by Axelrod [33]. He attempted to evolve cooperation by imposing a second-order punishment on those who do not impose a first-order punishment when one is called for. His model linked punishments against non-punishers with punishments against non-cooperators. This assumption, that first-order incentives and second-order ones are linked, or FO-SO-linkages, is critical for our study. Yamagishi and Takahashi [34] were the first to point out the linkage issue and demonstrated that a cooperative regime emerges if it is assumed that players have linkages between cooperation and first-order incentives, or C-FO-linkages. Related studies have developed models that assume C-FO-linkages have been analyzed [35, 36]. The existence of a C-FO-linkage, however, is still not a foregone conclusion [37]. Some experimental studies [38, 39] showed that sanctions enhance norm and cooperative behavior. An experimental study [40], conversely, concluded there is a negative correlation between cooperative behaviors in prisoner’s dilemma games and refusal behaviors in ultimatum games. Moreover, an analysis of large-scale panel data of Germany [41] concluded that rewards and punishments have no relationship. Another experiment [42] found that cooperation is not correlated with norm-enforcing punishments. An experimental study [37], on the other hand, showed a significant correlation between cooperation and punishment; however, they unostentatiously admitted that their result was insufficient evidence for the existence of C-FO-linkages.
The C-FO-linkage issue is a relationship between two behaviors which differ qualitatively. Considering this point, the relationship between first-order incentives and second-order incentives is an alternative issue because they are both incentives. Kiyonari and Barclay [43] focused on the FO-SO-linkage and showed its existence in their experiments on one-shot public good games. Their study opened the door on analyses of models that assume FO-SO-linkages. Hilbe et al [44] experimentally showed the rationality of a second-order punishment in an authorized sanction system. We, in this paper, test a model that assumes a FO-SO-linkage in a peer-to-peer incentive system.
We have developed a model of a meta-incentive game (MIG) that has second-order incentives, i.e., incentives for other players to provide incentives. This analytical model can describe the carrot-and-stick issue uniformly and comparatively. The model targets an individual incentive system that integrates a positive side (reward) and a negative side (punishment) [45]. The incentive we consider is an ex-post type applied after players engage in donation games, and thus, no centralized institution with incentives is needed. An institution requires an ex-ante commitment among players. Players should decide whether or not they will participate in the institution before playing the donation games [23, 46]. In order to resolve the second-order free-rider problem, we suppose that there are three types of players in the MIG, i.e., a non-cooperative incentive-non-provider as a first-order free-rider, a cooperative incentive-non-provider as a second-order free-rider, and a cooperative incentive-provider, and we will explore the conditions under which cooperative incentive-providers survive.
Results
MIG players first play donation games and then provide incentives in answer to their actions in the games. Incentives are provided not only for or against the others’ cooperative or non-cooperative actions but also for the others’ incentive behaviors or lack thereof on third-party players.
Players are divided into three sorts of strategist: a cooperative incentive-provider (CI), a cooperative incentive-non-provider (CN), and a non-cooperative incentive-non-provider (NN). Fig 1 shows an illustration of the MIG. The MIG consists of three stages (games): a donation game (DG), a first-order incentive game (FIG), and a second-order incentive game (SIG). Note that each player has perfect information, so each one knows all the players’ actions.
The population is infinitely large and well-mixed. The frequencies of the three strategies follow replicator dynamics describing a natural law whereby the higher one’s payoff is the more frequent one’s strategy becomes. Let x, y, and z be the frequencies of CI, CN, and NN, respectively. Naturally, x+y+z = 1. The equations are formulated as
(1) |
where U CI, U CN, U NN, and are, respectively, the average payoffs of CI, CN, NN, and all the players. is given by
(2) |
Now let us describe the parameter notations needed to calculate a player’s (expected) payoff. Let c be the cost of donation, b be the receiver’s benefit, F 1 be the fine imposed as a first-order punishment, P 1 be the cost of a first-order punishment, A 1 be the amount of a first-order reward, R 1 be the cost of a first-order reward, F P be the fine for freeriding for the first-order punishment, P P be the cost for freeriding for the first-order punishment, A P be the amount of the reward for the punisher, R P be the cost of the reward for the punisher, F R be the fine for freeriding for giving a reward, P R be the cost for freeriding for giving a reward, and A R be the amount of rewarding a rewarder, and R R be the cost of rewarding a rewarder. All these values should be non-negative constants.
We will avoid analytical difficulties due to the usage of many parameters by defining a simple meta-incentive game (S-MIG) using two parameters: the incentive cost-effect ratio (μ), which represents the proportion of a fine or award that incentive-receivers should pay or receive relative to its cost, and the discount factor of costs on the level of incentive (δ), where
and
We assume that μ > 1 and 0 < δ < 1. We can set c = 1 without loss of generality. An S-MIG is perfectly described by a duplet (μ, δ).
Finally, before analyzing our model, we define all 24 possible configurations of MIG in Fig 2. For example, the P-type MIG has only first-level punishments. In this type, players can give or receive neither a first-level reward nor a second-level incentive. The PR type has first-level punishments for non-cooperators and second-level rewards for punishers.
We explore the conditions under which a cooperative equilibrium (x > 0) emerges by analyzing the replicator dynamics on different types of S-MIG. As shown in Methods, the dynamics of S-MIGs can be classified into four groups. Fig 2 illustrates the existence condition for the basin of attraction and the local stabilities on the point (x, y, z) = (1, 0, 0) of all types, and Fig 3 shows the phase portraits of the representative S-MIGs on a 2-dimensional simplex.
We identified certain features of the second-order free-rider problem. First, as the second-order free-rider problem warns, cooperation cannot be maintained in any system with only first-order incentives. Similarly, cooperation does not arise even if the system has second-order incentives but no second-order incentives for first-order rewards. This is because rewarding others is equivalent to a prisoner’s dilemma game or a donation game, so incentive-providers as well as cooperators diminish over time.
Second, because of the neutral drift effect, no system without a reward side can keep its position when cooperation dominates. This is another aspect of the second-order free-rider problem. These results reveal two important facts about reward systems: a second-order incentive system for first-order rewards is better than a non-incentive system on the reward side for promoting cooperation, and a system without second-order incentives for first-order rewards is worse than the non-incentive system.
Third, any system with second-order incentives for first-order rewards can produce a stable cooperative regime under specific conditions. To do so, a certain number of cooperative incentive-providers are needed. If their numbers are small, they cannot survive.
Fourth, the conditions under which a cooperative regime emerges depend on the system. A system with second-order rewards has the strictest condition, and the conditions become less stringent for a system with second-order punishments and a system with second-order both rewards and punishments.
Finally, the condition on the frequency of incentive-providers under which a cooperative regime can be sustained also depends on the system. When the cost-effect ratio of incentives is sufficiently large, the lower limit of the frequency of incentive-providers in a system with second-order rewards for first-order incentives is lower than that in a system with second-order punishments for first-order incentives.
Discussion
What can resolve the second-order free-rider problem? Sigmund et al [23] achieved stabilizing cooperative regimes by using pool punishments instead of peer punishments. Is it possible to maintain such regimes without any proactive institution? Our model demonstrates that assuming first-order and second-order incentives are linked can lead to a solution without any social costs or a punishment fund. The assumed linkage means that individuals who are willing to provide incentives would automatically provide meta-incentives as well. The consequences are that although the model allows second-order free riders, it does not allow third-order free riders. (If they were allowed, they would again destabilize cooperation.) Moreover, efficiency is traded for stability in Sigmund et al’s model, because individuals pay for a punishment fund without as yet knowing who the free riders to be punished are [23]. In cotrast, our model can search for an efficient incentive level for maintaining cooperative regimes, because individuals can reactively incur costs for incentives with knowing who is to be given an incentive. Thus, one of the implications of the linkage assumption is that a more efficient incentive system for stabilizing cooperation would be an intermediate of the traditional peer and pool incentive systems.
We note that a similar amount of cooperation could also be achieved if there was a linkage between cooperation and first-order incentives (C-FO linkage, see Methods). This linkage means an alternative model with two strategies, i.e., defectors and cooperators who automatically also provide first-order incentives. As shown in Methods, our model can indeed cover a case that assumes C-FO linkages instead of FO-SO linkages in specific parameters.
For our model incentives for reward are necessary for cooperative regimes. If players play in a non-incentive system on the reward side, a punishment function does not work when cooperation is achieved. That is to say, they cannot respond to an invasion of neutral mutants who do not provide incentives. As a result, cooperation suddenly collapses. This is why Axelrod’s simulation [33] cannot keep a cooperative regime for a long time [47–49]. Therefore, if players play in a non-incentive system on the reward side, another mechanism is needed, e.g., a social vaccine proposed by Yamamoto and Okada [50], to maintain cooperative regimes.
How about a system with first-order reward and without second-order incentives for the reward? Here, worse comes to worst because it becomes free from the possibility of staying a cooperative regime temporally. When a cooperative regime is achieved, players do not need their punishment functions, and thus, only the first-order reward function works. As the second-order free-rider problem indicates, cooperative incentive-non-providers beat cooperative incentive-providers because they don’t bear the burden of paying for rewards.
Second-order incentives for first-order rewards are necessary for achieving and maintaining robust cooperative regimes and to resolve the second-order free-rider problem. That is, a mechanism is needed to make it beneficial for a player to give a first-order reward. Assuming that players who tend to provide incentives for other players’ cooperative or non-cooperative behaviors also tend to provide incentives for their incentive behaviors, the second-order free-rider problem can be completely resolved without any third-order or higher (meta)incentive. In our model, moreover, incentives are performed ex-post and individually, and thus the system does not need a centralized institution or ex-ante commitment. Many studies on non-meta-level incentives have shown that punishment is more effective than a reward. We have identified a possible explanation as to why people prefer second-order rewards to second-order punishments. Kiyonari and Barclay’s experimental study [43] supports this—they found that people readily provided second-order rewards toward those who rewarded cooperators while they did not administer second-order punishments to non-punishers because the reward systems were more easily supported by higher order incentives and were thus more likely to persist.
Which incentive should a designer of an MIG choose? If the cost-effect ratio of incentives is sufficiently large, even a handful of cooperative incentive-providers can beat non-cooperative incentive-non-providers if the designer uses a system with rewards for incentives instead of a system with punishments for incentives.
Our work differs from Sasaki et al’s model of integrating rewards and punishments, which was designed for an institutional system with a compulsory entrance fee and thus no option of second-order free riders [51]. Kendal et al [52] analyzed a model of second-order peer rewards for punishers, but did not consider rewards for cooperators who contribute in the game. Our paper is a pioneering analysis of MIGs; as such, we only dealt with a minimum deviation and analyze pro-social incentives and left anti-social punishments [53] and anti-social rewards as topics for future study. Our model assumes an infinitely well-mixed population, and this assumption should be loosened in the future. Szolnoki and Perc [54] compared the effect of a reward with that of a punishment in a spatial public goods game. Chen and Perc [55] studied the optimal distribution of institutional incentives in a public goods game on a scale-free network.
The asymmetry of the effects of rewards and punishments might be evident in the cost-effect ratio. Although it may be natural that a fine handed out as punishment should be larger than its cost, can the reward be larger than its cost? We believe that this balances out the risk of second-order free-riders emerging. Of course, we leave open the possibility of incorporating asymmetric reward and punishment effects as a future extension of a system.
Methods
In this section, we analyze a meta-incentive game (MIG) by solving replicator equations. Table 1 gives brief descriptions of all 24 types of the game.
Table 1. Types of MIG.
Type | Brief description | Parameters used |
---|---|---|
P | Punishment for non-cooperators on 1st level | F 1, P 1 |
R | Reward for cooperators on 1st level | A 1, R 1 |
P+R | Both reward and punishment on 1st level (as well as P-type plus R-type) | F 1, P 1, A 1, R 1 |
PP | Punishment for non-cooperators on 1st level and punishment for non-punishers on 2nd level | F 1, P 1, F P, P P |
PR | Punishment for non-cooperators on 1st level and reward for punishers on 2nd level | F 1, P 1, A P, R P |
PB | Punishments on both levels (as well as PP-type) and reward for punishers on 2nd level | F 1, P 1, F P, P P, A P, R P |
RP | Reward for cooperators on 1st level and punishment for non-rewarders on 2nd level | A 1, R 1, F R, P R |
RR | Reward for cooperators on 1st level and reward for rewarders on 2nd level | A 1, R 1, A R, R R |
RB | Rewards on both levels (as well as RR-type) and punishment for non-rewarders on 2nd level | A 1, R 1, F R, P R, A R, R R |
P+RP | Both reward and punishment on 1st level and punishment for non-rewarders on 2nd level | A 1, R 1, F 1, P 1, F R, P R |
P+RR | Both reward and punishment on 1st level and reward for rewarders on 2nd level | A 1, R 1, F 1, P 1, A R, R R |
P+RB | Both reward and punishment on 1st level, and both punishment for non-rewarders and reward for rewarders on 2nd level | A 1, R 1, F 1, P 1, A R, R R, F R, P R |
PP+R | Both reward and punishment on 1st level and punishment for non-punisher on 2nd level | F 1, P 1, A 1, R 1, F P, P P |
PR+R | Both reward and punishment on 1st level and reward for punisher on 2nd level | F 1, P 1, A 1, R 1, A P, R P |
PB+R | Both reward and punishment on 1st level, and both punishment for non-punishers and reward for punishers on 2nd level | F 1, P 1, A 1, R 1, F P, P P, A P, R P |
PP+RP | Both reward and punishment on 1st level, and punishments for both non-punishers and non-rewarders on 2nd level | F 1, P 1, A 1, R 1, F P, P P, F R, P R |
PP+RR | Both reward and punishment on 1st level, and both punishment for non-punishers and reward for rewarders on 2nd level | F 1, P 1, A 1, R 1, F P, P P, A R, R R |
PP+RB | Full-type excluded reward for punishers on 2nd level | All except A P, R P |
PR+RP | Both reward and punishment on 1st level, and both punishment for non-rewarders and reward for punishers on 2nd level | F 1, P 1, A 1, R 1, F R, P R, A P, R P |
PR+RR | Both reward and punishment on the 1st level, and rewards for both punishers and rewarders on the 2nd level | A 1, R 1, F 1, P 1, A R, R R, A P, R P |
PR+RB | The Full-type excluded punishment for non-punishers on 2nd level | All except F P, P P |
PB+RP | Full-type excluded reward for rewarders on 2nd level | All except A R, R R |
PB+RR | Full-type excluded punishment for non-rewarders on 2nd level | All except F R, P R |
PB+RB (Full) | MIG itself | All |
Expected payoffs of the players
The expected payoffs of the players are
(3) |
Let us explain the terms of U CI. The first term represents the benefit of donation, whereas the second term gives the cost of donation (because one is a cooperator) in the DG stage. The third term represents the costs of the first-order incentivization: rewarding a cooperator and punishing a non-cooperator. The fourth term represents the first-order reward for cooperating. The fifth term represents the costs of the second-order incentivization and consists of four parts, i.e., rewarding those who have rewarded a cooperator, punishing those who have not rewarded a cooperator, rewarding those who have punished a non-cooperator, and punishing those who have not punished a non-cooperator. Finally, the last term describes the second-order rewards for rewarding a cooperator and punishing a non-cooperator. Similar explanations can be applied to the other expected payoffs, U CN and U NN.
Replicator dynamics analysis for various types of S-MIG
Using Eqs (1), (2), and (3), the replicator equations are calculated as , and , where
(4) |
We prove that there is no equilibrium at any internal point (x, y, z) in S-MIG in the following subsection, and thus, here, it is enough to conduct an analysis of the borders. Moreover, there is no equilibrium at any point on the line x = 0, except two corners ((y, z) = (0, 1), (1, 0)). This is because, on the line x = 0, and thus, y always decreases. Hence, we will only deal with the function on the lines y = 0 and z = 0. On those two lines, that function is calculated as and , where
(5) |
In order to exemplify how to analyze each type of the S-MIG, we deal with the RB type as a representative and derive their equations of f(x) and g(x). In the type, F 1, P 1, F P, P P, A P and R P should be zero because there has no incentive system on the punishment side. Moreover, using c = 1 and definitions of μ and δ, R 1 = δ, A 1 = μδ, R R = P R = δ 2 and A R = F R = μδ 2. Substituting them into f 1 in Eq (4), we get
Therefore, we derive Eq (5) of the RB-type S-MIG as
and
Table 2 shows the equations for and for each type of S-MIG. Using Eq (5), we can calculate an existence condition for the basin of attraction on the point (x, y, z) = (1, 0, 0). If and are satisfied, the point (x, y, z) = (1, 0, 0) is asymptotically stable, and thus, a cooperative regime emerges. The dynamics of each type of S-MIG are classified into four groups. In what follows, we will deal with one representative type in each group. The remainder can be similarly derived.
Table 2. functions on the lines y = 0 and z = 0 for each type of S-MIG. Here, f(x) and g(x) are defined in Eq (5).
Type | f(x) | g(x) |
---|---|---|
P | −1−δ(1−x)+μδx | 0 |
R | −1−δx+μδx | −δ |
P+R | −1−δ+2μδx | −δ |
PP | −1−δ(1−x)+μδx−δ 2(1−x)2+μδ 2 x(1−x) | 0 |
PR | −1−δ(1−x)+μδx−δ 2 x(1−x)+μδ 2 x(1−x) | 0 |
PB | −1−δ(1−x)+μδx−δ 2(1−x)+2μδ 2 x(1−x) | 0 |
RP | −1−δx+μδx−δ 2 x(1−x)+μδ 2 x 2 | −δ−δ 2(1−x)+μδ 2 x |
RR | −1−δx+μδx−δ 2 x 2+μδ 2 x 2 | −δ−δ 2 x+μδ 2 x |
RB | −1−δx+μδx−δ 2 x+2μδ 2 x 2 | −δ−δ 2+2μδ 2 x |
P+RP | −1−δ+2μδx−δ 2 x(1−x)+μδ 2 x 2 | −δ−δ 2(1−x)+μδ 2 x |
P+RR | −1−δ+2μδx−δ 2 x 2+μδ 2 x 2 | −δ−δ 2 x+μδ 2 x |
P+RB | −1−δ+2μδx−δ 2 x+2μδ 2 x 2 | −δ−δ 2+2μδ 2 x |
PP+R | −1−δ+2μδx−δ 2(1−x)2+μδ 2 x(1−x) | −δ |
PR+R | −1−δ+2μδx−δ 2 x(1−x)+μδ 2 x(1−x) | −δ |
PB+R | −1−δ+2μδx−δ 2(1−x)+2μδ 2 x(1−x) | −δ |
PP+RP | −1−δ+2μδx−δ 2(1−x)+μδ 2 x | −δ−δ 2(1−x)+μδ 2 x |
PP+RR | −1−δ+2μδx−δ 2[x 2+(1−x)2]+μδ 2 x | −δ−δ 2 x+μδ 2 x |
PP+RB | −1−δ+2μδx−δ 2[1−x(1−x)]+μδ 2 x(x+1) | −δ−δ 2+2μδ 2 x |
PR+RP | −1−δ+2μδx−2δ 2 x(1−x)+μδ 2 x | −δ−δ 2(1−x)+μδ 2 x |
PR+RR | −1−δ+2μδx−δ 2 x+μδ 2 x | −δ−δ 2 x+μδ 2 x |
PR+RB | −1−δ+2μδx−δ 2 x(2−x)+μδ 2 x(x+1) | −δ−δ 2+2μδ 2 x |
PB+RP | −1−δ+2μδx−δ 2(1−x)(x+1)+μδ 2 x(2−x) | −δ−δ 2(1−x)+μδ 2 x |
PB+RR | −1−δ+2μδx−δ 2[1−x(1−x)]+μδ 2 x(2−x) | −δ−δ 2 x+μδ 2 x |
PB+RB | −1−δ+2μδx−δ 2+2μδ 2 x | −δ−δ 2+2μδ 2 x |
The R, P+R, PP+R, PR+R, and PB+R types belong in the first group. Each type of this group has a globally stable point (x, y, z) = (0, 0, 1), so cooperation is never achieved regardless of the values of (μ, δ). is satisfied; thus, (1, 0, 0) is unstable. Therefore, the only stable equilibrium point is (0, 0, 1), as shown in Fig 3(A).
The P, PP, PR, and PB types belong in the second group. The whole line z = 0 consists of fixed points because is always satisfied. Each type of this group has two patterns of behavior depending on the values of (μ, δ). In one, all the fixed points on the line z = 0 are unstable, and thus, there is a globally stable point (0, 0, 1), as in the first group. In the other, some of the fixed points are stable if when , as shown in the following subsection. Note that this behavior also satisfies an existence condition for the basin of attraction on the point (x, y, z) = (1, 0, 0) on y = 0. Let us verify the PR case as a representative example. and prove that f(x) is an increasing function. Moreover, f(1) = −1+μδ > 0 implies that the point (x, y, z) = (1, 0, 0) is asymptotically stable. Even in the second pattern, however, the point (1, 0, 0) is not stable for a long time. This is because, the whole line z = 0 consists of fixed points, and thus, neutral drift is possible. Occasionally, x moves to an unstable equilibrium, and this type eventually reaches (0, 0, 1). The phase diagram of the PP-type S-MIG is shown in Fig 3(B).
The third group, consisting of types which include either the RP or RR, and P+RB, PP+RB, PR+RB, and PB+RB (Full) types, has two patterns of behavior depending on the values of (μ, δ). In one, there is a globally stable point (0, 0, 1), as in the first group. In the other, another locally asymptotically stable point (1, 0, 0) can exist. This group has three existence conditions for the basin of attraction on the point (x, y, z) = (1, 0, 0): for the types which include RR, for the types which include RP, and for the P+RB, PP+RB, PR+RB, and PB+RB (Full) types. Let us examine the PB+RB(Full)-type S-MIG as a representative type. Here, ; hence, the dynamics on the line z = 0, on one hand, are bistable and is a fixed point that separates the two basins of attraction when is satisfied. On the other hand, the dynamics on y = 0 depends on f(x). shows that is an increasing function. f(0) < 0 and f(1) = −1−δ−δ 2+2μδ(1+δ) > 0 when the existence condition for the basin of attraction on the point (x, y, z) = (1, 0, 0) is satisfied. Therefore, the dynamics on y = 0 are also bistable, and is a fixed point that separates the two basins of attraction. Fig 3(C) shows the phase diagram of the Full-type S-MIG.
The final group, consisting of only the RB type, is the same as the third except for the direction of the dynamics in the internal space of the basin of attraction for cooperation (see Fig 3(D)). In this type, the unstable equilibrium point on the line z = 0 is a source while those in the third group are saddles. Likewise, the unstable equlibrium point on y = 0 in the RB type is a saddle, while those in the third group are sources. Using the analytical method of the following subsections, we can calculate the eigenvalues of the matrices derived by linearization of the dynamics around the equilibrium point. The eigenvalues of the equilibrium on the line z = 0 are and , and both are positive. Local stability theory says that an equilibrium with two positive eigenvalues is unstable and an equilibrium with one positive eigenvalue and one negative eigenvalue is a saddle. We can verify that the types of the third group have inverse stabilities. First, we deal with the Full-type S-MIG. Let λ 1 and λ 2 be the eigenvalues of the matrix derived by linearization of the dynamics around the equilibrium point (x*, y*, 0) on the line z = 0. λ 1 λ 2 = −y* δ 2(1+δ) < 0. Moreover, in the case of the RR-type S-MIG, λ 1 λ 2 = −x* y* δ 2 < 0. The other cases are omitted.
Finally, we compare the lower limits of x of the basins of attraction for cooperation (x, y, z) = (1, 0, 0) on y = 0. Let f D(x) and x D be and the lower limit of x of the basin of attraction on the line y = 0 in the D-type S-MIG for D ∈ {P, R, …, PB+RB}. Basically, the more complex the type is, the lower its lower limit of x becomes. For example, we will prove x P+RB < x P+RR. f P+RB(x) = −1−δ+2μδx−δ 2 x+2μδ 2 x 2 and f P+RR(x) = −1−δ+2μδx−δ 2 x 2+μδ 2 x 2 are in Table 2. and imply that both functions are increasing. Note that f D(x D) = 0. f P+RB(x P+RR) = δ 2 x P+RR[x P+RR(1+μ)−1] and then f P+RB(x P+RR) > 0. Therefore, x P+RB < x P+RR. Moreover, some equivalence relations can be derived. For example, let us compare x RR with x RP. f RR(x) and f RP(x) are increasing functions because their partial derivatives are positive in 0 < x < 1. f RR(x RP) = δ 2 x RP(1−2x RP) and are derived. Therefore, if , then and x RR < x RP.
where , and μ 0 > μ 10 > μ 11 > μ 12 > μ 13.
A model of C-FO-linkage
The MIG assumes FO-SO-linkages. In this subsection, we analyze the replicator dynamics of a model that assumes C-FO-linkages instead of FO-SO-linkages. Accordingly, cooperators automatically provide first-order incentives, and thus, there is no CN, and no second-order incentive is needed in the MIG. When y is set to zero and all the parameters for the second-order incentives are also zero (F P = P P = A P = R P = F R = P R = A R = R R = 0), this new version, or an incentive game (IG), can be regarded as a model of the C-FO-linkages. We denote the three possible IG configurations as the P-type, R-type, and P+R type. The replicator equation of IG is
We can devise a simple incentive game (S-IG) by using μ, δ, and c = 1. All S-IG types have two patterns of behavior depending on the values of (μ, δ). In one, there is a globally stable point (x, z) = (0, 1). In the other, another locally asymptotically stable point (x, z) = (1, 0) can exist. The existence condition for the basin of attraction at the stable point (1, 0) is for a P-type, for an R-type, or for a P+R-type S-IG.
Nonexistence of internal equilibrium
In this subsection, we prove that there is no equilibrium at any internal point on the 2-dimensional simplex Δ = {(x, y, z) : x, y, z ≥ 0, x+y+z = 1}. Assume that (x*, y*, z*) is an internal equilibrium. On that point, Eq (4) should be 0. This is because x* (y*, z*) is not zero, and thus, the function () can be divided by x*(y*, z*). Using , one gets
Similarly, using , one gets
(6) |
Table 3 shows the equations and solutions of z* in Eq (6) for all 24 types of S-MIG. The solutions of z* in the PP and RP types are not unique when μδ = 1. At that time, however, x* must be 1. Therefore, there is no equilibrium for both types when 0 < δ < 1.
Table 3. Equations and solutions of z* in Eq (6) for each type.
Type | Equation of z* in MIG | Solution of z* in S-MIG |
---|---|---|
P | −P 1 z* = 0 | 0 |
R | R 1 z* = R 1 | 1 |
P+R | (R 1−P 1)z* = R 1 | no solution |
PP | not unique when μδ = 1 | |
PR | 0 | |
PB | not unique when δ = 1 | |
RP | not unique when μδ = 1 | |
RR | 1 | |
RB | not unique when δ = 1 | |
P+RP | no solution | |
P+RR | no solution | |
P+RB | no solution | |
PP+R | ||
PR+R | ||
PB+R | ||
PP+RP | no solution | |
PP+RR | ||
PP+RB | ||
PR+RP | ||
PR+RR | no solution | |
PR+RB | ||
PB+RP | ||
PB+RR | ||
PB+RB | no solution |
If δ = 1 is permitted, the PB and RB types have internal equilibria at all points on . However, they are not stable. We will prove this and deal with the PB type as a representative. When δ = 1, Eq (1) in the PB type is
Now let us analyze the local stability of the point (x, y, z) around the equilibrium point (x*, y*, z*). Let ε x = x−x*, ε y = y−y* and ɛ = (ε x, ε y)T, where T means transposition. As a result of the linearization of the dynamics, where
because μx*−1 = 0. The eigenvalues of M are λ = 0, z*(3−2x*). These are non-negative, and thus, any internal equilibrium is non-isolated and unstable. The eigenvalues in the case of the RB type are λ = 0, z*+2(x*+y*)(y*+z*), and thus, the same conclusion is reached.
Local stability of the line z = 0 in the P, PP, PR, and PB types
In this subsection, we analyze the local stabilities of the line z = 0 in the P, PP, PR, and PB types of S-MIG. Here, we deal with the P-type S-MIG as a representative. Eq (1) becomes
As is shown in the previous subsection, the linearization of the dynamics around the equilibrium point (x, y, z) = (x*, y*, 0) leads to , where ε x = x−x*, ε y = y−y*, ɛ = (ε x, ε y)T, and
The eigenvalues of M are λ = 0, 1−μδx*, and thus, the whole line z = 0 consists of fixed points and it is stable (unstable) when () as well as the PP, PR, and PB types.
Acknowledgments
We thank Alfred Taudes, Vienna University of Economics and Business, and Ichiro Takahashi, Soka University, for their helpful comments.
Data Availability
All relevant data are within the paper.
Funding Statement
IO acknowledges support by the Grants-in-aid for Scientific Research from the Japan Society for the Promotion of Science (KAKENHI) 22520160. HY acknowledges support by the Grants-in-aid for Scientific Research from the Japan Society for the Promotion of Science (KAKENHI) 26330387. TS acknowledges support by the Foundational Questions in Evolutionary Biology Fund: RFP-12-21 and the Austrian Science Fund (FWF): P27018-G11. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1. Fehr E, Fischbacher U, Gächter S (2002) Strong reciprocity, human cooperation, and the enforcement of social norms. Human Nature 13, 1–25. 10.1007/s12110-002-1012-7 [DOI] [PubMed] [Google Scholar]
- 2. Nowak MA, Sigmund K (2005) Evolution of indirect reciprocity. Nature 437, 1291–1298. 10.1038/nature04131 [DOI] [PubMed] [Google Scholar]
- 3. Riolo RL, Cohen MD, Axelrod R (2001) Evolution of cooperation without reciprocity. Nature 414, 441–443. 10.1038/35106555 [DOI] [PubMed] [Google Scholar]
- 4. Ohtsuki H, Iwasa Y (2007) Global analyses of evolutionary dynamics and exhaustive search for social norms that maintain cooperation by reputation. J Theor Biol 244(3), 518–531. 10.1016/j.jtbi.2006.08.018 [DOI] [PubMed] [Google Scholar]
- 5. Helbing D, Szolnoki A, Perc M, Szabó G (2010) Evolutionary establishment of moral and double moral standards through spatial interactions. PLoS Comput Biol 6(4), e100075 10.1371/journal.pcbi.1000758 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Nakamaru M, Iwasa Y (2005) The evolution of altruism by costly punishment in the lattice structured population: Score-dependent viability versus score dependent fertility. Evol Ecol Res 7, 853–870. [Google Scholar]
- 7. Hauert C, De Monte S, Hofbauer J, Sigmund K (2002) Replicator dynamics for optional public good games. J Theor Biol 218, 187–194. 10.1006/jtbi.2002.3067 [DOI] [PubMed] [Google Scholar]
- 8. Sasaki T, Okada I, Unemi T (2007) Probabilistic participation in public goods games. Proc Biol Sci 274, 2639–2642. 10.1098/rspb.2007.0673 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Arenas A, Camacho J, Cuesta JA, Requejo RJ (2011) The joker effect: Cooperation driven by destructive agents. J Theor Biol 279, 113–119. 10.1016/j.jtbi.2011.03.017 [DOI] [PubMed] [Google Scholar]
- 10. Axelrod R (1984) The evolution of cooperation. New York: Basic Books. [Google Scholar]
- 11. Nowak MA (2012) Evolving cooperation. J Theor Biol 299, 1–8. 10.1016/j.jtbi.2012.01.014 [DOI] [PubMed] [Google Scholar]
- 12. Balliet D, Mulder LB, Van Lange PAM (2011) Reward, punishment, and cooperation: A meta-analysis. Psychol Bull 137(4), 594–615. 10.1037/a0023489 [DOI] [PubMed] [Google Scholar]
- 13. Fehr E, Gächter S (2002) Altruistic punishment in humans. Nature 415, 137–140. 10.1038/415137a [DOI] [PubMed] [Google Scholar]
- 14. Fehr E, Fischbacher U (2003) The nature of human altruism. Nature 425, 785–791. 10.1038/nature02043 [DOI] [PubMed] [Google Scholar]
- 15. Gürerk Ö, Irlenbusch B, Rockenbach B (2006) The competitive advantage of sanctioning institutions. Science 312, 108–111. 10.1126/science.1123633 [DOI] [PubMed] [Google Scholar]
- 16. Rockenbach B, Milinski M (2006) The efficient interaction of indirect reciprocity and costly punishment. Nature 444, 718–723. 10.1038/nature05229 [DOI] [PubMed] [Google Scholar]
- 17. Dreber A, Rand DG, Fudenberg D, Nowak MA (2008) Winners don’t punish. Nature 452, 348–351. 10.1038/nature06723 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Egas M, Riedl A (2008) The economics of altruistic punishment and the maintenance of cooperation. Proc Biol Sci 275, 871–878. 10.1098/rspb.2007.1558 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Wu JJ, Zhang BY, Zhou ZX, He QQ, Zheng XD, Cressman R, Tao Y (2009) Costly punishment does not always increase cooperation. Proc Natl Acad Sci USA 106(41), 17448–17451. 10.1073/pnas.0905918106 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Ohtsuki H, Iwasa Y, Nowak MA (2009) Indirect reciprocity provides only a narrow margin of efficiency for costly punishment. Nature 457, 79–82. 10.1038/nature07601 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Milinski M, Rockenback B (2012) On the interaction of the stick and the carrot in social dilemmas. J Theor Biol 299, 139–143. 10.1016/j.jtbi.2011.03.014 [DOI] [PubMed] [Google Scholar]
- 22. Szolnoki A, Szabó G, Perc M (2011) Phase diagrams for the spatial public goods game with pool punishment. Phys Rev E 83, 036101 10.1103/PhysRevE.83.036101 [DOI] [PubMed] [Google Scholar]
- 23. Sigmund K, De Silva H, Traulsen A, Hauert C (2010) Social learning promotes institutions for governing the commons. Nature 466, 861–886. 10.1038/nature09203 [DOI] [PubMed] [Google Scholar]
- 24. Rand DG, Dreber A, Ellingsen T, Fudenberg D, Nowak MA (2009) Positive interactions promote public cooperation. Science 325, 1272–1275. 10.1126/science.1177418 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Sutter M, Haigner S, Kocher MG (2010) Choosing the carrot or the stick? Endogenous institutional choice in social dilemma situations. Rev Econ Stud 77(4), 1540–1566. 10.1111/j.1467-937X.2010.00608.x [DOI] [Google Scholar]
- 26. Charness G, Rabin M (2002) Understanding social preferences with simple tests. Q J Econ 117(3), 817–869. 10.1162/003355302760193904 [DOI] [Google Scholar]
- 27. Hilbe C, Sigmund K (2010) Incentives and opportunism: from the carrot to the stick. Proc Biol Sci 277, 2427–2433. 10.1098/rspb.2010.0065 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Gächter S, Renner E, Sefton M (2008) The long-run benefits of punishment. Science 322, 1510 10.1126/science.1164744 [DOI] [PubMed] [Google Scholar]
- 29. Sefton M, Shupp R, Walker JM (2007) The effect of rewards and sanctions in provision of public goods. Econ Inquiry 45(4), 671–690. 10.1111/j.1465-7295.2007.00051.x [DOI] [Google Scholar]
- 30. Chaudhuri A (2011) Sustaining cooperation in laboratory public goods experiments: a selective survey of the literature. Experim Econ 14(1), 47–84. 10.1007/s10683-010-9257-1 [DOI] [Google Scholar]
- 31. Ule A, Schram A, Riedl A, Cason TN (2009) Indirect punishment and generosity toward strangers. Science 326, 1701–1704. 10.1126/science.1178883 [DOI] [PubMed] [Google Scholar]
- 32. Fehr E (2004) Human behaviour: Don’t lose your reputation. Nature 432, 449–450. 10.1038/432449a [DOI] [PubMed] [Google Scholar]
- 33. Axelrod R (1986) An evolutionary approach to norms. Amer Polit Sci Rev 80(4), 1095–1111. 10.2307/1960858 [DOI] [Google Scholar]
- 34. Yamagishi T, Takahashi N (1994) Evolution of norms without metanorms In: Schulz U, Albers W, Mueller U, editors. Social Dilemmas and Cooperation, Berlin: Springer, pp. 311–326. [Google Scholar]
- 35. Boyd R, Richerson P (1992) Punishment allows the evolution of cooperation (or anything else) in sizable groups. Ethol Sociobiol 13, 171–195. 10.1016/0162-3095(92)90032-Y [DOI] [Google Scholar]
- 36. Panchanathan K, Boyd R (2004) Indirect reciprocity can stabilize cooperation without the second-order free rider problem. Nature 432, 499–502. 10.1038/nature02978 [DOI] [PubMed] [Google Scholar]
- 37. Li Y, Yamagish T (2014) A test of the strong reciprocity model: Relationship between cooperation and punishment. Jap J Psychol 85(1) 100–105. 10.4992/jjpsy.85.100 [DOI] [PubMed] [Google Scholar]
- 38. Horne C (2001) The enforcement of norms: Group cohesion and meta-norms. Soc Psychol Q 64(3), 253–266. 10.2307/3090115 [DOI] [Google Scholar]
- 39. Horne C (2007) Explaining norm enforcement. Rational Soc 19(2), 139–170. 10.1177/1043463107077386 [DOI] [Google Scholar]
- 40. Yamagishi T, Horita Y, Mifune N, Hashimoto H, Li Y, Shinada M, et al. (2012) Rejection of unfair offers in the ultimatum game is no evidence of strong reciprocity. Proc Natl Acad Sci USA 109(50) 20364–8. 10.1073/pnas.1212126109 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Egloff B, Richter D, Schmukle SC (2013) Need for conclusive evidence that positive and negative reciprocity are unrelated. Proc Natl Acad Sci USA 110(9) E786 10.1073/pnas.1221451110 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Peysakhovich A, Nowak MA, Rand DG (2014) Humans display a ‘cooperative phenotype’ that is domain general and temporally stable. Nat Commun 5, 4939 10.1038/ncomms5939 [DOI] [PubMed] [Google Scholar]
- 43. Kiyonari T, Barclay P (2008) Cooperation in social dilemmas: Free riding may be thwarted by second-order reward rather than by punishment. J Pers Soc Psychol 95(4), 826–842. 10.1037/a0011381 [DOI] [PubMed] [Google Scholar]
- 44. Hilbe C, Traulsen A, Röhl T, Milincki M (2014) Democratic decisions establish stable authorities that overcome the paradox of second-order punishment. Proc Natl Acad Sci USA 111(2), 752–756. 10.1073/pnas.1315273111 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Matsumoto Y, Jin N (2010) Co-evolution of leader traits and member traits in social dilemmas. Jap J Experim Soc Psychol 50(1), 15–27. 10.2130/jjesp.50.15 [DOI] [Google Scholar]
- 46. Yamagishi T (1986) The provision of a sanctioning system as a public good. J Pers Soc Psychol 51, 110–116. 10.1037/0022-3514.51.1.110 [DOI] [Google Scholar]
- 47. Oda T (1990) Evolutional approach to the emergence problem of order—application of metanorms game -. Sociol Theo Met 5(1), 81–99. [Google Scholar]
- 48. Galán JM, Izquierdo LR (2005) Appearances can be deceiving: Lessons learned re-implementing axelrod’s ‘evolutionary approach to norms’. J Art Soc Soc Sim 8(3), No.2. [Google Scholar]
- 49. Galán JM, Latek MM, Rizi SMM (2011) Axelrod’s metanorm games on networks. PLoS ONE 6(5), e20474 10.1371/journal.pone.0020474 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Yamamoto H, Okada I (2011) Effect of a social vaccine. Proc 7th European Social Simulation Association Conference.
- 51. Sasaki T, Brännström Å, Dieckmann U, Sigmund K (2012) The take-it-or-leave-it option allows small penalties to overcome social dilemmas. Proc Natl Acad Sci USA 109(4), 1165–1169. 10.1073/pnas.1115219109 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52. Kendal J, Feldman MW, Aoki K (2006) Cultural coevolution of norm adoption and enforcement when punishers are rewarded or non-punishers are punished. Theor Popul Biol 70(1), 10–25. 10.1016/j.tpb.2006.01.003 [DOI] [PubMed] [Google Scholar]
- 53. Herrmann B, Thoni C, Gächter S (2008) Antisocial punishment across societies. Science 319, 1362–1367. 10.1126/science.1153808 [DOI] [PubMed] [Google Scholar]
- 54. Szolnoki A, Perc M (2013) Correlation of positive and negative reciprocity fails to confer an evolutionary advantage: Phase transitions to elementary strategies. Phys Rev X 3, 041021. [Google Scholar]
- 55. Chen X, Perc M (2014) Optimal distribution of incentives for public cooperation in heterogeneous interaction environments. Front Behav Neurosci 8, 248 10.3389/fnbeh.2014.00248 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
All relevant data are within the paper.