Skip to main content
PLOS One logoLink to PLOS One
. 2024 Feb 9;19(2):e0297213. doi: 10.1371/journal.pone.0297213

The art of compensation: How hybrid teams solve collective-risk dilemmas

Inês Terrucha 1,2,*, Elias Fernández Domingos 3,4, Francisco C Santos 5,6, Pieter Simoens 1, Tom Lenaerts 2,3,4,7,*
Editor: The Anh Han8
PMCID: PMC10857581  PMID: 38335192

Abstract

It is widely known how the human ability to cooperate has influenced the thriving of our species. However, as we move towards a hybrid human-machine future, it is still unclear how the introduction of artificial agents in our social interactions affect this cooperative capacity. In a one-shot collective risk dilemma, where enough members of a group must cooperate in order to avoid a collective disaster, we study the evolutionary dynamics of cooperation in a hybrid population. In our model, we consider a hybrid population composed of both adaptive and fixed behavior agents. The latter serve as proxies for the machine-like behavior of artificially intelligent agents who implement stochastic strategies previously learned offline. We observe that the adaptive individuals adjust their behavior in function of the presence of artificial agents in their groups to compensate their cooperative (or lack of thereof) efforts. We also find that risk plays a determinant role when assessing whether or not we should form hybrid teams to tackle a collective risk dilemma. When the risk of collective disaster is high, cooperation in the adaptive population falls dramatically in the presence of cooperative artificial agents. A story of compensation, rather than cooperation, where adaptive agents have to secure group success when the artificial agents are not cooperative enough, but will rather not cooperate if the others do so. On the contrary, when risk of collective disaster is low, success is highly improved while cooperation levels within the adaptive population remain the same. Artificial agents can improve the collective success of hybrid teams. However, their application requires a true risk assessment of the situation in order to actually benefit the adaptive population (i.e. the humans) in the long-term.

Introduction

As Artificial Intelligence (AI) systems are making more and more decisions autonomously, we are relinquishing decision control, for example by allowing intelligent machines to accomplish some of our goals independently or alongside us (e.g., using Google translate to enable business opportunities across different languages [1]), within the context of hybrid human-machine socio-technical systems (e.g., sharing the road with self-driving cars [2]). Given the extraordinary difficulties humans have demonstrated when trying to overcome global crises, such as the COVID-19 pandemic [3] or climate change [4], the question can be raised on whether and how AI agents may help to resolve the problems in coordinating the efforts in those and similar mixed-motive situations.

Even though many different works have advocated for the introduction of beneficial AI to promote human prosociality [57], others have pointed out that humans may be keen to exploit this benevolent AI behavior in their own favor [811]. Thus, before flooding society with AI applications with the promise that they could solve some of the most pressing societal issues, it is worth asking: What behavioral responses can be expected in the presence of AI partners? How may decision-making potentially be affected? Will hybrid groups involving AI agents with predefined decision processes (and humans capable of adapting their own behavior freely) actually achieve greater collective success?

We frame here these questions within the context of the Collective Risk Dilemma (CRD) [4], a game that abstracts the conflict between helping the group to achieve a future goal at a personal cost, or free ride on the efforts of others and just collect the gains associated with achieving the goal (see Methods). The CRD is a public goods game with a delayed reward that is associated with societal problems like pandemic control through vaccination, climate negotiation to achieve CO2 reduction and energy-grid usage by prosumers. Many experiments to assess human behavior have been performed [4, 1215], where [16] found that replacing all human action by artificial delegates has a positive impact on the success rate in the CRD. They also showed that hybrid groups, i.e. groups made of both humans and artificial agents with a pre-defined, non-adaptive behavior, do not achieve higher success rates than non-hybrid (human-only) groups of the same total group size.

Here we expand on that experimental work and hypothesize that the way one evaluates the performance of the hybrid team, namely, the control group used to compare it with, will influence the resulting observations. In the experimental work carried out in [16], on the one hand, the control group is the original non-hybrid team of the same size, which by the substitution of some group members by artificial agents forms the hybrid team. On the other hand, one may consider also the notion of hybrid group formation via addition: artificial agents are introduced in non-hybrid teams without substituting any pre-existing group members. Changing the original non-hybrid team, or the context of team formation, in this manner, may have repercussions on the performance metrics of the resulting hybrid team (see Methods). This manuscript aims to unravel the behavioral dynamics of hybrid human-agent groups, and provide knowledge on how to better assess and curate human-AI teams to tackle mixed-motive and competitive scenarios of varied risk probabilities in accordance to the non-hybrid context from which they derive (substitution or addition).

Two types of participants are considered in the hybrid groups, i.e. adaptive individuals that can change their behavior over time based on the outcome of their interactions (a proxy for potential human responses) and artificial agents that capture an automated probabilistic response that does not change as a result of the interactions (a proxy for average artificial agent behavior). We apply a social learning approach (see Methods) to alter the strategy of the first type of individuals. They can switch between the possible actions in function of their success in the interactions; both when there are other adapting individuals or stochastic artificial agents in groups of a given size. Such social adaptation can be achieved in different ways (e.g. Roth-Erev learning [17], Q-learning [18]), but here an evolutionary game theoretical approach is considered wherein strategic behaviors change in the adaptive individuals population by imitating those individuals that are performing the best [1923].

In our model, the automated response of every artificial agent does not consider continuous learning; rule-based systems are often used in AI products and they are hard-wired in the systems, as learning on the fly might be costly or even dangerous. Thinking about real-world AI applications, one should always consider that producers of AI products want to give guarantees on what the product does (also on what its limitations are, which is why we model agent behavior as a stochastic process that includes errors), and that allowing for extensive adaptation while in use may be very risky. It is also important to note that in this work, we are not considering the artificial agent designer, neither the dynamics involved behind them, whose self-interest might lead to the implementation of different behaviors in the artificial agents. We are simply probing: If we consider this space of behaviors for artificial agents, what kind of human behaviors may potentially emerge given constant hybrid interactions? Even though CRD scenarios may be used to model very high risk events like a pandemic or the climate change, the same kind of non-linearity could be observed within many industrial or software hybrid teams, where, if the project is not delivered, may suffer the consequences of losing their bonuses or even their jobs. Moreover, many teams may already could be considered hybrid if one takes into account the productivity software agents that are already used to make development or business processes more efficient.

We will show that, at least when the risk is high, the adaptive individuals in hybrid teams will exploit the benevolence of the artificial agents in their group, by avoiding to contribute with cooperative efforts when their artificial counterparts are already meeting the threshold needed, as previously hinted to in [11]. On the contrary, when the artificial agents added to each group are associated with a lower capacity to contribute for the collective endeavor, because the risk of loss is high, the adaptive population will boost its cooperation levels to compensate for the bad behavior of the agents. However, when we assess whether or not the hybrid team performs better than the original non-hybrid team, we generally find that success rates only increase in lower risk situations. The exception being when the original non-hybrid team was actually unable to reach the threshold on its own, in which case adding artificial members to its group will always increase the success rate in the long-term.

Related work

In [8] it is pointed out that more experimental research is needed to really understand how human strategic decision-making changes when interacting with autonomous agents. Following on this [9], compiles a review of more than 90 experimental studies that have made use of computerized players. Its main conclusions validate that indeed, human behavior changes when some of the other players are artificial, and furthermore, the behavior deviates to become more rational (or in other words, selfish), where humans are observed to actually try to exploit the artificial players.

This last conclusion was both supported by [10, 11]. The first finds that humans cheat more against machines than against other humans, and thus prefer to play with machines, in an experiment that tested honesty in opposition to the possibility of higher financial gains. The latter recently published an experimental study that concludes that humans are keen on exploiting benevolent AI in various different classical social dilemma games. Within the context of the CRD used for the present work [16], groups participants in hybrid teams with AI agents. Even though 3 out of the 6 group members were AI agents that were successful in avoiding the risky outcome in previous treatments, the hybrid groups were not more successful than only-human groups. Looking closer at the results, one can see that the average payoff of the humans in hybrid teams actually increases. These experimental results already hint towards the adoption of a compensatory behavior on part of the human members of the group once they are informed about the addition of somewhat collaborative agents to the group.

In contrast with aforementioned works [5, 6] point, towards the possibility of engineering prosociality in human behaviour through the use of pro-social AI agents. In the pursuit of this idea [7], assembles a comprehensive review on the use of robots and virtual agents to trigger pro-social behaviour. Out of 23 studies included, 52% reported positive effects in triggering such cooperative behavior. However, 22% were inconclusive and 26% reported mixed results. Moreover, while recent experimental works show that programming autonomous agents [24] that include emotion [25] or some form of communication [26] may positively impact human cooperation, it is still unclear what are the mechanisms facilitating this effect.

More directly related to our theoretical study, there are different works on the dynamics of how evolving populations adapt their behavioral profile given the introduction of agents with a fixed behavior (usually cooperative) either at the group level or at the population level [2732]. With our research questions, we also aim at understanding how the introduction of agents with a fixed behavior, not necessarily cooperative, affects the evolution of cooperation.

Methods

The one-shot Collective Risk Dilemma (CRD)

In this manuscript we adopt the N person one-shot CRD [18, 3338]. Here, a group of N individuals must each decide whether to Cooperate (C), by contributing a fraction c of their initial endowment b, or to Defect (D) and contribute nothing. If the group contains at least M C players, i.e., the group contributes in total Mcb (MN) to the public good, then each player may keep whatever is left of their initial endowment. Otherwise, there is a probability r that all players will loose all their savings and receive a payoff of 0, hence the dilemma. Thus, the expected payoff of a D and a C player can be defined in functions of the number of Cs in the group, j:

πD(j)=b(1-r+rθ(j-M)) (1)
πC(j)=πD-cb (2)

where θ(x) is the Heaviside unit step function, with θ(x) = 0 if x < 0 and θ(x) = 1 otherwise.

CRD with hybrid interactions

We consider a population H of Z adaptive individuals which are randomly sampled into groups of size Na to play the CRD with a artificial agents from population A (whose individuals display a fixed averaged behavior). This allows us, as explained in the section below, to investigate the population dynamics of this dilemma. When engaging in group interactions, each adaptive individual can either cooperate C or defect D. The state of the population is then defined by the number of cooperators k ∈ [0, Z]. The behavior of the fixed artificial agents is defined by their probability of cooperating in each interaction, p ∈ [0, 1], thus, they implement a stochastic (or mixed) strategy. In each group we can calculate the expected payoff of Ds or Cs in function of the number of cooperators from the adaptive population, i, the number of fixed agents a and the payoff of a D (C) πD(C):

ΠD(C)(i,a,p)=pπD(C)(j=i+a)+(1-p)πD(C)(j=i) (3)

The behavioral dynamics exhibited by the population of adaptive individuals are governed by a social learning mechanism, where two randomly chosen individuals compare their fitness and imitate the one who is more successful within their social environment [1923]. Their fitness is the measure of the success of their current strategy (their payoff) averaged over all different group interactions. It can be defined as a function of the aforementioned variables by taking into account the population state and the payoffs given by Eqs (1) and (2). Following on this, the fitness equations for cooperative (C) and defective (D) strategies, can be written as:

fC=(Z-1N-a-1)-1i=0N-a-1(k-1i)(Z-kN-a-1-i)ΠC(i+1,a,p) (4)
fD=(Z-1N-a-1)-1i=0N-a-1(ki)(Z-k-1N-a-1-i)ΠD(i,a,p). (5)

Each individual in the adaptive population may change its strategy profile at a given evolutionary step in the following way: an agent with a D (C) strategy is randomly selected from the population H to adapt. With probability μ it will mutate into a C (D) strategy, otherwise, with probability 1 − μ, it will compare its fitness with another randomly selected individual (assuming the newly selected individual has a different strategy) [1923, 33, 39]. In case imitation is selected, a D (C) strategy will turn into a C (D) with a probability

P(DC)=11+e-β(fC-fD) (6)

described by the Fermi function. This changes the state of the population H of adaptive agents from k to k + 1. This probability becomes higher with a larger difference between the fitness of the two agents, fCfD, or with a larger selection strength of the process, β.

The transition probabilities that regulate the stochastic dynamics of population H, by defining the probability of increasing (+) or decreasing (-) the number of cooperators within a population are given by:

T+(k)=Z-kZ((1-μ)kZ-1P(DC)) (7)
T-(k)=kZ((1-μ)Z-kZ-1P(CD)) (8)

where P(CD) is obtained by replacing C with D, and D with C in Eq (6).

From these equations, we can construct the complete Markov chain of the Z + 1 different states that fully describe the evolutionary process of the population H. From this Markov Chain we can compute the stationary distribution P(k), the average cooperation level C¯ and the average group success s¯G of each population configuration.

To compute the stationary distribution P(k), we retrieve the eigenvector corresponding to the eigenvalue 1 of the tridiagonal transition matrix S = [pij]T [20, 22, 23]. The values pij are defined by the equations:

pk,k±1=T±(k) (9)
pk,k=1-pk,k-1-pk,k+1 (10)

where the formulas that define T±(k) can be consulted in Eqs (7) and (8).

From this it follows that the cooperation level C¯ of population H (for a given set of parameters N, M, r, a and p) by averaging the fraction of cooperators in each population state, k/Z, over the stationary distribution of states P(k) as given by:

C¯=k=0ZP(k)kZ (11)

As already mentioned, within the context of the CRD [4, 18, 3338] another relevant quantity to derive is the probability of success of each group in reaching the threshold necessary of M cooperators to avoid the collective risk.

At the population level, we compute the fraction of groups in each population state that are successful by resorting to the multivariate hypergeometric sampling, as follows

sG(k)=(ZN-a)-1h=0N-a(kh)(Z-kN-h-a)××(pθ(h+a-M)+(1-p)θ(h-M)) (12)

where θ(x) is the Heaviside unit step-function as in Eqs (1) and (2) of the main text.

Finally, similarly to what was done in Eq (11), we calculate the averaged group success by weighing the group success of each population state of Eq (12) over the stationary distribution of the evolutionary process P(k):

s¯G=k=0ZP(k)sG(k) (13)

The code used to compute all the aforementioned quantities and reproduce the results that constitute this manuscript is made available at [40].

Forming hybrid teams through substitution or addition

As mentioned above, hybrid teams are formed by Na adaptive individuals and a artificial agents with a fixed behavior. Using Eqs (11) and (13) we are able to calculate the average cooperation level and success rate of a population H that evolves by playing the CRD in hybrid teams with that configuration. To determine how well the adaptive population fares when playing in hybrid teams, we should be able to compare these metrics with the results obtained for non-hybrid teams, i.e. teams with a = 0, therefore only made of N adaptive individuals.

To this end, we introduce the idea of a control population, corresponding to a population HC that evolves by playing the CRD in non-hybrid teams, and for which we compute both its cooperation level C¯C and its success rate s¯C (subscript C refers to the control) following Eqs (11) and (13), respectively, by adjusting the parameters N and a according to the control configuration of its non-hybrid teams. We assume that hybrid teams are a result of either one of two different processes—substitution or addition -, so that the control chosen to evaluate the relative results of the hybrid team configuration differ, depending on the process used to form the hybrid team.

The process of substitution refers to the formation of a hybrid team out of a non-hybrid team of N group members by substituting a number a of these group members by artificial agents. In this case, the non-hybrid team used as control will have N of adaptive individuals only (a total group size NC = N), whereas the hybrid team whose performance we are evaluating will also have N group members, where Na are adaptive individuals and a are artificial agents. To calculate cooperation level and success rate for the control population with a non-hybrid team group configuration—C¯S and s¯S (with a subscript S for substitution) -, we simply take Eqs (11) and (13) and substitute a → 0, maintaining all the other variables already used to compute C¯ and s¯G for the case of hybrid teams. In this case, we can write the relative gain in cooperation level ΔC¯S and the relative gain in success rate Δs¯S by the process of substitution as shown below in Eqs (14) and (15), respectively.

ΔC¯S=C¯G-C¯S=C¯G-C¯G(a0) (14)
Δs¯S=s¯G-s¯S=s¯G-s¯G(a0) (15)

Another way to form hybrid teams, rather than substituting group members by artificial agents, is to simply add new members to the non-hybrid team which are artificial agents—a process we define as addition. In this case, the non-hybrid team used as control will be made of NC = Na adaptive individuals only, and it is to be compared with hybrid teams made of Na adaptive individuals and a artificial agents, so N = NC + a. We consider this case to be a more useful comparison for real applications, since from the ethical standpoint on the future of work, hybrid teams should be created as an attempt to enhance human performance, rather than to substitute human labour [41]. In this case, to calculate the cooperation level and success rate of the control population, C¯A and s¯A (with a subscript A for addition), we use the same Eqs (11) and (13) as for the accompanying hybrid scenario but we substitute a → 0 and NNa, since for this control there are no artificial agents and the group size of the team is considered to be the same as the number of adaptive individuals that are in the hybrid team (NC = Na in this case). For this case of addition, Eqs (16) and (17) show how to calculate the gain in cooperation level (ΔC¯A) and in success rate (Δs¯A) obtained by the formation of hybrid teams.

ΔC¯A=C¯G-C¯A=C¯G-C¯G(a0,NN-a) (16)
Δs¯A=s¯G-s¯A=s¯G-s¯G(a0,NN-a) (17)

Results and discussion

Compensation ensures high success, but adaptive individuals do not cooperate when artificial agents do

To probe the question of whether or not to introduce artificial agents and form hybrid teams to tackle a CRD, one of the first metrics to study is the success rate (Eq (13)) obtained by such groups. Fig 1 shows how the average group success of an evolving population of adaptive individuals fares if they were playing the CRD within hybrid teams in three conditions; i) a < M < Na (Fig 1A), ii) a = M = Na (Fig 1B) and iii) a > M > Na (Fig 1C). The difference between all three panels is in which type of individuals, i.e. adaptive (who are proxies for humans) or fixed (who are proxies for artificial agents) behaviour individuals, can achieve group success by themselves: in Fig 1A the threshold can be solely met by the adaptive ones, in Fig 1B both the adaptive and the artificial agents can reach the threshold alone, while in Fig 1C only the artificial agents can reach the threshold M by themselves. In all three scenarios, artificial and adaptive agents can cooperate together to reach the threshold M because (Na) + a = 6 which is always bigger than M = 3.

Fig 1. Average group success obtained for an adaptive population of individuals playing a CRD in hybrid groups.

Fig 1

All images portray hybrid groups of group size N = 6 where at least M = 3 must be cooperators in order to avoid risking losing their endowments with risk probability r in the yaxis. The xaxis shows the probability of the a artificial agents in the group, who implement a fixed stochastic strategy, cooperating: p. Average group success is illustrated following the color bar on the left side of the figure. In (A) hybrid groups are made of Na = 4 adaptive individuals and a artificial agents, so that a < M < Na and adaptive individuals are able to meet the threshold on their own. In this case, for higher risk r, the hybrid groups achieve higher success, a trend that is fairly independent from the probability of the artificial agents cooperating p. In (B) a = M = Na = 3, again adaptive individuals can meet the threshold M on their own, but the dilemma is harder since all of them would have to cooperate in order to do so. Two regions of high success are surrounded by a shadow of lower success rates: when p is very high, independently of risk probability r; and a semi-circle on the top left corner of the picture where risk is high and p is low. (C) represents hybrid groups made of Na = 2 adaptive individuals and a = 4 artificial agents, so that adaptive individuals would not be able to meet the threshold M on their own as a > M > Na. In this case, average group success appear to change only in accordance to the probability of the artificial agents cooperating p, independently from the risk of collective loss r. These results follow Eq (13), using the parameters: Z = 100, μ = 0.01, β = 2, b = 1, c = 0.1.

Starting by Fig 1A, one can immediately distinguish between high-risk scenarios where success rate is almost always very high, and low-risk scenarios where reaching success appears to be more difficult (almost) independently of other variables. Prior work on the CRD [18, 3338] identified this non-linearity in the game, which is related to the fact that in high-risk situations, individual payoffs are highly dependent on collective success whereas in lower risk CRDs, it is more profitable to take the risk than to cooperate. In Fig 1A, where a < M and therefore the success of the hybrid team depends on the Na adaptive participants of the group, this trend is clearly identifiable. In this case, variable p (see Eq (3)), that identifies the given average level of cooperative behavior of the artificial agents in the group, appears to simply dim the risk boundary that separates high and low success rates, with the transition being very abrupt for lower p but smoother for higher p, in which case the success rate is never very low, even for low-risk scenarios.

When a = Na = M in Fig 1B, and therefore the threshold could be met by either the artificial or the adaptive agents (or a combination of both, as in all cases), we see that indeed when p is high—the artificial agents are very cooperative -, high success rates are uniformly achieved for any risk factor. This is due to the fact that cooperation by the artificial agents alone is sufficient to secure collective success for this case of aM. However, for high risk, there is a shadow of lower success rates for decreasing p. A semi-circle of high success can also be observed in the top left region (high risk and low p). Since within this semi-circle the cooperative level of the agents p would not be sufficient to ensure collective success, this semi-circle must be associated with high levels of cooperation on the part of the adaptive individuals. Yet, and this is highly intriguing, a shadow of lower success separates this semi-circle from the region where the artificial agents are cooperative enough to achieve success by themselves. When comparing that region of reduced success between the two extremes with the results in Fig 1A, the adaptive individuals seem to have become less effective in achieving the goal, even for high risk.

Finally, Fig 1C depicts a situation where the adaptive population by itself would not be able to meet the required threshold to avoid collective disaster: Na < M. The figure shows how collective success is completely dependent on the cooperation level encoded in the behavior of the artificial agents in the group, i.e. p. When p is sufficiently high, then maximum success rates are observed, and when p drops towards 0, success rate plummets accordingly.

To explain high success rates in regions of low p, or when a < M, in Fig 1, we hypothesize that the cooperation level of the adaptive population must be compensating the insufficient cooperative effort of the artificial agents in these cases. To support this hypothesis, Fig 2 shows the average cooperation level observed for the adaptive population only, following Eq (11).

Fig 2. Average cooperation level obtained for an adaptive population of individuals playing a CRD in hybrid groups.

Fig 2

The same set-up (using the same parameter choices) used in Fig 1 is recreated, but now we calculate average cooperation level following Eq (11). In (A) where Na = 4 are adaptive individuals and a artificial agents, so a < M < Na, we see that the average level of cooperation in the adaptive population appears to solely explain the average group success observed in Fig 1A. In (B), where a = M = Na = 3, adaptive individuals can again meet the threshold M on their own, but the dilemma is harder than in A), since all of them would have to cooperate in order to do so. High cooperation levels associated with the adaptive population are restricted to an enclosed semi-circle on the top left of the figure, for high risk probability r and for lower cooperation levels on the part of the artificial agents p. (C) represents hybrid groups made of Na = 2 adaptive individuals and a = 4 artificial agents, so that adaptive individuals would not be able to meet the threshold M on their own as a > M > Na. In this case, the adaptive population does not cooperate at all.

In Fig 2A, where adaptive individuals can reach the threshold on their own Na > M, we observe that their cooperation level alone is able to explain how the success rate changes with respect to r and p, showing an almost fully correlated result to the one of Fig 1A. Interestingly the cooperation level for the adaptive population presented in Fig 2B can only account for one of the regions of high success previously identified in Fig 1B, given that the other one is connected with the cooperative efforts of the artificial agents. Assigning an important role to the artificial agents in the group, alters the overall behavior of the adaptive individuals: they become more cooperative in the high-risk region but restrict their effort to a specific area of the parameter space. Finally, since the adaptive population presents no cooperation whatsoever in Fig 2C, we can confirm that the success rates observed in Fig 1C are fully justified by the efforts of the artificial agents in this case. Overall, we see that success can be obtained for a wide array of different parameters (r and p) through a mechanism of compensation developed by the adaptive population to cooperate whenever the artificial agents are not cooperating enough.

However, the same mechanism will lead adaptive individuals to fully refrain from cooperation when the number of artificial agents in the group and their cooperative effort is perceived to be sufficient to attain collective success. In a way, the compensation mechanism can then almost be described as an exploit of the artificial agents by part of the adaptive population. Indeed, a negative consequence of this evolution of the adaptive population, is that even for high risk scenarios, less than optimal success rates are obtained in Fig 1B and 1C. In Fig 1B, the shadow of less than optimal average success rate could have been avoided if the adaptive population was to cooperate more outside of the semi-circle observed in Fig 2B; and in Fig 1C success could have extended for a wider range of p values if the adaptive population would have cooperated a bit to complement the efforts of less cooperative artificial agents. In both cases, a possible expectation mismatch about the artificial agents’ efforts led the adaptive population to relax too much their own cooperative load.

Hybrid teams achieve higher success rates in low-risk CRDs than teams without artificial agents

One of the main arguments for introducing artificial agents in a group to form hybrid teams with adaptive agents, is precisely to test it as a solution to enhance group success in the CRD, a question experimentally analysed in [16]. We thus analyse the two types of hybrid group formation mechanisms discussed in the Introduction (see also Methods), and ask the question of when do the results reported in Figs 1 and 2 actually signify an improvement in success and cooperation when compared with non-hybrid teams (see next Section). In Fig 3 the results for the substitution approach (see Methods) are provided wherein a of the non-hybrid group members are replaced by artificial agents. The expected gain in group success is calculated using Eq (15), where the observations as in Fig 1 are subtracted from the success of fully non-hybrid groups.

Fig 3. The effect of substitution, as a process to form hybrid teams, on success rate.

Fig 3

The parameters referring to each hybrid group configuration are shown on top of each panel, as is the control parameters that allow us to calculate s¯S (success rate for the non-hybrid groups before substitution). As in Fig 1: in (A) Na = 4 and a = 2, in (B) Na = a = 3 and in (C) Na = 2 and a = 4. The plot shows that substituting part of the group by artificial agents with different cooperative levels only improves success when the risk probability is low. For high-risk CRD scenarios, substituting a high number of group members by artificial agents will greatly reduce the success rate that would otherwise be obtained by adaptive individuals playing the game alone. All images are produced following Eq (15) and use the same other parameters as the previous figures in this manuscript.

A general pattern can be identified where we observe that success is generally enhanced in low risk scenarios or in high risk scenarios where there are few agents (Naa, in panels A and B) and they are not very cooperative (low p). However, when risk is high and agents are cooperative, success appears to decrease in comparison with groups that would be made of N = 6 humans with the same threshold M = 3. For the special case of Fig 3C where Na < a, when the agents are not cooperative and risk is high there is a big decrease in success when compared with groups made only of adaptive individuals, probably because the latter do not cooperate at all as we have observed in Fig 2C when grouped in hybrid teams of such a configuration.

Of course, substituting adaptive individuals by artificial agents is not the ideal application in a real-world scenario. AI ethics and the concerns for the future of work [41] indicate that AI agents (for which the artificial agents in this model try to account for) should be introduced in working teams to enhance the work that is already being performed, rather than to substitute human labour. With this in mind, we investigate the effect of addition (see Methods) as another process of forming hybrid teams. In Fig 4, we therefore compare the hybrid team configuration with a non-hybrid team made of the same number of NC adaptive agents, thus without the added a artificial agents.

Fig 4. The effect of addition, as a process to form hybrid teams, on success rate.

Fig 4

Again we re-use the same hybrid team configuration for each of the panels as in the previous figures. The control is now given by addition, changing the control parameters to the ones indicated in each panel title: in (A) the control group configuration consists of N = 4 adaptive agents, in (B) of N = 3 adaptive agents and in (C) of N = 2 adaptive agents. Note that unlike in Fig 3, the group configuration of the control in panel C is unable to meet the threshold without the added artificial agents. The results plotted follow Eq (17) and show that adding artificial agents to the group only improves success when the risk probability is low or when the adaptive individuals would not be able to reach the threshold M on their own (as in panel C where Na < M). For high-risk CRD scenarios where NaM, adding artificial agents will only reduce or, at best, maintain the success rate they would otherwise already achieve on their own.

Similarly to Fig 3, we observe that when the risk of collective disaster is low, adding artificial agents to the group will always benefit collective success. However, when the risk is high and we add a small number of cooperative agents to the group, this can hurt group achievement (Fig 4A and 4B). The main difference between substitution as shown in Fig 3 and addition as now plotted in Fig 4, as a mechanism to better group performance in the CRD, is that adding artificial agents to a group that would otherwise not be able to meet the threshold M will always increase the success of that group as long as the artificial agents are somewhat cooperative (p > 0). We also observe that for low p in Fig 4 there is no slight improvement of success in high risk scenarios, something that was also observed in Fig 3A and 3B, but a situation that we do not consider of relevance since it would be unlikely for an institution or organization to purposefully add non-cooperative agents to a hybrid team tackling a high-risk CRD.

At high risk, highly cooperative artificial agents put the cooperative capacity of adaptive individuals at risk

Even though success in avoiding the risky outcome is an important metric when studying the CRD, we find it relevant to also understand how the introduction of artificial agents to form hybrid teams affects the evolution of cooperation. As previously mentioned in the introduction, the world is growing more dependent on the use of AI applications to support human decision-making and performance, though we still have little understanding of how this machine dependence might affect human skills in the long-term. Limiting our study to social dilemma situations, the skill in question is precisely the human capacity for cooperation. Fig 5 presents what the effect of forming teams through a process of substitution might have on the average cooperation level obtained by the population of adaptive individuals. The figure shows again the results for the 3 different cases, as was done in Fig 3, now following Eq (14) to calculate the gain in cooperation.

Fig 5. The effect of substitution, as a process to form hybrid teams, on cooperation level.

Fig 5

All images were produced using the same parameters as the ones in Fig 3, now following Eq (14) to calculate the gain in cooperation level, rather than the effect on success rate. As indicated in the titles of each panel, the hybrid team configuration once again refers to: in (A) Na = 4 and a = 2, in (B) Na = a = 3 and in (C) Na = 2 and a = 4. The plot shows that substituting part of the group by cooperative artificial agents will generally decrease the cooperation exhibited by the population if it was grouped in non-hybrid groups of the same group size N. The exception being the case of panel A, where Na > M > a for the region of low-risk. Substituting part of the group by less cooperative artificial agents might increase cooperation in the adaptive population when Naa and risk is high, although it is an unlikely scenario in real-life applications for these to be introduced in high-risk CRD scenarios.

Examining Fig 5, we can see that cooperation can be enhanced by the constitution of hybrid teams through the substitution of some group members by artificial agents, but only if certain conditions are met. When risk is high, we see that cooperation can be enhanced when the artificial agents in the group are not so cooperative (low p) and there are still at least as many humans in the group as the number of cooperators needed to meet the threshold (NaM), as in panels A and B of Fig 5. However, when the artificial agents are cooperative (high p), risk is high (high r)—reference to panels A and B—or the number of adaptive agents in the group would not suffice the cooperation requirements to achieve success on their own (Na < M)—panel C -, then the adaptive population would exhibit higher levels of cooperation if not playing the CRD in a hybrid team. Again, following the analysis of Figs 3 and 4, we find that when the risk of collective disaster is low enough, substituting adaptive individuals by artificial agents has essentially no effect on cooperation (the exception being the small boost found in the case of Na > M > a in Fig 5A).

As previously discussed however, the hybrid team configuration might spur from the addition of artificial agents to a team made of Na adaptive agents, instead of substituting a adaptive agents from a team of N. To study this effect of addition in the forming of hybrid teams, we produce Fig 6, where we use Eq (16) to plot the difference between the results already shown in Fig 2 and the results that an adaptive population would have obtained if playing in teams made only of Na = 4 (panel A), Na = 3 (panel B) and Na = 2 (panel C) with a = 0.

Fig 6. The effect of addition, as a process to form hybrid teams, on cooperation level.

Fig 6

To reproduce this figure, the same parameters already used in Fig 4 are used as input into Eq (16). Therefore, as indicated in the subtitles of each panel: in (A) there are Na = 4 adaptive agents in both the control and the hybrid case, with a = 0 in the control and a = 2 otherwise; in (B) Na = a = 3 for the hybrid teams and Na = 3 but a = 0 in the control; and in (C) Na = 2 for both scenarios, but a = 4 in the hybrid case and a = 0 in the control. The plot shows that adding artificial agents to the group can only boost cooperation in the adaptive population when the risk r is low and Na > a. Otherwise, their addition will only decrease or the level of cooperation observed for a population where groups of the same Na number of adaptive individuals try to tackle the CRD on their own. When the risk is high and Naa, adding a cooperative artificial agents to the group can greatly impair the cooperation level observed in the adaptive population, as shown in panels A and B.

Fig 6 further restricts the positive impact that hybrid teams might have in the evolution of cooperation on the adaptive population: only when the risk is low (low r) and the artificial agents added to the team could not reach the threshold on their own (a < M)—focus on Fig 6A -, does the addition of artificial agents benefit the cooperation capacity of the adaptive population. Whereas for low p and high risk r there was also a positive impact in the cooperation capacity when addressing the case of substitution in Fig 5, in this case of addition no difference is observed within this set of parameters. Similarly, even though substituting the greater part of a team by artificial agents would seriously impair that team’s cooperative capacity when the risk is high (Fig 5C), adding artificial agents to a team that would not be able to meet the threshold on their own (Na < M) does not change the cooperation level of that cooperative capacity (Fig 6C).

Conclusions

In this work we investigated what behavior is selected by social learning in the context of the one-shot CRD, when interactions occur in hybrid groups made of adapting individuals (a proxy for human decision-making) and artificial agents with fixed probabilistic behaviors (a proxy for average AI behavior). This model is used as a thought experiment to reason about the behavior one could expect in hybrid groups of humans and AI agents. It focuses on mixed-motive situations where there is a conflict between individual and common interests, but also a risky outcome where the lack of cooperation does not immediately correspond to collective loss. One of the main contributions of this work is to reveal the importance of context and choice of the control group: even though it appears in hindsight as an obvious result, the addition of artificial agents to a group that would not be able to reach the threshold on its own always increases success (Fig 4C). However, the wrong choice of control, for example, if one would compare the hybrid team with a non-hybrid team of the same group size, its performance appears to be very negative both in terms of group success in the high-risk region (Fig 3C) and in terms of the cooperation level (Fig 5C). To our knowledge, this work is the first concerned with this discussion on how the context of the hybrid team formation may influence its relative performance.

The model allowed us to understand that the benefits of introducing artificial agents in teams to help solve a CRD highly depends on the risk associated with the failure to do so: when the risk of loss is low, the introduction of artificial agents in the group always increases success; when it is high, this is not true. We find this model of special interest since most social dilemmas faced by our society also have a probabilistic, rather than deterministic, nature (for example, dealing with pandemics [3] or climate change [4]). Even so, humans will rely on biases and heuristics to evaluate risk, and will often misinterpret it [42]. Within this context, the proposed model is ever more relevant as it shows that an under- or overestimation of the risk of collective loss may lead to a wrong decision of whether or not to form hybrid teams to tackle the dilemma.

Moreover, with the model, we have disentangled how the changes in success rate are related to the effort produced by the artificial agents: whenever the latter are perceived as highly cooperative, the behavior of the adaptive population evolves to exploit the artificial agents’ benevolence. An effect that has also been shown experimentally in [11]. However, when the artificial agents are low contributors, the adaptive population shifts to compensate those low contributions. Yet, this is only true for higher risk levels, as in this case, reaching the goal is the only means to secure payoff. These results are supported experimentally: in [16], hybrid teams were not significantly more successful than only-human teams, but humans were observed to contribute less at the expense of the artificial agents in their groups in a high-risk collective-risk dilemma. On the other hand, when risk is low and there is not enough incentive for the adaptive population to change their behavior, the addition of artificial agents to the group does seem to increase the group success rate. If the hybrid team does not solely depend on the artificial agents to meet the threshold, their addition might even contribute to a spike in the adaptive population in the long-term.

Even though previous experimental work [11, 16] corroborates some of the results reported here, additional behavioral experiments are needed to truly validate all the conclusions. Such experiments are left for future research given that a thorough experiment design with multiple controls (in this case, by using different human-only group sizes to test the hypothesis of addition vs substitution) would require a significant effort and resources.

In general, the discussed research suggests that there is potential benefit of using AI to increase the success rates of human-AI groups working together in low-risk scenarios. Nevertheless, our findings clearly show that the addition of artificial agents will not always be beneficial if the AI designer is also concerned with at least maintaining the cooperation level of the adaptive individuals before the hybrid team was formed. By working alongside cooperative AI, humans might eventually adapt to relax their own cooperative efforts. Hence, we must either identify AI policies that avoid this scenario and still promote cooperation to avoid collective risks or promote other modes of interaction in-between hybrid teams. Our model provides a tool to explore different risk scenarios simulated with different compositions of hybrid teams. Such a tool could become the first step in assessing how to safely constitute hybrid teams to effectively solve CRD-like problems.

Data Availability

All computational files used to reproduce the results contained in this work are available from the Zenodo database at https://doi.org/10.5281/zenodo.10406595.

Funding Statement

I.T., P.S. and T.L are supported by an FWO (Fond Wetenschappelijk Onderzoek) project with grant no. G054919N. E.F.D. is supported by an F.R.S-FNRS (Fonds de la Recherche Scientifique) Chargé de Recherche grant (nr. 40005955). T.L. is furthermore supported by two F.R.S.-FNRS PDR (grant numbers 31257234 and 40007793) and acknowledges also the support by TAILOR, a project funded by EU Horizon 2020 research and innovation programme under GA No 952215. E.F.D. and T.L. are supported by Service Public de Wallonie Recherche under grant n° 2010235–ariac by digitalwallonia4.ai. T.L. and P.S. acknowledge the support by the Flemish Government through the Flanders AI Research Program. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Hasyim M, Yolanda Latjuba A, Akhmar AM, Kaharuddin K, Jihad Saleh N. Human-Robots And Google Translate: A Case Study Of Translation Accuracy In Translating French-Indonesian Culinary Texts. Turkish Journal of Computer and Mathematics Education. 2021;. [Google Scholar]
  • 2. Nair GS, Bhat CR. Sharing the road with autonomous vehicles: Perceived safety and regulatory preferences. Transportation Research Part C: Emerging Technologies. 2021;122:102885. doi: 10.1016/j.trc.2020.102885 [DOI] [Google Scholar]
  • 3.Kunicova J. Driving the COVID-19 Response from the Center: Institutional Mechanisms to Ensure Whole-of-Government Coordination (English). Governance and Institutions Responses to COVID-19 Washington, D C. 2020;.
  • 4. Milinski M, Sommerfeld RD, Krambeck HJ, Reed FA, Marotzke J. The collective-risk social dilemma and the prevention of simulated dangerous climate change. Proceedings of the National Academy of Sciences. 2008;105(7):2291–2294. doi: 10.1073/pnas.0709546105 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Harvey J, Golightly D, Smith A. HCI as a means to prosociality in the economy. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems; 2014. p. 2955–2964.
  • 6.Paiva A, Santos FP, Santos FC. Engineering pro-sociality with autonomous agents. In: 32nd AAAI Conference on Artificial Intelligence; 2018. p. 7994–7999.
  • 7. Oliveira R, Arriaga P, Santos FP, Mascarenhas S, Paiva A. Towards prosocial design: A scoping review of the use of robots and virtual agents to trigger prosocial behaviour. Computers in Human Behavior. 2021;114:106547. doi: 10.1016/j.chb.2020.106547 [DOI] [Google Scholar]
  • 8.Camerer CF. Artificial Intelligence and Behavioral Economics. In: The Economics of Artificial Intelligence: An Agenda. University of Chicago Press; 2019. p. 587–608. Available from: http://www.nber.org/chapters/c14013.
  • 9.March C. The Behavioral Economics of Artificial Intelligence: Lessons from Experiments with Computer Players. Munich: Center for Economic Studies and the ifo institute; 2019. 7926. Available from: https://ssrn.com/abstract=3485475.
  • 10. Cohn A, Gesche T, Maréchal MA. Honesty in the digital age. Center for Economic Studies and the ifo institute; 2018. 6996. [Google Scholar]
  • 11. Karpus J, Krüger A, Verba JT, Bahrami B, Deroy O. Algorithm exploitation: Humans are keen to exploit benevolent AI. iScience. 2021;24(6):102679. doi: 10.1016/j.isci.2021.102679 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Tavoni A, Dannenberg A, Kallis G, Löschel A. Inequality, communication, and the avoidance of disastrous climate change in a public goods game. Proceedings of the National Academy of Sciences. 2011;108(29):11825–11829. doi: 10.1073/pnas.1102493108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Milinski M, Röhl T, Marotzke J. Cooperative interaction of rich and poor can be catalyzed by intermediate climate targets. Climatic change. 2011;109(3):807–814. doi: 10.1007/s10584-011-0319-y [DOI] [Google Scholar]
  • 14. Dannenberg A, Löschel A, Paolacci G, Reif C, Tavoni A. On the provision of public goods with probabilistic and ambiguous thresholds. Environmental and Resource economics. 2015;61(3):365–383. doi: 10.1007/s10640-014-9796-6 [DOI] [Google Scholar]
  • 15. Cadsby CB, Maynes E. Voluntary provision of threshold public goods with continuous contributions: experimental evidence. Journal of Public Economics. 1999;71(1):53–73. doi: 10.1016/S0047-2727(98)00049-8 [DOI] [Google Scholar]
  • 16. Fernández Domingos E, Terrucha I, Suchon R, Grujić J, Burguillo JC, Santos FC, et al. Delegation to artificial agents fosters prosocial behaviors in the collective risk dilemma. Scientific reports. 2022;12(1):8492. doi: 10.1038/s41598-022-11518-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Macy MW, Flache A. Learning dynamics in social dilemmas. Proceedings of the National Academy of Sciences. 2002;99(suppl 3):7229–7236. doi: 10.1073/pnas.092080099 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Domingos EF, Grujić J, Burguillo JC, Santos FC, Lenaerts T. Modeling behavioral experiments on uncertainty and cooperation with population-based reinforcement learning. Simulation Modelling Practice and Theory. 2021;109:102299. doi: 10.1016/j.simpat.2021.102299 [DOI] [Google Scholar]
  • 19. Hofbauer J, Sigmund K, et al. Evolutionary games and population dynamics. Cambridge university press; 1998. [Google Scholar]
  • 20. Nowak MA. Evolutionary Dynamics: Exploring the Equations of Life. Harvard University Press; 2006. [Google Scholar]
  • 21. Traulsen A, Pacheco JM, Nowak MA. Pairwise comparison and selection temperature in evolutionary game dynamics. Journal of theoretical biology. 2007;246(3):522–529. doi: 10.1016/j.jtbi.2007.01.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Traulsen A, Hauert C, De Silva H, Nowak MA, Sigmund K. Exploration dynamics in evolutionary games. Proceedings of the National Academy of Sciences. 2009;106(3):709–712. doi: 10.1073/pnas.0808450106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Hindersin L, Wu B, Traulsen A, García J. Computation and simulation of evolutionary game dynamics in finite populations. Scientific reports. 2019;9(1):1–21. doi: 10.1038/s41598-019-43102-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. de Melo CM, Marsella S, Gratch J. Human cooperation when acting through autonomous machines. Proceedings of the National Academy of Sciences. 2019;116(9):3482–3487. doi: 10.1073/pnas.1817656116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.de Melo CM, Khooshabeh P, Amir O, Gratch J. Shaping cooperation between humans and agents with emotion expressions and framing. In: Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems; 2018. p. 2224–2226.
  • 26. Crandall JW, Oudah M, Ishowo-Oloko F, Abdallah S, Bonnefon JF, Cebrian M, et al. Cooperating with machines. Nature communications. 2018;9(1):1–12. doi: 10.1038/s41467-017-02597-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Vasconcelos VV, Santos FC, Pacheco JM, Levin SA. Climate policies under wealth inequality. Proceedings of the National Academy of Sciences. 2014;111(6):2212–2216. doi: 10.1073/pnas.1323479111 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Mao A, Dworkin L, Suri S, Watts DJ. Resilient cooperators stabilize long-run cooperation in the finitely repeated prisoner’s dilemma. Nature communications. 2017;8(1):1–10. doi: 10.1038/ncomms13800 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Oakley B, Knafo A, Madhavan G, Wilson DS. Pathological altruism. Oxford University Press; 2011. [Google Scholar]
  • 30.Santos FP, Pacheco JM, Paiva A, Santos FC. Evolution of collective fairness in hybrid populations of humans and agents. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 33; 2019. p. 6146–6153.
  • 31. Shirado H, Christakis NA. Locally noisy autonomous agents improve global human coordination in network experiments. Nature. 2017;545(7654):370–374. doi: 10.1038/nature22332 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Sharma G, Guo H, Shen C, Tanimoto J. Small bots, big impact: solving the conundrum of cooperation in optional Prisoner’s Dilemma game through simple strategies. Journal of The Royal Society Interface. 2023;20(204):20230301. doi: 10.1098/rsif.2023.0301 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Santos FC, Pacheco JM. Risk of collective failure provides an escape from the tragedy of the commons. Proceedings of the National Academy of Sciences. 2011;108(26):10421–10425. doi: 10.1073/pnas.1015648108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Abou Chakra M, Traulsen A. Evolutionary Dynamics of Strategic Behavior in a Collective-Risk Dilemma. PLOS Computational Biology. 2012;8(8):1–7. doi: 10.1371/journal.pcbi.1002652 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Vasconcelos VV, Santos FC, Pacheco JM. A bottom-up institutional approach to cooperative governance of risky commons. Nature Climate Change. 2013;3(9):797–801. doi: 10.1038/nclimate1927 [DOI] [Google Scholar]
  • 36. Pacheco JM, Vasconcelos VV, Santos FC. Climate change governance, cooperation and self-organization. Physics of life reviews. 2014;11(4):573–586. doi: 10.1016/j.plrev.2014.02.003 [DOI] [PubMed] [Google Scholar]
  • 37. Hagel K, Abou Chakra M, Bauer B, Traulsen A. Which risk scenarios can drive the emergence of costly cooperation? Scientific reports. 2016;6(1):1–9. doi: 10.1038/srep19269 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Domingos EF, Grujić J, Burguillo JC, Kirchsteiger G, Santos FC, Lenaerts T. Timing uncertainty in collective risk dilemmas encourages group reciprocation and polarization. iScience. 2020;23(12):101752. doi: 10.1016/j.isci.2020.101752 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Pacheco JM, Santos FC, Souza MO, Skyrms B. Evolutionary dynamics of collective action in N-person stag hunt dilemmas. Proceedings of the Royal Society B: Biological Sciences. 2009;276(1655):315–321. doi: 10.1098/rspb.2008.1126 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Terrucha I. Supporting computational notebook for manuscript: “The art of compensation: how hybrid teams solve collective-risk dilemmas”; 2023. Available from: 10.5281/zenodo.10406595. [DOI] [PMC free article] [PubMed]
  • 41. Akata Z, Balliet D, De Rijke M, Dignum F, Dignum V, Eiben G, et al. A research agenda for hybrid intelligence: augmenting human intellect with collaborative, adaptive, responsible, and explainable artificial intelligence. Computer. 2020;53(08):18–28. doi: 10.1109/MC.2020.2996587 [DOI] [Google Scholar]
  • 42. Kahneman D, Slovic P, Tversky A. Judgment under uncertainty: Heuristics and biases. Cambridge university press; 1982. [DOI] [PubMed] [Google Scholar]

Decision Letter 0

The Anh Han

19 Dec 2023

PONE-D-23-33002The art of compensation: how hybrid teams solve collective-risk dilemmasPLOS ONE

Dear Dr. Terrucha,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

ACADEMIC EDITOR: The received feedback from the referee and the associate editor is that the paper is well written and addresses an important and timely problem. It would make an excellent contribution to the theoretical literature of the evolution of cooperation and provides useful insights for the design of prosocial AI. The reviewer has provided constructive suggestions -- please carefully take them into account when preparing the revised version.

==============================

Please submit your revised manuscript by Feb 02 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

The Anh Han, Ph.D.

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for stating the following financial disclosure:

“I.T., P.S. and T.L are supported by an FWO (Fond Wetenschappelijk Onderzoek) project with grant no. G054919N. E.F.D. is supported by an F.R.S-FNRS (Fonds de la Recherche Scientifique) Chargé de Recherche grant (nr. 40005955). T.L. is furthermore supported by two F.R.S.-FNRS PDR (grant numbers 31257234 and 40007793) and acknowledges also the support by TAILOR, a project funded by EU Horizon 2020 research and innovation programme under GA No 952215.. E.F.D. and T.L. are supported by Service Public de Wallonie Recherche under grant n° 2010235–ariac by digitalwallonia4.ai. T.L. and P.S. acknowledge the support by the Flemish Government through the Flanders AI Research Program.”

Please state what role the funders took in the study.  If the funders had no role, please state: "The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

If this statement is not correct you must amend it as needed.

Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf.

3. Thank you for stating the following in the Acknowledgments Section of your manuscript:

“I.T., P.S. and T.L are supported by an FWO (Fond Wetenschappelijk Onderzoek) project with grant no. G054919N. E.F.D. is supported by an F.R.S-FNRS (Fonds de la Recherche Scientifique) Charg´e de Recherche grant (nr. 40005955). T.L. is furthermore supported by two F.R.S.-FNRS PDR (grant numbers 31257234 and 40007793) and acknowledges also the support by TAILOR, a project funded by EU Horizon 2020 research and innovation programme under GA No 952215. E.F.D. and T.L. are supported by Service Public de Wallonie Recherche under grant n° 2010235–ariac by digitalwallonia4.ai. T.L. and P.S. acknowledge the support by the Flemish Government through the Flanders AI Research Program.”

We note that you have provided additional information within the Acknowledgements Section that is currently declared in your Funding Statement. Please note that funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

“I.T., P.S. and T.L are supported by an FWO (Fond Wetenschappelijk Onderzoek) project with grant no. G054919N. E.F.D. is supported by an F.R.S-FNRS (Fonds de la Recherche Scientifique) Chargé de Recherche grant (nr. 40005955). T.L. is furthermore supported by two F.R.S.-FNRS PDR (grant numbers 31257234 and 40007793) and acknowledges also the support by TAILOR, a project funded by EU Horizon 2020 research and innovation programme under GA No 952215.. E.F.D. and T.L. are supported by Service Public de Wallonie Recherche under grant n° 2010235–ariac by digitalwallonia4.ai. T.L. and P.S. acknowledge the support by the Flemish Government through the Flanders AI Research Program.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. We are unable to open your Supporting Information file [supporting information.ipynb]. Please kindly revise as necessary and re-upload.

5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

6. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments:

The received feedback from the referee and the associate editor is that the paper is well written and addresses an important and timely problem. It would make an excellent contribution to the theoretical literature of the evolution of cooperation and provides useful insights for the design of prosocial AI. The reviewer has provided constructive suggestions -- please carefully take them into account when preparing the revised version.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: In this manuscript, the authors developed an agent-based model where two types of agents play the one-shot collective risk dilemma (CRD). One type is supposed to be a human-like agent. Those agents optimize their strategies for the CRD based on the framework of evolutionary games. The other type is supposed to be AI-like agents. Those agents cooperate in the CRD with probability p. Using this model, the authors find that interesting phenomena shown in Fig. 1B and 2B. When the AI-like agents cooperate with high probability (bottom right area in Fig. 1B), the human-like agents do not cooperate. They free ride on the cooperation of the AI-like agents. On the other hand, when the AI-like agents cooperate with low probability (top left area in in Fig. 1B), the human-like agents actively cooperate to accomplish the goal of the CRD. The authors argued that those two types of agents compensate with each other. Thus, these human-AI hybrid teams works well to prevent the collective risks. Then, the authors considered the situation how the hybrid team works well by substitution or addition of AI-like agents to human-like agents. The model is new and worth investigating from the view points of not only evolutionary games but also human-AI interaction. I suggest two minor comments below. Please take them into account for improving the manuscript more.

1. Obviously, readers would like to know whether real humans and real AI (AI bots) actually compensate with each other when playing the CRD. It can be tested by conducting online experiments. But this is out of the scope of this manuscript because the authors take a model-based approach here. However, it is worth mentioning this point in the Conclusions by using one paragraph. If the authors agree with my suggestion, please do so.

2. It is worth citing the following royal society interface paper because they also simulated human-AI evolutionary games based on a model-based approach.

Small bots, big impact: solving the conundrum of cooperation in optional Prisoner’s Dilemma game through simple strategies, Gopal Sharma, Hao Guo, Chen Shen and Jun Tanimoto

Typos:

1. Ref. [32] is double cited on lines 127 and 258.

2. line 280: with the results in 1 --> with the results in 1A

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Feb 9;19(2):e0297213. doi: 10.1371/journal.pone.0297213.r002

Author response to Decision Letter 0


22 Dec 2023

We thank the reviewer for the positive assessment of our work. Please find below the answers to your two remaining comments.

1. Thank you for your comment, we absolutely agree with the importance of validating our theoretical results with behavioral experiments and have added this remark as a paragraph to our Conclusions. The new paragraph should appear highlighted in the document 'Revised Manuscript with Track Changes' in lines 453-458.

2. Thank you for making us aware of this reference, it is indeed of great interest to our work and we have added it in the Related Work section in line 122 as the new reference [32], as highlighted in the document 'Revised Manuscript with Track Changes'.

And finally, we gladly confirm that the typos found were corrected, and we thank the reviewer again for their time and attention in revising our manuscript and helping us improve our work.

Attachment

Submitted filename: Response_to_Reviewers.pdf

Decision Letter 1

The Anh Han

2 Jan 2024

The art of compensation: how hybrid teams solve collective-risk dilemmas

PONE-D-23-33002R1

Dear Dr. Terrucha,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

The Anh Han, Ph.D.

Academic Editor

PLOS ONE

Acceptance letter

The Anh Han

31 Jan 2024

PONE-D-23-33002R1

PLOS ONE

Dear Dr. Terrucha,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. The Anh Han

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response_to_Reviewers.pdf

    Data Availability Statement

    All computational files used to reproduce the results contained in this work are available from the Zenodo database at https://doi.org/10.5281/zenodo.10406595.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES