Skip to main content
PLOS One logoLink to PLOS One
. 2022 Dec 14;17(12):e0278150. doi: 10.1371/journal.pone.0278150

Rationality and cognitive bias in captive gorillas’ and orang-utans’ economic decision-making

Penelope Lacombe 1,*, Sarah Brocard 1, Klaus Zuberbühler 1,2,, Christoph D Dahl 1,3,4,‡,*
Editor: Elsa Addessi5
PMCID: PMC9749992  PMID: 36516129

Abstract

Human economic decision-making sometimes appears to be irrational. Partly, this is due to cognitive biases that can lead to suboptimal economic choices and context-dependent risk-preferences. A pertinent question is whether such biases are part of our evolutionary heritage or whether they are culturally acquired. To address this, we tested gorillas (Gorilla gorilla gorilla) and orang-utans (Pongo abelii) with two risk-assessment experiments that differed in how risk was presented. For both experiments, we found that subjects increased their preferences for the risky options as their expected gains increased, showing basic understanding of reward contingencies and rational decision-making. However, we also found consistent differences in risk proneness between the two experiments, as subjects were risk-neutral in one experiment and risk-prone in the other. We concluded that gorillas and orang-utans are economically rational but that their decisions can interact with pre-existing cognitive biases which modulates their risk-preference in context-dependent ways, explaining the variability of their risk-preference in previous literature.

Introduction

Economic theories and mathematical modelling of decision-making are increasingly used to understand and predict human behaviour. Classic models, such as ’Expected Utility Theory’ (EUT) [1], assume that humans are rational decision-makers, such that, when faced with a choice of options, they compare the utilities of all available options in order to gain the maximal utility [2]. ’Expected Utility Theory’ is based on the premise that the value of an option is defined by its utility and that humans are capable of evaluating utilities to make well-deliberated decisions.

However, consistent experimental findings show that human economic behaviours often appear irrational, i.e., not aimed at maximising utility [3]. Investigating the factors that lead to seemingly or real irrational choices shows that the way utility is assessed can highly modulate decision-making. Indeed, assessing and comparing utilities appears to be a difficult process, and decision-makers rarely have a perfect knowledge of all options they have to choose from, which can lead to apparently irrational behaviours (i.e., subjects do not maximise utility). For instance, humans can forego high-profit options in order to explore unfamiliar alternative options, even if they risk ending up with less profit ([4, 5]). This explorative behaviour aims to evaluate the utilities of all options, which is necessary for the decision-maker, and is especially important in unstable environments where continued sampling of options is essential for success [6].

Thus, apparently irrational behaviours can sometimes be explained by sampling strategies (or sampling difficulties) aimed at evaluating utilities, and do not conflict with EUT.

However, a large part of the apparently irrational behaviours described in the literature show a more severe discrepancy between human decision-making and what is predicted by EUT (e.g., violation of the independence axiom of EUT [7]), which cannot be explained by issues in utility evaluation or perception. This led to the development of behavioural economics, and the establishment of ’Prospect Theory’ [8] that has been particularly influential in explaining intuitive and emotional choices rather than rational aspirations to maximise gain.

’Prospect Theory’ states that options are not only assessed by their (true) utilities but also compared to a reference as gains or losses, and losses are valued more heavily than gains [9]. Furthermore, it states that probabilities are perceived subjectively, such that small probabilities are over-estimated for gains and under-estimated for losses [9]. Importantly, ’Prospect Theory’ also describes several cognitive biases, i.e., systemic patterns that affect human decision-making and induce violations to the EUT expectation of utility-maximising strategies. Such biases include the ’Endowment Effect’, a preference for already-owned objects over non-owned items of the same value [10], or the ’Framing Effect’, which explains that humans are risk-seeking when options are presented with positive connotations (’200 of 600 people lived’) and risk-averse when they have negative connotations (’400 of 600 people died’) [9].

Thereafter, the current consensus in behavioural economics is that human decision-making is a result of a rational utility-maximising strategy, as described by EUT, that interacts with diverse cognitive biases, which are consequences of the social [11], emotional [12], motivational [13], personal [14], and experimental [15] environment of the subject on its economic strategies.

To what extent do human rational utility-maximising strategy and cognitive bias balance compare to animals? What are the evolutionary roots of human decision-making?

There is a good evidence, across species, for utility maximising behaviour, which has led to optimal foraging theory in behavioural ecology ([16, 17]), which states that animal foraging behaviour is maximised in terms of cost-benefit ratio, suggesting that natural selection favours decision-making that leads to economically advantageous foraging behaviour. Indeed, many animal species have very efficient optimisation strategies and appear to be very good at sampling their environment and gaining maximal profit, leading to ideal-free distributions [18]. However, optimal foraging theory is concerned with population-level patterns and not with cognitive mechanisms driving individual decisions, which typically remain unexplored.

The fact that humans often do not make economically optimal choices, especially in risky environments, raises the question of whether the underlying biases may have older, phylogenetically evolved origins before the advent of modern humans. Indeed, previous research with great apes has discovered cognitive biases similar to the ’Framing Effect’ (chimpanzees, bonobos [19]) and the ’Endowment Effect’ (chimpanzees [20], gorillas [21]) in humans. Further indications for biases and suboptimal decision-making in primates derive from the fact that there can be considerable variability in performance between studies. Indeed, while risk-aversion is frequently described in humans [22] and non-primates [23], risk-preference is harder to generalise in primates, due to inter-species variability (for instance, bonobos appear more risk-averse than chimpanzees [2426]) and between-study variance within a given primate species: Rhesus macaques appear risk-prone in [2729] and risk-averse in [30], chimpanzees appear risk-prone in [3133] and risk-averse in [3436], and bonobos appear risk-prone in [19, 32] and risk-averse in [24, 25, 31]. One major issue in this line of research is the diversity of the experimental designs that are used to test primates (see [37] for a review). For instance, a standard and simple design to assess risk-preference is to present a choice between a ‘safe’ option (providing a fixed and known reward) and a ’risky’ option, that may or may not provide a reward (e.g. [24]). In this ’single cup’ design, subjects have to learn the different possible outcomes of the risky option and the probability to obtain them by memorising outcomes of previous trials. Across species, this experimental design generally (but not always [38]) leads to risk-aversion: bonobos showed 28% of risky choices when both options were equivalent [24], and around 30% of risky choices when the safe option is an intermediately-preferred food reward and the risky option either preferred of non-preferred food reward, with equal probabilities [25]. Chimpanzees showed around 40% of risky choices when the safe option is an intermediately-preferred food reward and the risky option either preferred of non-preferred food reward, with equal probabilities [25], and 34% of risky choices when both options were equivalent [36]. Alternatively, this experimental design can lead to weak risk-proneness: Chimpanzees showed 64% of risky choices when both options were equivalent [24], and chimpanzees and bonobos showed 57% of risky choices when the safe option is an intermediately-preferred food reward and the risky option either preferred of non-preferred food reward, with equal probabilities [39].

More complex designs can be used to assess risk-preference: in [40] for instance, the risky reward was not presented under one risky cup, but in a series of cups, one of which hid the reward (see Fig 1). In this ’multiple cups’ design, subjects do not have to learn the different outcomes and infer their respective probabilities, but they have to understand the relationship between the number of risky cups and the probability to obtain the risky reward.

Fig 1. ’Single cup’ design and ’multiple cups’ design.

Fig 1

Apparatus for the ’single cup’ design (A) as used in [24], leading to risk-aversion or weak risk-proneness. Subjects have to choose between one safe cup and one risky cup yielding an unknown amount of reward, with an unknown probability of yield. Apparatus for the ’multiple cups’ design (B) as used in [40], leading to high risk-proneness. Subjects have to choose between one safe cup and one to four risky cups, where only one risky cup contains the reward.

In this design, all four species of great apes were risk-prone, and, compared to the weak risk-proneness or the risk-aversion observed in the ’single cup’ design, exhibited extremely high levels of risk-proneness (for bonobos 75% and for chimpanzees 100% of risky choices [40]) when both options were equivalent.

This relationship between experimental design and risk-preference appears to be an indicator of one or more cognitive biases that affect how subjects perceive the safe or the risky option in a given design. If verified, this extreme shift from risk-aversion or weak risk-proneness (’single cup’ design) to high risk-proneness (’multiple cups’ design) in bonobos and chimpanzees would indicate that, like humans, they are not rational decision-maker (at least in one of the designs) who choose based on expected gain, but that decisions are context-dependant and subjected to cognitive biases.

However, this conclusion remains preliminary for the following reasons. First, in [40], authors kept the quantity of safe rewards smaller than the risky reward, which in itself may have favoured economically irrational decision-making and led subjects to prefer the risky option. Second, subjects were tested with high-quality rewards, which may have introduced a cognitive bias towards risk-taking ([41, 42]). Third, test trials were regularly interspersed with refresher trials (showing subjects the location of the larger, risky reward), which could have introduced a further cognitive bias towards the risky option. Whether or not these design features were sufficient to account for the risk-proneness in [40] remains to be tested. Another concern, raised by the authors themselves, is that subjects might simply have failed to understand the task. In particular, [40] wrote that subjects might be unable ’…to infer the chances of the risky option without experience and therefore being biased towards a risky choice’.

Indeed, if subjects are tested with variable options, they will inevitably need a number of trials to sample and compute the reward probabilities before being able to take rational decisions. Hence, randomisation of reward probabilities (i.e., changing the number of risky cups) between trials makes it impossible for subjects to learn the probability of the risky option. Instead, apes would have to figure out the relationship between the probability to win and the number of cups by pure inductive reasoning, which may be difficult to do without prior explicit training, even though apes have the cognitive ability to do so ([32, 43, 44]).

In this study, we revisited the question of whether great apes are rational economic decision-makers, by analysing how consistent risk-preference were across the two main experimental designs discussed before (’single cup’ vs ’multiple cups’ design). We were able to test two species of great apes, gorillas and orang-utans, which have not contributed much to the literature on primate behavioural economics, despite their close phylogenetic relatedness with humans. Indeed, only two studies compared them directly, and they turned out inconsistent: in the study by [40] gorillas and orang-utans were risk-prone; in the study by [45] gorillas and orang-utans were risk-averse, further contributing to the inconsistent patterns already reported in chimpanzees and bonobos. However, a preliminary common conclusion is that orang-utans are less risk-averse than gorillas, which may be linked to differences in feeding, as orang-utans feed on seasonally variable resources [46], using active and costly foraging behaviour [47], while gorillas feed on more steady resources which requires low foraging effort [48], a line of argument already invoked to explain differences between bonobos and chimpanzees [24]. Focusing on these two species (gorillas and orang-utan), in other words, could help clarify various longstanding problems in the behavioural economic literature.

Aside from the species comparison, a second main objective of our study was to determine the impact of experimental design on rational decision-making. In particular, we directly compared two different procedures (’single cup’ derived from [24] and ’multiple cups’ derived from [40]) with identical reward contingencies, in order to establish whether gorillas and orang-utans made rational choices in both experimental designs. Importantly, we modified the ’multiple cups’ design and tested both low- and high-valued reward (as reward value had an impact in earlier studies), added a condition where the safe and risky option were of equal value, avoided refresher trials, counterbalanced the side of the safe and risky options and applied double-blind testing to rule out experimenter bias. To ensure that subjects understood the economic nature of the task, we systematically varied both the amount of reward and the probability to win.

The expectation here is that a rational decision-maker should be relatively more inclined to make a risky choice if the risky option increases or similarly if the probability to win increases. We also tested subjects’ rationality by comparing their performance between the two set-ups. The expectation is that a rational decision-maker should have a comparable strategy whatever the set-up, provided that the potential gains and probabilities are identical between the set-ups. However, in line with the literature, we predicted risk-aversion or weak risk-proneness in the ’single cup’ design and high risk-proneness in the ’multiple cups’ design, hence deviating from rational decision-making.

Materials and methods

Subjects

Subjects were five orang-utans (one adult male, three adult females and one juvenile female) and three gorillas (one adult female, one juvenile male and one juvenile female), between 4 and 20 years of age, born in captivity and housed at Basel Zoo, Switzerland, see S1 Table in S1 File. These eight subjects participated in two experiments (experiment 1 and experiment 2 with low-value reward, see Procedure below), and six participated in experiment 2 with high-valued reward, see S1 Table in S1 File. They were kept in a mixed social group (gorillas) or in pairs (orang-utans). Except for one orang-utan, all subjects were naive to any kind of cognition studies.

Procedure

Experiments were conducted in the indoor part of the enclosures. Subjects were exposed to choice options presented on a trolley that could be moved between enclosures. The trolley was positioned around 30cm from the enclosures. Subjects decided by themselves whether and when they wanted to participate in the trials. Subjects were neither food- nor water-deprived and the regular feeding schedule was continued during data collection. The main research was approved by the cantonal veterinary office of Basel Stadt (permit 2983).

Experiments were conducted during 4-hours morning sessions, five days a week, between February 2018 and February 2020. A maximum of two sessions of 10 trials were conducted every day per subject. The reward we used were 0.5cm3 cubes of food (up to seven pieces at once). Two types of food were used, one low-value food type (vegetables, such as beetroots or carrots, depending on subject preference) and one high-value food type (grapes). All sessions were videotaped, using a HC V500M Panasonic camcorder mounted on a tripod.

Each subject was exposed to one training and two experimental conditions, whose order was counterbalanced between subjects. For the two experiments, the side on the trolley of the safe and the risky options were counterbalanced each day. Finally, we used a double-blind procedure to make sure that the experimenter did not influence the choice of the subjects.

Training

In the initial training session, subjects were trained to indicate a cup, remember the content of a cup, and discriminate between different quantities of reward. For this, two transparent little saucers were first placed on the trolley, filled with different quantities of reward. During the first part of the training (i.e., training the subjects to indicate a cup), the experimenter filled one of the saucer with three pieces of food, and the other remained empty. In order to obtain the reward, the subject had to point with their hands (juveniles) or with a stick (adults, as their hand did not fit inside the grid of the enclosure) to the corresponding saucer. After the subject chose a saucer, the experimenter handed the reward (if the chosen cup contained reward) to the subject on a skewer through the mesh of the enclosure.

We collected the pointing stick from the subject’s hand after every trial and repositioned it in the enclosure close to the subject at the end of the trial. For each subject, sessions of 10 trials were run. If, during two consecutive sessions, the subject chose the correct cup in over eight trials out of 10, the first part of the training was completed. During the second part of the training (i.e., training the subjects to remember the content of a cup), the experimental design was similar except that, after filling one of the saucer with three pieces of food, the experimenter hid both saucers with opaque cups (of different shape and colour) in full view of the subject. To complete that part of the training, subjects had to choose the correct cup in over eight trials out of 10, during two consecutive sessions. Finally, during the last part of the training (i.e., training the subjects to discriminate food quantities), the experimenter filled the two saucers with two different quantities of food and covered them with opaque cups (of different shape and colours) in full view of the subject. The two different quantities of food were in that order: 2 vs 0, 2 vs 4, 2 vs 6, 6 vs 7. For one pair of food quantities (ex 2 vs 0) subjects had to choose the correct cup in over eight trials out of 10, during two consecutive sessions, to move on to the next pair of food quantities (ex 2 vs 4). During each part of the training the side of the large and small food quantity was randomised. The duration of the training phase was variable between the subjects (between one and 10 months). Four subjects stopped the experiment during the training phase because of a lack of motivation.

Experiment 1 –’single cup’ design

In Experiment 1, Experimenter A placed two opaque cups on the trolley, a yellow (safe) cup and a pink (risky) cup, see S1 Fig in S1 File and S1 Movie in S2 File. Next to each cup, she placed a transparent saucer. A wooden occluder was then placed over the risky cup and corresponding saucer. Experimenter A placed 2 pieces of reward (vegetables) in the safe saucer, in full view of the subject. Then she covered that saucer with the safe cup. After that, she either put, or pretended to put, a variable quantity of reward in the risky saucer and covered it with the risky cup, all of which behind the wooden occluder. Then she removed the occluder and left from behind the trolley. Experimenter B, naïve to the condition, then replaced Experimenter A. She waited for the subject to choose one of the cups, revealed the content of both cups, and gave the reward under the chosen cup to the subject (if there was one).

Within the experiment we varied the probability to win and the value of the risky option (see Table 1), while the safe option was fixed. The probability to win corresponded to the probability that a reward was placed under the risky cup (i.e., if P = 0.5, there was one chance out of two that the reward was under the risky cup).

Table 1. Expected values (EV) of the risky option across the different conditions of probability to win (P) and of risky reward value (V) in Experiment 1.

The safe option was always set at EV = 2.

P
V 0.25 0.33 0.5 1
2 -- -- 1 (-) --
4 1 (-) 1.33 (-) 2 (=) 4 (+)
6 -- -- 3 (+) --

P = probability of obtaining the reward when choosing the risky option; V = value of the reward of the risky option; EV = expected value of the risky option (EV = P*V). The symbol next to the EV indicates whether the risky option has a smaller (-) EV than the safe option, a larger EV (+) than the safe option, or whether the safe and risky option have the same EV (=). Sample sizes: each condition (box) consisted of 40 trials (4 consecutive sessions of 10 trials) per subject. Twenty trials were run per day so each condition took two days to be tested. — = the condition was not tested.

In this design, subjects have to learn the value and the probability of the risky option through the analysis and the memorisation of trials’ feedback, so all trials corresponding to the same P*V combinations were performed in a row. For each subject, four consecutive sessions of 10 trials for each of the six combinations of P*V (see the six cells in Table 1) were performed, for a total of 24 sessions (N = 240 trials per subject). The order of the six combinations was randomised between subjects.

Experiment 2 –’multiple cups’ design

Experimenter A displayed a light-blue (safe) cup and one to four orange (risky) cups on the trolley, and hid the risky cups with the occluder (see S2 Fig in S1 File and S2 Movie in S2 File). Then, she placed two transparent saucers on top of the occluder and filled the ‘safe’ saucers with two pieces of reward, and one of the ‘risky’ saucers with two to seven pieces of reward. The safe saucer was then placed under the safe cup in full view of the subject. The risky saucer was placed under one of the risky cups, behind the occluder so that the subject did not know under which cup the saucer was. The position of the baited cup was randomised between trials so that each risky cup held the reward equally often. Experimenter A then removed the occluder and left from behind the trolley. Experimenter B, naïve to the conditions, replaced Experimenter A. She waited for the subject to choose one of the cups, revealed the content of all cups, and gave to the subject the reward that was under the chosen cup (if there was one). Experiment 2 was conducted twice, first using low-valued food reward (vegetables), then after completion, a second time using high-value food type (grapes).

Within the experiment we varied the probability to win (by altering the number of risky cups) and the value of the risky option (see Table 2), while the safe option was fixed.

Table 2. Expected values of the risky option across the different conditions of probability to win (P) and of risky reward value (V) in Experiment 2.

The safe option was always set at EV = 2.

P
V 0.25 0.33 0.5 1
2 0.5 (-) 0.66 (-) 1 (-) 2 (=)
4 1 (-) 1.33 (-) 2 (=) 4 (+)
6 1.5 (-) 2 (=) 3 (+) 6 (+)
7 1.75 (-) 2.33 (+) 3.5 (+) 7 (+)

P = probability of obtaining the reward when choosing the risky option (i.e., 1/number of risky cups); V = value of the reward of the risky option; EV = expected value of the risky option (EV = P*V). The symbol next to the EV indicates whether the risky option has a smaller (-) EV than the safe option, a larger EV (+) than the safe option, or whether the safe and risky option have the same EV (=). Sample sizes: each condition (box) consisted of 10 non-consecutive trials per subject. Twenty trials were run per day.

In this design, subjects have to understand the probability of the risky option through the analysis of the number of risky cups, therefore, to avoid that subjects learn the probability to win over repetition of trials we did not perform consecutively all trials corresponding to the same P*V combination. Each risky reward amount (two, four, six and seven) was tested in two non-consecutive sessions of 10 trials. Among these 20 trials for each risky reward, each cup number (one, two, three or four) was tested five times. The order of the eight sessions (two sessions per risky reward) was randomised between subjects. For each type of reward, the experiment was run twice, to check for learning effects during the experiment, so that each subject was exposed to 16 sessions, i.e., N = 160 trials total (for one type of reward), see S3 Fig in S1 File.

Statistical analysis

Both experiments were videotaped and subjects’ choices (safe or risky option) were coded live during the test. A second observer coded the subjects’ choices from the video. Reliability was assessed by recoding 20% of trials which led to excellent reliability (Cohen’s k = 0.98). Response times were recorded for each trial (interval between the time when the experimenter showed the cups to the subject and the time when the subject pointed unequivocally to one cup).

For our analysis of subjects’ choices, we fitted a generalised linear mixed model with binomial error structure and logit link function to our data. The response variable was whether subjects chose the risky option or not. The data were analysed using the glmer function of the lme4 package in R. We checked the normality and the homoscedasticity of plotted residuals and their independence with respect to fitted and other predictors to ensure we met the assumptions of the model. The significance of each predictor variable in explaining variation in rate of risky choices was tested by an analysis of deviance (type II Wald chi-square test).

As our data set originated from a small number of subject, we checked model stability using the influence.ME package in R, that allows to detect influential data in mixed effects models: first, we calculated the estimates of our models when iteratively excluding the influence of each subject (using the function influence of the package), then we computed the Cook’s distances measure of every subject on our models (using the cooks.distance function of the package) in order to check whether our models were influenced by certain subjects. If every Cook’s distances were inferior to one we concluded that our models were stable.

When analysing risk-preference data in Experiment 1, we fitted the value and the probability of the risky option as continuous variables and the species as a categorical variable. In order to check whether subjects changed their preference throughout the four sessions of each P*V combination, we also fitted the session number as a continuous variable, and the three-way interactions between session, species and probability of the risky option, and between session, species and value of the risky option. In order to test side preference, we fitted the side of the safe cup as a categorical variable. We fitted random intercepts for session within individuals and random slopes of all predictor terms of interest, random slopes of session within individual, and random slopes of all predictor terms of interest within session. We compared this full model to a null model (no fixed effects and the same random structure as the full model) using a Likelihood Ratio Test. If the comparison showed a significant difference, we assessed the significance of each predictor variable in explaining variation in rate of risky choices by an analysis of deviance (type II Wald chi-square test). Then, starting with the highest-level interaction terms, we removed non-significant terms one by one. The final model where all non-significant interaction terms were removed is presented in the result section.

Then, on this final model, we ran post-hoc tests to calculate estimated marginal means or estimated trends (functions emtrends or emmeans of the emmeans package). Finally, in order to check whether the level of risky choice for EV = 2 was significantly different from 50%, we calculated the estimated marginal means for EV = 2 (equivalence point) for each species and each session if session was a significant predictor. If the level of risky choice was not significantly different from 50%, that would indicate risk-neutrality (subjects have no preference between a safe and a risky option of equal expected values). On the contrary, a level of risky choice for EV = 2 significantly higher than 50% would indicate risk-proneness (subjects prefer the risky option), and significantly lower than 50% would indicate risk-aversion (subjects prefer the safe option).

To analyse risk-preference data in Experiment 2 we ran a similar analysis as previously described but we added a predictor variable: the type of reward as a categorical variable. Finally, in order to investigate any positional preference or side bias in Experiment 2, we again fitted a generalised linear mixed model on our data where the response variable was the proportion of selection for each cup (each cup represented by its position on the trolley from left to right, irrespectively of whether it was a risky or a safe cup). We fitted the position of the cup, the total number of cups, the side of the safe cup on the trolley and the species, as well as the two-way interactions between those predictors.

In order to compare the data of Experiment 1 and 2 (with low-valued reward) and to understand the effect of experimental design on risk-preference we ran a similar analysis with a new predictor variable: the experimental design (categorical variable), as well as the two-way interactions between experimental design and economic parameters (probability and value of the risky option). As in the analysis above, we fitted random intercept for session within individuals and random slopes of all predictor terms of interest, random slopes of session within individual, and random slopes of all predictor terms of interest within session, and a random slope of experimental design within individual.

Finally, we analysed the response times in both experiments by fitting a linear mixed-effect model with the response time as the response variable. We log-transformed the response times and fitted a model with the same random structure as before, and with species, economic parameters, type of reward and experimental design as predictors.

Results

Experiment 1 (’single cup’ design)

In Experiment 1, we varied the probability to win and the value of the risky option (see Table 1, where P refers to the probability of the risky option, V to the value of the risky option, and EV to the expected value of the risky option, while EV = P*V). The results of this experiment are shown in Fig 2 (and S6 Fig in S1 File for individual data). Our full model with all predictors terms and three-way-interactions was significantly different from the null model (LRT: χ2(24) = 138.10, p < .001), see S2 and S3 Tables in S1 File for the random and fixed structure of our models. Table 3 shows our final model where non-significant interaction terms were removed. Cook’s distances for the data set of each subject were all smaller than 1; our final model was then evaluated stable across subjects.

Fig 2. Results of Experiment 1.

Fig 2

Mean percentage of trials where subjects selected the risky option according to the value of the risky option (A) the probability to win (B) and the expected value of the risky option (C) for gorillas (black) and orang-utans (grey). Error bars indicate 95% confidence intervals.

Table 3. Fixed effects of the final model investigating subjects’ risky preference in Experiment 1.

The table reports the results of the analysis of deviance (type II Wald chi-square tests).

Chi-square Df p-value
species 3.21 1 0.074
side of the safe cup 36.00 1 < .001
risky probability 90.10 1 < .001
risky value 64.72 1 < .001
session 12.45 3 < .01
species: session 7.26 3 0.064
risky probability: session 12.06 3 < .01
risky value: session 12.75 3 < .01
species: risky value 0.03 1 0.850
species: risky value: session 12.30 3 < .01

Subjects picked the risky cup more often with increasing probability of the risky option (χ2(1) = 90.10, p < .001) and more often with increasing value of the risky option (χ2(1) = 64.72, p < .001). Their level of risky choices was also affected by the session (χ2(3) = 12.45, p < .01) and by the side of the safe cup (χ2(1) = 36.00, p < .001). More importantly, there was a significant interaction between the session and the economic parameters. Firstly, there was a significant interaction between the session and the probability of the risky option (χ2(3) = 12.06, p < .01). Post-hoc tests showed that trend estimates were smaller in sessions 1 and 3 than in session 2 and 4 (see S4 Table in S1 File), i.e., subjects’ sensitivity to the probability to win increased between the two daily sessions (sessions 1 and 2 being performed the same day, and 3 and 4 the next day). Secondly, there was a significant interaction between session, value of the risky option, and species (χ2(3) = 12.30, p < .01). Post-hoc tests showed that this is mainly because, for gorillas, trend estimates for the value of the risky option differed between session 1 and every other session.

The study of the percentage of risky choices at the indifference point (EV = 2, equality of the expected value of the safe and of the risky option) showed that, for gorillas, the percentage of risky choices when EV = 2 was never significantly different than 50% (interval for estimated marginal means: [0.31;0.64] for session 1, [0.39; 0.73] for session 2, [0.39; 0.72] for session 3, [0.27; 0.60] for session 4). For orang-utans, the percentage of risky choices at EV = 2 was not significantly different from 50% for the two first sessions (interval for estimated marginal means: [0.32; 0.58] for session 1, [0.35; 0.61] for session 2) and was significantly lower than 50% for the last two (interval for estimated marginal means: [0.16; 0.37] for session 3, [0.19; 0.42] for session 4).

Experiment 2 (’multiple cups’ design)

In Experiment 2, we varied the probability to win and the value of the risky option (see Table 2, where P refers to the probability of the risky option, V to the value of the risky option, and EV to the expected value of the risky option, while EV = P*V), as well as the type of reward (low-valued food type: vegetables, and high-valued food type: grapes). The results of this experiment are shown in Fig 3 (and S6 Fig in S1 File for individual data). Our full model with all predictors terms and three-way-interactions was significantly different from the null model (LRT: χ2(25) = 64.94, p < .001), see S5 and S6 Tables in S1 File for the random and fixed structure of our models. Table 4 shows our final model without interactions terms as they were non-significant in the full model.

Fig 3. Results of Experiment 2.

Fig 3

Mean percentage of trials where subjects selected the risky option according to the value of the risky option (A) the probability to win (B) and the expected value of the risky option (C) for the high-valued reward (grapes, in thick line) and low-valued reward (vegetables, in narrow line). Error bars indicate 95% confidence intervals. Only trials with P<1 were considered for (C).

Table 4. Fixed effects of the final model investigating subjects’ risky preference in Experiment 2.

The table reports the results of the analysis of deviance (type II Wald chi-square tests). Only trials with P<1 were considered.

Chi-square Df p-value
species 1.15 1 0.280
side of the safe cup 11.46 1 < .001
risky probability 13.4 1 < .001
risky value 47.66 1 < .001
reward 7.68 1 < .05
session 10.31 3 0.01

Cook’s distances for the data set of each subject were all smaller than 1, except for one juvenile orang-utan (Ketawa: Cook’s D = 1.14, see S6 Fig in S1 File for individual data); our final model was then evaluated relatively stable across subjects.

Subjects picked the risky cup more often with increasing probability of the risky option (χ2(1) = 13.40, p < .001) and more often with increasing value of the risky option (χ2(1) = 47.66, p < .001). Their level of risky choices was also affected by the session (χ2(3) = 10.431 p = .01), by the side of the safe cup (χ2(1) = 11.46, p < .001) and by the reward (χ2(1) = 7.68, p < .05). Investigating the effect of session on subjects’ choices, post-hoc tests (see S7 Table in S1 File), showed that the level of risky choices increased throughout session (level of risky choices ± SE: 0.72 ± 0.07 for session 1, 0.76 ± 0.07 for session 2, 0.80 ± 0.06 for session 3, 0.85 ± 0.05 for session 4). Contrary to what we described in Experiment 1, in Experiment 2 there was no interaction between sessions and risk parameters, which indicates that the effect of session on performance was a general increase of risk-proneness over the experiment and not a learning effect of the probabilities to win over repeated testing.

Finally, our model showed that there was no differences between gorillas and orang-utans (χ2(1) = 1.15 p = .28). The overall percentage of risky choices at the indifference point was 75% (interval for estimated marginal means: [0.63; 0.88]) for low-valued food type) and 81% (interval for estimated marginal means: [0.71; 0.92]) for high-valued food type, i.e., both significantly higher than 50%.

To investigate which parameters influenced the rate of selection of each cup on the trolley (safe and risky cups), we fitted a GLMM model to the percentage of selection of each cup, with the same random structure as before, and with the following predictors: position of the cup on the trolley, side of the safe cup, total number of cups (i.e., risky probability), species. Our full model with all predictors terms and two-way-interactions was significantly different from the null model (LRT: χ2(9) = 54.45 p < .001).

Our final model, see Table 5, showed that percentage of choice of each cup was affected by the overall number of cups (χ2(1) = 52.30 p < .001), and more importantly, that the impact of the position was affected by the total number of risky cups (significant two-way interaction: χ2(1) = 15.65, p < .001), and the species (significant two-way interaction: χ2(1) = 11.40, p < .001). Indeed, see S8 Fig in S1 File, both species showed positional bias: gorillas exhibited a preference for the left location and orang-utans for the central position.

Table 5. Fixed effects of the final model investigating positional biases in Experiment 2.

The table reports the results of the analysis of deviance (type II Wald chi-square tests) for Experiment 2.

Chi-square Df p-value
species 1.83 1 0.18
side of the safe cup 0.80 1 0.37
risky probability 52.30 1 < .001
position of the cup on the trolley 1.52 1 0.22
position of the cup on the trolley: side of the safe cup 3.06 1 0.08
position of the cup on the trolley: risky probability 15.65 1 < .001
position of the cup on the trolley: species 11.40 1 < .001

Performance in ’single cup’ vs ’multiple cups’ designs

In order to investigate the impact of the experimental design on subjects’ choices, we fitted a full model with the same random structure as the ones we used to analyse Experiment 1 and Experiment 2 with the addition of a random slope of experimental design within individual. We fitted the two economic parameters, the species, the side of the safe cup, the experimental design and the two-way-interactions between experimental design and economic parameters as predictors (see S8 and S9 Tables in S1 File for the random and fixed structure of our models). Our full model was significantly different from the null model (LRT: χ2(7) = 101.10, p < .001). Table 6 shows our final model where the non-significant interaction terms between risky probability and experimental design was removed. Cook’s distances for the data set of each subject were all smaller than 1, except for one juvenile orang-utan (Ketawa: Cook’s D = 1.14, see S6 Fig in S1 File for individual data); our final model was then evaluated relatively stable across subjects.

Table 6. Fixed effects of the final model.

The table reports the results of the analysis of deviance (type II Wald chi-square tests) for Experiment 1 and 2 combined. Only trials with P<1 were considered.

Chi-square Df p-value
species 0.89 1 0.350
side of the safe cup 34.35 1 < .001
risky probability 21.43 1 < .001
risky value 94.07 1 < .001
experimental design 13.65 1 < .001
risky value: experimental design 15.95 1 < .001

Subjects picked the risky cup more often with increasing probability of the risky option (χ2(1) = 21.43, p < .001), more often with increasing value of the risky option (χ2(1) = 94.07, p < .001), and more often in Experiment 2 (χ2(1) = 13.676 p < .001). More importantly, the interaction between the value of the risky option and the experimental design was significant (χ2(1) = 15.95, p < .001), but the interaction between the probability to win and the experimental design was not, as we considered only trials with P<1, see Fig 4.

Fig 4. Performance comparison between experiments.

Fig 4

Mean percentage of trials where subjects selected the risky option for Experiment 1 (solid line) and Experiment 2 (dotted line, low-valued food type): according to the value of the risky option (A) and the probability to win (B). Error bars indicate 95% confidence intervals.

To investigate the impact of the experimental design on subject choices, we ran post-hoc tests (see S10 Table in S1 File) that showed that the level of risky choices was overall higher in Experiment 2 (mean of risky choices for all trials: 78% with a 95% CI: [66%; 90%]) than in Experiment 1 (mean of risky choices for all trials: 39% with a 95% CI: [32%; 47%]), and that the trend estimates for the value of the risky option were steeper in Experiment 1 (trend estimate: 0.1, SE: 0.01) than in Experiment 2 (trend estimate: 0.03, SE: 0.01), see S10 Table in S1 File.

Finally, we analysed the response times in both experiments to assess whether the higher levels of risk-proneness in Experiment 2 could be due to high impulsivity levels, as the risky reward was shown to the subjects before each trial in this design, which could have led subjects to pick the risky option.

We used response time as the response variable after a log-transformation and fitted a model with the same random structure as before, and with species, economic parameters, type of reward and experimental design as predictors. Our model was significantly different from the null model (LRT: χ2(7) = 75.58, p < .001), and showed that response times were longer in Experiment 2, and longer when the reward was of low-value (respectively χ2(1) = 72.99, p < .001 and χ2(1) = 8.09, p < .005), see S9 Fig in S1 File.

Discussion

We carried out a series of choice experiments in which captive gorillas and orang-utans were asked to choose between two options that differed in risk and reward. In two different experiments, subjects could either choose a safe option that always yielded a predetermined reward or choose a risky option, which could yield a higher reward or no reward at all. In the first experiment, the ’single cup’ design, the safe reward was reliably presented under a safe cup, whereas the risky reward was intermittently presented under a risky cup. Here, subjects had to learn from the feedback of previous trials the potential gain and its probability to be hidden under the risky cup. In the second experiment, the ’multiple cups’ design, the safe reward was again reliably presented under a safe cup, whereas the risky reward was always presented under one of several risky cups. Here, subjects had to understand the relationship between the number of risky cups and the probability to gain the reward if selecting a risky cup.

We analysed subjects’ performance in each of the two experimental designs and found that orang-utans and gorillas generally acted as rational decision-makers, as their preference for the risky option depended on its potential gain and on the probability to win. We found for both experiments that the proportion of risky choices had a linear relationship with the probability to win, and the amount of reward to win. However, we also found that risk preference was additionally affected by non-economic parameters, as subjects’ strategies were not stable across experimental designs. In particular, at the indifference point, subjects were generally risk-neutral in Experiment 1 and risk-prone in Experiment 2. Concerning species comparison, we did not find a performance difference across species in Experiment 2, however, we found that gorillas were more risk-prone than orang-utans in Experiment 1. Our results conflict with previous studies, that reported a species difference in risky context, and concluded that orang-utans are more risk-prone than gorillas ([32, 40]); however, this could be due to the small sample size of our study.

Our major finding, that experimental design affects subjects’ risk-preference, is in line with our predictions, as Experiment 1 was derived from an experimental design usually (though not always [38]) leading to risk-aversion ([24, 25, 36]) or weak risk-proneness ([24, 39]) and Experiment 2 was derived from an experimental design that led to high risk-proneness in the four species of great apes [40]. However, the effect of experimental design on subjects’ economic strategy had never been directly tested, and the increased risk-proneness triggered by the ’multiple cups’ design was so far speculative due to experimental issues, i.e., the safe option was always smaller than the risky option, the side of the safe and risky option were not counterbalanced, visible refreshers trials could have biased subjects toward the risky option [40]. Additionally, authors only used a high-valued reward, which could have increased the levels of risky choices (see [41, 42, 49] for debate on impact of reward on economic decision-making in humans and animals). Finally, [40] could not reject that subjects failed to understand the task and to properly infer the probabilities to win without experience.

In this study, we addressed the experimental issues of the ’multiple cup’ design: we used both a low-valued and a high-valued reward, removed refreshers trials, added a condition where the safe and risky option were identical, alternated the side of the safe and risky options. Additionally, we used a double-blind procedure to prevent any kind of experimenter bias. Finally, we investigated how the probability to win affected the level of risky choices (by computing a GLMM with probability to win as a fixed effect) in order to evaluate whether subjects based their strategy on the probability to win and thus whether they were able to assess it. As a result, first, we showed that subjects chose the risky option less often when the probability to win decreased (i.e., when the number of risky cups increased), showing that they understood the economic nature of the task. Second, we found lower levels of risky choices than in [40], indicating that the above-mentioned experimental issues could indeed have artificially increased risk-proneness in the original study.

In sum, our replication of the ’multiple cups’ design showed that gorillas and orang-utans understood the economical nature of the task and expressed risk-prone behaviour in this design, even after we removed factors that artificially increased risk-preference in the original design.

As a conclusion, our data confirmed that gorillas and orang-utans were risk-prone in the ’multiple cups’ design, while they were risk-neutral in the ’single cup’ design. This indicated that, while for each experimental design subjects used a rational strategy and compared potential gain and probabilities to make a decision, apparently irrational contextual factors interacted with this strategy and affected gorillas’ and orang-utans’ risk-assessment. The irrational finding that the same economic choice (for instance between ’choosing a sure two pieces of carrots’ and ’a 50% chance of winning four pieces of carrots’) leads to risk-neutrality (no preference) in Experiment 1 and to risk-proneness (preference for the risky choice) in Experiment 2 demonstrated that some aspects inherent in the experimental design of one (or both) experiments triggered one or several cognitive processes affecting how options were perceived.

We put forward several hypothesis accounting for this shift of strategy between the ’single cup’ design and the ’multiple cups’ design.

A first hypothesis is that this shift was due to a ’description-experience gap’ (D-E gap) as described in [50], which states that decision-makers could experience probability distortions depending on how probabilities are presented to them: either from description (i.e., when probabilities are described to them) or from experience (i.e., they have to sample every option and figure out the probabilities). Indeed, in our study, the ’single cup’ design corresponded to an experiment based on experience and the ’multiple cups’ design to an experiment based on description, and thus the difference in performance could be explained by such a D-E gap, where subjects were more risk-prone in a design from description (’multiple cups’) than from experience (’single cup’). This contradicts previous findings in humans [51] and apes [52] which concluded that designs from experience increased risk-proneness. However, analysing D-E gap is yet subject to debate as several authors question its validity and cause [53], and more importantly, its effect on decision-making seems unclear. Whether the D-E gap induces probabilities distortion for small probabilities (under 20% [54]) or whether it induces risk-preference shift with increased risk-proneness in experiment from experience [51], or the opposite (our findings) remains elusive. Importantly, and contrary to previous studies using visual cues with non-human primates (the probability to gain a reward if choosing a risky option was represented on a screen by the length ratio of coloured bars, for instance a coloured bar divided equally in two colours indicated a 50% probability of winning [52]), we described probabilities in a highly intuitive fashion (by a number of risky cups), and avoided mixed described designs where subjects could also learn probabilities by sampling. Indeed, in designs that are intermediate between description and experience, all trials with the same rewards contingencies are performed in a row, and subjects have feedback from the chosen option. Finally, both designs provided the same level of feedback (full feedback: the content of each option was shown after each trial), while often experiments from experience provide partial feedback (only the content of the chosen option is shown after each trial, e.g. [52, 55]), which could lead the decision-maker to pick the risky option solely to gain information, and thus increase the level of risky choices in experiment from experience. In all these respects, this makes our experimental design a highly appropriate one to study D-E gap in non-human primates.

A second hypothesis, the ’impulsivity hypothesis’, states that higher levels of risky choices in the ’multiple cups’ design were due to the fact that the risky reward was shown to the subjects before each trial in this design only. This could have induced an urge to choose that risky reward, regardless of any underlying economic rationality, because of the difficulty to exert self-control over one’s impulsive choice ([56, 57] for examples in children, and [58] for an example of how inhibitory control difficulties impact economic choices in mangabeys). Indeed, [40] wrote that ’…this bias towards the risky option could be explained, for instance, by a failure to inhibit a subject’s inherent tendency to choose the large reward, [as] several studies have shown that great apes (and other primates) need a large number of trials to overcome their initial tendency to choose a higher valued food […] even when the reward is no longer visible’. Such a hypothesis could be difficult to test directly, as impulsivity is often measured by proxies such as response times. The analysis of the response times in both experimental designs indicated, on the contrary, longer response times in the ’multiple cups’ design than in the ’single cup’ design, which do not necessarily reflect differences in impulsivity levels but merely that decision-making in the ’multiple cups’ design took longer as there are more cups to choose from and that pointing needs to be more precise, which also takes time. Testing the ’impulsivity hypothesis’ should thus use other means than measuring reaction or responses times, and further studies could try to control subjects’ impulsive responses and examine the degree to which the level of impulsivity correlates with risk preference. For instance, previous work in children [59] showed that if a transparent barrier is placed between subjects and the items they have to pick from (so that they cannot point directly at the desired item but have to make a detour over the barrier to point), automatic and impulsive responses were inhibited. Another study [60] showed similar result when subjects had to wear a weighted bracelet, which also reduced the level of impulsive responses. Such experimental designs (placing transparent barrier between the enclosure grid and the trolley, or providing subjects with weighted sticks,…) could be tested in the ’multiple cups’ design to investigate if it would lower gorillas and orang-utans observed risk-proneness.

A third hypothesis to explain the risk-preference shift between the ’single cup’ and the ’multiple cups’ design is the ’exploration hypothesis’ which states that performance is best explained as a bias towards exploring rather than exploiting the risky option, as these two strategies can compete in risk-assessment tasks [61]. In the ’multiple cups’ design, subjects were confronted with an array of risky cups, and each one could or could not contain the reward, while in the standard ’single cup’ design, only one cup was available. It is therefore conceivable that the ’multiple cups’ design was more salient to apes and so triggered more curiosity and exploratory behaviour, even though subjects had full feedback on the content of the cups after each trial. In previous studies using a similar ’multiple cups’ design, the difference in economic strategies between risk-neutral capuchins and risk-prone young human children [62] were attributed to difference in exploitation/exploration strategies, as capuchins were argued to rely more on an exploitation strategy of safe options while young children rely more on an exploration strategy of uncertain options [63]. However, in our study, we found no support for that ’exploration hypothesis’, due to the fact that both species exhibited positional biases in Experiment 2: rather than exploring each risky cup, gorillas exhibited a strong preference for the one that was at their left, while orang-utans preferred the central cups.

This leads to a fourth hypothesis, the ’lateral hypothesis’, stating that a positional physical bias in Experiment 2 could account for the high levels of risky choices in that experiment only. However, as we counterbalanced the sides of the risky and safe option, only the orang-utans central bias (and not the gorillas left preference) could have had an impact on their choices, and artificially increased the level of risky choices (as the central cup is always a risky cup). This would explain why orang-utans showed lower levels of risky choices than gorillas in Experiment 1 but showed comparable to higher rates of risky choices than gorillas in Experiment 2. Importantly, the interaction between a rational economic decision-making and gorillas’ and orang-utans’ positional bias could explain the inexplicable drop of risky choices to 50% when the probability to win was P = 1, i.e., when there was only one risky cup in Experiment 2. Indeed, gorillas selected preferentially the cup to their left, which corresponded in 50% of the trials to the risky cup, and orang-utans selected indifferently one of the cup as they could not select a central one. This would suggest that when P = 1, subjects were fully driven by their lateral bias. However, the ’lateral hypothesis’ cannot explain the preference shift from risk-neutrality in Experiment 1 to risk-proneness in Experiment 2 for gorillas, so, even if it could explain a portion of orang-utan increased risk-proneness in Experiment 2, it cannot be the sole explanation for our results.

Finally, a last hypothesis is the ’framing hypothesis’, stating that the main difference between the ’single cup’ and the ’multiple cups’ design was how subjects framed the safe and risky option. In the ’single cup’ design, as the risky option was not necessarily baited with a reward (it was only baited at a certain probability), subjects could have perceived the safe reward as their ’reference’ reward quantity, and they either could win (if they picked the risky option and gained the reward) more than that reference quantity, or they could lose it (if they picked the risky option and did not gain anything). On the contrary, in the ’multiple cups’ design, the risky reward (equal or larger than the safe reward) was shown before each trial and the risky option was necessarily baited with a reward (one of the risky cups was always hiding the risky reward); it is therefore possible that subjects have considered that reward quantity as the reference. In that case, choosing the safe reward meant losing a portion of that reference, and choosing the risky option meant either getting that reference quantity or losing it all. In other words, the ’single cup’ design could be viewed as a risk-assessment task in the gain domain, and the ’multiple cups’ design could be viewed as a risk-assessment task in the loss domain. As the ’Prospect Theory’ [8] (and several empirical studies in humans [64] and non-human primates [19]) showed that primate decision-makers are risk-averse in the domain of gain and risk-prone in the domain of loss, the different framing of the safe and risky reward quantities in the ’single cup’ design versus ’multiple cups’ design could explain subjects preference shift between those two designs: risk-neutrality in the ’single cup’ design (risk-aversion/neutrality in the domain of gain) and risk-proneness in the ’multiple cups’ design (risk-proneness in the domain of loss).

Whatever the cognitive biases are, our results showed that assessing whether non-human primates are risk-prone or risk-averse is challenging, and that it depends on the experimental design used to answer that question. This is consistent with what has been shown in human research, where slight variations of experimental design led to significant variations of economic preferences [15]. The first research effort, as shown in this manuscript, was to investigate which cognitive biases are triggered by design modifications of the task and led to such drastic variations of economic preferences between ’single cup’ and ’multiple cups’ design.

The second research effort was to investigate which of the two experimental designs we presented in our study was the more similar to the natural context in which gorillas and orang-utans have to make economical choices and, hence, which one contained higher ecological validity. Gorillas and orang-utans are mostly herbivores and consume preferentially ripe fruits [65], therefore in the wild their economical choices are related to the search and acquisition of those fruits (risky option) over the consumption of less preferred food, such as young sprouts, leaves, or even bark (safe option). Previous studies investigated the relationship between feeding ecology and risk-preference ([24, 40]), but analysing such relationship is challenging. However, preliminary conclusions indicated that ape species with a higher appetite for ripe fruit (such as chimpanzees [66] or orang-utans [65]) were more risk-prone than ape species who are more incline to switch to less preferred food when fruit production is low (such as bonobos [67] or gorillas [65]). Our study does not allow to conclude similarly, as, maybe due to our small sample size, we did not show species differences between gorillas and orang-utans in Experiment 2, and we showed that gorillas were more risk-prone than orang-utans in Experiment 1. However, we can still use information on species feeding ecology to examine which of the experimental designs corresponded more accurately to the natural context where gorillas and orang-utans have to make economical choices. As mentioned above, gorillas and orang-utans have to choose between a preferred risky option (travel to distant patches to look for ripe fruit, and risk returning empty-handed), or a safe option (consume less preferred food, as leaves or even bark): they have to make economical choices based on their estimation of several parameters of their environment (probability to encounter fruit tree, average production of fruits, direction to go,…), and they only have a partial knowledge of the environment. As such, the economical context they have to face is more similar to that of the ’single cup’ design, as in the ’multiple cups’ design subjects theoretically have a perfect knowledge of the options they have to choose from (probability to gain the reward, amount of reward). However, wild gorillas and orang-utans also have to make a choice between different patches to explore, and not only between a less-preferred option and a variable preferred option, which resembles more to the ’multiple cups’ design. Together, gorillas and orang-utans natural economical context appears as a mix of the two designs: a ’multiple cups’ design in which each cup could yield a preferred reward (i.e., where each cup resembles the risky cup of the ’single cup’ design), and where subjects would have to evaluate the average quantity of that preferred reward as well as the probability to win each quantity of preferred reward, and where these two parameters (probability and quantity) would be variable over time (i.e., a risky but also ambiguous option, which is not at all an aspect we integrated in our task).

As a conclusion, all the above shows that due to the complexity of the economic decisions that gorillas and orang-utans face in the wild, experimental research necessarily uses simplified designs that fail to reproduce that complexity, hence the difficulty to have a clear picture of their economic strategy and preference, and the variability of economical data induced by the variability of experimental designs.

Future work is needed to explore the different potential cognitive biases we laid out in the above hypothesis and test whether one or several could explain our data. Additionally, it would be useful to confirm our findings with a more significant sample size, especially to deeper investigate the species comparison between gorillas and orang-utans. Finally, other primate species (including humans) should be tested with the same designs, as it is currently unclear whether there are systematic species differences in economic decision making. With simple experimental designs, as presented in this study, it should be possible to develop a more comprehensive phylogeny of economic decision making and cognitive biases in human and non-human primates, a further piece in the overall puzzle of the evolutionary origins of intelligence.

Conclusions

We carried out two experiments to investigate the rationality and consistency of economic decision-making in captive gorillas and orang-utans, in order to fill a gap in the literature on those two species. We chose experimental designs that have already been used in previous research and were suspected to induce different economic strategies, but we added important modifications to rule out lower-level explanations for potential differences in performance between them. Our results show that even though gorillas and orang-utans rely on rational cues (potential gain, probability to win) to establish risk-preference, they are subject to context-based cognitive biases that affect their preference. Indeed, we demonstrated a risk-preference shift from risk-neutrality to risk-proneness depending on the experimental design. The cognitive biases responsible for this shift, such as description-experience gap, impulsivity, positional bias, framing effect, are currently unknown and require targeted research.

Supporting information

S1 File. Contains all the supporting tables and figures.

(PDF)

S2 File

(DOCX)

Acknowledgments

We are grateful to Basel Zoo for allowing us to conduct research with their great apes. We are thankful to Dr. Radu Slobodeanu for statistical advice.

Data Availability

All data are available (http://doi.org/10.5281/zenodo.4709798).

Funding Statement

This work was supported with funding by the Swiss National Science Foundation (grant PZ00P3_154741 (CDD), 310030_185324 (KZ), and NCCR Evolving Language (Agreement #51NF40_180888 (KZ)), and the Taipei Medical University (Startup-funding, grant 108-6402-004-112 (CDD)). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Harrison RW, von Neumann J, Morgenstern O. The Theory of Games and Economic Behavior. Journal of Farm Economics. 1945;27: 725. doi: 10.2307/1232672 [DOI] [Google Scholar]
  • 2.Harrod RF, Stigler GJ. The Theory of Price. Economica. 1948;15: 224. doi: 10.2307/2549405 [DOI] [Google Scholar]
  • 3.Schoemaker PJH. The Expected Utility Model: Its Variants, Purposes, Evidence and Limitations. Journal of Economic Literature. 1982;20: 529–563. [Google Scholar]
  • 4.Banks J, Olson M, Porter D. An experimental analysis of the bandit problem. Economic Theory. 1997;10: 55–77. doi: 10.1007/s001990050146 [DOI] [Google Scholar]
  • 5.Meyer RJ, Shi Y. Sequential Choice Under Ambiguity: Intuitive Solutions to the Armed-Bandit Problem. Management Science. 1995;41: 817–834. doi: 10.1287/mnsc.41.5.817 [DOI] [Google Scholar]
  • 6.Sumner E, Li AX, Perfors A, Hayes B, Navarro D, Sarnecka BW. The Exploration Advantage: Children’s instinct to explore allows them to find information that adults miss. PsyArXiv; 2019. Sep. Available: https://osf.io/h437v [Google Scholar]
  • 7.Machina M. The Economic Theory of Individual Behavior Toward Risk: Theory, Evidence and New Directions. Stanford. 1983. [Google Scholar]
  • 8.Kahneman D, Tversky A. Prospect Theory: An Analysis of Decision under Risk. Econometrica. 1979;47: 263. doi: 10.2307/1914185 [DOI] [Google Scholar]
  • 9.Kahneman D, Tversky A, editors. Choices, values, and frames. New York: Cambridge, UK: Russell sage Foundation; Cambridge University Press; 2000. [Google Scholar]
  • 10.Kahneman D, Knetsch JL, Thaler RH. Anomalies: The Endowment Effect, Loss Aversion, and Status Quo Bias. Journal of Economic Perspectives. 1991;5: 193–206. doi: 10.1257/jep.5.1.193 [DOI] [Google Scholar]
  • 11.Ermer E, Cosmides L, Tooby J. Relative status regulates risky decision making about resources in men: evidence for the co-evolution of motivation and cognition. Evolution and Human Behavior. 2008;29: 106–118. doi: 10.1016/j.evolhumbehav.2007.11.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Morawetz C, Mohr PNC, Heekeren HR, Bode S. The effect of emotion regulation on risk-taking and decision-related activity in prefrontal cortex. Social Cognitive and Affective Neuroscience. 2019;14: 1109–1118. doi: 10.1093/scan/nsz078 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Atkinson JW. Motivational determinants of risk-taking behavior. Psychological Review. 1957;64: 359–372. doi: 10.1037/h0043445 [DOI] [PubMed] [Google Scholar]
  • 14.Lauriola M, Levin IP. Personality traits and risky decision-making in a controlled experimental task: an exploratory study. Personality and Individual Differences. 2001;31: 215–226. doi: 10.1016/S0191-8869(00)00130-6 [DOI] [Google Scholar]
  • 15.De Petrillo F, Paoletti M, Bellagamba F, Manzi G, Paglieri F, Addessi E. Contextual factors modulate risk preferences in adult humans. Behavioural Processes. 2020;176: 104137. doi: 10.1016/j.beproc.2020.104137 [DOI] [PubMed] [Google Scholar]
  • 16.Werner EE, Hall DJ. Optimal Foraging and the Size Selection of Prey by the Bluegill Sunfish (Lepomis Macrochirus). Ecology. 1974;55: 1042–1052. doi: 10.2307/1940354 [DOI] [Google Scholar]
  • 17.Pyke GH. Optimal Foraging Theory: A Critical Review. Annu Rev Ecol Syst. 1984;15: 523–575. doi: 10.1146/annurev.es.15.110184.002515 [DOI] [Google Scholar]
  • 18.Tregenza T. Building on the Ideal Free Distribution. Advances in Ecological Research. Elsevier; 1995. pp. 253–307. doi: 10.1016/S0065-2504(08)60067-7 [DOI] [Google Scholar]
  • 19.Krupenye C, Rosati AG, Hare B. Bonobos and chimpanzees exhibit human-like framing effects. Biol Lett. 2015;11: 20140527. doi: 10.1098/rsbl.2014.0527 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Brosnan SF, Jones OD, Lambeth SP, Mareno MC, Richardson AS, Schapiro SJ. Endowment Effects in Chimpanzees. Current Biology. 2007;17: 1704–1707. doi: 10.1016/j.cub.2007.08.059 [DOI] [PubMed] [Google Scholar]
  • 21.Drayton LA, Brosnan SF, Carrigan J, Stoinski TS. Endowment effects in gorillas (Gorilla gorilla). Journal of Comparative Psychology. 2013;127: 365–369. doi: 10.1037/a0031902 [DOI] [PubMed] [Google Scholar]
  • 22.Holt CA, Laury SK. Risk Aversion and Incentive Effects. American Economic Review. 2002;92: 1644–1655. doi: 10.1257/000282802762024700 [DOI] [Google Scholar]
  • 23.Kacelnik A, Bateson M. Risky Theories—The Effects of Variance on Foraging Decisions. Am Zool. 1996;36: 402–434. doi: 10.1093/icb/36.4.402 [DOI] [Google Scholar]
  • 24.Heilbronner SR, Rosati AG, Stevens JR, Hare B, Hauser MD. A fruit in the hand or two in the bush? Divergent risk preferences in chimpanzees and bonobos. Biol Lett. 2008;4: 246–249. doi: 10.1098/rsbl.2008.0081 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Rosati AG, Hare B. Chimpanzees and Bonobos Exhibit Emotional Responses to Decision Outcomes. di Pellegrino G, editor. PLoS ONE. 2013;8: e63058. doi: 10.1371/journal.pone.0063058 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Rosati AG, Hare B. Decision making across social contexts: competition increases preferences for risk in chimpanzees and bonobos. Animal Behaviour. 2012;84: 869–879. doi: 10.1016/j.anbehav.2012.07.010 [DOI] [Google Scholar]
  • 27.Xu ER, Kralik JD. Risky business: rhesus monkeys exhibit persistent preferences for risky options. Front Psychol. 2014;5. doi: 10.3389/fpsyg.2014.00258 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.McCoy AN, Platt ML. Risk-sensitive neurons in macaque posterior cingulate cortex. Nat Neurosci. 2005;8: 1220–1227. doi: 10.1038/nn1523 [DOI] [PubMed] [Google Scholar]
  • 29.Hayden BY, Platt ML. Temporal Discounting Predicts Risk Sensitivity in Rhesus Macaques. Current Biology. 2007;17: 49–53. doi: 10.1016/j.cub.2006.10.055 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Smith TR, Beran MJ, Young ME. Gambling in rhesus macaques (Macaca mulatta): The effect of cues signaling risky choice outcomes. Learn Behav. 2017;45: 288–299. doi: 10.3758/s13420-017-0270-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Rosati AG, Hare B. Reward currency modulates human risk preferences. Evolution and Human Behavior. 2016;37: 159–168. doi: 10.1016/j.evolhumbehav.2015.10.003 [DOI] [Google Scholar]
  • 32.Broihanne M-H, Romain A, Call J, Thierry B, Wascher CAF, De Marco A, et al. Monkeys (Sapajus apella and Macaca tonkeana) and great apes (Gorilla gorilla, Pongo abelii, Pan paniscus, and Pan troglodytes) play for the highest bid. Journal of Comparative Psychology. 2019;133: 301–312. doi: 10.1037/com0000153 [DOI] [PubMed] [Google Scholar]
  • 33.Calcutt SE, Proctor D, Berman SM, de Waal FBM. Chimpanzees Are More Averse to Social Than Nonsocial Risk. Psychol Sci. 2019;30: 105–115. doi: 10.1177/0956797618811877 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Leinwand JG, Huskisson SM, Egelkamp CL, Hopper LM. Within- and between-species variation in the responses of three primate species to a touchscreen gambling task. Learning and Motivation. 2020;71: 101635. doi: 10.1016/j.lmot.2020.101635 [DOI] [Google Scholar]
  • 35.Haux LM, Engelmann JM, Herrmann E, Hertwig R. How chimpanzees decide in the face of social and nonsocial uncertainty. Animal Behaviour. 2021;173: 177–189. doi: 10.1016/j.anbehav.2021.01.015 [DOI] [Google Scholar]
  • 36.Keupp S, Grueneisen S, Ludvig EA, Warneken F, Melis AP. Reduced risk-seeking in chimpanzees in a zero-outcome game. Phil Trans R Soc B. 2021;376: 20190673. doi: 10.1098/rstb.2019.0673 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.De Petrillo F, Rosati AG. Variation in primate decision-making under uncertainty and the roots of human economic behaviour. Phil Trans R Soc B. 2021;376: 20190671. doi: 10.1098/rstb.2019.0671 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.De Petrillo F, Ventricelli M, Ponsi G, Addessi E. Do tufted capuchin monkeys play the odds? Flexible risk preferences in Sapajus spp. Animal Cognition. 2015;18: 119–130. doi: 10.1007/s10071-014-0783-7 [DOI] [PubMed] [Google Scholar]
  • 39.Rosati AG, Hare B. Chimpanzees and bonobos distinguish between risk and ambiguity. Biol Lett. 2011;7: 15–18. doi: 10.1098/rsbl.2010.0927 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Haun DBM, Nawroth C, Call J. Great Apes’ Risk-Taking Strategies in a Decision Making Task. Santos L, editor. PLoS ONE. 2011;6: e28801. doi: 10.1371/journal.pone.0028801 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Craft BB, Church AC, Rohrbach CM, Bennett JM. The effects of reward quality on risk-sensitivity in Rattus norvegicus. Behavioural Processes. 2011;88: 44–46. doi: 10.1016/j.beproc.2011.07.002 [DOI] [PubMed] [Google Scholar]
  • 42.Ludvig EA, Madan CR, Pisklak JM, Spetch ML. Reward context determines risky choice in pigeons and humans. Biol Lett. 2014;10: 20140451. doi: 10.1098/rsbl.2014.0451 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Hanus D, Call J. When maths trumps logic: probabilistic judgements in chimpanzees. Biol Lett. 2014;10: 20140892. doi: 10.1098/rsbl.2014.0892 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Rakoczy H, Clüver A, Saucke L, Stoffregen N, Gräbener A, Migura J, et al. Apes are intuitive statisticians. Cognition. 2014;131: 60–68. doi: 10.1016/j.cognition.2013.12.011 [DOI] [PubMed] [Google Scholar]
  • 45.Broihanne M-H, Dufour V. Risk-taking in children and primates in a comparable food gambling game. Advances in Psychology Research. 2018; 78-1-53613-948–8. [Google Scholar]
  • 46.Brockman D, Schaik C, editors. Seasonality in Primates: Studies of Living and Extinct Human and Non-Human Primates. Cambridge University Press; 2005. doi: 10.1017/CBO9780511542343 [DOI] [Google Scholar]
  • 47.Koops K, Visalberghi E, van Schaik CP. The ecology of primate material culture. Biol Lett. 2014;10: 20140508. doi: 10.1098/rsbl.2014.0508 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Rogers ME, Abernethy K, Bermejo M, Cipolletta C, Doran D, Mcfarland K, et al. Western gorilla diet: A synthesis from six sites. Am J Primatol. 2004;64: 173–192. doi: 10.1002/ajp.20071 [DOI] [PubMed] [Google Scholar]
  • 49.Weber BJ, Chapman GB. Playing for peanuts: Why is risk seeking more common for low-stakes gambles? Organizational Behavior and Human Decision Processes. 2005;97: 31–46. doi: 10.1016/j.obhdp.2005.03.001 [DOI] [Google Scholar]
  • 50.Hertwig R, Erev I. The description–experience gap in risky choice. Trends in Cognitive Sciences. 2009;13: 517–523. doi: 10.1016/j.tics.2009.09.004 [DOI] [PubMed] [Google Scholar]
  • 51.Ludvig EA, Spetch ML. Of Black Swans and Tossed Coins: Is the Description-Experience Gap in Risky Choice Limited to Rare Events? Sirigu A, editor. PLoS ONE. 2011;6: e20262. doi: 10.1371/journal.pone.0020262 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Heilbronner SR, Hayden BY. The description-experience gap in risky choice in nonhuman primates. Psychon Bull Rev. 2016;23: 593–600. doi: 10.3758/s13423-015-0924-2 [DOI] [PubMed] [Google Scholar]
  • 53.Fox CR, Hadar L. “Decisions from experience” = sampling error + prospect theory: Reconsidering Hertwig, Barron, Weber & Erev (2004). Judgment and Decision Making. 2006;1: 159–161. [Google Scholar]
  • 54.Rakow T, Rahim S B.. Developmental insights into experience-based decision making. J Behav Decis Making. 2010;23: 69–82. doi: 10.1002/bdm.672 [DOI] [Google Scholar]
  • 55.Barron G, Erev I. Small feedback-based decisions and their limited correspondence to description-based decisions. J Behav Decis Making. 2003;16: 215–233. doi: 10.1002/bdm.443 [DOI] [Google Scholar]
  • 56.Russell J, Mauthner N, Sharpe S, Tidswell T. The ‘windows task’ as a measure of strategic deception in preschoolers and autistic subjects. British Journal of Developmental Psychology. 1991;9: 331–349. doi: 10.1111/j.2044-835X.1991.tb00881.x [DOI] [Google Scholar]
  • 57.Steelandt S, Thierry B, Broihanne M-H, Dufour V. The ability of children to delay gratification in an exchange task. Cognition. 2012;122: 416–425. doi: 10.1016/j.cognition.2011.11.009 [DOI] [PubMed] [Google Scholar]
  • 58.Rivière J, Stomp M, Augustin E, Lemasson A, Blois-Heulin C. Decision-making under risk of gain in young children and mangabey monkeys. Dev Psychobiol. 2018;60: 176–186. doi: 10.1002/dev.21592 [DOI] [PubMed] [Google Scholar]
  • 59.Rivière J, David E. Perceptual-motor constraints on decision making: the case of the manual search behavior for hidden objects in toddlers. J Exp Child Psychol. 2013;115: 42–52. doi: 10.1016/j.jecp.2012.11.006 [DOI] [PubMed] [Google Scholar]
  • 60.Rivière J, Lécuyer R. Effects of arm weight on C-not-B task performance: implications for the motor inhibitory deficit account of search failures. J Exp Child Psychol. 2008;100: 1–16. doi: 10.1016/j.jecp.2008.01.005 [DOI] [PubMed] [Google Scholar]
  • 61.Cohen JD, McClure SM, Yu AJ. Should I stay or should I go? How the human brain manages the trade-off between exploitation and exploration. Phil Trans R Soc B. 2007;362: 933–942. doi: 10.1098/rstb.2007.2098 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Rivière J, Kurt A, Meunier H. Choice under risk of gain in tufted capuchin monkeys (Sapajus apella): A comparison with young children (Homo sapiens) and mangabey monkeys (Cercocebus torquatus torquatus). Journal of Neuroscience, Psychology, and Economics. 2019;12: 159–168. doi: 10.1037/npe0000109 [DOI] [Google Scholar]
  • 63.Roig A, Meunier H, Poulingue E, Marty A, Thouvarecq R, Rivière J. Is economic risk proneness in young children (Homo sapiens) driven by exploratory behavior? A comparison with capuchin monkeys (Sapajus apella). Journal of Comparative Psychology. 2022;136: 140–150. doi: 10.1037/com0000314 [DOI] [PubMed] [Google Scholar]
  • 64.Malenka DJ, Baron JA, Johansen S, Wahrenberger JW, Ross JM. The framing effect of relative and absolute risk. J Gen Intern Med. 1993;8: 543–548. doi: 10.1007/BF02599636 [DOI] [PubMed] [Google Scholar]
  • 65.Conklin–Brittain NL, Knott CD, Wrangham RW. The feeding ecology of apes. The Apes: Challenges for the 21st Century. 2001. pp. 167–174. [Google Scholar]
  • 66.Wrangham RW, Conklin-Brittain NL, Hunt KD. Dietary Response of Chimpanzees and Cercopithecines to Seasonal Variation in Fruit Abundance. International Journal of Primatology. 1998;19: 949–970. doi: 10.1023/A:1020318102257 [DOI] [Google Scholar]
  • 67.Wrangham RW, Chapman CA, Clark-Arcadi AP, Isabirye-Basuta G. Social ecology of Kanyawara chimpanzees: implications for understanding the costs of great ape groups. 1st ed. In: McGrew WC, Marchant LF, Nishida T, editors. Great Ape Societies. 1st ed. Cambridge University Press; 1996. pp. 45–57. doi: 10.1017/CBO9780511752414.006 [DOI] [Google Scholar]

Decision Letter 0

Elsa Addessi

29 Oct 2021

PONE-D-21-29220Rationality and cognitive bias in captive gorillas and orang-utans economic decision-makingPLOS ONE

Dear Dr. Lacombe,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please carefully address all the points raised by both reviewers, paying particular attention to (i) correctly citing the literature on previous risk preference studies in nonhuman primates (as evidenced by both reviewers, not all nonhuman primate groups tested so far were risk averse; for instance, capuchins tested by the De Petrillo et al 2015 – your reference # 22 – were risk prone when the risky and the safe options had the same EV); (ii) better address in the Discussion the limits of your study in terms of the small sample size, which does not allow to properly evaluate age and sex effects; (iii) take into account all the methodological and statistical concerns raised by Reviewer 1.

Please take also into account the following issues:

  • Please carefully proof-read the manuscript; I spotted some typos and misspellings.

  • ll 253-55 the graphical solution doesn’t seem ideal, maybe it would be better to use plain text with three different symbols beside each number for indicating whether the EV of the risky option was higher, equal or lower than that of the safe option

  • l 286 shouldn’t it be EV=2?

  • L 304 “to check for within learning effects”: something seems to be missing here

  • L 327 “paired t-test”: didn’t you use single-sample t tests to assess whether risky choices significantly differed from the chance level?

  • L 361: ‘Multiple cup’ design

  • L 377 “Results of Experiment 2” is repeated twice

  • L 432 perhaps “had” rather than “formed” here?

  • L 495 “with feedback from the chosen option”

  • L 560 “do not gain”

  • L 562 please delete “that”

  • L 566 “a” rather than “an”

  • L 596 “species”

  • I suggest to delete, or extensively rephrase, the paragraph on ll 596-615, as it does not point out at the feeding ecology differences between gorillas and orangutans (and, indeed, a thorough discussion on the species differences is not warranted given the small sample sizes) and I am not sure whether it constructively contributes to the Discussion. 

  • L 630 please delete “of” (repeated twice)

  • L 636 “on” rather than “towards”?

  • LL 672-3 “respectively”

  • Fig 2 “(B)” is missing

  • Supplementary material: is it the spelling of the video location correct? Shouldn’t it be “zenodo” rather than “zenedo”?

Please submit your revised manuscript by Dec 13 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Elsa Addessi

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: In this study, the authors report two experiments to test rational decision-making in orangutans and gorillas in two risky choice paradigms. The topic is relevant, timely and of interest to the primate decision-making community. It is important to develop a more differentiated view on risk preferences of different species and what modulates them. Adding a sample of Gorillas and Orangutans to the picture is a valuable addition. Having said this, I have three main points of critique that I think need to be addressed carefully.

1. Premise isn’t solid. The authors state “Generally, most primates and non-primates are risk-averse” (l 99). This is contradictory to what they say prior to this (cognitive biases etc, showing context specificity of risk preferences in humans) and, importantly, this generic statement is also not adequate for nonhuman primates as a group. For example, several studies found chimpanzees to be risk-seeking (sometimes in comparison to bonobos) with procedures similar to the two-cup method (e.g., Heilbronner et al., 2008; Rosati & Hare, 2011, 2012, 2013) and in another study the authors cite already, different apes and monkeys were not throughout risk-avoidant (Broihanne et al., 2018). Since the introduction, in its current form, relies on this premise to identify the knowledge gap that the current study aims to fill, it doesn’t seem fit. There are also very different procedures, e.g. eye gaze instead of pointing to a cup, that find nonhuman primates to be risk-seeking sometimes.

In sum, I think the introduction needs some work to adequately reflect the state of current literature and integrate the current study coherently into this picture.

Some examples of relevant literature re risk-proneness in chimpanzees and eye gaze method and two recent studies showing chimps to be more risk-averse, one using a two-cup setup and one using a more complicated apparatus

• Rosati AG, Hare B. 2012 Decision making across social contexts: competition increases preferences for risk in chimpanzees and bonobos. Anim. Behav. 84, 869–879. (doi:10.1016/j.anbehav.2012.07.010

• Rosati AG, Hare B. 2013 Chimpanzees and bonobos exhibit emotional responses to decision outcomes. PLoS ONE 8, e0063058. (doi:10.1371/journal.pone. 0063058)

• Rosati AG, Hare B. 2011 Chimpanzees and bonobos distinguish between risk and ambiguity. Biol. Lett. 7, 15–18. (doi:10.1098/rsbl.2010.0927)

• McCoy, Allison N., and Michael L. Platt. "Risk-sensitive neurons in macaque posterior cingulate cortex." Nature neuroscience 8.9 (2005): 1220-1227.

• Haux, L. M., Engelmann, J. M., Herrmann, E., & Hertwig, R. (2021). How chimpanzees decide in the face of social and nonsocial uncertainty. Animal Behaviour, 173, 177-189.

• Keupp, S., Grueneisen, S., Ludvig, E. A., Warneken, F., & Melis, A. P. (2021). Reduced risk-seeking in chimpanzees in a zero-outcome game. Philosophical Transactions of the Royal Society B, 376(1819), 20190673.

2. I felt that, in general, the results and conclusions should be phrased more carefully, considering the small sample size. Don’t get me wrong, small sample size is not a reason not to publish a study. However, we all have the responsibility to point out obvious limitations in the scope of our analysis and conclusions. Sample sizes of n = 5 and n = 3 neither warrant analysis of individual factors such as sex per species nor statements such as “Our results are in stark contrast to the literature, …” (L 438-440), especially considering that the referred papers also had relatively small sample sizes. The current results add an interesting piece to the puzzle, but I wouldn’t go as far as claiming they show a huge discrepancy, yet. In this respect, I was also wondering if all the individuals showed similar patterns or behaved very differently with perhaps one individual driving the effect? (I raise this point further below in reference to result presentation).

3. Methods & statistics & results: the sections need to be substantially fleshed out. Important pieces of information are missing, the rationale for the statistical approach is not sufficiently explained and model descriptions & results need more detailed information.

Regarding statistics:

I could access the data files, but without accompanying R-code it wasn’t possible to reconstruct which analyses were run. Also, what is a “discontent analysis”? – this is mentioned in the description of the data files but has no match in the manuscript.

The information provided in the manuscript is not sufficient to understand which models were run, which predictors were included, how these predictors were treated (as categorical or continuous) and which model set out to answer which particular question.

For example, l. 309-311 – what does “risky option” refer to (the probability of the option or its size?) and which individual characteristics were assessed?

L. 319-320 – is it correct that ‘session’ entered the model as both fixed and random effect? Also, in none of the models, “protocol” seems to be included as a predictor, but wasn’t this the main goal of the study?

L 318-319: What are the fixed effects? What do you mean with “adding a significant interaction”? If the model is specified as including the interaction term as a predictor, then running the model and comparing it to respective null model will reveal which terms explain considerable parts of the variance. But you can’t only add a predictor term after you found it somehow to be significant – maybe this is a misunderstanding due to the way the sentence is phrased in the current version of the manuscript. But similar wording is in l 335-336, so I wonder what exactly were the steps that the authors performed during their analysis?

Currently, the analysis section reads like a long list of GLMMs and t-tests, but it’s difficult to follow which question each of them answers and which part of the results section refers to which of these models/t-tests. Why do you run additional t-tests when you already test for risk-preference with the GLMMs? In addition, from the information that is provided, random slopes are missing from the GLMMs. However, these are important to model effects of fixed effects among levels of random effects and to avoid inflated Type 1 error rates.

I would like to see some more information on the paired-samples t-Test that was performed to test for side bias (as stated in L 334) and provided the results reported in L 383-390 – I am struggling to see how a t-Test can account simultaneously for species differences, different number of possible positions to choose depending on number of presented cups, and side of safe reward. Or were several t-Tests run on the same data to test for these effects? In this case, which method was used to account for multiple testing, and was there a reason why these aspects were not considered as predictor terms in the GLMM in the first place?

I didn’t find information about inter-rater reliability.

Results:

It was difficult to follow the results description and judge its appropriateness, given the missing information as outlined above. A table specifying the model output would help a lot, specifying for all predictor terms the respective estimates, standard errors, confidence intervals, and test results (likelihood ratio test, degrees of freedom, p-value). And information if the model was overall different from a model excluding the predictor terms of interest; consider including effect size in result reporting, as well.

L 369: this is the first time that is mentioned that not all subjects were tested with all food rewards.

L 355: is this 45.7% for Gorillas or Orangutans? And the other species?

I don’t think assessing effects of age and sex statistically is warranted, given the samples only have 1 male per species, only 1 adult for the orangs and only 1 juvenile for the Gorillas.

L 400-406: which data and which model corresponds to these results?

Figures: It would be nice to see the individual data, especially because there are so few subjects. X-axis labelling should be adapted so it shows the possible values for the different figures (e.g., A only allows values of 2,4,6 and 7 but shows also integers in between, whereas C allows multiple values but only shows 1,2,3, and 4). Confidence intervals overlap for different lines and hide the CI of the other line (species or reward value); applying a jitter function might help.

Caption Fig.2: missing (B). Delete “for low valued reward” – its not necessary because no high value reward condition was presented.

Caption Fig.3: delete “Results of Experiment 2”.

Methods:

L 189-192: please, provide more details about subjects here (at the very least that only 1 male was in each group) and refer to supplementary table.

In the discussion, a pilot study is mentioned; which is contradictory to subjects being completely naïve to testing prior to procedure of Exp.1. Was the pilot part of this study and what was it about? Is it of relevance to the subjects’ testing history?

L 215-28: which quantities were used? What does “success rate of >=80%” mean? Did they have to reach this criterion for different quantity discriminations, or overall, or in a specified number of consecutive test sessions? How many trials did it take the individuals to reach this criterion? These informations are important to get an idea how stable the 80% performance rate was or whether it might have been a lucky accident.

L224-226: Did the subjects first see where food was placed and then pointed to the now hidden rewards, or did they have to learn to associate a cup colour with a specific content by sampling information?

L232-242: did subjects see in advance which quantity might be hidden under the risky cup? Why did you use two experimenters? On the videos it looks like E2 is watching E1 hiding the food, was this always the case?

L260-265: How was number trials decided? (for example, based on a simulation to find out necessary number of trials to find an effect, if there is one; or based on previous literature)

Design table: in a 10-trial session, how were the different win probabilities realized? How many wins and losses were presented per session per condition? It doesn’t add up for me. For example, .33 -> did you bait 3 or 4 wins within 10 trial session?

Exp. 2: did all subjects of Exp1. Participate in Exp. 2? Was there a break between the experiments?

Did they receive a familiarization training with the new procedure to learn that even when there are more cups than previously, they still only get to pick 1 and not more?

Why was the order of food type not counterbalanced between individuals?

Why is there an additional value (7 rewards) for Exp 2?

L297-299: I don’t understand this reasoning. Why was it necessary to prevent subjects from learning about the win probability? Isn’t it necessary for the subject to understand the probabilities to make an informed decision about whether to play it safe or not?

L301-303: this sentence is confusing – how is each reward amount, all numbers of cups be tested 5 times within a session that only contains 10 trials?

L297-305: Its very hard to follow the description of number of trials, sessions, and experiment repetitions. E.g. l303-> could you just say that you presented 16 sessions per reward type, i.e. 32 sessions in total? It might help to present an example of one of the orders (either in the main text or as part of the supplementary material).

I was missing an explanation why not all expected values tested in Exp.1. And why not both high and low quality food types? Seems to make it difficult to directly compare Exp. 1 and Exp.2.

Discussion

The authors raise interesting points and embed the results well, raising several possibilities for the differences they found between Exp. 1 and Exp. 2. Despite the discussion being already in a good shape, I have some remaining questions and remarks.

L 443-444: “an experimental design usually leading to risk-aversion [26]”. But Ref 26 found risk aversion in bonbos and risk proneness in chimps and thus doesn’t fit the statement very well

L 452-454: “Finally, [28] could not reject that subjects understood the task and properly inferred the probabilities to win without experience. In this study, we fixed these experimental issues” -> explanation required how exactly the current study fixed the issue.

L 456-457: On the videos provided, it looks like the baiting procedure is well in sight of Experimenter 2. Was this always the case?

L 457: “levels of risky choices” -> does this refer to the size of the risky option?

L 458-459: “probability to win to verify that subjects based their strategy on the value of that probability and were thus able to compute it” -> its not fully clear to which of the listed results this is referring.

L 512. What pilot study? It is mentioned the first time here, should the reader know about it already at this point?

The “lateral hypothesis”. The position bias is interesting, and also that the two species apparently had different kind of bias. However, what does this tell us about the risky choice results? Are they meaningful at all when subjects had a clear position bias? I think this point deserves mention and being discussed.

Reviewer #2: The article aims to investigate how gorillas and orangutans behave in two risk-assessment tasks. In experiment 1, the apes experienced a single-choice scenario where they had to decide between a cup containing a secure reward and a cup that may contain a high-value reward. Importantly, they did not know beforehand the quantity of reward at stake and whether a reward would be present or not at any given trial—although they might learn the probability that the reward was present after some experience. In experiment 2, apes experienced a multiple-choice scenario where again they had to decide between a cup containing a secure reward and a choice between a maximum of four risk options—only one being baited with a reward. Importantly, in the second study, the apes knew the risky reward at stakes and that it was always baited inside one of the cups. They found that both species seem to act rationally, increasing their choices towards the risk option when the expected value for the risk option is higher, when the probability to obtain it is higher and when its value is higher. However, they find differences between experiments, with both species being more risk-prone in experiment 2 (in those trials in which the expected value of the secure and the risky choice was the same) and differences between species, with gorillas being more risk-prone than orangutans in experiment 1.

The studies are well-conducted, and the statistical analysis is correct. In the discussion, the authors do a good job describing potential hypotheses to explain their results. However, I would like the authors to clarify some points before final acceptance.

Uncertainty vs. risk tasks

My main comment concerns the difference between uncertainty/ambiguity and risk scenarios. The authors interpret their two studies as tasks suitable to test risk preferences. However, in their discussion, they comment that during the single-cup experiment 1, apes learn the probabilities by experience (as opposed to experiment 2, in which the exact probabilities are described since apes can observe the risky reward and the number of cups). While I agree with the distinction, I am surprised that the authors do not discuss differences in terms of ambiguity/risk. One possibility is that in experiment 1, it is harder to learn the probability of appearance despite the experience. Although the study reports that a rise in the probability to win had an increasing effect on apes risk choices, one could imagine that apes were often deciding under conditions of ambiguity (especially at the beginning of experiment 1 session), and that could partially explain differences between experiments 1 and 2 overall preferences towards risk. If that were the case, the results would also align with previous studies by Rosati & Hare, 2011 and Haux et al., 2021.

Introduction

L 99 and 105: I am not sure if the ref. 26 interpretation is accurate. The chimpanzees were risk-prone in that study. Furthermore, risk options always provided rewards (either less or more grapes than the secure option).

L 132: I would stick with rewards instead of awards.

L 180: Check if the hypothesis are correct. The authors mention that they expect risk proneness in both tasks, but in L 99, they argue that primates are mostly risk-averse. I guess the authors expect more risk aversion or neutrality in experiment 1, as they have found.

Training:

In general, some details are missing. Which were the quantities involved in the training? Did they need to reach a rate of 80% correct over how many trials?

Also, what was the purpose of the hidden condition, to demonstrate that they can remember the quantities through the transparent saucer before they are covered? Did they then also need to reach at least an 80% of success in those trials? Please specify this in the manuscript.

Experiment 1:

L 260: I wonder if the authors could analyze the session effect in experiment 1 as they did in experiment 2 (instead of having session as a random effect). A session effect could tell us whether apes were learning the EV across time—increasing or decreasing their risk choices depending on the probability of obtaining the reward (P) and its value (V). The last comment relates to my previous one on the difference between ambiguity/risk choices.

It might also help to specify whether the presentation order varied between the four sessions. For instance, if the probability of food present in the risk choice was 0.5, was food present in the risk option every second trial, or was the presentation randomized as long as there would be a total of 5 trials with and without food? If trial presentation order was blocked within sessions, then ambiguity was reduced since the four sessions would be exactly the same.

Statistical analysis:

The description of the models is very clear except for the fixed-effects part. The authors describe the random effects and the additional fixed effects for experiment 2, but it is unclear which are the shared fixed effects between studies (e.g., 318). It is only apparent in the result section (e.g., the value of the risky option).

L 323: another instead of an other.

Results:

I would remind the readers that the probability of the risky option refers to P in the table, that the value of the risky option refers to V and that the expected value refers to EV =P*V.

Discussion:

L 444: Ref 26 leads to risk aversion only in bonobos.

L 487: Close parenthesis after reference 43.

L 492: either "described" or 'described'.

L 493 to 495: I find these sentences slightly unclear (e.g., "with feedback the chosen option").

L 512: It might be interesting to report some descriptives of the pilot regarding the latencies discussed.

L 563: I understand the logic in the possible distinction between the risk assessment task in the domain of gains (experiment 1) and the domain of losses (experiment 2). However, how do we know whether apes took the risk choice as the reference in experiment 2? Perhaps the authors could cite some prospect theory literature showing whether better options are usually taken as reference points?

Also, the authors mention that, given that the risk option is the reference, choosing it leads to the loss of all the reference quantity—instead of a proportion of it when choosing the safe option. However, that might not be true for all trial constellations since sometimes the risk option is equally probable, and they can also obtain it from time to time. In other words, they do not permanently lose all the “reference” quantity. Otherwise, they would always take the safe option.

L 596: close parenthesis after Experiment 2.

L 610: I would suggest making the potential role of single cup design more explicit.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Review PLOS ONE 2021.docx

PLoS One. 2022 Dec 14;17(12):e0278150. doi: 10.1371/journal.pone.0278150.r002

Author response to Decision Letter 0


13 Dec 2021

Cover letter and detailed answers to the editor’s and the reviewers’ comments

Ref.: Ms. No. PONE-D-21-29220

Rationality and cognitive bias in captive gorillas and orang-utans economic decision-making

PLOS ONE

Dear Dr Elsa Addessi, Academic Editor of PLOS ONE,

We would like to express our appreciation for your and the reviewers' useful comments and contribution, especially concerning the literature and the methodological aspects (figures and statistical analyses). We implemented all changes as suggested by the editor and the reviewers. In the following we addressed all comments and questions in details.

Our best regards,

Pénélope Lacombe, Christoph Dahl.

Editor:

correctly citing the literature on previous risk preference studies in nonhuman primates (as evidenced by both reviewers, not all nonhuman primate groups tested so far were risk averse; for instance, capuchins tested by the De Petrillo et al 2015 – your reference # 22 – were risk prone when the risky and the safe options had the same EV);

Modification of the manuscript : lines 95 to 119.

better address in the Discussion the limits of your study in terms of the small sample size, which does not allow to properly evaluate age and sex effects;

We modified the manuscript following this suggestion (the effect of age and sex are no longer analysed) and we addressed the limitation of our conclusions (lines 677 and 706).

take into account all the methodological and statistical concerns raised by Reviewer

Modification of the manuscript : lines 326 to 369.

253-55 the graphical solution doesn’t seem ideal, maybe it would be better to use plain text with three different symbols beside each number for indicating whether the EV of the risky option was higher, equal or lower than that of the safe option

Modification of the manuscript : table 1 and 2.

l 286 shouldn’t it be EV=2?

Modification of the manuscript : line 305.

L 304 “to check for within learning effects”: something seems to be missing here

Modification of the manuscript : line 322.

L 327 “paired t-test”: didn’t you use single-sample t tests to assess whether risky choices significantly differed from the chance level?

Indeed, this was an error in the manuscript. Modification of the manuscript : line 370.

L 361: ‘Multiple cup’ design

Modification of the manuscript : line 415.

L 377 “Results of Experiment 2” is repeated twice

Modification of the manuscript : line 437.

L 432 perhaps “had” rather than “formed” here?

Modification of the manuscript : line 513.

L 495 “with feedback from the chosen option”

Modification of the manuscript : line 579.

L 560 “do not gain”

Modification of the manuscript : line 644.

L 562 please delete “that”

Modification of the manuscript : line 646.

L 566 “a” rather than “an”

Modification of the manuscript : line 650.

L 596 “species”

Modification of the manuscript : line 677.

I suggest to delete, or extensively rephrase, the paragraph on ll 596-615, as it does not point out at the feeding ecology differences between gorillas and orangutans (and, indeed, a thorough discussion on the species differences is not warranted given the small sample sizes) and I am not sure whether it constructively contributes to the Discussion. 

The aim of this paragraph was not to discuss species differences but to discuss the relevance of each protocol in the study of economic strategy in great apes. Our point was that, in order to investigate great apes’ economic decision-making, one should confront the subjects with similar choices as the ones they would face in the wild. The conclusion is that, given the ecology of gorillas and orang-utans, a mixed of both designs would have more ecological relevance.

L 630 please delete “of” (repeated twice)

Modification of the manuscript : line 712.

L 636 “on” rather than “towards”?

Modification of the manuscript : line 718.

LL 672-3 “respectively”

Modification of the manuscript : lines 761, 762.

Fig 2 “(B)” is missing

Modification of the manuscript : Fig 2.

Supplementary material: is it the spelling of the video location correct? Shouldn’t it be “zenodo” rather than “zenedo”?

Modification of the manuscript :lines 769, 772.

Reviewer #1: 

Premise isn’t solid. The authors state “Generally, most primates and non-primates are risk-averse” (l 99). This is contradictory to what they say prior to this (cognitive biases etc, showing context specificity of risk preferences in humans) and, importantly, this generic statement is also not adequate for nonhuman primates as a group. For example, several studies found chimpanzees to be risk-seeking (sometimes in comparison to bonobos) with procedures similar to the two-cup method (e.g., Heilbronner et al., 2008; Rosati & Hare, 2011, 2012, 2013) and in another study the authors cite already, different apes and monkeys were not throughout risk-avoidant (Broihanne et al., 2018). Since the introduction, in its current form, relies on this premise to identify the knowledge gap that the current study aims to fill, it doesn’t seem fit. There are also very different procedures, e.g. eye gaze instead of pointing to a cup, that find nonhuman primates to be risk-seeking sometimes.

In sum, I think the introduction needs some work to adequately reflect the state of current literature and integrate the current study coherently into this picture.

Some examples of relevant literature re risk-proneness in chimpanzees and eye gaze method and two recent studies showing chimps to be more risk-averse, one using a two-cup setup and one using a more complicated apparatus

Modification of the manuscript : lines 95 to 119.

I felt that, in general, the results and conclusions should be phrased more carefully, considering the small sample size. Don’t get me wrong, small sample size is not a reason not to publish a study. However, we all have the responsibility to point out obvious limitations in the scope of our analysis and conclusions. Sample sizes of n = 5 and n = 3 neither warrant analysis of individual factors such as sex per species nor statements such as “Our results are in stark contrast to the literature, …” (L 438-440), especially considering that the referred papers also had relatively small sample sizes. The current results add an interesting piece to the puzzle, but I wouldn’t go as far as claiming they show a huge discrepancy, yet. In this respect, I was also wondering if all the individuals showed similar patterns or behaved very differently with perhaps one individual driving the effect? (I raise this point further below in reference to result presentation).

We modified the manuscript following this suggestion (the effect of age and sex are no longer analysed) and we addressed the limitation of our conclusions (lines 677 and 706). For individual data, see below.

Methods & statistics & results: the sections need to be substantially fleshed out. Important pieces of information are missing, the rationale for the statistical approach is not sufficiently explained and model descriptions & results need more detailed information.

Modification of the manuscript : lines 326 to 369.

Also, what is a “discontent analysis”? – this is mentioned in the description of the data files but has no match in the manuscript.

We removed that analysis of our study.

The information provided in the manuscript is not sufficient to understand which models were run, which predictors were included, how these predictors were treated (as categorical or continuous) and which model set out to answer which particular question.

For example, l. 309-311 – what does “risky option” refer to (the probability of the option or its size?) and which individual characteristics were assessed?

L. 319-320 – is it correct that ‘session’ entered the model as both fixed and random effect? Also, in none of the models, “protocol” seems to be included as a predictor, but wasn’t this the main goal of the study?

L 318-319: What are the fixed effects? What do you mean with “adding a significant interaction”? If the model is specified as including the interaction term as a predictor, then running the model and comparing it to respective null model will reveal which terms explain considerable parts of the variance. But you can’t only add a predictor term after you found it somehow to be significant – maybe this is a misunderstanding due to the way the sentence is phrased in the current version of the manuscript. But similar wording is in l 335-336, so I wonder what exactly were the steps that the authors performed during their analysis?

Currently, the analysis section reads like a long list of GLMMs and t-tests, but it’s difficult to follow which question each of them answers and which part of the results section refers to which of these models/t-tests. Why do you run additional t-tests when you already test for risk-preference with the GLMMs? In addition, from the information that is provided, random slopes are missing from the GLMMs. However, these are important to model effects of fixed effects among levels of random effects and to avoid inflated Type 1 error rates.I would like to see some more information on the paired-samples t-Test that was performed to test for side bias (as stated in L 334) and provided the results reported in L 383-390 – I am struggling to see how a t-Test can account simultaneously for species differences, different number of possible positions to choose depending on number of presented cups, and side of safe reward. Or were several t-Tests run on the same data to test for these effects? In this case, which method was used to account for multiple testing, and was there a reason why these aspects were not considered as predictor terms in the GLMM in the first place?

Modification of the manuscript : lines 326 to 369.

I didn’t find information about inter-rater reliability.

Modification of the manuscript : lines 326 to 328.

Results:

It was difficult to follow the results description and judge its appropriateness, given the missing information as outlined above. A table specifying the model output would help a lot, specifying for all predictor terms the respective estimates, standard errors, confidence intervals, and test results (likelihood ratio test, degrees of freedom, p-value). And information if the model was overall different from a model excluding the predictor terms of interest; consider including effect size in result reporting, as well.

Modification of the manuscript : tables 3, 4, 5 and 6. The result sections were also re-written following your suggestions (lines 382 to 397, 416 to 432, 452 to 455, and 473 to 475).

L 355: is this 45.7% for Gorillas or Orangutans? And the other species?

Modification of the manuscript : lines 400 to 403.

I don’t think assessing effects of age and sex statistically is warranted, given the samples only have 1 male per species, only 1 adult for the orangs and only 1 juvenile for the Gorillas.

We modified the manuscript following this suggestion (the effect of age and sex are no longer analysed).

L 400-406: which data and which model corresponds to these results?

Modification of the manuscript : table 6.

Figures: It would be nice to see the individual data, especially because there are so few subjects.

Individual data for Experiment 1

Individual data for Experiment 2 (low-level reward)

X-axis labelling should be adapted so it shows the possible values for the different figures (e.g., A only allows values of 2,4,6 and 7 but shows also integers in between, whereas C allows multiple values but only shows 1,2,3, and 4). Confidence intervals overlap for different lines and hide the CI of the other line (species or reward value); applying a jitter function might help.

Modification of the manuscript : figures 2, 3, 4.

Caption Fig.2: missing (B). Delete “for low valued reward” – its not necessary because no high value reward condition was presented.

Modification of the manuscript : figure 2.

Caption Fig.3: delete “Results of Experiment 2”.

Modification of the manuscript : figure 3.

Methods:

L 189-192: please, provide more details about subjects here (at the very least that only 1 male was in each group) and refer to supplementary table.

Modification of the manuscript : lines 195 to 200.

In the discussion, a pilot study is mentioned; which is contradictory to subjects being completely naïve to testing prior to procedure of Exp.1. Was the pilot part of this study and what was it about? Is it of relevance to the subjects’ testing history?

The wording “pilot study” was confusing. We just analysed the response times in both experiments and received the following results (see figure below).

L 215-28: which quantities were used? What does “success rate of >=80%” mean? Did they have to reach this criterion for different quantity discriminations, or overall, or in a specified number of consecutive test sessions? How many trials did it take the individuals to reach this criterion? These informations are important to get an idea how stable the 80% performance rate was or whether it might have been a lucky accident.

Modification of the manuscript : lines 234 to 247.

L224-226: Did the subjects first see where food was placed and then pointed to the now hidden rewards, or did they have to learn to associate a cup colour with a specific content by sampling information?

Modification of the manuscript : lines 234 to 247.

L232-242: did subjects see in advance which quantity might be hidden under the risky cup?

No, they did not.

Why did you use two experimenters?

To conduct a double-blind procedure.

On the videos it looks like E2 is watching E1 hiding the food, was this always the case?

No, this was not the case: E2 stayed close to E1 but was not looking at where the food was hidden.

L260-265: How was number trials decided? (for example, based on a simulation to find out necessary number of trials to find an effect, if there is one; or based on previous literature).

In Heilbronner et al (2008), subjects were tested in an analogous design of Experiment 1 and the learning effect was analyzed: subjects risk-preference were stable between the 3 blocks of 3 sessions (10 trials per session). 30 trials were then sufficient to establish risk-preference. In Hayden et al (2008), and Long et al (2009) similar tasks (risk-assessment in a design similar to the “single-cup” design) were conducted using blocks of 25 to 40 trials per condition.

As we conducted 20 trials per day, we performed 40 trials (2 sessions of 10 trials) for each PxV combination.

Design table: in a 10-trial session, how were the different win probabilities realized? How many wins and losses were presented per session per condition? It doesn’t add up for me. For example, .33 -> did you bait 3 or 4 wins within 10 trial session?

For P=0.3, we assigned 1 (baited cup) in every 3 trials, ex : 0 1 0 0 0 1 0 1 0 1 0 0 0 1 0 1 0 0 1 (0 0). The last two trials were not performed as we conducted 20 trials per day.

For P=0.25, we assigned 1 (baited cup) in every 4 trials, ex : 0 1 0 0 0 0 1 0 etc.

For P=5, we assigned 1 (baited cup) in every 2 trials, ex : 0 1 0 1 1 0 0 11 0 0 1 etc.

Exp. 2: did all subjects of Exp1. Participate in Exp. 2? Was there a break between the experiments?

No, one subject (female adult gorilla: Adira) dropped the experiment between E1 and E2, and one subject (male adult orang-utan: Bagus) only did E2+ (E2 with high-reward food), see table S1.These two subjects were removed from the whole experiment. Two subjects that completed E1 only completed E2- (E2 with low-reward food). They were included in the results of E1 and E2- and in the comparison of E1 and E2- (as both experiments used low-reward food), but not in the results of E2+.

There was a break between the experiments of one week.

Did they receive a familiarisation training with the new procedure to learn that even when there are more cups than previously, they still only get to pick 1 and not more?

No, but they did not try to pick several cups.

Why was the order of food type not counterbalanced between individuals?

This would have been preferable, but we obtained the authorisation from the veterinary to use high-valued food after we started the experiment. For one subject (Bagus), we started E2 with high-valued food, but he did not performed E2 with low-valued food so he was removed from the experiment.

Why is there an additional value (7 rewards) for Exp 2?

We wanted to increase the number of conditions where the expected value of the risky option was generally higher than the expected value of the safe option (without the value of 7 rewards we would have tested only one such condition where P<1 and EV > 2). This was important especially for Exp 2 as the design of Exp 1 has been studied and tested on several species before, and previous work showed that subjects understood the economic nature of the task, while Exp 2 has only been performed once (in the same design that we run it) and, prior to our study, it was still unclear weather subjects understood the task.

We were allowed a maximum value of 7 rewards per trial by the zoo veterinarian.

L297-299: I don’t understand this reasoning. Why was it necessary to prevent subjects from learning about the win probability? Isn’t it necessary for the subject to understand the probabilities to make an informed decision about whether to play it safe or not?

In E2 we did not want subjects to learn the probabilities to win, but to understand and infer probabilities without training from the number of risky cups. This is why the number of risky cups was semi-randomised within a session.

L301-303: this sentence is confusing – how is each reward amount, all numbers of cups be tested 5 times within a session that only contains 10 trials?

Each reward amount is tested during 2 non-consecutive sessions of 10 trials: among these 20 trials, each cup number (1, 2, 3 or 4) is tested 5 times.

L297-305: Its very hard to follow the description of number of trials, sessions, and experiment repetitions. E.g. l303-> could you just say that you presented 16 sessions per reward type, i.e. 32 sessions in total? It might help to present an example of one of the orders (either in the main text or as part of the supplementary material).

For one subject and one type of reward:

We performed two runs of the experiment to check for learning effects. For each run of Exp 2, we conducted 16 sessions. During each sessions we tested one reward quantity, and varied the number of risky cups between trials. Throughout the 16 sessions of one run, each PxV combination was tested 5 times (for instance in the table the grey cells corresponds to the 5 occurrences of the V=2, P=0.25 combination for the first run).

Modification of the manuscript : figure S3.

I was missing an explanation why not all expected values tested in Exp.1. And why not both high and low quality food types? Seems to make it difficult to directly compare Exp. 1 and Exp.2.

It would have been preferable to test all values in Exp.1 and to test both quality food types. Unfortunately, the experiment was already quite complex and very long to run (2.5 years), therefore we could not test the combinations of every parameters (EV, type of food) in Experiment 1 (testing one condition takes 4 times longer in Experiment 1 than in Experiment 2).

Discussion

L 443-444: “an experimental design usually leading to risk-aversion [26]”. But Ref 26 found risk aversion in bonbos and risk proneness in chimps and thus doesn’t fit the statement very well

Modification of the manuscript : lines 522-523.

L 452-454: “Finally, [28] could not reject that subjects understood the task and properly inferred the probabilities to win without experience. In this study, we fixed these experimental issues” -> explanation required how exactly the current study fixed the issue.

We modified the experimental design (counterbalancing the sides of safe and risky options, removing the refreshers trials, adding of a condition where the risky option is as large as to the safe option, using a low-level reward,...) which resulted in lower levels of risky choices, as expected, and allows to address [28] concern that subject did not understand the task, as we showed that they responded rationality to variations in probability or potential gain. This is mentioned in detail in the introduction of the manuscript, lines 144 to 155, and 176 to 180.

L 456-457: On the videos provided, it looks like the baiting procedure is well in sight of Experimenter 2. Was this always the case?

No, this was not the case: E2 stayed close to E1 but was not looking at where the food was hidden.

L 457: “levels of risky choices” -> does this refer to the size of the risky option?

No, it refers to the percentage of risky choices (this expression was used before in the result section, but can be modified if not clear).

L 458-459: “probability to win to verify that subjects based their strategy on the value of that probability and were thus able to compute it” -> its not fully clear to which of the listed results this is referring.

Modification of the manuscript : lines 537 to 544.

L 512. What pilot study? It is mentioned the first time here, should the reader know about it already at this point?

The wording “pilot study” was incorrect, we are referring to the analysis of response times in both experiments: we measured subjects' response times for every trial of Experiment 1 and 2 to test whether there was a relationship between RT and experimental design. As the result was inconclusive (longer response times in Experiment 2, which could be interpreted as due to the fact that subjects had more cups to choose from).

The “lateral hypothesis”. The position bias is interesting, and also that the two species apparently had different kind of bias. However, what does this tell us about the risky choice results? Are they meaningful at all when subjects had a clear position bias? I think this point deserves mention and being discussed.

We addressed this point in the discussion (lines 244 to 247, and 255 to 258) : even though subjects had a clear position bias, this doesn't affect risk-assessment results in gorillas (as the side of the safe option was counterbalanced between trials). For orang-utans, the lateral bias apparently interacts with subjects' choice between the safe and risky option, but as the total number of risky cups is a good predictor of levels of risky choice, this lateral bias is not the only predictors of subjects' choices, as they are making rational economic decisions.

Reviewer #2: 

Uncertainty vs. risk tasks

My main comment concerns the difference between uncertainty/ambiguity and risk scenarios. The authors interpret their two studies as tasks suitable to test risk preferences. However, in their discussion, they comment that during the single-cup experiment 1, apes learn the probabilities by experience (as opposed to experiment 2, in which the exact probabilities are described since apes can observe the risky reward and the number of cups). While I agree with the distinction, I am surprised that the authors do not discuss differences in terms of ambiguity/risk. One possibility is that in experiment 1, it is harder to learn the probability of appearance despite the experience. Although the study reports that a rise in the probability to win had an increasing effect on apes risk choices, one could imagine that apes were often deciding under conditions of ambiguity (especially at the beginning of experiment 1 session), and that could partially explain differences between experiments 1 and 2 overall preferences towards risk. If that were the case, the results would also align with previous studies by Rosati & Hare, 2011 and Haux et al., 2021.

This figure shows the level of risky choices in E1 for the 4 sessions of each PxV combination. We cannot explain the difference between E1 and E2 by the sole hypothesis that subjects make their choices under ambiguity in E1, because, as subjects learn the probability (and the quantity of reward) throughout the 4 sessions of 10 trials for each PxV condition of E1, the level of risky choices does not drastically increase (on the contrary, the level of risky choices tends to decrease).

L 99 and 105: I am not sure if the ref. 26 interpretation is accurate. The chimpanzees were risk-prone in that study. Furthermore, risk options always provided rewards (either less or more grapes than the secure option).

Modification of the manuscript : lines 95 to 119.

L 132: I would stick with rewards instead of awards.

Modification of the manuscript : line 146.

L 180: Check if the hypothesis are correct. The authors mention that they expect risk proneness in both tasks, but in L 99, they argue that primates are mostly risk-averse. I guess the authors expect more risk aversion or neutrality in experiment 1, as they have found.

Modification of the manuscript : lines 186, 187.

Training:

In general, some details are missing. Which were the quantities involved in the training? Did they need to reach a rate of 80% correct over how many trials?

Also, what was the purpose of the hidden condition, to demonstrate that they can remember the quantities through the transparent saucer before they are covered? Did they then also need to reach at least an 80% of success in those trials? Please specify this in the manuscript.

Modification of the manuscript : lines 234 to 247.

Experiment 1:

L 260: I wonder if the authors could analyze the session effect in experiment 1 as they did in experiment 2 (instead of having session as a random effect). A session effect could tell us whether apes were learning the EV across time—increasing or decreasing their risk choices depending on the probability of obtaining the reward (P) and its value (V). The last comment relates to my previous one on the difference between ambiguity/risk choices.

Modification of the manuscript : lines 388 to 393.

It might also help to specify whether the presentation order varied between the four sessions. For instance, if the probability of food present in the risk choice was 0.5, was food present in the risk option every second trial, or was the presentation randomized as long as there would be a total of 5 trials with and without food? If trial presentation order was blocked within sessions, then ambiguity was reduced since the four sessions would be exactly the same.

For P=0.3, we assigned 1 (baited cup) in every 3 trials, ex : 0 1 0 0 0 1 0 1 0 1 0 0 0 1 0 1 0 0 1 (0 0). The last two trials were not performed as we conducted 20 trials per day.

For P=0.25, we assigned 1 (baited cup) in every 4 trials, ex : 0 1 0 0 0 0 1 0 etc.

For P=5, we assigned 1 (baited cup) in every 2 trials, ex : 0 1 0 1 1 0 0 11 0 0 1 etc.

Ambiguity was then reduced as we limited the variability of the occurrence of baiting and baited the risky cup with a frequency that was highly similar to the probability.

Statistical analysis:

The description of the models is very clear except for the fixed-effects part. The authors describe the random effects and the additional fixed effects for experiment 2, but it is unclear which are the shared fixed effects between studies (e.g., 318). It is only apparent in the result section (e.g., the value of the risky option).

Modification of the manuscript : lines 326 to 369.

Results:

I would remind the readers that the probability of the risky option refers to P in the table, that the value of the risky option refers to V and that the expected value refers to EV =P*V.

Modification of the manuscript : lines 382 to 384, and 416 to 418.

Discussion:

L 444: Ref 26 leads to risk aversion only in bonobos.

Modification of the manuscript : lines 522, 523.

L 487: Close parenthesis after reference 43.

Modification of the manuscript : line 572.

L 492: either "described" or 'described'.

Modification of the manuscript : line 577.

L 493 to 495: I find these sentences slightly unclear (e.g., "with feedback the chosen option").

Modification of the manuscript : line 579.

L 512: It might be interesting to report some descriptives of the pilot regarding the latencies discussed.

We analysed the response times in both experiments and received the following results (see figure below).

L 563: I understand the logic in the possible distinction between the risk assessment task in the domain of gains (experiment 1) and the domain of losses (experiment 2). However, how do we know whether apes took the risk choice as the reference in experiment 2? Perhaps the authors could cite some prospect theory literature showing whether better options are usually taken as reference points?

We don't know whether the apes took the risk choice as the reference, but if they did, that would explain our result. We just brought up that hypothesis, but it would indeed have to be tested.

Also, the authors mention that, given that the risk option is the reference, choosing it leads to the loss of all the reference quantity—instead of a proportion of it when choosing the safe option. However, that might not be true for all trial constellations since sometimes the risk option is equally probable, and they can also obtain it from time to time. In other words, they do not permanently lose all the “reference” quantity. Otherwise, they would always take the safe option.

Indeed, the sentence was incorrect.

L 610: I would suggest making the potential role of single cup design more explicit.

Modification of the manuscript : line 693.

Attachment

Submitted filename: review_13dec_final.docx

Decision Letter 1

Elsa Addessi

21 Jan 2022

PONE-D-21-29220R1Rationality and cognitive bias in captive gorillas' and orang-utans' economic decision-makingPLOS ONE

Dear Dr. Lacombe,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses all the points raised by both reviewers during the review process.

Please submit your revised manuscript by Mar 07 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Elsa Addessi

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: (No Response)

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: First of all, I would like to thank the authors for addressing my previous comments thoroughly. While I am going to list a number of points that I still think should be addressed, I want to emphasize that I find the manuscript is already much improved and the authors’ effort to accommodate Editor’s and Reviewers’ previous comments is clearly visible.

In brief, my main points concern the analysis. Particularly missing random slopes and additional t-tests. If the results hold when random slopes are included in the models, or if the authors can justify why they do not need a random slopes structure in their analysis (maybe I overlooked something in the design of the study?), then that’s great and I have no other major concerns.

In the following, I list my points of concern and questions in more detail.

Statistical models

• L 339-340: “The significance of each predictor variable in explaining variation in rate of risky choices was tested with ANOVA” -> Could the authors clarify what they mean with this in the context of a full-null model comparison approach? In my understanding, when model comparison found that two models differ then the model output statistics tell about contribution of each predictor to explain variation and no additional ANOVAs are needed.

• L 335-343: to which analysis is this model referring? If this is meant as a general paragraph to report that assumption checks etc were done for each of the analyses, please clarify accordingly. The paragraph about full-null model comparisons should be moved up here.

• L 345-354: description of the full model to analyze Exp.1. Just minor, but this was a bit difficult to understand, because on the one hand the authors count predictor variables (up to 5), but on the other hand they have more predictor terms in their full model, namely the additional two 2-way interactions. Maybe this could be phrased a bit differently, something like this: Our full model included x and y … as well as their interaction as fixed predictor terms of interest. To control for z and xx…, we also included these as fixed effects but kept them in the null model.

• L 373: Shouldn’t species be a predictor of interest, why was it kept in the null model? I understood a potential species difference was one of the main research questions.

• L 357-361: Is this part of the original confirmatory analysis plan to test the research question of risk preference? It sounds more like a position/side bias check. It would be helpful if this was clarified explicitly.

o This reminds me of a methodological question: according to which logic were hidden reward positions determined? I assume this was somehow balanced such that each of the four possible risky positions held a risky cup equally often and a reward was hidden at each position equally often across the different number of risky cup trials? Or were risky cups always positioned immediately next to the safe cup with no empty positions when fewer than 4 risky cups were presented? It information would be useful for readers who want to use the paradigm and replicate the study.

• L 376-381: The point of modelling risk preference as a function of the above specified predictor variables is to account for the influence of these predictor variables. The described t-tests ignore these aspects. Assuming, for example, the t-test reveals risk neutrality, but the GLMM shows a clear effect of session (for example, apes are risk averse at the beginning and become increasingly risk-seeking throughout the study) – Then what have you gained with this t-test? It provides less and potentially misleading information. Therefore, I find the described t-tests inappropriate here.

• I am still missing random slopes in the model structure. Random effects need to be considered when individuals provide repeated observations. But here, individuals contribute repeated observations under different repeated conditions (i.e., the different values & probabilities of risky option, different sessions in which these conditions are tested for every individual, etc), which should be modeled via appropriate random slopes. In this case, I think one would want to include random slopes of all predictor terms of interest as well as session within individual and of all predictor terms of interest within session. (See e.g. Barr et al, 2013, for why random slopes are important to draw reliable inference about fixed effects 10.1016/j.jml.2012.11.001)

Results

• In general, I would like to see the models run with random slope structure to see if the results hold; nonetheless, I have some comments on the current results.

• L 390 – 398: The authors report about main effects of value of risky option, probability to win and session but then report the two interactions including these variables turned out significant. This means the main effects should not be interpreted because of limited use when they interact with other variables.

• L 400-404: I suppose these results are based on the additional t-tests not on the model comparison, because the authors report earlier that species remained a predictor in the null model. It would help if it was mentioned again to the reader to which analysis the respective results belong. (but see my general critique of the t-tests above)

• L 481-486: If risky choice was affected by an interaction of type of experiment and value of the risky option, then the main effect of experimental design shouldn’t be interpreted (l 481-482).

• L 485-487 this sentence doesn’t seem to reflect the model results and Fig. 4B. Model results show no interaction between experimental design and probability to win, and Fig. 4B depicts something that looks more like a main effect of experiment (with more risky choice in Exp. 2) with the exception of when p =1, to which individuals seem to be sensitive in Exp. 2 but not Exp. 1.

Discussion

• L 524-526: “in particular, if controlled for expected gains, subjects were generally risk-neutral in Experiment 1 and risk-prone in Experiment 2.” I don’t know what the authors mean by this and how they controlled for expected gains. Is this referring to section “Performance in 'single cup' vs 'multiple cup' design”? If yes, please clarify whether main effect of experimental design can be interpreted despite interacting with risk probability. Also, why is expected gain a non-economic parameter?

• L 543: For clarification that this paragraph refers specifically to Exp. 2, add “In this study, we addressed these experimental issues of the multiple cup design”

• 551: remove “did”

• 628-630: “However we found no support due to the fact that both species exhibited positional biases (gorillas: lateral bias; orang-utans: central bias) in Experiment 2”. For clarification, add a sentence explaining what you mean. For example, something like “Specifically, rather than exploring the full range of possible locations, gorillas exhibited a strong preference for the left-most location and orangutans for the central locations.”

Figure S3: there is no figure being displayed for me in the supplementary document, there is a white picture but nothing else

The authors have provided figures of individual performance in their response letter, but why not include them in the ESM for interested readers? For the figure of individual response Exp. 2: are error bars displayed correctly? It seems some values do have error bars while others don’t have them – or are they so small that they don’t display well? Its especially apparent in Figure C.

Reviewer #2: Overall, the manuscript is clearer now. However, I still find inconsistencies in how the statistical analysis are described and the results reported. I also have a few other comments in other manuscript sections that should be addressed before the final acceptance.

Introduction

L 135. The authors stated in ref. 39 that bonobos chose the risky option in 87.5% of trials. The numbers do not seem to coincide with figure 2 of ref. 39 unless the authors are averaging between small and medium sizes. Is this the case? Please clarify.

Methods

L 274. Table 1 seems to miss a minus value "(-)" after EV 1.

Statistical analysis

L 341: "we reached a model with interpretable terms".

L 356: I believe the added predictor is a sixth rather than a third predictor in the model. Experiment 2 analysis contains the same 5 predictors of Experiment 1 (risk probability, risk value, specie, session and side of the safe cup) + the addition of type of reward.

L 360: There seems to be an interaction between the position of the cup and species, as indicated in the results and in Table 5. This information is missing here, together with the fact that species was also a predictor.

L 372: I am confused because the null models included the species variable. To my understanding, that means the authors did not analyze the effect of species statistically (a full model including species against a null model without the variable) in any of their models, and therefore species differences should not be reported.

Results

General comment: The results are more precise than in the previous version of the manuscript, but I still have considerations about how they are reported. The authors constantly report main effects even when interactions between those main effects are significant. This is especially salient in Experiment 1. Probability to win, value of the risky option and session are all significant. Similarly, the interactions between session * probability and session * value are also significant. The authors should only report the interaction' effects and plot the results accordingly. In other words, the predictors' effect only makes sense in interaction with each other, not on their own anymore.

L 397: Please report more information on the significant two-way interaction. In which direction did they modify their choices?

Table 3: The inclusion of tables is helpful, but I would suggest the authors report more information regarding degrees of freedom, confidence intervals (values that are already calculated since they are plotted), and the model's estimates.

L 431: "..we found that the the session"

L 433: I would inform readers about the interaction between value and session. Also, I would suggest first reporting the effect of session in L 432 (that subjects increase their risk tendency over the experiment sessions) and then the non-significant effect of the interaction, clarifying that the apes did not learn the probabilities to win over repeated testing.

L 465: Same general comment I made earlier. If an interaction between the position of the cup and species is significant, the position of the cup on the trolley per se adds no value to the interpretation of the results.

L 483: The significant interaction between the type of experiment and the value of the risky option is not reported in table 6. In addition, what do the authors exactly mean by "subjects were more attentive to the value of the risky option in exp 1 than 2". It is unclear that attentive means that apes chose the risky option less in exp 1 compared to exp 2, as shown in figure 4A.

L 488; To clarify, trials with P = 1 refer to any trial with just two cups regardless of the number of rewards in the risky cup, right?

Discussion

L 525: I would add that controlled for expected gains means when EV = 2.

L 592: References missing for those experiments from experience providing partial feedback.

L 606: I would include the analysis in the manuscript. The authors reported the resulting plots to both reviewers in the previous review round. I think they should report the results of the analysis and the plots, at least in the supplementary materials, with a reference in the results section.

L 686: I would say that there were no apparent differences between species since the authors previously report an almost significant effect of species (p = 0.052) in Experiment 1 (although the authors may not require reporting species differences if their effect was controlled for and thus not statistically analyzed).

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Review_PLOS_Lacombe et al.pdf

PLoS One. 2022 Dec 14;17(12):e0278150. doi: 10.1371/journal.pone.0278150.r004

Author response to Decision Letter 1


7 Apr 2022

Reviewer #1:

In brief, my main points concern the analysis. Particularly missing random slopes and additional t-tests. If the results hold when random slopes are included in the models, or if the authors can justify why they do not need a random slopes structure in their analysis (maybe I overlooked something in the design of the study?), then that’s great and I have no other major concerns.

Random slopes were added to our model and we compared the full model with random slopes to simpler model removing random slopes one by one in order to keep the most parsimonious one. See lines 339->344 and supplementary tables S2, S6 and S9.

L 339-340: “The significance of each predictor variable in explaining variation in rate of risky choices was tested with ANOVA” -> Could the authors clarify what they mean with this in the context of a full-null model comparison approach? In my understanding, when model comparison found that two models differ then the model output statistics tell about contribution of each predictor to explain variation and no additional ANOVAs are needed.

The phrasing we used was confusing. We did not performed an “ANOVA” but an analysis of deviance (type II Wald chisquare test) in order to test the significance of predictors. We removed the comparison to the null model from our analysis.

• L 335-343: to which analysis is this model referring? If this is meant as a general paragraph to report that assumption checks etc were done for each of the analyses, please clarify accordingly. The paragraph about full-null model comparisons should be moved up here.

Is is indeed a general paragraph We made it more clear (line 324).

L 345-354: description of the full model to analyze Exp.1. Just minor, but this was a bit difficult to understand, because on the one hand the authors count predictor variables (up to 5), but on the other hand they have more predictor terms in their full model, namely the additional two 2-way interactions. Maybe this could be phrased a bit differently, something like this: Our full model included x and y … as well as their interaction as fixed predictor terms of interest. To control for z and xx…, we also included these as fixed effects but kept them in the null model.

We made it more clear (lines 334 to 339)

L 373: Shouldn’t species be a predictor of interest, why was it kept in the null model? I understood a potential species difference was one of the main research questions.

Indeed species should not be kept in the null model. In any case, the comparison to the null model was removed (we tested the significance of predictors with analysis of deviance).

L 357-361: Is this part of the original confirmatory analysis plan to test the research question of risk preference? It sounds more like a position/side bias check. It would be helpful if this was clarified explicitly.

We made it more clear (line 362).

This reminds me of a methodological question: according to which logic were hidden reward positions determined? I assume this was somehow balanced such that each of the four possible risky positions held a risky cup equally often and a reward was hidden at each position equally often across the different number of risky cup trials? Or were risky cups always positioned immediately next to the safe cup with no empty positions when fewer than 4 risky cups were presented? It information would be useful for readers who want to use the paradigm and replicate the study.

Indeed, we balanced the position of the baited risky cup so that each risky cup help the reward equally often. We made it more clear (lines 290->291).

L 376-381: The point of modelling risk preference as a function of the above specified predictor variables is to account for the influence of these predictor variables. The described t-tests ignore these aspects. Assuming, for example, the t-test reveals risk neutrality, but the GLMM shows a clear effect of session (for example, apes are risk averse at the beginning and become increasingly risk-seeking throughout the study) – Then what have you gained with this t-test? It provides less and potentially misleading information. Therefore, I find the described t-tests inappropriate here.

Indeed, we removed that analysis from our manuscript. We tested risk-preference at the indifference point using post-hoc tests on our models (see lines 345->352, and result section).

I am still missing random slopes in the model structure. Random effects need to be considered when individuals provide repeated observations. But here, individuals contribute repeated observations under different repeated conditions (i.e., the different values & probabilities of risky option, different sessions in which these conditions are tested for every individual, etc), which should be modeled via appropriate random slopes. In this case, I think one would want to include random slopes of all predictor terms of interest as well as session within individual and of all predictor terms of interest within session. (See e.g. Barr et al, 2013, for why random slopes are important to draw reliable inference about fixed effects 10.1016/j.jml.2012.11.001)

Random slopes were added to our model and we compared the full model with random slopes to simpler model removing random slopes one by one in order to keep the most parsimonious one. See lines 339->344 and supplementary tables S2, S6 and S9.

In general, I would like to see the models run with random slope structure to see if the results hold; nonetheless, I have some comments on the current results.

Random slopes were added to our model and we compared the full model with random slopes to simpler model removing random slopes one by one in order to keep the most parsimonious one. See lines 339->344 and supplementary tables S2, S6 and S9.

L 390 – 398: The authors report about main effects of value of risky option, probability to win and session but then report the two interactions including these variables turned out significant. This means the main effects should not be interpreted because of limited use when they interact with other variables.

Indeed, we modified the manuscript and report the main effects but do not interpret it anymore.

L 400-404: I suppose these results are based on the additional t-tests not on the model comparison, because the authors report earlier that species remained a predictor in the null model. It would help if it was mentioned again to the reader to which analysis the respective results belong. (but see my general critique of the t-tests above)

We modified that analysis and do not perform t-test anymore (see lines 406-413).

L 481-486: If risky choice was affected by an interaction of type of experiment and value of the risky option, then the main effect of experimental design shouldn’t be interpreted (l 481-482).

We modified the manuscript and removed that interpretation.

L 485-487 this sentence doesn’t seem to reflect the model results and Fig. 4B. Model results show no interaction between experimental design and probability to win, and Fig. 4B depicts something that looks more like a main effect of experiment (with more risky choice in Exp. 2) with the exception of when p =1, to which individuals seem to be sensitive in Exp. 2 but not Exp. 1.

Indeed there was a mistake in the sentence, we modified the manuscript (line 471->474).

L 524-526: “in particular, if controlled for expected gains, subjects were generally risk-neutral in Experiment 1 and risk-prone in Experiment 2.” I don’t know what the authors mean by this and how they controlled for expected gains.

We made it more clear (see line 515).

Is this referring to section “Performance in 'single cup' vs 'multiple cup' design”? If yes, please clarify whether main effect of experimental design can be interpreted despite interacting with risk probability. Also, why is expected gain a non-economic parameter?

Yes it is referring to that section. In our opinion, the effect of experimental design is multiple :

1) there is an overall higher level of risky choices in Experiment 2 than in Experiment 1 (see Fig. 4 and post-hoc tests)

2) the experimental design affects subjects' perception of the P=1 condition (see Fig. 4)

3) the experimental design affects subjects' perception of the value of the risky option (analysis of deviance in table 6 and Fig. 4)

We did not mean than expected gain is a non-economic parameter. The wording was not clear, we were only referring to experimental design when we mentioned “non-economic parameters”.

L 543: For clarification that this paragraph refers specifically to Exp. 2, add “In this study, we addressed these experimental issues of the multiple cup design”

We made it more clear (line 533)

551: remove “did”

Done

628-630: “However we found no support due to the fact that both species exhibited positional biases (gorillas: lateral bias; orang-utans: central bias) in Experiment 2”. For clarification, add a sentence explaining what you mean. For example, something like “Specifically, rather than exploring the full range of possible locations, gorillas exhibited a strong preference for the left-most location and orangutans for the central locations.”

We made it more clear (lines 618-620).

Figure S3: there is no figure being displayed for me in the supplementary document, there is a white picture but nothing else

Fixed

The authors have provided figures of individual performance in their response letter, but why not include them in the ESM for interested readers? For the figure of individual response Exp. 2: are error bars displayed correctly? It seems some values do have error bars while others don’t have them – or are they so small that they don’t display well? Its especially apparent in Figure C.

Indeed we fixed the error bars. The figures are now in the SM.

Reviewer #2: 

L 135. The authors stated in ref. 39 that bonobos chose the risky option in 87.5% of trials. The numbers do not seem to coincide with figure 2 of ref. 39 unless the authors are averaging between small and medium sizes. Is this the case? Please clarify.

We selected the trials in ref. 39 where the safe and the risky option had the same expected value (in this article, the indifference point is at EV=1). According to table 1 of ref. 39, this only corresponds to trials where the safe option was medium (V=3) and their was 3 risky cups (P=0.33). Going back to table 2 this corresponds to a level of risky choice of 75% (indeed, not 87,5%) for bonobos and 100% for chimpanzees.

L 274. Table 1 seems to miss a minus value "(-)" after EV 1.

Done

L 341: "we reached a model with interpretable terms".

Done

L 356: I believe the added predictor is a sixth rather than a third predictor in the model. Experiment 2 analysis contains the same 5 predictors of Experiment 1 (risk probability, risk value, specie, session and side of the safe cup) + the addition of type of reward.

We modified it (line 354).

L 360: There seems to be an interaction between the position of the cup and species, as indicated in the results and in Table 5. This information is missing here, together with the fact that species was also a predictor.

We modified it.

L 372: I am confused because the null models included the species variable. To my understanding, that means the authors did not analyze the effect of species statistically (a full model including species against a null model without the variable) in any of their models, and therefore species differences should not be reported.

Indeed species should not be kept in the null model. In any case, the comparison to the null model was removed (we tested the significance of predictors with analysis of deviance).

General comment: The results are more precise than in the previous version of the manuscript, but I still have considerations about how they are reported. The authors constantly report main effects even when interactions between those main effects are significant. This is especially salient in Experiment 1. Probability to win, value of the risky option and session are all significant. Similarly, the interactions between session * probability and session * value are also significant. The authors should only report the interaction' effects and plot the results accordingly. In other words, the predictors' effect only makes sense in interaction with each other, not on their own anymore.

Indeed, we modified the manuscript following your suggestion (see lines 395->405).

L 397: Please report more information on the significant two-way interaction. In which direction did they modify their choices?

See lines 395->405 and tables S4 and S5.

Table 3: The inclusion of tables is helpful, but I would suggest the authors report more information regarding degrees of freedom, confidence intervals (values that are already calculated since they are plotted), and the model's estimates.

See tables 3, 4, 5, 6 and tables S3, S8.

L 431: "..we found that the the session"

Done

L 433: I would inform readers about the interaction between value and session. Also, I would suggest first reporting the effect of session in L 432 (that subjects increase their risk tendency over the experiment sessions) and then the non-significant effect of the interaction, clarifying that the apes did not learn the probabilities to win over repeated testing.

Modified following your suggestion (434->441).

L 465: Same general comment I made earlier. If an interaction between the position of the cup and species is significant, the position of the cup on the trolley per se adds no value to the interpretation of the results.

We modified it.

L 483: The significant interaction between the type of experiment and the value of the risky option is not reported in table 6. In addition, what do the authors exactly mean by "subjects were more attentive to the value of the risky option in exp 1 than 2". It is unclear that attentive means that apes chose the risky option less in exp 1 compared to exp 2, as shown in figure 4A

We were referring to the slope of the curve Choice = f(value risky option), that is steeper in Experiment 1 than in Experiment 2, suggesting that subjects were more attentive to variations of the value of risky option in Experiment 1. However, we removed that interpretation.

L 488; To clarify, trials with P = 1 refer to any trial with just two cups regardless of the number of rewards in the risky cup, right?

Yes, any trial with only the safe and one risky cup.

L 525: I would add that controlled for expected gains means when EV = 2.

Done

L 592: References missing for those experiments from experience providing partial feedback.

Done (line 581).

L 606: I would include the analysis in the manuscript. The authors reported the resulting plots to both reviewers in the previous review round. I think they should report the results of the analysis and the plots, at least in the supplementary materials, with a reference in the results section.

Done (lines 487 to 493 and figure S9).

L 686: I would say that there were no apparent differences between species since the authors previously report an almost significant effect of species (p = 0.052) in Experiment 1 (although the authors may not require reporting species differences if their effect was controlled for and thus not statistically analyzed).

Done (lines 673 to 675).

Attachment

Submitted filename: letter-revision-6 avril.odt

Decision Letter 2

Elsa Addessi

19 May 2022

PONE-D-21-29220R2Rationality and cognitive bias in captive gorillas' and orang-utans' economic decision-makingPLOS ONE

Dear Dr. Lacombe,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

 Please very carefully take into account the additional remarks on statistical analyses provided by both reviewers, along with some minor suggestions, reported below. 

Please submit your revised manuscript by Jul 03 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Elsa Addessi

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments:

- abstract: Pongo abelii should be in italics

- l 275: “Probability of obtaining the reward when choosing the risky option” rather than “Probability of winning the reward in the risky option”

ll 278-9: “each condition (box) consisted of 40 trials”

l 315: “each condition (box) consisted of 10 non-consecutive trials”

l 320: “we did not perform consecutively” (rather than “in a row”)

l 321: “was tested in two consecutive sessions”

l 348: why was session number fitted as a categorical variable rather than a continuous variable?

l 376: “where the response variable was”

l 397: “showed that the main effects of the probability and the value of the risky options were significant (…) as well as those of the species (…) (please make a similar change also on ll 445-7 and 485-7)

ll 399 & 402: Actually, p = 0.051 or 0.052 are marginally significant, please rephrase on ll 399 and 402

l 413: please erase “that” before “the trend estimates”

l 447: “When investigating the effect of session on subjects’ choices”

l 456: “there was no difference” (please erase “specific”)

l 472: “left location” (please erase “most”)

l 486: please replace “the value to win” with “the value of the risky option”, here and wherever else applicable

l 497: “To investigate”

l 504: please erase “the” (typo)

ll 533-6: please add here a caveat concerning the small sample size of the present study

l 539 Actually, capuchin monkeys were risk prone in this task (please check “F De Petrillo, M Ventricelli, G Ponsi, E Addessi 2015 Do tufted capuchin monkeys play the odds? Flexible risk preferences in Sapajus spp. Animal Cognition 18 (1), 119-130)

ll 577-8: please rephrase as follows: “decision-makers could experience probability distortions depending on how probabilities are presented to them”

l 581: “experiment based on experience”…”experiment based on description”

l 690: please erase “their” before “preliminary conclusions”

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I have given general and specific comments to all parts of the manuscript in previous rounds of the review process and will focus on the statistics in this review.

I can see the authors worked on taking on board my and the other Reviewer’s previous comments, but still have concerns about the reported analysis.

I still can’t see a global model comparison to evaluate the contribution of the fixed predictor terms. The authors report they performed an analysis of deviance and did a model reduction by excluding random terms and nonsignificant fixed effects interaction terms. However, this process is not documented sufficiently. For example, in Table 3 we see that the three-way interactions are „significant“, but apparently some lower-level two-way interactions have been removed. It is unclear for me how the final model looked like. When higher-level interactions are included in a model, lower-level interaction terms are automatically included as well. In l 340 it says „We removed non-significant interaction terms one by one until we reached a model with interpretable terms (without non-significant interaction terms)“. But this process is not sufficiently documented (which terms were removed; at what point/which order; based on comparison with which model/criterion?).

The authors compare models with different random slopes structures, but I can’t see for what purpose this is done. Generally, a maximum random slopes structure is recommended and for comparison of models to evaluate the fixed predictor terms, the compared models should have the same random slopes structure (Barr et al., 2013. https://doi.org/10.1016/j.jml.2012.11.001).

Hence, what I would like to see is a comparison of the full model (fixed effects of interest [+ optional fixed effects of variables to only control for] + full random effect/slope structure) with a null model ([optional fixed effects to only control for] + full random effect/slope structure). If this comparison indicates that the full model doesn’t explain the data better than the null model, then that is the result. If this comparison indicates that the full model explains the data better than the null model, then the effects of the predictor terms can be assessed and potentially one continues to reduce fixed effects structure by removing nonsignificant interactions, starting with the highest-level interaction. But this global full-null model comparison is not reported, so we don’t even know if those significant terms even matter to explain the data better. See here, for example, for why model full model comparisons are important (Forstmeier & Schielzeth, 2011. 10.1007/s00265-010-1038-5 10.1007/s00265-010-1038-5).

Why is effect of species tested as part of an interaction in Exp. 1 but not in Exp. 2?

L 355: „If needed we ran post-hoc tests to calculate estimated marginal means or estimated trends“. Could you specify what „if needed“ means, please.

In general, I am a bit concerned that the authors might want too much of their data. Considering that only five resp. three individuals per species were tested, including two three-way-interactions and corresponding random slopes structure constitutes in a massively complex model! I would be interested in information about effect sizes and model stability (i.e., comparing the estimates from the model based on all data with those from models with the levels of the random effects excluded one at a time). Should it turn out that not all questions can be answered at once with the existing data set, then it would be better to simplify the analysis and reduce the number of research questions.

L 388-391 „The final model without random slopes, and with individual and session within individual as random intercepts (see supplementary table S2) was not significantly different from the full model (LRT: χ2(4)=3.83, p=.43)“. This sounds circular to me. It’s intrinsic to the process of removing terms previously and seems kind of circular. When I have a model and remove stuff that is „not significant“ and then compare this reduced model to the previous model, it’s expected that the difference is not significant again (because this was the reason why the term was removed in the first place).

L 340-341: Non-significant interaction terms should not be called „uninterpretable“.

Supplementary Figures. Thank you for including Figures on individual data. Figures S6 and S7 are very busy. Would it be possible to split the data of the two species into two parallel plots respectively? In addition, it would help to have two different symbols representing the species level in addition to different line/symbol colours, wherever both species are presented in the same plot. And it would help if a colour legend would be displayed along with every Figure. The information given in Table S1 but it‘s very unintuitive having to look this up some pages down in a table.

Fig. S4+S5: I would find it more intuitive to present session on the x-axis and display the value/probability to win of risky option as different lines.

I have noted a few spelling and grammatical errors throughout, which I personally wouldn’t mind (not being a native English speaker myself); but since Plos One specifically advices to comment on this, I recommend to run a spell and grammar check across the paper bevor a final submission.

Reviewer #2: The paper has improved and I am willing to accept the manuscript, after few minor comments and one clarification is addressed.

Main clarification point: I applaud the inclusion of the random slopes, the clarity in the interpretation of the results, and in general all statistical changes the authors have addressed. The methods and results sections are much clearer in the current version. However, although I do not think the following information is necessary to be included in an accepted version of the manuscript—I leave this decision to the editor, I would like to know why the authors did no longer use the full-null model comparison LRT method (as they do with full and final models) and instead tested the significance of their predictors with an analysis of deviance. Such a drastic change is worth an explanation before acceptance, especially when functions such as drop1 in lme4 package allows to obtain the p-values for single and interaction predictors with the LRT framework.

Besides, in L 376 there seems to be a typo: “with the response variable was the proportion”.

Also, in conclusions L 744 “are currently unknown are require..”

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Review PLOS ONE 2022 round 3.docx

PLoS One. 2022 Dec 14;17(12):e0278150. doi: 10.1371/journal.pone.0278150.r006

Author response to Decision Letter 2


6 Jul 2022

Editor :

- abstract: Pongo abelii should be in italics

- l 275: “Probability of obtaining the reward when choosing the risky option” rather than “Probability of winning the reward in the risky option”

ll 278-9: “each condition (box) consisted of 40 trials”

l 315: “each condition (box) consisted of 10 non-consecutive trials”

l 320: “we did not perform consecutively” (rather than “in a row”)

l 321: “was tested in two consecutive sessions”

Done

l 348: why was session number fitted as a categorical variable rather than a continuous variable?

Session was indeed fitted as a continuous variable, we corrected it.

l 376: “where the response variable was”

l 397: “showed that the main effects of the probability and the value of the risky options were significant (…) as well as those of the species (…) (please make a similar change also on ll 445-7 and 485-7)

ll 399 & 402: Actually, p = 0.051 or 0.052 are marginally significant, please rephrase on ll 399 and 402

l 413: please erase “that” before “the trend estimates”

l 447: “When investigating the effect of session on subjects’ choices”

l 456: “there was no difference” (please erase “specific”)

l 472: “left location” (please erase “most”)

l 486: please replace “the value to win” with “the value of the risky option”, here and wherever else applicable

l 497: “To investigate”

l 504: please erase “the” (typo)

ll 533-6: please add here a caveat concerning the small sample size of the present study

Done

l 539 Actually, capuchin monkeys were risk prone in this task (please check “F De Petrillo, M Ventricelli, G Ponsi, E Addessi 2015 Do tufted capuchin monkeys play the odds? Flexible risk preferences in Sapajus spp. Animal Cognition 18 (1), 119-130)

We specified it (line 107 and 531).

ll 577-8: please rephrase as follows: “decision-makers could experience probability distortions depending on how probabilities are presented to them”

l 581: “experiment based on experience”…”experiment based on description”

l 690: please erase “their” before “preliminary conclusions”

Done

Reviewer 1 :

I still can’t see a global model comparison to evaluate the contribution of the fixed predictor terms.

We reverted to the LRT analysis in order to compare our full model to a null model (same random structure without fixed effects), then obtained a final model by removing non-significant interaction terms based on an analysis of deviance (type II Wald chisquare test).

The authors report they performed an analysis of deviance and did a model reduction by excluding random terms and nonsignificant fixed effects interaction terms. However, this process is not documented sufficiently. In l 340 it says „We removed non-significant interaction terms one by one until we reached a model with interpretable terms (without non-significant interaction terms)“. But this process is not sufficiently documented (which terms were removed; at what point/which order; based on comparison with which model/criterion?).

We made it more clear in the current version of the manuscript.

For example, in Table 3 we see that the three-way interactions are „significant“, but apparently some lower-level two-way interactions have been removed. It is unclear for me how the final model looked like. When higher-level interactions are included in a model, lower-level interaction terms are automatically included as well.

We made it more clear. For every analysis (and for the null, full and final model), the random and fixed structures are presented in the supplemental material.

The authors compare models with different random slopes structures, but I can’t see for what purpose this is done. Generally, a maximum random slopes structure is recommended and for comparison of models to evaluate the fixed predictor terms, the compared models should have the same random slopes structure (Barr et al., 2013. https://doi.org/10.1016/j.jml.2012.11.001).

We did this in order to select the model with the more parsimonious random structure, but we removed that process in the current version and kept the complete random structure as you suggested.

Hence, what I would like to see is a comparison of the full model (fixed effects of interest [+ optional fixed effects of variables to only control for] + full random effect/slope structure) with a null model ([optional fixed effects to only control for] + full random effect/slope structure). If this comparison indicates that the full model doesn’t explain the data better than the null model, then that is the result. If this comparison indicates that the full model explains the data better than the null model, then the effects of the predictor terms can be assessed and potentially one continues to reduce fixed effects structure by removing nonsignificant interactions, starting with the highest-level interaction. But this global full-null model comparison is not reported, so we don’t even know if those significant terms even matter to explain the data better. See here, for example, for why model full model comparisons are important (Forstmeier & Schielzeth, 2011. 10.1007/s00265-010-1038-5 10.1007/s00265-010-1038-5).

We followed your suggestion and did this process for every analysis.

Why is effect of species tested as part of an interaction in Exp. 1 but not in Exp. 2?

The effect of species was tested as part of an interaction in both experiments but the effect was non-significant in Experiment 2 only, thus was removed from the final model. See supplemental tables S3 and S5 for the full models that we used in both experiments.

L 355: „If needed we ran post-hoc tests to calculate estimated marginal means or estimated trends“. Could you specify what „if needed“ means, please.

We meant that if the main effect of a predictor (ex : session) was significant, we calculated EMM (for instance) for every level of that predictor.

In general, I am a bit concerned that the authors might want too much of their data. Considering that only five resp. three individuals per species were tested, including two three-way-interactions and corresponding random slopes structure constitutes in a massively complex model!

We added three-way interactions in order to answer questions from the reviewer in the previous rounds of reviews. The random slopes structure was also added at reviewers' suggestions. Would you wish that we simplify our models ?

I would be interested in information about effect sizes and model stability (i.e., comparing the estimates from the model based on all data with those from models with the levels of the random effects excluded one at a time).

We followed your suggestion and selected the model with the more complete random structure.

L 388-391 „The final model without random slopes, and with individual and session within individual as random intercepts (see supplementary table S2) was not significantly different from the full model (LRT: χ2(4)=3.83, p=.43)“. This sounds circular to me. It’s intrinsic to the process of removing terms previously and seems kind of circular. When I have a model and remove stuff that is „not significant“ and then compare this reduced model to the previous model, it’s expected that the difference is not significant again (because this was the reason why the term was removed in the first place).

We followed your suggestion and selected the model with the more complete random structure.

L 340-341: Non-significant interaction terms should not be called „uninterpretable“.

Modified

Figures S6 and S7 are very busy. Would it be possible to split the data of the two species into two parallel plots respectively? In addition, it would help to have two different symbols representing the species level in addition to different line/symbol colours, wherever both species are presented in the same plot. And it would help if a colour legend would be displayed along with every Figure. The information given in Table S1 but it‘s very unintuitive having to look this up some pages down in a table.

Done

Fig. S4+S5: I would find it more intuitive to present session on the x-axis and display the value/probability to win of risky option as different lines.

Done

Reviewer #2

I would like to know why the authors did no longer use the full-null model comparison LRT method (as they do with full and final models) and instead tested the significance of their predictors with an analysis of deviance. Such a drastic change is worth an explanation before acceptance, especially when functions such as drop1 in lme4 package allows to obtain the p-values for single and interaction predictors with the LRT framework.

We reverted to the LRT analysis in order to compare our full model to a null model (same random structure without fixed effects), then obtained a final model by removing non-significant interaction terms based on an analysis of deviance (type II Wald chisquare test).

Besides, in L 376 there seems to be a typo: “with the response variable was the proportion”.

Also, in conclusions L 744 “are currently unknown are require..”

Done

Attachment

Submitted filename: reviewers comment (round 3).odt

Decision Letter 3

Elsa Addessi

26 Jul 2022

PONE-D-21-29220R3Rationality and cognitive bias in captive gorillas' and orang-utans' economic decision-makingPLOS ONE

Dear Dr. Lacombe,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

Please address very carefully Reviewer's 1 final remars on statistical analyses, namely provide an explicit assessment of model stability (and, as required, a caveat in the Discussion section in case your models shouldn't be stable), a revised analysis on log-transformed latencies and the required revisions on the Results section. Additionally, please proof-read the manuscript once again, as there are a few typos here and there. For instances, on l 373, "low-level reward" should be "low-valued reward".  Finally, I strongly advise to check out and cite the following papers, which are relevant for your literature review and relative hypotheses:

J Roig, A., Meunier, H., Poulingue, E., Marty, A., Thouvarecq, R., & Rivière (2022) Is economic risk proneness in young children (Homo sapiens) driven by exploratory behavior ? A comparison with capuchin monkeys (Sapajus apella). Journal of Comparative Psychology

J Rivière, A Kurt, H Meunier (2019) Choice under risk of gain in tufted capuchin monkeys (Sapajus apella): A comparison with young children (Homo sapiens) and mangabey monkeys (Cercocebus torquatus torquatus). Journal of Neuroscience, Psychology, and Economics 12 (3-4), 159                

J Rivière, M Stomp, E Augustin, A Lemasson, C Blois‐Heulin (2018) Decision‐making under risk of gain in young children and mangabey monkeys. Developmental Psychobiology 60 (2), 176-186

  ==============================

Please submit your revised manuscript by Sep 09 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Elsa Addessi

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I would like to thank the authors for taking the Reviewers’ suggestions on board. I think the manuscript is in good shape now, and I only have a couple of minor comments.

No model stability assessment is provided, yet. The authors responded to my comment to provide a model stability assessment by saying they now selected the model with complete random structure. But this wasn’t the point of the question. My point related to information on influence diagnostics. The concern is that complex models such as these (including 3-way and several 2-way interaction terms) on data that stem from a small number of cases might result in low model stability, meaning the results could be unreliable even if the model converged and tests to compare the effects of the predictor terms reveal significant effects. If influence diagnostics should reveal that this was the case, it’s still interesting to report the results, but they should be accompanied by a note of caution that model stability was limited.

Results

Exp. 1: More appropriate to start with reporting the interaction, because otherwise the reader is tempted to think these main effects are interpretable on their own. Lines 401-404: its misleading to say that you found main effects of value & probability of risky option because these variables explain the results only in relation to other variables in the interactions. These lines can be deleted.

Exp.2: Move sentence 449-451 up to beginning of the paragraph (l 443), to fill the reader in what the final model looks like, before reading about the model results in more detail.

Comparison Exp1 and Exp2: like for Exp. 1, please don’t report main effects that are part of significant interactions as stand-alone significant effects; they are not interpretable outside the interaction (l 491-494). In this paragraph (l 491-497), you switch between talking about final and full model; is it necessary here to refer to the non-significant interaction in the full model (l 496)?

Latency analysis: l 510-511: “Finally, we analysed the response times in both experiments (to assess whether the higher levels of risk-proneness in Experiment 2 could be due to high impulsivity levels)” -> Could you elaborate on this a bit, it’s not clear why you suspect potentially more impulsivity in Exp. 2 than in Exp. 1.

It didn’t catch my eye in previous rounds, but apparently no transformation was applied to the response time for the analysis. Usually, raw data of response times don’t meet model assumptions and its common to log-transform them for analysis. Was this not necessary here, were raw response times sufficiently normally distributed?

L 446: replace investigation with investigating

L452: increased -> increase

In general, I would suggest to slightly re-phrase the result descriptions throughout, to make the readers life easier. Instead of referring to the significance of an effect, describe the effect and report the stats behind the statement.

For example, instead of saying “the main effects of the probability and the value of the risky option were significant (respectively χ2(1)=13.40, p<.001 and χ2(1)=47.66, p<.001)”, re-phrase to something along the lines of “subjects picked the risky cup more/less often with increasing probability of the risky option (χ2(1)=13.40, p<.001) and more/less often with increasing value of the risky option (χ2(1)=47.66, p<.001).

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Dec 14;17(12):e0278150. doi: 10.1371/journal.pone.0278150.r008

Author response to Decision Letter 3


23 Sep 2022

We accepted all changes requested by the editor or reviewer.

Decision Letter 4

Elsa Addessi

6 Oct 2022

PONE-D-21-29220R4Rationality and cognitive bias in captive gorillas' and orang-utans' economic decision-makingPLOS ONE

Dear Dr. Lacombe,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please carefully address the following minimal changes required before the manuscript can be accepted: 

L 402, 444, 501 “Cook's distances”

L 403 please delete a space before the semicolon

L 443 please replace “as there were” with “as they were”

L 444, 501  “expect” should be “except”

L 445, 502 please delete a space after Ketawa and before the semicolon; Cook's should be with initial capital letter

L 463 “increase” rather than “increased”

LL 523-5 “as the large, risky reward was shown to the subjects before each trial in this design only which could have led subjects to choose that risky reward”: this sentence looks problematic, please clarify

L 631 “induced”

L 633 “for an example”

L 665 please move “[62]” after children

L 666 “an exploitation”

L 667 “an exploration”

L 698 please replace the comma with a semicolon

Please proof-read the entire article once more, as I kept spotting some typos.

Please submit your revised manuscript by Nov 20 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Elsa Addessi

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 5

Elsa Addessi

11 Nov 2022

Rationality and cognitive bias in captive gorillas' and orang-utans' economic decision-making

PONE-D-21-29220R5

Dear Dr. Lacombe,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Elsa Addessi

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Elsa Addessi

23 Nov 2022

PONE-D-21-29220R5

Rationality and cognitive bias in captive gorillas' and orang-utans' economic decision-making

Dear Dr. Lacombe:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Elsa Addessi

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Contains all the supporting tables and figures.

    (PDF)

    S2 File

    (DOCX)

    Attachment

    Submitted filename: Review PLOS ONE 2021.docx

    Attachment

    Submitted filename: review_13dec_final.docx

    Attachment

    Submitted filename: Review_PLOS_Lacombe et al.pdf

    Attachment

    Submitted filename: letter-revision-6 avril.odt

    Attachment

    Submitted filename: Review PLOS ONE 2022 round 3.docx

    Attachment

    Submitted filename: reviewers comment (round 3).odt

    Data Availability Statement

    All data are available (http://doi.org/10.5281/zenodo.4709798).


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES