Abstract
Social information use is widespread in the animal kingdom, helping individuals rapidly acquire useful knowledge and adjust to novel circumstances. In humans, the highly interconnected world provides ample opportunities to benefit from social information but also requires navigating complex social environments with people holding disparate or conflicting views. It is, however, still largely unclear how people integrate information from multiple social sources that (dis)agree with them, and among each other. We address this issue in three steps. First, we present a judgement task in which participants could adjust their judgements after observing the judgements of three peers. We experimentally varied the distribution of this social information, systematically manipulating its variance (extent of agreement among peers) and its skewness (peer judgements clustering either near or far from the participant's judgement). As expected, higher variance among peers reduced their impact on behaviour. Importantly, observing a single peer confirming a participant's own judgement markedly decreased the influence of other—more distant—peers. Second, we develop a framework for modelling the cognitive processes underlying the integration of disparate social information, combining Bayesian updating with simple heuristics. Our model accurately accounts for observed adjustment strategies and reveals that people particularly heed social information that confirms personal judgements. Moreover, the model exposes strong inter-individual differences in strategy use. Third, using simulations, we explore the possible implications of the observed strategies for belief updating. These simulations show how confirmation-based weighting can hamper the influence of disparate social information, exacerbate filter bubble effects and deepen group polarization. Overall, our results clarify what aspects of the social environment are, and are not, conducive to changing people's minds.
Keywords: social information use, consensus, polarization, socail learning, heuristics, cultural evolution
1. Introduction
Social information guides decision making across a broad range of animal taxa [1–3]. By interacting with others and observing their behaviour, individuals can often glean useful cues helping them to learn the location of resources, acquire new skills, and adjust to novel circumstances [4–7]. The sources of social information available to individuals are largely determined by the structure of their social network [8,9]. How individuals gather and integrate information from their social environment shapes a number of key ecological and (cultural) evolutionary processes, including the transmission of knowledge through these social networks, the dynamics of social behaviour, and the emergence and persistence of local traditions [10–16]. In humans, social information use facilitates the accumulation of cultural knowledge across generations, which is widely deemed central to the ecological success of our species [17–19]. In recent decades, technological advances (most notably the Internet) have exponentially increased the number of potential sources of social information. While this affords instant access to a wealth of useful knowledge, it is more likely than ever to encounter social sources holding disparate or conflicting views. In this paper, we examine the strategies that individuals use when they are confronted with such disparate social information coming from multiple sources.
In humans, social information use often involves changing one's mind after observing the behaviour of other individuals [20–23]. This process is commonly investigated using estimation tasks in which people are allowed to revise their initial estimates after observing the estimate of a peer (e.g. [24–28]). Studies using this approach give a detailed and quantified account of the effects of social cues on behaviour, primarily focusing on how individuals incorporate a single piece of social information [20–22,24,28,29]. Studies considering multiple peers have mainly described the effect of the central tendency (e.g. the mean of the pieces of social information provided [25,30–32]; but see [33]). In most real-world environments, however, people are confronted with multiple sources of social information at the same time, in various degrees of extremeness. Currently, it is unclear how people integrate such disparate social information. Here, we will address this issue in three steps.
First, we experimentally investigate how basic characteristics of the distribution of social information shape social information use. Specifically, we systematically manipulate the variance (reflecting the agreement among peers) and skewness (reflecting the clustering of peers close to or far away from the focal participant) of the distribution, while holding its mean constant. We show that the impact of social information strongly depends on its distribution. Disagreement among peers decreases its overall influence. Furthermore, the direction of the skew substantially alters the impact of social information: participants adjust their first estimate more when the majority of peers moderately agree with them and one peer strongly disagrees, compared to a situation in which a single peer strongly agrees with them, but the majority of peers strongly disagrees. This highlights the impact of confirmation-based weighting.
Second, we introduce a formal model to explain the strategies underlying these adjustments. This model is informed by previous research on individuals' strategies for incorporating a single piece of social information. This research has identified three distinct strategies: (i) keeping one's initial belief, (ii) adopting the behaviour of others, or (iii) ‘compromising' between personal and social information [21–23,26,27,34–40]. We develop a modelling framework that extends these insights to situations with multiple social sources, accommodating both simple heuristics (keeping and adopting), and more complex strategies (compromising). Our model successfully recovers the main patterns in the observed data: it accurately predicts how the distribution of social information impacts the relative frequencies of adjustment strategies, and accounts for the strong between-individual heterogeneity in strategy use. Our modelling results reveal that social information receives more weight when it is in line with people's initial beliefs (reflecting confirmation effects), and when it is in close agreement with other social information (reflecting peer consensus).
Finally, we use our model to predict how the observed adjustment strategies may shape belief dynamics in exemplary social environments that vary in the like-mindedness of peers (e.g. due to the social network structure). These simulations reveal how and when people's prioritising of confirmatory social information can exacerbate filter bubble effects. Moreover, they illustrate how individual differences in confirmation-based weighting can render beliefs to become more moderate (fostering group consensus) or more extreme (fuelling group polarization).
2. Experimental design
To examine how people integrate disparate information from multiple social sources, we used an adapted version of the BEAST (Berlin Estimate AdjuStment Task): a validated perceptual judgement task known to reliably measure individuals' social information use (figure 1) [29]. In the task, participants are shown several images of animal groups and have to estimate the number of animals (figure 1a,b). They then observe the estimates of three previous participants, and make a second estimate (figure 1c). The relative degree of adjustment quantifies an individual's social information use (figure 1d).
We study participants' social information use across four conditions that systematically differ in variance and skewness, while controlling for the mean deviation from a participant's first estimate (figure 1e): (i) low variance, not skewed (LN); (ii) high variance, not skewed (HN); (iii) high variance, with a cluster of two peers relatively far from the participant's first estimate (HF); and (iv) high variance, with a cluster of two peers relatively close to a participant's first estimate (HC). These conditions encompass a broad range of distributions individuals may encounter when sampling their social environment. Across all conditions, the three pieces of social information always point in the same—and correct—direction (i.e. avoiding situations in which the social information brackets the personal estimate). Importantly, holding constant the mean relative deviation from a participant's first estimate across conditions implies that a participant weighting all peer estimates equally should make similar adjustments across all conditions.
Prior to the main experiment, we pre-recorded individual estimates for each of the images by 100 individuals recruited from Amazon Mechanical Turk (MTurk), rewarding them for accuracy (electronic supplementary material, §3a). We used these estimates as social information in the main experiment. In a given round of the main experiment, the three pieces of pre-recorded social information were selected based on the participant's first estimate and the experimental condition of that round. That is, we selected those pieces of social information that most closely matched the experimental condition. This procedure allowed us to achieve experimental control without using deception (for full details and screenshots, see electronic supplementary material, §§3a and 4).
Ninety-five participants (all from the USA; 57% male; mean ± s.d. age: 35.8 ± 10.7 years) were recruited from MTurk for the main experiment, and completed 30 rounds of the judgement task. These 30 rounds included five rounds of each condition and 10 ‘filler' rounds. The social information in the filler rounds consisted of three randomly drawn estimates of the pre-recorded participants (for a given image). This procedure ensured that across all rounds, social information was sometimes higher and sometimes lower than a participant's first estimate, and sometimes bracketed the participant's own estimate. Reducing the regularity of the presented social information was expected to increase its trustworthiness. The 30 rounds were shown in a random order (and this order was the same for all participants). Throughout the task, participants did not receive feedback about their accuracy, impeding opportunities to learn about their own performance or the quality of the social information. Participants were rewarded for ccuracy: at the end of the experiment, one (first or second) estimate from the 30 rounds was selected and used for payment (see electronic supplementary material, §3a for details).
As a control, participants completed an additional block of five rounds (order counterbalanced) in which they did not observe the stimulus, but only the estimates of four peers. The distribution of these peer estimates emulated the distributions of the four experimental conditions (i.e. one of each condition), plus one filler round. This enabled us to compare how participants integrate four pieces of information of which none is their own, versus four pieces of information of which one is their own [41,42].
3. Results
(a). Experimental results
Participants' use of social information strongly depended on its distribution (figure 1f). Participants adjusted their estimates most when social information had low variance and no skewness (LN condition; figure 1f, yellow), shifting, on average, 42% towards the mean social information. In the high variance and no skew condition (HN condition; figure 1f, purple), average adjustments were credibly lower (mean adjustment: 29%; see electronic supplementary material, table S1 for statistics). Although adjustments in both conditions with skewed distributions were credibly lower than in the LN condition, the direction of the skew affected the relative adjustment: participants adjusted credibly more when two peers clustered relatively close to (HC condition; mean: 37%; figure 1f, red) rather than far from the participant is own estimate (HF condition; mean: 28%; figure 1f, green). Overall, these results demonstrate that variance and skewness in peer behaviour markedly affect social information use.
These results also show that average adjustments tended to be much smaller than what would be expected when participants weigh each of the three peer estimates as much as their own personal estimate (in which case they would adjust to the arithmetic mean of the four estimates, shifting 75% towards the mean social information in each condition). Moreover, participants substantially varied in their average adjustments, and these average adjustments strongly correlated across conditions (all pairwise Pearson correlations r ≥ 0.76; electronic supplementary material, figure S1), indicating consistent inter-individual differences in social information use (see also below).
Figure 2 zooms in on the strategies underlying behavioural adjustments across rounds, differentiating between three distinct strategies: (1) keeping the first estimate, (2) adopting the estimate of one of the three peers or (3) compromising between the first estimate and the peer estimates. The relative frequency of these strategies differed markedly between the four conditions (figure 2a–d; see electronic supplementary material, table S2 for statistics). When participants observed a single peer that closely agreed with them (HN and HF conditions; figure 2b,c), participants were more likely to either keep their first estimate, or to adopt the estimate of this near peer. When none of the peers was in close agreement with them (LN and HC conditions; figure 2a,d), participants were more likely to compromise, adjusting their estimate towards—but rarely beyond—the nearest peer. These results demonstrate that variance and skewness in peer behaviour have strong effects on the strategies people use to integrate social information. Figure 2e shows the frequency of strategies per participant across all conditions, illustrating that participants ranged from almost exclusively compromising, to exclusively keeping, with compromising being the most frequent strategy.
In all four control conditions—in which participants did not observe the stimulus, but four peer estimates, emulating the four distributions of the experimental conditions—responses were close to the arithmetic mean of the four peer estimates (electronic supplementary material, figures S2 and S3). Participants did, however, assign more weight to estimates closer to each other (electronic supplementary material, figure S4). This indicates that the observed deviations from the arithmetic mean in the experimental conditions (figures 1f and 2a–d) are not due to an inability to integrate multiple pieces of information. Rather, the stark differences between the experimental and control conditions show that people down-weight social information that is more distant from their own first estimate, an effect known as ‘egocentric discounting' [22,23,25,27,28,33,36–38,43].
(b). Cognitive model
To investigate potential cognitive mechanisms underlying individuals' integration of disparate social information, we developed a set of models unifying simple heuristics (i.e. keeping and adopting) and more complex strategies (i.e. compromising; figure 3a). Based on our behavioural findings and previous literature, we assume that an individual selects an adjustment strategy (keep, adopt or compromise) depending on the distance between its own first estimate and the estimate of the nearest peer [33,43]. We further assume that, when compromising, individuals take a weighted average of their own first estimate and social information. We model compromising as a process of Bayesian updating (figure 3a). In this process, the weight of each of the peer estimates can depend on its distance to an individual's own first estimate (confirmation-based weighting [23,25,27,28]), and its distance to other peer estimates (proximity-based weighting; electronic supplementary material figure S4 [33,43]). These assumptions are reflected in four model features, capturing the selection of (i) the keep heuristic, or (ii) the adopt heuristic, and, when compromising, the weighting of social information based on (iii) confirmation or (iv) proximity. In electronic supplementary material, §3b, we provide full details of the model implementation and analysis.
We fit a series of models to simultaneously estimate the parameter values defining participants' selection of adjustment strategies (keeping, adopting or compromising), and confirmation- and proximity-based weighting. To account for individual differences in strategy use (figure 2e), we implement hierarchical models. We evaluate the importance of the four model features: the ‘keep' and ‘adopt' heuristics, as well as ‘confirmation-' and ‘proximity-based weighting', by calculating the leave-one-out cross-validation information criterion (looic [44]) of the 16 models comprising all possible combinations of these features (electronic supplementary material, table S3). A parameter recovery analysis confirmed that our model fitting procedure yielded robust and interpretable parameter values (electronic supplementary material figure S5 and §3b).
Figure 3b compares the fits of models including versus excluding each feature. It reveals that all features—except the adopt heuristic—reliably improve the model fit. Accordingly, the best-fitting model includes the keep heuristic, and compromising with confirmation- and proximity-based weighting (electronic supplementary material, table S3). Figure 3c–e shows the effects of these three features in this best-fitting model (see electronic supplementary material, table S4 for the parameter estimates). Figure 3c shows that participants were more likely to apply a heuristic of ‘keeping' when the nearest peer was in close agreement with them. Figure 3d–e illustrates the process of compromising, showing how confirmation and proximity impact the weight that is assigned to social information.
When compromising, participants tended to weigh personal information more than social information (electronic supplementary material, table S4), and participants assigned more weight to peers who more strongly agreed with them (figure 3d). This result is indicative of confirmation-based weighting (i.e. favouring information that affirms one's beliefs). In addition, participants assigned more weight to peers who showed more agreement with other peers (figure 3e). This ‘proximity' effect was, however, weaker than the ‘confirmation' effect (as indicated by the shallower slope in figure 3e than in figure 3d). For each of these features, the model detects substantial individual differences (indicated by the thin lines in figure 3c–e), thus capturing the high inter-individual differences in mean adjustment and strategy use (figure 2f; electronic supplementary material, figure S1). Finally, we note that the absence of an effect of adopting the nearest peer estimate (figure 3b) may be because adopting can be mimicked by adjustment through compromising.
Importantly, the best-fitting model closely predicts the mean adjustment across conditions (figure 1f; red vertical lines) as well as the distributions of adjustments in rounds across conditions (figure 2a–d; red diamonds). This shows that the model can account for the main patterns observed in our experimental conditions. Our model also accurately predicts out-of-sample participants' mean adjustment and keep probability in the filler rounds (where peer estimates were randomly selected from the pre-recorded pool and frequently bracketed the participant's first estimate; electronic supplementary material, figure S6). Furthermore, the model can recover a commonly observed phenomenon in estimation tasks, namely that mean adjustments are highest when social information is at intermediate distance from first estimates (electronic supplementary material, figure S7 [25,27]). Taken together, these results suggest that our model can generalize to cases that are qualitatively different from our experimental conditions (on which the model was fitted).
(c). Simulations
The identified strategies of social information use allow us to predict how they may shape belief shifts in settings where individuals encounter peers with various levels of like-mindedness. These settings can reflect individuals' access to information from their social network being local or global [45], their personal preferences for homophily [46], or, in case of human online interactions, being controlled by algorithms prioritizing similar (or dissimilar) social sources over others, biasing the available social information [47]. In the following, we use simulations to explore how the social environment and social information use may foster consensus, or, alternatively, lead to polarization. We simulate agents who, as in the experiment, observe other estimates and adjust their first estimate. Agents observe 10 pieces of social information, across five qualitative different exemplary settings. We start with three settings in which an agent's first estimate is confirmed by either a (i) large majority, (ii) half of the peers, or (iii) only a small minority. We further simulate settings in which the focal agent is leaning towards one of two strongly disagreeing groups, and compare adjustment of agents with (iv) strong or (v) weak confirmation-based weighting. In each setting, we simulate 1000 agents whose adjustment strategies (i.e. their parameter setting) were sampled from the group-level distributions parameterized by the mean posterior estimates of the best-fitting model (see electronic supplementary material, §3c for details).
Figure 4 shows the predicted adjustments for the five scenarios. (i) When agents predominantly observe social information that agrees with their prior beliefs (as might happen in a ‘filter bubble' [48,49]), they predominantly keep their first estimate or, at most, make very small adjustments (figure 4a). (ii) Even when only half of the peers agree with them—and the other half strongly disagrees (reflecting a typical attempt to ‘de-bias' individuals [47])—agents only shift little, remaining far away from the global mean estimate (figure 4b). This suggests that even regular exposure to opposing information (e.g. from outside one's filter bubble) is unlikely to lead to substantial adjustments. (iii) Even when only a small minority agrees with them, agents are still prone to keep their first estimate, although adjustments towards the majority become more substantial (figure 4c). Overall, these simulation results illustrate how confirmation-based weighting can curb belief updating: confirmatory social information reinforces people's prior beliefs and prompts them to retain these beliefs, even if they reflect minority views.
To further examine the possible implications of individual differences in confirmation-based weighting, we simulated adjustments for agents showing weak and strong confirmation effects (i.e. individuals with steep and shallow slopes in figure 3d). Agents were initially located in between two clusters of peers, but slightly closer to one of the clusters. In this setting, agents with strong confirmation-based weighting tend to adjust towards the local cluster—moving away from the global mean (figure 4d). Weak confirmation-based weighting tends to lead to adjustments towards the global mean estimate (figure 4e). These results suggest that strong confirmation-based weighting can drive people to more extreme views, and may increase polarization over time.
4. Discussion
This paper makes three novel contributions. First, we experimentally show that the impact of multiple sources of social information strongly depends on its distribution. Increased variance in social information reduces participants' adjustments, and skewness decreases adjustments if a single peer confirms their first estimate. Second, our cognitive model provides a unified account for how people integrate disparate social information, showing that people rely on a combination of simple heuristics of keeping their initial beliefs and compromising towards social information. The model captures how the weight of social information is determined by both its degree of confirmation of people's initial beliefs, and its proximity to other pieces of social information. Third, the model made accurate out-of-sample predictions of adjustments in qualitatively different judgement situations, and our simulations illustrate how prioritising confirmatory social information may lead individuals to take up more extreme beliefs.
Overall, people assigned more weight to their personal initial beliefs than to social information (figures 1f and 2a–d; electronic supplementary material, figure S3, figure S4 and table S4). It may seem somewhat puzzling why they would do so in a task in which social information consists of judgements of people incentivised to accurately solve the same problem. Indeed, there is no reason to assume that one's first estimate would be more accurate than those of others. One rationale for prioritising personal estimates is the lack of access to others' reasons for holding their beliefs, which may lead participants to discount social information, especially when it is very distinct from one's personal beliefs [28,50] (but see [41]).
Our cognitive model provides a detailed account of people's social information use, contributing to understanding its underlying computational and cognitive mechanisms [51]. The model presents a unified framework laying out how people integrate personal and social information, by combining a heuristic strategy (keeping) with compromising (weighted averaging). We further obtain a detailed picture of the process of compromising by formalizing how people weigh several pieces of social information, the combined effects of which would be hard to understand without a model. Interestingly, while social information that is consistent with own personal beliefs is weighted more when people compromise (confirmation-based weighting; figure 3d), it also increases the chance that people simply keep their initial beliefs (figure 3c). The combined result of these effects is that social information tends to have the strongest impact on overall adjustments when it is at intermediate distance (electronic supplementary material, figure S7). Our results reveal that people also weigh social information based on its consistency with other social information (‘proximity-based weighting'; figure 3e; electronic supplementary material, figure S4). One rationale for prioritising social information consistent with other social information is relatively straightforward: when people are motivated and able to make valid judgements, agreement among peers reliably signals accuracy [52,53]. The concerted action of the three mechanisms (heuristics of keeping, and compromising based on confirmation and peer proximity) explain our experimental observations that the impact of social information strongly depends on the variance and skewness of its distribution.
Our model's accurate predictions of behaviour in the filler rounds underscores its ability to go beyond mere redescription of the data it was fitted to, and suggests that our model can be generalized to settings that are qualitatively different from our experimental conditions (electronic supplementary material, figure S5 and figure S6). Our simulations go beyond the limited set of distributions of social information studied in our experiments, generating predictions about how social information use may be shaped by distributions resembling important real-world settings. The advent of the Internet has dramatically changed the structure and dynamics of social interactions; at times being (algorithmically) biased towards like-minded sources [47,49,54], but also giving people access to diverse social sources with potentially conflicting views. Our simulations predict that disparate social information changes people's minds only to a limited degree, even when this social information signals that people hold minority views (figure 4). More importantly, under certain conditions, observing balanced social information can even lead individuals with strong confirmation-based weighting to take more extreme views (figure 4d). These findings have direct implications for interventions. For instance, they suggest that efforts to de-bias online information that present people with balanced views [47] might not suffice to break filter bubble effects and dynamics of polarization. Future work could more explicitly address the role of the social network structure, testing how the distribution of beliefs across social networks may interact with network structure in governing the dynamics of belief updating processes in a population (e.g. [55]). We believe our cognitive modelling framework can help in achieving a mechanistic understanding of how social information use may contribute to the formation of group consensus or the risk of polarization, and more generally, how the distribution of individual strategies of social information use in a population drives the transmission of information across social networks and shapes the course of cultural evolution.
The current study provides a robust template for understanding social information use in a range of (complex) social environments that people encounter in their day-to-day lives. Future empirical work should test the predictions of our simulations, as well as the extent to which our findings—obtained with a stylised perceptual judgement task with anonymous peers—generalize to other domains of decision making. Our results demonstrate that confirmation-based weighting can strongly reduce people's willingness to change their minds even in the minimal setting of a task with an objectively correct solution. It seems plausible that confirmation-based weighting—and its polarizing consequences—are even stronger in many important real-world contexts involving emotive, moral or political issues. In these (often controversial) contexts, the integration of disparate social information may be further hampered due to ‘motivated reasoning' [56] or when observed individuals belong to an out-group [57,58]. Conversely, disparate social information might impact behaviour more strongly when it stems from peers who are familiar [59], similar [11,60], prestigious [61,62], or known to have expertise in the task at hand [7,63,64]. Moreover, the effects of each of these factors are likely to substantially differ between individuals (cf. figures 2e and 3c–e) and between societies [65–68]. Our experimental design and modelling framework are flexible enough to include adaptations to accommodate each of these elements, helping understand the features of social sources that may influence weighting. For example, if social sources vary in their expertise, individuals might assign more weight to social information provided by an expert rather than by a non-expert. By explicitly accounting for the amount of variance people assign to social information from experts versus non-experts, an extended version of our model could distinguish how the impact of social information depends on the source's expertise, beyond its degree of alignment with people's initial beliefs. Furthermore, our hierarchical modelling approach allows accounting for individual differences in social information use, a regularly observed but poorly understood phenomenon in humans and other animals [65,69–73]. Linking individuals' strategies and their underlying cognitive mechanisms to genetic, developmental and cultural processes may help unearth the causes of between-individual and between-society differences in social information use.
To conclude, our findings contribute to a growing literature on how people integrate social information to update their beliefs. We go beyond previous experimental work on the effects of social information (which focused on single social cues or the mean of multiple cues) by showing that the variance and skewness of its distribution strongly modulates its impact on behaviour. Our cognitive model provides a detailed picture of the cognitive mechanisms that underlie the integration of disparate social information, highlighting the role of heuristics of keeping one's initial beliefs, and the importance of confirmation- and proximity-based weighting. Finally, our simulations consider various exemplary social environments to illustrate how confirmation-based weighting can markedly exacerbate filter bubble dynamics and polarization. We anticipate that these findings will provide a useful point of departure for future work aiming to understand the nature of human social information use and its implications for group dynamics, and to inform interventions to effectively de-bias individuals and help them forming accurate beliefs about the world.
Supplementary Material
Acknowledgements
We thank Casper Hesp and the members of the Connected Minds Lab at the University of Amsterdam for useful comments and discussions.
Ethics
Ethical approval was obtained from the Institutional Review Board of the Max Planck Institute for Human Development Berlin (ARC 2017/18).
Data accessibility
All code associated with this paper (i.e. experimental software, cognitive model, simulations and analyses) is publicly available in the public OSF repository https://osf.io/rmcuy/.
Authors' contributions
L.M., A.G., R.H.J.M.K. and W.v.d.B. designed the study. L.M. and A.G. programmed the experiment, A.G. collected the data, L.M., A.N.T., S.H. and W.v.d.B. developed and analysed the model. All authors were involved in interpreting the results, and writing and editing the manuscript.
Competing interests
We declare we have no competing interests.
Funding
L.M. and W.v.d.B. are supported by an Open Research Area grant (ID 176). L.M. is further supported by an Amsterdam Brain and Cognition Project grant 2018. W.v.d.B. is further supported by the Jacobs Foundation European Research Council grant no. (ERC-2018-StG-803338) and the Netherlands Organization for Scientific Research grant no. (NWO-VIDI 016.Vidi.185.068).
References
- 1.Dall SRX, Giraldeau L-A, Olsson O, McNamara JM, Stephens DW. 2005. Information and its use by animals in evolutionary ecology. Trends Ecol. Evol. 20, 187–193. ( 10.1016/j.tree.2005.01.010) [DOI] [PubMed] [Google Scholar]
- 2.Seppänen J-T, Forsman JT, Mönkkönen M, Thomson RL. 2007. Social information use is a process across time, space, and ecology, reaching heterospecifics. Ecology 88, 1622–1633. ( 10.1890/06-1757.1) [DOI] [PubMed] [Google Scholar]
- 3.Danchin É, Giraldeau L-A, Valone TJ, Wagner RH. 2004. Public information: from nosy neighbors to cultural evolution. Science 305, 487–491. ( 10.1126/science.1098254) [DOI] [PubMed] [Google Scholar]
- 4.Laland KN. 2004. Social learning strategies. Anim. Learn. Behav. 32, 4–14. ( 10.3758/BF03196002) [DOI] [PubMed] [Google Scholar]
- 5.Rendell L, Fogarty L, Hoppitt WJE, Morgan TJH, Webster MM, Laland KN. 2011. Cognitive culture: theoretical and empirical insights into social learning strategies. Trends Cogn. Sci. 15, 68–76. ( 10.1016/j.tics.2010.12.002) [DOI] [PubMed] [Google Scholar]
- 6.Boyd R, Richerson PJ, Henrich J. 2011. The cultural niche: why social learning is essential for human adaptation. Proc. Natl Acad. Sci. USA 108, 10 918–10 925. ( 10.1073/pnas.1100290108) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Kendal RL, Boogert NJ, Rendell L, Laland KN, Webster M, Jones PL. 2018. Social learning strategies: bridge-building between fields. Trends Cogn. Sci. 22, 651–665. ( 10.1016/j.tics.2018.04.003) [DOI] [PubMed] [Google Scholar]
- 8.Kurvers RH, Krause J, Croft DP, Wilson AD, Wolf M. 2014. The evolutionary and ecological consequences of animal social networks: emerging issues. Trends Ecol. Evol. 29, 326–335. ( 10.1016/j.tree.2014.04.002) [DOI] [PubMed] [Google Scholar]
- 9.Aplin LM, Farine DR, Morand-Ferron J, Sheldon BC. 2012. Social networks predict patch discovery in a wild population of songbirds. Proc. R. Soc. B 279, 4199–4205. ( 10.1098/rspb.2012.1591) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Cavalli-Sforza LL, Feldman MW. 1981. Cultural transmission and evolution: a quantitative approach. Princeton, NJ: Princeton University Press. [PubMed] [Google Scholar]
- 11.Boyd R, Richerson PJ. 1985. Culture and the evolutionary process. Chicago, IL: University of Chicago Press. [Google Scholar]
- 12.Mesoudi A. 2011. Cultural evolution: how Darwinian theory can explain human culture and synthesize the social sciences. Chicago, IL: University of Chicago Press. [Google Scholar]
- 13.Aoki K, Lehmann L, Feldman MW. 2011. Rates of cultural change and patterns of cultural accumulation in stochastic models of social transmission. Theor. Popul. Biol. 79, 192–202. ( 10.1016/j.tpb.2011.02.001) [DOI] [PubMed] [Google Scholar]
- 14.Hoppitt W, Laland KN. 2013. Social learning: an introduction to mechanisms, methods, and models. Princeton, NJ: Princeton University Press. [Google Scholar]
- 15.Aplin LM, Farine DR, Morand-Ferron J, Cockburn A, Thornton A, Sheldon BC. 2015. Experimentally induced innovations lead to persistent culture via conformity in wild birds. Nature 518, 538–541. ( 10.1038/nature13998) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Lehmann L, Feldman MW. 2008. The co-evolution of culturally inherited altruistic helping and cultural transmission under random group formation. Theor. Popul. Biol. 73, 506–516. ( 10.1016/j.tpb.2008.02.004) [DOI] [PubMed] [Google Scholar]
- 17.Henrich J. 2015. The secret of our success: how culture is driving human evolution, domesticating our species, and making us smarter. Princeton, NJ: Princeton University Press. [Google Scholar]
- 18.Tomasello M. 2009. The cultural origins of human cognition. Cambridge, MA: Harvard University Press. [Google Scholar]
- 19.Richerson PJ, Boyd R. 2004. Not by genes alone: how culture transformed human evolution. Chicago, IL: University of Chicago Press. [Google Scholar]
- 20.Bednarik P, Schultze T. 2015. The effectiveness of imperfect weighting in advice taking. Judgment Decis. Mak. 10, 265–276. [Google Scholar]
- 21.Soll JB, Larrick RP. 2009. Strategies for revising judgment: how (and how well) people use others' opinions. J. Exp. Psychol. Learn. Memory Cogn. 35, 780–805. ( 10.1037/a0015145) [DOI] [PubMed] [Google Scholar]
- 22.Yaniv I. 1997. Weighting and trimming: heuristics for aggregating judgments under uncertainty. Org. Behav. Hum. Decis. Process. 69, 237–249. ( 10.1006/obhd.1997.2685) [DOI] [Google Scholar]
- 23.Yaniv I. 2004. Receiving other people's advice: Influence and benefit. Org. Behav. Hum. Decis. Process. 93, 1–13. ( 10.1016/j.obhdp.2003.08.002) [DOI] [Google Scholar]
- 24.Bonaccio S, Dalal RS. 2006. Advice taking and decision-making: an integrative literature review, and implications for the organizational sciences. Org. Behav. Hum. Decis. Process. 101, 127–151. ( 10.1016/j.obhdp.2006.07.001) [DOI] [Google Scholar]
- 25.Jayles B, Kim H, Escobedo R, Cezera S, Blanchet A, Kameda T, Sire C, Theraulaz G. 2017. How social information can improve estimation accuracy in human groups. Proc. Natl Acad. Sci. USA 114, 12 620–12 625. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Moussaïd M, Herzog SM, Kämmer JE, Hertwig R. 2017. Reach and speed of judgment propagation in the laboratory. Proc. Natl Acad. Sci. USA 114, 201611998 ( 10.1073/pnas.1611998114) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Moussaïd M, Kämmer JE, Analytis PP, Neth H. 2013. Social influence and the collective dynamics of opinion formation. PLoS ONE 8, e78433 ( 10.1371/journal.pone.0078433) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Yaniv I, Kleinberger E. 2000. Advice taking in decision making: egocentric discounting and reputation formation. Org. Behav. Hum. Decis. Process. 83, 260–281. ( 10.1006/obhd.2000.2909) [DOI] [PubMed] [Google Scholar]
- 29.Molleman L, Kurvers RHJM, van den Bos W. 2019. Unleashing the BEAST: a brief measure of human social information use. Evol. Hum. Behav. 40, 492–499. ( 10.1016/j.evolhumbehav.2019.06.005) [DOI] [Google Scholar]
- 30.Larrick RP, Soll JB. 2006. Intuitions about combining opinions: misappreciation of the averaging principle. Manage. Sci. 52, 111–127. ( 10.1287/mnsc.1050.0459) [DOI] [Google Scholar]
- 31.Mannes AE. 2009. Are we wise about the wisdom of crowds? The use of group judgments in belief revision. Manage. Sci. 55, 1267–1279. ( 10.1287/mnsc.1090.1031) [DOI] [Google Scholar]
- 32.Park SA, Goïame S, O'Connor DA, Dreher J-C. 2017. Integration of individual and social information for decision-making in groups of different sizes. PLoS Biol. 15, e2001958 ( 10.1371/journal.pbio.2001958) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Yaniv I, Milyavsky M. 2007. Using advice from multiple sources to revise and improve judgments. Org. Behav. Hum. Decis. Process. 103, 104–120. ( 10.1016/j.obhdp.2006.05.006) [DOI] [Google Scholar]
- 34.Aitchison L, Bang D, Bahrami B, Latham PE. 2015. Doubly Bayesian analysis of confidence in perceptual decision-making. PLoS Comput. Biol. 11, e1004519 ( 10.1371/journal.pcbi.1004519) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Bahrami B, Olsen K, Latham PE, Roepstorff A, Rees G, Frith CD. 2010. Optimally interacting minds. Science 329, 1081–1085. ( 10.1126/science.1185718) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Budescu DV, Rantilla AK, Yu H-T, Karelitz TM. 2003. The effects of asymmetry among advisors on the aggregation of their opinions. Org. Behav. Hum. Decis. Process. 90, 178–194. ( 10.1016/S0749-5978(02)00516-2) [DOI] [Google Scholar]
- 37.Budescu DV, Yu H-T. 2007. Aggregation of opinions based on correlated cues and advisors. J. Behav. Decis. Mak. 20, 153–177. ( 10.1002/bdm.547) [DOI] [Google Scholar]
- 38.Harries C, Yaniv I, Harvey N. 2004. Combining advice: the weight of a dissenting opinion in the consensus. J. Behav. Decis. Mak. 17, 333–348. ( 10.1002/bdm.474) [DOI] [Google Scholar]
- 39.Shea N, Boldt A, Bang D, Yeung N, Heyes C, Frith CD. 2014. Supra-personal cognitive control and metacognition. Trends Cogn. Sci. 18, 186–193. ( 10.1016/j.tics.2014.01.006) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Toelch U, Dolan RJ. 2015. Informational and normative influences in conformity from a neurocomputational perspective. Trends Cogn. Sci. 19, 579–589. ( 10.1016/j.tics.2015.07.007) [DOI] [PubMed] [Google Scholar]
- 41.Soll JB, Mannes AE. 2011. Judgmental aggregation strategies depend on whether the self is involved. Int. J. Forecast. 27, 81–102. ( 10.1016/j.ijforecast.2010.05.003) [DOI] [Google Scholar]
- 42.Yaniv I, Choshen-Hillel S. 2012. Exploiting the wisdom of others to make better decisions: suspending judgment reduces egocentrism and increases accuracy. J. Behav. Decis. Mak. 25, 427–434. ( 10.1002/bdm.740) [DOI] [Google Scholar]
- 43.Schultze T, Rakotoarisoa A-F, Schulz-Hard S. 2015. Effects of distance between initial estimates and advice on advice utilization. Judgment Decis. Mak. 10, 144–172. [Google Scholar]
- 44.Vehtari A, Gelman A, Gabry J. 2016. loo: efficient approximate leave-one-out cross-validation (LOO) and WAIC for Bayesian models. R package version 1. [Google Scholar]
- 45.Croft DP, James R, Krause J. 2008. Exploring animal social networks. Princeton, NJ: Princeton University Press. [Google Scholar]
- 46.McPherson M, Smith-Lovin L, Cook JM. 2001. Birds of a feather: homophily in social networks. Ann. Rev. Sociol. 27, 415–444. ( 10.1146/annurev.soc.27.1.415) [DOI] [Google Scholar]
- 47.Bozdag E, van den Hoven J. 2015. Breaking the filter bubble: democracy and design. Ethics Info. Technol. 17, 249–265. ( 10.1007/s10676-015-9380-y) [DOI] [Google Scholar]
- 48.Bakshy E, Messing S, Adamic LA. 2015. Exposure to ideologically diverse news and opinion on Facebook. Science 348, 1130–1132. ( 10.1126/science.aaa1160) [DOI] [PubMed] [Google Scholar]
- 49.Pariser E. 2011. The filter bubble: what the internet is hiding from you. London, UK: Penguin. [Google Scholar]
- 50.Ravazzolo F, Røisland Ø. 2011. Why do people place lower weight on advice far from their own initial opinion? Econ. Lett. 112, 63–66. ( 10.1016/j.econlet.2011.03.032) [DOI] [Google Scholar]
- 51.Heyes C. 2016. Blackboxing: social learning strategies and cultural evolution. Phil. Trans. R. Soc. B 371, 20150369 ( 10.1098/rstb.2015.0369) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Kurvers RH, Herzog SM, Hertwig R, Krause J, Moussaid M, Argenziano G, Zalaudek I, Carney P, Wolf M. 2019. How to detect high-performing individuals and groups: decision similarity predicts accuracy. Sci. Adv. 5, eaaw9011 ( 10.1126/sciadv.aaw9011) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Mercier H, Morin O. 2019. Majority rules: how good are we at aggregating convergent opinions? Evol. Hum. Sci. 1 ( 10.1017/ehs.2019.6) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Sunstein C. 2003. republic. Princeton, NJ: Princeton University Press. [Google Scholar]
- 55.Smolla M, Akçay E. 2019. Cultural selection shapes network structure. Sci. Adv. 5, eaaw0609 ( 10.1126/sciadv.aaw0609) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Bail CA, et al. 2018. Exposure to opposing views on social media can increase political polarization. Proc. Natl Acad. Sci. USA 115, 9216–9221. ( 10.1073/pnas.1804840115) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Guilbeault D, Becker J, Centola D. 2018. Social learning and partisan bias in the interpretation of climate trends. Proc. Natl Acad. Sci. USA 115, 9714–9719. ( 10.1073/pnas.1722664115) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Votruba AM, Kwan VSY. 2015. Disagreeing on whether agreement is persuasive: perceptions of expert group decisions. PLoS ONE 10, e0121426 ( 10.1371/journal.pone.0121426) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Corriveau K, Harris PL. 2009. Choosing your informant: weighing familiarity and recent accuracy. Dev. Sci. 12, 426–437. ( 10.1111/j.1467-7687.2008.00792.x) [DOI] [PubMed] [Google Scholar]
- 60.Aoki K, Feldman MW. 2014. Evolution of learning strategies in temporally and spatially variable environments: a review of theory. Theor. Popul. Biol. 91, 3–19. ( 10.1016/j.tpb.2013.10.004) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Henrich J, Gil-White FJ. 2001. The evolution of prestige: freely conferred deference as a mechanism for enhancing the benefits of cultural transmission. Evol. Hum. behav. 22, 165–196. ( 10.1016/S1090-5138(00)00071-4) [DOI] [PubMed] [Google Scholar]
- 62.Brand CO, Heap S, Morgan TJH, Mesoudi A. 2020. The emergence and adaptive use of prestige in an online social learning task. Sci. Rep. 10, 12095 ( 10.1038/s41598-020-68982-4) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Henrich J, Broesch J. 2011. On the nature of cultural transmission networks: evidence from Fijian villages for adaptive learning biases. Phil. Trans. R. Soc. B 366, 1139–1148. ( 10.1098/rstb.2010.0323) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Wood LA, Kendal RL, Flynn EG. 2015. Does a peer model's task proficiency influence children's solution choice and innovation? J. Exp. Child Psychol. 139, 190–202. ( 10.1016/j.jecp.2015.06.003) [DOI] [PubMed] [Google Scholar]
- 65.Mesoudi A, Chang L, Dall SR, Thornton A. 2016. The evolution of individual and cultural variation in social learning. Trends Ecol. Evol. 31, 215–225. ( 10.1016/j.tree.2015.12.012) [DOI] [PubMed] [Google Scholar]
- 66.Mesoudi A, Chang L, Murray K, Lu HJ. 2015. Higher frequency of social learning in China than in the West shows cultural variation in the dynamics of cultural evolution. Proc. R. Soc. Lond. B 282, 20142209 ( 10.1098/rspb.2014.2209) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Molleman L, Kanngiesser P, van den Bos W. 2019. Social information use in adolescents: the impact of adults, peers and household composition. PLoS ONE 14, e0225498 ( 10.1371/journal.pone.0225498) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Henrich J, Heine SJ, Norenzayan A. 2010. Beyond WEIRD: towards a broad-based behavioral science. Behav. Brain Sci. 33, 111–135. ( 10.1017/S0140525X10000725) [DOI] [Google Scholar]
- 69.Efferson C, Lalive R, Richerson PJ, McElreath R, Lubell M. 2008. Conformists and mavericks: the empirics of frequency-dependent cultural transmission. Evol. Hum. Behav. 29, 56–64. ( 10.1016/j.evolhumbehav.2007.08.003) [DOI] [Google Scholar]
- 70.Molleman L, van den Berg P, Weissing FJ. 2014. Consistent individual differences in human social learning strategies. Nat. Commun. 5, 3570 ( 10.1038/ncomms4570) [DOI] [PubMed] [Google Scholar]
- 71.Kurvers RHJM, Van Oers K, Nolet BA, Jonker RM, Van Wieren SE, Prins HHT, Ydenberg RC.. 2010. Personality predicts the use of social information. Ecol. Lett. 13, 829–837. ( 10.1111/j.1461-0248.2010.01473.x) [DOI] [PubMed] [Google Scholar]
- 72.Kurvers RH, Prins HH, van Wieren SE, van Oers K, Nolet BA, Ydenberg RC.. 2010. The effect of personality on social foraging: shy barnacle geese scrounge more. Proc. R. Soc. B 277, 601–608. ( 10.1098/rspb.2009.1474) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Watson SK, et al. 2018. Chimpanzees demonstrate individual differences in social information use. Anim. Cogn. 21, 639–650. ( 10.1007/s10071-018-1198-7) [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
All code associated with this paper (i.e. experimental software, cognitive model, simulations and analyses) is publicly available in the public OSF repository https://osf.io/rmcuy/.