Abstract
The extent to which nonhuman primate vocalizations are amenable to modification through experience is relevant for understanding the substrate from which human speech evolved. We examined the vocal behaviour of Guinea baboons, Papio papio, ranging in the Niokolo Koba National Park in Senegal. Guinea baboons live in a multi-level society, with units nested within parties nested within gangs. We investigated whether the acoustic structure of grunts of 27 male baboons of two gangs varied with party/gang membership and genetic relatedness. Males in this species are philopatric, resulting in increased male relatedness within gangs and parties. Grunts of males that were members of the same social levels were more similar than those of males in different social levels (N = 351 dyads for comparison within and between gangs, and N = 169 dyads within and between parties), but the effect sizes were small. Yet, acoustic similarity did not correlate with genetic relatedness, suggesting that higher amounts of social interactions rather than genetic relatedness promote the observed vocal convergence. We consider this convergence a result of sensory–motor integration and suggest this to be an implicit form of vocal learning shared with humans, in contrast to the goal-directed and intentional explicit form of vocal learning unique to human speech acquisition.
Keywords: Papio, primate communication, vocal learning, sensory–motor integration, speech evolution, implicit learning
1. Introduction
One of the key preconditions for the development of speech is the ability to adjust vocal output in response to auditory input. Humans are exceptionally proficient at vocal learning. Although effortless speech learning is confined to the early years [1], humans still possess the ability to imitate sounds voluntarily and acquire further languages throughout their lives. Numerous comparative studies have aimed at elucidating the evolutionary origins of vocal learning within the primate lineage, to uncover the extent to which nonhuman primates reveal evidence for vocal plasticity, and whether such plasticity may be conceived as a pre-adaptation for the evolution of speech [2,3].
Despite considerable research effort, it appears that the ability to learn sounds from auditory experience in most nonhuman primate species is limited. Unlike humans or some songbird species, nonhuman primates are not obligatory vocal learners that require species-specific auditory input to develop their normal vocal repertoires [4,5]. Early attempts to train a young chimpanzee to produce speech sounds yielded disappointing results and prompted most of the ‘ape language' projects to turn to another modality, using either symbol systems or sign languages [6]. Studies of the neural basis of vocal production in different monkey species found that the animals lack the neural connections necessary for the volitional control over the fine structure of vocalizations, although they exert greater control over the usage of calls (reviewed in [2]). One exception to the rule of limited vocal plasticity may be orangutans, which have greater control over their vocal apparatus [7,8].
In addition to the limited ontogenetic plasticity, a range of comparative studies within different nonhuman primate species strongly suggest that the motor patterns underlying vocalizations are evolutionarily highly conserved within genera (reviewed in [2]). For instance, the structure of alarm calls of members of the genus Chlorocebus differs only marginally between East African vervets, Chlorocebus pygerythrus, and West African green monkeys, Chlorocebus sabaeus. Moreover, in response to a drone, naive West African green monkeys spontaneously uttered calls that structurally were highly similar to East African vervet ‘eagle alarm calls', indicating that the link between the perception of a specific (potentially dangerous) stimulus and the activation of a given motor programme is also conserved [9].
At the same time, subtle modifications in vocal output as a result of auditory experience appear to be possible. For instance, common marmosets, Callithrix jacchus, increase the amplitude of their calls in noisy environments (‘Lombard effect'; [10,11]). More importantly, a range of species show group-specific variations or ‘dialects' in their vocalizations (reviewed in [12]), while Japanese macaques matched some of the acoustic features of calls presented in playbacks [13]. These instances of vocal plasticity have been described as ‘vocal accommodation’ [12,14,15] or ‘social shaping' [15], similar to the observation that humans may involuntarily match the pitch, temporal patterning and prosody of the people they are talking to.
Following the idea that auditory input may lead to vocal convergence, subjects that interact more frequently with one another using vocalizations should produce calls that are more similar to each other than those that interact less frequently. A higher acoustic similarity may also result from genetic relatedness, however. For instance, highly related subjects may also have a similar morphology of the vocal production apparatus [16]. Before conclusions about the role of experience can be drawn, it is necessary to assess whether potential acoustic variation between individuals can (also) be explained by genetic distance.
To date, few studies have investigated the effects of genetic relatedness and interaction frequency at the same time. Lemasson and colleagues reported that interaction frequency but not genetic relatedness accounted for acoustic variation in the calls of Campbell's monkeys, Cercopithecus campbelli campbelli [17]. The reported correlation of acoustic similarity with grooming frequency may be spurious, however, as data from two groups (with N = 6 and 4 females, respectively) were pooled, and the correlation was largely driven by the differences between the groups. Levréro and colleagues [15] studied the acoustic structure of contact calls in 36 male and female mandrills living in three social groups. Both genetic relatedness and familiarity impacted acoustic similarity of the species' ‘kakak' calls, while retaining cues to kin memberships: playback experiments showed that subjects responded significantly more strongly to calls recorded from related kin, irrespective of familiarity [15].
We here set out to assess the impact of social interaction while controlling for genetic relatedness by comparing the acoustic variation in the grunts of male Guinea baboons, Papio papio. Guinea baboons are an interesting model to examine the influence of auditory experience and social group membership, as they live in a nested multi-level society with male philopatry [18]. At the base of the society are ‘units' comprising one adult male, one to six females and young. A small number of units, together with bachelor males, form a ‘party', and two or three parties make up a ‘gang' (figure 1a). Assignments to parties and gangs are based on spatial proximity and affiliative interactions [19]. During affiliative interactions with other group members, males produce low-frequency tonal grunts (figure 1b). The Guinea baboons' social structure allowed us to assess vocal convergence at two social levels, namely within parties and within gangs.
If the frequency of interaction affects the structure of calls, subjects that interact frequently with one another should produce calls that are more similar to each other. Thus, members of the same party should have the greatest similarity, while members of the same gang should produce calls that are more similar to each other than to calls produced by members of another gang. If genetic relatedness affects the vocal structure, dyads that are more highly related should reveal greater acoustic similarity. Note that these two effects (interaction frequency and relatedness) are not mutually exclusive.
2. Methods
We obtained recordings of grunts from a total of 27 male baboons in 2010/11, 2014 and 2016. Thirteen of the males were members of the ‘Mare' gang and 14 were members of the ‘Simenti' gang. The Mare gang comprised two parties of 6 and 7 males each; the Simenti gang comprised two parties of 5 and 9 males each. Twenty-three of the 27 males were confirmed or assumed to be present throughout the study period (see electronic supplementary material, table S1).
Vocalizations were recorded using Marantz PMD 661 recorders (D&M Professional, Longford, UK) with Sennheiser directional microphones (K6 power module + ME66 recording head; Sennheiser, Wedemark, Germany) equipped with Rycote windshields (Rycote Windjammer, Rycote Ltd, Stroud, UK). We used Avisoft-SASLab Pro (Avisoft Bioacoustic, Berlin, Germany) to check the recording quality and to label and extract grunts with sufficient quality and low background noise. We only used calls recorded at a maximum distance of 3 m. To maximize the number of grunts per male, we included grunts from different contexts (electronic supplementary material, table S2). In total, we included 756 grunts in the acoustic analysis. On average, we used 28 calls per subject in the analysis (range: 5–127). The Mare and Simenti gang males were represented by 390 and 366 grunts, respectively. Ideally, one would have liked to include further gangs with additional subjects to assess whether the observed pattern holds beyond our study population, but adding further groups was beyond our capacities.
We reduced the sampling frequency from 44.1 to 5 and 1 kHz to obtain an appropriate frequency resolution for the estimation of acoustic features and calculated two 1024-pt fast Fourier transformations (FFTs), one resulting in a frequency range of 2500 Hz (frequency resolution 5 Hz, temporal resolution 6.4 ms) and a second FFT resulting in a frequency range of 500 Hz (frequency resolution of 1 Hz, and a temporal resolution of 16 ms). Calculating two FFTs allowed us to maximize the temporal resolution for the entire call type, and estimate the fundamental frequency at a higher frequency resolution. The resulting frequency–time spectra were analysed with a custom software program LMA 2019, which allows visual control of the accuracy of parameter estimation [20,21]. LMA outputs a total of 82 acoustic parameters.
To identify which parameters would be informative to distinguish between individuals (and thus, social levels), we entered all 82 acoustic features from the LMA output into a stepwise discriminant function (DFA) with subject identity as a grouping variable. The selection criterion for acoustic features to enter the discriminant function analysis was Pin = 0.05 and to be removed Pout = 0.1. The DFA used 31 acoustic features for individual discrimination (electronic supplementary material, table S3). To quantify the acoustic distance between males, we used the average pairwise F-value from the discriminant function analysis as a dissimilarity score for each dyad. The dissimilarity score provides an assessment of the similarity of calls, with higher values indicating greater dissimilarity and lower values greater similarity. In the following, we will simply refer to the similarity of calls. The average pairwise F-value has been used in different studies examining relationships between acoustic structure and genetic or geographic distance [22,23]. The discriminant function analysis was performed using IBM SPSS v. 26.0 (IBM, Armonk, NY). To assess whether the classification result of the individual discrimination of male grunts is higher than would be expected by chance, we additionally performed a permuted DFA [24], which controls for variation in individual contributions of grunts.
We extracted DNA from faecal samples using the First-DNA all tissue kit (Genial®) and characterized genetic variation by assessing the individual allele variation on 24 polymorphic autosomal microsatellite markers. The 24 markers were amplified using the Multiplex PCR Kit (QIAGEN) and fluorescent-labelled primers. PCR products were separated and detected through capillary gel electrophoresis on an ABI 3130xL Genetic Analyzer (Applied Biosystems®, USA). Microsatellite allele sizes were evaluated using GeneMapper 5 (Applied Biosystems®). One locus (D1s548) showed signs of null alleles and significant deviations from Hardy–Weinberg equilibrium and was therefore excluded, resulting in a total of 23 loci included in the relatedness estimation (calculated with MICRO-CHECKER v. 2.2.3 [25] and the PopGenReport R package v. 3.0.0 [26]. We used the R package ‘related' v. 1.0 [27,28] to estimate relatedness using R v. 3.4.4 and RStudio v. 1.1.456. The Wang estimator (hereafter W) appeared to be most suitable for the present analysis (see electronic supplementary material, table S4). W ranges from −1 to 1. Negative values indicate that dyads are less related than on average, while positive values indicate that they are more highly related than on average (see [29] for a detailed description of the analysis).
These and the following statistical analyses were conducted in the R environment v. 3.6.3 [30], using the RStudio interface v. 1.3.959 [31]. We used a Mantel matrix correlation test (package ‘vegan'; v. 2.5.6) to test the correlation between acoustic and genetic variation. To test whether calls within a gang were more similar to each other than between gangs, we applied a categorical Mantel test, using ‘same gang membership' (yes/no) as the categorical predictor variable, and W or F (transformed as ln(1 + F)) as the continuous variable. The analysis of the effect of gang membership was based on 351 dyads. To study the effect of party membership, we also used a categorical Mantel test, but only considered pairs of males that lived in the same gang (e.g. SNE and BAA, both members of the Mare gang, or BEN and WLD, both members of the Simenti gang; total number of dyads within both gangs, N = 169). We used a restricted permutation approach where males were permuted between parties within gangs. We used 1000 permutations in all analyses, except the one for the variation between parties within gangs, where we used 10 000 permutations. Effect sizes were calculated with the package ‘effsize' version 0.8.0. The data and code for statistical analysis are deposited at https://osf.io/h7q5r/.
3. Results
Confirming previous analyses, males were more highly related within gangs than between gangs (categorical Mantel test, p = 0.001, N = 351; figure 2a). The effect size (Cohen's d) was 0.52 (CIlower −0.73, CIupper, −0.31; medium effect size). Within gangs, males in the same party were more highly related on average than males that were not members of the same party (p = 0.035, N = 169; figure 2b), with a small effect size (d = 0.24, CIlower −0.54, CIupper −0.07).
Grunts could be assigned to the correct individual significantly more frequently than by chance, with an average correct assignment of 34.5% using the procedure in SPSS (chance level 3.7%, leave-one-out validation: 21.0% correct classification). The classification in the permuted DFA (pDFA) with a reduced set of N = 135 calls (see electronic supplementary material, table S5) yielded an average classification of 11.2% (p < 0.001). Acoustic similarity did not correlate with genetic similarity (r = −0.006, p = 0.515). Because of the inherent uncertainty with which dyadic relatedness can be estimated [32], we ran an additional analysis in which we compared the acoustic similarity of dyads in the top quartile (W > 0.125) versus the bottom quartile (W < −0.117). Again, we found no effect of relatedness on acoustic similarity (categorical Mantel test, p = 0.933; figure 3).
Grunts of males within gangs were more similar to each other than between gangs (categorical Mantel test, p = 0.012; figure 4a), and grunts of males within a party were also more similar to each other than between parties in the same gang (p = 0.001; figure 4b). The effect sizes were modest, however (d = 0.177, CIlower −0.03, CIupper 0.38 between gangs and 0.152, CIlower −0.15, CIupper 0.46 between parties, respectively). When we compared the mean acoustic similarity of males that resided in the same party (mean logF = 0.29) with those that were part of a different gang (mean logF = 0.33), the effect size was small (d = 0.24, CIlower −0.02, CIupper 0.50). Grunts varied with social level (party/gang) mostly in parameters that are related to the filter function of the vocal tract (electronic supplementary material, table S6). The fundamental frequency or call duration did not vary systematically between social levels.
4. Discussion
The structure of male grunts varied between members of different gangs, and also between members of parties within a gang. The effect sizes of these two comparisons were modest, however. Males in the same gang were also more highly related to one another, but this did not account for the acoustic variation between parties and gangs, as evidenced by the lack of an effect of genetic relatedness on acoustic similarity. In this regard, the Guinea baboons differ from mandrills, where both relatedness and interaction frequency predicted the structure of the vocalizations [15].
It may seem puzzling at first that genetic relatedness did not account for the higher vocal similarity in Guinea baboons despite the fact that genetic relatedness and acoustic similarity were both higher within parties and gangs than between. This can be explained by the fact that not all dyads within a social level are indeed more highly related than across these social levels. Acoustic similarity thus appears mainly to be driven by social interaction, which is not restricted to highly related dyads. To a certain degree, relatedness and acoustic similarity vary independently of one another.
How may auditory input affect vocal production? One mechanism that may support the reported minor adjustments in vocal output with experience may rest on sensory–motor integration [33]. According to the idea of a ‘common coding' framework, specific sensorimotor areas represent both sensory input and motor commands generating that corresponding pattern [34]. In humans, neuroimaging studies identified specific motor activations when subjects listened to speech sounds [35]. If such sensory–motor integration exists in the auditory–vocal domain of nonhuman primates, the exposure to specific auditory input may increase the likelihood to produce the corresponding motor pattern via co-activation.
A recent study provided compelling evidence for the integration of auditory input with vocal output in a nonhuman primate species. In common marmosets, activity in the auditory cortex directly affected the monkeys' control of vocal production [36]. Firstly, a shift in the auditory feedback of the monkeys' vocalization led to compensatory changes in the frequency patterns of the subsequent vocalizations. Secondly, microstimulation of the auditory cortex during vocalization led to abrupt shifts in vocal frequency [36]. In a translocation experiment, common marmosets (N = 4) adjusted the structure of their vocalizations in response to auditory input from conspecifics, even if the individuals did not interact directly [37]. Beyond the immediate effects of auditory experience, there is also evidence that feedback from parents affects the trajectory of vocal development in marmosets [38–40].
It has been argued that the human ability to imitate the utterances of others gradually evolved from the vocal plasticity observed in nonhuman primates [17,41]. We contend that vocal learning may be based on a variety of different mechanisms, including vocal convergence, ‘learning from success', a form of usage learning that comprises the use of specific call variants because they are more likely to yield the desired response, as well as the spontaneous imitation of a recently formed auditory template [42]. Instead of conceiving vocal learning capacities as a continuum [43], we agree with other authors that vocal learning may be supported by a variety of different mechanisms [44]. Future studies should aim to distinguish between these mechanisms, and also consider the effect size of vocal plasticity.
Taking the extent of plasticity as well as the mechanisms that support them into account will contribute to overcoming futile debates about whether or not nonhuman primates reveal evidence for vocal learning [45]. The vast majority of studies in nonhuman primates that reported evidence for vocal convergence observed only minor changes within the species-specific range of calls. Thus, the small effect sizes reported here are important aspects of the results. Humans, instead, are not only able to work on their accents, as Eliza Doolittle in ‘My fair lady', but they can also sing ‘supercalifragilisticexpialidocious' with Mary Poppins. The spontaneous imitation of new words is open-ended, while it is much more difficult to change one's accent once a certain age has been reached. Vocal convergence in nonhuman primates appears to be more similar to the formation of an accent than the acquisition of novel phonetic combinations that make up new words. An interesting open question is whether vocal convergence is simply a by-product of the sensory experience or whether it has been selected for, since it may signal ‘in-group' membership and thus have an important social function [46].
Irrespective of whether vocal convergence has been selected for or not, we propose that it constitutes an implicit form of motor learning shared between nonhuman primates and humans, while speech production constitutes an explicit form of motor learning. Implicit and explicit processes are not entirely dichotomous: explicitly acquired motor skills can become automatic (as when you learn to drive a car), while implicit processes may be made explicit [47]. Yet, it has proven useful to distinguish between implicit and explicit forms of knowledge and knowledge acquisition [48]. Taatgen suggested that implicit learning is a by-product of general learning mechanisms, while explicit learning is tied to learning goals and thus intentionality [47]. This definition appears useful for the distinction between vocal convergence as a result of sensory–motor integration on the one hand and the goal-directed acquisition of the patterns that result in the production of speech, on the other.
A further open question is whether the observed acoustic variation is salient to the animals themselves. In a previous study [49], we tested male responses to the playbacks of grunts of males that share the same home range as the study males (neighbours) versus to grunts of males living 50 km away (strangers). As a control, we played back the grunts of males from their own gang. Surprisingly, males responded strongly only to the grunts from males of their own gang, but largely ignored neighbour or stranger males' calls. In principle, these responses could be explained by the recognition of the males' voice characteristics. Yet, it might also be the case that males recognize the ‘sound’ of their subgroup. Playbacks presenting artificially created grunts bearing the own gang's characteristics versus another gang's characteristics would be needed to test this conjecture.
In summary, we find evidence for a moderate degree of vocal convergence in the grunts of male Guinea baboons. The magnitude of the change is difficult to compare with those of other studies on nonhuman primates mentioned above, given the differences in methodological approaches, but broadly appears to be in a similar range. Our findings add to the body of evidence that within species-specific constraints, subtle and potentially meaningful variation can be found in nonhuman primate vocalizations. This variation does not compare with the open-ended possibility of vocal imitation found in human speech, however.
Supplementary Material
Acknowledgements
We thank the Diréction des Parcs Nationaux (DPN) and Ministère de l'Environnement et de la Protéction de la Nature (MEPN) de la République du Sénégal for permission to work in the Niokolo Koba National Park. Roger Mundry provided invaluable statistical advice. We are grateful to two anonymous reviewers and the editors for very helpful comments on the manuscript.
Ethics
This research complies with the Association for the Study of Animal Behaviour Guidelines for the Use of Animals in Research (Animal Behaviour, 2018, 135, I–X), the legal requirements of the country in which the work was carried out, and all institutional guidelines. The research has been approved by the Diréction des Parcs Nationaux of the Republic of Senegal (22 April 2019).
Data accessibility
The data and code for statistical analysis are deposited at https://osf.io/h7q5r/.
Authors' contributions
J.F., K.H. and F.W. conceived the study, F.W. collected the data, F.W. and K.H. conducted the acoustic analysis, F.D.P. and F.T. conducted the genetic analysis, J.F. and K.H. ran the statistical analyses and prepared the figures, J.F. wrote the manuscript with input from all authors.
Competing interests
We declare we have no competing interests.
Funding
This research was supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) – Project no. 254142454/GRK 2070, the Leibniz ScienceCampus Primate Cognition of the Leibniz Association, and the Christian-Vogel Fonds of the Gesellschaft für Primatologie.
References
- 1.Westermann G, Mani N. 2017. Early word learning. Oxford, UK: Taylor & Francis. [Google Scholar]
- 2.Fischer J, Hage SR. 2019. Primate vocalization as a model for human speech: scopes and limits. In Human language: from genes and brains to behavior (ed. Hagoort P.), pp. 639–656. Cambridge, MA: MIT Press. [Google Scholar]
- 3.Jarvis ED. 2019. Evolution of vocal learning and spoken language. Science 54, 50–54. ( 10.1126/science.aax0287) [DOI] [PubMed] [Google Scholar]
- 4.Nowicki S, Searcy WA. 2014. The evolution of vocal learning. Curr. Opin. Neurobiol. 28, 48–53. ( 10.1016/j.conb.2014.06.007) [DOI] [PubMed] [Google Scholar]
- 5.Tyack PL. 2020. A taxonomy for vocal learning. Phil. Trans. R. Soc. B 375, 20180406 ( 10.1098/rstb.2018.0406) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Wallman J. 1992. Aping language. Cambridge, UK: Cambridge University Press. [Google Scholar]
- 7.Wich SA, Swartz KB, Hardus ME, Lameira AR, Stromberg E, Shumaker RW. 2009. A case of spontaneous acquisition of a human sound by an orangutan. Primates 50, 56–64. ( 10.1007/s10329-008-0117-y) [DOI] [PubMed] [Google Scholar]
- 8.Lameira AR, Hardus ME, Bartlett AM, Shumaker RW, Wich SA, Menken SBJ. 2015. Speech-like rhythm in a voiced and voiceless orangutan call. PLoS ONE 10, e116136 ( 10.1371/journal.pone.0116136) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Wegdell F, Hammerschmidt K, Fischer J. 2019. Conserved alarm calls but rapid auditory learning in monkey responses to novel flying objects. Nature Ecol. Evol. 3, 1039–1042. ( 10.1038/s41559-019-0903-5) [DOI] [PubMed] [Google Scholar]
- 10.Brumm H, Voss K, Köllmer I, Todt D. 2004. Acoustic communication in noise: regulation of call characteristics in a New World monkey. J. Exp. Biol. 207, 443–448. ( 10.1242/jeb.00768) [DOI] [PubMed] [Google Scholar]
- 11.Egnor SER, Hauser MD. 2006. Noise-induced vocal modulation in cotton-top tamarins (Saguinus oedipus). Am. J. Primatol. 68, 1183–1190. ( 10.1002/ajp.20317) [DOI] [PubMed] [Google Scholar]
- 12.Ruch H, Zürcher Y, Burkart JM. 2018. The function and mechanism of vocal accommodation in humans and other primates. Biol. Rev. 93, 996–1013. ( 10.1111/brv.12382) [DOI] [PubMed] [Google Scholar]
- 13.Sugiura H. 1998. Matching of acoustic features during the vocal exchange of coo calls by Japanese macaques. Anim. Behav. 55, 673–687. ( 10.1006/anbe.1997.0602) [DOI] [PubMed] [Google Scholar]
- 14.Fischer J. 2008. Transmission of acquired information in nonhuman primates. In Learning and memory: a comprehensive reference (eds Menzel R, Byrne J), pp. 299–313. Oxford, UK: Elsevier. [Google Scholar]
- 15.Levréro F, Carrete-Vega G, Herbert A, Lawabi I, Courtiol A, Willaume E, Kappeler PM, Charpentier MJE. 2015. Social shaping of voices does not impair phenotype matching of kinship in mandrills. Nat. Commun. 6, 7609 ( 10.1038/ncomms8609) [DOI] [PubMed] [Google Scholar]
- 16.Hauser MD. 1992. Articulatory and social factors influence the acoustic structure of rhesus monkey vocalizations: a learned mode of production? J. Acoust. Soc. Am. 91, 2175–2179. ( 10.1121/1.403676) [DOI] [PubMed] [Google Scholar]
- 17.Lemasson A, Ouattara K, Petit EJ, Zuberbühler K. 2011. Social learning of vocal structure in a nonhuman primate? BMC Evol. Biol. 11, 362 ( 10.1186/1471-2148-11-362) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Fischer J, et al. 2017. Charting the neglected West: the social system of Guinea baboons. Am. J. Phys. Anthropol. 162, 15–31. ( 10.1002/ajpa.23144) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Patzelt A, Kopp GH, Ndao I, Kalbitzer U, Zinner D, Fischer J.. 2014. Male tolerance and male–male bonds in a multilevel primate society. Proc. Natl Acad. Sci. USA 111, 14 740–14 745. ( 10.1073/pnas.1405811111) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Fischer J, Noser R, Hammerschmidt K. 2013. Bioacoustic field research: a primer to acoustic analyses and playback experiments with primates. Am. J. Primatol. 75, 643–663. ( 10.1002/ajp.22153) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Schrader L, Hammerschmidt K. 1997. Computer-aided analysis of acoustic parameters in animal vocalizations: a multi-parametric approach. Bioacoustics 7, 247–265. ( 10.1080/09524622.1997.9753338) [DOI] [Google Scholar]
- 22.Thinh VN, Hallam C, Roos C, Hammerschmidt K. 2011. Concordance between vocal and genetic diversity in crested gibbons. BMC Evol. Biol. 11, 36 ( 10.1186/1471-2148-11-36) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Meyer D, Hodges JK, Rinaldi D, Wijaya A, Roos C, Hammerschmidt K. 2012. Acoustic structure of male loud-calls support molecular phylogeny of Sumatran and Javanese leaf monkeys (genus Presbytis). BMC Evol. Biol. 12, 16 ( 10.1186/1471-2148-12-16) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Mundry R, Sommer C. 2007. Discriminant function analysis with nonindependent data: consequences and an alternative. Anim. Behav. 74, 965–976. ( 10.1016/j.anbehav.2006.12.028) [DOI] [Google Scholar]
- 25.van Oosterhout C, Hutchinson WF, Wills DPM, Shipley P.. 2004. MICRO-CHECKER: software for identifying and correcting genotyping errors in microsatellite data. Mol. Ecol. Notes 4, 535–538. ( 10.1111/j.1471-8286.2004.00684.x) [DOI] [Google Scholar]
- 26.Gruber B, Adamack AT. 2015. landgenreport: A new R function to simplify landscape genetic analysis using resistance surface layers. Mol. Ecol. Resour. 15, 1172–1178. ( 10.1111/1755-0998.12381) [DOI] [PubMed] [Google Scholar]
- 27.Wang J. 2011. Coancestry: a program for simulating, estimating and analysing relatedness and inbreeding coefficients. Mol. Ecol. Resour. 11, 141–145. ( 10.1111/j.1755-0998.2010.02885.x) [DOI] [PubMed] [Google Scholar]
- 28.Pew J, Muir PH, Wang J, Frasier TR. 2015. related: An R package for analysing pairwise relatedness from codominant molecular markers. Mol. Ecol. Resour. 15, 557–561. ( 10.1111/1755-0998.12323) [DOI] [PubMed] [Google Scholar]
- 29.Dal Pesco F, Trede F, Zinner D, Fischer J. . 2020. Analysis of genetic relatedness and paternity assignment in wild Guinea baboons (Papio papio) based on microsatellites. protocols.io ( 10.17504/protocols.io.bfdmji46) [DOI] [PMC free article] [PubMed]
- 30.R Core Team. 2020. R: a language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; See https://www.R-project.org. [Google Scholar]
- 31.RStudio Team. 2020. RStudio: integrated development for R. Boston, MA: RStudio; See https://www.rstudio.com/. [Google Scholar]
- 32.van Horn RC, Altmann J, Alberts SC.. 2008. Can't get there from here: inferring kinship from pairwise genetic relatedness. Anim. Behav. 75, 1173–1180. ( 10.1016/j.anbehav.2007.08.027) [DOI] [Google Scholar]
- 33.Pickering MJ, Garrod S. 2013. An integrated theory of language production and comprehension. Behav. Brain Sci. 36, 329–347. ( 10.1017/S0140525X12001495) [DOI] [PubMed] [Google Scholar]
- 34.Prinz W, Neumann O. 1990. A common-coding approach to perception and action. In Relationships between perception and action: current approaches, (eds O Neumann, W Prinz), pp. 167–201. Berlin, Germany: Springer. [Google Scholar]
- 35.Pulvermüller F, Fadiga L. 2010. Active perception: sensorimotor circuits as a cortical basis for language. Nat. Rev. Neurosci. 11, 351–360. ( 10.1038/nrn2811) [DOI] [PubMed] [Google Scholar]
- 36.Eliades SJ, Tsunada J. 2018. Auditory cortical activity drives feedback-dependent vocal control in marmosets. Nat. Commun. 9, 2540 ( 10.1038/s41467-018-04961-8) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Zürcher Y, Willems EP, Burkart JM. 2019. Are dialects socially learned in marmoset monkeys? Evidence from translocation experiments. PLoS ONE 14, e0222486 ( 10.1371/journal.pone.0222486) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Gultekin YB, Hage SR. 2017. Limiting parental feedback disrupts vocal development in marmoset monkeys. Nat. Commun. 8, 14046 ( 10.1038/ncomms14046) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Gultekin YB, Hage SR. 2018. Limiting parental interaction during vocal development affects acoustic call structure in marmoset monkeys. Sci. Adv. 4, 11 ( 10.1126/sciadv.aar4012) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Takahashi DY, Liao DA, Ghazanfar AA. 2017. Vocal learning via social reinforcement by infant marmoset monkeys. Curr. Biol. 27, 1844– 1852.e6 ( 10.1016/j.cub.2017.05.004) [DOI] [PubMed] [Google Scholar]
- 41.Watson SK, Townsend SW, Schel AM, Wilke C, Wallace EK, Cheng L, West V, Slocombe KE. 2015. Vocal learning in the functionally referential food grunts of chimpanzees. Curr. Biol. 25, 495–499. ( 10.1016/j.cub.2014.12.032) [DOI] [PubMed] [Google Scholar]
- 42.Fischer J, Hammerschmidt K. 2020. Towards a new taxonomy of primate vocal production learning. Phil. Trans. R. Soc. B 375, 20190045 ( 10.1098/rstb.2019.0045) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Arriaga G, Jarvis ED. 2013. Mouse vocal communication system: are ultrasounds learned or innate? Brain Lang. 124, 96–116. ( 10.1016/j.bandl.2012.10.002) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Martins PT, Boeckx C. 2020. Vocal learning: beyond the continuum. PLoS Biol. 18, e3000672 ( 10.1371/journal.pbio.3000672) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Wirthlin M, et al. 2019. A modular approach to vocal learning: disentangling the diversity of a complex behavioral trait. Neuron 104, 87–99. ( 10.1016/j.neuron.2019.09.036) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Crockford C, Herbinger I, Vigilant L, Boesch C. 2004. Wild chimpanzees produce group-specific calls: a case for vocal learning? Ethology 110, 221–243. ( 10.1111/j.1439-0310.2004.00968.x) [DOI] [Google Scholar]
- 47.Taatgen NA. 1999. Learning without limits: from problem solving towards a unified theory of learning. Groningen, The Netherlands: University of Groningen. [Google Scholar]
- 48.Frensch PA, Rünger D. 2003. Implicit learning. Curr. Dir. Psychol. Sci. 12, 13–18. ( 10.1111/1467-8721.01213) [DOI] [Google Scholar]
- 49.Maciej P, Patzelt A, Ndao I, Hammerschmidt K, Fischer J. 2013. Social monitoring in a multilevel society: a playback study with male Guinea baboons. Behav. Ecol. Sociobiol. 67, 61–68. ( 10.1007/s00265-012-1425-1) [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data and code for statistical analysis are deposited at https://osf.io/h7q5r/.