Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2019 Feb 14;9:2062. doi: 10.1038/s41598-019-38514-w

Positive and negative contexts predict duration of pig vocalisations

Mary Friel 1,4, Hansjoerg P Kunc 1, Kym Griffin 2, Lucy Asher 3, Lisa M Collins 4,5,
PMCID: PMC6375976  PMID: 30765788

Abstract

Emotions are mental states occurring in response to external and internal stimuli and thus form an integral part of an animal’s behaviour. Emotions can be mapped in two dimensions based on their arousal and valence. Whilst good indicators of arousal exist, clear indicators of emotional valence, particularly positive valence, are still rare. However, positively valenced emotions may play a crucial role in social interactions in many species and thus, an understanding of how emotional valence is expressed is needed. Vocalisations are a potential indicator of emotional valence as they can reflect the internal state of the caller. We experimentally manipulated valence, using positive and negative cognitive bias trials, to quantify changes in pig vocalisations. We found that grunts were shorter in positive trials than in negative trials. Interestingly, we did not find differences in the other measured acoustic parameters between the positive and negative contexts as reported in previous studies. These differences in results suggest that acoustic parameters may differ in their sensitivity as indicators of emotial valence. However, it is important to understand how similar contexts are, in terms of their valence, to be able to fully understand how and when acoustic parameters reflect emotional states.

Introduction

Emotions are mental states occurring in response to external or internal stimuli which influence an individual’s fitness1. In the dimensional model of emotions, emotional states can be mapped in two-dimensional space, where valence (positive versus negative) and arousal (intensity) characterise each dimension1,2. Emotional states in non-human animals have been demonstrated in a broad range of species across the taxonomic scale (e.g. insects:3,4; birds:5,6; mammals:7,8). However, clear indicators of positively valenced emotion are rare, because identifying consistent correlates on the valence dimension is difficult9,10. Emotions are multi-componential and each component can vary across the two dimensions of arousal and valence1. The four pillars of emotion assessment in humans are physiological, behavioural, cognitive and subjective verbal report11,12. Quantifying emotions in non-human animals relies on indirect assessment of three of these pillars: physiology, behaviour and cognition; the fourth pillar is solely accessible in humans. Cognitive bias testing assesses the effect of emotion on cognition in animals1315. However, this test requires lengthy specialised training of animals to implement, which itself could potentially alter the emotional state of the indivudals being assessed16,17. Vocalisations, however, can be reliably measured and may allow the quantification of emotions in non-human animals as they reflect the inner state of the caller, providing information about an individuals’ affective state10,18,19. Thus, emotions expressed in vocalisations may play an important role in shapping social interactions among indiviuduals.

Morton’s motivational-structural rules state that acoustic characteristics of vocalisations are predictable from the context in which they are produced20. Vocalisations produced in more ‘hostile’ contexts are expected to have lower frequencies and longer durations than vocalisations produced in ‘friendly’ or ‘fearful’ contexts20. However, in addition to between-context variation in acoustic characteristics, variation within vocalisation types may also be context-dependent, and may reflect the internal state of the caller2123. In the domestic pig, the most common vocalisation is the grunt, which functions predominantly as a contact call24,25. Information encoded in grunts includes body size25, individual identity26, personality type27 and emotional state28. Pigs’ vocal responses to mental and physical stressors differ depending on the nature of the stressor29. The cognitive bias paradigm has been applied to assess of the effects of various factors on emotional valence in pigs7,30,31.

The aim of this study was to test whether vocalisations produced in oppositely valenced contexts differ in their acoustic characteristics. We based the choice of acoustic parameters to measure on previous studies (see Table 1) and measured these in grunts produced in two different situations. In the first situation, the valence of the context was experimentally manipulated using cognitive bias positive and negative training trials. In the second situation, response to the probe in ambiguous test trials provided the measure of the valence of the context for the individual. The valence of each context was defined as approaching a stimulus in positive contexts and avoiding a stimulus in negative contexts32. We then analysed the effect of optimistic versus pessimistic responding on the acoustic characteristics of grunts, within each situation, in two separate analyses.

Table 1.

Overview of the acoustic parameters measured and the underlying rationale for each parameter included in the analysis.

Variable Rationale Selected references
Duration Expected to be longer in negative compared to positive contexts, longer duration may facilitate communication of more urgent negative emotions 36,50, for review see10
Mean F0 Increased tension of vocal muscles during negative emotion may cause higher F0 in negative contexts than positive contexts 58
AMextent Less variability due to increased amplitude modulations in negative contexts compared to positive contexts 37
AM rate Found to be higher in calls from negative contexts compared to calls from positive contexts 22
% Time Max Intensity Rarely investigated parameter, increased subglottal pressure during negative emotion may increase percent time intensity is maximum
Q25 Found to be lower in grunts produced in negative context than in a positive context 28
Q50 Found to be lower in grunts produced in negative context than in a positive context 28
Q75 Increased tension of vocal muscles during negative emotion may cause higher Q75 37
HNR Can convey information about emotional arousal, decreased in grunts during higher arousal 42
F1mean Found to be higher in grunts produced in negative contexts compared to positive contexts 22
F1 range Found to be higher in calls produced in negative contexts compared to positive contexts 22

Parameter descriptions from Briefer (2012): Duration = duration of the call; F0 mean = mean fundamental frequency across the call; AM extent = mean peak-to-peak variation of each amplitude modulation; AM rate = Number of complete cycles of amplitude modulation per second; % Time Max Intensity = Percentage of total call duration when intensity is maximum; Q25 = frequency value at the upper limit of the first quartile of energy; Q50 = frequency value at the upper limit of the second quartile of energy; Q75 = frequency at the upper limit of the third quartile of energy; Harmonic-to-noise ratio (HNR) = ratio of amplitude peaks of detectable harmonics to noise threshold; F1 mean = mean frequency value of the first formant; F1 range = difference between the minimum and maximum F1 frequencies.

Methods

Animals and housing

We studied crossbred PIC 337, Large White x Landrace pigs, allocated into groups of 18 per pen, balanced for weight and sex, where they remained from 4–10 weeks of age at a research farm (Agri-Food and Biosciences Institute, Hillsborough, Northern Ireland). All test subjects were 10 weeks old at the time of cognitive bias testing. The experiment consisted of three replicates. Within each replicate there were two types of housing environment treatment; a barren environment and an enriched environment (cf.27,33 and see electronic supplementary material for full details). The barren environment had a partially slatted concrete floor and the minimum legal space allowance of 0.41 m2 per pig, whereas the enriched environment had a solid floor with straw bedding and a greater space allowance of 0.62 m2 per pig.

Ethical note

Ethical approval for this research was granted from University of Lincoln’s College of Science Ethics Committee (COSREC62, 8/9/15). All procedures conformed to the ASAB/ABS Guidelines for the Use of Animals in Research.

Cognitive bias testing

Twenty-seven pigs were tested in a spatial cognitive bias task (see Supplementary Methods for full details of training). On the cognitive bias test day, each pig was exposed once individually to each of the ambiguous probe locations: near positive (NP), middle (M) and near negative (NN), with pseudo-randomization between two positive (P) and negative (N) training trials resulting in 9 trials per test (i.e. P, N, M, N, P, NN, N, P, NP). Ambiguous locations were unrewarded but in positive and negative training trials the locations were baited as in training. Latency to reach the bowl was recorded in all trials. Individuals were given 30 seconds to approach the bowl and if they did not approach in that time they were recorded as having a latency of 30 sec and returned to the start box for the next trial.

The raw latency to reach the probe data violated the assumptions of the linear models, therefore we classified responses to the probe within each as “optimistic”, “pessimistic” or “neutral” (cf.31). Pigs’ responses in each trial were classified in relation to the mean speed of approach to the trained positive and negative targets from the final training session using the following formula:

adjustedscore=x[(y+z)/2]×100

where x = latency to contact the bowl during the cognitive bias test, y = mean latency to contact the bowl during the positive training trials, z = mean latency to contact the bowl during the negative training trials. As per Carreras et al. (2016), adjusted scores were calculated as percentages; if the adjusted score was ≤75% the response was classified as “optimistic”; if the score was >75 and <125% the response was classified as “neutral”; and if the score was ≥125% the response was classified as “pessimistic”.

Acoustic recording and analysis

Vocalisations produced during the cognitive bias test were recorded at a distance of 1–4 meters using a Sennheiser ME66/K6 (Sennheiser Electronic GmbH & Co. KG, Wedemark, Germany) directional microphone connected to a Marantz PMD660 (D&M Professional, Kanagawa, Japan) solid state audio recorder (.wav format, sampling frequency: 44.1 kHz, resolution: 16 bit). The microphone was placed on one corner of the arena at a height of 1.5 m and pointed into the test arena (see Fig. S1 for microphone location above test arena). The recorder was switched on at the start of the test session for each individual and all trials were recorded.

All good quality grunts with a high signal-to-noise ratio from the positive, negative and ambiguous trials were selected for analysis (see Fig. 1 for spectrograms). Grunts were chosen for analysis on the condition that they were at least three calls apart, if produced in a bout, to avoid pseudoreplication. A total of 125 calls were produced when a subject responded optimistically; 153 calls were produced when a subject responded pessimistically; 4 calls were produced when the subject’s response was neutral and these 4 calls were excluded from subsequent analyses. A further 14 calls that had been produced in positive and negative trials where the subject had responded ‘incorrectly’ (i.e. responded pessimistically in the positive trial and optimistically in the negative trial) were also excluded. This resulted in a total of 264 calls from 23 individuals for analysis (78 calls from ambiguous test trials and 186 calls from the positive and negative training trials: see Table S1 for full details of number of calls contributed per individual). Using a custom built script in PRAAT34, we measured 11 acoustic parameters from each call (see Table 1 for list of parameters measured and supplement for further details of acoustic analysis). Some parameters (amplitude modulations and rate) could not be measured in every call and this resulted in some missing values, hence the sample size differs between acoustic parameters (Table 2).

Figure 1.

Figure 1

Grunt vocalisations of domestic pig. Grunt vocalisations from individual 71: (a) is from a positive trial with duration = 0.372 seconds and (b) is from a negative trial with duration = 0.411 seconds. F0 = fundamental frequency and F1 = the first formant (calls are available as audio files “Positive context grunt” and “Negative context grunt” respectively in the supplementary material). Visualisation settings: view range = 0–8000 Hz, window length = 0.03 sec, dynamic range = 65 dB, time steps = 700, frequency steps = 250, Gaussian window.

Table 2.

Raw means and SDs, along with results from models for the effect of training trial type, i.e. Negative or Positive, controlled for sex, environment, and individual identity along with statistical results, sample size (N) and P values.

Row No. Parameter 1. Trained positive/negative cue trials χ12 (N) P 2. Ambiguous cue trials χ12 (N) P
Negative Positive Pessimistic Optimistic
Mean SD Mean SD Mean SD Mean SD
1 Duration (s) 0.43 0.15 0.35 0.15 9.86 (186) 0.0017 0.42 0.21 0.34 0.12 3.16 (78) 0.076
2 F0 mean (Hz) 49.3 6.47 49.7 6.43 0.09 (186) 0.760 49.36 6.12 50.28 7.04 1.93 (78) 0.164
3 AM extent (dB) 4.30 3.23 3.61 2.32 1.99 (147) 0.158 3.61 2.60 4.16 3.48 0.05 (56) 0.832
4 AM rate (s-1) 3.26 1.56 3.24 1.55 0.05 (147) 0.831 3.41 1.68 2.77 1.34 2.03 (56) 0.154
5 % Time Max Intensity 46.82 16.13 50.26 15.38 1.78 (174) 0.182 44.22 13.87 51.25 14.92 3.46 (66) 0.063
6 HNR (dB) 1.79 1.78 1.79 1.57 0.37 (186) 0.545 1.82 1.47 1.77 1.66 0.41 (78) 0.520
7 Q25 (Hz) 156.34 42.05 164.44 47.05 2.75 (186) 0.097 178.76 53.42 160.49 43.46 1.06 (78) 0.303
8 Q50 (Hz) 287.82 72.49 301.4 87.57 2.17 (186) 0.140 323.11 98.60 305.84 99.08 0.01 (78) 0.927
9 Q75 (Hz) 668.79 372.9 653.38 376.8 0.41 (186) 0.524 794.21 547.94 720.97 454.32 0.03 (78) 0.858
10 F1 mean (Hz) 321.62 19.86 325.19 22.26 1.45 (185) 0.229 325.21 20.52 322.49 20.59 0.03 (78) 0.865
11 F1 range (Hz) 64.68 28.21 57.5 26.88 2.84 (185) 0.092 53.51 24.54 54.36 19.37 0.15 (78) 0.701

Statistical analysis

Data analysis was conducted using R v. 3.0.2 (R Development Core Team 2008). The statistical analysis of the acoustic parameters was split into two analyses: (1) the trained positive/negative trials, and; (2) the ambiguous trials. This was because these two trial types, trained versus ambiguous, were fundamentally different. In the trained trials the bowls were baited with either a positive or negative stimulus and thus, were expected to elicit either a positive or negative valence state during that trial. The ambiguous trials, on the other hand, were not baited and in addition, in each of the 3 ambiguous trials, this was the first time the individual had encountred the probe in that location.

Linear mixed models were used to investigate the effect of response type (optimistic or pessimistic) on the acoustic parameters in the positive/negative trials and the ambiguous trials separately, accounting for experimental design (see electronic supplementary material for full details of models). Model residuals were visually inspected for normality and homoscedasticity. Data are presented as mean ± SD.

Results

Positive and negative training trials

The duration of calls produced in the negative training trials was significantly longer than the duration of calls produced during the positive training trials (raw mean ± SD: negative = 0.43 ± 0.15, positive = 0.35 ± 0.15, p = 0.0017). There was no significant difference between the positive and negative training trials in the other acoustic parameters investigated (see Table 2). To correct for multiple testing, we applied a Bonferroni correction; this led to a corrected value of p < 0.0046 (i.e. 0.05/11), which did not affect the interpretation of any of the results.

Ambiguous trials

There were no significant effect of response type (optimistic or pessimistic) on the acoustic parameters of the grunts from the ambiguous trials (see Table 2).

Discussion

We found that emotional valence is encoded in the duration of calls. Grunts produced during contexts of positive valence were shorter than those produced during contexts of negative valence. Call duration is affected by respiration, and changes in the action or tension of the respiratory muscles may explain the shorter duration of vocalisations in positive compared to negative contexts10. Longer calls produced in response to negative stimuli may function to increase the salience of these, presumably more urgent, calls to conspecifics. Previous studies in pigs have also found shorter call duration in positive contexts compared to negative contexts28,35, with similar findings in horses23, elephants21 and dogs36. Thus, call duration appears to provide a consistent indicator of emotional valence in mammalian vocalisations10.

Interestingly, we did not find significant differences in the other measured acoustic parameters between positive and negative contexts, which differs from recent studies22,37,38. There are several non-mutually exclusive explantions why the outcomes differ among studies: firstly, the emotional valence experienced by indiviudals may not have differed enough to be reflected in the measured acoustic parameters. The only difference between the two contexts was the position of the bowl and whether a reward was received or not, and the trials were carried out in quick succession. The dimensional model views valence as a continuum. Other studies may have induced emotions further apart on the continuum, whilst our experimental contexts may have induced emotions that were closer to each other on the continuum. Thus, an acoustic parameter would need to be very sensitive to subtle differences in valence to be used as an indicator of valences that are closer together on the continuum. Secondly, the experimental design of previous studies differed from ours in several other aspects, including day of recording, number of individuals present during recording and the functional relevance of the context e.g. social isolation, startling and aggression as negative contexts and food anticipation and affiliative interactions as positive contexts22,28,37,38. Thirdly, it may be that the calls produced here were influenced more by unmeasured factors than the valence of the contexts. Leliveld and colleagues found that the context accounted for only 1.5–11.9% of the variability in the acoustic parameters they measured28, suggesting that many other factors influence the acoustic characteristics of vocalisations.

Vocalisations may be more associated with the expression of arousal than valence39,40. Indeed, one of the main challenges in assessing emotional valence is eliciting oppositely valenced states whilst not affecting arousal, since negatively valenced states also tend to be higher arousal states37. For assessing the characteristics of vocalisations in such conditions, this is problematic, as high arousal is known to impact the acoustic characteristics of vocalisations41. Fundamental frequency parameters and energy quartiles are robust measures of emotional arousal in mammals10,42,43. We did not detect any statistically significant differences in these parameters between the positively and negatively valenced contexts (see Table 2). This suggests that level of arousal did not differ between the positive and negative contexts used here.

In the ambiguous trials, individuals were presented with an ambiguous cue, i.e. the bowl located in the near positive, middle and near negative locations. In contrast to the positive and negative trained trials, we did not find any significant associations between optimistic/pessimistic response type and any of the acoustic parameters measured during the ambiguous trials. However, there was a trend for calls to be shorter when produced during an optimistic response than a pessimistic response. As the training trials were rewarded/unrewarded, we predicted they would induce emotional states of positive and negative valence respectively44. In contrast, responses in ambiguous trials provide a measure of the animals’ underlying emotional state14, rather than inducing any valenced emotional state per se. Thus, optimistic and pessimistic responding in the ambiguous trials may not have been directly comparable to optimistic and pessimistic responding in the training trials.

Emotions are typically measured through behaviour, for example through body posture45, facial expression46,47, and vocalisations48. For social species, the ability to recognise the emotional state of conspecifics is likely to be adaptive as it may enable them to predict the future behaviour of the individual and adjust their behaviour accordingly4951. Individuals of different species are able to perceive, and respond to, differences in emotional state based on differences in the acoustic structure of the same type of vocalisations produced in different contexts5154. Changes in the characteristics of vocalisations could potentially function as a signal of emotional state to conspecifics, however further research is required to fully understand the effect of valence on acoustic parameters in animal vocalisations. Recent studies on emotional contagion suggest that cues relating to emotional state can elicit similar changes in behaviour in individuals who have no direct experience of the original cue5557. Thus, changes in vocalisations corresponding with emotions could function to signal emotional state to conspecifics and to regulate social interactions.

In conclusion, duration of vocalisation appears to be a consistent indicator of valence across species10, and our results suggest that it may also be a sensitive indicator of small differences in valence. This is encouraging, as call duration has the advantage of being easy to measure and apply in a wide variety of contexts. To assess whether valence affects the acoustic parameters of calls, it is necessary to understand how similar or dissimilar oppositely valenced states are to each other. This could help uncover which acoustic parameters are sensitive to small changes in valence, and which others may only be affected by large differences in valence.

Supplementary information

Supplementary methods (209.9KB, pdf)

Acknowledgements

We thank Luigi Baciadonna and Elodie Briefer for their invaluable help and advice on the acoustic analysis and we acknowledge David Reby who wrote the PRAAT script used in the acoustic analysis. We would also like to thank the staff at AFBI Hillsborough for their care of the pigs and use of the experimental rooms and Grace Carroll for her help with data collection. We thank the three anonymous reviewers for their helpful comments on the manuscript. The work presented here was funded as part of a BBSRC Grant No. (BB/K002554/2) and M. Friel was funded by DEL-NI and Queen’s University Belfast.

Author Contributions

Conceived and designed the experiment (M.F., L.M.C., H.K., L.A.). Performed the experiment (M.F., K.G.). Statistical analyses (M.F.). Contributed to reagents, materials and analytical tools (L.M.C., H.K.). Contributed to the writing of the manuscript (M.F., K.G., H.K., L.A. and L.M.C.).

Data Availability

The datasets generated and analysed during this study are available from the corresponding author on request.

Competing Interests

The authors declare no competing interests.

Footnotes

Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Supplementary information accompanies this paper at 10.1038/s41598-019-38514-w.

References

  • 1.Mendl M, Burman OHP, Paul ES. An integrative and functional framework for the study of animal emotion and mood. Proc. R. Soc. B Biol. Sci. 2010;277:2895–2904. doi: 10.1098/rspb.2010.0303. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Russell JA. Core Affect and the Psychological Construction of Emotion. Psychol. Rev. 2003;110:145–172. doi: 10.1037/0033-295x.110.1.145. [DOI] [PubMed] [Google Scholar]
  • 3.Bateson M, Desire S, Gartside SE, Wright GA. Agitated honeybees exhibit pessimistic cognitive biases. Curr. Biol. 2011;21:1070–1073. doi: 10.1016/j.cub.2011.05.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Perry CJ, Baciadonna L, Chittka L. Unexpected rewards induce dopamine-dependent positive emotion-like state changes in bumblebees. Science. 2016;353:1529–1531. doi: 10.1126/science.aaf4454. [DOI] [PubMed] [Google Scholar]
  • 5.Wichman A, Keeling LJ, Forkman B. Cognitive bias and anticipatory behaviour of laying hens housed in basic and enriched pens. Appl. Anim. Behav. Sci. 2012;140:62–69. [Google Scholar]
  • 6.Cussen VA, Mench JA. Personality predicts cognitive bias in captive psittacines, Amazona amazonica. Anim. Behav. 2014;89:123–130. [Google Scholar]
  • 7.Douglas C, Bateson M, Walsh C, Bédué A, Edwards SA. Environmental enrichment induces optimistic cognitive biases in pigs. Appl. Anim. Behav. Sci. 2012;139:65–73. [Google Scholar]
  • 8.Schino G, Massimei R, Pinzaglia M, Addessi E. Grooming, social rank and ‘optimism’ in tufted capuchin monkeys: a study of judgement bias. Anim. Behav. 2016;119:11–16. [Google Scholar]
  • 9.Belin P, et al. Human cerebral response to animal affective vocalizations. Proc. R. Soc. B Biol. Sci. 2008;275:473–481. doi: 10.1098/rspb.2007.1460. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Briefer EF. Vocal expression of emotions in mammals: Mechanisms of production and evidence. J. Zool. 2012;288:1–20. [Google Scholar]
  • 11.Scherer KR, Grandjean D. Facial expressions allow inference of both emotions and their components. Cogn. Emot. 2008;22:789–801. [Google Scholar]
  • 12.Shuman V, Clark-Polner E, Meuleman B, Sander D, Scherer KR. Emotion perception from a componential perspective. Cogn. Emot. 2017;31:47–56. doi: 10.1080/02699931.2015.1075964. [DOI] [PubMed] [Google Scholar]
  • 13.Baciadonna L, McElligott AG. The use of judgement bias to assess welfare in farm livestock. Anim. Welf. 2015;24:81–91. [Google Scholar]
  • 14.Harding EJ, Paul ES, Mendl M. Cognitive bias and affective state. Nature. 2004;427:312. doi: 10.1038/427312a. [DOI] [PubMed] [Google Scholar]
  • 15.Paul ES, Harding EJ, Mendl M. Measuring emotional processes in animals: The utility of a cognitive approach. Neurosci. Biobehav. Rev. 2005;29:469–491. doi: 10.1016/j.neubiorev.2005.01.002. [DOI] [PubMed] [Google Scholar]
  • 16.Roelofs S, Boleij H, Nordquist RE, van der Staay FJ. Making Decisions under Ambiguity: Judgment Bias Tasks for Assessing Emotional State in Animals. Front. Behav. Neurosci. 2016;10:1–16. doi: 10.3389/fnbeh.2016.00119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Novak J, Bailoo JD, Melotti L, Rommen J, Würbel H. An exploration based cognitive bias test for mice: Effects of handling method and stereotypic behaviour. PLoS One. 2015;10:1–16. doi: 10.1371/journal.pone.0130718. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Gogoleva SS, Volodin IA, Volodina EV, Kharlamova AV, Trut LN. Sign and strength of emotional arousal: Vocal correlates of positive and negative attitudes to humans in silver foxes (Vulpes vulpes) Behaviour. 2010;147:1713–1736. [Google Scholar]
  • 19.Manteuffel G, Puppe B, Schön PC. Vocalization of farm animals as a measure of welfare. Appl. Anim. Behav. Sci. 2004;88:163–182. [Google Scholar]
  • 20.Morton ES. On the occurrence and significance of motivational-structural rules in some bird and mamal sounds. Am. Nat. 1977;111:855–869. [Google Scholar]
  • 21.Soltis J, Blowers TE, Savage A. Measuring positive and negative affect in the voiced sounds of African elephants (Loxodonta africana) J. Acoust. Soc. Am. 2011;129:1059–1066. doi: 10.1121/1.3531798. [DOI] [PubMed] [Google Scholar]
  • 22.Maigrot A-L, Hillmann E, Briefer E. Encoding of Emotional Valence in Wild Boar (Sus scrofa) Calls. Animals. 2018;8:85. doi: 10.3390/ani8060085. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Briefer EF, et al. Segregation of information about emotional arousal and valence in horse whinnies. Sci. Rep. 2015;4:1–11. doi: 10.1038/srep09989. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Kiley M. The vocalizations of ungulates, their causation and function. Z. Tierpsychol. 1972;31:171–222. doi: 10.1111/j.1439-0310.1972.tb01764.x. [DOI] [PubMed] [Google Scholar]
  • 25.Garcia M, Wondrak M, Huber L, Fitch WT. Honest signaling in domestic piglets (Sus scrofa domesticus): vocal allometry and the information content of grunt calls. J. Exp. Biol. 2016;219:1913–1921. doi: 10.1242/jeb.138255. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Blackshaw JK, Jones DN, Thomas FJ. Vocal individuality during suckling in the intensively housed domestic pig. Appl. Anim. Behav. Sci. 1996;50:33–41. [Google Scholar]
  • 27.Friel M, Kunc HP, Griffin K, Asher L, Collins LM. Acoustic signalling reflects personality in a social mammal. R. Soc. Open Sci. 2016;3:1–9. doi: 10.1098/rsos.160178. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Leliveld LMC, Düpjan S, Tuchscherer A, Puppe B. Behavioural and physiological measures indicate subtle variations in the emotional valence of young pigs. Physiol. Behav. 2016;157:116–124. doi: 10.1016/j.physbeh.2016.02.002. [DOI] [PubMed] [Google Scholar]
  • 29.Düpjan S, Schön PC, Puppe B, Tuchscherer A, Manteuffel G. Differential vocal responses to physical and mental stressors in domestic pigs (Sus scrofa) Appl. Anim. Behav. Sci. 2008;114:105–115. [Google Scholar]
  • 30.Brajon S, Laforest JP, Schmitt O, Devillers N. The way humans behave modulates the emotional state of piglets. PLoS One. 2015;10:1–17. doi: 10.1371/journal.pone.0133408. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Carreras R, et al. Effect of gender and halothane genotype on cognitive bias and its relationship with fear in pigs. Appl. Anim. Behav. Sci. 2016;177:12–18. [Google Scholar]
  • 32.Bateson M. Optimistic and pessimistic biases: a primer for behavioural ecologists. Curr. Opin. Behav. Sci. 2016;12:115–121. [Google Scholar]
  • 33.Asher L, Friel M, Griffin K, Collins LM. Mood and personality interact to determine cognitive biases in pigs. Biol. Lett. 2016;12:20160402. doi: 10.1098/rsbl.2016.0402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Reby D, McComb K. Anatomical constraints generate honesty: Acoustic cues to age and weight in the roars of red deer stags. Anim. Behav. 2003;65:519–530. [Google Scholar]
  • 35.Tallet, C. et al. Encoding of Situations in the Vocal Repertoire of Piglets (Sus scrofa): A Comparison of Discrete and Graded Classifications. PLoS One8, (2013). [DOI] [PMC free article] [PubMed]
  • 36.Taylor AM, Reby D, McComb K. Context-related variation in the vocal growling behaviour of the domestic dog (canis familiaris) Ethology. 2009;115:905–915. [Google Scholar]
  • 37.Briefer EF, Tettamanti F, McElligott AG. Emotions in goats: Mapping physiological, behavioural and vocal profiles. Anim. Behav. 2015;99:131–143. [Google Scholar]
  • 38.Maigrot AL, Hillmann E, Anne C, Briefer EF. Vocal expression of emotional valence in Przewalski’s horses (Equus przewalskii) Sci. Rep. 2017;7:1–11. doi: 10.1038/s41598-017-09437-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Bachorowski J-A, Owren MJ. The sounds of emotion: Production and perception of affect-related vocal acoustics. Ann. N. Y. Acad. Sci. 2003;1000:244–265. doi: 10.1196/annals.1280.012. [DOI] [PubMed] [Google Scholar]
  • 40.Laukka P, Juslin PN, Bresin R. A dimensional approach to vocal expression of emotion. Cogn. Emot. 2005;19:633–653. [Google Scholar]
  • 41.Goudbeek M, Scherer K. Beyond arousal: Valence and potency/control cues in the vocal expression of emotion. J. Acoust. Soc. Am. 2010;128:1322. doi: 10.1121/1.3466853. [DOI] [PubMed] [Google Scholar]
  • 42.Linhart, P., Ratcliffe, V. F., Reby, D. & Špinka, M. Expression of emotional arousal in two different piglet call types. PLoS One10, (2015). [DOI] [PMC free article] [PubMed]
  • 43.Briefer, E. F. et al. Horse vocalisations supplementary material. Sci. Rep. 4, (2015).
  • 44.Rolls ET. Emotion and decision-making explained: A précis. Cortex. 2014;59:185–193. doi: 10.1016/j.cortex.2014.01.020. [DOI] [PubMed] [Google Scholar]
  • 45.Wemelsfelder F, Haskell M, Mendl MT, Calvert S, Lawrence AB. Diversity of behaviour during novel object tests is reduced in pigs housed in substrate-impoverished conditions. Anim. Behav. 2000;60:385–394. doi: 10.1006/anbe.2000.1466. [DOI] [PubMed] [Google Scholar]
  • 46.McLennan KM, et al. Development of a facial expression scale using footrot and mastitis as models of pain in sheep. Appl. Anim. Behav. Sci. 2016;176:19–26. [Google Scholar]
  • 47.Finlayson K, Lampe JF, Hintze S, Würbel H, Melotti L. Facial indicators of positive emotions in rats. PLoS One. 2016;11:1–24. doi: 10.1371/journal.pone.0166446. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Burgdorf J, Panksepp J. The neurobiology of positive emotions. Neurosci. Biobehav. Rev. 2006;30:173–187. doi: 10.1016/j.neubiorev.2005.06.001. [DOI] [PubMed] [Google Scholar]
  • 49.Müller CA, Schmitt K, Barber ALA, Huber L. Dogs can discriminate emotional expressions of human faces. Curr. Biol. 2015;25:601–605. doi: 10.1016/j.cub.2014.12.055. [DOI] [PubMed] [Google Scholar]
  • 50.Brudzynski SM. Ethotransmission: Communication of emotional states through ultrasonic vocalization in rats. Curr. Opin. Neurobiol. 2013;23:310–317. doi: 10.1016/j.conb.2013.01.014. [DOI] [PubMed] [Google Scholar]
  • 51.Filippi P, Gogoleva SS, Volodina EV, Volodin IA, de Boer B. Humans identify negative (but not positive) arousal in silver fox vocalizations: Implications for the adaptive value of interspecific eavesdropping. Curr. Zool. 2017;63:445–456. doi: 10.1093/cz/zox035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Briefer EF, et al. Perception of emotional valence in horse whinnies. Front. Zool. 2017;14:1–12. doi: 10.1186/s12983-017-0193-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Faragó T, Pongrácz P, Range F, Virányi Z, Miklósi Á. ‘The bone is mine’: affective and referential aspects of dog growls. Anim. Behav. 2010;79:917–925. [Google Scholar]
  • 54.Filippi P, et al. Humans recognize emotional arousal in vocalizations across all classes of terrestrial vertebrates: Evidence for acoustic universals. Proc. R. Soc. B Biol. Sci. 2017;284:1–9. doi: 10.1098/rspb.2017.0990. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Reimert I, Fong S, Rodenburg TB, Bolhuis JE. Emotional states and emotional contagion in pigs after exposure to a positive and negative treatment. Appl. Anim. Behav. Sci. 2017;193:37–42. [Google Scholar]
  • 56.Quervel-Chaumette M, Faerber V, Faragó T, Marshall-Pescini S, Range F. Investigating empathy-like responding to conspecifics’ distress in pet dogs. PLoS One. 2016;11:1–15. doi: 10.1371/journal.pone.0152920. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Edgar JL, Lowe JC, Paul ES, Nicol CJ. Avian maternal response to chick distress. Proc. R. Soc. B Biol. Sci. 2011;278:3129–3134. doi: 10.1098/rspb.2010.2701. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Maruščáková IL, et al. Humans (Homo sapiens) judge the emotional content of piglet (Sus scrofa domestica) calls based on simple acoustic parameters, not personality, empathy, nor attitude toward animals. J. Comp. Psychol. 2015;129:121–131. doi: 10.1037/a0038870. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary methods (209.9KB, pdf)

Data Availability Statement

The datasets generated and analysed during this study are available from the corresponding author on request.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES