Skip to main content
Current Zoology logoLink to Current Zoology
. 2016 Aug 22;63(1):21–26. doi: 10.1093/cz/zow091

What artifice can and cannot tell us about animal behavior

Daniel L Powell a,b,*, Gil G Rosenthal a,b
PMCID: PMC5804151  PMID: 29491959

Abstract

Artifice—the manipulation of social and environmental stimuli—is fundamental to research in animal behavior. State-of-the-art techniques have been developed to generate and present complex visual stimuli. These techniques have unique strengths and limitations. However, many of the issues with synthetic animation and virtual reality are common to playback experiments in general, including those using unmanipulated video or auditory stimuli. Playback experiments, in turn, fall into the broader category of experiments that artificially manipulate the array of stimuli experienced by a subject. We argue that the challenges of designing and interpreting experiments using virtual reality or synthetic animations are largely comparable to those of studies using older technologies or addressing other modalities, and that technology alone is unlikely to solve these challenges. We suggest that appropriate experimental designs are the key to validating behavioral responses to artificial stimuli and to interpreting all studies using artifice, including those that present complex visual displays.

Keywords: animation, animal communication, artifice, experimental design, mate choice, playback.

Overview

Contemporary techniques for presenting complex visual stimuli (Chouinard-Thuly et al. 2017) are the vanguard of an experimental tradition that goes back at least to ter Pelkwijk and Tinbergen’s (1937) use of wax dummies to study aggression in sticklebacks. Artifice, defined here as the controlled manipulation of social or environmental stimuli, is the heart of experimental animal behavior. Artifice encompasses not only artificial social partners, but also artificially contrived social situations and environments, like pairwise choice tests for potential live partners isolated by physical barriers. In this issue, Chouinard-Thuly et al. (2017) detail the benefits and challenges of producing synthetic and interactive stimuli. The goal of this paper is to place these into the broader context of benefits and challenges that face experiments in animal behavior. The animations and virtual reality discussed elsewhere in this special column are kinds of stimuli used in playback experiments. Playback experiments, in turn, are a subset of techniques that involve artifacts. Many of the key issues facing animations and virtual reality are ubiquitous in the experimental analysis of behavior.

Playback

Animations and virtual reality environments are playback stimuli. Playback, the broadcasting of recorded or synthetic stimuli to elicit receiver responses, is a powerful tool with three main advantages over using live signalers. First, playback presentation provides control over stimuli, allowing experimenters to quantitatively manipulate and systematically vary specific aspects of signal variation while holding others constant (McGregor 1992). Second, playback allows one to create stimuli that are unavailable in nature, notably through de novo digital synthesis (Chouinard-Thuly et al. 2017). This is particularly useful when we want to test receiver responses to ancestral or novel signals (e.g., Phelps et al. 2001) or when we want to characterize how receiver responses align with multivariate signal variation (e.g., Fisher et al. 2009).

The third main advantage is consistency. Live signalers will inevitably vary in motivation, and their signaling behavior will inevitably interact with that of the subject. Playback controls these factors, allowing the experimenter to present and manipulate cues independent of identifiable confounding variables associated with signaler morphology and behavior, and has the potential to increase subject response repeatability (McGregor 1992; Kodric-Brown and Nicoletto 1997). Further, the repeatability of playback stimuli sequences allows us to track changes in behavioral responses over ontogeny, as the same subjects can be tested for response to the exact same stimuli at different life stages (Seyfarth and Cheney 1986; Mateo and Holmes 1997; Rosenthal 1999).

For synthetic animations and virtual reality to reap these advantages, it might be useful to consider a long-established technique in animal behavior, playback of acoustic stimuli (e.g., Marler and Tamura 1962; Bell 1974; Rand and Ryan 1981). Acoustic playbacks can be used to elicit context-appropriate responses, like aggressive displays or proceptive behaviors, in all major groups of animals known to signal in the auditory modality (King 2015; Rosenthal forthcoming 2016). Acoustic playback has always been an aspirational model for what can and cannot be accomplished in the visual modality, and video playback in vertebrates was first developed by birdsong researchers (Evans and Marler 1991). There are 2 main advantages of acoustic over video playback: they are easier to manipulate, and they retain more faithful representations of the natural stimuli they mimic due to the lower dimensionality of acoustic signals (Rosenthal and Ryan 2000).

Ease of Manipulation

In its simplest form, acoustic playback consists of calls or songs recorded in the field or laboratory and played back through commercially available, high dynamic range speakers in order to assay responses from a receiver. This is analogous to recording video of an animal performing a behavior, editing it, and playing it back on a monitor. What distinguishes acoustic playback, however, is the relative ease with which signals can be analyzed, manipulated, and synthesized. If a sound is assumed to originate from a point source, we can describe it as a function of frequency and intensity over time. Recorded signals can be broken into constituent parts through spectral decomposition techniques like Fourier and wavelet analyses (Beeman 1998). The same analytical approaches can be used to create synthetic stimuli that are mimetic to the extent that subjects fail to distinguish between a synthetic stimulus and a natural recording. For example, for túngara frogs, a synthetic signal described by 12 quantitative parameters is indistinguishable from recordings of natural calls to females (Ryan and Rand 1990, 1995). Likewise, mathematical decomposition of calls allows the experimenter to manipulate stimuli in diverse ways including “cutting and pasting” signal components (e.g., Clemens et al. 2014) or altering spectral and temporal properties. The modular nature of acoustic signals even allows experimenters to synthesize putative ancestral calls, thereby testing hypotheses about the coevolutionary history of signals and mating preferences (Ryan and Rand 1995).

Playback in the visual modality, which encompasses artificial stimuli ranging from video recordings of live exemplars to synthetic computer animations (Rosenthal 1999; Chouinard-Thuly et al. 2017), offers similar opportunities to manipulate stimuli but is inherently more challenging. In its technically simplest form, video playback uses recordings of live stimuli which are edited to vary in some attribute and then played back to test subjects. In the context of mate choice, for example, a subject might be tested for preference between 2 videos, presented either sequentially or simultaneously in a dichotomous choice arena, of the same courter performing different behaviors (e.g., Rosenthal et al. 1996). This allows the experimenter some control over variation in both the signal variable of interest and potentially confounding covariates.

Signal Fidelity

The technology used for recording, manipulating, and broadcasting stimuli is better for reproducing acoustic signals than it is for reproducing visual signals. Most of the longstanding criticisms of video playback methods revolve around the issue of signal fidelity (for more comprehensive reviews see D’Eath 1998; Fleishman et al. 1998; Fleishman and Endler 2000). Many of these issues still apply to contemporary techniques in the visual modality (Chouinard-Thuly et al. 2017).

In some important ways, video technology has caught up with acoustics. High-resolution monitors, high-speed video, and high monitor refresh rates have largely eliminated earlier concerns about the temporal and spatial resolution of stimuli (Baldauf et al. 2008; Chouinard-Thuly et al. 2017). Hyperspectral imaging, or just point-sampling of reflectance or radiance, allows us to quantify continuous color spectra (reviewed in Stevens et al. 2007), and multi-camera recording allows us to record 3-dimensional representations of visual scenes (Peters and Evans 2007). Signal fidelity becomes limited with the output technology. We first discuss 2 peculiarities of video, its 2-dimensional nature and the way it represents color, before turning to deficiencies that video playback shares with other experimental approaches in visual behavior.

Virtual reality holds promise for addressing the first limitation of video, the representation of a 3-dimensional scene on a 2-dimensional cue. Two-dimensional playback fails to mimic cues of depth reviewed in (D’Eath 1998; Zeil 2000). So-called pictoral cues to depth such as angular size, occlusion, and texture gradients can be incorporated into recorded video or mimicked automatically by animation software (Peters and Evans 2007). However, non-pictoral cues cannot be simulated in a 2-dimensional scene. These include binocular disparity cues by which depth is judged by the difference in retinal image between eyes, motion parallax, and accommodation or focusing cues (Zeil 2000). Further, stimuli can become distorted at close range, an issue that can partially be addressed by not allowing the subject to come into close contact with the video screen (D’Eath 1998). Virtual reality systems, however, have the potential to provide non-pictoral cues and correct for distortion in real time (Stowers et al. 2014).

The biggest deficiency of video relative to acoustic playback lies in the latter’s ability to represent a continuous frequency spectrum, even for ultra- and infrasonic stimuli (Pye and Langbauer 1998; Bradbury and Vehrencamp 2011). With video, by contrast, the color spectrum is down-sampled into 3 values that address human color perception. Video monitors therefore do not adequately mimic natural colors for non-human animals (Fleishman and Endler 2000; Chouinard-Thuly et al. 2017). It is possible in principle to adjust the white balance (the relative weight of the 3 color outputs) to the spectral sensitivities of other animals, but matching these sensitivities accurately is compromised since photoreceptor sensitivity can shift over ontogeny and seasonally (e.g., Cronly-Dillon and Sharma 1968; Spady et al. 2006; Temple et al. 2008). Further, commonly used animals will often have at least one photoreceptor whose sensitivity lies outside the spectral range of the monitor (Fleishman et al. 1998), notably in the ultraviolet (Cuthill et al. 2000; Fleishman and Endler 2000).

It should be noted that the failure to reproduce the ultraviolet portion of the visual spectrum is not a problem peculiar to video playback. Many, perhaps most, experiments using live exemplars or 3 dimensional models use a UV-deficient lighting environment or UV-absorbent materials that effectively abolish ultraviolet signals.

The same is true of polarization signals. Recent studies have shown that polarization sensitivity is important to the visual ecology of fish (Brady et al. 2015), cephalopods (Mäthger et al. 2009), and stomatopod crustaceans (Gagnon et al. 2015). With liquid crystal display monitors, all emitted light is linearly polarized while other monitor types such as cathode ray tube (CRT) produce virtually no polarized light (Chouinard-Thuly et al. 2017). The inability to manipulate polarized light is a potentially serious limitation of video stimuli, but again one that has been rarely considered in other types of manipulative experiments.

Finally, the background against which a stimulus is viewed is crucial for whether it is detected and how it is perceived by the subject. Most acoustic playbacks in the laboratory control for background–signal interactions by presenting a signal against a background of silence, but a “silent” visual background of black makes some signalers much more conspicuous than others. Since visual conspicuousness is defined relative to background, no such standard is possible for visual stimuli (Rosenthal 1999; Chouinard-Thuly et al. 2017). The spectral and spatiotemporal properties of visual backgrounds are of primary importance to receivers responding to visual stimuli (Uy and Endler 2004; Rosenthal 2007), and it is critical that researchers incorporate ecologically relevant background variation into their study designs.

Artifice is useful in animal behavior only to the extent that we can use it to draw conclusions about how animals would interact with natural stimuli. Signal fidelity poses the critical concern that a technique may introduce unanticipated confounding variables that may mislead our interpretation about what subjects attend to (D’Eath 1998; Fleishman et al. 1998; McGregor 2000).

Interactivity

An important feature of artifice is whether or not it is interactive—that is, whether the behavior of a model changes in response to the behavior of the subject. Playback can be interactive or not, as can the experiments using live exemplars (e.g., using 1-way mirrors). Most playback experiments are non-interactive; stimulus presentation does not depend on subject behavior. This can be advantageous, because it controls for receiver behavior and temporal variation in stimulus behavior (Kodric-Brown and Nicoletto 1997). However, courtship and agonistic displays, the most common signals simulated in video playback, are almost always interactive, with signalers responding dynamically with receiver responses. In some cases simulating such interactivity is useful for eliciting biologically meaningful subject responses (Ord et al. 2002; Chouinard-Thuly et al. 2017). In the case of courtship, the chooser interacts with the courter and often modifies its behavior according to the chooser’s response. It is possible to generate real-time interaction between stimulus and subjects in 2-dimensional space. For example, Butkowski et al. (2011) used data from tracking software to determine the position and behavior of a courting stimulus in real time. For longer-range interactions where a signaler broadcasts visual signals from one spot, such 2-dimensional interactions may mimic the natural dynamics of communication, for example, threat displays in Jacky dragon lizards (Ord et al. 2002).

Many of the behaviors simulated by playback, however, involve interactions in 3 dimensions, and as such playback may be inadequate to mimic complex interactive behaviors. For example, male sheepshead swordtail fish often prevent females from swimming away by blocking her exit path (Powell DL and Rosenthal GG, unpublished data). An entity that is trapped in 2 dimensions like a video playback stimulus cannot perform such 3 dimensional maneuvers. It should be noted that the same limitation is present when live signalers are blocked from directly interacting with subjects.

These kinds of interactions present a further problem when we want to interpret data from dichotomous choice tests, a very common design in experimental animal behavior. Here, it may be difficult to disentangle a subject moving away from stimulus A from a subject moving toward stimulus B. We may minimize this problem in part, by employing a Y-maze design in which moving away from one stimulus does not necessitate moving toward the other and by refining behavioral assays to characterize both aversive and proceptive responses to stimuli (Rosenthal forthcoming, in 2016). Some of these issues can also be circumvented by single-choice designs, whereby subjects are presented with a single stimulus at a time (Wagner 1998). However, single-choice assays tend to reduce choosiness, that is, the extent to which subjects respond differently to different stimuli. Single-choice assays may or may not be appropriate depending on whether subjects are likely to encounter signals simultaneously or sequentially in nature (Dougherty and Shuker 2015).

Robotic stimuli can provide 3-dimensional interaction. Robots have been used to explore several classes of behavioral responses e.g., differential responses to displays according to receiver sex in lizards (Martins et al. 2005), male courtship in birds (Patricelli et al. 2006; Patricelli and Krakauer 2009), female preference for color pattern (Phamduy et al. 2014) and schooling behavior in fish (Butail et al. 2013). However, the use of interactive robots presents a host of considerations not relevant for other playback experiments, notably incidental sensory cues that arise as a constraint of the mechanics of presentation, a response to any of which may complicate interpretation of results. For example, the inflation and deflation of artificial vocal sacs used in many anuran behavior studies produces mechanical sounds to which receivers may respond (Taylor et al. 2008; Klein et al. 2012).

Interactivity, or the lack thereof, presents a difficult problem in the experimental study of animal communication. Even so-called open-field interactions may introduce artifacts arising from spatial confinement or features of the habitat. One approach is to measure subject behavior across multiple contexts. For example, Kingston et al. (2003) used open field and dichotomous choices test with live stimuli, accompanied by dichotomous choice tests using color-filtered monochrome animated stimuli to show that females pygmy swordtails prefer blue morph males over the more aggressive gold morph males. Animated stimuli allowed them to decouple behavior from color.

Pseudoreplication

While workers have long articulated the problem of pseudoreplication (Hurlbert 1984; McGregor et al. 1992), many studies using synthetic, natural, and live stimuli continue to use limited numbers of exemplars to draw conclusions about responses to general properties of signals. It is worth mentioning issues specific to artifice here. Manipulated complex stimuli, whether digitally altered video or hand-painted dummies, are expensive and time-consuming to produce. Therefore, researchers often produce a handful at most of stimuli, and commit pseudoreplication by assuming that a given pair of artificial stimuli will elicit the same response as another similarly manipulated pair. One of the most attractive aspects of synthetic stimuli is that they help avoid pseudoreplication by allowing experimenters to frame hypotheses around defined stimulus properties, for example the population sample mean and standard deviation of a call parameter. With complex visual stimuli like animations, it can be challenge to measure all relevant parameters. This means that efforts to represent mean values for some parameters are often compromised by pseudoreplication with respect to others. For example, many of the animations we have generated (e.g., Rosenthal et al. 2002; Fisher et al. 2009; Verzijden and Rosenthal 2011) are compromised by the use of individual photographs for body texture and the use of single motion paths for rotoscoping stimulus display. This does nothing to eliminate pseudoreplication, unless many different animations are constructed each using unique texture files and motion paths.

Validating Experimental Designs Using Artifice

The hypotheses we test with artifice are usually ones that we want to extend to fitness consequences in natural stimuli. It is therefore necessary to demonstrate that subjects respond to artifice in ways that are directly comparable to how they behave in nature. If we have access to wild populations of the study organism, we can directly compare responses in experimental settings to responses to natural stimulus variation in the wild. Short of this, we should compare responses to artificial stimuli to responses to live stimuli. As Fleishman and Endler (2000) point out such comparison should be between stimuli of the same kind. That is, it is valid only to compare a specific class of synthetic stimulus to a live stimulus of the same type (i.e., an animation of a blue-morph male vs. a live blue-morph male).

It is also necessary to carefully consider whether the behaviors being assayed are reliable proxies for natural social behaviors associated with the relevant phenomenon, whether it be mate choice, aggression, or antipredator response. Measuring behavior across multiple contexts can increase confidence in the interpretation of subject responses to artifice (Kingston et al. 2003), as can testing appropriate negative controls such as subject responses to playback background in the absence of a stimulus. Furthermore, measures of attention such as gaze tracking (Yorzinski et al. 2013) represent a promising avenue for validating assayed behaviors. Ideally, we could measure behaviors or physiological responses specific to the process of interest, for example, proceptive behaviors like lordosis for mate choice. Such ground-truthing of methods is often overlooked or at least over-generalized from one taxon to another. When possible, we should directly compare behavioral proxies to natural behaviors. For example, the commonly used proxy of mate choice in small freshwater fishes, association time, can be directly validated by actual mate choice outcomes (Walling et al. 2009), but generalizing the utility of this proxy from one population or species to related groups may not be straightforward.

Where possible, experimental results should be validated using multivariate measures in wild populations. For the case of mate choice, this might include parentage analysis whereby responses to stimuli can be compared with actual mating decisions as quantified by the genetic makeup of offspring (Culumber et al. 2014). Another, behavioral approach might be to conduct social-network analyses of interactions involving proceptive or aversive behaviors of receivers as a function of variation in signalers (Wey et al. 2008). Analysis of agonistic interactions or responses to referential signals in the wild can further validate video stimuli designed to test such signals. The efficacy of synthetic stimuli is supported if they elicit responses that are congruent with responses in the wild to natural stimulus variation.

Synthesis and Future Directions

New technologies in animal behavior increase the scope and complexity of the stimuli we can represent and therefore the questions we can ask. With increasing complexity, however, comes increased risk that our experiments elucidate responses that don’t address the questions we ask. With proper design of stimuli and experiments, techniques like synthetic animations and virtual reality are powerful tools. For contemporary technologies with visual stimuli, the “best practices” paper by Chouinard-Thuly et al. (2017) addresses the important points that we need to be aware of with respect to signal fidelity with complex visual stimuli. If we cannot correct for these “known unknowns”, our hypotheses and interpretations need to be carefully circumscribed. For example, video playback cannot easily address hypotheses regarding color signaling for most organisms, but is nevertheless useful for testing responses to spatiotemporal patterns, given appropriate caveats about possibly confounding effects of distorted color representation (Fleishman and Endler 2000).

Complex visual manipulations, like any artifice, can address many hypotheses not tractable using other methods provided our hypotheses are sufficiently focused to accommodate its technical limitations (Oliveira et al. 2000). This of course is an important consideration for any study, just as the validation of stimulus efficacy as well as suitability of assayed behaviors as proxies for natural responses to stimuli are necessary for reliable interpretation of results. Technology alone cannot make our studies more interpretable. The major technical improvements to video output have removed limitations of spatial and temporal acuity, but we still have no way to present signals on video in a way that approaches the fidelity of acoustic signals. The increased complexity of animations and virtual reality has provided quantitative rigor, and mitigated pseudoreplication in some respects, but has at the same time raised a host of new problems that we must now address. By deploying an array of methodologies for assaying subject responses, and by appropriately framing and limiting our hypotheses, we can use this ever-expanding toolkit to gain unprecedented insight into animal communication.

We are especially indebted to Laura Chouinard-Thuly, Stefanie Gierszewski, and Klaudia Witte for organizing the symposium and putting together this special column. Pablo Delclós, Chris Holland, Mateo Garcia, Elizabeth Marchio, and 3 anonymous reviewers provided valuable comments that greatly improved this manuscript.

Funding

Manuscript preparation was supported by the National Science Foundation [1354172 to G.G.R. and [DGE 1252521 to D.L.P.].

References

  1. Beeman K, 1998. Digital signal analysis, editing and synthesis In: Hopp SL, Owren MJ, Evans CS, editors. Animal Acoustic Communication. Berlin: Springer, 59–101. [Google Scholar]
  2. Bell RW, 1974. Ultrasounds in small rodents: arousal-produced and arousal-producing. Dev Psychobiol 7:39–42. [DOI] [PubMed] [Google Scholar]
  3. Brady PC, Gilerson AA, Kattawar GW, Sullivan JM, Twardowski MS. et al. , 2015. Open-ocean fish reveal an omnidirectional solution to camouflage in polarized environments. Science 350:965–969. [DOI] [PubMed] [Google Scholar]
  4. Baldauf SA, Kullmann H, Bakker T, 2008. Technical restrictions of computer-manipulated visual stimuli and display units for studying animal behaviour. Ethology 114:737–751. [Google Scholar]
  5. Bradbury JW, Vehrencamp SL, 2011. Principles of Animal Communication. Sunderland: Sinauer. [Google Scholar]
  6. Butail S, Bartolini T, Porfiri M, 2013. Collective response of zebrafish shoals to a free-swimming robotic fish. PLoS ONE 8:e76123.. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Butkowski T, Yan W, Gray AM, Cui R, Verzijden MN. et al. , 2011. Automated interactive video playback for studies of animal communication. J Visual Exp 48:3–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Clemens J, Krämer S, Ronacher B, 2014. Asymmetrical integration of sensory information during mating decisions in grasshoppers. Proc Natl Acad Sci USA 111:16562–16567. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Chouinard-Thuly L, Gierszewski S, Rosenthal GG, Reader S, Rieucau G. et al. , 2017. Technical and conceptual considerations for using animated stimuli in studies of animal behavior. Curr Zool63:5–19. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Cronly-Dillon J, Sharma SC, 1968. Effect of season and sex on the photopic spectral sensitivity of the three-spined stickleback. J Exp Biol 49:679–687. [DOI] [PubMed] [Google Scholar]
  11. Culumber ZW, Ochoa OM, Rosenthal GG, 2014. Assortative mating and the maintenance of population structure in a natural hybrid zone. Am Nat 184:225–232. [DOI] [PubMed] [Google Scholar]
  12. Cuthill IC Hart NS Partridge JC Bennett ATD,Hunt S. et al. , 2000. Avian colour vision and avian video playback experiments. Acta Ethol 3:29–37. [Google Scholar]
  13. D’Eath RB, 1998. Can video images imitate real stimuli in animal behaviour experiments? Biol Rev 73:267–292. [Google Scholar]
  14. Dougherty LR, Shuker DM, 2015. The effect of experimental design on the measurement of mate choice: a meta-analysis. Behav Ecol 26:311–319. [Google Scholar]
  15. Evans CS, Marler P, 1991. On the use of video images as social stimuli in birds: audience effects on alarm calling. Anim Behav 41:17–26. [Google Scholar]
  16. Fisher HS, Mascuch SJ, Rosenthal GG, 2009. Multivariate male traits misalign with multivariate female preferences in the swordtail fish Xiphophorus birchmanni. Anim Behav 78:265–269. [Google Scholar]
  17. Fleishman LJ, Endler JA, 2000. Some comments on visual perception and the use of video playback in animal behavior studies. Acta Ethol 3:15–27. [Google Scholar]
  18. Fleishman LJ, McClintock W, D’eath RB, Brainard D, Endler JA, 1998. Colour perception and the use of video playback experiments in animal behaviour. Anim Behav 56:1035–1040. [DOI] [PubMed] [Google Scholar]
  19. Gagnon YL, Templin RM, How MJ, Marshall NJ, 2015. Circularly polarized light as a communication signal in mantis shrimps. Curr Biol 25:3074–3078. [DOI] [PubMed] [Google Scholar]
  20. Hurlbert S, 1984. Pseudoreplication and the design of ecological field experiments. Ecol Monogr 54:187–211. [Google Scholar]
  21. King SL, 2015. You talkin’ to me? Interactive playback is a powerful yet underused tool in animal communication research. Biol Lett 11:pii: 20150403. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Kingston JJ, Rosenthal GG, Ryan MJ, 2003. The role of sexual selection in maintaining a colour polymorphism in the pygmy swordtail Xiphophorus pygmaeus. Anim Behav 65:735–743. [Google Scholar]
  23. Klein BA, Stein J, Taylor RC, 2012. Robots in the service of animal behavior. Commun Integr Biol 5:466–472. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Kodric-Brown A, Nicoletto PF, 1997. Repeatability of female choice in the guppy: response to live and videotaped males. Anim Behav 54:369–376. [DOI] [PubMed] [Google Scholar]
  25. Marler P, Tamura M, 1962. Song “dialects” in three populations of white-crowned sparrows. Condor 64:368–377. [Google Scholar]
  26. Martins EP, Ord TJ, Davenport SW, 2005. Combining motions into complex displays: playbacks with a robotic lizard. Behav Ecol Sociobiol 58:351–360. [Google Scholar]
  27. Mateo JM, Holmes WG, 1997. Development of alarm-call responses in Belding’s ground squirrels: the role of dams. Anim Behav 54:509–524. [DOI] [PubMed] [Google Scholar]
  28. Mäthger LM, Shashar N, Hanlon RT, 2009. Do cephalopods communicate using polarized light reflections from their skin? J Exp Biol 212:2133–2140. [DOI] [PubMed] [Google Scholar]
  29. McGregor PK, 1992. Preface In: McGregor P, editor. Playback and Studies of Animal Communication. New York: Plenum Press, v–vii. [Google Scholar]
  30. McGregor PK, 2000. Playback experiments: design and analysis. Acta Ethol 3:3–8. [Google Scholar]
  31. McGregor PK, Catchpole CK, Dabelsteen T, Falls JB, Fusani L. et al. , 1992. Design of playback experiments: the Thornbridge Hall NATO ARW consensus In: McGregor P, editor. Playback and Studies of Animal Communication. New York: Plenum Press, 1–9. [Google Scholar]
  32. Oliveira RF, Rosenthal GG, Schlupp I, McGregor PK, Cuthill IC. et al. , 2000. Considerations on the use of video playbacks as visual stimuli: the Lisbon workshop consensus. Acta Ethol 3:61–65. [Google Scholar]
  33. Ord TJ, Peters RA, Evans CS, Taylor AJ, 2002. Digital video playback and visual communication in lizards. Anim Behav 63:879–890. [Google Scholar]
  34. Patricelli GL, Coleman SW, Borgia G, 2006. Male satin bowerbirds Ptilonorhynchus violaceus adjust their display intensity in response to female startling: an experiment with robotic females. Anim Behav 71:49–59. [Google Scholar]
  35. Patricelli GL, Krakauer AH, 2009. Tactical allocation of effort among multiple signals in sage grouse: an experiment with a robotic female. Behav Ecol 21:97–106. [Google Scholar]
  36. Phamduy P, Polverino G, Fuller RC, Porfiri M, 2014. Fish and robot dancing together: bluefin killifish females respond differently to the courtship of a robot with varying color morphs. Bioinspir Biomim 9:036021.. [DOI] [PubMed] [Google Scholar]
  37. Phelps SM, Ryan MJ, Rand AS, 2001. Vestigial preference functions in neural networks and túngara frogs. Proc Natl Acad Sci USA 98:13161–13166. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Peters RA, Evans CS, 2007. Active space of a movement-based signal: response to the Jacky dragon Amphibolurus muricatus display is sensitive to distance, but independent of orientation. J Exp Biol 210:395–402. [DOI] [PubMed] [Google Scholar]
  39. Pye JD, Langbauer WR, 1998. Chapter 7: Ultrasound and infrasound In: Hopp SL, Owren MJ, Evans CS, editors. Animal Acoustic Communication. Berlin: Springer, 222–248. [Google Scholar]
  40. Rand AS, Ryan MS, 1981. The adaptive significance of a complex vocal repertoire in a Neotropical frog. Z Tierpsychol 57:209–214. [Google Scholar]
  41. Rosenthal GG, 1999. Using video playback to study sexual communication. Environ Biol Fish 56:307–316. [Google Scholar]
  42. Rosenthal GG, 2007. Spatiotemporal aspects of visual signals in animal communication. Annu Rev Ecol Evol Syst 38:155–178. [Google Scholar]
  43. Rosenthal GG, Forthcoming 2016. Mate Choice: the Evolution of Sexual Decision Making from Microbes to Humans. Princeton: Princeton University Press. [Google Scholar]
  44. Rosenthal GG, Evans CS, Miller WL, 1996. Female preference for dynamic traits in the green swordtail Xiphophorus helleri. Anim Behav 51:811–820. [Google Scholar]
  45. Rosenthal GG, Ryan MJ, 2000. Visual and acoustic communication in non-human animals: a comparison. J Biosci 25:285–290. [DOI] [PubMed] [Google Scholar]
  46. Rosenthal GG, Ryan MJ, Wagner WE, Jr, 2002. Secondary loss of preference for swords in the pygmy swordtail Xiphophorus nigrensis (Pisces: Poeciliidae). Anim Behav 63:37–45. [Google Scholar]
  47. Ryan MJ, Rand AS, 1990. The sensory basis of sexual selection for complex calls in the túngara frog Physalaemus pustulosus (sexual selection for sensory exploitation). Evolution 44:305–314. [DOI] [PubMed] [Google Scholar]
  48. Ryan MJ, Rand AS, 1995. Female responses to ancestral advertisement calls in túngara frogs. Science 269:390–392. [DOI] [PubMed] [Google Scholar]
  49. Spady TC, Parry JW, Robinson PR, Hunt DM, Bowmaker JK. et al. , 2006. Evolution of the cichlid visual palette through ontogenetic subfunctionalization of the opsin gene arrays. Mol Biol Evol 23:1538–1547. [DOI] [PubMed] [Google Scholar]
  50. Seyfarth RM, Cheney DL, 1986. Vocal development in vervet monkeys. Anim Behav 34:1640–1658. [Google Scholar]
  51. Stevens M, Parraga CA, Cuthill IC, Partridge JC, Troscianko TS, 2007. Using digital photography to study animal coloration. Biol J Linn Soc 90:211–237. [Google Scholar]
  52. Stowers JR, Fuhrmann A, Hofbauer M, Streinzer M, Schmid A. et al. , 2014. Reverse engineering animal vision with virtual reality and genetics. Computer 47:38–45. [Google Scholar]
  53. Taylor RC, Klein BA, Stein J, Ryan MJ, 2008. Faux frogs: multimodal signalling and the value of robotics in animal behaviour. Anim Behav 76:1089–1097. [Google Scholar]
  54. Temple SE, Veldhoen KM, Phelan JT, Veldhoen NJ, Hawryshyn CW, 2008. Ontogenetic changes in photoreceptor opsin gene expression in coho salmon (Oncorhynchus kisutch, Walbaum). J Exp Biol 211:3879–3888. [DOI] [PubMed] [Google Scholar]
  55. ter Pelkwijk JJ, Tinbergen N, 1937. Eine reizbiologische Analyse einiger Verhaltensweisen von Gasterosteus aculeatus L. Z Tierpsychol 1:193–200. [Google Scholar]
  56. Uy J, Endler AC, 2004. Modification of the visual background increases the conspicuousness of golden-collared manakin displays. Behav Ecol 15:1003–1010. [Google Scholar]
  57. Verzijden MN, Rosenthal GG, 2011. Effects of sensory modality on learned mate preferences in female swordtails. Anim Behav 82:557–562. [Google Scholar]
  58. Wagner WE, 1998. Measuring female mating preferences. Anim Behav 55:1029–1042. [DOI] [PubMed] [Google Scholar]
  59. Walling CA, Royle NJ, Lindström J, Metcalfe NB, 2009. Do female association preferences predict the likelihood of reproduction? Behav Ecol Sociobiol 64:541–548. [Google Scholar]
  60. Wey T, Blumstein DT, Shen W, Jordán F, 2008. Social network analysis of animal behaviour: a promising tool for the study of sociality. Anim Behav 75:333–344. [Google Scholar]
  61. Yorzinski JL, Patricelli GL, Babcock JS, Pearson JM, Platt ML, 2013. Through their eyes: selective attention in peahens during courtship. J Exp Biol 216:3035–3046. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Zeil J, 2000. Depth cues, behavioural context, and natural illumination: some potential limitations of video playback techniques. Acta Ethol 3:39–48. [Google Scholar]

Articles from Current Zoology are provided here courtesy of Oxford University Press

RESOURCES