Skip to main content
PLOS One logoLink to PLOS One
. 2024 Aug 28;19(8):e0307189. doi: 10.1371/journal.pone.0307189

How do soundboard-trained dogs respond to human button presses? An investigation into word comprehension

Amalia P M Bastos 1,2, Ashley Evenson 3,4, Patrick M Wood 1,2,4, Zachary N Houghton 5, Lucas Naranjo 4,6, Gabriella E Smith 7,8, Alexandria Cairo-Evans 1, Lisa Korpos 1,4, Jack Terwilliger 1, Sarita Raghunath 1, Cassandra Paul 1, Hairou Hou 1, Federico Rossano 1,*
Editor: Brenton G Cooper9
PMCID: PMC11355551  PMID: 39196871

Abstract

Past research on interspecies communication has shown that animals can be trained to use Augmentative Interspecies Communication (AIC) devices, such as soundboards, to make simple requests of their caretakers. The recent uptake in AIC devices by hundreds of pet owners around the world offers a novel opportunity to investigate whether AIC is possible with owner-trained family dogs. To answer this question, we carried out two studies to test pet dogs’ ability to recognise and respond appropriately to food-related, play-related, and outside-related words on their soundboards. One study was conducted by researchers, and the other by citizen scientists who followed the same procedure. Further, we investigated whether these behaviours depended on the identity of the person presenting the word (unfamiliar person or dog’s owner) and the mode of its presentation (spoken or produced by a pressed button). We find that dogs produced contextually appropriate behaviours for both play-related and outside-related words regardless of the identity of the person producing them and the mode in which they were produced. Therefore, pet dogs can be successfully taught by their owners to associate words recorded onto soundboard buttons to their outcomes in the real world, and they respond appropriately to these words even when they are presented in the absence of any other cues, such as the owner’s body language.

Introduction

The use of Augmentative Interspecies Communication (AIC) devices has gained popularity among pet owners in recent years, especially among dog owners [13]. Most of these AIC devices consist of soundboards with buttons that, when pressed, produce pre-recorded words in the owner’s voice. Most owners report that they train their dogs using a technique that has become known as modelling: owners demonstrate the actions associated with each button repeatedly until their animals make presses themselves, at which point the same actions are performed by the owners, regardless of whether the animal’s press was intentional or accidental [2]. Over time, animals may press the buttons on their soundboards more frequently, which may in turn provide them with more control over their daily lives and environments.

Although interspecies communication through soundboard-like AIC devices has been adopted to varying degrees of success with apes, dolphins, and other species (for a review, [1]), this has primarily been undertaken by researchers and professional animal trainers. In the scientific literature, only a single dog named Sofia has been shown to use buttons to request actions such as going on a walk or playing, following training by a professional dog trainer [4, 5]. The training methods used with Sofia are considerably different from the modelling approach that is most common among pet owners: Sofia was first taught to tap her AIC device with her paw on command, which was eventually phased out once she began doing so spontaneously [4]. Given the nature of Sofia’s training, it is also possible that Sofia experienced less cueing from her trainer than other owner-trained pet dogs, whose performance may be more susceptible to the Clever Hans effect [6, 7].

Crucially, communicative production of words, such as that demonstrated by Sofia, must be preceded by the comprehension of said words. Although Sofia is the only dog ever shown to produce requests through an AIC device, many more dogs are capable of responding appropriately to spoken human signals. Most pet dogs are trained to respond to at least a few vocal signals from their owners such as “sit” and “lie down” [8, 9], but some dogs such as Rico and Chaser can learn tens or hundreds of names for individual objects ([1017]; although learning object names may be more challenging to dogs than learning words associated with actions, see [18]). Given that owner-trained soundboard-using dogs have not yet been tested in controlled experimental contexts, we currently do not have any evidence to suggest that they have successfully associated the words produced by their soundboards to their respective consequences, let alone that they produce said words communicatively.

Another potential issue surrounding owner-trained soundboard-using dogs concerns the reliability of citizen science data [19]. Although thousands of pet owners currently contribute regular data on their animals’ soundboard use to scientific research [2], the extent to which reported interactions are cued, inaccurately reported, or cherry picked by owners is unclear and a matter under current investigation. Consequently, testing soundboard-trained dogs under controlled experimental conditions is a necessary step towards assessing this form of interspecies communication.

Given the novelty of this approach, the nature of most of the data on this form of interspecies communication, and our limited knowledge of dogs’ capacity to acquire associations in this context, in the present study we pursue a thorough and multifaceted investigation into the comprehension of button presses by humans. We begin from the premise that owner-trained soundboard-using dogs are probably more likely to associatively learn the words for routine activities or events (such as learning that the word “outside”, when spoken or pressed by a human, is usually followed by a door being opened to the backyard) than some of the more abstract words sometimes provided on owner-trained dogs’ soundboards (such as “tomorrow” or “later”). Given that dogs’ soundboards are determined by their owners and therefore are composed of buttons with different words, and have different layouts (for a detailed explanation, see [2]), here we select three of the most commonly occurring routine-related word concepts across all participants (namely, food-related, play-related, and outside-related words) and choose the most common variants within these concepts (OUT or OUTSIDE for outside-related words, PLAY or TOY for play-related words, and FOOD or EAT or DINNER or HUNGRY for food-related words) for our experiments (as pre-registered in the Open Science Framework: https://osf.io/tskcq). Over the course of two complementary experiments, we use these three word concepts, as well as a nonsense word (the nonce word “DAXING”) to answer five fundamental questions about owner-trained soundboard-using dogs’ comprehension of the words on their soundboards.

First, using both a researcher-led in-person experiment and a remotely conducted citizen science experiment, we test whether dogs can recognise and respond appropriately to the words recorded onto their soundboards. We hypothesise that, if dogs have associatively learnt the connection between these words and the outcomes they usually entail, then they should behave in anticipation of the actions or events indicated by these words. For example, upon hearing the word “PLAY”, dogs should be more likely to walk to their toy box and pick up a toy, than if they had heard “FOOD”. Conversely, we should see fewer looks to a food bowl following the word “PLAY” than following “FOOD”. Given that these words might typically be produced by owners concurrently with many other contextual cues (for example, “FOOD” might be produced most commonly at specific mealtimes, while the owner fills up the dog’s bowl), in our experimental trials we strip all words of any additional contextual cues, to ensure that dogs’ responses occur specifically in response to the words themselves.

Second, we use the two experiments to compare dogs’ responses to a word when it is produced by the owner and by an unfamiliar person (a researcher). We hypothesise that, if dogs expect words to precede their accompanying actions, and can do so without any additional contextual or owner-provided cues, then they should respond similarly regardless of the person producing the word. Given that words on dogs’ buttons usually consist of recordings of their owner’s speech (as was the case for every dog participating in this study), and that differences in the acoustic properties of different people’s speech can conceivably generate differences in dogs’ behavioural responses unrelated to their comprehension of words, we compare their responses to button presses by the owner and presses of those same buttons by an unfamiliar person.

Third, a comparison between the two experiments–particularly in terms of dogs’ responses to button presses by their owners and an unfamiliar person–can also inform whether the results of citizen science studies on soundboard-trained dogs are broadly comparable to in-person researcher-led studies, and therefore establish whether future work with this population could be carried out through data collected remotely by citizen scientists.

Fourth, we also compare dogs’ responses to owners’ button presses and owners’ vocal production of these same words. If dogs respond appropriately and equivalently to words regardless of their mode of production (whether they are spoken directly by the owner or produced by a button press), then this would suggest that dogs are associating words’ consequences (e.g., the actions following words, such as a play interaction following “PLAY”) to the words (speech or speech recordings) themselves, rather than attending primarily to other cues such as the location of the buttons (which would predict contextually-appropriate responses when buttons are pressed, but not when the same words are spoken out loud). On the other hand, if dogs respond appropriately to owners’ button presses of words but not to their vocal production, then it is likely that dogs associate button locations–or some other property of the buttons–to their accompanying actions or events, disregarding button audio.

Finally, because of recent research suggesting that dogs tilt their heads in response to familiar words more so than unfamiliar ones [20], and that this is a lateralized behaviour, we investigated subjects’ head tilts throughout the two experiments. We hypothesise that, if dogs do in fact tilt their heads in response to words that they recognise, and soundboard-trained dogs recognise the words recorded onto their buttons, then we should see more head-tilting behaviour in response to the known words (i.e., outside-related, play-related, and food-related words) than to a nonce word (“DAXING”). This result should hold regardless of whether words are produced by button presses or spoken by their owner. If this is true, then this study can also provide a larger sample size to inform whether this head-tilting behaviour is more commonly left-lateralized or right-lateralized, or, as in the original study, this is split evenly across subjects. If, on the other hand, we find evidence to suggest that soundboard-trained dogs do respond appropriately to known words but do not tilt their heads upon hearing them, this might instead suggest that there are population or breed differences between soundboard-trained dogs and the “gifted world learner” [15] dogs in the study conducted by Sommese and colleagues [20].

In sum, the present study aims to investigate five principal questions: (1) whether dogs can recognise and respond appropriately to the words recorded on their soundboards; (2) whether they exhibit these responses even in the absence of other contextual or owner-produced cues, even when they are produced by an unfamiliar person; (3) whether citizen science studies and in-person studies can produce comparable results in this population of subjects; (4) whether dogs attend to speech specifically in this context, or other cues; and (5) whether owner-trained soundboard-using dogs tilt their heads in response to familiar words in the same way that dogs trained to recognise object names do.

Methods

Experiment 1: In person study

Subjects

For the in-person study (IPS), subjects were 30 dogs (14 males, age: M = 3.52 years, SD = 2.91; see Table 1 for subject information). No subjects were excluded from analyses. All dogs were family pets trained by their owners to use soundboard AIC devices, and whose soundboards contained the words OUT/OUTSIDE, PLAY/TOY, and FOOD/EAT/DINNER/HUNGRY. Owners were asked to habituate their dogs to people wearing sunglasses, hats, and other head accessories prior to the study date, in their own time, given that the study would involve owners wearing multiple accessories on their faces. Dogs were tested in their home settings with their owners present; owners were not told the purpose of the study until testing was concluded. All subjects lived primarily indoors and were fed indoors and the study received ethics approval from the UCSD IACUC (protocol no. S21098).

Table 1. List of subjects that participated in Experiment 1.

Age at time of testing is provided in years, rounded to the nearest decimal point, as per owner reports. For mixed breed dogs, up to three primary component breeds (based on owner reports or genetic testing) are provided in brackets.

Subject Breed Sex Age (Years) Eligible Buttons Testing Location
Adora Border Collie F 3.0 OUTSIDE, PLAY, EAT Living Room
Alvin Mixed Breed (Golden Retriever, Pug, Beagle) M 9.6 OUTSIDE, PLAY, FOOD Living Room
Bastian Mixed Breed (Rat Terrier, Pomeranian, Miniature Poodle) M 4.1 OUTSIDE, PLAY, HUNGRY Living Room
Buster Mixed Breed (French Bulldog, Boston Terrier) M 5.8 OUTSIDE, PLAY, EAT Living Room
Charlie Yorkshire Terrier M 3.2 OUTSIDE, PLAY TOY, FOOD Living Room
Chester Mixed Breed (Labrador Retriever, Boston Terrier, Pomeranian) M 14.0 OUTSIDE, PLAY, FOOD Living Room
Chuckles Mixed Breed (American Cocker Spaniel, Miniature Poodle) M 2.9 OUTSIDE, PLAY, FOOD Living Room
Dillon Mixed Breed (Labrador Retriever, German Shepherd Dog, Chow Chow) M 2.8 OUTSIDE, PLAY, FOOD Living Room
Eleanor German Shepherd Dog F 7.5 OUTSIDE, PLAY, HUNGRY Living Room
Eleanor Mixed Breed (Bernese Mountain Dog, Standard Poodle) F 1.8 OUTSIDE, PLAY, FOOD Living Room
Frank Mixed Breed (Affenpinscher, Miniature Schnauzer) M 2.8 OUTSIDE, PLAY, FOOD Living Room
Gibson Mixed Breed (Bernese Mountain Dog, Miniature Poodle) M 1.3 OUTSIDE, PLAY, HUNGRY Living Room
Gracie Mixed Breed (Miniature Poodle, Labrador Retriever) F 3.2 OUTSIDE, PLAY, HUNGRY Living Room
Izzy Miniature American Shepherd F 2.7 OUTSIDE, PLAY, HUNGRY Living Room
Jumbo Miniature American Shepherd F 2.2 OUTSIDE, PLAY, HUNGRY Entrance Hallway
Juniper Mixed Breed (Miniature Poodle, Chihuahua, other) F 8.5 OUTSIDE, PLAY, FOOD Home Office
Juno Mixed Breed (Pug, American Bully, American Pit Bull Terrier) F 1.3 OUTSIDE, PLAY, EAT Living Room
Kitsune Mixed Breed (Bernese Mountain Dog, Miniature Poodle) F 0.6 OUTSIDE, PLAY, HUNGRY Living Room
Lobo Alaskan Klee Kai M 5.2 OUTSIDE, PLAY, HUNGRY Living Room
Luna Miniature American Shepherd F 2.7 OUTSIDE, PLAY, HUNGRY Living Room
Misty Mixed Breed (Siberian Husky, Australian Cattle Dog) F 0.6 OUTSIDE, PLAY, EAT Living Room
Mokai Border Collie M 2.5 OUTSIDE, PLAY, EAT Dining Room
Oski Pug M 3.5 OUTSIDE, PLAY, FOOD Living Room
Parker Mixed Breed (Beagle, American Eskimo Dog, Chihuahua) F 1.5 OUTSIDE, PLAY, HUNGRY Living Room
Pepita Mixed Breed (American Foxhound, Australian Cattle Dog, Beagle) F 1.9 OUTSIDE, PLAY, FOOD Dining Room
PJ Mixed Breed (Standard Poodle, Golden Retriever) M 1.3 OUTSIDE, PLAY, EAT Living Room
Rey Mixed Breed (Beagle, Dachshund, Miniature Pinscher) F 3.5 OUTSIDE, PLAY, EAT Living Room
Scout Mixed Breed (German Shepherd Dog, Boxer, American Pit Bull Terrier) F 2.7 OUTSIDE, PLAY, HUNGRY Living Room
Simon Standard Poodle M 0.6 OUTSIDE, PLAY, FOOD Living Room
Zero Shiba Inu F 2.3 OUTSIDE, PLAY, HUNGRY Living Room

Procedure

The study was conducted in dogs’ homes, in the room where their soundboard was typically located. Upon arrival, one researcher (Experimenter 1, hereafter E1) waited outside the house whilst the other (Experimenter 2, hereafter E2) went into the home and greeted the owner and their dog, before placing the dog either in the backyard or in a room in the house other than the one where the soundboard was located.

Once the dog was out of sight, E2 placed three large coloured stickers over the three buttons containing one of each of the words of interest: (1) “OUT” or “OUTSIDE”; (2) “PLAY” or “TOY”; and (3) “FOOD” or “EAT” or “DINNER” or “HUNGRY”. This ensured that, at test, E1 could not tell the buttons’ identity based on any writing or symbols visible on the buttons. E2 then asked that the owner record their voice onto a new button (that matched the size and style of the dog’s existing soundboard buttons) for the word “DAXING” and placed that on the soundboard, with another coloured sticker. Sticker colours were randomly assigned to buttons between subjects, such that E1 would not have known the identity of any buttons based on the colour of the stickers placed on them. Stickers were made from green, red, yellow, and blue cardstock paper and adhesive putty. Owners were shown a novel action–placing their hands on their head and spinning in a slow circle–which they were told they would have to perform at some point later in the study.

The dog was then brought back into the room where their soundboard was located, and E1 was invited into the home and allowed to greet the dog and owner. E2 then set up the camera recording equipment (Panasonic Full HD Camcorder HC-V180K, GoPro Hero7) in the room, and an Anmeate Baby Monitor facing the soundboard, such that multiple camera angles of the room were recorded (Fig 1).

Fig 1. Set up of a trial for the in-person study.

Fig 1

The door used for the button concept “OUTSIDE” is outlined in blue, the dog’s toys are outlined in orange, and the food bowl is outlined in green. During trials for this subject, for example, the owner sat in the reclining chair on the bottom left of the frame, facing away from the soundboard, while wearing noise-cancelling headphones, a face mask (as a Covid-19 precaution), and a black sleep mask.

While E1 explained the experimental procedure to the owner and ensured the dog was comfortable and behaving normally in their presence, E2 identified a different room in the house to hide so they could remotely monitor the study (through a live feed of the baby monitor). Next, the owner was asked to wear a sleep mask and noise-cancelling headphones and ensure that the dog was not fearful of either accessory. Finally, E2 handed E1 their headphones and set both headphone volumes such that that neither E1 nor the dog’s owner could hear E2 speaking loudly from a short distance, therefore ensuring that they would also be unable to hear the recordings on any of the buttons when they were pressed during a trial.

Trials began with E1 and the owner sitting in the soundboard room whilst the dog ranged freely around the room. The owner wore a sleep mask and listened to music on noise-cancelling headphones so they could not see any of the trial procedures. This ensured that owners were blind to the study’s hypotheses and predictions but were still present in the room to ensure that dogs were comfortable and behaving normally.

At the start of each trial, E2 remotely triggered a sound file to E1’s headphones stating the colour of the sticker for the button they would press in that trial, then remotely started playing music into the headphones of both E1 and the dog’s owner. E1 looked straight ahead until they heard a beep in their headphones, again triggered via Bluetooth by E2. At this point, E1 stood up and began the trial: E1 walked to the same place by the soundboard each time, pressed the button covered by the sticker colour they were given, stepped away from the button again, and looked in the direction of the dog. E1 froze in this position for 1 minute while they waited for the next beep in their headphones, at which point they turned around to face the nearest wall and closed their eyes. This final beep and turn concluded the trial.

Following the end of the trial, E2 triggered an audio file in the owner’s headphones stating that they could remove their headphones and sleep mask and gave the owner instructions on an action to perform. Unbeknownst to the owner, this action always matched the button recently pressed by E1. For “OUT” or “OUTSIDE”, the owner was asked to let their dog outside as they normally would (e.g., open the door to the backyard so the dog could go out). For “PLAY” or “TOY”, the owner was asked to briefly engage their dog in play with a toy. For “FOOD”, “EAT”, “DINNER”, or “HUNGRY”, the owner was asked to place a small amount of their dog’s usual food in their food bowl. For “DAXING”, the owner performed the novel action they practiced earlier, namely placing both hands on the top of their head and spinning in a slow circle. While the owner performed their action, E1 remained with their eyes closed, listening to music through noise-cancelling headphones, so that they would not be unblinded to the button they just pressed. Once the owner performed their action, E2 came out of their hiding place and asked the owner to return the room to its starting state (e.g., closing any doors that were opened, placing any toys back where they were at the start of the trial), and returning to their starting position in the room. E2 then walked over to E1 and tapped them on the shoulder, so that E1 knew they could remove their headphones, open their eyes, and turn back around.

Experiment 2: Citizen science study

Subjects

In the citizen science study (CSS), subjects were a new group of 29 dogs (9 males, age: M = 2.95 years, SD = 1.91; see Table 2 for subject information). Another 3 subjects (3 males) completed the study but had their data excluded due to procedural errors, and a further 8 subjects (4 males) were recruited for the study but never completed it. All dogs were owner-trained family pets with soundboards containing the same words as required in Experiment 1. Owners were asked to habituate their dogs to sunglasses prior to beginning the study, since owners were asked to wear mirrored sunglasses throughout all trials. Dogs were tested in their home settings by their owners, who were not told the purpose of the study until testing was concluded. As before, all dogs lived primarily indoors and were fed their meals indoors, and the study received ethics approval from the UCSD IACUC (protocol no. S21098).

Table 2. List of subjects that participated in Experiment 2.

Age at time of testing is provided in years, rounded to the nearest decimal point, as per owner reports. For mixed breed dogs, up to three primary component breeds (based on owner reports or genetic testing) are provided in brackets.

Subject Breed Sex Age (Years) Eligible Buttons
BeBe Ivy Mixed Breed (American Foxhound, Boxer, American Pit Bull Terrier) F 3.3 OUTSIDE, PLAY, HUNGRY
Birdy Mixed Breed (American Pit Bull Terrier, other) F 1.8 OUTSIDE, PLAY, EAT
Dexadrine German Shepherd Dog F 3.0 OUTSIDE, PLAY, EAT
Dio Miniature American Shepherd F 2.2 OUTSIDE, PLAY, FOOD
Freya Australian Shepherd F 2.5 OUTSIDE, PLAY, HUNGRY
Ila Mixed Breed (Standard Poodle, Golden Retriever) F 3.7 OUTSIDE, PLAY, HUNGRY
Juno Mixed Breed (Springer Spaniel, Labrador Retriever) F 2.2 OUTSIDE, PLAY, DINNER
Juno Mixed Breed (American Pit Bull Terrier, Labrador Retriever, Australian Shepherd) F 1.1 OUTSIDE, PLAY, EAT
Kida Australian Shepherd F 4.7 OUTSIDE, PLAY, FOOD
Kiki Border Collie F 2.1 OUTSIDE, PLAY, FOOD
Kipo French Bulldog M 1.3 OUTSIDE, PLAY, FOOD
Lina Mixed Breed (Australian Kelpie, Pembroke Welsh Corgi) F 1.6 OUTSIDE, PLAY, FOOD
Lola Mixed Breed (Border Collie, Labrador Retriever, Standard Poodle) F 2.2 OUTSIDE, PLAY, FOOD
Lucy Havanese F 1.7 OUTSIDE, PLAY, FOOD
Lulu Mixed Breed (American Pit Bull Terrier, Labrador Retriever, Australian Shepherd) F 1.8 OUTSIDE, PLAY, FOOD
Monty Staffordshire Bull Terrier M 0.8 OUTSIDE, PLAY, EAT
Moxie German Shepherd Dog F 2.2 OUTSIDE, PLAY, FOOD
Neko Siberian Husky M 4.9 OUTSIDE, PLAY, FOOD
Nita German Shepherd Dog F 5.2 OUTSIDE, PLAY, FOOD
Oliver Golden Retriever M 2.4 OUTSIDE, PLAY, FOOD
Pepper Mixed Breed (Maltese, Miniature Poodle) F 1.9 OUTSIDE, PLAY, FOOD
Pinot Mixed Breed (Australian Shepherd, Labrador Retriever, Standard Poodle) M 2.1 OUTSIDE, PLAY, EAT
Rhea Rose Standard Poodle F 1.0 OUTSIDE, PLAY, FOOD
Roscoe Mixed Breed (Staffordshire Bull Terrier, Cavalier King Charles Spaniel) M 5.0 OUTSIDE, PLAY, FOOD
Sadie Mixed (German Shepherd Dog, Staffordshire Bull Terrier, Chow Chow) F 4.9 OUTSIDE, PLAY, EAT
Samwise Labrador Retriever M 1.7 OUTSIDE, PLAY, EAT
Waffle Mixed Breed (Chihuahua, Miniature Poodle, Pug) M 2.0 OUTSIDE, PLAY, HUNGRY
Zodiak Mixed Breed (American Pit Bull Terrier, Staffordshire Bull Terrier, Chow Chow) M 9.3 OUTSIDE, PLAY, FOOD
Zoe German Shepherd Dog F 7.0 OUTSIDE, PLAY, FOOD

Procedure

Owners were given study instructions in three formats: an explanatory video demonstrating study procedures; written materials with step-by-step instructions; and a 15-minute zoom call with a researcher who talked them through the study instructions and answered any remaining questions. As with the in-person study, owners were not told the purpose of the experiment until after they concluded their participation and submitted video data of all trials to the research team. Debriefs were also done by an experimenter through another 15-minute zoom call. Given that they administered the experimental trials to their own dogs, owners could not be made blind to conditions.

In each trial, owners either pressed the button for, or spoke, one of the words used in Experiment 1, namely: (1) “OUT” or “OUTSIDE”; (2) “PLAY” or “TOY”; and (3) “FOOD” or “EAT” or “DINNER” or “HUNGRY”, and (4) the nonce word “DAXING”. Each word was produced twice by the owner over the course of the study, once spoken and once pressed. Owners carried out two trials per day, allowing an interval of at least 30 minutes between the two trials within a day, over the course of four days, which were spread over a maximum of 2 weeks. Therefore, dogs in this study experienced a total of eight conditions: a spoken outside word, a pressed outside word, a spoken play word, a pressed play word, a spoken food word, a pressed food word, the word “DAXING” produced by a button press, and “DAXING” spoken by their owner. Trial orders were randomised across subjects.

Before the start of each trial, owners were asked to move their dog to a different room of the house where they could not see their soundboard. Owners then recorded a new button for the word “DAXING” and placed it in an empty slot on their dog’s soundboard. The dog was brought back into the room where their soundboard was located for the start of the trial.

Each trial began with the owner standing in the same location next to the dog’s soundboard. They then either pressed the button or spoke the word specified for that trial. Owners then refrained from interacting with their dog for the next minute; owners were specifically instructed not to speak to, point to, react to, gesture at, or otherwise interact with their dog or their environment for a full minute. At the end of each trial, owners were asked to perform the matching action for the pressed or spoken word, as specified by their participant sheet (i.e., as in Experiment 1, owners were asked to open the door and let their dog into the backyard following the production of the word “OUTSIDE”, and spin slowly with their hands on their heads following the word “DAXING”). Following their post-trial action, owners again took their dogs out of the room and returned the soundboard to its usual configuration, removing the button for the word “DAXING”, such that dogs could not press the “DAXING” button in the intervals between trials and on non-trial days. The full trial time was recorded on video by the owners, sometimes from multiple camera angles.

Coding and analyses

Videos for both studies (Experiment 1 and Experiment 2) were processed prior to blind coding to ensure that blind coders could never observe the button press produced by E1 (Experiment 1) or the button press or word spoken by the owner (Experiment 2), as well as the post-trial action performed by the owner. Video files were renamed so as to remove any trial information. Videos from different camera angles of the same trial were time-matched and aligned side-by-side into a single video file. E1 and the owner were occluded by a black rectangle such that blind coders could not see any humans on the screen, and the video was cut so that it began following the moment that E1 stood upright after pressing a button and ended 60 seconds later.

Blind coders were naïve to experimental design, experimental hypotheses, and conditions. They were initially trained on a set of videos which included trials of the study piloted with non-eligible dogs and non-study videos crowdsourced online of non-subjects. Coders annotated videos following a pre-determined ethogram (Table 3). Interrater agreement was substantial throughout training (Kappa coefficient of 0.81, with a confidence interval of [0.76; 0.86] for 100% of training set) and remained satisfactory throughout the coding of the full dataset (Kappa coefficient of 0.70, with a confidence interval of [0.64; 0.76] for a randomly selected subset comprising 10% of all data).

Table 3. Ethogram of behaviours coded for both experiments.

We predicted that dogs would show more outside-directed behaviours in response to “OUT” or “OUTSIDE” words, more toy-directed behaviours in response to “PLAY” or “TOY” words, and more food-directed behaviours in response to “FOOD”, “EAT”, “DINNER”, or “HUNGRY” words.

Undirected Behaviours
Behaviour Modifiers Behaviour Description
Lip/Nose Licking N/A Contact between tongue and dog’s nose or lips (including tongue flicks).
Head Tilting Left, Right Dog momentarily holds its head diagonally rather than vertically, to either side.
Door-Directed Behaviours
Behaviour Behaviour Description
Moving Towards Door Distance between dog and door decreases, as dog moves with their head oriented towards the door. It should be marked from the moment the dog lifts their paw to make the movement towards the door.
Looking at Door Dog faces the door and looks at it for >1 second.
Touching Door Dog intentionally physically interacts with the door: namely by scratching at, pawing at, or nudging it.
Sitting/Lying by Door Dog changes their position to sitting by or laying down by the door.
Toy-Directed Behaviours
Behaviour Behaviour Description
Moving Towards Toy Distance between dog and toy (or toy basket) decreases, as dog moves with their head oriented towards the toy (or toy basket). It should be marked from the moment the dog lifts their paw to make the movement towards the toy / toy basket.
Looking at Toy Dog faces the toy (or toy basket) and looks at it for >1 second.
Picking up Toy Dog picks up toy (or toy basket) with their muzzle.
Interacting with Toy Dog intentionally physically interacts with a toy (or toy basket): namely by scratching at, pawing at, sniffing it, or nudging it. Sniffing involves a dog moving its nostrils while holding their muzzle still and in very close proximity (< 5cm) to the object.
Food-Directed Behaviours
Behaviour Behaviour Description
Moving Towards Bowl Distance between dog and food bowl decreases, as dog moves with their head oriented towards the food bowl. It should be marked from the moment the dog lifts their paw to make the movement towards the food / food bowl.
Looking at Bowl Dog faces the food bowl and looks at it for >1 second.
Sitting/Lying by Bowl Dog changes their position to sitting by or laying down by the food bowl.
Interacting with Bowl Dog intentionally physically interacts with empty food bowl: namely by scratching at, pawing at, sniffing it, or nudging it. Sniffing involves a dog moving its nostrils while holding their muzzle still and in very close proximity (< 5cm) to the object.

To determine whether dogs responded as expected to E1’s button presses (in Experiment 1) and the owners’ button presses (in Experiment 2), we investigated whether their behavioural responses to button presses were consistent with individual buttons’ words. For example, upon observing a button press for “OUTSIDE”, a dog should be more likely to move toward the door than upon observing a press for “FOOD”.

The data was analysed using Bayesian linear mixed-effects models implemented in brms [21] in R [22]. Specifically, we used a Bernoulli model which requires the dependent variable to be binary. In our case, we ran three separate Bernoulli models, one for each type of behaviour (food-directed behaviours, outside-directed behaviours, and play-directed behaviours). Our reasoning was that if dogs have created associations between these words and their meanings, then they should show the appropriate behaviour associated with that button in the condition.

The dependent variable for each model was whether, on any given trial, the dog showed the target behaviour. For example, in the model for the FOOD condition, the dependent variable was 1 for a given trial if the dog showed a food-related behaviour, and 0 if not (where “trial” here refers to a coded behaviour, so each food-related behaviour would be coded 1 in the FOOD model, and each non-food behaviour would be coded 0). A complete model description can be found in the analysis script (which is provided in full at https://github.com/znhoughton/comprehension).

The independent variables were the three conditions (Food Condition, Outside Condition, or Play Condition), Experiment (IPS or CSS), and the interaction between the two, with maximal random effects [23]. For each model, we used weakly informative priors. The model syntax for each model is included below:

  • Model 1: Play behaviours ~ Condition*Experiment + (1 + Condition|Subject)

  • Model 2: Outside behaviours ~ Condition*Experiment + (1 + Condition|Subject)

  • Model 3: Food behaviours ~ Condition*Experiment + (1 + Condition|Subject)

Next, using only data from Experiment 2 (CSS), we tested whether dogs behave comparably in response to a spoken word and a word produced by a button press. To do so, we once again used three Bayesian logistic regression models implemented in brms in R, the same dependent variable as above, but with the fixed-effects of condition and mode (i.e., spoken or pressed; and their interaction), random slopes for condition by subject. This is described in the following three equations:

  • Model 4: Food behaviours ~ Condition*Mode + (1 + Condition*Mode|Subject)

  • Model 5: Outside behaviours ~ Condition*Mode + (1 + Condition*Mode|Subject)

  • Model 6: Play behaviours ~ Condition*Mode + (1 + Condition*Mode|Subject)

Finally, given recent work to suggest head-tilting is a lateralized behaviour which occurs when dogs recognise and process known words [18], we planned to compare the number of head tilts (to both sides first, then separately for the right and left sides) dogs made in the nonce (“Daxing”) condition compared to the three meaningful known-word conditions (Outside, Play, and Food conditions) in both the researcher-led experiment and the citizen science component of this study. In order to assess this, we set out to use a Bayesian linear mixed-effects model with a Poisson distribution implemented in brms in R, with number of head tilts as the dependent variable and condition and context (known word vs. nonce word) as well as their interaction as fixed-effects and random intercepts for subject and random slopes for condition by subject. Our model specification is listed in the following equation: Number of head tilts ~ Condition * Experiment + (1|Subject) + (Condition|Subject).

All analyses were pre-registered in the Open Science Framework (https://osf.io/tskcq).

Results

Question 1: Can dogs recognise and respond appropriate to the words recorded on their soundboards?

Across both experiments, dogs exhibited approximately seven times more play-directed behaviours in the Play Condition (Table 4), and approximately seven times more outside-directed behaviours in the Outside Condition (Table 5), suggesting that they recognised and responded appropriately to these two words. We found no conclusive evidence to suggest that dogs exhibited food-directed behaviours in the Food Condition compared to the other two conditions (Table 6). Note that, since all models used sum coding, the intercept in all tables represents the grand mean, and the coefficient values represent the distance, in log-odds, between the effect and the intercept. Dogs’ behaviours across the three familiar-word conditions of both experiments are shown in Fig 2.

Table 4. Output for Model 1: Play behaviours ~ Condition*Experiment + (1 + Condition|Subject).

In the Play Condition, the odds of dogs displaying play-directed behaviours was approximately seven times greater than average across the three meaningful-word conditions.

Estimate Est.Error Q2.5 Q97.5
Intercept -1.771 0.953 -3.746 -0.021
Play Condition 1.968 0.943 0.314 4.026
Food Condition 0.099 0.889 -1.695 1.828
Outside Condition -2.067 1.355 -5.104 0.270
Experiment CSS 0.555 0.816 -0.994 2.277
Play:CSS -0.758 0.848 -2.591 0.744
Food:CSS -0.157 0.794 -1.740 1.381

Table 5. Output for Model 2: Outside behaviours ~ Condition*Experiment + (1 + Condition|Subject).

In the Outside Condition, the odds of dogs displaying outside-directed behaviours was approximately seven times greater than average across the three meaningful-word conditions.

Estimate Est.Error Q2.5 Q97.5
Intercept -1.725 0.875 -3.572 -0.061
Outside Condition 1.999 0.852 0.423 3.806
Food Condition -1.513 0.839 -3.326 0.029
Play Condition -0.485 0.812 -2.145 1.039
Experiment CSS -1.906 0.780 -3.583 -0.503
Outside:CSS -0.030 0.865 -1.885 1.563
Food:CSS -0.220 0.790 -1.811 1.290

Table 6. Output for Model 3: Food behaviours ~ Condition*Experiment + (1 + Condition|Subject).

Dogs were not more likely to perform food-directed behaviours in the Food Condition compared to Play or Outside Conditions.

Estimate Est.Error Q2.5 Q97.5
Intercept -4.110 0.823 -5.943 -2.715
Food Condition 1.279 0.828 -0.260 3.008
Outside Condition -0.040 0.977 -2.134 1.778
Play Condition -1.239 1.122 -3.698 0.703
Experiment CSS -0.199 0.557 -1.311 0.880
Food:CSS 0.050 0.609 -1.132 1.296
Outside:CSS -0.303 0.746 -1.883 1.082

Fig 2. The plot of the average proportion of each behaviour type in each condition.

Fig 2

Proportion here is the number of behaviours of a specific type (e.g., outside behaviours) in a given condition (e.g., play condition) divided by the total number of behaviours in that condition. The x-axis is Condition. The y-axis is the average proportion of each behaviour type across participants. Each of the points represents the proportion of behaviours of the corresponding type of behaviour in the respective condition. The error bars represent standard errors.

Question 2: Do dogs respond equivalently to words produced by their owner compared to an unfamiliar person?

We found no effect of button presser identity (owner or unfamiliar person) on dog’s behaviours for any of the three conditions (see confidence intervals overlapping zero in Tables 57 for IPS and CSS comparisons). This suggests that dogs respond appropriately to button presses even in the absence of other contextual cues or owner-produced cues.

Table 7. Output for Model 4: Food behaviours ~ Condition*Mode + (1 + Condition*Mode|Subject).

“Play:Press”, “Food:Press” and “Outside:Press” correspond to the interactions between each condition and mode. For example, the effect of “Play:Press” is the extent to which the effect of “Play” differs when the dog hears a recording from a button press as opposed to a spoken word.

Estimate Est.Error Q2.5 Q97.5
Intercept -5.749 1.373 -8.740 -3.494
Play 1.680 1.299 -0.855 4.388
Food 0.315 1.262 -2.269 2.759
Press 0.261 0.775 -1.137 1.915
Play:Press -1.155 1.059 -3.377 0.846
Food:Press -0.417 1.136 -2.687 1.890

Question 3: Can citizen science studies and in-person studies of soundboard-trained dogs produce comparable results?

We found no difference in dogs’ behaviours in response to owner-produced and experimenter-produced button presses (see Tables 46), and therefore the results for both experiments were comparable. In the present investigation, dogs behaved similarly in response to button presses in the in-person study (IPS) and the citizen science study (CSS), suggesting that both experiments produced equivalent results.

Question 4: Do soundboard-trained dogs attend to speech?

Within Experiment 2 (IPS), dogs’ responses to spoken and pressed words across the three conditions were comparable for food-related behaviours (Table 7), outside-related behaviours (Table 8), and play-related behaviours (Table 9). Overall, these results suggest that the two modes of word-production led to equivalent behavioural responses by subjects.

Table 8. Output for Model 5: Outside behaviours ~ Condition*Mode + (1 + Condition*Mode|Subject).

“Play:Press”, “Food:Press” and “Outside:Press” correspond to the interactions between each condition and mode. For example, the effect of “Play:Press” is the extent to which the effect of “Play” differs when the dog hears a recording from a button press as opposed to a spoken word.

Estimate Est.Error Q2.5 Q97.5
Intercept -5.978 1.941 -10.253 -2.603
Outside 3.251 1.436 0.580 6.307
Food -2.180 1.539 -5.391 0.741
Press 0.245 1.178 -2.034 2.618
Outside:Press -0.328 1.495 -3.386 2.599
Food:Press 2.842 1.563 0.197 6.344

Table 9. Output for Model 6: Play behaviours ~ Condition*Mode + (1 + Condition*Mode|Subject).

“Play:Press”, “Food:Press” and “Outside:Press” correspond to the interactions between each condition and mode. For example, the effect of “Play:Press” is the extent to which the effect of “Play” differs when the dog hears a recording from a button press as opposed to a spoken word.

Estimate Est.Error Q2.5 Q97.5
Intercept -1.027 1.317 -3.754 1.482
Food 2.154 1.170 0.085 4.699
Outside -0.188 0.949 -2.139 1.710
Press 1.275 0.711 -0.044 2.755
Food:Press -0.437 0.775 -2.020 1.099
Outside:Press -0.784 0.737 -2.263 0.625

Question 5: Do soundboard-trained dogs tilt their heads in response to familiar words? Ultimately, our dataset contained too few head-tilting observations (9 instances) for this analysis to be completed, so we cannot ascertain whether head-tilting was more common in meaningful known-word conditions (Play, Food, and Outside Conditions) compared to the nonce word condition (Daxing Condition).

Discussion

Our study suggests that dogs were more likely to perform play-related behaviours after an experimenter or their owner produced a play-related word, and were more likely to exhibit outside-related behaviours in response to an experimenter or their owner producing an outside-related word. This demonstrates that dogs are, at the very least, capable of learning an association between these words or buttons and their outcomes in the world. On the other hand, dogs did not produce preferentially food-related behaviours in response to the relevant words. It is possible that this may have occurred because several of the dogs taking part in the study may have been satiated before the start of test trials, or because dogs did not expect that they would be served a meal outside of their usual feeding times. Our in-person study was typically conducted during work hours, which do not typically overlap with adult dogs’ mealtimes in the early mornings and/or evenings.

Given that dogs responded equivalently to their owners’ button presses and an unfamiliar person’s button presses, our results also demonstrate that dogs attend to and respond to the buttons or words themselves, rather than behaving solely based on unrelated unintentional cues provided by their owners (e.g., as would be expected from a Clever Hans effect). In both experiments, even when all word-related contextual cues were removed from the human’s interaction with the AIC device, dogs still responded with contextually appropriate behaviours. For example, if dogs typically observe their owners picking up a toy before they press a button for a play-related word, any play-related behaviours the dogs exhibit in turn might occur in response to the sight of the toy, the owner’s button press, or a combination of both events. Since our citizen science experiment specifically required that owners press buttons without performing any other actions, the fact that dogs still displayed contextually appropriate responses in the absence of other cues demonstrates that they attended specifically to the words recorded onto their buttons. Additionally, dogs responded appropriately to an unfamiliar human making button presses, even though that person was unaware of which button they were pressing.

Both word sounds, and the location of their respective buttons on dogs’ soundboards, are highly correlated with their effects in the world. For example, when owners press the button OUTSIDE, dogs might equally attend to the location of that button on their AIC device, or the sound (word) produced by the button press. Our citizen science study addressed this question by comparing dogs’ responses to their owners’ speech and button presses. We found no differences in dogs’ responses to either mode of word production, suggesting that most dogs do not associate button location alone with the outcomes of button presses. If they did, they should perform considerably more context-appropriate behaviours in response to the relevant button presses than to owners’ spoken words. We note that, although this does not necessarily preclude that dogs may have formed two separate associations–one for the button location and another for the speech sound of the word–, it nevertheless demonstrates that most dogs can and do attend to and appropriately respond to the auditory properties of words.

Crucially, we found no differences in dogs’ behaviours across the two experiments, suggesting that owners’ and researchers’ conduct of the methods was sufficiently equivalent to yield comparable results. This suggests that results from the citizen science version of the study were comparable to those performed by researchers during visits to owners’ homes. This is important because soundboard-trained dogs are spread all over the world and we do not yet know how moving their soundboards outside of their home environment might affect their soundboard use [2]. Our findings offer a promising outlook for future citizen science studies with this population of owners and their dogs: remotely conducted citizen science studies could be a critical tool for studying this large and geographically widespread population and maintaining long-term engagement of pet owners with this research program. However, we advise that other future studies should again validate citizen science methods by comparing them against researcher-led in-person experiments, therefore replicating our findings, before scientists can rely more heavily on citizen science for this and similar studies. Additionally, we highlight that all procedures and analyses of this study were pre-registered in advance of data collection, and that future studies extending on this work would also benefit from pre-registration. Pre-registration offers an important tool for ensuring transparency and reproducibility of research, and its use is critical to large scale exploratory studies and citizen science studies.

In sum, our findings provide the first evidence of button word comprehension by owner-trained soundboard-using dogs, and demonstrate that dogs’ contextually appropriate responses to button presses were comparable regardless of the identity of the person using the soundboard, and the absence of other environmental cues related to that word. Our findings also suggest that dogs attend to the sounds recorded onto their buttons, given that they responded equivalently to words when they were produced by button presses and when they were spoken by their owners. A summary of our findings is tabulated in Table 10 below.

Table 10. Summary of the conclusions drawn for each of the five questions addressed in the present study.

Research Question Results of Study
1. Can dogs recognise and respond appropriate to the words recorded on their soundboards? Yes; dogs demonstrated context-appropriate responses to two of the three word categories tested, namely “play” and “outside”. However, we found no evidence for context-appropriate responses to “food”-related words.
2. Do dogs respond equivalently to words produced by their owner compared to an unfamiliar person? Yes; we did not find any differences in dogs’ responses to button presses produced by their owners compared to an unfamiliar person.
3. Can citizen science studies and in-person studies of soundboard-trained dogs produce equivalent results? Yes; in this case, the citizen science study yielded comparable results to the in-person study carried out by researchers.
4. Do soundboard-trained dogs attend to speech? Yes; dogs responded equivalently to their owners pressing words on their soundboards compared to their owners speaking these same words.
5. Do soundboard-trained dogs tilt their heads in response to familiar words? Inconclusive; we did not observe enough head tilting in our dataset to investigate this question.

In our study, dogs responded to spoken or pressed food-related and outside-related words with contextually appropriate responses slightly more often than expected by chance. The accuracy of dog’s responses in our study is comparable to dogs’ accuracy in responding to human pointing [24]. However, dogs are capable of much greater accuracy when purposely trained to respond to stimuli, as is the case for dogs trained to detect wildlife [25], individual people [26], and chemical substances [27, 28] through scent. This discrepancy in performance could be due to the nature of the stimuli: perhaps dogs find it more difficult to form associations between words and their respective outcomes compared to scents and a consistent reward. Alternately, communicative contexts may produce less predictable responses due to the weaker correlation between the perception of relevant stimuli (a pointed finger, or a word) and their outcomes, therefore leading to less strongly conditioned responses to the stimuli. While scent detection dogs undergoing training will likely experience reinforcement almost every single time the target stimulus is present, a pet dog may not receive any reinforcement on a great number of occasions when they hear familiar words or observe human pointing. Relatedly, scent detection dogs’ accuracy is much higher in contexts where target odours are present at very high rates, with implications for their performance in real-world scenarios [29]. Therefore, it is also possible that task-oriented trained contexts, such as scent detection, are much more motivating to dogs, and therefore more likely to trigger consistent responses, than day-to-day contexts involving more variable reinforcement. Further, measures of performance in working dogs such as scent detection dogs are typically based on a small subset of the population that passes stringent standards of training [30], and typically involves animals that are selectively bred for the purpose of scent detection and further selected based on temperament or cognitive traits [3032], whereas the soundboard-trained population comprises owner-trained pet dogs, whose temperaments and cognitive traits are likely to vary considerably.

Having established that soundboard-trained dogs can and do attend to and comprehend words, future work should also disambiguate the extent to which spatial information about button positions, or other potential cues for button identity, might aid dogs’ ability to use AIC devices. Additionally, more research is needed investigating soundboard-trained dogs’ responses to a wider range of words, particularly in comparison to a population of non-soundboard-trained pet dogs. Although owners anecdotally report that owner-trained pet dogs spontaneously acquire comprehension of large spoken vocabularies [9], there are no fully controlled experiments investigating whether dogs exhibit contextually appropriate spontaneous responses to familiar words in the absence of other contextual cues, as in the present study. This is crucial because, although owners may, for example, report that their dogs respond appropriately to a food-related word, this word is typically presented alongside a myriad of other confounding cues, such as the time of day when the dog’s meals are served, the presence of a bowl, or the behaviours their owner might perform before serving their dog’s dinner. Finally, our ongoing work is investigating dogs’ word production [2]. In order to determine whether dogs’ performance at word comprehension is reflected in their button pressing, carefully controlled future studies must investigate whether dogs can spontaneously produce contextually appropriate button presses in experimentally induced situations (as in [4]). Not only would such a study be helpful in understanding the depth of soundboard-trained dogs’ word comprehension, but it would also establish the extent to which AIC devices can be used for two-way interspecies communication, involving both presses made by the owner for their dog, and by the dog for their owner.

Acknowledgments

We are grateful to all our dog participants and their owners for contributing their time and data to our research project.

Data Availability

Study pre-registration is deposited in the Open Science Framework, https://osf.io/tskcq Data and analysis scripts are available from Github, https://github.com/znhoughton/comprehension.

Funding Statement

A.P.M.B. is supported by the Johns Hopkins Provost’s Postdoctoral Fellowship Program.

References

  • 1.Smith GE, Bastos APM, Evenson A, Trottier L, Rossano F. Use of Augmentative Interspecies Communication devices in animal language studies: A review. WIREs Cognitive Science. 2023;14: e1647. doi: 10.1002/wcs.1647 [DOI] [PubMed] [Google Scholar]
  • 2.Bastos APM, Rossano F. Soundboard-using pets? Introducing a new global citizen science approach to interspecies communication. Interaction Studies. 2023;24: 311–334. doi: 10.1075/is.22050.pin [DOI] [Google Scholar]
  • 3.Hunger C. How Stella Learned to Talk: The Groundbreaking Story of the World’s First Talking Dog. New York, NY, US: William Morrow; 2021. [Google Scholar]
  • 4.Rossi AP, Ades C. A dog at the keyboard: using arbitrary signs to communicate requests. Anim Cogn. 2008;11: 329–338. doi: 10.1007/s10071-007-0122-3 [DOI] [PubMed] [Google Scholar]
  • 5.Ramos D, Ades C. Two-item sentence comprehension by a dog (Canis familiaris). PLOS ONE. 2012;7: e29689. doi: 10.1371/journal.pone.0029689 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Pfungst O. Clever Hans (the horse of Mr. Von Osten): A contribution to experimental animal and human psychology. Holt, Rinehart and Winston; 1911. doi: 10.5962/bhl.title.56164 [DOI] [Google Scholar]
  • 7.Sebeok TA, Rosenthal R, editors. The Clever Hans phenomenon: Communication with horses, whales, apes, and people. Annals of the New York Academy of Sciences. 1981;364: 309–309. [PubMed] [Google Scholar]
  • 8.Eckstein G. Concerning a Dog’s Word Comprehension. Science. 1949;109: 494. doi: 10.1126/science.109.2837.494 [DOI] [PubMed] [Google Scholar]
  • 9.Reeve C, Jacques S. Responses to spoken words by domestic dogs: A new instrument for use with dog owners. Applied Animal Behaviour Science. 2022;246: 105513. doi: 10.1016/j.applanim.2021.105513 [DOI] [Google Scholar]
  • 10.Kaminski J, Call J, Fischer J. Word Learning in a Domestic Dog: Evidence for “Fast Mapping.” Science. 2004;304: 1682–1683. doi: 10.1126/science.1097859 [DOI] [PubMed] [Google Scholar]
  • 11.Pilley JW, Reid AK. Border collie comprehends object names as verbal referents. Behav Processes. 2011;86: 184–195. doi: 10.1016/j.beproc.2010.11.007 [DOI] [PubMed] [Google Scholar]
  • 12.Griebel U, Oller DK. Vocabulary learning in a Yorkshire terrier: slow mapping of spoken words. PLoS One. 2012;7: e30182. doi: 10.1371/journal.pone.0030182 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Dror S, Miklósi Á, Sommese A, Temesi A, Fugazza C. Acquisition and long-term memory of object names in a sample of Gifted Word Learner dogs. Royal Society Open Science. 2021;8: 210976. doi: 10.1098/rsos.210976 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Dror S, Sommese A, Miklósi Á, Temesi A, Fugazza C. Multisensory mental representation of objects in typical and Gifted Word Learner dogs. Anim Cogn. 2022;25: 1557–1566. doi: 10.1007/s10071-022-01639-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Fugazza C, Dror S, Sommese A, Temesi A, Miklósi Á. Word learning dogs (Canis familiaris) provide an animal model for studying exceptional performance. Sci Rep. 2021;11: 14070. doi: 10.1038/s41598-021-93581-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Boros M, Magyari L, Morvai B, Hernández-Pérez R, Dror S, Andics A. Neural evidence for referential understanding of object words in dogs. Current Biology. 2024;0. doi: 10.1016/j.cub.2024.02.029 [DOI] [PubMed] [Google Scholar]
  • 17.Kőszegi H, Fugazza C, Magyari L, Iotchev IB, Miklósi Á, Andics A. Investigating responses to object-labels in the domestic dog (Canis familiaris). Sci Rep. 2023;13: 3150. doi: 10.1038/s41598-023-30201-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Ramos D, Mills DS. Limitations in the learning of verbal content by dogs during the training of OBJECT and ACTION commands. Journal of Veterinary Behavior. 2019;31: 92–99. doi: 10.1016/j.jveb.2019.03.011 [DOI] [Google Scholar]
  • 19.Stewart L, MacLean EL, Ivy D, Woods V, Cohen E, Rodriguez K, et al. Citizen Science as a New Tool in Dog Cognition Research. PLOS ONE. 2015;10: e0135176. doi: 10.1371/journal.pone.0135176 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Sommese A, Miklósi Á, Pogány Á, Temesi A, Dror S, Fugazza C. An exploratory analysis of head-tilting in dogs. Anim Cogn. 2022;25: 701–705. doi: 10.1007/s10071-021-01571-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Bürkner P-C. brms: An R Package for Bayesian Multilevel Models Using Stan. J Stat Soft. 2017;80. doi: 10.18637/jss.v080.i01 [DOI] [Google Scholar]
  • 22.R Core Team R. R: A language and environment for statistical computing. 2013. [Google Scholar]
  • 23.Barr DJ, Levy R, Scheepers C, Tily HJ. Random effects structure for confirmatory hypothesis testing: Keep it maximal. J Mem Lang. 2013;68: 10.1016/j.jml.2012.11.001. doi: 10.1016/j.jml.2012.11.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.ManyDogs, Espinosa J, Stevens JR, Alberghina D, Barela J, Bogese M, et al. ManyDogs 1: A multi-lab replication study of dogs’ pointing comprehension. 2023. doi: 10.31234/osf.io/4kncr [DOI] [Google Scholar]
  • 25.McKeague B, Finlay C, Rooney N. Conservation detection dogs: A critical review of efficacy and methodology. Ecol Evol. 2024;14: e10866. doi: 10.1002/ece3.10866 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Marchal S, Bregeras O, Puaux D, Gervais R, Ferry B. Rigorous Training of Dogs Leads to High Accuracy in Human Scent Matching-To-Sample Performance. Egles C, editor. PLoS ONE. 2016;11: e0146963. doi: 10.1371/journal.pone.0146963 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Jezierski T, Adamkiewicz E, Walczak M, Sobczyńska M, Górecka-Bruzda A, Ensminger J, et al. Efficacy of drug detection by fully-trained police dogs varies by breed, training level, type of drug and search environment. Forensic Science International. 2014;237: 112–118. doi: 10.1016/j.forsciint.2014.01.013 [DOI] [PubMed] [Google Scholar]
  • 28.Furton KG, Myers LJ. The scientific foundation and efficacy of the use of canines as chemical detectors for explosives1. Talanta. 2001;54: 487–500. doi: 10.1016/S0039-9140(00)00546-4 [DOI] [PubMed] [Google Scholar]
  • 29.DeChant MT, Aviles-Rosa E, Prada-Tiedemann P, Hall NJ. Part, III: Increasing odor detection performance after training with progressively leaner schedules of odor prevalence. Journal of the Experimental Analysis of Behavior. 2023;120: 137–152. doi: 10.1002/jeab.841 [DOI] [PubMed] [Google Scholar]
  • 30.Troisi CA, Mills DS, Wilkinson A, Zulch HE. Behavioral and Cognitive Factors That Affect the Success of Scent Detection Dogs. CCBR. 2019;14: 51–76. doi: 10.3819/CCBR.2019.140007 [DOI] [Google Scholar]
  • 31.MacLean EL, Hare B. Enhanced Selection of Assistance and Explosive Detection Dogs Using Cognitive Measures. Front Vet Sci. 2018;5. doi: 10.3389/fvets.2018.00236 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Lazarowski L, Thompkins A, Krichbaum S, Waggoner LP, Deshpande G, Katz JS. Comparing pet and detection dogs (Canis familiaris) on two aspects of social cognition. Learn Behav. 2020;48: 432–443. doi: 10.3758/s13420-020-00431-8 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Brenton G Cooper

30 May 2024

PONE-D-24-15704How do soundboard-using dogs respond to human button presses? An investigation into label comprehensionPLOS ONE

Dear Dr. Bastos,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.  Please carefully address the reviewer's concerns regarding the presentation of your results and conclusions of your study in your revision.  This includes potentially altering the title of your manuscript as you revise based on the thoughtful critique provided by the reviewer. 

Please submit your revised manuscript by Jul 14 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Brenton G. Cooper, Ph.D.

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. Thank you for stating the following in the Funding Section of your manuscript:

“A.P.M.B. is supported by the Johns Hopkins Provost’s Postdoctoral Fellowship Program.”

We note that you have provided funding information that is currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

“The author(s) received no specific funding for this work.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. Thank you for stating the following in the Competing Interests section:

“A.P.M.B., P.M.W., Z.N.H., J. T., G.E.S., & L.K. have previously consulted for FluentPet, Inc., a company that produces AIC devices for pets. A.E. & L.N. are employees of FluentPet, Inc.”

Please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials, by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests).  If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include your updated Competing Interests statement in your cover letter; we will change the online submission form on your behalf.

5. Your ethics statement should only appear in the Methods section of your manuscript. If your ethics statement is written in any section besides the Methods, please move it to the Methods section and delete it from any other section. Please ensure that your ethics statement is included in your manuscript, as the ethics statement entered into the online submission form will not be published alongside your manuscript.

6. Please remove your figures from within your manuscript file, leaving only the individual TIFF/EPS image files, uploaded separately. These will be automatically included in the reviewers’ PDF.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Bastos et al.

How do soundboard-using dogs respond to human button presses?...

May 14, 2024

At its heart this is a very simple paper, but, because of the context of this work, it is likely to excite a lot of attention. Much of it misguided. Consequently, a great deal of care needs to be taken in its presentation. A lot of care has already been taken. I’m going to suggest some additional care.

In essence, Bastos et al. tested pet dogs for their ability to respond appropriately to three words: “Out,” “play,” and “food.” (The exact word used depended on the dog’s prior training). “Respond appropriately” means here, move towards the room exit in response to “out,” move towards toys in response to “play,” and move towards food in response to “food.” The authors found above chance tendencies to move towards the exit and toys in response to “out,” and “play” respectively. There was no tendency to move towards food in response to “food.” These tendencies to respond appropriately to “out” and “play” were not massive effects. On average the dogs showed about 50% appropriate behaviors in response to the relevant word. So a detectable tendency but far from the levels expected, say, of explosive detection dogs.

The study stands out for extreme care in design and experimental controls. It was preregistered, every imaginable attempt was made to blind the experimenters and video coders to what was going on. So the risks of Clever-Hans-type unintentional cueing are very small. The data analysis also appears careful and fully appropriate.

Thus far this might seem like a very minor study. What will surely lead to it attracting a great deal of attention is the fact that the dogs that were tested had all been trained (by their owners) on the button systems that are currently very popular in social media and elsewhere. These are buttons programmed so that, when pressed, they say a word recorded by the dog’s owner. The widespread interest in these buttons stems from the fact that people train their dogs to press them, however, in this study, the dogs were not given an opportunity to press the buttons. All button presses were carried out by humans. There were two sets of conditions: one where the owner or experimenter said the relevant word; the other where the owner or experimenter generated the word by pressing the button that activated a prior recording of the owner saying the word. No significant differences were found between these conditions. Furthermore, there is no basis to believe that the fact that the dogs had all been trained to press these buttons themselves has anything to do with their performance. It wouldn’t surprise me if most pet dogs respond 50% appropriately to someone saying “out” or “play.” In fact, I’m rather surprised that the dogs in this study didn’t respond correctly to a word standing for “food.” My dog does and we haven’t had to train her explicitly to respond to “din dins!”

There were also a further two sets of conditions, one in which the dogs were tested by experimenters, and one where the dogs were tested by their owners. No difference was found between these conditions.

I think this is an interesting little paper, but, as I started out by stating, its presentation and conclusions need to be carefully expressed to avoid the possibility of misunderstandings.

First, I think the fact that these dogs had been trained to press “talking” buttons should be downplayed. In this study the dogs were not pressing the buttons. There are no comparative data to indicate that the prior button-press-training has anything to do with the performance observed, and the generally low (albeit statistically significantly above chance) levels of performance provide no grounds for believing that the prior training had had any positive effect.

Figure 2 should be redrawn to cover the whole range over which values were free to vary (0 – 100%). The present truncated formatting gives a visual impression of a more powerful effect than was actually observed.

I wonder about including “soundboard-using” in the title. As noted above, there is no evidence that having been trained on the soundboard impacted the dogs’ performance.

Tables 8-10 lack full explanations. What are “Play:Press” and “Food:Press” etc.?

I’m not sure why, in the Discussion, the authors stated that they possibly failed to detect a respond to the word “food” because food-related behaviors occurred at a high rate in all conditions. Figure 2 indicates the opposite; Food related behaviors occur at the lowest rate of any analyzed behaviors in all three conditions.

I recognize that this isn’t an error, but I think it would be simpler to refer to the stimuli as ‘words” rather than “labels.” That will be easier for more readers to understand.

I would like to see consideration in the Discussion of other contexts in which dogs respond to stimuli. This might include scenarios like sniffer dogs, which are trained explicitly and to far higher criteria than were observed here, and possibly other everyday examples of dogs developing appropriate responses in daily interaction with humans, such as following pointing gestures.

Overall a worthy piece of work that ought to be published after some revision.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Aug 28;19(8):e0307189. doi: 10.1371/journal.pone.0307189.r002

Author response to Decision Letter 0


10 Jun 2024

Reviewer Comment: At its heart this is a very simple paper, but, because of the context of this work, it is likely to excite a lot of attention. Much of it misguided. Consequently, a great deal of care needs to be taken in its presentation. A lot of care has already been taken.

I’m going to suggest some additional care. In essence, Bastos et al. tested pet dogs for their ability to respond appropriately to three words: “Out,” “play,” and “food.” (The exact word used depended on the dog’s prior training). “Respond appropriately” means here, move towards the room exit in response to “out,” move towards toys in response to “play,” and move towards food in response to “food.” The authors found above chance tendencies to move towards the exit and toys in response to “out,” and “play” respectively. There was no tendency to move towards food in response to “food.” These tendencies to respond appropriately to “out” and “play” were not massive effects. On average the dogs showed about 50% appropriate behaviors in response to the relevant word. So a detectable tendency but far from the levels expected, say, of explosive detection dogs.

The study stands out for extreme care in design and experimental controls. It was preregistered, every imaginable attempt was made to blind the experimenters and video coders to what was going on. So the risks of Clever-Hans-type unintentional cueing are very small. The data analysis also appears careful and fully appropriate.

Author Response: We thank the reviewer for their positive comments.

Reviewer Comment: Thus far this might seem like a very minor study. What will surely lead to it attracting a great deal of attention is the fact that the dogs that were tested had all been trained (by their owners) on the button systems that are currently very popular in social media and elsewhere. These are buttons programmed so that, when pressed, they say a word recorded by the dog’s owner. The widespread interest in these buttons stems from the fact that people train their dogs to press them, however, in this study, the dogs were not given an opportunity to press the buttons. All button presses were carried out by humans. There were two sets of conditions: one where the owner or experimenter said the relevant word; the other where the owner or experimenter generated the word by pressing the button that activated a prior recording of the owner saying the word. No significant differences were found between these conditions. Furthermore, there is no basis to believe that the fact that the dogs had all been trained to press these buttons themselves has anything to do with their performance.

Author Response: The reviewer is correct in that this is a study of comprehension rather than production: at no point did the dogs have to press any buttons. We agree that some non-soundboard-trained dogs also likely recognise certain words (e.g., “food”), but we cannot at this point generalise this finding to all pet dogs. Soundboard-trained dogs experience human speech directed at them very regularly. This may not be true of many pet dogs (e.g., particularly those living outside the home in a yard, or in cultures where pets may not be typically spoken to).

Reviewer Comment: It wouldn’t surprise me if most pet dogs respond 50% appropriately to someone saying “out” or “play.” In fact, I’m rather surprised that the dogs in this study didn’t respond correctly to a word standing for “food.” My dog does and we haven’t had to train her explicitly to respond to “din dins!”

Author Response: We thank the reviewer for this excellent example. We have used it to improve our discussion, raising the points the reviewer made about comparing soundboard-trained dogs and non-soundboard-trained dogs with regards to their word comprehension: “Additionally, more research is needed investigating soundboard-trained dogs’ responses to a wider range of words, particularly in comparison to a population of non-soundboard-trained pet dogs. Although owners anecdotally report that owner-trained pet dogs spontaneously acquire comprehension of large spoken vocabularies [9], there are no fully controlled experiments investigating whether dogs exhibit contextually appropriate spontaneous responses to familiar words in the absence of other contextual cues, as in the present study. This is crucial because, although owners may, for example, report that their dogs respond appropriately to a food-related word, this word is typically presented alongside a myriad of other confounding cues, such as the time of day when the dog’s meals are served, the presence of a bowl, or the behaviours their owner might perform before serving their dog’s dinner. Finally, our ongoing work is investigating dogs’ word production [2]. In order to determine whether dogs’ performance at word comprehension is reflected in their button pressing, carefully controlled future studies must investigate whether dogs can spontaneously produce contextually appropriate button presses in experimentally induced situations (as in [4]).” (L520-534)

Reviewer Comment: There were also a further two sets of conditions, one in which the dogs were tested by experimenters, and one where the dogs were tested by their owners. No difference was found between these conditions.

I think this is an interesting little paper, but, as I started out by stating, its presentation and conclusions need to be carefully expressed to avoid the possibility of misunderstandings.

Author Response: We agree with the reviewer and thank them for their caution. We also believe this work requires careful presentation.

Reviewer Comment: First, I think the fact that these dogs had been trained to press “talking” buttons should be downplayed. In this study the dogs were not pressing the buttons. There are no comparative data to indicate that the prior button-press-training has anything to do with the performance observed, and the generally low (albeit statistically significantly above chance) levels of performance provide no grounds for believing that the prior training had had any positive effect.

Author Response: Although we agree that we cannot provide a direct comparison, it is very likely that owners who train their dogs to press buttons also direct a great deal of speech to their dogs – probably more so than the average pet owner. As such, these dogs may experience far more opportunities in which to learn the associations between words and their outcomes, compared to the average pet dog. We also cannot make any claims about how non-soundboard-trained pet dogs might react to a word – even if it is a familiar word – being produced by a button containing a voice recording. The training experienced by soundboard-using dogs at the very least habituates them to this context and enables them to respond to the voice recordings on buttons equivalently to speech.

Reviewer Comment: Figure 2 should be redrawn to cover the whole range over which values were free to vary (0 – 100%). The present truncated formatting gives a visual impression of a more powerful effect than was actually observed.

Author Response: We have now redrawn Figure 2 as per the reviewer’s comments.

Reviewer Comment: I wonder about including “soundboard-using” in the title. As noted above, there is no evidence that having been trained on the soundboard impacted the dogs’ performance.

Author Response: In line with the reviewer’s comment, we have changed “soundboard-using” to “soundboard-trained” in the title and at several points throughout the manuscript, as appropriate. Although there is no evidence that being trained on a soundboard increases a dog’s ability to learn the associations between words and their outcomes, it is quite likely that dogs whose owners train them to press buttons experience two considerably different sets of circumstances compared to the average pet dog: (1) they experience many more opportunities for word-to-outcome associative learning, because their owners regularly direct speech at them – and anecdotally we know from owner reports that this speech is simplified, with particular features such as exaggerated pronunciation and repetitions (this is a topic for future investigation); (2) they are additionally at the very least habituated to, if not directly associatively learning from, the voice recordings being produced by buttons. As such, it would be precipitous for us to generalise our findings to all pet dogs, regardless of training history.

Reviewer Comment: Tables 8-10 lack full explanations. What are “Play:Press” and “Food:Press” etc.?

Author Response: We have now added the following description to the three tables’ descriptive captions, to clarify these rows: ““Play:Press”, “Food:Press” and “Outside:Press” correspond to the interactions between each condition and mode. For example, the effect of “Play:Press” is the extent to which the effect of “Play” differs when the dog hears a recording from a button press as opposed to a spoken word.”

Reviewer Comment: I’m not sure why, in the Discussion, the authors stated that they possibly failed to detect a respond to the word “food” because food-related behaviors occurred at a high rate in all conditions. Figure 2 indicates the opposite; Food related behaviors occur at the lowest rate of any analyzed behaviors in all three conditions.

Author Response: The reviewer is correct. We apologise for this mistake and have rephrased our discussion accordingly: “It is possible that this may have occurred because several of the dogs taking part in the study may have been satiated before the start of test trials, or because dogs did not expect that they would be served a meal outside of their usual feeding times. Our in-person study was typically conducted during work hours, which do not typically overlap with adult dogs’ mealtimes in the early mornings and/or evenings.” (L443-447)

Reviewer Comment: I recognize that this isn’t an error, but I think it would be simpler to refer to the stimuli as ‘words” rather than “labels.” That will be easier for more readers to understand.

Author Response: We thank the reviewer for this suggestion. We have now rephrased “labels” as “words” throughout the manuscript.

Reviewer Comment: I would like to see consideration in the Discussion of other contexts in which dogs respond to stimuli. This might include scenarios like sniffer dogs, which are trained explicitly and to far higher criteria than were observed here, and possibly other everyday examples of dogs developing appropriate responses in daily interaction with humans, such as following pointing gestures.

Author Response: We thank the reviewer for this excellent suggestion. We have now included a paragraph on this topic in our discussion: “In our study, dogs responded to spoken or pressed food-related and outside-related words with contextually appropriate responses slightly more often than expected by chance. The accuracy of dog’s responses in our study is comparable to dogs’ accuracy in responding to human pointing [24]. However, dogs are capable of much greater accuracy when purposely trained to respond to stimuli, as is the case for dogs trained to detect wildlife [25], individual people [26], and chemical substances [27,28] through scent. This discrepancy in performance could be due to the nature of the stimuli: perhaps dogs find it more difficult to form associations between words and their respective outcomes compared to scents and a consistent reward. Alternately, communicative contexts may produce less predictable responses due to the weaker correlation between the perception of relevant stimuli (a pointed finger, or a word) and their outcomes, therefore leading to less strongly conditioned responses to the stimuli. While scent detection dogs undergoing training will likely experience reinforcement almost every single time the target stimulus is present, a pet dog may not receive any reinforcement on a great number of occasions when they hear familiar words or observe human pointing. Relatedly, scent detection dogs’ accuracy is much higher in contexts where target odours are present at very high rates, with implications for their performance in real-world scenarios [29]. Therefore, it is also possible that task-oriented trained contexts, such as scent detection, are much more motivating to dogs, and therefore more likely to trigger consistent responses, than day-to-day contexts involving more variable reinforcement. Further, measures of performance in working dogs such as scent detection dogs are typically based on a small subset of the population that passes stringent standards of training [30], and typically involves animals that are selectively bred for the purpose of scent detection and further selected based on temperament or cognitive traits [30–32], whereas the soundboard-trained population comprises owner-trained pet dogs, whose temperaments and cognitive traits are likely to vary considerably..” (L493-516)

Reviewer Comment: Overall a worthy piece of work that ought to be published after some revision.

Author Response: We again thank the reviewer for their positive comments.

Attachment

Submitted filename: Bastos et al. 2024 Reviewer Comments Responses 10-06-24.docx

pone.0307189.s001.docx (21.9KB, docx)

Decision Letter 1

Brenton G Cooper

2 Jul 2024

How do soundboard-trained dogs respond to human button presses? An investigation into word comprehension

PONE-D-24-15704R1

Dear Dr. Bastos,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Brenton G. Cooper, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: My thanks to the authors for taking my comments on board. I think the revised version is an excellent paper and it will be very interesting to see the responses it engenders.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

Acceptance letter

Brenton G Cooper

19 Jul 2024

PONE-D-24-15704R1

PLOS ONE

Dear Dr. Rossano,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Brenton G. Cooper

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Bastos et al. 2024 Reviewer Comments Responses 10-06-24.docx

    pone.0307189.s001.docx (21.9KB, docx)

    Data Availability Statement

    Study pre-registration is deposited in the Open Science Framework, https://osf.io/tskcq Data and analysis scripts are available from Github, https://github.com/znhoughton/comprehension.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES