Skip to main content
The Journal of the Acoustical Society of America logoLink to The Journal of the Acoustical Society of America
. 2022 May 10;151(5):3116–3128. doi: 10.1121/10.0010422

FORUM: Remote testing for psychological and physiological acoustics

Z Ellen Peng 1,a),, Sebastian Waz 2, Emily Buss 3, Yi Shen 4, Virginia Richards 2, Hari Bharadwaj 5, G Christopher Stecker 1, Jordan A Beim 6, Adam K Bosen 1, Meredith D Braza 3, Anna C Diedesch 7, Claire M Dorey 8, Andrew R Dykstra 9, Frederick J Gallun 10, Raymond L Goldsworthy 11, Lincoln Gray 12, Eric C Hoover 13, Antje Ihlefeld 14, Thomas Koelewijn 15, Judy G Kopun 1, Juraj Mesik 6, Daniel E Shub 16, Jonathan H Venezia 17
PMCID: PMC9305596  PMID: 35649891

Abstract

Acoustics research involving human participants typically takes place in specialized laboratory settings. Listening studies, for example, may present controlled sounds using calibrated transducers in sound-attenuating or anechoic chambers. In contrast, remote testing takes place outside of the laboratory in everyday settings (e.g., participants' homes). Remote testing could provide greater access to participants, larger sample sizes, and opportunities to characterize performance in typical listening environments at the cost of reduced control of environmental conditions, less precise calibration, and inconsistency in attentional state and/or response behaviors from relatively smaller sample sizes and unintuitive experimental tasks. The Acoustical Society of America Technical Committee on Psychological and Physiological Acoustics launched the Task Force on Remote Testing (https://tcppasa.org/remotetesting/) in May 2020 with goals of surveying approaches and platforms available to support remote testing and identifying challenges and considerations for prospective investigators. The results of this task force survey were made available online in the form of a set of Wiki pages and summarized in this report. This report outlines the state-of-the-art of remote testing in auditory-related research as of August 2021, which is based on the Wiki and a literature search of papers published in this area since 2020, and provides three case studies to demonstrate feasibility during practice.

I. BACKGROUND AND INTRODUCTION

Research studies that investigate human behaviors have largely been conducted in laboratory environments, which provide researchers with strict experimental controls such as limited ambient noise and visual distractions. Due to the COVID-19 global pandemic, laboratory closures and safety restrictions on in-person data collection led many investigators to develop and implement remote data collection protocols. When shifting laboratory-based data collection with human participants to remote testing, investigators must consider numerous factors that impact the accuracy, reliability, validity, and ethical compliance of the work (Reips, 2002a,b).

The Remote Testing Task Force, established by the Acoustical Society of America (ASA) Psychological and Physiological (PP) Acoustics Technical Committee in the summer of 2020, sought information from the broader scientific community on enhancing the practicality and quality of perceptual research conducted via remote testing. The collective results regarding issues and best practices related to remote testing were curated on a web site, the PP Remote Testing Wiki.1 This paper provides a summary of this information, which is organized following the flow chart illustrated in Fig. 1.

FIG. 1.

FIG. 1.

(Color online) (A) Two main solutions for remote testing identified by the task force, as well as an (B) illustration of the major components in a remote testing study: compliance, recruitment, consenting, design, devices/hosting, calibration, stimulus presentation, data and safety monitoring, experiment administration, special performance consideration, and data management (Anglada-Tort et al., 2021; Cognition, https://www.cognition.run/; Harrison, 2020; Hartshorne et al., 2019; Lange et al., 2015; de Leeuw, 2015; Peirce et al., 2019; Peirce, 2007; Stoet, 2010) are depicted.

In auditory research, remote testing has received increased interest for research continuation during the global pandemic. However, remote testing has many advantages that have made it an attractive option for data collection for psychological research long before the pandemic. There has been extensive work done via web-based testing in general psychology (Hartshorne et al., 2019). In music cognition research, many studies have repeatedly shown validity of data collection outside of the laboratory by using web-based platforms (Mehr et al., 2017; Mehr et al., 2018; Mehr et al., 2019; Peretz et al., 2008; Vuvan et al., 2018) and much larger sample sizes (Larrouy-Maestri et al., 2019; Mehr et al., 2019; Müllensiefen et al., 2014; Peretz and Vuvan, 2017). In audiology, the Digits-in-Noise test has been validated as a remote testing solution for hearing screens outside of sound booths (Swanepoel et al., 2019). The practice of remote testing is likely to continue to be important. For instance, remote testing can be used to collect data from a large sample of participants while they are immersed in real-world environments, supporting increased validity of research findings as related to everyday listening situations (Bottalico et al., 2020; Brungart et al., 2020; Massonnié et al., 2019; Parker et al., 2020; Traer and McDermott, 2016; Yancey et al., 2021). It can overcome geographical constraints, allowing individuals to participate in research studies in the comfort of their own homes (Bobb et al., 2016; Graaff et al., 2018) and providing researchers with access to broader demographics outside specific regions. Remote testing also provides flexibility with respect to scheduling and alleviates the stress for participants and their caregivers, which can be associated with travel for in-person visits to the laboratory. This is particularly important for research that engages special populations, such as individuals with rare diseases and families with young children (Dial et al., 2019; Merchant et al., 2021). When research participation occurs in the home environment, it also enhances the feasibility of longitudinal studies such as those that include training (Whitton et al., 2017). Beyond research applications, remote testing will likely play an increasingly important role in audiological telehealth practice, which will promote patients' access to hearing healthcare in the future (Molini-Avejonas et al., 2015).

II. SURVEY OF RECENT STUDIES USING REMOTE DATA COLLECTION

We conducted a brief review of remote testing designs used by auditory researchers in published works between January 1, 2020 and January 31, 2022. The purpose was not to systematically review publications on remote testing but to survey the range of current and emerging approaches to remote testing—particularly in response to the COVID-19 pandemic—within the ASA community. The list of keywords used is provided in the Appendix. The keywords were taken from the titles of research articles listed in the bibliography of the PP Remote Testing Wiki.1 A small number of keywords were added manually.

Each keyword was combined with either “Journal of the Acoustical Society of America” or “JASA Express Letters” and entered as a query into the Google Scholar search engine (Google, Mountain View, CA). The option “since January 1, 2020” was selected to limit the search to articles published within the last year. The option “sort by relevance” was also selected, and only the first 3 pages of results from each query were considered (a total of 30 results per query). S.W. reviewed the search results and inspected any paper that reported on auditory behavioral science (as judged by the title and abstract). Papers that met this criterion were read, and if the reviewer confirmed that a paper made use of remote testing methods, the paper was added to Table I. Although only ASA journals were explicitly queried, a selection of relevant papers published in other journals was also included. A small number of papers published after May 2021 was added during the process of manuscript preparation. Conference proceedings and preprints were excluded. Four of the articles included in Table I were identified directly during task-force activities (i.e., they appeared in the PP Remote Testing Wiki bibliography) and were not discovered using a search engine.

TABLE I.

A list of published studies conducted via remote testing since 2020. NA, Not available.

Published study Journal Web-based vs take-home Test environment Researcher's vs bring-your-own (BYO) device Platform for web-based testing Calibrated stimulus or hardware Validation with in-laboratory testing? Compensation Sample size
Mealings et al. (2020) International Journal of Audiology Take-home Quiet rooms at the schools Researcher's Yes No Not described 297
Whittico et al. (2020) JASA Express Letters Take-home Subjects' daily environments Researcher's NA No Not described 17
Chen et al. (2020) a Otology and Neurotology Take-home Subjects' homes Researcher's Yes No Not described 25
Steffman (2021) JASA Express Letters Web-based Subjects' homes BYO Not described No No Not described 38
Merchant et al. (2021) JASA Express Letters Web-based Subjects' homes BYO Downloadable matlab script No Yes $15 per hour 39
Milne et al. (2021) Behavior Research Methods Web-based Subjects' homes BYO Gorilla.sc No No Yes, amount not described 242
Mullin et al. (2021) PLoS One Web-based Subjects' homes BYO Amazon Mechanical Turk; Sona Systems; Qualtrics for experiment No No Yes, $0.25 for study completion 641
Jaggers and Baese-Berk (2020) JASA Express Letters Web-based Subjects' homes BYO Amazon Mechanical Turk No No Not described 180
Melguy and Johnson (2021) JASA Web-based Subjects' homes BYO Amazon Mechanical Turk No No Not described 139
Davidson (2020) Language Web-based Subjects' homes BYO Amazon Mechanical Turk No No Yes, $2.25 for 10-min study completion 71
Toscano and Toscano (2021) PLoS One Web-based Subjects' homes BYO Amazon Mechanical Turk for recruitment; Qualtrics for experiment No No Yes, amount not described 181
Nagle and Huensch (2020) Journal of Second Language Pronunciation Web-based Subjects' homes BYO Amazon Mechanical Turk No No Yes, $4 for 32-min study completion 30
Huang and Elhilali (2020) eLife Web-based Subjects' homes BYO Amazon Mechanical Turk for recruitment; jsPsych and Psiturk for experiment No Yes Yes, amount not described 93
Strori et al. (2020) JASA Web-based Subjects' homes BYO Amazon Mechanical Turk No Yes Yes, amount not described 222
Getz and Toscano (2021) Attention, Perception, and Psychophysics Web-based Subjects' homes BYO Amazon Mechanical Turk No Yes Yes, $3.63 for 30-min study completion 238
Lahdelma et al. 2022 Musicae Scientiae Web-based Subjects' homes BYO Prolific.co for recruitment; PsyToolkit for experiment No No Not described 40
Düvel et al. (2020) a Music and Science Web-based Subjects' homes BYO SoSci Survey No No Not described 177
Bottalico et al. (2020) a JASA Web-based Subjects' homes BYO SurveyGizmo No No Yes, $15 for 1-hr study completion 40
McPherson et al. (2020) a Nature Communications Web-based Subjects' homes with examiner supervision for in-person experiment and no supervision for web-based experiment Researcher's for in-person experiment; BYO for web-based experiment Amazon Mechanical Turk Yes for in-person experiment; no for web-based experiment Yes Yes, amount not described 194
Shafiro et al. (2020) American Journal of Audiology Web-based Subjects' homes BYO iSpring for experiment on web-browser of subjects' choices No Yes Not described 67
Giovannone and Theodore (2021) Journal of Speech, Language, and Hearing Research Web-based Subjects' homes BYO Prolific.co for recruitment; Gorilla.sc for experiment No Yes Yes, $5.33 for study completion 190
Cooke and Lecumberri (2021) JASA Web-based Subjects' homes BYO Custom-built platform using Flask web framework No Yes Yes, amount not described 252
Gijbels et al. (2021) Frontiers in Psychology Web-based and take-home Subjects' homes BYO and researcher's Zoom No Yes Yes, amount not described 172
Viswanathan et al. (2022) Journal of Neuroscience Web-based Subjects' homes BYO Prolific for recruitment; custom web application No Yes Not described 191
Viswanathan et al. (2021) JASA Web-based Subjects' homes BYO Prolific for recruitment; custom web application No Yes Not described 286
Tripp and Munson (2021) JASA Express Letters Web-based Subjects' homes BYO Prolific for recruitment; Qualtrics for experiment No No Not described 162
Mitchell et al. (2021) JASA Web-based Subjects' homes BYO Gorilla.sc Yes, participants matched level to reference sound No Not described 86
Mimani and Singh (2021) JASA Web-based Subjects' homes BYO Google Form NA No Not described 1068
Kothinti et al. (2021) JASA Web-based Subjects' homes BYO Amazon Mechanical Turk for recruitment; jsPsych and Psiturk for experiment Yes, participants adjusted level to comfort Yes Yes, amount not described 325
Armitage et al. (2021) JASA Web-based Subjects' homes BYO Prolific for recruitment; Psytoolkit for experiment No No Not described 397
Otčenášek et al. (2022) JASA Web-based Subjects' homes and remote classroom BYO and researcher's Combination No Yes Not described 21
Zhang et al. (2021) JASA Combination Subjects' homes BYO and researcher's Zoom NA No Not described 7
McLaughlin et al. (2022) Attention, Perception, and Psychophysics Web-based Subjects' homes BYO Gorilla.sc No Yes Prolific 370
van Brenk et al. (2021) American Journal of Speech-Language Pathology Web-based Subjects' homes BYO Amazon Mechanical Turk No No Payment on Amazon Mechanical Turk, $0.8–$1.2 885
Perreau et al. (2021) American Journal of Audiology Take-home Subjects' homes and in-laboratory Researcher's NA Yes No Not described 19
a

Articles identified during PP Remote Testing Wiki development.

In total, our literature review identified 35 studies published since 2020 that made use of remote testing methods. Eight of these studies were published in the Journal of the Acoustical Society of America or JASA Express Letters. Table I describes each of the reviewed studies in terms of several design dimensions. The majority of the studies listed in Table I was conducted on web-based platforms at participants' homes with Amazon Mechanical Turk (MTurk; Amazon, Seattle, WA) being the most popular web-based platform. Only a few of these studies involved any form of supervision. Attempts to calibrate stimuli were more common among take-home studies than web-based studies. The majority of web-based studies used selection criteria to remove noncompliant or poorly performing participants, and roughly one-half of web-based studies compared data collected online to data collected in person for validation. None of the take-home studies listed in Table I made use of selection criteria or data collected in person. There was a large range of compensation across the studies. On average, web-based studies involved larger sample sizes than take-home studies; however, our search yielded only a small number of take-home studies which may be insufficient for comparison.

III. ISSUES AND SUGGESTED APPROACHES

This section provides a summary of the key issues involved in designing research studies for remote testing.

A. Compliance and administration

In the United States, there are three primary areas related to compliance when carrying out research involving human subjects: (i) study protocol reviewed by the Institutional Review Board (IRB) overseeing the research, including plans for protecting health-related information under the federal Health Insurance Portability and Accountability Act (HIPAA) statute; (ii) informed consent of the participants; and (iii) data and safety monitoring. There may be additional local requirements depending on the investigator's hosting institution. Although research compliance is not unique to remote testing, remote testing may require more considerations than in-person testing to ensure research compliance.

1. IRB and HIPAA

The IRB is tasked with protecting the rights and welfare of human subjects who participate in research (Bankert and Amdur, 2006). HIPAA is a federal law that protects a patient's health information from disclosure without consent (HIPAA Compliance Assistance, 2003). Some research studies may include collecting health information as part of the study protocol, such as obtaining case history from participants with hearing loss. In these cases, obtaining a HIPAA release from the participant or their legal guardian will be required to gain permission to access protected health information (PHI).

A study protocol that includes remote testing may be required to have plans for additional precautions and considerations beyond those for in-person testing, such as

  • (1)

    modified procedures for recruiting subjects and obtaining informed consent remotely, including obtaining PHI;

  • (2)

    additional risks of harm to the subject (e.g., sounds that are presented at higher intensities than intended);

  • (3)

    additional risks with respect to loss of confidentiality associated with transferring data from the remote test site;

  • (4)

    procedures for providing hardware or verifying hardware already in the subject's possession to meet required stimulus quality;

  • (5)

    liability associated with asking a subject to download software onto a personal computer (PC) or remotely accessing a subject's computer; and

  • (6)

    procedures for subject payment.

2. Participant recruitment

Participants in a behavioral experiment involving remote testing may be recruited locally using traditional recruitment methods or virtually via online recruitment services (e.g., see examples indicated in Fig. 1). Local recruitment may be appropriate for experiments that require sending calibrated test equipment to the participants and collecting the equipment after the completion of the experiment. On the other hand, online recruitment has the potential advantage of recruiting participants outside of the experimenter's geographical region. Some online recruitment services (e.g., Prolific, London, UK) maintain a relatively large subject pool with recorded demographic information, which allows researchers to conduct targeted recruitment (for instance, recruiting subjects within a certain age range). Contingent on IRB approval, some studies may also be advertised via social media, newsletters, or other online forums.

It is worth pointing out that despite its numerous advantages, online recruitment may also bring unique challenges. Online recruitment is largely based on participants' self-selection to enter the study. Some segments of the population may be less willing to participate in online studies compared to in-person studies; for example, some older adults may lack the confidence or ability to carry out multiple self-guided steps in an online protocol, likely resulting in recruiting older adults who are more technologically capable or confident. Self-selection may, therefore, result in certain populations being underrepresented in the participant sample, leading to lower generalizability in the research findings (Bethlehem, 2010; Turner et al., 2020). Targeted recruitment (e.g., using Prolific) may be useful in such circumstances to manually rebalance the representativeness of the sample. Moreover, even with explicitly stated inclusion and exclusion criteria, online participants may misrepresent themselves and subsequently impact data quality (Kan and Drummey, 2018). Additional challenges can be found in a recent review for behavioral studies, in general, along with recommendations for best practices to recruit participants on online platforms (Newman et al., 2021).

3. Consent procedure

The general guidelines for obtaining informed consent to participate in a research study are similar for remote and in-person testing. If documentation of the informed consent is required, there may be specific procedures for obtaining electronic consent (e.g., e-signature) or verbal consent through phone, video chat, or web-based application, as approved by the local IRB. In some cases, there may be additional procedures for research conducted in the context of telehealth practice (Bobb et al., 2016; Welch et al., 2016).

In contrast to in-person testing, where a signature is typically required to document consent, IRBs may waive the signature requirement entirely or ask for consent to be documented using a checkbox option on the consent form if the research presents no more than minimal risk of harm to the participant.

4. Data and safety monitoring

In a typical research study, data are exchanged between participants and investigators. Participants may provide demographic information or PHI. Experimenters provide instructions, and data collection often entails stimuli or prompts as well as participant response data. While considerations related to data safety are not unique to remote testing, collecting data remotely may introduce additional risks for data quality and loss of confidentiality that do not typically apply to in-person testing. Introducing additional security procedures may mitigate such risks by encrypting data, de-identifying data, and using HIPAA-compliant communication software. A plan to ensure data security and integrity should be included in the study protocol submitted to the IRB. Several approaches to data handling are detailed in Sec. III E.

5. Compensation

Compensation for research participation in remote testing can be administered using the same procedures for in-person testing in the laboratory (e.g., cash and checks), or it can be entirely electronic, contingent on IRB approval. Currently available electronic payment methods include electronic gift cards [e.g., Visa (Visa, Inc., Foster City, CA) or Mastercard (Purchase, NY)], third-party payers [e.g., Venmo (New York, NY) or Paypal (New York, NY)], and electronic checks. If recruitment and testing occur through online services that incorporate payment features (e.g., Amazon MTurk or Prolific), then participants can receive payment directly from those services. These online recruitment services typically charge a fee that is proportional to the amount paid to each participant.

B. Platforms

In this paper, a “platform” refers to any combination of hardware, software, and network system that can be used to support remote data collection outside of the laboratory. A remote testing platform can be broadly categorized into two solutions based on the logistics involved: take-home or web-based [Fig. 1(A)]. Take-home remote testing involves the delivery of at least one piece of equipment to the participant, such as headphones or a tablet. Relatively speaking, investigators have better control of the system functionality and behavior with the take-home approach by delivering pre-calibrated devices to the participant (Gallun et al., 2018; de Larrea-Mancera et al., 2020). In contrast, web-based remote testing allows less system control, although system requirements can be defined in the inclusion criteria for the study. In general, web-based testing offers more flexibility with respect to participant recruitment, logistics, and device choices (e.g., across computer, tablets, and smartphones) than the take-home approach (Grootswagers, 2020; Sauter et al., 2020). The choice of an appropriate platform should be guided by the research question and the appropriate software/hardware components as listed in Fig. 1(B).

Software development may be similar for take-home remote testing and in-person testing [e.g., use of matlab (MathWorks, Natick, MA) or Python (The Python Software Foundation, Wilmington, DE)]; however, investigators should consider including additional features to enhance the user experience during unsupervised testing and ensure security during data handling (e.g., using password protection or limiting the functionality of the device to only those features needed for testing). By contrast, designing a web-based experiment will likely involve a new programming language or data format for some investigators [e.g., JavaScript (PluralSight, Jerusalem, Israel) or JSON]. Investigators who are translating existing Mac (Apple, Cupertino, CA)/PC-based matlab or Python scripts to JavaScript-based online platforms should be mindful of the potential dependencies on the browser (e.g., Mozilla Contributors, Mozilla Corporation, Mountain View, CA) and hardware used by the participants. For non-programmers, some web-based platforms (Anwyl-Irvine et al., 2020) provide modular building functionality to reduce the burden of learning a new programming language.

C. Stimulus presentation and calibration

For in-person testing in the laboratory, investigators can select and calibrate audio hardware (i.e., earphone/headphone, loudspeaker, and sound card) to present stimuli with high fidelity and consistency across participants. There is a collection of national and international standards developed to define the equipment, environments, and procedures for clinical tests that are often adapted for experiments in hearing research. For many remote testing scenarios, particularly web-based applications (Reimers and Stewart, 2016), this level of stimulus control may not be feasible (see Fig. 1 and Table I). Hence, a potential obstacle for remote testing is variability in the stimulus quality due to different audio hardware used by participants. Depending on the platform chosen, the variability may be introduced along the continuum from take-home deployment with specific devices delivered to participants (de Larrea-Mancera et al., 2020) to web-based deployment, which participants can access with their own hardware and software (Merchant et al., 2021; Shapiro et al., 2020).

On one end of the continuum, when tight stimulus controls are required, a take-home solution for remote testing should be considered, and all deployed audio equipment should be calibrated as during in-person testing, following relevant standards. On the other end of the continuum, when supra-threshold phenomena with very limited level-dependencies are studied, calibration may not be required, and the participant may self-adjust the stimulus presentation to a comfortable level. More commonly, some limited calibration may be needed when participants use their own devices. This may be done by

  • (1)

    participant reports: collecting information about the make and model of the specific devices used by participants allows experimenters to present stimuli based on calibration conducted on the same device (e.g., using inverse transfer functions for select headphones); and

  • (2)

    psychophysical techniques: perceptual verification can be used to confirm device output fidelity (e.g., confirming independent input to the two ears through a binaural task (Milne et al., 2021; Woods et al., 2017).

Independent reviews of commercially available audio hardware (e.g., earphones) can be found in several online locations, including audiophile-oriented web sites (Headphones: Reviews, https://www.rtings.com/headphones/reviews) and online repositories of acoustic measurements suitable for equalization (Audio Group Download Server, http://audiogroup.web.th-koeln.de/ku100hrir.html; Pasaen, https://github.com/jaakkopasanen/AutoEq/blob/master/results/INDEX.md) In some cases, these reviews include extensive acoustic measurements using head and torso simulators. There may be additional variability in sound quality during remote testing. For instance, the spectral and temporal properties of audio presented from loudspeakers may be affected by reverberation in different home environments. There may be additional spectro-temporal transformations associated with audio hardware characteristics and placement. For some auditory stimuli with sharp onsets, such as pulse trains, there may be operating system-specific distortions associated with internal audio processing (e.g., Windows 10 systems, Microsoft, Redmond, WA). In many cases, the dynamic range of undistorted sound output will be limited, which reduces the ability to perform audiometric testing. Similar considerations apply to the presentation of visual stimuli in the context of remote testing. For example, studies of audiovisual integration should consider possible sources of audiovisual asynchrony associated with hardware, software, and operating systems (Bridges et al., 2020).

Besides the calibration of audio and video hardware, additional steps may be taken to reduce sources of variability that are critical for the research question. For example, additional task instructions or guidance on hardware configuration may be provided (e.g., a table mat with a map drawn on it for fixed loudspeaker placement or step-by-step instructions for verifying audio/video outputs). When participants are allowed to use their own devices for remote testing and the experimenter is not present for troubleshooting, extensive use-case investigations with combinations of hardware and software prior to data collection will help improve data quality.

For remote testing, calibration of participants' responses prior to administering the task should also be considered as part of the study design. For most studies involving responses collected via button clicks or drawing on a scale, clear instructions will suffice. In some cases, a simple response calibration routine can be used to confirm that participants have access to and understand the response interface. In other cases, more stringent response calibration may be required. For example, in motion-tracking systems and head-mounted displays, initial calibration of head orientation may be necessary to compute absolute orientation from relative measurements obtained during the task.

Even though many commercially available audio/video devices can be used for remote testing, depending on the study-specific tolerance on output quality, not all of them are compatible with hearing devices (e.g., hearing aids and cochlear implants). Some hearing devices now have the capabilities to stream audio directly through proprietary audio cables with promising evidence to suggest similar task performance between remote testing and in-person/laboratory testing (Sevier et al., 2019). Another audio-streaming method is through Bluetooth connection with the caveat that the audio signal quality using this approach may be challenging to verify. In addition, temporal delays associated with wireless transmission can introduce asynchronies with visual information displayed on a tablet or computer connected to the audio device, such as virtual buttons that are intended to light up when a sound plays. If the participants do not already have the audio-streaming features activated, additional device programming by a licensed clinician may be required (Sevier et al., 2019).

D. Participant response and task performance

The adaptation of a psychophysical task for remote testing depends on the functionality of the chosen platform. For instance, button-click responses can be reliably collected across most platforms. Several browser-based platforms [e.g., Lookit (MIT, Cambridge, MA) and Gorilla.sc (Cauldron Science Ltd, Cambridge, UK)] support video-recording with eye-tracking features. But many other physiological data (e.g., heart rate, EEG) and tracked motion responses may not be feasible to collect without specialized instrumentation, calibration procedures, and other considerations outside the scope of this manuscript.

There may be factors in home environments that affect task performance. Besides the potential effects of room acoustics mentioned in Sec. III C, factors such as ambient noise and environmental distractions can divert attention away from the experimental task. For example, data loss may occur on individual trials due to momentary interruptions or distractions in the test environment or technical issues, such as dropped audio, when streaming stimuli or responses. Future work is needed to compare task performance using in-person and remote testing procedures to address this issue (Merchant et al., 2021; Whitton et al., 2016).

Investigators may maximize the quality of data obtained remotely by (1) providing detailed instructions with step-by-step verifications when applicable; (2) designing age-, skill- and user-friendly technological interfaces; (3) considering linguistic knowledge of the subject when creating instructions (e.g., for children or non-native speakers); and (4) providing accessible supervision (e.g., via video calls) and/or verification if needed (e.g., catch-trials, monitoring, or inquiring about ambient noise levels or other activities taking place in the environment). It may also be useful to ask participants whether or not they were distracted after the test session has ended so that potentially unreliable data can be flagged.

E. Data management, handling, and analysis

Remote testing by definition involves data generated outside of the laboratory. Management and storage become critical to ensure research compliance and prevent data loss. There are two main approaches for data handling: client-side or server-side. Each platform used for remote testing falls somewhere in the wide spectrum between fully server-side data handling and fully client-side data handling. For client-side handling, the study protocol may include hardware and software delivery to the participants, precluding the need to put data online (device examples appear in Fig. 1). This management and handling protocol provides more exclusive access to the data by the investigator and may be required by some oversight bodies to ensure that no data are stored on systems accessible by anyone outside of the study team. Most oversight boards limit this requirement to participant identifiers as defined by HIPAA,2 but some also include participant codes that could be used to reidentify coded data if someone had the decoding information. For server-side handling, data are uploaded on a server (i.e., storage “cloud”; Grootswagers, 2020; Sauter et al., 2020). There are many different options when selecting a server to host data (e.g., Amazon Web Services), but additional steps may be necessary to ensure compliance with local IRB/HIPAA requirements, and it is particularly important to ensure that tdata security protocols being employed (if any) are communicated to and approved by the relevant oversight bodies (e.g., the IRB). Automatic data upload protocols to a cloud server at different time points during a task (e.g., at the end of each trial, condition, or full task) may be beneficial to minimize data loss in the event of unexpected internet or task interruption (Reips, 2002b; Sauter et al., 2020). Manual uploads by participants to a cloud may be another option, but server-side handling of data typically allows for easier deployment (e.g., on a browser), continuous data logging to a common repository, and a simpler experience for the participants. Server-side data handling need not be limited to the web-based solution for remote testing. It is possible to build calibrated, noise-monitoring take-home systems that have continuous communication with a data server.

Because of variability in hardware and software systems, as well as differences in the local environment and participants' attentional states (see also Secs. III C and III D), greater variance in experimental data may be expected for unsupervised remote testing as compared to supervised in-person testing. The magnitude of variance within subjects may vary between tasks (Bright and Pallawela, 2016; Dandurand et al., 2008). Greater variance in outcome measures may influence the interpretation of the results. Associated challenges could include greater differences in baseline task performance, reduction in effect size, and poor test-retest reliability. When possible and practical, including conditions that replicate a similar participant sample previously collected from in-laboratory testing can serve as controls for validation (Eerola et al., 2021).

Several analytical approaches may be considered in data analysis to handle elevated across-subject variance in data collected remotely. As an initial pass, removal of outliers may be warranted, provided that the procedures and rationale are clearly defined prior to data collection. Specific procedures for outlier removal and subsequent statistical analysis should be considered case-by-case, depending on the experimental task and associated pilot data. Analysis approaches that are robust in dealing with elevated variance across subjects may be appropriate, but all such procedures must be clearly described in any publication or presentation. These approaches include bootstrapping to create confidence intervals of group statistics or Bayesian analyses that constrain parameter estimates by incorporating prior knowledge about subject-level variance (Rouder et al., 2019) or the likelihood of latent group membership (e.g., malingerers or distracted listeners; Ortega et al., 2012).

IV. DISCUSSION

A. Case studies

Three case studies of experiments conducted remotely are presented here to illustrate some of the steps outlined in Fig. 1. These studies illustrate the different platforms chosen as appropriate for the specific investigations.

1. A case study using the take-home approach

Chen et al. (2020) presented a study using a self-directed repeated measure for subjective tinnitus characterization employing calibrated take-home equipment. The participants were adults with complaints of tinnitus. The consent procedures and instructions were given in person during the initial visit to the clinic. During this visit, the participants received a calibrated system that included a tablet and consumer-grade headphones to take home for five remote testing sessions over a two-week period. The tablet software implemented standard clinical procedures, including a health questionnaire and automated audiometry described in another validation study by the same authors (Whitton et al., 2016). The data were stored on the tablet, which was returned to the experimenters after the study's completion. Three behavioral tasks were tested for each remote session: (1) a tinnitus matching task, in which subjects changed the test stimulus' center frequency, level, modulation rate, and bandwidth to match their tinnitus; (2) a tinnitus intensity rating using a visual analog scale; and (3) a task for estimating loudness discomfort levels of pure tones.

The repeated remote sessions characterized fluctuations in individual subjective tinnitus ratings over time. The authors found that within-subject variability of tinnitus intensity scores and loudness discomfort levels reduced over time, which might be due to increasing familiarity with reporting symptoms using the study-specific instruments (i.e., tasks and surveys). The authors concluded that characterization of subjective markers of tinnitus can benefit from multiple test sessions over time as compared to a single session in the clinic. The data obtained using these remote testing procedures with calibrated devices provide strong evidence of feasibility for this type of investigation.

2. A case study using downloadable experimental software

Merchant et al. (2021) presented a study on binaural intelligibility level difference (BILD) for school-age children and adults, for which the data were collected in their homes. All subjects had normal hearing as reported by themselves or a parent. Most child participants had a sibling or parent from the same household who also participated. The consent procedure was conducted over Webex video-conference software (Cisco, San Jose, CA). The experiment was programmed using compiled matlab scripts. Participants downloaded the software, ran it on their PCs, and listened to stimuli using personal headphones. At the beginning of the experiment, instructions were provided for volume adjustment, and subjects were asked to use the same settings throughout the testing. Participants were encouraged to ask questions during the video conference or over electronic mail. Parents or caregivers were provided with instructions for how to assist their children when running the experiment. Payment was provided in the form of electronic gift cards.

The task was three-alternative forced-choice word recognition in quiet or speech-shaped noise; the masker was always diotic, and the target was either in phase or out of phase across ears. The procedures followed those of a previous study with in-person data collection (Schneider et al., 2018). After each experimental run, participants were asked to report any distractions, and ambient noise levels were measured using a sound level meter application that runs on a smartphone. The data were collected at three time points (days 1, 2, and 7) and uploaded to the REDCap (Vanderbilt, Nashville, TN) cloud server hosted at the authors' institution.

The authors concluded that remote testing of the BILD in children is both feasible and generally reliable. Multiple approaches were used to understand data variability; data were collected on multiple days for each subject, and the results were compared with data from a previous in-person experiment conducted with different participants. The authors suggested that the use of personal hardware may increase the incidence of outlier data and there may be a benefit to supplying or specifying specific headphones to improve reliability.

3. A case study using the browser-based approach

Milne et al. (2021) presented a web-based study on psychoacoustic methods for detecting improper headphone use. They compared performances on a psychoacoustic task using three types of stimulus that can only be performed with precision when the subject has access to binaural cues: Huggins Pitch, detection of an out-of-phase tone (Woods et al., 2017), and binaural beats. The experiments included protocols implemented using Gorilla. All of the tasks from the study are publicly available in the “Gorilla Open Materials” repository. The consent process was integrated into Gorilla, providing participants with an information sheet and IBR-approved consent form, which required participants to tick a checkbox for each clause of the form (Milne, 2021).

In experiment 1, over 100 adult subjects with a background in hearing science were recruited. These “trusted” subjects were drawn from the auditory science community via mailing lists and direct electronic mailings and, thus, assumed to have good compliance with audio device use and task instructions (i.e., using good audio devices and following instructions about which kind of device to use). These trusted subjects performed the Huggins Pitch and anti-phasic tone detection tasks using headphones and, again, with stereo loudspeakers. In experiment 2, another group of 100 adult subjects was recruited from Prolific. This group represented “naive” subjects who might not have good compliance with headphone use. Their data were compared with baseline data from the trusted subjects to estimate the frequency of naive subjects using loudspeakers or headphones with low quality. In experiment 3, a smaller number of trusted subjects were tested using binaural beats. Across all three experiments, the task was three-alternative forced-choice with the same response interface for each stimulus type (e.g., Huggins Pitch, anti-phase, and beat).

Milne et al. (2021) concluded that (1) the experiments using Huggins Pitch and beat stimuli are more sensitive than anti-phase stimuli for determining improper headphone use, and (2) approximately 40% of naive subjects in a random sample of remote participants did not follow the instructions to use headphones (vs speakers) or used headphones with poor quality. Further, they recommended using a two-step headphone screening protocol to ensure that the subject's audio device meets minimum requirements for auditory testing. The Huggins Pitch task paired with either the binaural beats or anti-phase test ensures a reasonably low, ∼7% false-positive rate of passing the headphone screening.

B. Considerations for peer review and identifying suggested approaches

By and large, remote testing methodologies offer greater flexibility and access to participants but poorer experimental control as compared to in-person data collection. With appropriate consideration of limitations and safeguards, these methods do allow for studying a broad range of scientific questions. Given the increased need for new research paradigms that better promote participant safety and inclusion, reviewers will need to think critically about the actual impacts of a new methodology compared to existing approaches. The key focus, when evaluating such work, should be on whether the hardware and test protocols are sufficient to support reliable and valid data that inform the specific research question rather than on whether they meet the standards of in-laboratory testing.

The long-term need to identify standards and best practices for remote testing is clear. We, as a scientific community formed by members of the task force, contribute toward that goal by describing possible approaches and identifying key challenges in this work. Future standards that describe minimum requirements should represent a consensus from within the hearing research community and involve interdisciplinary groups such as the ASA Standards Committee. From our survey of current remote testing approaches, it seems likely that much of the recommended methodology will be task specific, thus, complicating the goal of identifying universal best practices. At this time, however, a few broad recommendations can be identified as general safeguards for data quality. Specifically, we recommend keeping the following in mind when designing remote experiments:

  • (1)

    selecting the platform and procedures best suited for the research question and practical constraints associated with recruitment and data collection;

  • (2)

    measuring and documenting calibration (or the range of calibrations) to characterize the stimuli and test environments;

  • (3)

    validating experimental data through in-person data comparison in laboratory settings, if appropriate; and

  • (4)

    including additional steps to identify outlier data (e.g., perceptual screening) and/or control data quality (e.g., catch-trials).

ACKNOWLEDGMENTS

An earlier version of this report has been presented at the 179th meeting of the Acoustical Society of America. A.R.D., F.J.G., and E.C.H. benefitted from National Institutes of Health (NIH) funding to F.J.G. (Grant No. R01 DC015051) to develop and validate PART, a platform for remote testing. S.W. is a volunteer at FindingFive, a web-based experiment platform featured in Fig. 1. G.C.S. is a member of Auditory Space, LLC, developers of software for virtual and remote testing. J.H.V. is an employee of the U.S. Department of Veterans Affairs; the contents of this manuscript do not represent the views of the U.S. Department of Veterans Affairs or the U.S. government.

APPENDIX

The search keywords for publication between January 1, 2020 and May 24, 2021 include online, internet, remote, self-testing, self-administered, portable, home, mobile, phone, web, Crowdsource, Amazon Mechanical Turk.

Footnotes

References

  • 1. Anglada-Tort, M. , Harrison, P. M. C. , and Jacoby, N. (2021). “ REPP: A robust cross-platform solution for online sensorimotor synchronization experiments,” bioRxiv. [DOI] [PMC free article] [PubMed]
  • 2. Anwyl-Irvine, A. L. , Massonnié, J. , Flitton, A. , Kirkham, N. , and Evershed, J. K. (2020). “ Gorilla in our midst: An online behavioral experiment builder,” Behav. Res. Methods 52, 388–407. 10.3758/s13428-019-01237-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Armitage, J. , Lahdelma, I. , and Eerola, T. (2021). “ Automatic responses to musical intervals: Contrasts in acoustic roughness predict affective priming in Western listeners,” J. Acoust. Soc. Am. 150, 551–560. 10.1121/10.0005623 [DOI] [PubMed] [Google Scholar]
  • 5. Bankert, E. A. , and Amdur, R. J. (2006). Institutional Review Board: Management and Function ( Jones and Bartlett Learning, Burlington, MA: ). [Google Scholar]
  • 6. Bethlehem, J. (2010). “ Selection bias in web surveys,” Int. Stat. Rev. 78, 161–188. 10.1111/j.1751-5823.2010.00112.x [DOI] [Google Scholar]
  • 7. Bobb, M. R. , Van Heukelom, P. G. , Faine, B. A. , Ahmed, A. , Messerly, J. T. , Bell, G. , Harland, K. K. , Simon, C. , and Mohr, N. M. (2016). “ Telemedicine provides noninferior research informed consent for remote study enrollment: A randomized controlled trial,” Acad. Emerg. Med. 23, 759–765. 10.1111/acem.12966 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Bottalico, P. , Murgia, S. , Puglisi, G. E. , Astolfi, A. , and Kirk, K. I. (2020). “ Effect of masks on speech intelligibility in auralized classrooms,” J. Acoust. Soc. Am. 148, 2878–2884. 10.1121/10.0002450 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Bridges, D. , Pitiot, A. , MacAskill, M. R. , and Peirce, J. W. (2020). “ The timing mega-study: Comparing a range of experiment generators, both lab-based and online,” PeerJ 8, e9414. 10.7717/peerj.9414 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Bright, T. , and Pallawela, D. (2016). “ Validated smartphone-based apps for ear and hearing assessments: A review,” JMIR Rehabil. Assist. Technol. 3, e13. 10.2196/rehab.6074 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Brungart, D. S. , Barrett, M. E. , Cohen, J. I. , Fodor, C. , Yancey, C. M. , and Gordon-Salant, S. (2020). “ Objective assessment of speech intelligibility in crowded public spaces,” Ear Hear. 41, 68S. 10.1097/AUD.0000000000000943 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Chen, J. X. , Whitton, J. P. , Parthasarathy, A. , Hancock, K. E. , and Polley, D. B. (2020). “ Fluctuations in subjective tinnitus ratings over time: Implications for clinical research,” Otol. Neurotol. 41, e1167–e1173. 10.1097/MAO.0000000000002759 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Cooke, M. , and Lecumberri, M. L. G. (2021). “ How reliable are online speech intelligibility studies with known listener cohorts?,” J. Acoust. Soc. Am. 150, 1390–1401. 10.1121/10.0005880 [DOI] [PubMed] [Google Scholar]
  • 15. Dandurand, F. , Shultz, T. R. , and Onishi, K. H. (2008). “ Comparing online and lab methods in a problem-solving experiment,” Behav. Res. Methods 40, 428–434. 10.3758/BRM.40.2.428 [DOI] [PubMed] [Google Scholar]
  • 16. Davidson, L. (2020). “ Contributions of modal and creaky voice to the perception of habitual pitch,” Language 96, e22–e37. 10.1353/lan.2020.0013 [DOI] [Google Scholar]
  • 17. de Graaff, F. , Huysmans, E. , Merkus, P. , Goverts, S. T. , and Smits, C. (2018). “ Assessment of speech recognition abilities in quiet and in noise: A comparison between self-administered home testing and testing in the clinic for adult cochlear implant users,” Int. J. Audiol. 57, 872–880. 10.1080/14992027.2018.1506168 [DOI] [PubMed] [Google Scholar]
  • 18. de Larrea-Mancera, E. S. L. , Stavropoulos, T. , Hoover, E. C. , Eddins, D. A. , Gallun, F. J. , and Seitz, A. R. (2020). “ Portable Automated Rapid Testing (PART) for auditory assessment: Validation in a young adult normal-hearing population,” J. Acoust. Soc. Am. 148, 1831–1851. 10.1121/10.0002108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. de Leeuw, J. R. (2015). “ jsPsych: A JavaScript library for creating behavioral experiments in a Web browser,” Behav. Res. Methods 47, 1–12. 10.3758/s13428-014-0458-y [DOI] [PubMed] [Google Scholar]
  • 20. Dial, H. R. , Hinshelwood, H. A. , Grasso, S. M. , Hubbard, H. I. , Gorno-Tempini, M.-L. , and Henry, M. L. (2019). “ Investigating the utility of teletherapy in individuals with primary progressive aphasia,” Clin. Interv. Aging 14, 453. 10.2147/CIA.S178878 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Düvel, N. , Kopiez, R. , Wolf, A. , and Weihe, P. (2020). “ Confusingly similar: Discerning between hardware guitar amplifier sounds and simulations with the Kemper Profiling Amp,” Music Sci. 3, 205920432090195. 10.1177/2059204320901952 [DOI] [Google Scholar]
  • 22. Eerola, T. , Armitage, J. , Lavan, N. , and Knight, S. (2021). “ Online data collection in auditory perception and cognition research: Recruitment, testing, data quality and ethical considerations,” Aud. Percept. Cognit. 3, 205920432090195. 10.1080/25742442.2021.2007718 [DOI] [Google Scholar]
  • 23. Gallun, F. J. , Seitz, A. , Eddins, D. A. , Molis, M. R. , Stavropoulos, T. , Jakien, K. M. , Kampel, S. D. , Diedesch, A. C. , Hoover, E. C. , Bell, K. , Souza, P. E. , Sherman, M. , Calandruccio, L. , Xue, G. , Taleb, N. , Sebena, R. , and Srinivasan, N. (2018). “ Development and validation of Portable Automated Rapid Testing (PART) measures for auditory research,” Proc. Mtgs. Acoust. 33, 050002. 10.1121/2.0000878 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Getz, L. M. , and Toscano, J. C. (2021). “ Rethinking the McGurk effect as a perceptual illusion,” Atten. Percept. Psychophys. 83, 2583–2598. 10.3758/s13414-021-02265-6 [DOI] [PubMed] [Google Scholar]
  • 25. Gijbels, L. , Cai, R. , Donnelly, P. M. , and Kuhl, P. K. (2021). “ Designing virtual, moderated studies of early childhood development,” Front. Psychol. 12, 4331. 10.3389/fpsyg.2021.740290 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Giovannone, N. , and Theodore, R. M. (2021). “ Individual differences in lexical contributions to speech perception,” J. Speech, Lang. Hear. Res. 64, 707–724. 10.1044/2020_JSLHR-20-00283 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Grootswagers, T. (2020). “ A primer on running human behavioural experiments online,” Behav. Res. Methods 52, 2283–2286. 10.3758/s13428-020-01395-3 [DOI] [PubMed] [Google Scholar]
  • 28. Harrison, P. M. C. (2020). “ psychTestR: An R package for designing and conducting behavioural psychological experiments,” J. Open Source Software 5(49), 2088. 10.21105/joss.02088 [DOI] [Google Scholar]
  • 29. Hartshorne, J. K. , de Leeuw, J. R. , Goodman, N. D. , Jennings, M. , and O'Donnell, T. J. (2019). “ A thousand studies for the price of one: Accelerating psychological science with Pushkin,” Behav. Res. Methods 51, 1782–1803. 10.3758/s13428-018-1155-z [DOI] [PubMed] [Google Scholar]
  • 32.HIPAA Compliance Assistance (2003). Summary of the HIPAA Privacy Rule ( Office for Civil Rights, Washington, DC: ). [Google Scholar]
  • 33. Huang, N. , and Elhilali, M. (2020). “ Push-pull competition between bottom-up and top-down auditory attention to natural soundscapes,” eLife 9, e52984. 10.7554/eLife.52984 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Jaggers, Z. , and Baese-Berk, M. M. (2020). “ Investigating a bias for cue preservation in loanword adaptation,” J. Acoust. Soc. Am. 147, EL511–EL516. 10.1121/10.0001375 [DOI] [PubMed] [Google Scholar]
  • 35. Kan, I. P. , and Drummey, A. B. (2018). “ Do imposters threaten data quality? An examination of worker misrepresentation and downstream consequences in Amazon's Mechanical Turk workforce,” Comput. Human Behav. 83, 243–253. 10.1016/j.chb.2018.02.005 [DOI] [Google Scholar]
  • 36. Kothinti, S. R. , Huang, N. , and Elhilali, M. (2021). “ Auditory salience using natural scenes: An online study,” J. Acoust. Soc. Am. 150, 2952–2966. 10.1121/10.0006750 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Lahdelma, I. , Armitage, J. , and Eerola, T. (2022). “ Affective priming with musical chords is influenced by pitch numerosity,” Musicae Sci. 26(1), 208–217. 10.1177/1029864920911127 [DOI] [Google Scholar]
  • 38. Lange, K. , Kühn, S. , and Filevich, E. (2015). “ ‘ Just Another Tool for Online Studies’ (JATOS): An easy solution for setup and management of web servers supporting online studies,” PLoS One 10, e0130834. 10.1371/journal.pone.0130834 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Larrouy-Maestri, P. , Harrison, P. M. C. , and Müllensiefen, D. (2019). “ The mistuning perception test: A new measurement instrument,” Behav. Res. Methods 51, 663–675. 10.3758/s13428-019-01225-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Massonnié, J. , Rogers, C. J. , Mareschal, D. , and Kirkham, N. Z. (2019). “ Is classroom noise always bad for children? The contribution of age and selective attention to creative performance in noise,” Front. Psychol. 10, 381. 10.3389/fpsyg.2019.00381 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. McLaughlin, D. J. , Brown, V. A. , Carraturo, S. , and van Engen, K. J. (2022). “ Revisiting the relationship between implicit racial bias and audiovisual benefit for nonnative-accented speech,” Attention, Percept., Psychophys. (published online). 10.3758/S13414-021-02423-W [DOI] [PubMed] [Google Scholar]
  • 42. McPherson, M. J. , Dolan, S. E. , Durango, A. , Ossandon, T. , Valdés, J. , Undurraga, E. A. , Jacoby, N. , Godoy, R. A. , and McDermott, J. H. (2020). “ Perceptual fusion of musical notes by native Amazonians suggests universal representations of musical intervals,” Nat. Commun. 11, 2786. 10.1038/s41467-020-16448-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Mealings, K. , Cameron, S. , and Dillon, H. (2020). “ Correlating performance on the Listening in Spatialised Noise–Sentences test (LiSN-S) with the Listening in Spatialised Noise–Universal test (LiSN-U),” Int. J. Audiol. 59, 519–523. 10.1080/14992027.2020.1753119 [DOI] [PubMed] [Google Scholar]
  • 44. Mehr, S. A. , Kotler, J. , Howard, R. M. , Haig, D. , and Krasnow, M. M. (2017). “ Genomic imprinting is implicated in the psychology of music,” Psychol. Sci. 28, 1455–1467. 10.1177/0956797617711456 [DOI] [PubMed] [Google Scholar]
  • 45. Mehr, S. A. , Singh, M. , Knox, D. , Ketter, D. M. , Pickens-Jones, D. , Atwood, S. , Lucas, C. , Jacoby, N. , Egner, A. A. , Hopkins, E. J. , Howard, R. M. , Hartshorne, J. K. , Jennings, M. V. , Simson, J. , Bainbridge, C. M. , Pinker, S. , O'Donnell, T. J. , Krasnow, M. M. , and Glowacki, L. (2019). “ Universality and diversity in human song,” Science (published online). 10.1126/science.aax0868 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Mehr, S. A. , Singh, M. , York, H. , Glowacki, L. , and Krasnow, M. M. (2018). “ Form and function in human song,” Curr. Biol. 28, 356–368.e5. 10.1016/J.CUB.2017.12.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Melguy, Y. V. , and Johnson, K. (2021). “ General adaptation to accented English: Speech intelligibility unaffected by perceived source of non-native accent,” J. Acoust. Soc. Am. 149, 2602–2614. 10.1121/10.0004240 [DOI] [PubMed] [Google Scholar]
  • 48. Merchant, G. R. , Dorey, C. , Porter, H. L. , Buss, E. , and Leibold, L. J. (2021). “ Feasibility of remote assessment of the binaural intelligibility level difference in school-age children,” JASA Express Lett. 1, 014405. 10.1121/10.0003323 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Milne, A. E. (2021). (personal communication).
  • 49. Milne, A. E. , Bianco, R. , Poole, K. C. , Zhao, S. , Oxenham, A. J. , Billig, A. J. , and Chait, M. (2021). “ An online headphone screening test based on dichotic pitch,” Behav. Res. Methods 53, 1551–1562. 10.3758/s13428-020-01514-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Mimani, A. , and Singh, R. (2021). “ Anthropogenic noise variation in Indian cities due to the COVID-19 lockdown during March-to-May 2020,” J. Acoust. Soc. Am. 150, 3216–3227. 10.1121/10.0006966 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Mitchell, A. , Oberman, T. , Aletta, F. , Kachlicka, M. , Lionello, M. , Erfanian, M. , and Kang, J. (2021). “ Investigating urban soundscapes of the COVID-19 lockdown: A predictive soundscape modeling approacha,” J. Acoust. Soc. Am. 150, 4474–4488. 10.1121/10.0008928 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Molini-Avejonas, D. R. , Rondon-Melo, S. , de la Higuera Amato, C. A. , and Samelli, A. G. (2015). “ A systematic review of the use of telehealth in speech, language and hearing sciences,” J. Telemed. Telecare 21, 367–376. 10.1177/1357633X15583215 [DOI] [PubMed] [Google Scholar]
  • 54. Müllensiefen, D. , Gingras, B. , Musil, J. , and Stewart, L. (2014). “ The musicality of non-musicians: An index for assessing musical sophistication in the general population,” PLoS One 9, e89642. 10.1371/journal.pone.0089642 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55. Mullin, H. A. C. , Norkey, E. A. , Kodwani, A. , Vitevitch, M. S. , and Castro, N. (2021). “ Does age affect perception of the Speech-to-Song Illusion?,” PLoS One 16, e0250042. 10.1371/journal.pone.0250042 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Nagle, C. L. , and Huensch, A. (2020). “ Expanding the scope of L2 intelligibility research: Intelligibility, comprehensibility, and accentedness in L2 Spanish,” JSLP 6, 329–351. 10.1075/jslp.20009.nag [DOI] [Google Scholar]
  • 57. Newman, A. , Lam Bavik, Y. , Mount, M. , and Shao, B. (2021). “ Data collection via online platforms: Challenges and recommendations for future research,” Appl. Psychol. 70, 1380–1402. 10.1111/apps.12302 [DOI] [Google Scholar]
  • 58. Ortega, A. , Wagenmakers, E.-J. , Lee, M. D. , Markowitsch, H. J. , and Piefke, M. (2012). “ A Bayesian latent group analysis for detecting poor effort in the assessment of malingering,” Arch. Clin. Neuropsychol. 27, 453–465. 10.1093/arclin/acs038 [DOI] [PubMed] [Google Scholar]
  • 59. Otčenášek, J. , Frič, M. , Dvořáková, E. , Otčenášek, Z. , and Ubik, S. (2022). “ The subjective relevance of perceived sound aspects in remote singing education,” J. Acoust. Soc. Am. 151, 428–433. 10.1121/10.0009143 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Parker, A. , Slack, C. , and Skoe, E. (2020). “ Comparisons of auditory brainstem responses between a laboratory and simulated home environment,” J. Speech, Lang. Hear. Res. 63, 3877–3892. 10.1044/2020_JSLHR-20-00383 [DOI] [PubMed] [Google Scholar]
  • 62. Peirce, J. , Gray, J. R. , Simpson, S. , MacAskill, M. , Höchenberger, R. , Sogo, H. , Kastman, E. , and Lindeløv, J. K. (2019). “ PsychoPy2: Experiments in behavior made easy,” Behav. Res. Methods 51, 195–203. 10.3758/s13428-018-01193-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63. Peirce, J. W. (2007). “ PsychoPy—Psychophysics software in Python,” J. Neurosci. Methods 162, 8–13. 10.1016/j.jneumeth.2006.11.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64. Peretz, I. , Gosselin, N. , TillmannI, B. , Gagnon, B. , Trimmer, C. G. , Paquette, S. , and Bouchard, B. (2008). “ On-line identification of congenital amusia,” Music Percept. 25, 331–343. 10.1525/mp.2008.25.4.331 [DOI] [Google Scholar]
  • 65. Peretz, I. , and Vuvan, D. T. (2017). “ Prevalence of congenital amusia,” Eur. J. Hum. Genet. 25, 625–630. 10.1038/ejhg.2017.15 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66. Perreau, A. E. , Tyler, R. S. , Frank, V. , Watts, A. , and Mancini, P. C. (2021). “ Use of a smartphone app for cochlear implant patients with tinnitus,” Am. J. Audiol. 30, 676–687. 10.1044/2021_AJA-20-00195 [DOI] [PubMed] [Google Scholar]
  • 67. Reimers, S. , and Stewart, N. (2016). “ Auditory presentation and synchronization in Adobe Flash and HTML5/JavaScript Web experiments,” Behav. Res. Methods 48, 897–908. 10.3758/s13428-016-0758-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68. Reips, U.-D. (2002a). “ Internet-based psychological experimenting,” Social Sci. Comput. Rev. 20, 241–249. 10.1177/089443930202000302 [DOI] [Google Scholar]
  • 69. Reips, U.-D. (2002b). “ Standards for internet-based experimenting,” Exp. Psychol. 49, 243–256. 10.1026/1618-3169.49.4.243 [DOI] [PubMed] [Google Scholar]
  • 70. Rouder, J. N. , Haaf, J. M. , Davis-Stober, C. P. , and Hilgard, J. (2019). “ Beyond overall effects: A Bayesian approach to finding constraints in meta-analysis,” Psychol. Methods 24, 606. 10.1037/met0000216 [DOI] [PubMed] [Google Scholar]
  • 71. Sauter, M. , Draschkow, D. , and Mack, W. (2020). “ Building, hosting and recruiting: A brief introduction to running behavioral experiments online,” Brain Sci. 10, 251. 10.3390/brainsci10040251 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72. Schneider, E. , Browning, J. , Buss, E. , and Leibold, L. J. (2018). “ Binaural intelligibility level difference (BILD) in children with normal hearing,” in The American Academy of Audiology Annual Convention, Nashville, TN. [Google Scholar]
  • 73. Sevier, J. D. , Choi, S. , and Hughes, M. L. (2019). “ Use of direct-connect for remote speech-perception testing in cochlear implants,” Ear Hear. 40, 1162–1173. 10.1097/AUD.0000000000000693 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74. Shafiro, V. , Hebb, M. , Walker, C. , Oh, J. , Hsiao, Y. , Brown, K. , Sheft, S. , Li, Y. , Vasil, K. , and Moberly, A. C. (2020). “ Development of the basic auditory skills evaluation battery for online testing of cochlear implant listeners,” Am. J. Audiol. 29, 577. 10.1044/2020_AJA-19-00083 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75. Shapiro, M. L. , Norris, J. A. , Wilbur, J. C. , Brungart, D. S. , and Clavier, O. H. (2020). “ TabSINT: Open-source mobile software for distributed studies of hearing,” Int. J. Audiol. 59, S12–S19. 10.1080/14992027.2019.1698776 [DOI] [PubMed] [Google Scholar]
  • 76. Steffman, J. (2021). “ Contextual prominence in vowel perception: Testing listener sensitivity to sonority expansion and hyperarticulation,” JASA Express Lett. 1, 045203. 10.1121/10.0003984 [DOI] [PubMed] [Google Scholar]
  • 77. Stoet, G. (2010). “ PsyToolkit: A software package for programming psychological experiments using Linux,” Behav. Res. Methods 42, 1096–1104. 10.3758/BRM.42.4.1096 [DOI] [PubMed] [Google Scholar]
  • 78. Strori, D. , Bradlow, A. R. , and Souza, P. E. (2020). “ Recognition of foreign-accented speech in noise: The interplay between talker intelligibility and linguistic structure,” J. Acoust. Soc. Am. 147, 3765–3782. 10.1121/10.0001194 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79. Swanepoel, D. W. , de Sousa, K. C. , Smits, C. , and Moore, D. R. (2019). “ Mobile applications to detect hearing impairment: Opportunities and challenges,” Bull. W. H. O. 97, 717. 10.2471/BLT.18.227728 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80. Toscano, J. C. , and Toscano, C. M. (2021). “ Effects of face masks on speech recognition in multi-talker babble noise,” PLoS One 16, e0246842. 10.1371/journal.pone.0246842 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81. Traer, J. , and McDermott, J. H. (2016). “ Statistics of natural reverberation enable perceptual separation of sound and space,” Proc. Natl. Acad. Sci. U.S.A. 113, E7856–E7865. 10.1073/pnas.1612524113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82. Tripp, A. , and Munson, B. (2021). “ Written standard sentence materials convey social information,” JASA Express Lett. 1, 125202. 10.1121/10.0007466 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83. Turner, A. M. , Engelsma, T. , Taylor, J. O. , Sharma, R. K. , and Demiris, G. (2020). “ Recruiting older adult participants through crowdsourcing platforms: Mechanical Turk versus Prolific Academic,” in AMIA Annual Symposium Proceedings, 2020, 1230. [PMC free article] [PubMed] [Google Scholar]
  • 84. van Brenk, F. , Kain, A. , and Tjaden, K. (2021). “ Investigating acoustic correlates of intelligibility gains and losses during slowed speech: A hybridization approach,” Am. J. Speech. Lang. Pathol. 30, 1343–1360. 10.1044/2021_AJSLP-20-00172 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85. Viswanathan, V. , Shinn-Cunningham, B. G. , and Heinz, M. G. (2021). “ Temporal fine structure influences voicing confusions for consonant identification in multi-talker babble,” J. Acoust. Soc. Am. 150, 2664–2676. 10.1121/10.0006527 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86. Viswanathan, V. , Shinn-Cunningham, B. G. , and Heinz, M. G. (2022). “ Speech categorization reveals the role of early-stage temporal-coherence processing in auditory scene Analysis,” J. Neurosci. 42, 240–254. 10.1523/JNEUROSCI.1610-21.2021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87. Vuvan, D. T. , Paquette, S. , Mignault Goulet, G. , Royal, I. , Felezeu, M. , and Peretz, I. (2018). “ The Montreal protocol for identification of amusia,” Behav. Res. Methods 50, 662–672. 10.3758/s13428-017-0892-8 [DOI] [PubMed] [Google Scholar]
  • 88. Welch, B. M. , Marshall, E. , Qanungo, S. , Aziz, A. , Laken, M. , Lenert, L. , and Obeid, J. (2016). “ Teleconsent: A novel approach to obtain informed consent for research,” Contemp. Clin. Trials Commun. 3, 74–79. 10.1016/j.conctc.2016.03.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89. Whittico, T. H. , Ortiz, A. J. , Marks, K. L. , Toles, L. E. , Van Stan, J. H. , Hillman, R. E. , and Mehta, D. D. (2020). “ Ambulatory monitoring of Lombard-related vocal characteristics in vocally healthy female speakers,” J. Acoust. Soc. Am. 147(6), EL552–EL558. 10.1121/10.0001446 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90. Whitton, J. P. , Hancock, K. E. , Shannon, J. M. , and Polley, D. B. (2016). “ Validation of a self-administered audiometry application: An equivalence study,” Laryngoscope 126, 2382–2388. 10.1002/lary.25988 [DOI] [PubMed] [Google Scholar]
  • 91. Whitton, J. P. , Hancock, K. E. , Shannon, J. M. , and Polley, D. B. (2017). “ Audiomotor perceptual training enhances speech intelligibility in background noise,” Curr. Biol. 27, 3237–3247.e6. 10.1016/j.cub.2017.09.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92. Woods, K. J. P. , Siegel, M. H. , Traer, J. , and McDermott, J. H. (2017). “ Headphone screening to facilitate web-based auditory experiments,” Atten. Percept. Psychophys. 79, 2064–2072. 10.3758/s13414-017-1361-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93. Yancey, C. M. , Barrett, M. E. , Gordon-Salant, S. , and Brungart, D. S. (2021). “ Binaural advantages in a real-world environment on speech intelligibility, response time, and subjective listening difficulty,” JASA Express Lett. 1, 014406. 10.1121/10.0003193 [DOI] [PubMed] [Google Scholar]
  • 94. Zhang, C. , Jepson, K. , Lohfink, G. , and Arvaniti, A. (2021). “ Comparing acoustic analyses of speech data collected remotely,” J. Acoust. Soc. Am. 149(6), 3910–3916. 10.1121/10.0005132 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from The Journal of the Acoustical Society of America are provided here courtesy of Acoustical Society of America

RESOURCES