Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2025 Mar 1.
Published in final edited form as: Laryngoscope. 2023 Sep 6;134(3):1023–1031. doi: 10.1002/lary.31042

Cough Sounds in Screening and Diagnostics: a Scoping Review

Siddhi Hegde 1,*, Shreya Sreeram 1,*, Isaac L Alter 2,*, Chaya Shor 3, Tulio A Valdez 4, Kara D Meister 4,, Anaïs Rameau 3,
PMCID: PMC10915103  NIHMSID: NIHMS1928573  PMID: 37672667

Abstract

Objective:

To examine applications of cough sounds towards screening tools and diagnostics in the biomedical and engineering literature, with particular focus on disease types, acoustic data collection protocols, data processing and analytics, accuracy, and limitations.

Data Sources:

PubMed, EMBASE, Web of Science, Scopus, Cochrane Library, IEEE Xplore, Engineering Village, and ACM Digital Library were searched from inception to August 2021.

Review Methods:

A scoping review was conducted on screening and diagnostic uses of cough sounds in adults, children, and animals, in English peer-reviewed and gray literature of any design.

Results:

Of 438 abstracts screened, 108 articles met inclusion criteria. Human studies were most common (77.8%); the majority focused on adults (57.3%). Single-modality acoustic data collection was most common (71.2%), with few multimodal studies, including plethysmography (15.7%) and clinico-demographic data (7.4%). Data analytics methods were highly variable, with 61.1% using machine learning, the majority of which (78.8%) were published after 2010. Studies commonly focused on cough detection (41.7%) and screening of COVID-19 (11.1%); among pediatric studies, the most common focus was diagnosis of asthma (52.6%).

Conclusion:

Though the use of cough sounds in diagnostics is not new, academic interest has accelerated in the past decade. Cough sound offers the possibility of an accessible, non-invasive, and low-cost disease biomarker, particularly in the era of rapid development of machine learning capabilities in combination with the ubiquity of cellular technology with high-quality recording capability. However, most cough sound literature hinges on non-standardized data collection protocols and small, non-diverse, single-modality datasets, with limited external validity.

Keywords: Machine learning, acoustics, cough

Lay Summary:

Cough sound analysis holds the promise of an accessible disease biomarker. We survey efforts in using cough sounds in adult, pediatric and animal populations, and examine evolution in hardware and analytics. Rapid acceleration is noted in the past decade, though with limitations.

Introduction

Cough is an essential defensive reflex of the respiratory tract to clear material from the airway, characterized by the precise coordination of laryngeal, abdominal and chest wall muscles, under the control of medullary and higher cortical regions of the brain.1,2 Cough is associated with a characteristic sound, reflecting a specific pattern of airway vibration carrying critical information on the condition of the respiratory system.3,4 Cough sounds are unique and recognizable acoustic events. They have a timbre associated with particular disease states, which have been used by clinicians for diagnostic purposes for centuries. However these characteristic acoustic biomarkers have not yet been translated into applied screening or diagnostic technology and are currently limited to diagnostic clues observed by clinicians.5 Objective measures and automated analyses of cough sounds could transform the screening, diagnosis, and management of many diseases treated by otolaryngologists, pulmonologists, and other health professionals.

The current ubiquitous availability of audio recording hardware via cellular technology and exponential advances in data analytics via deep learning have created unprecedented opportunities for collecting and analyzing cough sounds. Several promising results have been published in the distinction of normal and abnormal cough sounds,6 screening of disease such as COVID-19,79 and prediction of disease severity in common respiratory conditions like asthma,10 chronic bronchitis,11 and croup.12,13 Diagnosis of respiratory conditions in low-resource regions is fraught with difficulties due to the lack of field-deployable imaging and laboratory facilities, as well as the scarcity of trained community healthcare workers, and cough sound biomarkers could help improve screening and diagnostics in such settings.

However, there is currently no consensus on protocols for cough sounds recording and analysis to guide research teams and aid clinical translation. This scoping review is presented to overview the existing literature and propose priorities for future research, including data collection standards. The primary objective is to survey the bioacoustics literature on cough sounds to highlight data collection protocols, acoustic analysis approaches, analytical accuracy, and limitations. The secondary objectives are to determine which disciplines and regions are active in cough sound analytics, which diagnoses and populations are a focus, and whether existing cough sound databases are open-access and represent a diverse population. Of note, we included animal studies, as use of cough sounds for livestock respiratory illness monitoring is a rich field of study in veterinary sciences with little reference in the medical literature, despite significant scientific overlap.

Methods

The protocol was executed in accordance with the Preferred Reporting Items for Systematic reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) checklist.14 The review protocol, including the objectives, inclusion criteria, and methods were specified in advance and made available on September 3rd, 2021 under the Open Science Framework platform (https://osf.io/gh9t2/).

Search strategy

A comprehensive search strategy was developed in consultation with a scholarly medical librarian combining relevant subject headings and keywords related to ‘cough sounds.’ The search was conducted in PubMed, EMBASE, Web of Science, Scopus, Cochrane Library, IEEE Xplore, Engineering Village, and ACM Digital Library from inception to August 4th, 2021. The search strategy is included in Supporting Information, Appendix 1. The following gray literature sources were also searched: arXiv, medRxiv, engrXiv, preprints.org, Open Science Framework preprints, TechRxiv, and Google Scholar.

Study Selection

Records were included if they analyzed cough sounds for screening or diagnosis of disease in human or animal subjects. There were no restrictions based on the year of publication, subject age, study techniques, or study location. Only studies with the full text in the English language were included. Narrative reviews, systematic and scoping reviews, research in progress, books, book chapters, conference or meeting abstracts, dissertations, theses, newspaper/magazine articles, editorials, commentaries, and letters to the editor were excluded.

Results were uploaded into Covidence systematic review screening software (Veritas Health Innovation, Melbourne, Australia) and de-duplicated. All abstracts (n=433) were reviewed independently by two researchers (S.H. and S.S.) for inclusion eligibility. First, titles and abstracts and then, full texts were independently screened and charted. Disagreements between reviewers were resolved by consensus, and the senior investigators (K.M. and A.R.) were consulted when needed. Inter-rater reliability of the reviewers’ assessment was reported in the form of Cohen’s kappa.15

Data Extraction

Following agreement on the data extraction fields and their definition, the research team developed a data extraction sheet, and tested it iteratively for consistent data capture.

Data extraction was performed in Covidence and then exported for analysis in Microsoft Excel. Data extraction definitions are included in Supporting Information, Appendix 2. Data were independently extracted from selected articles by two team members (S.H, S.S), and verified by a senior author (K.M., A.R.). Descriptive statistics were used to highlight the frequency of modalities of cough sound collection, storage and analysis, and the pathologic conditions represented in the included results.

Results

The literature search yielded 433 unique results after deduplication. Following title and abstract screening, 183 studies were excluded. Ultimately, 108 studies underwent a full-text review. The reviewer agreement was moderate for title and abstract screening (Cohen’s kappa=0.53) and high for full-text review (Cohen’s kappa=0.95). The PRISMA flow diagram outlining the study selection process is delineated in Figure 1. Citations for all included studies is available in Supporting Information, Appendix 3. A summary of all included studies can be found in Table S1.

Figure 1.

Figure 1.

Preferred reporting items for Systematic reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) flow diagram.

Author Affiliations and Journals

The body of literature was primarily authored by science, technology, engineering, and mathematics (STEM) specialists (68/108, 62.9%), followed by medical (35/108, 32.4%) and veterinary science specialists (5/108, 4.6%) (Figure 2A). A study was included in an expertise category if at least one author’s appointment was in that category. Co-authorship of STEM specialists and physicians and veterinarians was present in 34.3% (37/108) and 9.3% (10/108) of articles respectively. The articles were most commonly published in engineering journals (35/108, 32.4%) (Figure 2B). A wide range of geographic regions were represented in the published data; Europe (47/108, 43.5%) and Asia (28/108, 25.9%) were the most common continents (Figure 2C).

Figure 2.

Figure 2.

Pie charts showing the specialties of first authors (A), the fields represented by journals in which reviewed studies were published (B), and the region in which data were collected for each study (C). STEM = science, technology, engineering, and math.

Demographics and Clinical Profiles

Human studies were most common (84/108, 77.8%); of the studies which disclosed the age of their study population, the majority focused on adult subjects (35/61, 57.3%), while 31.1% (19/61) focused on pediatric patients (Figure 3B). The most common livestock species studied were pigs (16/23, 66.7%) and bovine cattle (3/23, 13.0%); guinea pigs, mice, and chickens were also subjects of investigation in very few studies (Figure 3C). Only one study involved both human and animal participants. Among human studies, combined adult and pediatric subjects were included in 7 studies. In human studies, demographic information was reported in 76.5% of studies (65/85). Age, sex, and body weight were the most reported clinico-demographic data, whereas race/ethnicity, gender and height were the least reported.

Figure 3.

Figure 3.

Pie charts illustrating type of study (A), population studied (B), animal species represented in animal studies (C), analytic focus of the studies (D), and the type of cough analyzed in each study (E).

Clinical information, such as past medical history and co-morbidities, were overall inconsistently described across the reviewed articles. In human studies, 36 (61/85, 71.7%) studies reported at least some patient co-morbidities of which, asthma (34/61, 55.7%), chronic obstructive pulmonary disease (COPD) (14/61, 22.9%) and bronchiectasis (7/61, 11.5%) were most common.

Screening/Diagnostic Focus

Cough detection tools were used to monitor subjects’ cough numbers throughout a given time frame. This is in contrast with cough quality classification tools, which aimed to detect the coughs first and then categorize the cough quality based on type (wet, dry), pitch (often for sex differentiation), and/or severity. Cough diagnosis classification studies aimed to determine the pathologic condition underlying cough, such as pneumonia, asthma, or COVID-19. Figure 3D describes the proportion of reviewed studies in each category, while Table 1 divides the studies by both category and authorship.

Table 1:

Distribution of author specialties organized by study type

Type of study Papers With At Least One Author in STEM Papers With At Least One Author in Medicine Papers With Physician and STEM Authors Papers With At Least One Author in Veterinary Science
Cough Detection (45) 37 (82.2%) 22 (48.9%) 12 (26.7%) 7 (15.6%)
Cough Quality Classification (16) 9 (56.3%) 13 (81.3%) 5 (31.3%) 0 (0%)
Cough Diagnosis Classification (47) 39 (83.0%) 26 (55.3%) 20 (42.6%) 5 (10.6%)

STEM = science, technology, engineering, and math.

A large number of studies focused on cough detection (45/108, 41.7%) in adults (19/35, 54.2%), children (2/19, 10.5%), and animals (14/23, 60.9%), as well as 43.5% (10/23) of human studies with unspecified age group. A wide number of pathologies were represented in human cough detection studies (Figure 4B). For the 38 studies focused on diagnostics in human subjects, the pathologies examined included most commonly COVID-19 (12/38, 31.6%), asthma (11/38, 28.9%), pneumonia (5/38, 13.2%), pertussis (4/38, 10.5%), croup (3/38, 7.9%), COPD (3/38, 7.9%), and tuberculosis (3/38, 7.9%); of note, many studies focused on diagnosing multiple diseases (Figure 4A). In pediatric studies, the most common diagnostic focus was asthma (10/19, 52.6%). Comparison to a diagnostic gold standard such as chest radiograph or microbiology result was explicitly reported in 21.1% (8/38) of diagnosis classification studies. Only 17 studies (17/108, 15.7%) attempted classification of cough severity. Veterinary diagnostic classification studies were mostly focused on infectious etiologies of cough (like porcine wasting disease, bovine respiratory disease, actinobacillus/pasteurella infections, bronchopneumonia, and other bacterial/viral infections (7/8, 87.5%) rather than chronic non-infectious etiologies, though one study focused on aerial pollutants.

Figure 4.

Figure 4.

Bar graphs showing pathologies diagnosed using cough classification models (A) and the pathologies included in cough detection studies (B).

Data acquisition protocols

Microphone and sensor types:

The use of an external microphone for data recording was most common (82/108, 75.9%). Twenty-nine studies mentioned the use of a condenser microphone (6 unidirectional, 5 omnidirectional, 11 electret). Three studies also relied on piezoelectric sensors. Sixteen studies used contact microphones, 6 studies mouthpieces and 9 studies lapel microphones. Twenty studies used smartphones, while 14 used publicly available or crowdsourced data, often of unknown or inconsistent recording methodology (Figure 5A). Twelve studies used multiple microphones, with some veterinary studies using as many as eight microphones. One study employed the use of earphones attached to the smartphone to record samples. All reviewed veterinary studies used non-contact microphones in order to avoid affecting the behavior of animals. Cough sounds were recorded from a group of animals rather than individual animals in 39% (9/23) veterinary studies, while the others recorded animals one by one; of these veterinary studies, 55.6% (5/9) mentioned the location of the microphone relative to the animal.

Figure 5.

Figure 5.

Graphs showing prevalence of data collection methods (A) and data analysis methods (B) over time. Crowdsourced and publicly available recordings usually had unspecified or variable recording methodology.

Recording protocol:

Voluntary cough sounds are defined as those produced on instruction from non-symptomatic human subjects,16,17 whereas reflexive cough sounds arise as a spontaneous primitive response to protect the airways.11 Reflexive coughs were captured in most of the studies (55/108, 50.9%), followed by voluntary coughs (28/108, 16.7%) and both types (11/108, 10.2%); 14 studies (13.0%) did not specify whether coughs were voluntary or reflexive (Figure 3E). Distance between the microphone and subject’s mouth was mentioned in 16 human studies. The average distance from the subject’s mouth and the microphone was 62.0 cm (median 50cm, range: 15–400 cm).

Recording environment:

Among human studies focused on cough detection, the most common setting was continuous ambulatory recording, in which patients went about their daily routines with a microphone or other cough monitor (11/31, 35.5%). Other environments included isolated or quiet rooms, in purely experimental settings (7/31, 22.6%); inpatient (3/31, 9.7%) or outpatient (2/31, 6.5%) clinical settings; fixed microphones in a bedroom or other space in patients’ homes (3/31, 9.7%); or public spaces, such as a hospital waiting room or library (2/31, 6.5%). Three or these studies (9.7%) did not specify the recording setting.

For cough quality classification studies in humans, the most common environment was a quiet room or other isolated experimental setup (10/14, 71.4%). Two studies (14.3%) involved recordings in an outpatient clinical setting, and one was recorded in an inpatient hospital setting. Finally, one study captured continuous recordings in patients’ bedrooms.

The most commonly represented environment among human studies focused on cough diagnosis classification was isolated rooms (11/39, 28.2%). Nine studies (23.1%) were recorded in inpatient clinical settings, and three (9.7%) were recorded in outpatient settings. Four studies (12.9%) used recordings from multiple settings, such as both inpatient and outpatient environments. One study (3.2%) obtained recordings from patients at home. Eleven additional studies (28.2%) did not specify a recording setting.

Finally, among animal studies, the most common recording environment was on farms (10/23, 43.5%), closely followed by laboratory settings (9/23, 39.1%). Four studies (17.4%) used recordings from both a laboratory and a farm.

Dataset characteristics

A minority of studies (9/108, 8.3%) did not report on dataset sizes, and more than a quarter of studies (31/108, 28.7%) did not report on the number of cough samples recorded. In those studies specifying dataset size and number of cough sounds collected, an average of 2357 cough samples (2357 ± 5086 coughs, median: 804.5, interquartile range: 1955, range: 3 to 30304) were collected in each study, with an average of 617 subjects (617 ± 2478 subjects, median: 32, interquartile range: 97, range: 2 to 20,000) and an average of 3.82 coughs were recorded per subject.

Single modality in data collection prevailed, with 71.2% (77/108) of studies only collecting acoustic recording for the analysis. Few studies described a multi-modal data collection approach including plethysmography (17/108, 15.7%) or clinical data such as temperature, respiratory rate, or body mass index (8/108, 7.4%). Patient reported outcome measure questionnaires such as the cough visual analogue scale and the Leicester Cough Questionnaire were utilized in four studies. Other multi-modal approaches relied on chest radiographs (1/108, 1%), thermal imaging (2/114, 1.8%), demographic features (3/108, 2.8%), video recording (1/108, 1%), and pressure/flow measurements (3/108, 2.8%).

Data analytics approach

Preprocessing

Preprocessing of the acoustic signal was done by de-clipping, de-noising, re-sampling, or windowing in (82/108) 75.9% of all the studies. Various methods of noise reduction, i.e., improving signal to noise ratio, were specifically studied in 9 (9/108, 8.3%) studies.

Data analytics

Data analytics were highly variable across studies and across time. The majority of studies relied on machine learning (66/108, 61.1%) for either cough detection or diagnosis classification, with the bulk of such studies (52/66, 78.8%) published after 2010 (Figure 5B).

The accuracy and sensitivity of the cough recognition models had wide ranges, 40%−100% and 32%−99.1% respectively. Human studies were more likely to report accuracy data (47/85, 55.3%) as compared to animal studies (10/23, 43.4%). Human studies had a mean accuracy of 88.2% whereas the average animal study accuracy was higher at 93.1%. Accuracy largely improved over time as nonlinear data analytics offered by machine learning models became more sophisticated. However, among the studies that reported accuracy, only two studies (2/57, 3.5%) used entirely separate datasets for training and testing of their algorithms. Much more common (27/57, 47.4%) was the subsetting of one central dataset for training and testing a machine learning algorithm, while four studies (8.5%) used a version of leave-one-out cross-validation. Studies that did not specify what data was used for training and testing of their machine learning model, or studies that used the exact same data for testing and training, were deemed to have unreliable accuracy measurement; such accuracy statistics were not included in Table S1.

Publication Years

Publication dates ranged from 1980 to 2021, with the majority of the studies being conducted in the recent decade of 2012–2021 (61/108, 56.5%). Figure 5 depicts the frequency of each modality for cough sound collection and data analysis segmented by decade.

Dataset origin

Single institution studies with closed-access datasets were more common (87.0%). A minority, but growing number of studies, (13.0%) relied on open-access datasets. Some of the popular datasets are described in Table S2.

Limitations

Fifty-seven (57/108, 52.8%) of all studies reported limitations in their approach or models. Several studies reported on a limited sample size (12/108, 11.1%). In the animal studies, several authors cited interference with ambient noise and limited number of microphones, leading to poor area coverage on big farms.

Discussion

This scoping review aimed to examine the utilization of cough sounds in screening and diagnostics. The number of publications analyzing cough acoustics over the past decade has been rapidly increasing, which possibly reflects technological advances in recording capabilities – especially with the ubiquitous availability of smartphone recording technology and improved aptitude for analysis of complex data with the advent of machine learning. Our study is the first scoping review of the cough sound literature that is agnostic to the analytic methodology. Other unique features of our scoping review is the incorporation of adult, pediatric, and animal studies, surveying the field broadly across disciplines. This is in contrast with Serrurier et al.,18 who performed a comparative literature review of the cough sound literature without standardized methodology. In addition, our publication includes studies from veterinary science, in which many cough analysis advances have been made in recent years that could be relevant to human health. We also specifically categorized the human literature into adult and pediatric studies.

The authors found the primary objective of the plurality of studies was diagnostic classification of respiratory illnesses such as COVID-19, COPD, asthma and tuberculosis in the adult study population, and asthma, pneumonia, pertussis and croup in the pediatric population. The number of studies focused on cough detection followed closely. Less commonly, some studies focused on classifying cough sounds based on character and severity.

Recording protocols were highly variable and had several limitations. For instance, though most studies relied on reflexive cough sounds, among those that collected voluntary cough sounds, only a fraction of studies described the specific instructions for subjects to produce cough. Though the presence of background noise may affect the quality of cough sound collection, only a minority of studies utilized a completely noise-free room and nearly a quarter of the studies (23%) did not specify the recording environment.

With respect to hardware for cough sound acquisition, we found the majority of studies used an external microphone. The type of external microphones varied with the aim of the study. For example, cough counting studies often used lapel microphones to ensure comfort of human subjects during continuous data collection. In contrast, studies focused on the quality of the cough sounds were more likely to use contact microphones to reduce noise by deriving audio signals directly from mechanical vibrations transmitted from the point of physical contact with the subject. Contact microphones may pick up otherwise inaudible acoustic data, potentially improving differentiation of cough sound types.

In recent years, cellular phones, with their wide availability and rapidly evolving hardware and software sophistication, have quickly replaced individual condenser microphones and associated recorder via non-contact recording. The use of smartphones was first noted in 2011, becoming more predominant from 2016 onwards. Smartphones offer the convenience of built-in microphones, data storage and processing software under one device. Smartphones also stand out as convenient and ubiquitous for mobile data collection at a population level, and hold the potential of continuous monitoring at the individual level. However, continuous audio sensing by smartphone poses complex privacy concerns, which have limited implementation and adoption.19

Nearly all studies used preprocessing techniques of the cough sound prior to running an acoustic analysis. Most acoustic analysis relied on spectrograms, which represent the spectro-temporal correlations present in the audio signal in the form of an image. Of particular concern for cough sounds, which are explosive acoustic events, is the phenomenon of clipping, i.e. the loss of high amplitude acoustic data when an audio signal is too loud. To avoid clipping in the recording of cough sounds, the distance between the subject’s mouth and the microphone becomes pertinent. Additionally, if the subject is too close to the microphone, gusts of air cause distortion of audio data, compromising its quality.20 Relatively few studies mentioned the exact distance between the subject and the microphone, with the average distance being 50 cm.

Irrespective of their objective, the majority of studies focused on single modality of cough acoustic data in their predictive models. The few studies that used a multi-modality approach combined cough sounds with demographic data or clinical data such as body temperature or respiratory rate. A handful of studies also supplemented their acoustic data with patient reported outcome measures, such as the cough visual analogue scale and the Leicester Cough Questionnaire. A multi-modality approach has been associated with an increase in the precision of disease predictive models based on cough sounds.21

Developments in the cough sound literature correlate with technological advancements in audio recording and data analytics. For instance, ambulatory audio recording devices appeared in the 1990s, and enabled continuous monitoring and real-time data capture of cough sounds. In the following decades, the introduction of the Fast Fourier Transform (FFT) underlying spectrograms and Mel-frequency cepstral coefficients (MFCCs) allowed researchers to capture more acoustic data dimensions.22 Additionally, in the past decade, the development of machine learning based on artificial neural networks expanded the ability to analyze vast amounts of complex digitized data, allowing researchers to exploit the potential of cough sounds at a much greater scale and with higher level of predictive precision. Researchers now have programs that simplify and increase the efficiency of data pre-processing, with such capabilities as distinguishing and tagging the contaminant background noises and allowing for drastic increase in the magnitude of data incorporated in predictive models. Data storage has also recently shifted online, mostly to the cloud, making accessibility of data easy and convenient. Availability of open-access large datasets of cough sounds has further contributed to the acceleration of publications in the past decade.

Livestock studies demonstrated that a one-to-one ratio of microphone to subject yields the best results in terms of audio acquisition quality. Factors which influenced hardware choice for animal cough sound recording included body weight and the species of animal involved. Most animal studies focused on field recording, with limitations related to ambient noise and poor discrimination between individual animals. This approach was particularly effective in respiratory infection spread monitoring and localization of sick animals. Interestingly, such field recordings have not been implemented for humans, likely due to challenges related to privacy concerns.

Our scoping review revealed trends of drawbacks amongst the studies reviewed. Included studies were authored by scientists in resource-affluent countries, despite the potential for higher impact in low-resource settings. In addition, study populations were consistently small, and there was a lack of prospective studies. Inclusion of demographic features were inconsistent across studies, and as highlighted in other scoping reviews in bioacoustics, we found a lack of reporting on race/ethnicity, gender, and socio-economic status of participants in human studies.23

This scoping review attempted to examine all relevant articles to cough sounds in screening and diagnostics in humans and animals published in English, though our selected search terms may not have captured all the relevant literature. The lack of critical appraisal of the included literature is inherent to scoping reviews and limits our ability to report on the quality of the medical evidence in this field of bioacoustics.

This study is also limited by the date range of the literature search, which does not include advances in the literature in the past two years as the field has continued to expand. Notably, diagnosis of COVID-19 from cough sounds using machine learning has remained a frequent focus.2426 For instance, a recent study by Davidson et al. sought to classify COVID-19-related pneumonia severity based on cough sounds.27 Kuluozturk et al. used a machine learning model to diagnose COVID-19 as well as heart failure and acute asthma based on cough sounds,28 while Yellapu et al. focused on the diagnosis of pulmonary tuberculosis.29 Novel cough detection algorithms have also continued to be developed,30,31 and Xu et al. proposed a smartphone-based cough sound analysis as a potential home-based pulmonary function test.32 Importantly, the effort to expand publicly available cough sound datasets has also continued, improving the landscape for better training and testing of future machine learning models.33

While no cough detection or diagnostic technology has yet achieved FDA approval, multiple cough monitoring devices, including VitaloJak and the Leicester Cough Monitor, have been used in research.34 More recently, several app-based ventures, notably ResApp (for cough sound diagnosis)35 and Hyfe (for cough counting),36 have continued to develop rapidly and are likely to apply for regulatory approval for clinical use in the near future. Should this technology become clinically available, it has the potential to significantly impact the practice of otolaryngologists. For example, cough diagnosis algorithms could play a role in the diagnosis of conditions such as croup, for which cough sounds have diagnostic value. Cough detection and cough counting technology could transform the management of neurogenic cough managed by otolaryngologists, allowing for reliable and objective assessment of response to treatment as measured by cough number and frequency.

Conclusion

Cough sound analysis holds the promise of accessible, non-invasive, and inexpensive diagnostic tools but is limited in practice currently. The majority of current studies hinge on non-standardized data collection protocols and small, non-diverse, private datasets from single institutions. To build a robust model with wide utility and applicability, researchers need a diverse population, in order to capture all possible cough sound variants. The advent of machine learning and the ubiquity of smartphone technology with high quality audio recording capabilities in the past decade has accelerated interest in cough acoustics for screening and diagnosis.

Supplementary Material

Supinfo

Table S2. Open access datasets with access links and their availability.

Supplementary Appendix 1. Search strategy.

Supplementary Appendix 2. Data extraction definitions.

Supplementary Appendix 3. Citations for all included articles.

Table S1

Table S1. Summary of included studies organized by subject area.

Funding:

Anaïs Rameau was supported by a Paul B. Beeson Emerging Leaders Career Development Award in Aging (K76 AG079040) from the National Institute on Aging and by the Bridge2AI award (OT2 OD032720) from the NIH Common Fund.

Footnotes

Conflicts of Interest:

Anaïs Rameau owns equity of Perceptron Health, Inc.

This project was presented as a poster at the 2022 American Broncho-Esophagological Association Meeting in April 2022 in Dallas, TX.

References

  • 1.Widdicombe J, Fontana G. Cough: what’s in a name? Eur Respir J. 2006;28(1):10–15. doi: 10.1183/09031936.06.00096905 [DOI] [PubMed] [Google Scholar]
  • 2.Brooks SM. Perspective on the human cough reflex. Cough Lond Engl. 2011;7:10. doi: 10.1186/1745-9974-7-10 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Leden Von, Isshiki N AN ANALYSIS OF COUGH AT THE LEVEL OF THE LARYNX. Arch Otolaryngol Chic Ill 1960. 1965;81:616–625. doi: 10.1001/archotol.1965.00750050631016 [DOI] [PubMed] [Google Scholar]
  • 4.Kosasih K, Abeyratne UR, Swarnkar V. High frequency analysis of cough sounds in pediatric patients with respiratory diseases. Conf Proc Annu Int Conf IEEE Eng Med Biol Soc IEEE Eng Med Biol Soc Conf. 2012;2012((Kosasih K) School of ITEE, The University of Queensland, Brisbane, Australia.):5654–5657. [DOI] [PubMed] [Google Scholar]
  • 5.Korpas J, Vrabec M, Sadlonova J, Salat D, Debreczeni LA. Analysis of the cough sound frequency in adults and children with bronchial asthma. Acta Physiol Hung. 2003;90(1):27–34. doi: 10.1556/APhysiol.90.2003.1.4 [DOI] [PubMed] [Google Scholar]
  • 6.Abaza AA, Day JB, Reynolds JS, et al. Classification of voluntary cough sound and airflow patterns for detecting abnormal pulmonary function. Cough. 2009;5(1):8. doi: 10.1186/1745-9974-5-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Melek M Diagnosis of COVID-19 and non-COVID-19 patients by classifying only a single cough sound. Neural Comput Appl. Published online 2021:1–12. doi: 10.1007/s00521-021-06346-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Khanzada A, Hegde S, Sreeram S, et al. Challenges and Opportunities in Deploying COVID-19 Cough AI Systems. J Voice Off J Voice Found. 2021;35(6):811–812. doi: 10.1016/j.jvoice.2021.08.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Melek Manshouri N Identifying COVID-19 by using spectral analysis of cough recordings: a distinctive classification study. Cogn Neurodyn. Published online July 29, 2021:1–15. doi: 10.1007/s11571-021-09695-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Toop LJ, Thorpe CW, Fright R. Cough sound analysis: A new tool for the diagnosis of asthma? Fam Pract. 1989;6(2):83–85. [DOI] [PubMed] [Google Scholar]
  • 11.Piirila P, Sovijarvi AR. Differences in acoustic and dynamic characteristics of spontaneous cough in pulmonary diseases. Chest. 1989;96(1):46–53. doi: 10.1378/chest.96.1.46 [DOI] [PubMed] [Google Scholar]
  • 12.Sharan RV, Abeyratne UR, Swarnkar VR, Porter P. Cough sound analysis for diagnosing croup in pediatric patients using biologically inspired features. Conf Proc Annu Int Conf IEEE Eng Med Biol Soc IEEE Eng Med Biol Soc Annu Conf. 2017;2017((Sharan RV; Abeyratne UR; Swarnkar VR; Porter P)):4578–4581. doi: 10.1109/EMBC.2017.8037875 [DOI] [PubMed] [Google Scholar]
  • 13.Sharan RV, Abeyratne UR, Swarnkar VR, Porter P. Automatic croup diagnosis using cough sound recognition. IEEE Trans Biomed Eng. 2019;66(2):485–495. doi: 10.1109/TBME.2018.2849502 [DOI] [PubMed] [Google Scholar]
  • 14.Kapadia MZ, Askie L, Hartling L, et al. PRISMA-Children (C) and PRISMA-Protocol for Children (P-C) Extensions: a study protocol for the development of guidelines for the conduct and reporting of systematic reviews and meta-analyses of newborn and child health research. BMJ Open. 2016;6(4):e010270. doi: 10.1136/bmjopen-2015-010270 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.McHugh ML. Interrater reliability: the kappa statistic. Biochem Medica. 2012;22(3):276–282. [PMC free article] [PubMed] [Google Scholar]
  • 16.Van Hirtum A, Berckmans D. Automated recognition of spontaneous versus voluntary cough. Med Eng Phys. 2002;24(7–8):541–545. doi: 10.1016/S1350-4533(02)00056-5 [DOI] [PubMed] [Google Scholar]
  • 17.Van Hirtum A, Berckmans D. Assessing the sound of cough towards vocality. Med Eng Phys. 2002;24(7–8):535–540. doi: 10.1016/S1350-4533(02)00055-3 [DOI] [PubMed] [Google Scholar]
  • 18.Serrurier A, Neuschaefer-Rube C, Röhrig R. Past and Trends in Cough Sound Acquisition, Automatic Detection and Automatic Classification: A Comparative Review. Sensors (Basel). 2022. Apr 10;22(8):2896. doi: 10.3390/s22082896. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Agu E, Pedersen P, Strong D, et al. The smartphone as a medical device: Assessing enablers, benefits and challenges. In: 2013 IEEE International Conference on Sensing, Communications and Networking (SECON). ; 2013:76–80. doi: 10.1109/SAHCN.2013.6644964 [DOI] [Google Scholar]
  • 20.Stinson MR, Daigle GA, Quaroni JF. Airflow noise in telephone handsets and methods for its reduction. J Acoust Soc Am. 2005;118(1):205–212. doi: 10.1121/1.1931087 [DOI] [PubMed] [Google Scholar]
  • 21.Kumar S, Nagar R, Bhatnagar S, Vaddi R, Gupta SK, Rashid M, Bashir AK, Alkhalifah T. Chest X ray and cough sample based deep learning framework for accurate diagnosis of COVID-19. Comput Electr Eng. 2022. Oct;103:108391. doi: 10.1016/j.compeleceng.2022.108391 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Küçükbay SE, Sert M. Audio-based event detection in office live environments using optimized MFCC-SVM approach. In: Proceedings of the 2015 IEEE 9th International Conference on Semantic Computing (IEEE ICSC 2015). ; 2015:475–480. doi: 10.1109/ICOSC.2015.7050855 [DOI] [Google Scholar]
  • 23.Bensoussan Y, Vanstrum EB, Johns MM 3rd, Rameau A. Artificial Intelligence and Laryngeal Cancer: From Screening to Prognosis: A State of the Art Review. Otolaryngol Head Neck Surg. 2023. Mar;168(3):319–329. doi: 10.1177/01945998221110839. [DOI] [PubMed] [Google Scholar]
  • 24.Celik G CovidCoughNet: A new method based on convolutional neural networks and deep feature extraction using pitch-shifting data augmentation for covid-19 detection from cough, breath, and voice signals. Comput Biol Med. 2023. Jun 8;163:107153. doi: 10.1016/j.compbiomed.2023.107153. Epub ahead of print. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Pavel I, Ciocoiu IB. COVID-19 Detection from Cough Recordings Using Bag-of-Words Classifiers. Sensors (Basel). 2023. May 23;23(11):4996. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Chowdhury NK, Kabir MA, Rahman MM, Islam SMS. Machine learning for detecting COVID-19 from cough sounds: An ensemble-based MCDM method. Comput Biol Med. 2022. Jun;145:105405. doi: 10.1016/j.compbiomed.2022.105405. Epub 2022 Mar 17. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Davidson C, Caguana OA, Lozano-García M, Arita Guevara M, Estrada-Petrocelli L, Ferrer-Lluis I, Castillo-Escario Y, Ausín P, Gea J, Jané R. Differences in acoustic features of cough by pneumonia severity in patients with COVID-19: a cross-sectional study. ERJ Open Res. 2023. May 2;9(3):00247–2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Kuluozturk M, Kobat MA, Barua PD, Dogan S, Tuncer T, Tan RS, Ciaccio EJ, Acharya UR. DKPNet41: Directed knight pattern network-based cough sound classification model for automatic disease diagnosis. Med Eng Phys. 2022. Dec;110:103870. doi: 10.1016/j.medengphy.2022.103870. Epub 2022 Aug 6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Yellapu GD, Rudraraju G, Sripada NR, Mamidgi B, Jalukuru C, Firmal P, Yechuri V, Varanasi S, Peddireddi VS, Bhimarasetty DM, Kanisetti S, Joshi N, Mohapatra P, Pamarthi K. Development and clinical validation of Swaasa AI platform for screening and prioritization of pulmonary TB. Sci Rep. 2023. Mar 23;13(1):4740. doi: 10.1038/s41598-023-31772-9. Erratum in: Sci Rep. 2023 Jun 26;13(1):10353. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.You M, Wang W, Li Y, Liu J, Xu X, Qiu Z. Automatic cough detection from realistic audio recordings using C-BiLSTM with boundary regression. Biomed Signal Process Control. 2022. Feb;72:103304. doi: 10.1016/j.bspc.2021.103304. Epub 2021 Nov 11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Barata F, Cleres D, Tinschert P, Iris Shih CH, Rassouli F, Boesch M, Brutsche M, Fleisch E. Nighttime Continuous Contactless Smartphone-Based Cough Monitoring for the Ward: Validation Study. JMIR Form Res. 2023. Feb 20;7:e38439. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Xu W, He G, Pan C, Shen D, Zhang N, Jiang P, Liu F, Chen J. A forced cough sound based pulmonary function assessment method by using machine learning. Front Public Health. 2022. Oct 25;10:1015876. doi: 10.3389/fpubh.2022.1015876. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Bhattacharya D, Sharma NK, Dutta D, Chetupalli SR, Mote P, Ganapathy S, Chandrakiran C, Nori S, Suhail KK, Gonuguntla S, Alagesan M. Coswara: A respiratory sounds and symptoms dataset for remote screening of SARS-CoV-2 infection. Sci Data. 2023. Jun 22;10(1):397. doi: 10.1038/s41597-023-02266-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Zhang M, Sykes DL, Brindle K, Sadofsky LR, Morice AH. Chronic cough-the limitation and advances in assessment techniques. J Thorac Dis. 2022. Dec;14(12):5097–5119. doi: 10.21037/jtd-22-874. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Pfizer offers AU$100 million for Australian company that developed COVID-19 cough diagnostic app. MobiHealthNews. Published April 21, 2022. Accessed July 26, 2023. https://www.mobihealthnews.com/news/anz/pfizer-offers-au100-million-australian-company-developed-covid-19-cough-diagnostic-app#:~:text=Biopharmaceutical%20giant%20Pfizer%20has%20offered [Google Scholar]
  • 36.Hammerand J Cough-counting is harder than it sounds, but AI and smartphones can help. Medical Design and Outsourcing. Published April 20, 2023. Accessed July 26, 2023. https://www.medicaldesignandoutsourcing.com/hyfe-chronic-cough-detection-software-counting-ai-smartphones/ [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supinfo

Table S2. Open access datasets with access links and their availability.

Supplementary Appendix 1. Search strategy.

Supplementary Appendix 2. Data extraction definitions.

Supplementary Appendix 3. Citations for all included articles.

Table S1

Table S1. Summary of included studies organized by subject area.

RESOURCES