Abstract
Goal: Official tests for COVID-19 are time consuming, costly, can produce high false negatives, use up vital chemicals and may violate social distancing laws. Therefore, a fast and reliable additional solution using recordings of cough, breathing and speech data for preliminary screening may help alleviate these issues. Objective: This scoping review explores how Artificial Intelligence (AI) technology aims to detect COVID-19 disease by using cough, breathing and speech recordings, as reported in the literature. Here, we describe and summarize attributes of the identified AI techniques and datasets used for their implementation. Methods: A scoping review was conducted following the guidelines of PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews). Electronic databases (Google Scholar, Science Direct, and IEEE Xplore) were searched between 1st April 2020 and 15th August 2021. Terms were selected based on the target intervention (i.e., AI), the target disease (i.e., COVID-19) and acoustic correlates of the disease (i.e., speech, breathing and cough). A narrative approach was used to summarize the extracted data. Results: 24 studies and 8 Apps out of the 86 retrieved studies met the inclusion criteria. Half of the publications and Apps were from the USA. The most prominent AI architecture used was a convolutional neural network, followed by a recurrent neural network. AI models were mainly trained, tested and run-on websites and personal computers, rather than on phone apps. More than half of the included studies reported area-under-the-curve performance of greater than 0.90 on symptomatic and negative datasets while one study achieved 100% sensitivity in predicting asymptomatic COVID-19 from cough-, breathing- or speech-based acoustic features. Conclusions: The included studies show that AI has the potential to help detect COVID-19 using cough, breathing and speech samples. The proposed methods (with some time and appropriate clinical testing) could prove to be an effective method in detecting various diseases related to respiratory and neurophysiological changes in the human body.
Keywords: COVID-19, artificial intelligence, machine learning, cough, speech signals, acoustics, breathing
I. Introduction
On March 11th 2020, the World Health Organization (WHO) announced that the COVID-19 outbreak had become a pandemic [1]. Fig. 1 depicts the weekly global report by the WHO on the pandemic in various regions from December 2019 to October 2020 [2].
Fig. 1.
Weekly reports of cases and deaths in various regions [2].
The most common symptoms of COVID-19 include high fever, new cough (may be dry), shortness of breath and loss or change to your sense of smell or taste [3], [4] as well as other neurological effects [5], [6]. Currently, COVID-19 can be detected in two ways: i) Virus detection through nose or throat swabs using a – RT-PCR test and ii) tests which deduce the immune response of the body such as - Antigen and Serology tests [7]. However, multiple limitations are involved with both of these tests. Firstly, high-false negative rates were seen with mass RT-PCR testing [8]. Secondly, tests often require physical contact which may disturb social distancing guidelines. Finally, time and expertise are required for every individual result [9]. Therefore, there is a need to augment the existing techniques to include preliminary screening to reduce these issues.
Since April 2020, many research organizations began to develop interest in changes to speech and acoustic alterations associated with the virus. Cough being the most common symptom of many diseases, can be differentiated among cases and controls [10] by Artificial Intelligence (AI) and machine learning algorithms. As AI can use deep learning or other machine learning, it could provide better efficiency in detecting viruses in comparison to the recommended tests if it finds novel and predictive patterns [10]. Changes in speech, both read and naturalistic, as due to breathing difficulty, stuffiness and inflammation, likewise showing promise in providing a basis to distinguish healthy from unhealthy cases [11]. The example of speech in Fig. 2 reveals the motivation for using AI as a preliminary screening technique, including better tracking and faster detection of COVID-19. Fig. 2 shows a comparison of a speech waveform and FFT (Fast Fourier transform) of COVID-19 versus healthy cases. The unhealthy subject recording, made soon after a positive COVID-19 diagnosis, exhibits a slower speaking rate (almost half that of the healthy case), greater pause length and duration, less natural and crisp articulation, lower frequency dynamics and high-frequency distribution.
Fig. 2.
(i), (ii): This figure data was collected and analyzed with ethical approval from University Hospital Galway, Ireland.
The objective of this review is to examine studies of COVID-19 detection via features derived from speech, breathing and cough recording as presented in the literature. The results may be useful for various institutions developing AI algorithms that use speech signal processing for future pandemics or respiratory-based illnesses.
II. Methods
To fulfill the aims of this study while certifying an iterative and transparent methodology, a scoping review was conducted following the guidelines of PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) [12]. Below are detailed methods used for this review.
A. Study Search Resources
In this review, the period chosen for relevant queries was between 1st April 2020 and 15th August 2021. The searches were completed via online databases such as PUBMED, Google Scholar, Science Direct, arXiv and IEEE xplore. In addition, reference lists of included studies were also screened.
B. Study Search Terminologies
The search terms used were chosen in accordance with the algorithm used, target diseases and symptoms of that disease. The terms were “AI cough”, “machine learning COVID-19”, “coronavirus-19 speech”, “coronavirus-19 cough”, “artificial intelligence for viruses”, “artificial intelligence for coughing” and “COVID-19 breathing”.
C. Study Eligibility Criteria
The main focus of this review was AI based technology that identifies or predicts the presence of COVID-19 in symptomatic, asymptomatic or negative patients by analyzing recordings of cough, breathing and speech signals. Therefore, studies on non-COVID cough and artificial intelligence alone were excluded. Furthermore, studies involving research or potential use of AI in speech signal processing were also excluded. The selected studies for this review were published in English between 1st April 2020 and 15st August 2021. These include publications, preprints, dissertations, peer-reviewed articles and mobile Apps, excluding overviews, proposals and editorials. This review does not impose any restriction on study design, outcome of the study or country of publication.
D. Selection of Study
The studies initially were screened using titles and abstracts. After finding a study with a relevant title and abstract, full screening was conducted. Furthermore, the relevant studies were considered after eliminating the possibility of type of publication, population and intervention. Moreover, bibliographies of the selected studies were checked for further studies not found in the initial search.
E. Data Extraction and Synthesis
The data extraction was performed based on two main aspects: i) Datasets used in studies and ii) AI approach and architecture used. Table I demonstrates the detailed description of data description.
TABLE I. Data Extraction Format.
| Datasets | Description |
|---|---|
| Data Sources | Sources of data used in the studies for detecting COVID-19 via cough/breathing/speech recording – Public platforms, clinical settings, hospitals, government sources etc. |
| Data Size | The amount of data used to train, test and validate the algorithm |
| Data Collection | The method used for data collection such as open sources like apps or websites. |
| AI Characteristics | Description |
| AI Architecture | The architecture of AI algorithm such as CNN, RNN etc. |
| AI Branches | The type of AI model used such as: linear, ensemble, deep learning. |
| AI Platform | The platform used for AI processing such as computer or mobile. |
After data extraction, the process of data synthesis for the selected studies was conducted. The data integration was concluded on the foundation of AI implementation. This included the type of AI architecture and its platform such as computer, tablets or mobile phones.
Moreover, the variety of branches used for the implementation of the models was also considered. On the other hand, the group of datasets were regarded as a tool for synthesis. These involved collection of data from various sources through multiple platforms. For instance, some studies obtained open-source recording of cough, breathing and speech through websites or mobile Apps. In addition, the sample size of the data used for training, testing and validating was also taken into account in this scoping review.
Although not used as a search criterion, another important consideration are cough, breathing and speech features extracted from the acoustic recordings of each dataset as well as their physiologic interpretation. These characteristics are summarized in Tables II and III for some of the included studies.
TABLE II. Features Used By Different Studies.
| Papers | Features Used |
|---|---|
| Thomas F. Quatieri | Formants(F1-F3), Fundamental Frequency (F0), Harmonic-to-noise ratio (HNR), Cepstral Peak Prominence, Creak, Envelope |
| Piyush Bagad | Log-Melspectrogram |
| Ali Imran | Mel Frequency Cepstral Coefficients (MFCC's), Mel-spectrogram |
| Lara Orlandic | MFFC's, Zero-crossing Rate, Energy Level, Spectral based features, Band Power, Signal strength, Power Spectral Density, Energy Envelope Peak |
| Jordi Laguarta | MFCC's |
| Ahmed Fakhry | MFCC's, Mel-spectrogram, Clinical information |
| Neeraj Sharma | MFCC's, Mean Square Energy, Polynomial Fit to the spectrum, Zero Crossing Rate, Spectral based features |
TABLE III. A Physiological Interpretation of Features.
| Feature | Physiology | Perception |
|---|---|---|
| Formants | Vocal tract articulator movements | Defines consonant and vowel perception |
| MFCCs | Vocal tract articulator movements | Derived from spectrum of speech |
| Fundamental Frequency (F0) | Vocal fold vibrations | “Pitch” of speech |
| Harmonic-to-Noise Ratio (HNR) | Noise at the glottis | Hoarseness |
| Cepstral Peak Prominence (CPP) | Stability of vocal fold vibration | Jittery pitch |
| Creak | Compression of vocal folds – become slack and compact | Voice sounds like a creaky door, also seen as vocal fry |
| Speech Envelope | Contributions of the respiratory system and resonance-harmonics interaction to amplitude modulation of speech | Can be reflected in intensity or loudness of speech |
III. Results
A. Search Results for Studies
Initially 86 studies were extracted and identified through a search in multiple databases and search engines. Out of these, 7 duplicates were eliminated for further screening. The titles and abstracts of the remaining 79 studies were screened. As a result, 41 studies were removed for reasons detailed in the Supplementary Material. The remaining 38 studies were examined by a full text review. However, 16 among these were excluded as they did not follow the study eligibility criteria. Consequently, 22 studies were incorporated. In addition, 2 more studies were identified and included after examining the reference list and reading literature reviews of the selected studies. Altogether, 24 studies and 8 websites/Apps were included in this review. websites/Apps which focused on cough, breathing or speech recordings were included (see Supplementary Material)
B. Attributes of the Included Studies
Among the included studies, 8 were preprints and 16 were published articles in peer-reviewed journals (Table IV). Most of the studies were published between April 2020 - October 2020 and January 2021- July 2021. The included studies and Apps were conducted and launched in 12 countries. Nonetheless, more than half of these studies and applications, were from The USA, followed by India and U.K. with 6 studies and 2 applications respectively.
TABLE IV. Attributes of Studies and Applications Combined.
| ATTRIBUTES-Studies (n*=29) | ||||
|---|---|---|---|---|
| Paper Status-Studies (n*) | Submission Months (for year 2020 and 2021 combined)-Studies (n) | Country of Origin-Studies (n*) | ||
| Published – 16 | January – 3 | July - 3 | USA – 10 | Luxembourg – 1 |
| Preprint – 8 | February – 1 | August – 3 | India – 7 | Singapore – 1 |
| Apps – 8 | Mach – 2 | September – 2 | Switzerland – 1 | Peru – 1 |
| April – 3 | October – 2 | Canada – 2 | Russia – 1 | |
| May – 2 | November – 3 | Germany – 2 | UAE – 2 | |
| June – 3 | December – 0 | U.K. – 3 | South Africa – 1 | |
Abbreviations: n = number of studies.
n* = number of studies and applications.
C. Characteristics of Datasets Utilized for AI Modelling
As shown in Table V, open-source public databases were the most commonly used resources for developing and modeling of AI architecture [11], [13]–[17], [21]–[27]. In addition, eight studies utilized both the clinical settings and publicly open datasets [13], [18], [20], [23], [27]. However, four of the included studies extracted the data from news interviews and social media platform such as Twitter, Instagram, Telegram or YouTube to train the AI algorithm [11], [20], [44].
TABLE VI. AI Architecture of Examined Studies (n).
| AIa Architecture (n=21) | |||
|---|---|---|---|
| AI Branches b- Studies (n) | Input features | AI Algorithm c- Studies (n) | Platforms- Studies (n) |
| Deep Learning –18 | Cepstral Peak Prominence – 1 | Convolutional Neural Network – 18 | Computer/Websites –19 |
| Other Machine Learning – 13 | Mel- Spectrogram – 6 | ResNet–50 Convolutional Neural Network – 4 | Mobile – 4 |
| Mel – Frequency Cepstral Coefficients – 10 | Rectifier Neural Network –2 | ||
a = Artificial Intelligence.
b = Numbers do not add up because several studies have similar and more than one AI branch.
c = Numbers do not add up because several studies have similar and more than one AI architecture.
TABLE VII. COVID-19 Apps (N) Based on Cough, Breathing and Speech and Artificial Intelligence.
| Apps/Websites (N=9) | ||
|---|---|---|
| Stage Reached- Studies (N) | Data Stored a Types- Studies (N) | Platforms- Studies (N) |
| Audio Collection – 8 | Cough – 6 | Website – 7 |
| Diagnosis – 1 | Speech – 6 | Mobile – 2 |
| Medical Advice – 2 | Breathing – 2 | |
| Medical Questionnaire – 3 | ||
| COVID-19 Symptoms – 3 | ||
a = Numbers do not add up because several Apps have similar and more than one data stored types.
TABLE V. Datasets of Included Studies (n) Only.
| Data Sets-Studies (n=21) | ||
|---|---|---|
| Data Set Size- Studies (n) | Data Types a- Studies (n) | Data Sources b- Studies (n) |
| <1000 - 5 | Cough recording - 15 | Public databases - 14 |
| 1000 to 9999 - 11 | Cough/Speech/Breathing recording – 6 | Clinical settings - 8 |
| ≥ 10000 - 7 | Speech recording – 6 | Social Media - 4 |
| N/A - 1 | Cough/Breathing recording - 3 | Interviews/News - 2 |
| Questionnaires - 3 | Isolation Wards - 1 | |
| Sound samples – 2 | ||
a = Numbers do not add up as several studies collected more than one type of data.
b = Numbers do not add up as several studies collected their data from more than one data source.
The type of data collected and stored as datasets were as follows: cough recordings (eg, recording for 3 s) [14], [15], [17], [23], [25]–[27] cough, breathing and speech recording (reciting a sentence) [16], [19], [22], [27]–[30], speech collection [11], [20], [21], [24] and sound samples (eg, recording using certain sounds) [18]. Moreover, some studies also collected a questionnaire based on a medical history or COVID-19 symptoms (eg, underlying conditions, age, sex, temperature) [15], [17].
The range of data size was observed from 5 to ≈32000. 10 of the studies used features like MFCCs [11], [13]–[18], [25]–[30] while 19 studies included CNN [11], [13]–[16], [18], [21]–[30] to extract features from the datasets. The data size was between 1000 and 9999 in half of the studies (n=11) [14], [16], [19], [22], [25]–[27], [28]–[30] whereas seven of the studies had the sample size greater than 10000 [15], [18], [20], [21], [28]. The Supplementary Material includes the datasets of all the examined studies.
D. Characteristics of AI Architecture
Deep learning architectures were implemented in 18 studies [11], [13]–[20], [23], [27], [30]. On the other hand, one-third of the studies used both deep learning and non-deep learning architectures [15], [16], [19], [24], [28]–[30]. Four studies used branching architecture [17], [25], [29].
The most common AI model architecture recorded in more than half of the studies (n=18), was the Convolutional Neural Network (CNN) [11], [14], [16], [18], [20]–[30]. Six studies used signal and embedded processing techniques along with the Mel- Spectrogram [15], [19], [20], [29]. Four studies developed their algorithm by using ResNet-50 CNN which has a multi branching system providing a more robust architecture [17], [23], [25], [29].
Four studies implemented AI on a mobile platform [11], [19], [26], [30] whereas computer/desktop/websites were the platform in the remaining studies. However, in one of the preprints, no platform was specified [19]. The Supplementary Material section contains all the attributes and characteristics of AI architecture of the studies included.
E. Search Results of COVID-19 Apps Based on Cough, Breathing and Speech Recordings
In the case of Apps and sponsored projects, all of them were at the audio collection stage [36]–[43]. On-board real-time AI processing had not yet been implemented in most systems.
However, one of the Apps developed gave a preliminary diagnosis of Influenza-like illnesses or COVID-19 screening utilizing features derived from cough, breathing and speech recordings [39] while two of the Apps recorded and collected cough data and returned recommendations for follow-up actions [40], [41].
Almost all the Apps (n=6) stored data in the form of cough, breathing and speech recordings. Nevertheless, 3 Apps required participants to fill in a medical questionnaire on the basis of recent medical history or any previous underlying condition to complete the data collection process [36], [37], [42].
In three included Apps, questions regarding COVID-19 symptoms were asked before the cough recording [39].
All the Apps (n=8) have their platforms on open-source websites. However, in two Apps the recordings can also be completed by using a mobile phone [36], [43]. The Supplementary Material provides the information of all the considered Apps.
F. COVID-19 Positive With Symptoms/Asymptomatic and COVID-19 Negative (Dataset Characteristics of Studies)
COVID-19 positive with symptomatic state: In all the included studies, positive COVID-19 with underlying symptoms had an area-under-the-curve (AUC) performance ranging from 0.67 to 0.98. One-third of the studies recorded area-under-the-curve (AUC) performance of greater than 0.92 [11], [16], [17]. The highest accuracy to detect COVID-19 positive with cough was 98% shown in the two included studies [16], [20]. The accuracy percentages were recorded much lower in the studies where open-source databases/websites were used.
COVID-19 positive asymptomatic state: One of the studies achieved 100% sensitivity in predicting the diagnosis of individuals asymptomatic COVID-19 with features derived from recording collected in an open-source database [16].
COVID-19 negative state: Negative COVID-19 with cough and other symptoms were detected with an accuracy of greater than 70% in nearly all the included studies. In addition, area-under-the-curve (AUC) performances were greater than 0.80 in detecting negative COVID-19.
The accuracy of one of the Apps [36] was lower with a range from 77% to 80% with AUC of 0.79. Another App reported to return preliminary result within two minutes with an accuracy of 92.64% [11].
IV. Discussion
In this study, a scoping review of AI detection of COVID-19 using cough, breathing and speech samples was conducted. The majority of the reported studies and launched Apps were published in the USA and India. Moreover, considering the lengthy publication process, one third of the included studies were preprints.
In the studies examined, the vital reason for AI modelling using cough, breathing and speech analysis was to identify the novel virus in asymptomatic patients and to make preliminary screening scalable, faster and reliable. The CNN architecture was used in most of the included studies. The most common platform used was computer based except for four studies [14], [19], [26], [30] that utilized mobile phones. Mobile phones can be used in more environments but with more background noise and this should be considered when observing a given model's performance.
Data sources in most of the included studies were crowd sourced (i.e., public datasets). Nonetheless, one study used news interviews and social media platforms for collecting voice samples which is likely to give low accuracy due to the disturbance and noise present in prerecorded samples. On the other hand, there were a few studies which were based on public databases or websites which collected more than 30000 recordings. However, there can be certain drawbacks in open-source databases such as discrepancy of recorded data, problems in recording device, incomplete recordings from the participants and lack of proper information in medical questionnaires. The cases of COVID-19 as of in August (2021) have surpassed 207 million [44].
A. Practical and Research Implications
The COVID-19 cough detection concept was derived from previously implemented architectures such as in detection of Alzheimer's. Whilst this review examines the most common and recent architectures and features which have aided in detecting COVID-19 from cough, breathing and speech samples, there may be alternate architectures and features that could prove more useful in tackling problems and diseases that may arise in the future.
Although official tests such as RT-PCR or Serology for the novel virus have been widely employed, the results take hours, use up vital chemicals, can violate social distancing laws, take up time for a medical worker, are costly and can produce high false negatives. AI with cough/breathing/speech analysis may be a potential solution for preliminary screening eliminating the need of going to test centers, thereby saving time, maintaining social distancing rules with better efficiency and high reliability. Lastly, with the advancement of AI in mobile phones, this tool can be used as a daily screening measure in schools/colleges, workplaces and for everyday commuters to more rapidly prevent outbreaks or super-spreaders events. The type of the approaches covered in this review provides a non-invasive, free and real-time analyzing tool which could be beneficial in future for COVID-19 as well as other respiratory based diseases.
This review mainly focuses on publications and Apps that integrate AI architecture with cough, breathing and speech signals to detect COVID-19. However, further reviews may be required to evaluate the quality, validation and functionality of these AI algorithms.
B. Strengths and Limitations
1). Strengths
This review includes a published papers up to June 2021 and active Apps for AI in detecting COVID-19 through cough, breathing and speech sample regardless of their characteristics, study design, study setting, and country of publication. This study shows the significance of AI in analyzing and detecting a crucial heath related issue. This review follows the full scientific rigor of PRISMA-ScR. Additionally, the risks of duplicate publication were minimized by using Google Scholar along with other search engines. Furthermore, we used the reference sections of all identified journal papers to ensure that our search strategy was not missing any relevant publications.
2). Limitations
The search study was restricted to English due to practical constraints. Due to this reason, some publications written in other languages might be missed, especially Chinese. The search terms were generalized such as artificial intelligence, COVID-19 cough, breathing and speech etc. but not the AI model names such as CNN, RNN etc. Lastly, some papers are preprints which may affect the accuracy of the information in the included studies.
V. Conclusion
In this manuscript, we provide a scoping review of 24 studies and 8 Apps for detecting COVID-19 by cough, breathing and speech recording using AI algorithms. Given, the initial results from multiple studies, this will be a promising research area, since a successful application would save time, reduce scarcity for official testing in small countries and maintain social distancing. In addition, with broad training, testing and validation of artificial intelligence, along with the neurophysiological understanding of human body; the proposed methods could bring a big difference in the fight against COVID-19 and future pandemics.
Acknowledgment
Any opinions, findings, conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the U.S. Under Secretary of Defense for Research and Engineering.
Funding Statement
The work of Mouzzam Husain, Andrew J Simpkin, Claire Gibbons and Derek T.O' Keeffe were supported in part by the Science Foundation Ireland (SFI) and in part by the European Regional Development Fund (ERDF) under Grants 13/RC/2073, 13/RC/2094_P2 and 20/COV/0225. The work of Daniel M. Low was supported in part by the RallyPoint Fellowship and in part by the NIH-NIDCD T32 (5T32DC000038-28). The work of Satrajit S. Ghosh was supported by the McGovern Institute for Brain Research at MIT. The work of Thomas F. Quatieri was supported by the Under Secretary of Defense for Research and Engineering under Air Force Contract FA8702-15-D-0001. The work of Tanya Talkar was supported in part by the NIDCD of the National Institutes of Health under Award F31DC019509, and in part by the Under Secretary of Defense for Research and Engineering under Air Force Contract FA8702-15-D-0001.
Contributor Information
Mouzzam Husain, Email: mouzzam.husain@outlook.com.
Andrew Simpkin, Email: andrew.simpkin@nuigalway.ie.
Claire Gibbons, Email: clairegibbons2017@gmail.com.
Tanya Talkar, Email: tanya.talkar@ll.mit.edu.
Daniel Low, Email: dlow@mit.edu.
Paolo Bonato, Email: pbonato@ieee.org.
Thomas Quatieri, Email: quatieri@ll.mit.edu.
Derek T. O'Keeffe, Email: derek.okeeffe@nuigalway.ie.
References
- [1].Who.int. 2021, “WHO director-general's opening remarks at the media briefing on COVID-19,” Mar. 2020. [Online]. Available: https://www.who.int/director-general/speeches/detail/who-director-general-s-opening-remarks-at-the-media-briefing-on-covid-19-11-march-2020
- [2].World Health Organization, 2020, Coronavirus disease (COVID-19). Accessed: Nov. 1, 2020. [Online]. Available: https://www.who.int/emergencies/diseases/novel-coronavirus-2019
- [3].Health Services Executive (HSE) Ireland, “Symptoms of COVID-19.” Accessed: Nov. 1, 2020. [Online]. Available: https://www2.hse.ie/conditions/coronavirus/symptoms.html
- [4].Centre for Disease Control (CDC) USA, “Symptoms of COVID19.” Accessed: Nov. 1, 2020. [Online]. Available: https://www.cdc.gov/coronavirus/2019-ncov/symptoms-testing/symptoms.html
- [5].Mao L. et al., “Neurological manifestations of hospitalized patients with COVID-19 in Wuhan, China: Retrospective case series study,” JAMA Neurol., vol. 77, pp. 683–690, 2020, doi: 10.1101/2020.02.22.20026500. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [6].Helms J. et al., “Neurologic features in severe SARS-CoV-2 infection,” New England J. Med., vol. 382, pp. 2268–2270, Apr. 2020, doi: 10.1056/nejmc2008597. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [7].HPRA, “In-Vitro diagnostic tests for COVID-19.” Accessed: Nov. 1, 2020. [Online]. Available: http://www.hpra.ie/homepage/medical-devices/covid-19-updates/in-vitro-diagnostic-(ivd)-tests-for-covid-19
- [8].Afzal A., “Molecular diagnostic technologies for COVID-19: Limitations and challenges,” J. Adv. Res., vol. 26, pp. 149–159, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [9].Visser L. E. et al., “Angiotensin converting enzyme inhibitor associated cough: A population-based case-control study,” J. Clin. Epidemiol., vol. 48, no. 6, pp. 851–857, 1995. [DOI] [PubMed] [Google Scholar]
- [10].Kumar M. and Rana L., “Artificial intelligence: A tool for COVID-19 surface detection,” Int. J. Sci. Res. Multidisciplinary Stud., vol. 6, no. 7, pp. 60–63, 2020. [Google Scholar]
- [11].Quatieri T. F., Talkar T., and Palmer J. S., “A framework for biomarkers of COVID-19 based on coordination of speech-production subsystems,” IEEE Open J. Eng. Med. Biol., vol. 1, pp. 203–206, 2020, doi: 10.1109/OJEMB.2020.2998051. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [12].Tricco A. C. et al., “PRISMA extension for scoping reviews (PRISMA-ScR): Checklist and explanation,” Ann. Int. Med., vol. 169, no. 7, pp. 467–473, Oct. 2018. [DOI] [PubMed] [Google Scholar]
- [13].Bagad P. et al., “Cough against COVID: Evidence of COVID-19 signature in cough sounds,” Sep. 17, 2020, arXiv:2009.08790.
- [14].Imran A. et al., “AI4COVID-19: AI enabled preliminary diagnosis for COVID-19 from cough samples via an app,” Inform. Med. Unlocked, Jan. 1, 2020, Art. no. 100378. [DOI] [PMC free article] [PubMed]
- [15].Orlandic L., Teijeiro T., and Atienza D., “The COUGHVID crowdsourcing dataset: A corpus for the study of large-scale cough analysis algorithms,” Sci. Data, vol. 8, no. 1, p. 1, Jun. 23, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Laguarta J., Hueto F., and Subirana B., “COVID-19 artificial intelligence diagnosis using only cough recordings,” IEEE Open J. Eng. Med. Biol., vol. 1, pp. 275–281, Sep. 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [17].Fakhry A. et al., “Virufy: A multi-branch deep learning network for automated detection of COVID-19,” Mar. 2, 2021, arXiv:2103.01806.
- [18].Sharma N. et al., “Coswara–A database of breathing, cough, and voice sounds for COVID-19 diagnosis,” May 21, 2020, arXiv:2005.10548.
- [19].Deshpande G. and Schuller B., “An overview on audio, signal, speech, & language processing for covid-19,” May 18, 2020, arXiv:2005.08579.
- [20].Hassan A., Shahin I., and Alsabek M. B., “Covid-19 detection system using recurrent neural networks,” in Proc. Int. Conf. Commun. Comput. Cybersecurity Inform., Nov. 2020, pp. 1–5. [Google Scholar]
- [21].Faezipour M. and Abuzneid A., “Smartphone-Based self-testing of COVID-19 using breathing sounds,” Telemed. J. E-Health, vol. 26, no. 10, pp. 1202–1205, Oct. 2020. doi: 10.1089/tmj.2020.0114. [DOI] [PubMed] [Google Scholar]
- [22].Ritwik K. V. S., Kalluri S. B., and Vijayasenan D., “COVID-19 patient detection from telephone quality speech data,” Nov 9, 2020, arXiv:2011.04299.
- [23].Dash T. K., Mishra S., Panda G., and Satapathy S. C., “Detection of COVID-19 from speech signal using bio-inspired based cepstral features,” Pattern Recognit., vol. 117, 2021, Art. no. 107999. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [24].Mouawad P., Dubnov T., and Dubnov S., “Robust detection of COVID-19 in cough sounds,” SN Comput. Sci., vol. 2, no. 1, pp. 1–13, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [25].Andreu-Perez J. et al., “A generic deep learning based cough analysis system from clinically validated samples for point-of-need Covid-19 test and severity levels,” IEEE Trans. Serv. Comput., vol. 15, no. 3, pp. 1220–1232, Feb. 23, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [26].Wei W., Wang J., Ma J., Cheng N., and Xiao J., “A real-time robot-based auxiliary system for risk evaluation of COVID-19 infection,” Aug. 18, 2020, arXiv:2008.07695.
- [27].Pahar M. and Niesler T., “Deep transfer learning based COVID-19 detection in cough, breath and speech using bottleneck features,” 2021. Accessed: Aug. 17, 2021. [Online]. Available: https://europepmc.org/article/ppr/ppr375383 [DOI] [PMC free article] [PubMed]
- [28].Coppock H., Gaskell A., Tzirakis P., Baird A., Jones L., and Schuller B., “End- to-end convolutional neural network enables COVID-19 detection from breath and cough audio: A pilot study,” BMJ Innovations, vol. 7, no. 2, pp. 356–362, 2021. [DOI] [PubMed] [Google Scholar]
- [29].Rodriguez C. R., Angeles D., Chafloque R., Kaseng F., and Pandey B., “Deep learning audio spectrograms processing to the early COVID-19 detection,” in Proc. IEEE 12th Int. Conf. Comput. Intell. Commun. Netw., 2020, pp. 429–434. [Google Scholar]
- [30].Ponomarchuk A. et al., “Project achoo: A practical model and application for COVID-19 detection from recordings of breath, voice, and cough,” IEEE J. Selected Topics Signal Process., Feb. 2022, vol. 16, no. 2, pp. 175–187. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [31].Mohammed E. A., Keyhani M., Sanati-Nezhad A., Hejazi S. H., and Far B. H., “An ensemble learning approach to digital corona virus preliminary screening from cough sounds,” Sci. Rep., vol. 11, no. 1, pp. 1–11, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [32].Kumar L. K. and Alphonse P. J. A., “Automatic diagnosis of COVID-19 disease using deep convolutional neural network with multi-feature channel from respiratory sound data: Cough, voice, and breath,” Alexandria Eng. J., vol. 61, no. 2, pp.1319–1334, Feb. 1, 2022. [Google Scholar]
- [33].Pal A. and Sankarasubbu M., “Pay attention to the cough: Early diagnosis of COVID-19 using interpretable symptoms embeddings with cough sound signal processing,” in Proc. 36th Annu. ACM Symp. Appl. Comput., 2021, pp. 620–628. [Google Scholar]
- [34].Lella K. K. and Alphonse P. J. A., “Automatic COVID-19 disease diagnosis using 1D convolutional neural network and augmentation with human respiratory sound based on parameters: Cough, breath, and voice,” AIMS Public Health, vol. 8, pp. 2327–8994, 2021, doi: 10.3934/publichealth.2021019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [35].Alqudaihi K. S. et al., “Cough sound detection and diagnosis using artificial intelligence techniques: Challenges and opportunities,” IEEE Access, vol. 9, pp. 102327–102344, 2021, doi: 10.1109/ACCESS.2021.3097559. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [36].Brown C. et al., “Exploring automatic diagnosis of COVID-19 from crowdsourced respiratory sound data,” Jun. 10, 2020, arXiv:2006.05919.
- [37].Breathefor Science, “Breathe for science,” 2021. [Online]. Available: https://www.breatheforscience.com/
- [38].Voca., “Corona voice detect - Voca,” 2021. [Online]. Available: https://voca.ai/corona-virus/
- [39].Voicemed.io, “VoiceMed – simplifying healthcare access with rapid & accurate disease screening,” 2021. [Online]. Available: https://www.voicemed.io/
- [40].DetectNow., “DetectNow,” 2021. [Online]. Available: https://detectnow.org/
- [41].Voiceome.org., “The COVID-19 voice study,” 2021. [Online]. Available: https://www.voiceome.org/covid19/index.html(Covid-19voicestudy)
- [42].Novoic.com, “Covid-19,” 2021. [Online]. Available: https://www.novoic.com/covid-19(Novoic)
- [43].Hyfeapp.com, “Hyfe cough tracker | stay on top of your health by tracking your cough,” 2021. [Online]. Available: https://www.hyfeapp.com/
- [44].Worldometers.info, “COVID live update,” 2021. Accessed: Aug. 1, 2021. [Online]. Available: https://www.worldometers.info/coronavirus/
- [45].MIT News | Massachusetts Institute of Technology, “Signs of Covid-19 may be hidden in speech signals,” 2021. [Online]. Available: https://news.mit.edu/2020/signs-covid-19-may-be-hidden-speech-signals-0708
- [46].Han H. et al., “Descriptive, retrospective study of the clinical characteristics of asymptomatic COVID-19 patients,” MSphere, vol. 5, no. 5, pp. e00922–e00920, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [47].Gao Z. et al., “A systematic review of asymptomatic infections with COVID-19,” J. Microbiol., Immunol. Infection, vol. 54, no. 1, pp. 12–16, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [48].Kim G. U. et al., “Clinical characteristics of asymptomatic and symptomatic patients with mild COVID-19,” Clin. Microbiol. Infection, vol. 26, no. 7, pp. 948–9e1, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]


