Abstract
Auscultation with stethoscope has been an essential tool for diagnosing the patients with respiratory disease. Although auscultation is non-invasive, rapid, and inexpensive, it has intrinsic limitations such as inter-listener variability and subjectivity, and the examination must be performed face-to-face. Conventional stethoscope could not record the respiratory sounds, so it was impossible to share the sounds. Recent innovative digital stethoscopes have overcome the limitations and enabled clinicians to store and share the sounds for education and discussion. In particular, the recordable stethoscope made it possible to analyze breathing sounds using artificial intelligence, especially based on neural network. Deep learning-based analysis with an automatic feature extractor and convoluted neural network classifier has been applied for the accurate analysis of respiratory sounds. In addition, the current advances in battery technology, embedded processors with low power consumption, and integrated sensors make possible the development of wearable and wireless stethoscopes, which can help to examine patients living in areas of a shortage of doctors or those who need isolation. There are still challenges to overcome, such as the analysis of complex and mixed respiratory sounds and noise filtering, but continuous research and technological development will facilitate the transition to a new era of a wearable and smart stethoscope.
Keywords: Auscultation, Digital stethoscope, Deep learning, Artificial intelligence, Neural network, Wearable or wireless device
Background
In the long-standing history of mankind, auscultation has long been widely used for the examination of patients [1]. A stethoscope is considered one of the most valuable medical devices because it is non-invasive, available in real-time, and much informative [2]. It is particularly useful in respiratory diseases, and abnormal respiratory sounds provide information on various pathological conditions of lungs and bronchi. In 1817, French doctor Rene Laennec invented an auscultation tool and it enabled him to listen to internal noises of patients [3, 4]. Since then, the stethoscope has gradually changed to a device with a binaural form, flexible tubing, and a rigid diaphragm [5, 6].
So far, the stethoscope has been widely used and adopted as the physician’s primary medical tool. However, as chest images are developed, the degree of dependence on auscultation is relatively decreasing [7]. This phenomenon may be caused by the inherent subjectivity. The ability to recognize and differentiate the abnormal sounds depends on the listener's experience and knowledge. This discrepancy can potentially lead to inaccurate diagnosis and mistreatment. To improve this problem, there have been efforts to implement a standardized system to record and share lung sounds to analyze them accurately. Recent technical advances have allowed the recording of lung sounds with a digital stethoscope by electronical intensification of the sounds, and the sharing of recorded sound via blue-tooth transmission [6]. Besides, there have been published studies on artificial intelligence (AI)-assisted auscultation which recognizes the pattern of sounds and identifies their abnormalities, and some digital stethoscopes already adopted machine learning (ML) algorithms [8–25].
Another drawback of auscultation is the impossibility of remote care. When doctors examine patients with a stethoscope, auscultation must be implemented by contacting the stethoscope on the body of patients. Many patients with chronic diseases or limited mobility stay in nursing facilities or at home often without a medical practitioner [24, 25]. Moreover, the demand of patients in hard-to-reach area for telemedicine is increasing nowadays. However, it is difficult for doctors to examine these patients and auscultation is hardly done. Advances in battery technology developed embedded processors with low power consumption and integrated sensors to make stethoscopes wearable and wireless [26–29], so that doctors can examine patients from a distance. Auscultation became possible even while wearing personal protective equipment when treating patients with infectious diseases such as Coronavirus disease-19 (COVID-19) [30–32].
In this review, we will check the limitations of the existing auscultation method by checking the types of abnormal breathing sounds and the accuracy of analysis through the existing stethoscope. Next, we will introduce the new auscultation methods developed so far (AI-assisted analysis and wireless or attached stethoscopes) and the current status of breath sound analysis using them. Furthermore, we will suggest further research directions in the future.
Classification of abnormal respiratory sounds
Respiratory sounds are produced by the airflow in the respiratory tract and are divided into two categories: Normal or abnormal sound. Normal respiratory sound is made when there is no pulmonary disorder and consist of tracheal, bronchial, bronchovesicular, and vesicular sounds [33]. Abnormal respiratory sounds are caused by diseases at the lung or bronchus [34]. They can be described by the mechanism of production, location they are detected in, characteristics (such as continuity, range of pitch, timing mostly heard), and acoustic features (Table 1) [35].
Table 1.
Classification of abnormal lung sounds and related diseases
Crackles are short, discontinuous, explosive sounds heard during inspiration and sometimes on expiration [36, 37]. Fine crackles are generated by inspiratory opening of small airways and associated with interstitial pneumonia or idiopathic pulmonary fibrosis (IPF), and congestive heart failure (CHF) [38]. Coarse crackles are produced by gas passing through intermittent airway opening and are related to secretory diseases such as chronic bronchitis and pneumonia [39].
Wheezes are generated in the narrowed or obstructed airway [36]. They have high frequency (> 100–5000 Hz) and sinusoidal oscillation in sound analysis [40]. They usually occur in obstructive airway diseases like asthma and chronic obstructive pulmonary disease (COPD) [39]. Rhonchi are induced by the narrowing of airways, caused by the production of secretions, so rhonchi can disappear after coughing (Table 1) [36].
Stridor is a high-pitched, continuous sound produced by turbulent airflow through a narrowed airway of upper respiratory tract [36]. It is usually a sign of airway obstruction that requires prompt intervention. In patients with pleural inflammation such as pleurisy or pleural tumor, a visceral pleura becomes rough, and its friction with the parietal pleura generates crackling sounds, a friction rub (Table 1) [41].
Although respiratory sounds are not difficult for a trained clinician to discern in usual cases, some sounds are ambiguous even for an expert to distinguish accurately. In addition, there are some cases where a mixture of several abnormal breathing sounds can be heard. Novel AI-assisted stethoscope can be useful for these challenging situations.
Limitation of conventional stethoscope and auscultation
As mentioned earlier, inherent subjectivity is considered as the largest drawback of auscultation. Many studies have been performed to assess the human’s ability to auscultate and identify respiratory sounds (Table 2).
Table 2.
Accuracy of human auscultation
| Topic of study | Results | References |
|---|---|---|
| The accuracy of lung auscultation | <Correct detection rate by sounds> | [42] |
| – Pulmonologists: 28% (abnormal bronchial sound), 90% (wheezes) | ||
| – Pediatricians: 16% (abnormal bronchial sound), 83% (wheezes) | ||
| – Interns: 13% (abnormal bronchial sound), 83% (wheezes) | ||
| Physicians’ classification of lung sounds from video recordings | <Multirater agreement (Fleiss’ κ) between observers> | [43] |
| – Detailed categories: 0.04 (rhonchi), 0.43 (high-pitched wheezes) | ||
| – Combined categories: 0.59 (wheezes), 0.62 (crackles) | ||
| Pulmonary auscultatory skills during training in internal medicine and family practice | <Identification rates by sounds> | [44] |
| – Trainees of family practice: 0% (whispered pectoriloquy), 84% (expiratory wheeze) | ||
| – Trainees of internal medicine: 1% (whispered pectoriloquy), 82% (expiratory wheeze) | ||
| – Pulmonary fellows: 5% (whispered pectoriloquy), 100% (expiratory wheeze) | ||
| Comparing the auscultatory accuracy of health care professionals | <Correct detection rates by sounds> | [45] |
| – Staff of internal medicine: 86.7% (wheezes), 96.7% (crackles) | ||
| – Resident of internal medicine: 59.0% (crackles), 80.0% (wheezes) | ||
| – Adult ICU nurses: 47.0% (crackles), 88.0% (wheezes) | ||
| The contribution of spectrogram for visualization sound in clinical practice | <Proper diagnosis rate of medical students> | [46] |
| Normal sounds: 57% (sound) → 63% (plus spectrogram) | ||
| Wheezes: 70% (sound) → 83% (plus spectrogram) | ||
| Crackles: 53% (sound) → 70% (plus spectrogram) | ||
| Stridor: 70% (sound) → 73% (plus spectrogram) |
Hafke-Dys et al. conducted a study comparing the skills of doctors and medical students in the auscultation of respiratory sounds. The pulmonologists performed remarkably better than the other groups and there was no significant difference in the rest of the groups [42]. Melbye et al. proceeded a study assessing the inter-observer variation in pediatricians and doctors for adults when classifying respiratory sounds into detailed or broader categories. The results indicated that descriptions of auscultation sounds in broader terms were more steadily shared between participants compared to more detailed descriptions [43]. Mangione et al. conducted a research assessing auscultatory skills of respiratory sounds among doctors and medical students. On average, trainees of internal medicine and family practice did not show significantly better performance than medical students. On the other hand, pulmonary fellows recorded the highest scores in all categories [44]. Mehmood et al. assessed the auscultatory accuracy of health care professionals working in medical intensive care unit (ICU). The sounds presented were wheezes, stridors, crackles, holosystolic murmur, and hyperdynamic bowel sounds. As expected, attending physicians performed best, followed by residents and subsequently nurses [45]. Andres measured the accuracy of medical students’ auscultation and investigated the efficacy of adding visual representation of sounds to support diagnosis and education. The results showed the potential of sound representation for increasing the accuracy of auscultation [46].
Overall, the studies have shown discrepancies in auscultation ability among doctors (especially in detailed classifications of respiratory sounds), suggesting that they may cause inaccurate diagnosis or incorrect treatment [47]. To reduce the subjective interpretation of sounds and complement the gap of auscultation capabilities between doctors, it would be helpful to establish a system which can record and share auscultated sounds.
Another drawback of auscultation is the impossibility of remote care. When doctors examine patients with a stethoscope, auscultation must be implemented by contacting the stethoscope on the body of patients. Many patients with chronic diseases or limited mobility stay in nursing facilities or at home often without a medical practitioner. Also, the demand of patients in hard-to-reach area for telemedicine is increasing nowadays. However, it is difficult for doctors to examine these patients and auscultation is hardly done. If a stethoscope that is easy to use even for non-specialists is developed using data transmission technology, doctors will be able to check the patient's condition from a distance.
Deep-learning based analysis of respiratory sounds
Development of a standardized system to analyze respiratory sounds accurately is required to overcome the subjectivity of human auscultation and the discrepancy in auscultation ability between doctors [8]. Recently, machine learning-based AI techniques are applied mainly by deep learning networks in many areas including chest radiograph or electroencephalography (EEG) [48–50]. These AI techniques enable us to obtain a new approach or more accurate analysis of respiratory sounds [9]. In order to satisfy the requirement, there have been many attempts to develop a new method of classifying and interpreting respiratory sounds automatically using deep learning-based analysis [10, 11]. However, because of the black box type algorithmic property of the deep learning algorithm, there is a certain lack of interpretability of detailed information of the analysis [51]. Though interpretability is an important factor for analysis, it is highly related to technical issues and data dependency. Moreover, it is not clearly defined nor stable yet [51]. For this reason, interpretability will be not covered in this review.
From the machine learning perspective, there are two main parts for respiratory sound analysis. The first is to develop predictive algorithms or models based on well-known machine learning methods (support vector machine [SVM], K-nearest neighbors [KNN], artificial neural network [ANN]) and deep learning architectures (convolutional neural networks [CNN], residual networks [ResNet], long short-term memory [LSTM], gated recurrent unit [GRU]) with multi-layers and the second is to define appropriate features explaining respiratory sound characteristics and extract them (short-time Fourier-transformed [STFT], wavelet transform [WT], Mel-frequency cepstrum coefficient [MFCC], singular spectrum analysis [SSA]) from given data and their ensembles. In this point of view, methods and algorithms for respiratory sound classification and prediction are summarized in more detail below (Table 3).
Table 3.
Deep learning-based analysis of respiratory sounds
| Topic of study | Number of subject/recording | Number of classes | Applied model | Result | References | |
|---|---|---|---|---|---|---|
| Recognition of pulmonary diseases from lung sounds using CNN and LSTM | 213/1483 | 6: Normal, asthma, pneumonia, bronchiectasis, COPD, heart failure | CNN, biLSTM | Accuracy | [12] | |
| biLSTM: 98.16% | ||||||
| CNN: 99.04% | ||||||
| CNN + biLSTM: 99.62% | ||||||
| Classification of respiratory sounds using OST and deep residual networks | Not available/489 | 3: Crackle, wheeze, normal | OST, ResNets | Accuracy | [13] | |
| STFT: 93.98% | ||||||
| ST: 97.79% | ||||||
| OST + ResNets: 98.79% | ||||||
| Detection of respiratory sounds based on wavelet coefficients and machine learning | 130/705 | 3: Crackles, rhonchi, normal | SVM, ANN, KNN | Accuracy | [14] | |
| SVM: 69.50% | ||||||
| ANN: 85.43% | ||||||
| KNN: 68.51% | ||||||
| Benchmarking of eight recurrent neural network variants for breath phase and adventitious sound detection | 279/9765 | 6: Inhalation, exhalation, wheeze, stridor, rhonchus, crackles | LSTM, GRU, BiLSTM, BiGRUc, CNN-LSTM, CNN-GRU, CNN-BiLSTM, and CNN-BiGRU | F1 score | [15] | |
| LSTM: 73.9% | GRU: 77.6% | |||||
| BiLSTM: 76.2% | BiGRU: 78.4% | |||||
| CNN-LSTM: 78.1% | CNN-GRU: 80.6% | |||||
| CNN-BiLSTM: 80.3% | CNN-BiGRU: 80.6% | |||||
| Classification of lung sounds through DS-CNN models with fused STFT and MFCC features | Not available/12,691 | 4: Normal, wheeze, crackle, unknown | DS-CNN, VGG-16, AlexNet, DS-AlexNet, LSTM, GRU, TCNd | Accuracy | [16] | |
| DS-CNN: 85.74% | VGG-16: 85.66% | |||||
| AlexNet: 79.92% | DS-AlexNet: 80.86% | |||||
| LSTM: 76.92% | GRU: 78.50% | |||||
| TCN: 75.51% | ||||||
| Implementation of AI algorithms in pulmonary auscultation examination | 50/522 | 4: Wheezes, rhonchi, fine and coarse crackles | Neural network | F1-score (%) | [17] | |
| Coarse crackles: Doctors (42.8%), NN (47.1%) | ||||||
| Fine crackles: Doctors (51.1%), NN (64.6%) | ||||||
| Wheezes: Doctors (61.8%), NN (66.4%) | ||||||
| Rhonchi: Doctors (61.0%), NN (72%) | ||||||
| AI accuracy in detecting pathological breath sounds in children | 25/192 | 2: Crackles, wheeze | Neural network | PPAa/NPAb | [18] | |
| Crackles: Clinicloud 0.95/0.99, Littman 0.82/0.96 | ||||||
| Wheezes: Clinicloud 0.90/0.97, Littman 0.80/0.95 | ||||||
| Classification of lung sounds using CNN | 1630/17,930 | ① 2: Healthy, pathological | CNN, SVM | Accuracy (CNN/SVM) | [19] | |
| ② 3: Rale, rhonchus, normal | ① 86%/86% | ② 76%/75% | ||||
| Feature extraction technique for pulmonary sound analysis based on EMD | 30/120 | 3: Normal, wheezes, crackles | ANN, SVM, GMM | Accuracy | [20] | |
| EMD with ANN: 94.16% | ||||||
| EMD with SVM: 93.75% | ||||||
| EMD with GMM: 88.16% | ||||||
| Application of deep Learning to classify the severity of COPD | Not available/120 | 2: Crackles, wheeze | NN, DBN | Accuracy | [21] | |
| NN: 77.60% | ||||||
| DBN: 95.84% | ||||||
| Application of deep Learning to detect early COPD | 50/600 | 1: Wheeze | DBN | Accuracy | [22] | |
| DBN: 93.67% | ||||||
| Application of semi-supervised deep learning to lung sound analysis | 284/11627 | 2: Crackles, wheeze | SVM | AUC | [23] | |
| Crackle: 0.74 | ||||||
| Wheeze: 0.86 | ||||||
aPPA positive percent agreement
bNPA negative percent agreement
cBiGRU bidirectional gated recurrent unit
dTCN temporal convolutional network, DBN deep belief networks, COPD chronic obstructive pulmonary disease
Fraiwan et al. conducted a study to explore the ability of deep learning algorithms in recognizing pulmonary diseases from recorded lung sounds. After several preprocessing steps (wavelet smoothing, displacement artifact removal, and z-sore normalization), two deep learning network architectures including CNN and bidirectional long short-term memory (biLSTM) units were applied. The resulting algorithm (CNN + biLSTM) achieved the highest accuracy [12]. Chen et al. proceeded with research to overcome the limitations of existing classification methods of lung sounds; artifacts and constrained feature extraction methods. The proposed method using optimized S-transform (OST) and deep ResNets outperformed the ensemble of CNN and the empirical mode decomposition (EMD)-based ANN [13]. Meng et al. combined the wavelet signal similarity with the relative wavelet energy and entropy as the feature vector to extract features of lung sounds. Applying the ANN to this system showed higher accuracy than the methods using SVM and KNN [14]. Hsu et al. applied eight kinds of AI-technique models and conducted a performance comparison between them. GRU-based models outperformed the LSTM-based models, and bidirectional models outperformed unidirectional counterparts. Moreover, adding CNN improved the accuracy of lung sounds analysis [15]. Jung et al. proposed a feature extracting process through the depthwise separable-convolution neural network (DS-CNN) to classify lung sounds accurately. Also, they found that the fusion of the STFT and the MFCC features and DS-CNN achieved a higher accuracy than other methods [16]. Grzywalski et al. compared the efficiency of auscultation of doctors and machine learning-based analysis based on neural networks and proposed that the efficiency could be improved by the implementation of automatic analysis [17]. Kevat et al. showed that a neural network-based AI algorithm detected respiratory sounds with a high accuracy [18]. Aykanat et al. found that CNN and SVM machine learning algorithms can be used to classify lung sounds, but the accuracy decreased as the number of sounds to be compared increased, as with humans [19]. Mondal et al. proposed a feature extraction technique based on EMD improving the performance of lung sound classification and the method was compared with WT, MFCC, and SSA method-based classification systems including ANN, SVM, and Gaussian mixture model (GMM) classifier. The proposed method gives a higher accuracy of 94.16 for an ANN classifier [20]. Altan applied deep belief networks (DBN) algorithm to diagnose early COPD and classify the severity of COPD, and the results showed significantly high accuracy. Since COPD is irreversible when it progresses, early diagnosis is important. In this regard, the results of their studies are groundbreaking and useful [21, 22]. Chamberlain et al. applied SVM with a semi-supervised deep learning algorithm and their algorithm achieved receiver operating characteristic (ROC) curves with a relatively high area under the ROC curve (AUC) [52].
Many studies have been conducted in collaboration with doctors and machine learning experts, and it has become possible to discriminate lung sounds with a considerable level of accuracy. However, there is still a limitation that the analysis becomes less accurate when noises caused by the stethoscope itself, surrounding environment, other organ activities, and so on are mixed among the recorded sounds or when two or more breathing sounds are present at the same time. This should be resolved through additional research in the future [53].
Development of digital stethoscopes
There are several available electronic stethoscopes: Littmann 3100, Stethee pro, Thinklabs one digital amplified medical stethoscope, Littman core digital stethoscope 8490, and StethoMe (Table 4). These digital stethoscopes overcome the low sound levels by electronically intensifying the respiratory sounds. Most importantly, recording of respiratory sounds with a digital stethoscope has allowed and facilitated the study of automatic respiratory sound analysis. Littmann 3100 is one of the most popular electronic stethoscopes, and many studies using respiratory sounds have been conducted with this stethoscope [54, 55]. It can save multiple sounds and transmit the data via Bluetooth transmission. Interestingly, Stethee Pro uses machine learning algorithms to capture and monitor both heart and lung sounds. This stethoscope can amplify the sound up to 96 times and visualize the sound data on the screen. Thinklabs One is the smallest digital stethoscope, and it can be used for personal protective equipment (PPE) auscultation in patients with infectious diseases such as COVID-19. StethoMe was developed for homecare service and installed AI can analyze the abnormality of respiratory sound. It is particularly specialized for monitoring airway diseases including asthma. These digital stethoscopes are continuously developing and becoming more useful for monitoring and diagnosing pulmonary disease.
Table 4.
Developing stethoscopes: digital, wireless, or wearable device
| Model/study | Characteristics | Manufacturer/references |
|---|---|---|
| Littmann 3100 Electronic Stethoscope | 24× amplification | Littmann® |
| Record and save | ||
| Bluetooth transmission | ||
| Stethee Pro | 96× amplification | M3DICINE Inc® |
| Machine learning algorithm | ||
| Ambient noise cancellation | ||
| Thinklabs One Digital Amplified Medical Stethoscope | 100× amplification | Thinklabs One® |
| Precision filtering | ||
| Personal protective equipment auscultation | ||
| StethoMe | Homecare service | StethoMe® |
| AI analyses the respiratory sounds | ||
| A wearable stethoscope for long-term ambulatory respiratory health monitoring | Long-term ambulatory | [26] |
| Respiratory health monitoring | ||
| Diaphragm-less acousto-electric transducer | ||
| Wearable multimodal stethoscope patch | Wearable biosignal acquisition | [27] |
| High quality cardiac and pulmonary auscultation | ||
| Wearable cardiorespiratory monitoring | Estimation of respiration using a phonocardiogram | [28] |
| Epidermal mechano-acoustic electrophysiological measurement device | Water-permeable, adhesive, biocompatible, and reversible device | [29] |
In addition, recent innovative advances in battery technology, embedded processors with low power consumption, and integrated sensors have made many medical devices wearable and wireless (Table 4). Some studies have applied these techniques to stethoscopes, and the researchers developed the stethoscopes that monitor cardiorespiratory signals through wireless bio-signal acquisition [26, 27]. Certain airway diseases, such as asthma, often get worse at night or early in the morning, so doctors often cannot detect them during the daytime. Just as in the diagnosis of arrhythmia disease, Holter monitoring is used to monitor a patient's heart rate for 24 h, continuous monitoring of respiratory sound through a wearable device in airway disease will be of great help in diagnosis and emergency treatment. Some groups developed water permeable, adhesive, biocompatible acoustic devices for electrophysiological recording [28, 29]. Technologies of recording of sounds clearly and filtering out noises need further improvement, but wearable stethoscopes are expected to be used to diagnose and monitor chronic pulmonary diseases soon.
Clinical application of digital stethoscopes and AI-assisted analysis
There are several clinical studies using distal stethoscopes and AI for respiratory analysis. One study showed that CNN can classify chronic disease, non-chronic disease, and healthy groups by automatically analyzing respiratory sounds. In addition, the CNN is able to subcategorize disease group to different types of diseases including COPD, bronchiectasis, pneumonia, and bronchiolitis (Table 5) [56]. Another study adopted the acoustic characteristics of fine crackles to predict honeycombing on chest computed tomography (CT). They concluded that the presence of honeycombing was independently associated with onset time, number of crackles in the inspiratory phase, and F99 value of fine crackles [57].
Table 5.
Clinical trials of novel digital stethoscope and AI-assisted analysis
| Topic of study | Results/characteristics | Condition/disease Study type (status) |
Reference and ClinicalTrial.gov. identifier |
|---|---|---|---|
| Detecting respiratory pathologies using CNN and variational autoencoders | CNN was used to classify chronic disease, non-chronic disease and healthy group | Chronic disease | [56] |
| Non-chronic disease | |||
| Healthy | |||
| Observational (completed) | |||
| Predicting honeycombing on HRCT by the acoustic characteristics of fine crackles | Acoustic properties of fine crackles distinguish the honeycombing from non-honeycombing group | Honeycombing | [57] |
| Non-honeycombing | |||
| Observational (completed) | |||
| Diagnosing interstitial pneumonia by analyzing inspiratory lung sounds recorded with phonopneumography | Spectral analysis of lung sounds is useful in the diagnosis and evaluation of the severity of IP | Interstitial pneumonia | [59] |
| Healthy | |||
| Observational (completed) | |||
| Evaluating Auscul-X, a Touch Free Digital Stethoscope | Multichannel, touch-free electronic stethoscope | COVID-19 | NCT04570189 |
| Observational (recruiting) | |||
| Collecting respiratory sound samples to diagnose COVID-19 patients | VOQX Electronic Stethoscope | COVID-19 | NCT04910191 |
| Sound signals are processed by machine learning algorithm | Interventional (recruiting) | ||
| Clinical Evaluation of Automatic Classification of Respiratory System Sounds | StethoMe stethoscope | Wheezing, rhonchi, crackle, lung sound | NCT04208360 |
| AI software application | Observational (not yet recruiting) | ||
| Digital auscultation test—IPF data collection | Littmann Digital Stethoscope | Idiopathic pulmonary fibrosis | NCT03503188 |
| 3M Littmann Steth Assist software | Interventional (completed) | ||
| Respiratory auscultation of an open real-time tele-stethoscope system | Open real-time tele-stethoscope system | Respiratory/heart disease | NCT03596541 |
| Respiratory crackle | Interventional (not yet recruiting) |
Many studies related to digital stethoscope and AI analysis of auscultation sound are still currently in progress. As the need to collect and analyze the auscultation sounds of patients in quarantine facilities increases due to the recent COVID-19 crisis, related research is being conducted more actively. Several studies are trying to find a typical pattern of auditory sounds in COVID-19 patients (Table 5). One study plans to evaluate the AI-aided auscultation with automatic classification of respiratory sounds by using StethoMe stethoscope. If these studies are conducted well and AI-equipped stethoscopes can detect wheezing, rhonchi, and crackle accurately, these stethoscopes will be useful in emergency room treatment, medical screening, and telemedicine fields [58]. These smart stethoscopes will be of great help in monitoring patients with chronic pulmonary diseases, and many studies are underway for patients with idiopathic pulmonary fibrosis (IPF) and COPD (Table 5).
Conclusion
Thanks to the development of digital stethoscope and sound transmission technology, we have already been able to record and share respiratory sounds. With deep learning-based breathing sound analysis algorithm, we can distinguish respiratory sounds to some extent without a pulmonologist. This makes it possible to overcome the subjectivity in interpretation of sounds, the biggest drawback of the stethoscope, and this smart stethoscope will help the rapid diagnosis and the choice of appropriate treatment methods of respiratory diseases.
In addition, current research on battery technology, embedded processors with low power consumption, and integrated sensors are expected to make stethoscopes and other medical devices wearable in addition to wireless. Through these advances, we will be able to get over another major limitation of the existing stethoscope, the impossibility of remote care. The latest medical demands such as non-face-to-face care due to COVID-19, the monitoring of chronic respiratory diseases, and telemedicine in the hard-to-reach area will be satisfied (Fig. 1).
Fig. 1.
Summary of new medical era using smart stethoscope
However, despite the innovative developments so far, there are still some problems for the smart stethoscope to overcome. Since noises exist in the actual medical field where auscultation is performed, careful attention is required in recording and interpreting respiratory sounds. Noise filtering is one of the most crucial and challenging points in the aspect of both mechanical devices and analyzing algorithms. Although respiratory sounds are sometimes heard alone, in many cases, two or more sounds are mixed. These problems suggest the need for processing sound data acquired under noisy conditions to improve the sound quality. This would help rationally classify a wider variety of automatically auscultated sounds. Now, with the development of chest imaging, the degree of dependence on auscultation is relatively decreasing. However, as the remaining challenges are solved through further researches and clinical feedbacks, the smart stethoscope will become a definitely useful and essential tool in the diagnosis and treatment of respiratory diseases.
Acknowledgements
Not applicable.
Abbreviations
- AI
Artificial intelligence
- ML
Machine learning
- COVID-19
Coronavirus disease-19
- IPF
Idiopathic pulmonary fibrosis
- CHF
Congestive heart failure
- COPD
Chronic obstructive pulmonary disease
- ICU
Intensive care unit
- EEG
Electroencephalography
- SVM
Support vector machine
- KNN
K-nearest neighbors
- ANN
Artificial neural network
- CNN
Convolutional neural networks
- ResNet
Residual networks
- LSTM
Long short-term memory
- GRU
Gated recurrent unit
- STFT
Short-time Fourier-transformed
- WT
Wavelet transform
- MFCC
Mel-frequency cepstrum coefficient
- SSA
Singular spectrum analysis
- biLSTM
Bidirectional long short-term memory
- OST
Optimized S-transform
- EMD
Empirical mode decomposition
- DS-CNN
Depthwise separable-convolution neural network
- GMM
Gaussian mixture model
- DBN
Deep belief networks
- ROC
Receiver operating characteristic
- AUC
Area under the ROC curve
- PPE
Personal protective equipment
- CT
Computed tomography
Authors' contributions
CC and TH suggested the idea for this article, YK, SL and SW performed the literature search and arrangement on this topic, and YH critically revised the work. All authors contributed to the elaboration and redaction of the final manuscript. All authors read and approved the final manuscript.
Funding
This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean Government (MSIT) (No. 2017R1A5A2015385), National Institute for Mathematical Sciences (NIMS) grant funded by the Korean government (No. B22910000) and Chungnam National University Hospital Research Fund, 2021 (No. 2021-CF-053). The funding bodies played no role in the design of the study and collection, analysis, and interpretation of data and in writing the manuscript.
Availability of data and materials
Not applicable.
Declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Yoonjoo Kim, YunKyong Hyon and Sunju Lee have contributed equally to this work
Contributor Information
Yoonjoo Kim, Email: misspims@naver.com.
YunKyong Hyon, Email: hyon@nims.re.kr.
Sunju Lee, Email: sjlee@nims.re.kr.
Seong-Dae Woo, Email: nextera@naver.com.
Taeyoung Ha, Email: tha@nims.re.kr.
Chaeuk Chung, Email: universe7903@gmail.com.
References
- 1.Coucke PA. Laennec versus Forbes: tied for the score ! How technology helps us interpret auscultation. Rev Med Liege. 2019;74(10):543–551. [PubMed] [Google Scholar]
- 2.Sarkar M, et al. Auscultation of the respiratory system. Ann Thorac Med. 2015;10(3):158–168. doi: 10.4103/1817-1737.160831. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Bloch H. The inventor of the stethoscope: René Laennec. J Fam Pract. 1993;37(2):191. [PubMed] [Google Scholar]
- 4.Roguin A. Rene Theophile Hyacinthe Laënnec (1781–1826): the man behind the stethoscope. Clin Med Res. 2006;4(3):230–235. doi: 10.3121/cmr.4.3.230. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Andrès E, et al. Respiratory sound analysis in the era of evidence-based medicine and the world of medicine 2.0. J Med Life. 2018;11(2):89–106. [PMC free article] [PubMed] [Google Scholar]
- 6.Swarup S, Makaryus AN. Digital stethoscope: technology update. Med Devices (Auckl) 2018;11:29–36. doi: 10.2147/MDER.S135882. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Arts L, et al. The diagnostic accuracy of lung auscultation in adult patients with acute pulmonary pathologies: a meta-analysis. Sci Rep. 2020;10(1):7347. doi: 10.1038/s41598-020-64405-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Gurung A, et al. Computerized lung sound analysis as diagnostic aid for the detection of abnormal lung sounds: a systematic review and meta-analysis. Respir Med. 2011;105(9):1396–1403. doi: 10.1016/j.rmed.2011.05.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Palaniappan R, Sundaraj K, Sundaraj S. Artificial intelligence techniques used in respiratory sound analysis—a systematic review. Biomed Tech (Berl) 2014;59(1):7–18. doi: 10.1515/bmt-2013-0074. [DOI] [PubMed] [Google Scholar]
- 10.Aras S, et al. Automatic detection of the respiratory cycle from recorded, single-channel sounds from lungs. Turk J Electr Eng Comput Sci. 2018;26(1):11–22. [Google Scholar]
- 11.Altan G, et al. Chronic obstructive pulmonary disease severity analysis using deep learning on multi-channel lung sounds. Turk J Electr Eng Comput Sci. 2020;28(5):2979–2996. [Google Scholar]
- 12.Fraiwan M, Fraiwan L, Alkhodari M, Hassanin O. Recognition of pulmonary diseases from lung sounds using convolutional neural networks and long short-term memory. J Ambient Intell Humaniz Comput. 2021:1–13. 10.1007/s12652-021-03184-y. [DOI] [PMC free article] [PubMed]
- 13.Chen H, et al. Triple-classification of respiratory sounds using optimized s-transform and deep residual networks. IEEE Access. 2019;7:32845–32852. [Google Scholar]
- 14.Meng F, et al. Detection of respiratory sounds based on wavelet coefficients and machine learning. IEEE Access. 2020;8:155710–155720. [Google Scholar]
- 15.Hsu FS, et al. Benchmarking of eight recurrent neural network variants for breath phase and adventitious sound detection on a self-developed open-access lung sound database-HF_Lung_V1. PLoS ONE. 2021;16(7):e0254134. doi: 10.1371/journal.pone.0254134. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Jung SY, et al. Efficiently classifying lung sounds through depthwise separable CNN models with fused STFT and MFCC features. Diagnostics (Basel) 2021;11(4):732. doi: 10.3390/diagnostics11040732. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Grzywalski T, et al. Practical implementation of artificial intelligence algorithms in pulmonary auscultation examination. Eur J Pediatr. 2019;178(6):883–890. doi: 10.1007/s00431-019-03363-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Kevat A, Kalirajah A, Roseby R. Artificial intelligence accuracy in detecting pathological breath sounds in children using digital stethoscopes. Respir Res. 2020;21(1):253. doi: 10.1186/s12931-020-01523-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Aykanat M, et al. Classification of lung sounds using convolutional neural networks. EURASIP J Image Video Process. 2017;2017(1):1–9. [Google Scholar]
- 20.Mondal A, Banerjee P, Tang H. A novel feature extraction technique for pulmonary sound analysis based on EMD. Comput Methods Prog Biomed. 2018;159:199–209. doi: 10.1016/j.cmpb.2018.03.016. [DOI] [PubMed] [Google Scholar]
- 21.Altan G, et al. Deep learning with 3D-second order difference plot on respiratory sounds. Biomed Signal Process Control. 2018;45:58–69. [Google Scholar]
- 22.Altan G, Kutlu Y, Allahverdi N. Deep learning on computerized analysis of chronic obstructive pulmonary disease. IEEE J Biomed Health Inform. 2019;24:1344–1350. doi: 10.1109/JBHI.2019.2931395. [DOI] [PubMed] [Google Scholar]
- 23.Chamberlain D, et al. Application of semi-supervised deep learning to lung sound analysis. In: 2016 38th Annual international conference of the IEEE engineering in medicine and biology society (EMBC); 2016. IEEE. [DOI] [PubMed]
- 24.Fernandez-Granero MA, Sanchez-Morillo D, Leon-Jimenez A. Computerised analysis of telemonitored respiratory sounds for predicting acute exacerbations of COPD. Sensors (Basel) 2015;15(10):26978–26996. doi: 10.3390/s151026978. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Murphy RL, et al. Automated lung sound analysis in patients with pneumonia. Respir Care. 2004;49(12):1490–1497. [PubMed] [Google Scholar]
- 26.Yilmaz G, et al. A wearable stethoscope for long-term ambulatory respiratory health monitoring. Sensors (Basel) 2020;20(18):5124. doi: 10.3390/s20185124. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Klum M, et al. Wearable multimodal stethoscope patch for wireless biosignal acquisition and long-term auscultation. In: Annual international conference of IEEE engineering in medicine and biology society, vol 2019; 2019. p. 5781–5. [DOI] [PubMed]
- 28.Klum M, et al. Wearable cardiorespiratory monitoring employing a multimodal digital patch stethoscope: estimation of ECG, PEP, LVET and respiration using a 55 mm single-lead ECG and phonocardiogram. Sensors (Basel) 2020;20(7):2033. doi: 10.3390/s20072033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Liu Y, et al. Epidermal mechano-acoustic sensing electronics for cardiovascular diagnostics and human–machine interfaces. Sci Adv. 2016;2(11):e1601185. doi: 10.1126/sciadv.1601185. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Vasudevan RS, et al. Persistent value of the stethoscope in the age of COVID-19. Am J Med. 2020;133(10):1143–1150. doi: 10.1016/j.amjmed.2020.05.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.White SJ. Auscultation without contamination: a solution for stethoscope use with personal protective equipment. Ann Emerg Med. 2015;65(2):235–236. doi: 10.1016/j.annemergmed.2014.11.021. [DOI] [PubMed] [Google Scholar]
- 32.Arun Babu T, Sharmila V. Auscultating with personal protective equipment (PPE) during COVID-19 pandemic—challenges and solutions. Eur J Obstet Gynecol Reprod Biol. 2021;256:509–510. doi: 10.1016/j.ejogrb.2020.11.063. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Bardou D, Zhang K, Ahmad SM. Lung sounds classification using convolutional neural networks. Artif Intell Med. 2018;88:58–69. doi: 10.1016/j.artmed.2018.04.008. [DOI] [PubMed] [Google Scholar]
- 34.Reichert S, et al. Analysis of respiratory sounds: state of the art. Clin Med Circ Respir Pulm Med. 2008;2:45–58. doi: 10.4137/ccrpm.s530. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Sengupta N, Sahidullah M, Saha G. Lung sound classification using cepstral-based statistical features. Comput Biol Med. 2016;75:118–129. doi: 10.1016/j.compbiomed.2016.05.013. [DOI] [PubMed] [Google Scholar]
- 36.Bohadana A, Izbicki G, Kraman SS. Fundamentals of lung auscultation. N Engl J Med. 2014;370(8):744–751. doi: 10.1056/NEJMra1302901. [DOI] [PubMed] [Google Scholar]
- 37.Serbes G, et al. Feature extraction using time-frequency/scale analysis and ensemble of feature sets for crackle detection. In: Annual international conference of IEEE engineering in medicine and biology society, vol 2011; 2011. p. 3314–7. [DOI] [PubMed]
- 38.Vyshedskiy A, et al. Mechanism of inspiratory and expiratory crackles. Chest. 2009;135(1):156–164. doi: 10.1378/chest.07-1562. [DOI] [PubMed] [Google Scholar]
- 39.Faustino P, Oliveira J, Coimbra M. Crackle and wheeze detection in lung sound signals using convolutional neural networks. In: Annual international conference of IEEE engineering in medicine and biology society, vol 2021; 2021. p. 345–8. [DOI] [PubMed]
- 40.Gavriely N, et al. Flutter in flow-limited collapsible tubes: a mechanism for generation of wheezes. J Appl Physiol (1985) 1989;66(5):2251–2261. doi: 10.1152/jappl.1989.66.5.2251. [DOI] [PubMed] [Google Scholar]
- 41.Pasterkamp H, et al. Towards the standardisation of lung sound nomenclature. Eur Respir J. 2016;47(3):724–732. doi: 10.1183/13993003.01132-2015. [DOI] [PubMed] [Google Scholar]
- 42.Hafke-Dys H, et al. The accuracy of lung auscultation in the practice of physicians and medical students. PLoS ONE. 2019;14(8):e0220606. doi: 10.1371/journal.pone.0220606. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Melbye H, et al. Wheezes, crackles and rhonchi: simplifying description of lung sounds increases the agreement on their classification: a study of 12 physicians' classification of lung sounds from video recordings. BMJ Open Respir Res. 2016;3(1):e000136. doi: 10.1136/bmjresp-2016-000136. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Mangione S, Nieman LZ. Pulmonary auscultatory skills during training in internal medicine and family practice. Am J Respir Crit Care Med. 1999;159(4 Pt 1):1119–1124. doi: 10.1164/ajrccm.159.4.9806083. [DOI] [PubMed] [Google Scholar]
- 45.Mehmood M, et al. Comparing the auscultatory accuracy of health care professionals using three different brands of stethoscopes on a simulator. Med Devices (Auckl) 2014;7:273–281. doi: 10.2147/MDER.S67784. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Andrès E. Advances and perspectives in the field of auscultation, with a special focus on the contribution of new intelligent communicating stethoscope systems in clinical practice, in teaching and telemedicine. In: El Hassani AH, editor. eHealth and remote monitoring. London: IntechOpen; 2012. [Google Scholar]
- 47.Pasterkamp H, Kraman SS, Wodicka GR. Respiratory sounds. Advances beyond the stethoscope. Am J Respir Crit Care Med. 1997;156(3 Pt 1):974–987. doi: 10.1164/ajrccm.156.3.9701115. [DOI] [PubMed] [Google Scholar]
- 48.Altan G, Yayık A, Kutlu YJNPL. Deep learning with ConvNet predicts imagery tasks through EEG. 2021. p. 1–16.
- 49.Tang YX, et al. Automated abnormality classification of chest radiographs using deep convolutional neural networks. NPJ Digit Med. 2020;3:70. doi: 10.1038/s41746-020-0273-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Thompson WR, et al. Artificial intelligence-assisted auscultation of heart murmurs: validation by virtual clinical trial. Pediatr Cardiol. 2019;40(3):623–629. doi: 10.1007/s00246-018-2036-z. [DOI] [PubMed] [Google Scholar]
- 51.Goodfellow I, Bengio Y, Courville AJCM. Deep learning (adaptive computation and machine learning series) Cambridge: MIT Press; 2017. pp. 321–359. [Google Scholar]
- 52.Chamberlain D, et al. Application of semi-supervised deep learning to lung sound analysis. In: Annual international conference of IEEE engineering in medicine and biology society, vol 2016; 2016. p. 804–7. [DOI] [PubMed]
- 53.Emmanouilidou D, et al. Computerized lung sound screening for pediatric auscultation in noisy field environments. IEEE Trans Biomed Eng. 2018;65(7):1564–1574. doi: 10.1109/TBME.2017.2717280. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Fontaine E, et al. In-flight auscultation during medical air evacuation: comparison between traditional and amplified stethoscopes. Air Med J. 2014;33(6):283–285. doi: 10.1016/j.amj.2014.06.008. [DOI] [PubMed] [Google Scholar]
- 55.Mamorita N, et al. Development of a Smartphone App for visualizing heart sounds and murmurs. Cardiology. 2017;137(3):193–200. doi: 10.1159/000466683. [DOI] [PubMed] [Google Scholar]
- 56.García-Ordás MT, et al. Detecting respiratory pathologies using convolutional neural networks and variational autoencoders for unbalancing data. Sensors (Basel) 2020;20(4):1214. doi: 10.3390/s20041214. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Fukumitsu T, et al. The acoustic characteristics of fine crackles predict honeycombing on high-resolution computed tomography. BMC Pulm Med. 2019;19(1):153. doi: 10.1186/s12890-019-0916-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Jaber MM, et al. A telemedicine tool framework for lung sounds classification using ensemble classifier algorithms. Measurement. 2020;162:107883. [Google Scholar]
- 59.Ono H, et al. Evaluation of the usefulness of spectral analysis of inspiratory lung sounds recorded with phonopneumography in patients with interstitial pneumonia. J Nippon Med Sch. 2009;76(2):67–75. doi: 10.1272/jnms.76.67. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Not applicable.


