Abstract
Recognizing specific heart sound patterns is important for the diagnosis of structural heart diseases. However, the correct recognition of heart murmur depends largely on clinical experience. Accurately identifying abnormal heart sound patterns is challenging for young and inexperienced clinicians. This study is aimed at the development of a novel algorithm that can automatically recognize systolic murmurs in patients with ventricular septal defects (VSDs). Heart sounds from 51 subjects with VSDs and 25 subjects without a significant heart malformation were obtained in this study. Subsequently, the soundtracks were divided into different training and testing sets to establish the recognition system and evaluate the performance. The automatic murmur recognition system was based on a novel temporal attentive pooling-convolutional recurrent neural network (TAP-CRNN) model. On analyzing the performance using the test data that comprised 178 VSD heart sounds and 60 normal heart sounds, a sensitivity rate of 96.0% was obtained along with a specificity of 96.7%. When analyzing the heart sounds recorded in the second aortic and tricuspid areas, both the sensitivity and specificity were 100%. We demonstrated that the proposed TAP-CRNN system can accurately recognize the systolic murmurs of VSD patients, showing promising potential for the development of software for classifying the heart murmurs of several other structural heart diseases.
Subject terms: Cardiovascular diseases, Cardiology, Diseases, Health care, Diagnosis, Software
Introduction
Ventricular septal defect (VSD), a type of congenital heart disease (CHD) caused by developmental defects of the interventricular septum, is the most common type of heart malformation present at birth. It occurs in approximately 2–6 of every 1000 live births and accounts for approximately 30% of all CHDs in children/adolescents1–4. The clinical presentation of a VSD is correlated with the size of the defect5. Mild VSDs are usually asymptomatic and commonly occur spontaneously within close proximity6. Patients with medium defects often suffer from dyspnea. Patients with severe VSDs exhibit cyanosis, dyspnea, syncope, or heart failure and require adequate surgeries unless the defects spontaneously decrease7–9. VSDs can also be classified according to the morphology and anatomical location of the defect. They can also be classified into four anatomical types: type I (outlet supracristal, subarterial, or infundibular), type II (perimembranous, paramembranous, or conoventricular), type III (inlet, atrioventricular canal, or atrioventricular septal defect), and type IV (muscular or trabecular)10–12. The perimembranous type is the most common (~ 80%), followed by the muscular (15–20%), inlet (~ 5%), and outlet (~ 5%) types.
Similar to many other heart malformations, heart murmurs can be heard in patients with VSD13. Patients with a VSD are known to commonly experience holosystolic murmurs, owing to the turbulence of the blood flow between the left and right ventricles14,15. Murmur recognition with auscultation is conventionally used for the screening and diagnosis of VSD16. However, the accuracy of this method largely depends on clinical experience and is a challenge for most young and inexperienced clinicians17. Therefore, the development of tools to automatically recognize heart-sound patterns can help physicians diagnose heart disease.
Artificial intelligence has recently been widely used in computer-aided diagnosis18,19. For example, many algorithms that claim to automatically recognize and classify medical images have been developed using deep learning20–22. Recent efforts have shown significant advances using artificial neural networks (ANNs) or deep neural networks (DNNs) to detect and classify heart sounds23–25. Convolutional neural networks (CNNs) have also been used to identify heart murmurs26. The aim of this study was to develop a novel algorithm that can automatically recognize the systolic murmurs of VSD patients using a novel temporal attentive pooling–convolutional recurrent neural network (TAP-CRNN) model27.
Results
Heart sounds from 76 subjects, including 51 VSD patients and 25 patients without significant heart malformations, were included in this study. Table 1 shows the mean age, height, weight, and sex distribution of these subjects. There were no statistically significant differences between the group suffering from VSD and the normal group, with regard to these clinical variables. Regarding the types of VSDs, most patients were diagnosed with the type 2 VSD (perimembranous type) and a minor VSD. The details of the VSD types are listed in Table 2.
Table 1.
Variables | VSD group (N = 51) | Normal group (N = 25) |
---|---|---|
Age (years) | 22.12 ± 16.96 (min: 2; max: 65) | 29.30 ± 18.67 (min: 4.3; max: 65) |
Sex | ||
Male [n; (%)] | 30 (58.82%) | 14 (56%) |
Female [n; (%)] | 21 (41.18%) | 11 (44%) |
Height (cm) | 147.23 ± 27.86 | 155.04 ± 23.91 |
Weight (kg) | 46.75 ± 22.70 | 55.82 ± 24.63 |
Table 2.
VSD types (N = 51) | Case number (%) |
---|---|
Type I: infundibular, outlet | 2 (3.92%) |
Type II: perimembranous | 42 (82.35%) |
Type III: inlet, atrioventricular | 0 |
Type IV: muscular, trabecular | 5 (9.80%) |
Unknown | 2 (3.92%) |
Size classification* | |
Small | 30 (58.82%) |
Medium | 13 (25.49%) |
Large | 4 (7.84%) |
Unknown | 4 (7.84%) |
*Small VSD: Qp/Qs < 1.5; medium VSD: 1.5≦Qp/Qs < 2; large VSD: 2≦Qp/Qs; where Qp indicates pulmonary blood flow, Qs indicates systemic blood flow.
Two repeated heart sound recordings were obtained at each of the five standard auscultation spaces. For some subjects whose recordings did not qualify, owing to the presence of noise, more than two recordings were obtained within the same auscultation space to confirm the quality of the soundtracks. A total of 776 heart soundtracks were recorded from 76 subjects, including 525 soundtracks from VSD patients and 251 soundtracks from normal subjects. The number of soundtracks in the training and test sets is shown in Table 3.
Table 3.
Variables | VSD group | Normal group |
---|---|---|
Number of subjects | 51 | 25 |
Number of sound recordings | 525 | 251 |
The TAP-CRNN model was used to recognize systolic murmurs in the current study. The structure of TAP-CRNN is described in Fig. 1, in which the phonocardiogram (PCG) signals were first converted into the spectral domain using a short-time Fourier transform, with a frame length of 512 and a frameshift of 256. Then, each frame of PCG signals is represented by a 257-dimentional log-power spectral feature vector. An input signal was classified as a systolic murmur or a normal signal for training the TAP-CRNN model, which consists of four parts: convolutional, recurrent, temporal attentive pooling (TAP), and dense layers. The convolutional layers extracted invariant spatial–temporal representations from the spectral features. The recurrent layers were used in the following step to extract the long temporal context information from the representations. The TAP layers were then used to assign importance weights to each frame in the systolic regions. Finally, the classified results were generated by the dense layers according to the temporal attentive pooling feature outputted from the TAP layers.
The performance of the algorithm for systolic murmur recognition was analyzed for two different tasks, namely, a train–test split and K-fold cross-validation. Table 4 shows the performance of TAP-CRNN for systolic murmur recognition in the train–test split task. The CNN and CRNN models for systolic murmur recognition were also analyzed for comparison. The CNN model comprised three convolutional layers, where the first layer consisted of 32 filters with a kernel size of 1 × 4, the second layer included 32 filters with a kernel size of 1 × 4, and the third layer contained 32 filters with a kernel size of 4 × 4; the model also comprised two dense layers, with each layer composed of 512 neurons. The CRNN model comprised two convolutional layers, with each layer consisting of 16 filters with a kernel size of 1 × 4; two recurrent layers (long short-term memory unit), with each layer including 256 neurons; and two dense layers, with each layer containing 256 neurons. Compared with the CRNN architecture, a TAP-CRNN comprises an additional TAP layer. The hyperbolic tangent units were used in all the models, and the softmax unit was used in the last output layer. Adaptive moment estimation (Adam)28 was used as the optimizer. For the train–test split task, the entire set of data were divided into 70% (191 normal sounds, and 351 systolic murmur sounds) and 30% (60 normal sounds and 178 systolic murmur sounds) for training the murmur recognition models and testing their performances, respectively. For this task, the sensitivity and specificity scores were 88% and 85% for CNN, 92% and 93% for CRNN, and 97% and 98% for TAP-CRNN, respectively. In supplementary Tables 1–3, 2 × 2 tables of positive and negative events are shown. The receiver operating characteristic (ROC) curves of CNN, CRNN, and TAP-CRNN are shown in Fig. 2. The results show that the use of the TAP-CRNN model achieves a better accuracy for systolic murmur recognition when compared to the use of the CNN and CRNN models. The K-fold cross-validation task was used to further verify the reliability of the system performance. We conducted experiments using a fourfold (K = 4) setup. We first divided the entire set of PCG data into four groups, and roughly equal numbers of VSD patients and normal people were assigned to each group. We used data belonging to three out of these four groups for training the TAP-CRNN model, and the remaining group was used for testing. There were no overlapping subjects in the training and test sets. We carried out this procedure four times, the results of which have been listed in Table 5. From Table 5, we can see that the fourfold results are quite consistent and share the same trends as the results reported in Table 4 (the train– test split task). The average sensitivity and specificity scores over 4-folds were 97.18% and 91.98% of TAP-CRNN, confirming that the proposed TAP-CRNN can reliably produce satisfactory results for all evaluation metrics.
Table 4.
Accuracy | Sensitivity | Specificity | PPV | NPV | |
---|---|---|---|---|---|
CNN | 87.0% | 87.6% | 85.0% | 94.5% | 69.9% |
CRNN | 92.0% | 91.6% | 93.3% | 97.6% | 78.9% |
TAP-CRNN | 97.1% | 96.6% | 98.3% | 99.4% | 90.1% |
Table 5.
Accuracy | Sensitivity | Specificity | PPV | NPV | |
---|---|---|---|---|---|
1st fold | 98.4 | 99.2% | 96.7% | 98.5% | 98.3% |
2nd fold | 96.8 | 96.2% | 98.3% | 99.2% | 92.2% |
3rd fold | 96.4 | 99.3% | 90.0% | 95.7% | 98.2% |
4th fold | 90.2 | 94.0% | 82.9% | 91.2% | 87.9% |
Average | 95.45 | 97.18 | 91.98 | 96.15% | 94.15% |
The capability of the TAP-CRNN model for recognizing the systolic murmurs at the five standard auscultation sites was also analyzed (Table 6, supplemental Tables 4–8). Both the second aortic and the tricuspid areas showed 100% sensitivity and 100% specificity. The sensitivity was decreased in the other spaces, including the aortic (95.5%), pulmonic (94.1%), and mitral (94.1%) areas.
Table 6.
Auscultation area | Accuracy | Sensitivity | Specificity | PPV | NPV |
---|---|---|---|---|---|
Aortic area | 94.6% | 95.5% | 91.7% | 97.7% | 84.6% |
Pulmonic area | 95.7% | 94.1% | 100% | 100% | 85.7% |
Second aortic area/ Erb’s point | 100% | 100% | 100% | 100% | 100% |
Tricuspid area | 100% | 100% | 100% | 100% | 100% |
Mitral area/apex | 95.7% | 94.1% | 100% | 100% | 85.7% |
Discussion
A murmur is a sound generated by the turbulent blood flow in the heart. Under normal conditions, the blood flow in a vascular bed is smooth and silent. However, blood flow can be turbulent and produce extra noise when the heart has a structural defect29. Murmurs can be classified based on their timing, duration, intensity, pitch, and shape. Specific murmur patterns may occur as a result of many types of structural heart diseases14. For example, holosystolic murmurs, which are characterized by uniform intensity during the systolic period, usually appear in patients with mitral regurgitation (MR), tricuspid regurgitation (TR), or VSD30–32. Murmurs that occur during the systolic period with a crescendo-decrescendo shape are called systolic ejection murmurs and are often heard in patients with aortic stenosis (AS), pulmonic stenosis (PS), and atrial septal defect (ASD)30. Experienced cardiologists may successfully distinguish these specific heart sound patterns during routine auscultation, and this capability is important in disease diagnosis. However, it is always a challenge for young and inexperienced physicians to make a correct diagnosis based on auscultation17,33. Therefore, the development of tools that can automatically classify specific murmur types is necessary and clinically significant34,35.
In recent years, CNNs have been widely used in computer-aided diagnosis36,37. Previous studies have used a CNN to classify pathological heart sounds38,39. A recurrent neural network is another model frequently used in computer-aided diagnosis40,41. In this study, we combined CNN and RNN models (forming a CRNN model) to recognize the systolic murmurs from VSD patients. We used a convolutional unit to extract invariant spatial–temporal representations and the recurrent unit to capture long temporal-context information for systolic murmur recognition. In addition, the TAP mechanism was also applied in the CRNN model to assign an importance weight for each frame within the murmur regions. Finally, the overall model is called TAP-CRNN. From our experimental results, the TAP-CRNN model demonstrated an accuracy of 96% for distinguishing systolic murmurs from normal heart sounds, outperforming both CNN and CRNN without TAP.
For heart sounds recorded in the tricuspid and second aortic areas (Erb’s point), both the sensitivity and specificity reached 100% when the TAP-CRNN model was used. A high accuracy in these two areas is reasonable because the murmurs caused by the blood flow between the right and left ventricles can be most clearly heard in the tricuspid area or the lower left sternal border, which overlies the defect42.
The intensity of the murmur is inversely proportional to the size of the VSD. The ability of the algorithm to recognize the murmurs caused by a moderate or large VSD was also tested in the current study. In the test set, 63 soundtracks from 6 patients with moderate/large VSDs were included. When using the TAP-CRNN model, the murmurs of these soundtracks from moderate/large VSDs can be accurately recognized, except for two soundtracks recorded in the mitral area. Although the results obtained by TAP-CRNN are encouraging, we will further test the performance using a larger dataset of heart sound in the future.
This study has several limitations. As a major limitation, this study focused on the specific heart sound patterns of VSD, while not considering other types of structural heart diseases. Although heart murmurs can be heard in many other congenital and valvular heart diseases, such as atrial septal defects, patent ductus arteriosus, mitral regurgitation, and aortic regurgitation, patients with these diseases were not included in this study. Harmless heart murmurs, which occasionally occur in normal subjects, were also not included43–45. A larger heart sound database is currently being established to comprehensively collect heart sounds from patients with all types of structural heart diseases. An advanced version of the proposed TAP-CRNN algorithm that can recognize the specific murmur types in such diseases is also under development.
Conclusions
We demonstrated that a TAP-CRNN model can accurately recognize the systolic murmur of VSD patients. As compared to CNN and CRNN without TAP, the proposed TAP-CRNN achieves higher sensitivity and specificity scores for systolic murmurs detections in patients with VSDs. The results suggest that by incorporating the attention mechanism, the CRNN-based model can more accurately detect murmur signals. We also noted that sounds recorded from the second aortic and the tricuspid areas can facilitate more accurate murmur detection results as compared to other auscultation sites. The experimental results from the present study confirmed that the proposed TAP-CRNN serves as a promising model for the development of software to classify the heart murmurs of many other types of structural heart diseases.
Methods
In this section, we introduce our data source, algorithm, and analysis method.
Data source
The sound dataset used in this study included heart sounds recorded from subjects at the National Taiwan University Hospital (NTUH) using an iMediPlus electronic stethoscope. This study was approved by the research ethics committee of NTUH, and informed consent was obtained from all subjects or, if subjects are under 18, from a parent and/or legal guardian in accordance with the Declaration of Helsinki. It is also confirmed that all methods were carried out in accordance with relevant guidelines and regulations.
Sounds from patients diagnosed with VSD were categorized as the VSD group and sounds from patients without a significant heart malformation were categorized as a normal group. Auscultation was applied for each subject by a cardiologist with 30 years of experience to confirm whether a pathological systolic murmur occurred in patients with VSD. Normal subjects with innocent murmurs were not included in this study. Echocardiography was conducted on all subjects to confirm the disease diagnosis46.
For each subject, two repeated heart sound recordings lasting 10 s each were made at each of the following sites: the aortic area (the second intercostal space on the right sternal border), the pulmonic area (the second intercostal space on the left sternal border), the secondary aortic area/Erb’s point (the third intercostal space on the left sternal border), the tricuspid area (the fourth intercostal space on the left sternal border), and the mitral area/apex (the fifth intercostal space to the left of the midclavicular line)30,47. The sounds were recorded by trained study nurses under the supervision of an experienced cardiologist. The soundtracks were saved as WAV files.
The soundtracks collected were divided into training and test sets. Notably, the training and test sets are two mutually exclusive sets without an overlap.
Algorithm characteristics
In this study, a short-time fast Fourier transformation was used to transform the phonocardiogram (PCG) signal into a time–frequency representation (spectral features), where denotes the input feature, and N is the number of frames of X. Each frame is represented by a 257-dimensional log-power spectral feature vector. The collection of frames in X forms a spectrogram, which is generally used to visualize the characteristics of temporal signals varying over time (Fig. 3). In this study, the TAP-CRNN structure was used for classification27. Figure 1 shows the network architecture of TAP-CRNN, in which convolutional layers48 were used to extract invariant frequency-shift features Recurrent layers49 were used to explore the global temporal feature of a sequence from the recurrent layer’s outputs, and the TAP layers then extracted the temporal attentive feature and weighed the spectral features when generating the classification results. Figure 4 shows CRNN with a TAP mechanism. The idea here is to focus on important features or regions by introducing attention blocks. Two different attention approaches, local and global, were used to exploit the effectiveness of the TAP mechanism. The back-propagation algorithm is adopted to train the TAP-CRNN parameters to minimize the cross entropy50. In terms of global attention, the model decides to focus equally on all regions (global). By contrast, local attention focuses on small regions (local). The idea of global attention is to consider all outputs of the convolutional layer and the temporal summarization of the output of the recurrent layer. For global attention of the TAP, we employ a simple concatenated layer to construct the global attentive vector c(n) by combining the information from the output of the convolutional layer and the output of the recurrent layer , such as in the following:
1 |
where is the output of the recurrent layer at the last time step, and are the parameter matrices used to concatenate and , i.e., and , where and are the output dimensions of the convolutional and recurrent layers, respectively.
The global attentive vector is subsequently fed into the global attention block to produce the global attention weights (scalar) and is shown as follows:
2 |
where is the vector used to calculate the global attention weight matrix shared by all time steps, and is the global bias matrix. The global attention weights are used to weight the local features from the convolutional layer at each time step as follows:
3 |
In addition to the global attention, the local attention is used to further refine the feature extraction and is calculated in the following manner:
4 |
where , , and are the parametric matrices used for the local attention weight calculation. These local attention weights are used to weight the features such as in the following:
5 |
where is the output weight vector for local attention. The final attentive context is calculated as the average of the weighted outputs and is shown as follows:
6 |
After obtaining the attentive context , we concatenate it with the last time step output of the CRNN as the input s of the dense layers, such as in the following:
7 |
The dense layers are constructed using fully connected units. The relationship between feature s and the output of the first hidden layer is described as follows:
8 |
where and correspond to the weight and bias vector in the first layer, and F(.) is the activation function. After obtaining the output of the first hidden layer, the relationship between the current and next hidden layer can be expressed as follows:
9 |
where L is the total number of layers of neurons in the output layer. Thus, the relationship for the classification layer or the output layer can be described as follows:
10 |
where G(.) is the softmax function, and o is the final output of TAP-CRNN.
The importance coefficients provided by the global and local attention were regarded as a frame-based event presence likelihood (EPL), i.e., . To determine the classified result, the frames with low EPLs were ignored while being emphasized with high EPLs. Figure 5 illustrates the spectrogram (Fig. 5a) and the EPL score (Fig. 5b) of heart sounds from subjects with VSD, in which the murmur regions showed high EPLs when the global attention coefficients and the local attention coefficients were calculated. The features of the murmur regions with a high EPL will be emphasized during the feature extraction.
Statistical analysis
The sex distribution, mean age, mean height, and mean weight of the subjects were calculated. An independent sample t-test and a chi-square test were conducted to compare the differences between the VSD and normal groups in terms of continuous and categorical variables, respectively.
The soundtracks used in the test set were applied to test the recognition performance. The accuracy, sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for the distinction of the systolic murmur of VSD patients from the normal heart sounds of healthy volunteers were calculated51–53. A diagnosis using echocardiography was applied as the gold standard for these calculations9,10,54. The equations are as follows:
11 |
12 |
13 |
14 |
15 |
where Tp indicates a true positive, Tn indicates a true negative, Fp indicates a false positive, and Fn indicates a false negative.
Supplementary information
Acknowledgements
This study was supported by iMediPlus, Inc.
Author contributions
J.-K.W. diagnosed disease, provided the experimental concepts and designed the experiment. W.-C.W., Y.-F.C. and Y.T. performed the experiments, developed the algorithm and wrote the paper. K.-H.T and C.-H.C. provided the experimental concepts and supplied an electronic stethoscope for recording data. C.-Y.T. analyzed and labeled data and wrote the paper.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
is available for this paper at 10.1038/s41598-020-77994-z.
References
- 1.Yeh SJ, et al. National database study of survival of pediatric congenital heart disease patients in Taiwan. J. Formos. Med. Assoc. 2015;114:159–163. doi: 10.1016/j.jfma.2012.10.006. [DOI] [PubMed] [Google Scholar]
- 2.Chaudhry TA, Younas M, Baig A. Ventricular septal defect and associated complications. J. Pak. Med. Assoc. 2011;61:1001–1004. [PubMed] [Google Scholar]
- 3.Wu MH, et al. Prevalence of congenital heart disease at live birth in Taiwan. J. Pediatr. 2010;156:782–785. doi: 10.1016/j.jpeds.2009.11.062. [DOI] [PubMed] [Google Scholar]
- 4.Hoffman JIE, Kaplan S. The incidence of congenital heart disease. J. Am. Coll. Cardiol. 2002;39:1890–1900. doi: 10.1016/S0735-1097(02)01886-7. [DOI] [PubMed] [Google Scholar]
- 5.Ito T, Okubo T, Kimura M, Ito S, Akabane J. Increase in diameter of ventricular septal defect and membranous septal aneurysm formation during the infantile period. Pediatr. Cardiol. 2001;22:491–493. doi: 10.1007/s002460010281. [DOI] [PubMed] [Google Scholar]
- 6.Zhang J, Ko JM, Guileyardo JM, Roberts WC. A review of spontaneous closure of ventricular septal defect. Baylor Univ. Med. Cent. Proc. 2015;28:516–520. doi: 10.1080/08998280.2015.11929329. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Ammash NM, Wames CA. Ventricular septal defects in adults. Ann. Internal Med. 2001;135:812–824. doi: 10.7326/0003-4819-135-9-200111060-00011. [DOI] [PubMed] [Google Scholar]
- 8.Minette MS, Sahn DJ. Congenital heart disease for the adult cardiologist-ventricular septal defects. Circulation. 2006;114:2190–2197. doi: 10.1161/CIRCULATIONAHA.106.618124. [DOI] [PubMed] [Google Scholar]
- 9.Dearani JA, et al. 2018 AHA/ACC guideline for the management of adults with congenital heart disease. Circulation. 2019;139:e698–e800. doi: 10.1161/CIR.0000000000000603. [DOI] [PubMed] [Google Scholar]
- 10.Baumgartner H, et al. ESC guidelines for the management of grown-up congenital heart disease (new version 2010) Eur. Heart J. 2010;31:2915–2957. doi: 10.1093/eurheartj/ehq249. [DOI] [PubMed] [Google Scholar]
- 11.Tchervenkov CI, Roy N. Congenital heart surgery nomenclature and database project: pulmonary atresia—ventricular septal defect. Ann. Thorac. Surg. 2000;69:S25–S35. doi: 10.1016/s0003-4975(99)01285-0. [DOI] [PubMed] [Google Scholar]
- 12.McDaniel NL. Ventricular and atrial septal defects. Pediatr. Rev. 2001;22:265–270. doi: 10.1542/pir.22-8-265. [DOI] [PubMed] [Google Scholar]
- 13.Etoom Y, Ratnapalan S. Evaluation of children with heart murmurs. Clin. Pediatr. (Phila) 2014;53:111–117. doi: 10.1177/0009922813488653. [DOI] [PubMed] [Google Scholar]
- 14.Lessard E, Glick M, Ahmed S, Saric M. The patient with a heart murmur: evaluation, assessment and dental considerations. J. Am. Dent. Assoc. 2005;136:347–356. doi: 10.14219/jada.archive.2005.0176. [DOI] [PubMed] [Google Scholar]
- 15.Frank JE, Jacobe KM. Evaluation and management of heart murmurs in children. Am. Fam. Physician. 2011;84:793–800. [PubMed] [Google Scholar]
- 16.Kang G, et al. Prevalence and clinical significance of cardiac murmurs in schoolchildren. Arch. Dis. Child. 2015;100:1028–1031. doi: 10.1136/archdischild-2014-307819. [DOI] [PubMed] [Google Scholar]
- 17.Kumar K, Thompson WR. Evaluation of cardiac auscultation skills in pediatric residents. Clin. Pediatr. (Phila) 2013;52:66–73. doi: 10.1177/0009922812466584. [DOI] [PubMed] [Google Scholar]
- 18.Erickson BJ, Bartholmai B. Computer-aided detection and diagnosis at the start of the third millennium. J. Digit. Imaging. 2002;15:59–68. doi: 10.1007/s10278-002-0011-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Bluemke DA. Radiology in 2018: are you working with AI or Being replaced by AI? Radiology. 2018;287:365–366. doi: 10.1148/radiol.2018184007. [DOI] [PubMed] [Google Scholar]
- 20.Shen D, Wu G, Suk H-I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017;19:221–248. doi: 10.1146/annurev-bioeng-071516-044442. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Litjens G, et al. A survey on deep learning in medical image analysis. Med. Image Anal. 2017;42:60–88. doi: 10.1016/j.media.2017.07.005. [DOI] [PubMed] [Google Scholar]
- 22.Suzuki K. Overview of deep learning in medical imaging. Radiol. Phys. Technol. 2017;10:257–273. doi: 10.1007/s12194-017-0406-5. [DOI] [PubMed] [Google Scholar]
- 23.Chen TE, et al. S1 and S2 heart sound recognition using deep neural networks. IEEE Trans. Biomed. Eng. 2017;64:372–380. doi: 10.1109/TBME.2016.2559800. [DOI] [PubMed] [Google Scholar]
- 24.DeGroff CG, et al. Artificial neural network—based method of screening heart murmurs in children. Circulation. 2001;103:2711–2716. doi: 10.1161/01.CIR.103.22.2711. [DOI] [PubMed] [Google Scholar]
- 25.Tsao Y, et al. Robust S1 and S2 heart sound recognition based on spectral restoration and multi-style training. Biomed. Signal Process. Control. 2019;49:173–180. doi: 10.1016/j.bspc.2018.10.014. [DOI] [Google Scholar]
- 26.Dominguez-Morales JP, Jimenez-Fernandez AF, Dominguez-Morales MJ, Jimenez-Moreno G. Deep neural networks for the recognition and classification of heart murmurs using neuromorphic auditory sensors. IEEE Trans. Biomed. Circ. Syst. 2018 doi: 10.1109/TBCAS.2017.2751545. [DOI] [PubMed] [Google Scholar]
- 27.Lu, X., Shen, P., Li, S., Tsao, Y. &Kawai, H. Temporal attentive pooling for acoustic event detection. In Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH (2018). doi:10.21437/Interspeech.2018-1552
- 28.Kingma, D. P. & Ba, J. L. Adam: A method for stochastic optimization. In 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings (2015).
- 29.Pelech AN. The physiology of cardiac auscultation. Pediatr. Clin. N. Am. 2004;51:1515–1535. doi: 10.1016/j.pcl.2004.08.004. [DOI] [PubMed] [Google Scholar]
- 30.Chizner MA. Cardiac auscultation: rediscovering the lost art. Curr. Probl. Cardiol. 2008;33:326–408. doi: 10.1016/j.cpcardiol.2008.03.003. [DOI] [PubMed] [Google Scholar]
- 31.Alpert MA. Systolic murmurs. In: Walker HK, Hall WD, editors. Clinical Methods: The History, Physical, and Laboratory Examinations. Oxford: Butterworth; 1990. [PubMed] [Google Scholar]
- 32.Naik RJ, Shah NC. Teenage heart murmurs. Pediatr. Clin. N. Am. 2014;61:1–16. doi: 10.1016/j.pcl.2013.09.014. [DOI] [PubMed] [Google Scholar]
- 33.Mattioli LF, Belmont JM, Davis AM. Effectiveness of teaching cardiac auscultation to residents during an elective pediatric cardiology rotation. Pediatr. Cardiol. 2008;29:1095–1100. doi: 10.1007/s00246-008-9265-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Satou GM, et al. Telemedicine in pediatric cardiology: a scientific statement from the American Heart Association. Circulation. 2017;135:e648–e678. doi: 10.1161/CIR.0000000000000478. [DOI] [PubMed] [Google Scholar]
- 35.Leng S, et al. The electronic stethoscope. Biomed. Eng. Online. 2015;14:66. doi: 10.1186/s12938-015-0056-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Yamashita R, Nishio M, Do RKG, Togashi K. Convolutional neural networks: an overview and application in radiology. Insights into Imaging. 2018 doi: 10.1007/s13244-018-0639-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Brinker TJ, et al. Skin cancer classification using convolutional neural networks: systematic review. J. Med. Internet Res. 2018 doi: 10.2196/11936. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Demir F, Şengür A, Bajaj V, Polat K. Towards the classification of heart sounds based on convolutional deep neural network. Health Inf. Sci. Syst. 2019;7:16. doi: 10.1007/s13755-019-0078-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Bozkurt B, Germanakis I, Stylianou Y. A study of time-frequency features for CNN-based automatic heart sound classification for pathology detection. Comput. Biol. Med. 2018 doi: 10.1016/j.compbiomed.2018.06.026. [DOI] [PubMed] [Google Scholar]
- 40.Faust O, et al. Automated detection of atrial fibrillation using long short-term memory network with RR interval signals. Comput. Biol. Med. 2018 doi: 10.1016/j.compbiomed.2018.07.001. [DOI] [PubMed] [Google Scholar]
- 41.Lee C, Kim Y, Kim YS, Jang J. Automatic disease annotation from radiology reports using artificial intelligence implemented by a recurrent neural network. Am. J. Roentgenol. 2019 doi: 10.2214/AJR.18.19869. [DOI] [PubMed] [Google Scholar]
- 42.Begic E, Begic Z. Accidental heart murmurs. Med. Arch. 2017;71:284–287. doi: 10.5455/medarh.2017.71.284-287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Saunders NR. Innocent heart murmurs in children. Taking a diagnostic approach. Can. Fam. Physician. 1995;41:1512. [PMC free article] [PubMed] [Google Scholar]
- 44.Doshi AR. Innocent heart murmur. Cureus. 2018;10:e3689. doi: 10.7759/cureus.3689. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Danford DA, Martin AB, Fletcher SE, Gumbiner CH. Echocardiographic yield in children when innocent murmur seems likely but doubts linger. Pediatr. Cardiol. 2002;23:410–414. doi: 10.1007/s00246-002-1390-y. [DOI] [PubMed] [Google Scholar]
- 46.Mcleod G, et al. Echocardiography in congenital heart disease. Prog. Cardiovasc. Dis. 2018;61:468–475. doi: 10.1016/j.pcad.2018.11.004. [DOI] [PubMed] [Google Scholar]
- 47.Bickley, L. S. Bates’ Guide to Physical Examination and History-Taking - Eleventh Edition. (LWW, 2012).
- 48.Krizhevsky A, Sutskever I, Hinton GE. ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012 doi: 10.1016/j.protcy.2014.09.007. [DOI] [Google Scholar]
- 49.Weninger, F. et al. Speech enhancement with LSTM recurrent neural networks and its application to noise-robust ASR. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)9237, 91–99 (2015).
- 50.Rumelhart DE, Hinton GE, Williams RJ. Learning representations by back-propagating errors. Nature. 1986;323:533–536. doi: 10.1038/323533a0. [DOI] [Google Scholar]
- 51.Fawcett T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006;27:861–874. doi: 10.1016/j.patrec.2005.10.010. [DOI] [Google Scholar]
- 52.Bradley AP. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997;30:1145–1159. doi: 10.1016/S0031-3203(96)00142-2. [DOI] [Google Scholar]
- 53.Diel R, Loddenkemper R, Niemann S, Meywald-Walter K, Nienhaus A. Negative and positive predictive value of a whole-blood interferon-γ release assay for developing active tuberculosis: an update. Am. J. Respir. Crit. Care Med. 2011;183:88–95. doi: 10.1164/rccm.201006-0974OC. [DOI] [PubMed] [Google Scholar]
- 54.Deri A, English K. EDUCATIONAL SERIES IN CONGENITAL HEART DISEASE: Echocardiographic assessment of left to right shunts: atrial septal defect, ventricular septal defect, atrioventricular septal defect, patent arterial duct. Echo Res. Pract. 2018;5:R1–R16. doi: 10.1530/ERP-17-0062. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.