Abstract
Bowel sound (BS) is receiving more attention as an indicator of gut health since it can be acquired non-invasively. Current gut health diagnostic tests require special devices that are limited to hospital settings. This study aimed to develop a prototype smartphone application that can record BS using built-in microphones and automatically analyze the sounds. Using smartphones, we collected BSs from 100 participants (age 37.6 ± 9.7). During screening and annotation, we obtained 5929 BS segments. Based on the annotated recordings, we developed and compared two BS recognition models: CNN and LSTM. Our CNN model could detect BSs with an accuracy of 88.9% andan F measure of 72.3% using cross evaluation, thus displaying better performance than the LSTM model (82.4% accuracy and 65.8% F measure using cross validation). Furthermore, the BS to sound interval, which indicates a bowel motility, predicted by the CNN model correlated to over 98% with manual labels. Using built-in smartphone microphones, we constructed a CNN model that can recognize BSs with moderate accuracy, thus providing a putative non-invasive tool for conveniently determining gut health and demonstrating the potential of automated BS research.
Keywords: sound analysis, bowel sound, machine learning, neural network, deep learning
1. Introduction
Digestive tract issues, including inflammatory bowel disease (IBS), have been reported to affect an increasing number of patients worldwide [1]. Although exercise and functional foods, such as probiotics and prebiotics, are often used as countermeasures, it can be difficult for patients to determine the cause of their bowel conditions, and therefore, which treatments would be beneficial. A tool that senses gut health would enable personalized patient care; however, existing diagnostic tools, such as gut microbiota analysis services, X-ray imaging, and endoscopy, are highly invasive, costly, and limited to a hospital environment. Since gut health plays an important role in overall health and well-being, this issue must be addressed urgently.
A simple and recognizable means of sensing gut health is through bowel sounds (BSs), which are used as an indicator of migrating motor complex [2] and can be obtained through auscultation and from electronically recorded voice data. Various BS parameters are known to indicate diseases and other conditions. For example, the loss of BSs can be used to diagnose ileus, and monitoring and quantifying BSs after surgery for gastric cancer can prevent postoperative ileus [3]. In addition, differences in the sound-to-sound interval (SSI) have been reported between healthy subjects and patients with IBS or Crohn’s disease, indicating that BSs can be used to diagnose bowel-related diseases and bowel motility [4,5]. With regard to food, BSs are also expected to be effectively used for food-intake related sensing, such as capturing changes in BSs with coffee and carbonated water intake [6] and changes in BS indices with the amount of energy contained in food [7].
BSs can be analyzed directly by a physician or using a learning model that recognizes sounds collected using a special recording instrument [4,6,8,9,10,11]. For instance, Craine et al. [4] used a method involving stationary-nonstationary filters with wavelet transformation, whereas Wang et al. [7] used an artificial neural network and Sakata and Suzuki [11] used a classification algorithm capturing sound features. Although BS measurement is not invasive and various analysis methods have been established, it requires specialized devices such as stethoscopes and contact-type recording devices that are only available in a hospital environment. This has limited the collection of BS data from healthy people, leading to a lack of basic information on BSs and their therapeutic applications [12]. Smartphones with recording capabilities are ubiquitous worldwide and have been used to acquire and analyze biological sounds. For instance, smartphones have been used to diagnose sleep apnea syndrome by measuring breath sounds and snoring during sleep [13]. The application of machine learning (ML) models has led to advancements in the use of biological sound as a source for medical diagnosis [14]. Recent examples include the detection of COVID 19 detection [15], and chronic heart failure [16]. However, because the accuracy of the measurement has an impact on medical diagnosis, the application of the models in a clinical setting has been challenging.
However, smartphones have not yet been used to measure and analyze BSs. In this study, we constructed a BS recognition model based on sounds obtained by subjects themselves using a smartphone and examined whether the model had similar accuracy to previous models based on BSs obtained using dedicated devices. Since general-purpose recording devices can result in strong noise, we used a convolutional neural network (CNN) that can automatically extract features from input data with high accuracy and low noise, which has previously been used for image, speech, and more recently, BS recognition [8,9]. The CNN model constructed using built-in smartphone microphones could recognize BS with moderate accuracy. Consequently, we developed a simple, less invasive method that allows anyone to easily measure BSs with minimal mental, physical, and economic burden. This method could improve public health by accelerating BS research and increasing our knowledge of gastrointestinal diseases. Furthermore, this study demonstrates the potential of collecting more BS data, which can advance the automated BS research.
2. Materials and Methods
2.1. Data Collection
This study was approved by the Institutional Review Board of Suntory Holdings (Kyoto, Japan). Study participants were recruited from employees of the Suntory World Research Center (Kyoto, Japan). BS recordings were collected from 100 participants, including 43 men and 57 women between the ages of 20 and 65. All participants provided informed consent.
Apple iPhone 7 (Apple Inc. 2017, Cupertino, CA, USA) smartphones were used for all experiments. A range of iPhone models (6s, 7, 8, SE2, and X) were assessed prior to the study to determine whether the models had different levels of electromagnetic interference. No significant differences in sound quality were observed among the models tested. All study participants were given iPhones with instructions for installing the BS recording app. Since fetal sounds may contaminate BS recordings, we excluded pregnant individuals and those intending to become pregnant during the study.
The test period consisted of four days of answering daily surveys and one day of BS recording. All participants agreed to completely abstain from eating food and drinking alcohol from 10:00 p.m. until the end of the BS recording the following morning. Each participant joined an online session at 9:00 a.m. and was instructed to adopt a seated position (Figure 1a) to measure their own BSs in a location where the noise level was less than 50 dB. This threshold was determined through prior experiments that tested the audible range of the iPhone7 microphone using white noise and the Voice Memo app (Apple Inc. 2016). The iPhone 7 was capable of recording audio signal from 40 Hz–16 kHz, and BSs were audible if the background sound was less than 60 dB. Each participant was asked to place the bottom microphone of the smartphone perpendicular to the abdomen, approximately 9 cm away from the navel in both directions (Figure 1b), and 60 s recordings were made on both sides of the abdomen in the morning and before bedtime, for a total of eight recordings.
Figure 1.
Bowel sound measurement positions. (a) Seated recording posture; (b) microphone placement over the lower right and left parts of the abdomen; (c) a smartphone screen showing the measurement instructions.
2.2. Data Analysis
BS recordings were downloaded from a server contracted by Suntory Holdings, Ltd., transformed into a spectrogram, and divided into multiple successive non-overlapping 100 ms segments due to the 48 kHz sampling rate and the 4800 non-uniform fast Fourier transformation (NFFT).
Three human raters trained in BS identification by a medical doctor performed double blinded annotations on all segments using Audacity(R), version 2.4.2, open-source sound editing software [17]. Annotations consisted of binary labeling based on the presence or absence of BSs. Sound data labeled identically by more than two raters was used as the training data set. If adjacent BS labels were more than 10 ms apart, they were treated as independent labels. If adjacent BS labels were within 10 ms, they were treated as one unified label.
2.3. Non-BS Characterization
Non-BSs were categorized as follows: (1) circuit noise originating from the smartphone; (2) biological noises originating from participant internal organs, such as heartbeats; (3) background noise originating from the surrounding environment. Only sound files in which the human raters could clearly identify BSs were used for further analysis.
2.4. CNN Model Design
ML models are effective for identifying BSs [7,12,14,15,16,18], among which CNNs have been reported to offer better BS recognition accuracy [8]. Another neural network model that has a good reputation for sound recognition is long short-term memory (LSTM), which can process not only single data points, but the entire data sequence [19,20]. Due to the seemingly random nature and dynamic range of BSs, we developed and compared CNN and LSTM-based BS recognition models.
The CNN model consisted of seven components: spectrogram formation (regular convolution), batch normalization and dropout, rectified linear unit activation, max pooling, and fully connected layers (Figure 2). Hyper-parameter experiments determined a frame length of 100 ms with 75% overlap allowed. The total length of the recording was 7740 s. For the input architecture, we used a mini batch size of 32, data height of 3, and data width of 11. One hundred thousand iterations were used to train the model. Dropout was set after the third layer.
Figure 2.
The 100 ms segment of spectrogram was fed to CNN building blocks where it consisted of two sets of three convolution layers, batch normalization, dropout, rectified linear unit (ReLU) activation function, max-pooling layer, and fully connected layer (FC).
First, the 60 s audio file with a sampling frequency of 48 kHz was converted into a WAV file using ffmpeg software (version 3.4.11) [21] and then converted into a mel scale spectrogram using the librosa package (version 0.8.0) in Python. The hanning window function was used to create the time axis component of the spectrogram using the Fourier transform process. A mel spectrogram, wherein the frequencies are converted to the mel scale, was used to display BS findings. Mel scale is ideal for detecting differences in lower frequencies than in higher frequencies. The frequencies of BS are reported to be in the range of 100–1000 Hz [12,20]; therefore, mel spectrogram was chosen for our study.
2.5. CNN Model Setup
The frame length was set to 100 ms with a 75% overlap allowed. It was determined that three building blocks, consisting of two convolution layers and a max pooling layer, were the most optimal for the model (Table S1). For the number of channels, an 8, 16, 16 combination was determined to be optimal along with a 3 × 3 kernel size (Table S2).
2.6. Model Evaluation
To accurately evaluate model performance, we selected a test set representing the average proportion of BSs and noise in all recordings. To determine how accurately the model could detect BSs over noise, holdout and cross evaluation methods were used. The dataset was divided into nine folds, each containing BS data for 15 subjects with no overlap. For holdout evaluation, a single train-test split was run, meaning that only the first split was tested (‘Split1’ in Figure 3). This test was repeated five times. For cross validation, which runs multiple train-test splits, nine fold cross validations were performed with eight splits and uncorrected labels, and the average results were calculated.
Figure 3.
Cross evaluation schematics.
2.7. SSI Correlation Analysis
To determine how accurately the CNN model could recognize BSs, we used the correct labels and prediction results to calculate the fasting SSI (Figure 4), which is the interval between BSs defined as the period between the end of one BS and the start of the next BS.
Figure 4.
Sound to sound interval was calculated by the length between the end of the bowel sound and the start of the consecutive bowel sound. The green colored area in the top horizontal bar labeled “actual” shows the bowel sound length that have been manually labeled by human raters. The bottom one with gray colored area labeled “prediction” shows the bowel sound length that have been predicted by the models. The interval of the bowel sounds is used as an index called SSI.
2.8. Software
Models were run using Python script version 3.8. Audio files were processed using librosa package (version 0.8.0), and pytorch (version 1.3.1) was used for network creation and model training. Hyper-parameter tuning was performed using Amazon Ec2 instance (g3xlarge 30.5 GB RAM with 8 GPU, Amazon.com Inc., Seattle, WA, USA) with a Linux operating system.
2.9. Statistical Analysis
To compare the performance of our two ML models, four metrics were employed to evaluate accuracy, precision, sensitivity, and the F measure, which provides the harmonic mean of precision and sensitivity, and is a better measure than accuracy when working with an imbalanced binary classification dataset. Pearson’s method was used to test the correlation between manually labeled and predicted BS, with p < 0.05 defined as significant.
Accuracy = (TP + TN)/(TP + TN + FP + FN) | (1) |
Precision = TP/(TP + FP) | (2) |
Sensitivity = TP/(TP + FN) | (3) |
F measure = 2∗(Precision∗Sensitivity)/(Precision + Sensitivity) | (4) |
TP and TN refer to correctly classified BS and non-BS counts, respectively. FP and FN refer to falsely recognized and falsely rejected BS counts, respectively.
3. Results
In this study, audio files were collected from 100 individuals. Details of participant age, height, weight, and body mass index (BMI) are included in Table 1 below. The study participants had a wide age range and a variety of body types. All the participants completed the experiments. However, some audio files were not audible by the human raters due to the noise contaminations (Figure 5), so we used 144 audio files from the smartphone for this study.
Table 1.
Study participant characteristics.
All (n = 100) | Male (n = 43) | Female (n = 57) | |
---|---|---|---|
Age | 37.6 ± 9.7 | 38 ± 9.7 | 37.3 ± 9.7 |
Height (cm) | 165.7 ± 8.2 | 172.5 ± 5.2 | 160.6 ± 6.0 |
Weight (kg) | 58.3 ± 9.6 | 65.9 ± 7.8 | 52.6 ± 6.4 |
Body Mass Index | 21.1 ± 2.2 | 22.1 ± 2.0 | 20.4 ± 2.1 |
Figure 5.
Flowchart of participant recruitment and data acquisition. All 100 participants completed the study. Due to noise contamination, the data for only 77 participants was used to build the machine learning model.
3.1. BSs Can Be Recorded Using a Smartphone Microphone
With the developed procedure, study participants were able to record BSs using a smartphone microphone (Figure 6). Smartphone microphones generally support a frequency range of 20–20,000 Hz. The mel spectrogram produced in this study showed various types of BSs, indicating that BSs occur within the audible range of smartphone built-in microphones.
Figure 6.
Spectrogram depicting bowel sounds recorded using the bottom microphone of a smartphone held in the right lower quadrant of the abdomen. Individual short BSs with multiple peaks are visible around the 100–800 Hz range.
3.2. Experimental Results
3.2.1. Holdout Evaluation Results
Our CNN model had an accuracy of 83.9%, precision of 75.7%, sensitivity of 78.6%, and F measure of 77.0%. The LSTM model reached an accuracy of 77.2%, precision of 53.7%, sensitivity of 87.9%, and F measure of 65.8% (Table 2, Figure 7). Thus, the CNN model had better accuracy and F measures than the LSTM model.
Table 2.
Holdout evaluation results for the two models.
Accuracy | Precision | Sensitivity | F Measure | |
---|---|---|---|---|
CNN | 0.839 ± 0.013 | 0.757 ± 0.062 | 0.786 ± 0.076 | 0.770 ± 0.016 |
LSTM | 0.772 ± 0.027 | 0.616 ± 0.035 | 0.887 ± 0.056 | 0.727 ± 0.024 |
Figure 7.
Two BS recognition models; CNN and LSTM are compared by using two evaluation methods; a holdout and a cross evaluation. (a) F measure results. (b) ACC results. (c) The correlation coefficient (%) for sound-to-sound interval (SSI) and sound duration (SD).
3.2.2. Cross Validation Results
Cross validation revealed that the CNN model had an accuracy of 88.9%, precision of 70.5%, sensitivity of 74.9%, and F measure of 72.3%. Meanwhile, the LSTM model reached an accuracy of 82.4%, precision of 53.7%, sensitivity of 87.9%, and F measure of 65.8%. Together, these results indicate that the CNN model had a better accuracy and F measure than the LSTM model (Table 3, Figure 7).
Table 3.
Cross evaluation results for the two models.
Accuracy | Precision | Sensitivity | F Measure | |
---|---|---|---|---|
CNN | 0.889 ± 0.038 | 0.705 ± 0.140 | 0.749 ± 0.142 | 0.723 ± 0.097 |
LSTM | 0.824 ± 0.080 | 0.537 ± 0.216 | 0.879 ± 0.067 | 0.658 ± 0.170 |
3.2.3. SSI Classification and Correlation Results
To compare the ability of the model for BS recognition against human hearings, the correlation coefficients of the SSI and the sound duration (SD) were calculated (Table 4, Figure 7).
Table 4.
The correlation coefficient of the sound-to-sound interval (SSI) and the sound duration (SD) between manual labeling and prediction.
SSI | SD | |
---|---|---|
CNN-holdout | 0.992 ± 0.004 | 0.960 ± 0.016 |
CNN-cross | 0.940 ± 0.047 | 0.905 ± 0.091 |
LSTM-holdout | 0.921 ± 0.022 | 0.822 ± 0.020 |
LSTM-cross | 0.872 ± 0.140 | 0.742 ± 0.195 |
As for the SSI, the holdout evaluation result of the CNN model showed 99.2% correlation between the manual labeling and the prediction, and the cross-evaluation result showed 94.0% of the correlation. The LSTM model results of the holdout and the cross evaluation showed 92.1% and 87.2%, respectively. As for the SD, which is the sum of the total BS sound in a single file, the CNN holdout result showed 96.0% correlation, and the cross evaluation was 90.5%. The LSTM holdout result showed 82.2% correlation, and the cross evaluation was 74.2% (Table 4, Figure 7). It was found that both SSI and SD had the best correlation coefficient with CNN holdout evaluation method.
4. Discussion
Gut health is closely related to mental and physical quality of life; however, current diagnostic tests are highly invasive, costly, and limited to hospital settings. Since BSs can be used as a physiological indicator of gut health, we developed a prototype BS recording application using built-in smartphone microphones. Using the smartphone application, 100 subjects recorded BSs at home with a background noise level below 50 dB. Although BS measurement was carried out remotely, live instruction was provided to ensure that all participants placed the smartphone in the correct position for exactly 60 s. Importantly, BSs were audible using the built-in microphone of the smartphone.
Based on the recording data obtained from these experiments, two ML-based models were built. The CNN model, which was tuned for the sound quality of the smartphone microphone, had a higher F measure (77.0%) and moderate accuracy (83.9%), suggesting that the sound input from the built-in microphone is adequate for recognizing BSs using the proposed CNN model. Cross evaluation further confirmed that the CNN model displayed better accuracy and F measures than the LSTM model. Therefore, we calculated the SSI using the CNN model results and compared the correct labels and prediction results to determine whether the model output made biological sense.
In this study, we found that it is possible to acquire BS data using a smartphone at a sensory evaluation level, as confirmed by a gastrointestinal surgeon. The frequency range of the acquired BSs was approximately 100–1000 Hz (Figure 6), consistent with the frequency range of BS acquired using stethoscopes and other dedicated devices [12]. To confirm the audible range of the iPhone 7, we checked the frequency range that can be acquired when white noise is played. The frequency range was found to be 10–10,000 Hz, validating our finding that the iPhone 7 can record BSs. In addition, the smartphone microphone is slightly recessed, meaning that noise is reduced when the microphone is placed close to the body, allowing BSs to be recorded accurately. Although previous studies have used specific devices to detect BSs [22], this study is the first to use smartphones. Notably, we found that the BS recordings made by smartphones were comparable to past studies in terms of accuracy and F measure (Table 5).
Table 5.
Summary of recent bowel sound algorithms. This summary is based largely on the results in Table 10 of Ficek et al. [9].
Author | Algorithm | Input Source | ACC | F Measure |
---|---|---|---|---|
Sato et al. [6] | ANN | non-contact microphone | 90% | 15% |
Kumar et al. [23] | SVM | contact microphone | 75% | N/A |
Liu et al. [19] | CNN | custom design contact microphone | 92% | N/A |
Ficek et al. [9] | CNN | custom design contact microphone | 97% | 66% * |
Ours | CNN | smartphone microphone | 89% ** | 72% |
* F measure for Ficek et al. was calculated by reproducing their models and using the same test dataset as ours (Table S3). ** Our ACC by a holdout evaluation was 83%, and the ACC by a cross-evaluation was 89%. Since the ACC reported by Ficek et al. is a cross-evaluation result, our cross-validation result instead of the holdout evaluation result was inserted.
Since BSs account for a small percentage of the total measurement time (i.e., low frequency), we created the CNN model to avoid false negatives so as not to miss BSs, with an emphasis on the F measure in addition to accuracy. The SSI obtained from the manual labeling of subjects in this study was 2.00 ± 7.92, which is similar to the results of a previous study conducted by Craine et al. [4]. The correlation between the SSI calculated from manually labeled BSs and that calculated from automatically extracted BSs was very high (R = 0.99, p < 0.05), indicating that the model developed in this study can calculate the SSI. Previous studies have reported that the SSI can be used to predict IBS and Crohn’s disease, and that the SSI can change before and after the consumption of beverages (e.g., carbonated water [6]). Thus, our technology could be used to diagnose these diseases and make suggestions to improve quality of life using smartphones.
Although we were able to construct a model to recognize BSs, it remains unclear what meaning can be assigned to different types of BSs. Future cross-sectional studies should therefore clarify the relationship between BSs and other clinical findings (e.g., bowel diseases, stool frequency and shape, and timing of acquisition). Recent studies have shown that the diversity of gut microbiota correlates with gastrointestinal motility [24], and that it may be possible to predict defecation from BSs [25]. Thus, we believe it is important to clarify the relationships between the SSI and other BSs. Since smartphones are not dedicated measuring devices, they detect a lot of background noise from the external environment, especially in the low-frequency band. This could bury important features by decreasing the signal-to-noise ratio. Consequently, it is necessary to consider how and where to apply the device and to keep the device away from nearby electronic devices.
During general auscultation, BS intervals of 2–15 s/min are considered normal, with a decrease in this interval indicating an increase in peristalsis and an increase in this interval indicating a decrease or absence of peristalsis. The mean value of the BSs obtained in this study was small, which is enough to be diagnosed as increased peristalsis during general auscultation. This labeling was conducted using fragments (0.1 s) so small that it is difficult for a person to recognize them without spectrogram visualization; therefore, to establish the extent to which the calculated SSI is related to peristalsis, BS measurements by manometry and colonic transit tests should be performed in parallel, as seen in a previous study [26]. This formula would allow clinical-level gut diagnosis to be performed at home using a smartphone.
5. Conclusions
Using a built-in smartphone microphone, we built a CNN model that can recognize BSs with moderate accuracy, thus providing a foundation for new non-invasive tools to conveniently measure gut health. This study has significant implications because it demonstrates the reliability of the CNN model for recognizing BSs using the microphone of a smartphone. Since many people own smartphones, more BS data can be collected; however, the prototype recording app needs to be improved in order to attract more users. In conclusion, we believe this smartphone application will bring gut health care to the next stage by offering a simple BS measurement tool to users, thus improving a healthy lifestyle.
Acknowledgments
We would like to thank Ryo Ogawa and Nobuya Tachimori from ONKYO Limited for their consultations on deep learning parameter tunings. Additionally, we would like thank Makoto Yoshida and Yusuke Kondo from ONKYO Limited for assisting with microphone testing prior to this study. We also would like to Jan Nowak from Poznan University of Medical Science and Robert Nowak from Warsaw University of Technology for giving us consultation on BS labeling protocol.
Supplementary Materials
The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s23010407/s1, Table S1: Parameter tuning results of the convolutional layers and the number of channels, Table S2: Parameter tuning results of the kernel size, Table S3: Model performance between the recent similar work by Ficek et al. [9] and ours with the F measure.
Author Contributions
N.K. and Y.K. contributed equally to this work. Conceptualization, N.K., Y.K. and M.Z.; methodology, Y.K.; software, Y.K.; validation, N.K. and Y.K.; formal analysis, Y.K.; investigation, Y.K. and N.K.; resources, M.Z.; data curation, Y.K.; writing—original draft preparation, Y.K.; writing—review and editing, N.K.; visualization, Y.K.; supervision, M.Z.; project administration, H.M.; funding acquisition, H.M. and N.M. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
The study was conducted in accordance with the Declaration of Helsinki and was approved by the Institutional Review Board of Suntory Holdings Limited (IRB No. 15000139) for studies involving humans.
Informed Consent Statement
Written informed consent was obtained from all study participants.
Data Availability Statement
Due to the nature of this research, data is not available at this point.
Conflicts of Interest
The authors declare no conflict of interest.
Funding Statement
This research received no external funding.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Ng S.C., Shi H.Y., Hamidi N., Underwood F.E., Tang W., Benchimol E.I., Panaccione R., Ghosh S., Wu J.C.Y., Chan F.K.L., et al. Worldwide incidence and prevalence of inflammatory bowel disease in the 21st century: A systematic review of population-based studies. Lancet. 2017;390:2769–2778. doi: 10.1016/S0140-6736(17)32448-0. [DOI] [PubMed] [Google Scholar]
- 2.Tomomasa T., Morikawa A., Sandler R.H., Mansy H.A., Koneko H., Masahiko T., Hyman P.E., Itoh Z. Gastrointestinal sounds and migrating motor complex in fasted humans. Am. J. Gastroenterol. 1999;94:374–381. doi: 10.1111/j.1572-0241.1999.00862.x. [DOI] [PubMed] [Google Scholar]
- 3.Namikawa T., Yamaguchi S., Fujisawa K., Ogawa M., Iwabu J., Munekage M., Uemura S., Maeda H., Kitagawa H., Kobayashi M., et al. Real-time bowel sound analysis using newly developed device in patients undergoing gastric surgery for gastric tumor. JGH Open. 2021;5:454–458. doi: 10.1002/jgh3.12515. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Craine B.L., Silpa M., O’Toole C.J. Computerized auscultation applied to irritable bowel syndrome. Dig. Dis. Sci. 1999;44:1887–1892. doi: 10.1023/A:1018859110022. [DOI] [PubMed] [Google Scholar]
- 5.Craine B.L., Silpa M., O’Toole C.J. Enterotachogram analysis to distinguish irritable bowel syndrome from Crohn’s disease. Dig. Dis. Sci. 2001;46:1974–1979. doi: 10.1023/A:1010651602095. [DOI] [PubMed] [Google Scholar]
- 6.Sato R., Emoto T., Gojima Y., Akutagawa M. Automatic bowel motility evaluation technique for noncontact sound recordings. Appl. Sci. 2018;8:999. doi: 10.3390/app8060999. [DOI] [Google Scholar]
- 7.Wang N., Testa A., Marshall B.J. Development of a bowel sound detector adapted to demonstrate the effect of food intake. Biomed. Eng. 2022;21:1. doi: 10.1186/s12938-021-00969-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Zhao K., Jiang H., Wang Z., Chen P., Zhu B., Duan X. Long-term bowel sound monitoring and segmentation by wearable devices and convolutional neural networks. IEEE Trans. Biomed. Circuits Syst. 2020;14:985–996. doi: 10.1109/TBCAS.2020.3018711. [DOI] [PubMed] [Google Scholar]
- 9.Ficek J., Radzikowski K., Nowak J.K., Yoshie O., Walkowiak J., Nowak R. Analysis of gastrointestinal acoustic activity using deep neural networks. Sensors. 2021;21:7602. doi: 10.3390/s21227602. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Du X., Allwood G., Webberley K.M., Osseiran A., Marshall B.J. Bowel sounds identification and migrating motor complex detection with low-cost piezoelectric acoustic sensing device. Sensors. 2018;18:4240. doi: 10.3390/s18124240. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Sakata O., Suzuki Y. Optimum Unit Time on Calculating Occurrence Frequency of Bowel Sounds for Real-Time Monitoring of Bowel Peristalsis. Int. J. Signal Process. Syst. 2016;4:465–468. doi: 10.18178/ijsps.4.6.465-468. [DOI] [Google Scholar]
- 12.Nowak J.K., Nowak R., Radzikowski K., Grulkowski I., Walkowiak J. Automated bowel sound analysis: An overview. Sensors. 2021;21:5294. doi: 10.3390/s21165294. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Kim D.H., Kim S.W., Hwang S.H. Diagnostic value of smartphone in obstructive sleep apnea syndrome: A systematic review and meta-analysis. PLoS ONE. 2022;17:e0268585. doi: 10.1371/journal.pone.0268585. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Peruzzi G., Galli A., Pozzebon A. A Novel Methodology to Remotely and Early Diagnose Sleep Bruxism by Leveraging on Audio Signals and Embedded Machine Learning; Proceedings of the 2022 IEEE International Symposium on Measurements & Networking (M&N); Padua, Italy. 18–20 July 2022; [DOI] [Google Scholar]
- 15.Hemdan E.E., El-Shafai W., Sayed A. CR19: A framework for preliminary detection of COVID-19 in cough audio signals using machine learning algorithms for automated medical diagnosis applications. J. Ambient. Intell. Humaniz. Comput. 2022;1:1–13. doi: 10.1007/s12652-022-03732-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Gjoreski M., Gradisek A., Budna B., Gams M., Poglajen G. Machine learning and end-to-end deep learning for the detection of chronic heart failure from heart sounds. IEEE Access. 2020;8:20313–20324. doi: 10.1109/ACCESS.2020.2968900. [DOI] [Google Scholar]
- 17.Audacity® Software Is Copyright The Name Audacity. [(accessed on 20 February 2021)]. Available online: https://audacityteam.org/
- 18.Yin Y., Jiang H., Feng S., Liu J., Chen P., Zhu B., Wang Z. Bowel sound recognition using SVM classification in a wearable health monitoring system. Sci. China Inf. Sci. 2018;61:222–224. doi: 10.1007/s11432-018-9395-5. [DOI] [Google Scholar]
- 19.Liu J., Yin Y., Jiang H., Kan H., Zhang Z., Chen P., Zhu B., Wang Z. Bowel sound detection based on MFCC feature and LSTM neural network; Proceedings of the Biomed Circuits System Conference; Cleveland, OH, USA. 17–19 October 2018; 2018. pp. 1–4. [DOI] [Google Scholar]
- 20.Hochreiter S., Schmidhuber J. Long short-term memory. Neural Comput. 1997;9:1735–1780. doi: 10.1162/neco.1997.9.8.1735. [DOI] [PubMed] [Google Scholar]
- 21.FFmpeg FFmpeg Documentation. [(accessed on 18 December 2020)]. Available online: https://ffmpeg.org/documentation.html.
- 22.Inderjeeth A.J., Webberley K.M., Muir J., Marshall B.J. The potential of computerised analysis of bowel sounds for diagnosis of gastrointestinal conditions: A systematic review. Syst. Rev. 2018;7:124. doi: 10.1186/s13643-018-0789-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Kumar T.S., Soiland E., Stavdahl O., Fougner A.L. Pilot study of early meal onset detection from abdominal sounds; Proceedings of the 2019 E-Health and Bioengineering Conference (EHB); Iasi, Romania. 21–23 November 2019; [DOI] [Google Scholar]
- 24.Asnicar F., Leeming E.R., Dimidi E., Mazidi M., Franks P.W., Al Khatib H., Valdes A.M., Davies R., Bakker E., Francis L., et al. Blue poo: Impact of gut transit time on the gut microbiome using a novel marker. Gut. 2021;70:1665–1674. doi: 10.1136/gutjnl-2020-323877. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Zhang T., Huang Z., Zou Y., Zhao J., Ke Y. A prediction model of defecation based on BP neural network and bowel sound signal features. Sensors. 2022;22:7084. doi: 10.3390/s22187084. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Kim K.S., Seo J.H., Song C.G. Non-invasive algorithm for bowel motility estimation using a back-propagation neural network model of bowel sounds. Biomed. Eng. 2011;10:69. doi: 10.1186/1475-925X-10-69. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Due to the nature of this research, data is not available at this point.