Abstract
To achieve the simultaneous and unobtrusive breathing rate (BR) and heart rate (HR) measurements during nighttime, we leverage a far-infrared imager and an infrared camera equipped with IR-Cut lens and an infrared lighting array to develop a dual-camera imaging system. A custom-built cascade face classifier, containing the conventional Adaboost model and fully convolutional network trained by 32K images, was used to detect the face region in registered infrared images. The region of interest (ROI) inclusive of mouth and nose regions was afterwards confirmed by the discriminative regression and coordinate conversions of three selected landmarks. Subsequently, a tracking algorithm based on spatio-temporal context learning was applied for following the ROI in thermal video, and the raw signal was synchronously extracted. Finally, a custom-made time-domain signal analysis approach was developed for the determinations of BR and HR. A dual-mode sleep video database, including the videos obtained under environment where illumination intensity ranged from 0 to 3 Lux, was constructed to evaluate the effectiveness of the proposed system and algorithms. In linear regression analysis, the determination coefficient (R2) of 0.831 had been observed for the measured BR and reference BR, and this value was 0.933 for HR measurement. In addition, the Bland-Altman plots of BR and HR demonstrated that almost all the data points located within their own 95% limits of agreement. Consequently, the overall performance of the proposed technique is acceptable for BR and HR estimations during nighttime.
Introduction
Sleep plays an important role in people's mental and physical health as well as well-being throughout life. Sleep disorders have been considered as a significant public health issue, with nearly 30 percent of the worldwide adults suffering from them [1]. It is noteworthy that a vast number of undiagnosed people are unaccounted due to the high costs and inconvenient diagnostic methods [2]. Some chronic health problems, such as obesity and diabetes, are reported to be correlative with the sleep disorders [3]. Sleep disorders can also influence the way one person thinks, works, reacts and learns [4], and thus may result in a series of terrible accidents, such as traffic crashes. According to the American Academy of Sleep Medicine, the specific sleep disorder will appear with some unique symptoms [5]. Therefore, it is essential to gather the various sleep related information including sleep time, body movement times, sleep position, heart rate (HR) and breathing rate (BR) for the diagnosis of sleep disorders. However, most current measurement technologies are intrusive, cumbersome, inconvenient and expensive, which in turn bring about discomfort, stress and even soreness of the patients. Furthermore, to record the sleep related data, the patients have to use many on-body sensors and sleep in the strange environments. This may have an impact on the collection of natural and accurate sleep information. For the above reasons, the diagnosis of sleep disorders has become an extremely difficult task.
Optical measurement techniques offer the potential for the contactless and unobtrusive estimation of sleep related information. The Microsoft Kinect sensor, an off-the-shelf depth camera, was utilized to remotely recognize the sleep movement and posture [6]. The same device was used by the other group of investigators who extracted frequency and regularity features from breathing curves for the sleep disorders analysis [7]. The capability of custom-made infrared imaging system for sleep monitoring had been also studied. A CMOS camera coupled with an infrared light operating at wavelength of 825 nm was applied for detecting the movement pattern during sleep [8]. Martinez and Stiefelhagen captured the fixed infrared dot matrix, and subsequently, the BR was estimated from the displacements of some dots caused by chest movements [9]. Apart from the methods based on movement detection, the variations of pixel intensities are also widely used signals to be served as sleep disorder symptoms. Thermal imaging technique can detect the radiation emitted by the objects whose temperature exceeds absolute zero, and therefore, it is very suitable for executing the tasks during night such as sleep monitoring. In terms of sleep apnea event detection, the thermal imagery was employed to extract the temperature variations around the nostril regions [10].
Among the sleep related data, BR and HR are two critically important factors reflecting the quality of sleep. For example, nocturnal paroxysmal dystonia, one of the sleep disorders, will occur with or be preceded by the breathing pause and slowed HR [5]. BR is defined as the number of breaths a person takes per minute. Some literature have reported to use the optical based approaches for measuring the BR from a long distance. Nam et al., applied the built-in camera on smartphone to film the motions of chest and abdomen, and the BR was then derived with the average median error below 2% [11]. In contrast to the displacement change, temperature variation is more significant and therefore more suitable for estimating breathing signature. Basu et al., used the infrared thermography to monitor such variation for the subsequent computation of BR [12]. HR is the speed of the heartbeat which is measured by the number of heart beats per minute. Under a certain illumination, subtle color changes caused by blood circulation allow us to detect HR through the camera [13]. Yan et al., took advantage of this point to measure HR from facial RGB video illuminated by natural light [14]. Although some investigators developed the filters for eliminating the effects of illumination variations, this is still a highly illumination-dependent solution for HR measurement. For measuring these bio-signals during night, the systems combining cameras in different wavelengths can provide more accurate and robust results than those using the single sensor.
Very limited attempts have been undertaken to implement multi-camera systems for measurements of BR and HR. Gupta et al., established a synchronized multimodal imaging system containing an RGB camera, a monochrome camera and a thermal camera for estimating the HR [15]. The results of their experiments demonstrated that the imaging system inclusive of more spectral channels could achieve more accurate and robust measurement. Procházka et al., used the RGB, depth and infrared sensors to concurrently measure BR and HR under the natural environment [13]. A dual-wavelength imaging system operating at 611 nm and 880 nm was applied for blood oxygen saturation estimation [16]. The feasibility of this system in BR and HR measurements requires further studies. In conclusion, due to their properties, the above systems may be unsuitable for measurements of vital physiological signals during night.
Therefore, considering the environmental characteristics of night, we leverage an infrared camera equipped with IR-cut lens and a far-infrared imager to develop a dual-camera imaging system, thus allowing to record the infrared and thermal videos. Furthermore, an infrared light array associated with a photosensitive sensor is adopted to switch the system to the night mode, in which the system can simultaneously obtain infrared and thermal image sequences.
The main contributions of this paper are to present: (1) a dual-camera setup that can carry out the simultaneous and unobtrusive BR and HR measurements during nighttime; (2) a collaborative image processing method to detect and track the region of interest (ROI) as well as register images; and (3) a novel signal extraction method in time domain to accurately determine BR and HR.
Methods
Conflict of interests and ethics statement
This study was approved by the institutional review board (IRB) of the Shanghai Jiao Tong University and all participants provided full written informed consent. The individual in this manuscript has given written informed consent (as outlined in PLOS consent form) to publish these case details.
Principle
The disadvantage compared to the infrared image is that, due to few geometric and textural facial details, the thermal image is at present inadequate to design fast and reliable face detection algorithms [17]. Therefore, in the current work, the infrared image is used to assist in automatically recognizing face and facial tissue in thermal image. For BR, the principle of thermal imager is based on the fact that the temperature around the nose and mouth fluctuates throughout the inspiration and expiration cycle. In terms of HR, we can catch the subtle color variations caused by blood circulation from thermal video. However, these two targeted signals are mixed together in the raw extracted signal. Luckily, the common BR and HR are between 10–40 bpm and 60–100 bpm, respectively [18, 19]. As shown in Fig 1, there is no overlap for these two vital physiological indicators in the Fourier power spectrum, thus enabling the BR and HR to be measured simultaneously.
A dual-camera imaging system
Fig 2demonstrates the experimental setup. A thermal imager (MAG62, Magnity Electronics Co.Ltd., Shanghai, PR China) with the resolution of 640×480 and the pixel pitch of 17 μm is fixed on a tripod to prevent vibration during the experiments. Its spectral range and thermal sensitivity are 7.5–14 μm and 0.5°C, respectively. An infrared camera with the resolution of 640×480 is stabilized on the top of the thermal imager. These two cameras are parallel to each other in such a way that the field of view is almost the same. A USB and patch cables are applied for infrared and thermal cameras to connect a computer, respectively. A custom-built image acquisition software is developed for the generation of two trigger signals, thereby allowing the simultaneous acquisition of thermal and infrared videos. The obtained videos are afterwards loaded into the Matlab R2014a (The Mathworks, Inc., Natick, MA, USA) for further analysis.
In order to record videos at night using the dual-camera system, a lens with IR-Cut filter is attached to the infrared camera, and this filter is triggered by sensor on the circuit board. In addition, an infrared lighting array having 36 infrared lamps is arranged below the imaging unit. The infrared lamps are arrayed in 6×6 matrix to realize the uniform illumination at night. A photosensitive sensor is installed in the middle of the infrared lighting array to control the working states of infrared lamps. The workflow of this dual-camera imaging system is elaborated in Fig 3to obtain the near-infrared and thermal videos during nighttime.
A collaborative image processing method
During the nighttime, the infrared lamps begin working (Fig 3). For the purpose of accurately measuring the BR and HR under the no or low lighting situations, we exploit the respective advantages of infrared and thermal modalities. The thermal image is first used as reference to register infrared image. Subsequently, we conduct cross-modality data processing to detect and track the ROI. The raw signal containing the wanted BR and HR data is then extracted. The following sub-sections will describe these in detail.
Image registration
The affine transformation [20] is applied for registering thermal and infrared images. The control point pairs, viz., the moving points in the infrared image and the fixed points in the thermal image, are defined by manually choosing the strongly correlated points from the first frame of bimodal videos. The Eq (1) is used to do this
(1) |
where Ivis is the original infrared image and its corresponding transformed infrared image is Ivis_r; T represents the transformation matrix; s, θ and b denote the scaling, rotation and translation vectors, respectively.
If the system keeps fixed during the experiment, there is no need to register it again.
Cross-modality ROI detection and tracking
The first infrared image from the registered infrared video is imported into the pre-trained Boosted Cascade classifier [21] for locating the face region. When the pre-trained classifier fail to seek out the face, we exert the fully convolutional network [22], one of the classification nets for deep learning, to accomplish this task (Fig 4). To train the deep classification network, a total of 16,000 positive images (face) and the same number of negative images (background) are randomly selected from the PASCAL dataset [23] as the inputs of network. The detailed architecture of fully convolutional network used in this work can be found in Fig 4. By this means, almost any faces with different postures (e.g. side face) can be found in the infrared images.
After the face detection, a total of 66 landmark points are acquired by the discriminative regression based approach [24]. As shown in Fig 5, the 29th, the 49th and the 55th landmark points are selected to generate ROI in such a way that ROI can cover both the nose and mouth regions, thereby allowing us to measure BR no matter whether the subjects breathe using the nose and mouth simultaneously, alternately or individually. For HR measurement, the region between the nose and mouth contains the strongest heartbeat information. The ROI is calculated from these three points according to
(2) |
where is the ROI formation vector containing the horizontal ordinate (X), vertical coordinate (Y), height (H), width (W) and inclination angle (θ). The (x29,y29), (x49,y49) and (x55,y55) are the coordinates for the 29th, the 49th and the 55th landmarks, respectively.
Once the ROI region has been confirmed, the linear coordinate mapping is used to determine the corresponding region in thermal image. Subsequently, the off-the-shelf tracking algorithm called spatio-temporal context learning (STC) [25] is employed to track the ROI in the thermal image sequences (Fig 5). The average pixel intensity of ROI in thermal image is computed using the equation below
(3) |
In equation, s(i,j,k) is the pixel intensity of thermal image at pixel (i, j) and video frame k; N is the vector of pixel coordinates in ROI and n is its number.
A signal extraction method in time domain
Based on Fig 1, we can find that BR and HR signals are located on the low and high frequency areas of Fourier spectrum, respectively. Also, there exist some noises from the other factors such as the movement and illumination variation in the original signals. To accurately and robustly estimate BR and HR, we develop a time-domain signal extraction method (Fig 6). One sample are chosen from our database to clearly describe this method.
With respect to BR measurement, the high frequency parts mainly caused by the heartbeat and movement are first removed by moving average filter with window size of 20. The other curve is formed by phase shift, and a set of cross points yield. In our preliminary experiments, we observe that there is a hysteresis in the obtained signals when changing the window size of the sliding average window. This hysteresis is also called phase shift in this study (Fig 6). The larger window size, the more historical information in each processed data point. The hysteresis of signal using large window size is obviously larger than that using small window size. This generates the cross points between two processed signals. To eliminate the cross points brought by the overlap peaks of these two curves, the moving average filter with the larger window size viz., 50 is applied for the phase shifted curve. Subsequently, BR can be finally determined from the number of cross points (Fig 6, left column).
For HR measurement, the breathing-related data in original signal is first wiped off by subtracting the filtered signal with window size of 20. For the reason that the most noise events are belonged to high frequency parts, the two cascade moving average filters with adaptive window sizes are adopted to extract the relatively purified high frequencies. These adaptive window sizes vary from person to person. The differential signal is then derived with data interval of 3 to avoid the local extreme values. HR can be finally calculated from the cross points between the differential signal and line of gradient equal to 1 (Fig 6, right column).
Compared to the previous methods such as the Matlab function called findpeaks and fast Fourier transform, our approach can give more accurate and robust results (results not shown). The results shown in next section demonstrate that this approach is useful for BR and HR estimations from the image-derived signal.
A dual-mode sleep video database
A database of thermal and near-infrared dual-mode sleep videos is constructed to quantitatively verify the proposed dual-mode imaging system for BR and HR measurements under sleep situations. A total of 12 participants including different genders (two females and ten males) and ages (from 21 to 38 years old) consented to be subjects and were instructed to sleep with some body movements in multimedia lab of Shanghai Jiao Tong University. These participants were randomly recruited from Shanghai Jiao Tong University, School of Electronic Information and Electrical Engineering. Before the experiment, we asked the participant to close her/his eyes and rest in bed for ten minutes, and one dual-mode video was recorded while the participant was resting. Some volunteers were asked to be tested several times, and this database therefore had 28 video pairs crossing two domains. The testing distances were set from 1 to 3 meter (Fig 2). To simulate the sleep situation, the illumination intensity of the testing room was controlled within 3 Lux, which was monitored by the digital light meter (TA8131, TASI ELECTRONIC CO.,LTD., Jiangsu, PR China). During the video collection, the reference BR and HR were recorded by the dedicated and qualified human observers and ECG Monitor (PC-80B, Shenzhen Creative Industry Co.,Ltd., Guangzhou, PR China), respectively. The effectiveness of our approach was checked by the use of Bland-Altman plot [26, 27] and linear correlation analysis.
Results and discussion
To test the performance of BR and HR measurements using bimodal imaging system in tandem with proposed algorithms in unobtrusive and contactless manner, two statistical analysis methods namely linear correlation analysis and Bland-Altman plot were applied for validating the data from small-scale pilot experiment.
BR measurement
The scatter plot and linear regression result of reference and measured BR are demonstrated in Fig 7A. As shown in Fig 7A, all scatter points located between the 95% upper and lower confidence intervals, and most of them were close to the line of perfect match, whose slope was equal to 1. From linear correlation analysis, the estimated BR was found to be relevant to the simultaneously-acquired reference BR with the determination coefficient (R2) of 0.831. The above results indicated that our approach was acceptable for BR measurement.
The corresponding Bland-Altman plot is shown in Fig 7B with the mean of difference of -0.635 bpm and limits of agreement of -3.134 and 1.865 bpm. The majority of scatter points were within the 95% limits of agreement, indicating that the measured BR was in good agreement with the reference BR. There was one black point out of upper limits of agreement with the offset approximately 0.635 bpm. The reason for this might be that the testing subject carried out one time very significant head motion during the experiment after we reviewed the original video. Several points were dispersed around the 95% limits of agreement, due in part to the highly low illumination (0~3 Lux) under simulated sleep situation. In conclusion, the result of Bland-Altman plot also revealed the capability of bimodal imaging system coupled with proposed algorithms for BR estimation during the night.
HR measurement
In respect to HR measurement, the scatter plot and regression line are presented in Fig 8A. Apart from the one black point, the other testing data fell on the region between the two 95% confidence intervals. The strong correlation viz., R2 = 0.933 was observed between the measured and reference HRs, suggesting that the overall performance of the developed system is good for HR estimation. Although our system was to great extent robust against the low illumination, some unknown factors from experimental setup and environment would still influence the testing results. For example, by checking the original video, we cannot figure out the reason causing the black point out of the 95% confidence interval.
From the Bland-Altman plot shown in Fig 8B, in addition to one black point, all testing points located within the 95% upper and lower limits of agreement, which were 4.293 and -5.293 bpm, respectively. The distribution of points in Fig 8B was very similar with that of scatter points in Fig 8A. Consequently, the results of Bland-Altman analysis further supported these obtained by linear regression analysis.
Discussion
Compared to the physiological signal measurement tools in the contact way during sleep [28,29], the contactless and unobtrusive optical approaches require more complex algorithms to analyze the obtained data. This is because the complicated imaging environments will increase the noise level of extracted signal and even deform the waveform. The most of published literature [30,31] used the depth camera to record the BR during sleep via the monitoring of the movement. However, the motion-based techniques such as active sonar [32] are sensitive to the significant body movement. Some investigators applied the RGB camera to estimate the BR and HR under the natural lighting conditions [14,33–35], but these system would not work when the subject slept during night. Hence, the state-of-art multi-camera system was developed for simultaneously measuring the BR and HR [15]. Despite the introduction of thermal camera, this system was infeasible for BR and HR measurement under the no or low lighting situations due to the 8×8 resolution of the thermal imagery. In contrast, we established the compact and effective dual-camera imaging system by adding the infrared camera equipped with IR-Cut filter and infrared lighting array (Fig 2), thus being able to measure BR and HR during nighttime.
Conclusion
A dual-mode imaging system operating on short-wave infrared and long-wave infrared wavelengths associated with the proposed algorithm can synchronously measure BR and HR during nighttime. Using the data from the small pilot study under simulated sleep situation, the measured BR had relatively good relationship with the reference BR with determination coefficient of 0.831 via linear regression analysis. In terms of HR measurement, the measured HR was found to be strongly relevant to the reference HR with the determination coefficient of 0.933. The Bland-Altman plots of BR and HR exhibited the good performance of the dual-camera imaging system with limits of agreement of -3.134 and 1.865 bpm and -5.293 and 4.293 bpm, respectively. In general, the overall performance of the proposed technique is acceptable for BR and HR estimations during nighttime.
Data Availability
All relevant data are available from the Figshare repository (https://figshare.com/articles/sleep_dataset_zip/5518996).
Funding Statement
This work was supported by the China Postdoctoral Science Foundation funded project (2016M600315), and in part by the National Science Foundation of China under Grants 61422112, 61371146, 61221001.
References
- 1.Bem RD, Bem RD, Bem RD, Miranda D. Non-invasive sleep-environment monitoring system. In the 8th ACM International Conference on PErvasive Technologies Related to Assistive Environments; 2015: pp. 75.
- 2.Papakostas M, Staud, J, Makedon F, Metsis V. Monitoring breathing activity and sleep patterns using multimodal non-invasive technologies. In the 8th ACM International Conference on PErvasive Technologies Related to Assistive Environments; 2015: pp. 78.
- 3.Collins J, Meng C, Eng A. Psychological Impact of Severe Obesity. Curr. Obes. Rep. 2016; 5: 435–440. doi: 10.1007/s13679-016-0229-4 [DOI] [PubMed] [Google Scholar]
- 4.Yamashita M, Yamamoto T. Excessive brain tryptophan and kynurenic acid lead to abnormal social behavior in central fatigue induced by chronic sleep disorder. Int. J. Psychophysiol. 2014; 94: 229–230. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Stuck BA, Weeß HG. The new “International Classification of Sleep Disorders”. Somnologie—Schlafforschung und Schlafmedizin 2015; 19: 126–132. [Google Scholar]
- 6.Lee J, Hong M, Ryu S. Sleep monitoring system using kinect sensor. Int. J. Distrib. Sens. N. 2015; 2015: 1–9. [Google Scholar]
- 7.Centonze F, Schatz M, Prochazka A, Kuchynka J. Feature extraction using MS Kinect and data fusion in analysis of sleep disorders. In 2015 International Workshop on Computational Intelligence for Multimedia Understanding; 2015: pp: 1–5.
- 8.Heinrich A, Heesch F. V, Puvvula B, Rocque M. Video based actigraphy and breathing monitoring from the bedside table of shared beds. J. Amb.Intel. Hum. Comp. 2016; 6: 107–120. [Google Scholar]
- 9.Martinez M, Stiefelhagen R. Breath rate monitoring during sleep using near-ir imagery and pca. In 1st International Conference on Pattern Recognition; 2012; pp: 3472–3475.
- 10.Fei J, Pavlidis I, Murthy J. Thermal Vision for Sleep Apnea Monitoring. In Miccai International Conference on Medical Image Computing & Computer-assisted Intervention; 2009; pp: 1084–1091. [DOI] [PubMed]
- 11.Nam Y, Kong Y, Reyes B, Reljin N, Chon KH. Monitoring of Heart and Breathing Rates Using Dual Cameras on a Smartphone. Plos One; 2016; 11: e0151013 doi: 10.1371/journal.pone.0151013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Basu A, Routray A, Mukherjee R, Shit S. Infrared Imaging based Hyperventilation Monitoring through Respiration Rate Estimation. Infrared Phys. Techn. 2016; 77: pp: 382–390. [Google Scholar]
- 13.Procházka A, Schätz M, Vyšata O, Vališ M. Microsoft Kinect Visual and Depth Sensors for Breathing and Heart Rate Analysis. Sensors; 2016; 16: 996. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Yan Y, Ma X, Yao L, Ouyang J. Noncontact measurement of heart rate using facial video illuminated under natural light and signal weighted analysis. Bio-med. Mater. Eng. 2015; 26: S903. [DOI] [PubMed] [Google Scholar]
- 15.Gupta O, McDuff D, Raskar R. Real-Time Physiological Measurement and Visualization Using a Synchronized Multi-Camera System. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops; 2016: pp: 46–53.
- 16.Shao D, Liu C, Tsow F, Yang Y, Du Z, Iriya R., et al. Noncontact Monitoring of Blood Oxygen Saturation Using Camera and Dual-Wavelength Imaging System. IEEE T. Bio-med. Eng. 2016; 63: 1091–1098. [DOI] [PubMed] [Google Scholar]
- 17.Sarfraz MS, Stiefelhagen R. Deep Perceptual Mapping for Thermal to Visible Face Recognition. 2015; arXiv preprint arXiv:1507.0287.
- 18.Jin F, Pavlidis I. Virtual Thermistor. In International Conference of the IEEE Engineering in Medicine and Biology Society; 2007: pp: 250–253. [DOI] [PubMed]
- 19.Telles S, Reddy SK, Nagendra HR. Oxygen consumption and respiration following two yoga relaxation techniques. Applied Psychophysiology and Biofeedback; 2000; 25: 221 [DOI] [PubMed] [Google Scholar]
- 20.Lamare F, Cresson T, Savean J, Cheze LRC, Reader AJ, Visvikis D. Respiratory motion correction for PET oncology applications using affine transformation of list mode data. Phys. Med. Biol. 2007: 52: 121 doi: 10.1088/0031-9155/52/1/009 [DOI] [PubMed] [Google Scholar]
- 21.Viola P, Jones M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Pattern Recognition; 2003: pp: 511.
- 22.Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2015: 3431–3440. [DOI] [PubMed]
- 23.Farfade SS, Saberian MJ, Li LJ. Multi-view face detection using deep convolutional neural networks. In Proceedings of the 5th ACM International Conference on Multimedia Retrieval; 2015: 643–650.
- 24.Asthana A, Zafeiriou S, Cheng S, Pantic M. Robust Discriminative Response Map Fitting with Constrained Local Models. In IEEE Conference on Computer Vision and Pattern Recognition; 2013: pp: 3444–3451.
- 25.Zhang K, Zhang L, Liu Q, Zhang D, Yang MH. Fast visual tracking via dense spatio-temporal context learning. In European Conference on Computer Vision; 2014; pp: 127–141.
- 26.Bland JM, Altman DG. Measuring agreement in method comparison studies. Stat. Methods Med. Res. 1999; 8: 135–160. doi: 10.1177/096228029900800204 [DOI] [PubMed] [Google Scholar]
- 27.Carkeet A. Exact Parametric Confidence Intervals for Bland-Altman Limits of Agreement. Optom. Vision Sci. 2015; 92: 71–80. [DOI] [PubMed] [Google Scholar]
- 28.Alshaer H. New Technologies for the Diagnosis of Sleep Apnea. Current hypertension reviews; 2016; 12: 48–56. [DOI] [PubMed] [Google Scholar]
- 29.Werth J, Atallah L, Andriessen P, Long X, Zwartkruis-Pelgrim E, Aarts R. M. Unobtrusive sleep state measurements in preterm infants—A review. Sleep Med. Rev. 2016; 32: 109–122. doi: 10.1016/j.smrv.2016.03.005 [DOI] [PubMed] [Google Scholar]
- 30.Alabbasi H, Gradinaru A, Moldoveanu F, Moldoveanu A. Human motion tracking & evaluation using Kinect V2 sensor. In E-Health and Bioengineering Conference, 2015; 2015: 1–4
- 31.Procházka A, Schätz M, Centonze F, Kuchyňka J, Vyšata O, Vališ M. Extraction of breathing features using MS Kinect for sleep stage detection," Signal, Image and Video P. 2016; 10: 1279–1286. [Google Scholar]
- 32.Nandakumar R, Gollakota S, Watson N. Contactless Sleep Apnea Detection on Smartphones. In Proceedings of the 13th Annual International Conference on Mobile Systems, Applications, and Services, 2015: pp: 45–57.
- 33.Reyes BA, Reljin N, Kong Y, Nam Y, Ha S, Chon KH. Towards the Development of a Mobile Phonopneumogram: Automatic Breath-Phase Classification Using Smartphones. Ann. Biomed Eng 2016: 44: 1–14. doi: 10.1007/s10439-015-1519-9 [DOI] [PubMed] [Google Scholar]
- 34.Tulyakov S, Alamedapineda X, Ricci E, Yin L, Cohn JF, Sebe N. Self-adaptive matrix completion for heart rate estimation from face videos under realistic conditions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2016; pp: 2396–2404.
- 35.Antink CH, Gao H, Brüser C, Leonhardt S. Beat-to-beat heart rate estimation fusing multimodal video and sensor data. Biomed. Opt. Express 2015; 6: 2895–2907. doi: 10.1364/BOE.6.002895 [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
All relevant data are available from the Figshare repository (https://figshare.com/articles/sleep_dataset_zip/5518996).