Abstract
Blood pressure monitoring is an essential component of hypertension management and in the prediction of associated comorbidities. Blood pressure is a dynamic vital sign with frequent changes throughout a given day. Capturing blood pressure remotely and frequently (also known as ambulatory blood pressure monitoring) has traditionally been achieved by measuring blood pressure at discrete intervals using an inflatable cuff. However, there is growing interest in developing a cuffless ambulatory blood pressure monitoring system to measure blood pressure continuously. One such approach is by utilizing bioimpedance sensors to build regression models. A practical problem with this approach is that the amount of data required to confidently train such a regression model can be prohibitive. In this paper, we propose the application of the domain-adversarial training neural network (DANN) method on our multitask learning (MTL) blood pressure estimation model, allowing for knowledge transfer between subjects. Our proposed model obtains average root mean square error (RMSE) of 4.80 ± 0.74 mmHg for diastolic blood pressure and 7.34 ± 1.88 mmHg for systolic blood pressure when using three minutes of training data, 4.64 ± 0.60 mmHg and 7.10 ± 1.79 respectively when using four minutes of training data, and 4.48±0.57 mmHg and 6.79±1.70 respectively when using five minutes of training data. DANN improves training with minimal data in comparison to both directly training and to training with a pretrained model from another subject, decreasing RMSE by 0.19 to 0.26 mmHg (diastolic) and by 0.46 to 0.67 mmHg (systolic) in comparison to the best baseline models. We observe that four minutes of training data is the minimum requirement for our framework to exceed ISO standards within this cohort of patients.
1. Introduction
Hypertension is a worldwide chronic disease that causes an estimated 7.6 million deaths every year. The diagnosis of hypertension is usually based on clinical blood pressure readings, but the measurement of blood pressure outside of a clinical visit (also known as ambulatory blood pressure measurement) can provide better prognostic guidance than measurements during a routine clinic visit (Olsen et al., 2016), due to well-known confounders such as masked hypertension (O’Brien et al., 2005), white coat hypertension (Pickering et al., 1988), and nocturnal non-dipping hypertension (Pickering et al., 1982). Ambulatory blood pressure monitoring has been shown to be more predictive of cardiovascular mortality than clinical monitoring in a study of 63,910 adults (Banegas et al., 2018), and nocturnal measurements are likely stronger predictors of cardiovascular risk than diurnal monitoring (Stergiou et al., 2018a; Whelton et al., 2018; Fan et al., 2010). Therefore, increased ambulatory measuring is desirable for public health. However, on-market ambulatory monitoring devices are not appropriate for extensive use for a number of reasons: they require specific patient postures, they are obtrusive, they disrupt sleep, and they result in poor adherence. Cuffless blood pressure monitoring devices are desirable for their possibility to overcome each of those shortcomings. Cuffless blood pressure estimation techniques utilize devices to monitor surrogates of blood pressure, and use these surrogates to build regression models to estimate diastolic and systolic blood pressure.
There are a variety of techniques that have recently been investigated for their potential application to cuffless blood pressure estimation. Chief among those techniques include photoplethysmography (PPG) in conjunction with electrocardiography (ECG) (Thomas et al., 2016), dual PPGs (Nabeel et al., 2018; Wang et al., 2018), Doppler radar technology (Shay and Dai, 2017), or bioimpedance (Ibrahim et al., 2018). Each of these techniques attempt to measure the pulse transit time (PTT) or pulse wave velocity (PWV), both of which are known surrogates for blood pressure (Luo et al., 2016; Ibrahim and Jafari, 2019; Zheng et al., 2014). Ibrahim and Jafari (2019) developed a bioimpedance-based sensor that locates arterial sites to measure these physiologic surrogates of blood pressure. Ibrahim and Jafari then used a window-based AdaBoost regression technique to measure personal diastolic and systolic blood pressure over windows of 10 consecutive beats to with respective errors of 2.6 mmHg and 3.4 mmHg. This finding falls within the ISO standard requiring errors less than 10 mmHg when comparing with a gold standard device (Stergiou et al., 2018b) for the particular cohort. We first develop a deep multitask learning (MTL) regression model using a version of the same dataset produced by Ibrahim and Jafari, but with an additional user. This model allows for more adaptable transfer learning than an AdaBoost regression model, and focuses on a beat-to-beat blood pressure estimation task as a new baseline.
One problem in training this deep neural network is that it requires a great amount of training data. Previous work on this dataset uses 80% of all available data, over 10 minutes on average, from each subject to train the personal models (Ibrahim and Jafari, 2019). However, this calibration period is burdensome and the goal of an independent device should be to minimize the amount of calibration time required to improve utility and align with clinical need. To that end, in this work we investigate techniques to reduce the amount of data involved in training a model. Directly training an MTL model on reduced training data fails with errors exceeding ISO standards. Therefore, to meet the ISO standards1 (in this cohort) while minimizing training data, we must utilize a technique to learn from other subjects. A trivial way to accomplish this task would appear to be to build a generalized model from many other subjects. However, due to inter-subject variability, generalized models we trained in leave-one-subject-out approaches produce errors greater than 10 mmHg, beyond the limits of the ISO standard. Transfer learning from a pretrained model is another solution, but the difference between subjects still impedes the learning process. Domain adaption (Duan et al., 2012; Blitzer et al.; Cook et al., 2013) is one solution to cross-domain problems, and has recently been applied with deep learning techniques (Long et al.; Tzeng et al., 2014) to minimize the maximum mean discrepancy distance between disparate outputs. Domain-adversarial neural networks (DANN) (Ganin et al., 2016) allows for using adversarial training to extract domain-invariant features, allowing for rapid model adaptation with minimal training data.
In this paper, we propose a DANN-based MTL model to estimate beat-to-beat blood pressure for the goal of maintaining accuracy to within ISO standards while minimizing the amount of required training data (Ganin et al., 2016). To maximize clinical utility, we aim to train this model with a maximum of five minutes of training data for a new user. Our base model, an MTL blood pressure (BP) estimation model, is composed of a long short-term memory (LSTM) coupled to a shared dense layer to extract heart beat features, and then two task-specific networks, one each for estimating diastolic and systolic blood pressure. When applying DANN, a domain (subject) classifier then attempts to classify a given beat as belonging to a particular subject. The adversarial training approach is then applied to this system with the goal of maximizing the performance of the BP estimator while minimizing the performance of the domain classifier. Throughout this process the BP estimator is trained with reduced data from the new subject until convergence is achieved.
Generalizable Insights about Machine Learning in the Context of Healthcare
Transfer learning can fail when individuals are very different. Differences in subjects (domains) make knowledge transfer or development of a generalized model infeasible as underlying physiologies are sufficiently different that no centralized representation is easily found nor an obvious transfer mechanism is apparent.
DANN allows for an adversarial approach to force a model to learn subject-invariant features. This approach allows for rapid model personalization with minimal new user data.
2. Related Work
2.1. Cuffless Blood Pressure
Utilizing ECG and PPG signals to predict blood pressure is an active area of research (Kachuee et al., 2016; Zheng et al., 2014). Similarly, some work has been done in using dual PPG systems for this predictive task (Nabeel et al., 2018; Wang et al., 2018). However, all of these methods suffer from errors introduced by variable pre-ejection periods which are highly influenced by stress, emotion, physical exertion, and age (Peter et al., 2014).
Another approach to cuffless blood pressure estimation, proposed by Ibrahim and Jafari (2019), involved the use of a wrist-worn array of bioimpedance signals consisting of 4 pairs of sensors on the ulnar and radial arteries. Variations in bioimpedance along these arteries are correlated with PTT, the time taken for the pressure pulse to travel between two points. PTT is one of the most prominent markers used for estimation of BP. By directly evaluating measurements along the wrist, this method does not suffer from timing errors as a result of the pre-ejection period. Ibrahim and Jafari (2019) extracted a total of 50 features from these four bioimpedance curves using four characteristic points representing diastolic peak, maximum slope, systolic foot, and inflection point before building separate AdaBoost regression models for each subject. The calculated features and reference labels were averaged over 10-beat windows with 50% overlap to reduce the effect of beat-to-beat variability. These models were trained using 80% of all available data. Their results show an average correlation coefficient (R) and RMSE of 0.77 and 2.6 mmHg for the diastolic blood pressure and 0.86 and 3.4 mmHg for the systolic blood pressure.
2.2. Cross-Domain Model Generalization
When individual systems in a problem have unique variations, it can become difficult to generalize a single model to provide good results across multiple individuals. Differences between domains may result from different root causes. For instance, different data collection methods (Getmantsev et al., 2018), different equipment settings (Yuan et al.), different learning tasks (Snell et al.), or unusual frequency of observed events (Kouw and Loog, 2018), can each lead to systems that superficially appear similar, and yet are unable to easily be generalized by a single model. Domain adaption techniques are one traditional solution applied for this type of cross-domain problem (Duan et al., 2012; Blitzer et al.; Cook et al., 2013). Another solution involves knowledge transfer, which transfers models between domains by reweighing probability distributions among instances (Chattopadhyay et al., 2012; Tan et al.), clustering trained source domains (Yao and Doretto; Zhao et al.), or feature transformation (Gong et al.; Wang et al., a). Many recent deep learning-based domain adaption techniques minimize the maximum mean discrepancy distance between the outputs (Long et al.; Tzeng et al., 2014). However, these techniques focus on learning explicitly the difference or transformation between domains, and do not focus on learning a domain-agnostic representation.
To address the cross-domain problem in inter-subject datasets, it is beneficial to learn useful knowledge representations while discarding subject-specific features that may hinder performance of the overall task. Domain-adversarial neural network (DANN) (Ganin et al., 2016), inspired by domain adaptation theory, uses an adversarial training approach to extract domain-invariant features. DANN has three parts: a feature extractor, a label predictor, and an additional domain classifier. A gradient reversal layer is applied to the feature extractor, targeting the loss from domain classification in order to adversarially train the feature extractor and domain classifier. In this way, the feature captured from source and target domains can be represented in a domain-invariant manner. DANN is widely applied in natural language processing and computer vision (Liu et al.; Taigman et al., 2016; Conneau et al., 2017), solving the cross-domain transfer learning problem. Anand and Kanhangad (2019) proposed a pore detection method from cross-sensor fingerprint images. Getmantsev et al. (2018) applied DANN in order to transfer knowledge of subject morbidity statistics from the UK Biobank into a smartphone-based HK dataset with drastically different demographic and life-style distributions, building a novel health risk model based on intraday physical activities. Faridee et al. and Ketykό and Kovács (2019) applied DANN on human activity and gesture recognition in order to account for unlabeled data and Wang et al. (b) extended the domain adaption algorithm to select disparate sources to use in activity recognition.
3. Adversarial Learning for Blood Pressure Estimation with Reduced Training Data
We describe our dataset and its preprocessing in Section 3.1. We design a baseline model for estimating blood pressure using 80% of the data to train, detailed in Section 3.2. Reducing the size of this dataset is an important clinical challenge. We introduce DANN in Section 3.3 to accomplish this reduction. This model is adversarial (Ganin et al., 2016), using a minmax optimization between the domain classifier and feature extractor: training the domain classifier for higher accuracy while minimizing the feature extractor to have low domain classification accuracy results in blinding the final model domains, forcing the model to rely on user-invariant features.
3.1. Dataset and Data Preprocessing
The BP dataset was collected using a wrist bioimpedance sensor (Ibrahim and Jafari, 2019) on 11 subjects. The sensor uses four channels to measure the impedance of skin surface moments at the ulnar and radial arteries. Each subject performed a variety of physical activities in order to achieve a range of blood pressures, rising and falling with exercise and rest. A Finapres NOVA device was used at the same time to capture beat-to-beat diastolic and systolic blood pressure as ground truth reference measurements. The signals were segmented by heartbeat and samples corrupted by motion artifacts or non-physically realistic values were removed. We downsampled the signals from the original 20 kHz to 100 Hz by equally sampling and applied zero padding to the beginning of each sequence. The first derivative of the four signal channels and the timing of each point are augmented as additional features, resulting in 9 input features in total. During training, all training features and labels are normalized between 0 and 1. Test features and labels were then scaled by factors learned from training normalization.
3.2. MTL BP Estimation Model
The base model consists of an LSTM layer, a shared dense layer, and two task-specific networks. The heartbeat data (and associated derived channels) are sent to an LSTM layer and then on to a shared dense layer. LSTM can memorize historical information, and therefore is applied to capture the patterns in the signal over time. The ability of an LSTM to retain historical information is valuable across the entirety of the heartbeat. We include a dropout layer following the LSTM. This layer allows the model to avoid overfitting and permits for some robustness to noise. Even if a part of the signal is corrupted, the model will still be able to perform with reasonable accuracy. We add a shared layer after the LSTM to further extract the relational information between channels. The extracted features are then passed on to the BP estimation network, consisting of two separate task-specific networks to estimate diastolic and systolic blood pressures. After each layer in these two task-specific networks, a dropout layer is applied to avoid overfitting. In order to build models for new subjects with reduced data, we further propose using DANN to transfer knowledge from other subjects and focus our attention on the beat-to-beat model as it provides higher potential clinicial utility.
3.3. Adversarial Training with Minimal Data
With enough data from a subject we are able to build a blood pressure regression model for that subject to within ISO standards. However, it is desirable to improve upon this and discover the minimal amount of training data that can provide for blood pressure while remaining within the ISO standard. For a device to be implemented in clinical practice, it should be widely adaptable to a variety of patients with minimal calibration time. Therefore, our objective here is to push the limits of training data utilized while remaining precise to the necessary standards.
When simply training with less data, the model quickly produces erroneous estimates that fall out of ISO standards after a small reduction in training data. To address this issue, we investigate transfer learning solutions to more rapidly adapt our model to a previously unknown subject. However, a chief challenge of model adaptation is that the difference in wearable sensor signal data between individuals is too large, and a single generalized model fails. Therefore, we need to learn from other subjects but discard the difference between subjects.
To build models for new subjects with reduced data, we utilize DANN to extract user-invariant features for the purpose of knowledge transfer. Figure 1 shows the implementation of DANN within our MTL model. Our DANN model has three key parts: a feature extractor, a BP estimator, and a domain classifier. The feature extractor and BP estimator are as described above: the feature extractor is an LSTM and the BP estimator is two task-specific networks. The domain classifier is described in detail below and serves as a new module that pushes learning of subject-agnostic features of the data.
Figure 1:
Adversarial training structure. There are three components: Feature extractor (blue), BP estimator (green), and Domain classifier (orange). The black solid lines represent data and arrows with dashed lines represent the Systolic and Diastolic loss, respectively, for gradient descent.
We treat each subject as an individual domain. The domain classifier is trained to maximize its accuracy in recognizing to which subject a beat belongs. The BP estimator is trained to maximize the BP regression accuracy. The feature extractor is trained using each of these losses, but the gradient is reversed for the domain classification. This gradient reversal pushes the feature extractor to be blind to subjects, causing the extracted features to be subject-invariant. This coupling of the BP regression with reversed domain classification is the key adversarial component of this model. The BP estimator, domain classifier, and feature extractor are updated using back propagation as follows:
(1) |
(2) |
(3) |
Here θ refers the parameters in a model: θBP, θd and θf indicate the parameters in the BP estimator, domain classifier, and feature extractor, respectively. LBP is the loss of the BP estimator and Ld is the cross-entropy loss from the domain classification. α is the learning rate, and λ is the loss weight, which balances the BP estimator and the domain classifier and is set to 1 in our experiments. LBP is given as
(4) |
where ED and ES represent the estimated diastolic and systolic pressures, respectively, and TS and TD are their target values. This loss function ensures that both feature regression networks are related. Using an adversarial training approach, the feature extractor is trained to be blind to the source of the samples. Using DANN, we try to discriminate the difference between subjects, and lead the feature extractor to obtain common information that is related to blood pressure among different subjects, so that the new subject can learn from other subjects with greater training data.
Initially, we use the new subject with reduced data as our target domain, and randomly choose another subject as source domain. However, in this case the domain classifier always predicts the domain to be the source. This results from the unbalanced data between source and target domains, and any actions to the domain classifier result in a decrease of temporal accuracy. Therefore, the domain classifier stays in the local minimum and can not be updated further. To solve this problem, we introduce a second training subject as the target domain which guides the network being trained toward the new subject. We select the subjects randomly because of a lack of feasible subject similarity metric, and discuss this limitation further in Section 4.1. After training DANN to have stable loss, we use the reduced training data from the new subject again to retrain the model, converting the obtained knowledge from the other two subjects to align better with the new subject. Finally, we train a model under a leave-one-subject-out scheme where all other subjects are used to train the DANN. However, this approach does not converge and no usable results are produced.
3.4. Experiments
To test the model performance with the reduced data, we initially limit training data to three minutes for each subject, using the remaining data as the test set. Three minutes was selected as a length of time that would be feasible for in-clinic calibration of the blood pressure system. We first train the model directly without any pretrained model loaded or technique applied during training, so that we can understand the performance from the limited training data. Then, in order to learn from other subjects, we load the pretrained model with 80% training data, and retrain the model with the reduced training set from the new subject. All layers of the pretrained model are retrained to adapt both the feature extraction and BP estimation functions to the new subject. For each subject, we test the pretraining approach from all the other subjects individually, and calculate the average RMSE and correlation. To evaluate the DANN model, we need two other subjects as source domain and target domain for the adversarial training approach other than the new subject. These two subjects are randomly picked from all other subjects, and we run the test 10 times for each subject as a new subject for robustness. The average RMSE and correlation are calculated as well after the 10 rounds of testing. We use the same model structure and hyperparameters in these experiments: three layers of task-specific networks with hidden size 30, learned from manually trained MTL models. This work is implemented in Python 3.6 with Tensorflow 1.15, Numpy 1.18, sklearn 0.21. The average computation time is 8.5 ± 0.5 minutes per subject without additional parallelization or fine-tuning on our server of 2 Xeon 2.2GHz CPUs, 8 GTX 1080ti GPUs, and 528 GB RAM. Code for this implementation can be found at https://github.com/stmilab/cufflessbp_dann.
The results of training with three minutes are not sufficient to reach ISO standards with this model. Therefore, we repeat these experiments with four and five minutes of training data. After analysis of training with four minutes, the DANN model performs within the ISO standards of 85% of all diastolic and systolic data points having less than 10 mmHg absolute error within this cohort.
The results of utilizing only three minutes of training data are shown in table 1, and table 2 and table 3 show results utilizing four and five minutes of training data respectively. From the results, using three minutes of training data obtains an RMSE of 4.80 ± 0.74 mmHg for diastolic blood pressure and 7.34±1.88 mmHg for systolic blood pressure. DANN improves RMSE over the pretrained model by 0.20 mmHg for diastolic blood pressure and 0.60 mmHg for systolic blood pressure. When utilizing four minutes of training data, the model obtains an RMSE of 4.64 ± 0.60 mmHg for diastolic blood pressure and 7.10 ± 1.79 mmHg for systolic blood pressure. DANN improves RMSE over the pretrained model by 0.19 mmHg for diastolic blood pressure and 0.46 mmHg for systolic blood pressure. For five minutes training data, DANN improves RMSE over the pretrained model by 0.26 mmHg to 4.48 mmHg for diastolic blood pressure, and improves by 0.67 mmHg to 6.79 mmHg for systolic blood pressure2. We also test the pretrained models on new users without any retraining process. The average RMSE for DBP is 6.94 mmHg and for SBP is 11.51 mmHg, and the average correlation for DBP is 0.07 and for SBP is −0.01.
Table 1:
Results using three minutes of subject-specific training data for diastolic and systolic blood pressure (DBP & SBP)
Subject | DANN | Pretrained | Directly Trained | ||||
---|---|---|---|---|---|---|---|
RMSE | R | RMSE | R | RMSE | R | ||
1 | DBP: | 4.56 ± 0.07 | 0.43 ± 0.05 | 4.93 ± 0.14 | 0.33 ± 0.07 | 4.93 | 0.16 |
SBP: | 5.98 ± 0.06 | 0.25 ± 0.03 | 6.19 ± 0.11 | 0.11 ± 0.05 | 12.88 | 0.00 | |
2 | DBP: | 5.39 ± 0.12 | 0.57 ± 0.03 | 5.72 ± 0.14 | 0.47 ± 0.04 | 6.44 | 0.00 |
SBP: | 8.45 ± 0.20 | 0.65 ± 0.02 | 9.24 ± 0.30 | 0.55 ± 0.04 | 12.91 | 0.02 | |
3 | DBP: | 4.08 ± 0.11 | 0.40 ± 0.02 | 4.22 ± 0.11 | 0.23 ± 0.11 | 13.65 | 0.00 |
SBP: | 6.06 ± 0.14 | 0.50 ± 0.03 | 6.81 ± 0.19 | 0.36 ± 0.10 | 7.41 | 0.00 | |
4 | DBP: | 4.21 ± 0.05 | 0.07 ± 0.05 | 4.29 ± 0.16 | 0.02 ± 0.04 | 4.12 | 0.05 |
SBP: | 7.63 ± 0.03 | 0.18 ± 0.03 | 8.11 ± 0.21 | 0.16 ± 0.07 | 17.26 | 0.00 | |
5 | DBP: | 5.15 ± 0.07 | 0.22 ± 0.06 | 5.52 ± 0.23 | 0.23 ± 0.04 | 5.61 | 0.20 |
SBP: | 5.95 ± 0.12 | 0.26 ± 0.09 | 6.22 ± 0.24 | 0.28 ± 0.02 | 6.02 | 0.30 | |
6 | DBP: | 6.25 ± 0.09 | 0.29 ± 0.04 | 6.41 ± 0.18 | 0.23 ± 0.04 | 7.26 | 0.19 |
SBP: | 7.59 ± 0.13 | 0.55 ± 0.02 | 8.16 ± 0.24 | 0.46 ± 0.04 | 9.16 | 0.00 | |
7 | DBP: | 5.20 ± 0.07 | 0.29 ± 0.05 | 5.60 ± 0.14 | 0.22 ± 0.05 | 6.09 | 0.25 |
SBP: | 8.21 ± 0.06 | 0.37 ± 0.07 | 8.76 ± 0.15 | 0.33 ± 0.05 | 8.89 | 0.00 | |
8 | DBP: | 5.50 ± 0.11 | 0.27 ± 0.10 | 5.77 ± 0.13 | 0.24 ± 0.11 | 5.74 | 0.20 |
SBP: | 12.06 ± 0.24 | 0.30 ± 0.03 | 12.88 ± 0.54 | 0.30 ± 0.09 | 12.82 | 0.30 | |
9 | DBP: | 4.02 ± 0.06 | 0.34 ± 0.02 | 4.22 ± 0.10 | 0.21 ± 0.05 | 4.72 | 0.21 |
SBP: | 5.47 ± 0.06 | 0.18 ± 0.08 | 5.81 ± 0.20 | 0.07 ± 0.04 | 5.56 | 0.00 | |
10 | DBP: | 4.23 ± 0.01 | 0.12 ± 0.02 | 4.34 ± 0.08 | 0.12 ± 0.05 | 4.24 | 0.07 |
SBP: | 5.86 ± 0.02 | 0.17 ± 0.02 | 6.00 ± 0.10 | 0.12 ± 0.04 | 5.93 | 0.06 | |
11 | DBP: | 4.24 ± 0.08 | 0.51 ± 0.02 | 4.61 ± 0.09 | 0.36 ± 0.06 | 4.44 | 0.49 |
SBP: | 7.42 ± 0.18 | 0.51 ± 0.03 | 8.14 ± 0.17 | 0.33 ± 0.07 | 8.47 | 0.00 | |
Mean | DBP: | 4.80 ± 0.74 | 0.32 ± 0.15 | 5.06 ± 0.78 | 0.24 ± 0.12 | 6.11 ± 2.56 | 0.16 ± 0.14 |
SBP: | 7.34 ± 1.88 | 0.36 ± 0.17 | 7.84 ± 2.06 | 0.28 ± 0.15 | 9.75 ± 3.57 | 0.06 ± 0.12 |
Table 2:
Results using four minutes of subject-specific training data for diastolic and systolic blood pressure (DBP & SBP)
Subject | DANN | Pretrained | Directly Trained | ||||
---|---|---|---|---|---|---|---|
RMSE | R | RMSE | R | RMSE | R | ||
1 | DBP: | 4.49 ± 0.08 | 0.45 ± 0.03 | 4.81 ± 0.11 | 0.32 ± 0.08 | 5.10 | 0.37 |
SBP: | 5.92 ± 0.08 | 0.26 ± 0.05 | 6.12 ± 0.10 | 0.12 ± 0.05 | 6.20 | 0.18 | |
2 | DBP: | 5.32 ± 0.11 | 0.58 ± 0.03 | 5.60 ± 0.15 | 0.51 ± 0.04 | 5.36 | 0.57 |
SBP: | 8.19 ± 0.29 | 0.68 ± 0.03 | 9.12 ± 0.30 | 0.58 ± 0.04 | 8.90 | 0.62 | |
3 | DBP: | 3.96 ± 0.06 | 0.42 ± 0.03 | 4.16 ± 0.10 | 0.23 ± 0.12 | 4.18 | 0.36 |
SBP: | 6.03 ± 0.28 | 0.57 ± 0.05 | 6.60 ± 0.29 | 0.43 ± 0.09 | 7.30 | 0.00 | |
4 | DBP: | 4.06 ± 0.06 | 0.09 ± 0.02 | 4.07 ± 0.05 | 0.06 ± 0.06 | 4.44 | 0.05 |
SBP: | 7.68 ± 0.14 | 0.25 ± 0.05 | 7.96 ± 0.21 | 0.17 ± 0.10 | 8.26 | 0.21 | |
5 | DBP: | 5.03 ± 0.18 | 0.23 ± 0.25 | 5.01 ± 0.12 | 0.21 ± 0.04 | 5.08 | 0.28 |
SBP: | 5.77 ± 0.09 | 0.28 ± 0.03 | 5.82 ± 0.14 | 0.26 ± 0.06 | 6.20 | 0.00 | |
6 | DBP: | 5.34 ± 0.23 | 0.33 ± 0.09 | 5.74 ± 0.06 | 0.20 ± 0.04 | 5.32 | 0.30 |
SBP: | 6.30 ± 0.19 | 0.63 ± 0.03 | 7.46 ± 0.39 | 0.53 ± 0.07 | 8.47 | 0.39 | |
7 | DBP: | 5.17 ± 0.10 | 0.33 ± 0.27 | 5.24 ± 0.11 | 0.26 ± 0.08 | 5.77 | 0.29 |
SBP: | 8.12 ± 0.16 | 0.43 ± 0.04 | 8.41 ± 0.18 | 0.34 ± 0.06 | 9.01 | 0.40 | |
8 | DBP: | 5.34 ± 0.12 | 0.39 ± 0.05 | 5.50 ± 0.15 | 0.34 ± 0.08 | 5.35 | 0.38 |
SBP: | 11.62 ± 0.34 | 0.44 ± 0.06 | 12.07 ± 0.25 | 0.38 ± 0.07 | 12.14 | 0.34 | |
9 | DBP: | 3.98 ± 0.07 | 0.33 ± 0.07 | 4.18 ± 0.06 | 0.21 ± 0.06 | 4.15 | 0.28 |
SBP: | 5.47 ± 0.08 | 0.24 ± 0.04 | 5.68 ± 0.08 | 0.06 ± 0.04 | 5.68 | 0.13 | |
10 | DBP: | 4.19 ± 0.03 | 0.12 ± 0.04 | 4.25 ± 0.07 | 0.13 ± 0.03 | 4.68 | 0.07 |
SBP: | 5.83 ± 0.02 | 0.13 ± 0.02 | 5.90 ± 0.09 | 0.13 ± 0.04 | 6.43 | 0.04 | |
11 | DBP: | 4.15 ± 0.06 | 0.52 ± 0.06 | 4.54 ± 0.12 | 0.38 ± 0.08 | 4.56 | 0.38 |
SBP: | 7.25 ± 0.13 | 0.52 ± 0.02 | 7.99 ± 0.16 | 0.37 ± 0.08 | 8.07 | 0.42 | |
Mean | DBP: | 4.64 ± 0.60 | 0.34 ± 0.15 | 4.83 ± 0.62 | 0.26 ± 0.12 | 4.90 ± 0.53 | 0.31 ± 0.15 |
SBP: | 7.10 ± 1.79 | 0.40 ± 0.18 | 7.56 ± 1.90 | 0.31 ± 0.17 | 7.88 ± 1.84 | 0.25 ± 0.20 |
Table 3:
Results using five minutes of subject-specific training data for diastolic and systolic blood pressure (DBP & SBP)
Subject | DANN | Pretrained | Directly Trained | ||||
---|---|---|---|---|---|---|---|
RMSE | R | RMSE | R | RMSE | R | ||
1 | DBP: | 4.39 ± 0.07 | 0.45 ± 0.04 | 4.79 ± 0.16 | 0.34 ± 0.09 | 4.87 | 0.37 |
SBP: | 5.85 ± 0.08 | 0.38 ± 0.03 | 6.04 ± 0.07 | 0.15 ± 0.06 | 6.03 | 0.00 | |
2 | DBP: | 5.26 ± 0.07 | 0.59 ± 0.03 | 5.63 ± 0.10 | 0.51 ± 0.02 | 5.34 | 0.56 |
SBP: | 7.98 ± 0.17 | 0.68 ± 0.02 | 9.00 ± 0.15 | 0.60 ± 0.02 | 8.58 | 0.66 | |
3 | DBP: | 3.89 ± 0.06 | 0.44 ± 0.03 | 4.13 ± 0.13 | 0.23 ± 0.15 | 3.90 | 0.44 |
SBP: | 5.77 ± 0.15 | 0.58 ± 0.02 | 6.26 ± 0.37 | 0.51 ± 0.08 | 7.19 | 0.00 | |
4 | DBP: | 4.06 ± 0.04 | 0.10 ± 0.02 | 4.04 ± 0.06 | 0.03 ± 0.07 | 4.37 | 0.11 |
SBP: | 7.69 ± 0.06 | 0.29 ± 0.04 | 8.02 ± 0.18 | 0.20 ± 0.10 | 7.87 | 0.23 | |
5 | DBP: | 4.83 ± 0.29 | 0.26 ± 0.07 | 4.87 ± 0.14 | 0.18 ± 0.10 | 5.11 | 0.25 |
SBP: | 5.61 ± 0.06 | 0.27 ± 0.04 | 5.85 ± 0.18 | 0.21 ± 0.08 | 6.05 | 0.19 | |
6 | - | - | - | - | - | - | - |
- | - | - | - | - | - | - | |
7 | DBP: | 5.04 ± 0.05 | 0.36 ± 0.02 | 5.21 ± 0.08 | 0.32 ± 0.03 | 5.61 | 0.30 |
SBP: | 7.94 ± 0.06 | 0.47 ± 0.02 | 8.28 ± 0.16 | 0.41 ± 0.04 | 8.69 | 0.43 | |
8 | DBP: | 5.27 ± 0.16 | 0.40 ± 0.03 | 5.50 ± 0.21 | 0.34 ± 0.11 | 5.49 | 0.37 |
SBP: | 10.83 ± 0.39 | 0.48 ± 0.05 | 12.04 ± 0.57 | 0.39 ± 0.12 | 12.98 | 0.35 | |
9 | DBP: | 3.84 ± 0.06 | 0.34 ± 0.03 | 4.04 ± 0.10 | 0.23 ± 0.07 | 4.29 | 0.24 |
SBP: | 5.29 ± 0.04 | 0.29 ± 0.03 | 5.53 ± 0.11 | 0.09 ± 0.03 | 5.63 | 0.03 | |
10 | DBP: | 4.20 ± 0.02 | 0.13 ± 0.03 | 4.31 ± 0.09 | 0.11 ± 0.05 | 4.44 | 0.18 |
SBP: | 5.87 ± 0.02 | 0.18 ± 0.04 | 5.98 ± 0.06 | 0.13 ± 0.05 | 6.32 | 0.16 | |
11 | DBP: | 4.02 ± 0.05 | 0.53 ± 0.02 | 4.30 ± 0.16 | 0.48 ± 0.05 | 4.13 | 0.54 |
SBP: | 6.91 ± 0.12 | 0.54 ± 0.02 | 7.62 ± 0.19 | 0.45 ± 0.05 | 8.48 | 0.48 | |
Mean | DBP: | 4.48 ± 0.57 | 0.36 ± 0.16 | 4.68 ± 0.60 | 0.28 ± 0.15 | 4.76 ± 0.58 | 0.33 ± 0.14 |
SBP: | 6.79 ± 1.70 | 0.41 ± 0.17 | 7.46 ± 2.00 | 0.31 ± 0.18 | 7.78 ± 2.05 | 0.25 ± 0.22 |
Figures 2, 3, and 4 show the Bland-Altman plots for the DANN model trained with three, four, or five minutes of training data, respectively. In the three-minute model, 96.0% of predictions have diastolic error less than 10 mmHg, however only 84.5% of predictions have systolic error less than 10 mmHg. This result is below the ISO standard, and prompt repeating the experiment with five minutes of training data. In the four-minute and five-minute model, this error improves to be within the ISO standard: 96.2% diastolic error and 85.9% systolic error are less than 10 mmHg in four-minute model, and 96.2% diastolic error and 85.5% systolic error are less than 10 mmHg in five-minute model. The decrease from the four-minute model to the five-minute model might be the missing subject in the five-minute model.
Figure 2:
Bland-Altman plot for DANN model using three minutes of subject-specific training data
Figure 3:
Bland-Altman plot for DANN model using four minutes of subject-specific training data
Figure 4:
Bland-Altman plot for DANN model using five minutes of subject-specific training data
Figure 5 is an example of estimated and target blood pressure with five minutes of training data applying DANN. Five minutes of training data is able to track the change of blood pressure, e.g. the systolic blood pressure in the figure. However, the reduced training data cannot always respond to changes in blood pressure, especially for cases with lower variability such as the estimation of diastolic blood pressure in the figure.
Figure 5:
Estimated and target blood pressure plots from a subject. The estimation here is provided by the DANN model and trained with five minutes of training data. This plot is not completely representative: for some subjects with lower variability, the model does not respond to changes in blood pressure and instead predicts a near constant blood pressure.
3.5. Analysis
From these results, we observe that model performance decreases significantly when reducing the training data, and less training data results in much lower accuracy (higher RMSE). With three minutes of training data the original MTL model without a pretrained model or the adversarial training process fails for many subjects. There are five subjects with RMSE over 10 mmHg, which is outside of the acceptable range for ISO standards in blood pressure. However, when training with a pretrained model from another subject, the model performance improves for both estimations. When applying the DANN-based training method, the RMSE further decreases, particularly for systolic blood pressure more so than for diastolic.
When training with four or five minutes of data, all three training approaches show an increase in performance. It is interesting to note that 10 out of 11 subjects obtain RMSE below 10 mmHg when directly training the MTL model without DANN. Comparing to training with the pretrained model and direct training approaches, our DANN-based model still has significant benefits. The DANN-based model has lower RMSE, lower standard deviation, and higher correlation, meaning that it performs better and more robustly for additional subjects. In comparison to direct training of the MTL model, both the DANN-based model and the pretrained model help improve the model performance, meaning that it is important to learn from other subjects when a new subject does not have enough training data. The advantage of the DANN-based model indicates that it is more useful to learn from other subjects and to discard the difference between subjects.
With less training data, the model tends to estimate blood pressure as closer to mean values, causing significant errors for extreme high and low blood pressure. When training the model with three minutes of data, the 85% absolute error for systolic blood pressure is 10.11 mmHg. The 85% error is greater than 10 mmHg of the ISO standard, even though it is already improved by applying DANN. Then, we extend the training data to meet the ISO standard requirement. In this model, four minutes of training data is the minimum required amount of training data to obtain confident blood pressure estimations within ISO standards, and while maximizing clinical convenience for future use.
4. Discussion
In the problem of minimizing patient training data for rapid adaptation of a blood pressure prediction model, this work shows the benefits of using DANN to transfer knowledge from other subjects. Considering the difference of physical signals between subjects, rather than building a generalized model or finding a mapping from one subject to another for transformation, DANN provides an approaching of using adversarial training to extract subject-invariant features for the purpose of transfer learning. When applying DANN to this MTL model, the shared layers can be treated as the feature extractor, and the task-specific networks are the label predictor. We have shown that this approach obtains improvements of RMSE 0.19 to 0.26 mmHg for diastolic blood pressure and 0.46 to 0.67 mmHg for systolic blood pressure in the comparison to our best baseline model. The adversarial training mechanism between feature extractor and domain predictor causes the shared layers in the MTL model to extract information which focus on blood pressure estimation while being trained to be insensitive to differences between subjects. In this manner, applying the model to a previously unknown subject allows for improved training while minimizing training data. We see here that the domain predictor is an important factor in the training process. The unbalanced dataset can cause failure of domain prediction by predicting every sample to the same class, collapsing the model to lack subject-specificity. Introducing the third subject solves this problem by balancing between source and target domains. DANN demonstrates strong performance in adaptation of the model to new individuals with minimal data. Notably, the model trained with DANN is less prone to memorizing individual subject features as compared to the isolated MTL model. We conduct additional experiments, evaluating the model’s ability to interpolate missing blood pressure values, demonstrating the more general regression DANN finds, by intentionally withholding certain blood pressure ranges from the training phase and find that the RMSE of the test set for values within the training set versus those within the generated gap are similar. This additional analysis is discussed in the Supplementary Appendix. The minimal training data for blood pressure with bioimpedance signals is four minutes in this system with the given dataset. This provides a guidance for future data collection to validate the device on more people and with a variety of clinical conditions. Our proposed model also provides a solution to the problem of modeling physical measurement tasks with reduced training data as well as a method of learning the minimum required amount of training data.
4.1. Limitations & Future Directions
One chief limitation is the lack of understanding here of the feature space in which subject variability exists. Given an arbitrarily large number of subjects, it is likely that similarities can be found to allow for more intelligent selection of subjects within the model. Therefore, a chief future direction will be to develop a more formalized understanding of this space to allow for a greater ability to choose which existing training sets should be utilized in adapting the model to a new subject. While DANN as presented in this paper is utilized with two random training subjects, this technique can be adapted to incorporate data from additional subjects. One approach will be to utilize a distance metric on demographics or physiological data to base the training on the most similar subjects. Although DANN improves the model performance by transfer learning from other subjects, the correlation for some subjects is very low, and the model cannot respond well to the data with less variability. In all cases, an exploration of the types of data (e.g. low and high blood pressure values) and their relative impact in training should be further explored.
4.2. Conclusion
In this paper, we propose a DANN-based MTL model to estimate beat-to-beat blood pressure from cuffless bioimpedance signals for new subjects with reduced training data. When reducing the training data to three, four, and five minutes, the base MTL model cannot directly be trained successfully to be within ISO standards. Therefore, in order to transfer knowledge from other subjects efficiently, we modify the DANN training approach to train the feature extractor for subject-invariant features. With DANN, the model obtains average RMSE 4.80 ± 0.74 mmHg for diastolic blood pressure and 7.34 ± 1.88 mmHg for systolic blood pressure when using three minutes training data, 4.64 ± 0.60 mmHg and 7.19 ± 1.79 for diastolic and systolic blood pressure from four minutes training data, and 4.48 ± 0.57 mmHg and 6.79±1.70 for diastolic and systolic blood pressure, respectively, when applying five minutes training data. DANN improves the knowledge transfer ability for three, four, and five minutes of training data in comparison to directly training or training with a pretrained model from another subject, decreasing RMSE by 0.19 to 0.26 mmHg for diastolic blood pressure and by 0.46 to 0.67 mmHg for systolic blood pressure in comparison to the best baseline model of utilizing a pretrained model from another subject. The model performance increases with additional data, and we conclude that four minutes is the minimum requirement to achieve the ISO standard with our proposed model and participant cohort. In the future, we consider expanding the DANN approach to include more subject features and to find a metric for the subject-domain space in order to choose similar subjects prior to adapting a model given new subject with targeted minimal training data.
Acknowledgements
This work was supported, in part, by the National Institutes of Health, under grant 1 R01 EB028106-01.
Appendix
A.1. Model Performance Relative to ISO Standard
For the development of a blood pressure device, ISO standards require that 85% of measurements be within 10 mmHg of a standardized reference value for a given cohort. For each subject in our dataset, Table A1 reports the percentage of measurements that fall within this range for varying lengths (3 minutes, 4 minutes, or 5 minutes) of training data. The mean values are reported as well, showing that with 4 minutes of training data, 96.1% of DBP and 85.2% of SBP measurements fall within this range for the cohort studied. This framework presented allows for further adaptation of DANN training times as data collection from future cohorts progresses.
Table A1:
The percentage (%) of results from the DANN model that fall within 10 mmHg of the reference value. To meet ISO standards in a given cohort, at least 85% of measurements must fall within that range.
3 mins | 4 mins | 5 mins | ||||
---|---|---|---|---|---|---|
Subject | DBP | SBP | DBP | SBP | DBP | SBP |
1 | 95.7% | 91.3% | 95.8% | 91.7% | 95.9% | 90.7% |
2 | 94.4% | 77.1% | 93.8% | 81.4% | 91.9% | 74.2% |
3 | 98.4% | 91.7% | 97.4% | 91.4% | 98.3% | 93.7% |
4 | 98.2% | 80.4% | 98.7% | 82.3% | 98.8% | 82.5% |
5 | 90.0% | 85.3% | 94.0% | 91.0% | 92.3% | 91.3% |
6 | 92.7% | 78.0% | 96.4% | 87.3% | - | - |
7 | 95.2% | 79.5% | 94.7% | 81.7% | 95.4% | 83.2% |
8 | 92.7% | 65.6% | 92.5% | 63.1% | 94.2% | 65.8% |
9 | 99.7% | 89.4% | 99.6% | 93.5% | 99.8% | 95.2% |
10 | 96.6% | 89.4% | 97.1% | 89.6% | 97.2% | 89.4% |
11 | 96.9% | 82.4% | 97.5% | 83.9% | 97.2% | 85.6% |
Mean | 95.5% | 83.1% | 96.1% | 85.2% | 96.1% | 85.2% |
A.2. Model Interpolation
To further evaluate DANN’s ability to generate a general regression model, which may aid in future reduction of needed training data, we test the ability of the model to interpolate blood pressures in specific ranges that are intentionally withheld from training. For each individual we adapt the model to using DANN, we first remove from all samples with either diastolic or systolic blood pressure within a specific range (for example, systolic blood pressure from 120–125 mmHg) from the training set. We then repeat model training (using 4-minutes of training data) and test on the full, held out test set. This is analogous to the experiment with results recorded in Table 2 but with a different distribution of training data, reducing the ranges of blood pressures seen from the new individual.
After training, we test the model with the full test set, which includes blood pressure values from the test individual withheld from training. We report both overall RMSE for all test data and in-gap RMSE where the test data exclusive comes from the omitted blood pressure range. Will illustrate model performance with diastolic gaps of 5 mmHg and systolic gaps of 6 mmHg. Specifically, we tested diastolic gaps of 55–60, 65–70, 70–75, 75–80, 80–85, and 85–90 mmHg, and systolic gaps of 90–96, 95–101, 100–106, 105–111, 110–116, 115–121, 120–126, 125–131, 130–136, 135–141, 140–146, and 145–151 mmHg. Due to variations between subjects, not all gaps were tested on all subjects. For instance, a subject whose systolic blood pressure never fell below 106 mmHg would not be included in a gap test for the systolic range of 100–106 mmHg. The overall RMSE and in-gap RMSEs were averaged over each subject, and those values are reported in Table A2. We test these gaps at intervals throughout the distribution of blood pressures present. We note that, even with the errors introduced from the missing values, DANN still outperforms the other models.
As seen in Table A2, the model error tends to increase slightly in the gaps of training data. This is expected given that this model is never trained on values from within those gaps. However, the mean of the error within the gap and overall is still small, showing that the model is able to successfully interpolate to unseen values.
Table A2:
Model results when trained using gaps in training data. DBP gap size is 5 mmHg and SBP gap size is 6 mmHg. Results shown are averaged across varying gap locations as described in the text.
DBP | SBP | |||
---|---|---|---|---|
Subject | Overall RMSE | In-Gap RMSE | Overall RMSE | In-Gap RMSE |
1 | 4.64 | 5.19 | 5.99 | 7.38 |
2 | 5.61 | 5.56 | 8.23 | 9.19 |
3 | 4.13 | 5.06 | 5.73 | 7.27 |
4 | 3.75 | 4.03 | 7.80 | 8.94 |
5 | 5.14 | 5.96 | 5.86 | 7.18 |
6 | 6.01 | 6.95 | 8.46 | 9.11 |
7 | 5.24 | 5.62 | 8.03 | 9.48 |
8 | 5.68 | 6.93 | 10.76 | 10.81 |
9 | 4.07 | 4.41 | 5.74 | 6.43 |
10 | 4.15 | 4.15 | 5.86 | 6.47 |
11 | 4.33 | 4.77 | 7.09 | 8.27 |
Mean | 4.80 ± 0.74 | 5.33 ± 0.96 | 7.23 ± 1.60 | 8.23 ± 1.40 |
Figure A1 further illustrates this finding, showing pooled test predictions for all users, including for diastolic gaps of 5 mmHg located at 70–75 mmHg. In these plots, the orange points represent samples that fell within the excluded range (in-gap samples) and the blue points represent samples from outside of the gaps. The left side of each figure shows the diastolic pressures, and the right side of each figure shows the systolic pressures. As can be seen, omitting a range of diastolic pressures does not clearly omit a range of systolic pressure, reflecting the lack of simple relationship between diastolic and systolic pressures.
Similarly, a plot of gaps in systolic blood pressures are shown in Figure A2 of 6 mmHg gaps located at 125–131 mmHg. As before, the orange points represent samples that fell within the excluded range (in-gap samples) and the blue points represent samples from outside of the gaps. The left side of each figure shows the diastolic pressures, and the right side of each figure shows the systolic pressures.
When a gap falls in the middle of the blood pressure distribution, both low and high blood pressures are equally trained, and when a gap falls at an extreme range, one side is trained well and the side with reduced data is trained poorly. This unbalanced training for an extreme gap results in higher difficulty of generalization. Therefore, the middle gap has fewer errors than gaps in low and high blood pressure ranges and why the in-gap RMSE reported in Table A2 sees a slight increase compared to the overall RMSE.
Figure A1:
Bland-Altman plot for DANN model using four minutes with middle DBP gap
Figure A2:
Bland-Altman plot for DANN model using four minutes with middle SBP gap
To further test the generalizability of this model, we explored the impact of other gap sizes. We tested diastolic gap sizes of 3 mmHg, 5 mmHg, 7 mmHg, and 10 mmHg and systolic gap sizes of 5 mmHg, 6 mmHg, 7 mmHg, and 10 mmHg. Results of this test on Subject 1 are shown in Table A3. As expected, increasing gap size results in higher RMSE for both overall and in-gap evaluations. In particular, this model struggles in generalizing across gaps of 7 mmHg or larger. For the samples tested here and the data length available, generalizations across this gap size appear to be ill-advised. While this paper centrally focuses on the duration of data needed, aiming to reduce data collection burdens on new users, additional work is needed to identify if further reductions are possible, including the type of blood pressure data needed, such as only low and high blood pressure values. While not an explicit goal of this work, DANN finds preliminary results indicating gaps are possible, and because of the distribution of training data available from other subjects, indicates low and high blood pressure values are more important to provide for the new user than middle (normal) blood pressure values.
Table A3:
Generalization results for varying gap sizes applied to Subject 1. As would be expected, increasing gap size results in poorer performance.
Gap Type | Gap Size | Overall RMSE | In-Gap RMSE |
---|---|---|---|
DBP | 3 | 4.58 | 4.87 |
DBP | 5 | 4.64 | 5.18 |
DBP | 7 | 4.85 | 6.69 |
DBP | 10 | 5.59 | 6.38 |
SBP | 5 | 5.96 | 6.24 |
SBP | 6 | 5.99 | 6.36 |
SBP | 7 | 6.21 | 7.07 |
SBP | 10 | 6.37 | 8.24 |
A.3. MTL Beat-to-Beat Performance Per Subject with 80% Training Data
While the primary focus of this work is on the performance of this model with the application of DANN, we separately studied the performance of the isolated MTL model. For each subject, we split the data to be 80% as training set, 10% as validation set, and 10% as test set. We test the whole dataset without repetition from 10-fold cross-validation. This experiment shows overfitting: while the training set is modeled with high average correlation and low average RMSE, the test set suffers significantly in comparison. Performances on the test set are shown in Table A4. These values still provide a basis for modeling of blood pressure but demonstrate the need for more intelligently trained models, such as DANN in this work.
Table A4:
MTL beat-to-beat performance per subject with 80% training data for diastolic and systolic blood pressure (DBP & SBP) RMSE (mmHg) and R.
Subject | DBP RMSE | SBP RMSE | DBP R | SBP R |
---|---|---|---|---|
1 | 4.40 | 5.84 | 0.48 | 0.29 |
2 | 5.40 | 8.55 | 0.54 | 0.63 |
3 | 3.95 | 5.86 | 0.42 | 0.58 |
4 | 4.14 | 7.55 | 0.11 | 0.33 |
5 | 5.29 | 5.73 | 0.23 | 0.27 |
6 | 6.15 | 8.16 | 0.25 | 0.49 |
7 | 4.94 | 7.84 | 0.40 | 0.48 |
8 | 5.30 | 10.93 | 0.40 | 0.50 |
9 | 3.70 | 5.49 | 0.42 | 0.22 |
10 | 4.06 | 5.65 | 0.25 | 0.23 |
11 | 4.30 | 7.18 | 0.47 | 0.54 |
Mean | 4.69 ± 0.73 | 7.16 ± 1.68 | 0.36 ± 0.13 | 0.41 ± 0.15 |
Footnotes
ISO standards need to be met for multiple cohorts which should be representative of different populations. In this work, only one cohort is studied. For the sake of brevity when referring to ISO standards we refer specifically to the studied cohort and not to future cohorts.
Subject 6 has only five minutes of data in total, and so is excluded from analyses with five minutes training data.
References
- Anand Vijay and Kanhangad Vivek. Cross-sensor pore detection in high-resolution fingerprint images using unsupervised domain adaptation. arXiv preprint arXiv:1908.10701, 2019.
- Banegas José R, Ruilope Luis M, de la Sierra Alejandro, Vinyoles Ernest, Gorostidi Manuel, de la Cruz Juan J, Ruiz-Hurtado Gema, Segura Julián, Rodríguez-Artalejo Fernando, and Williams Bryan. Relationship between clinic and ambulatory blood-pressure measurements and mortality. New England Journal of Medicine, 378(16):1509–1520, 2018. ISSN 0028–4793. [DOI] [PubMed] [Google Scholar]
- Blitzer John, McDonald Ryan, and Pereira Fernando. Domain adaptation with structural correspondence learning. In Proceedings of the 2006 conference on empirical methods in natural language processing, pages 120–128. [Google Scholar]
- Chattopadhyay Rita, Sun Qian, Fan Wei, Davidson Ian, Panchanathan Sethuraman, and Ye Jieping. Multisource domain adaptation and its application to early detection of fatigue. ACM Transactions on Knowledge Discovery from Data (TKDD), 6(4):1–26, 2012. ISSN 1556–4681. [Google Scholar]
- Conneau Alexis, Lample Guillaume, Ranzato Marc’Aurelio, Denoyer Ludovic, and Jégou Hervé. Word translation without parallel data. arXiv preprint arXiv:1710.04087, 2017.
- Cook Diane, Feuz Kyle D, and Krishnan Narayanan C. Transfer learning for activity recognition: A survey. Knowledge and information systems, 36(3):537–556, 2013. ISSN 0219–1377. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Duan Lixin, Tsang Ivor W, and Xu Dong. Domain transfer multiple kernel learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(3):465–479, 2012. ISSN 0162–8828. [DOI] [PubMed] [Google Scholar]
- Fan Hong-Qi, Li Yan, Thijs Lutgarde, Hansen Tine W, Boggia José, Kikuya Masahiro, Björklund-Bodegård Kristina, Richart Tom, Ohkubo Takayoshi, and Jeppesen Jørgen. Prognostic value of isolated nocturnal hypertension on ambulatory measurement in 8711 individuals from 10 populations. Journal of hypertension, 28(10):2036–2045, 2010. ISSN 0263–6352. [DOI] [PubMed] [Google Scholar]
- Faridee Abu Zaher Md, Al Hafiz Khan Md Abdullah, Pathak Nilavra, and Roy Nirmalya. Augtoact: scaling complex human activity recognition with few labels. In Proceedings of the 16th EAI International Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services, pages 162–171. [Google Scholar]
- Ganin Yaroslav, Ustinova Evgeniya, Ajakan Hana, Germain Pascal, Larochelle Hugo, Laviolette François, Marchand Mario, and Lempitsky Victor. Domain-adversarial training of neural networks. The Journal of Machine Learning Research, 17(1):2096–2030, 2016. ISSN 1532–4435. [Google Scholar]
- Getmantsev Evgeny, Zhurov Boris, Pyrkov Timothy V, and Fedichev Peter O. A novel health risk model based on intraday physical activity time series collected by smartphones. arXiv preprint arXiv:1812.02522, 2018.
- Gong Boqing, Shi Yuan, Sha Fei, and Grauman Kristen. Geodesic flow kernel for unsupervised domain adaptation. In 2012 IEEE Conference on Computer Vision and Pattern Recognition, pages 2066–2073. IEEE. ISBN 1467312282. [Google Scholar]
- Ibrahim Bassem and Jafari Roozbeh. Cuffless blood pressure monitoring from an array of wrist bio-impedance sensors using subject-specific regression models: Proof of concept. IEEE transactions on biomedical circuits and systems, 13(6):1723–1735, 2019. ISSN 1932–4545. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ibrahim Bassem, McMurray Justin, and Jafari Roozbeh. A wrist-worn strap with an array of electrodes for robust physiological sensing. In 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pages 4313–4317. IEEE, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kachuee Mohammad, Kiani Mohammad Mahdi, Mohammadzade Hoda, and Shabany Mahdi. Cuffless blood pressure estimation algorithms for continuous health-care monitoring. IEEE Transactions on Biomedical Engineering, 64(4):859–869, 2016. ISSN 0018–9294. [DOI] [PubMed] [Google Scholar]
- Ketykó István and Kovács Ferenc. On the metrics and adaptation methods for domain divergences of semg-based gesture recognition. arXiv preprint arXiv:1912.08914, 2019.
- Kouw Wouter M and Loog Marco. An introduction to domain adaptation and transfer learning. arXiv preprint arXiv:1812.11806, 2018.
- Liu Ming-Yu, Breuel Thomas, and Kautz Jan. Unsupervised image-to-image translation networks. In Advances in neural information processing systems, pages 700–708. [Google Scholar]
- Long Mingsheng, Zhu Han, Wang Jianmin, and Jordan Michael I. Deep transfer learning with joint adaptation networks. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 2208–2217. JMLR.org. [Google Scholar]
- Luo Ningqi, Dai Wenxuan, Li Chenglin, Zhou Zhiqiang, Lu Liyuan, Poon Carmen CY, Chen Shih-Chi, Zhang Yuanting, and Zhao Ni. Flexible piezoresistive sensor patch enabling ultralow power cuffless blood pressure measurement. Advanced Functional Materials, 26(8):1178–1187, 2016. ISSN 1616–301X. [Google Scholar]
- Nabeel PM, Karthik Srinivasa, Joseph Jayaraj, and Sivaprakasam Mohanasankar. Arterial blood pressure estimation from local pulse wave velocity using dual-element photoplethysmograph probe. IEEE Transactions on Instrumentation and Measurement, 67(6):1399–1408, 2018. ISSN 0018–9456. [Google Scholar]
- O’Brien Eoin, Asmar Roland, Beilin Lawrie, Imai Yutaka, Mancia Giuseppe, Mengden Thomas, Myers Martin, Padfield Paul, Palatini Paolo, and Parati Gianfranco. Practice guidelines of the european society of hypertension for clinic, ambulatory and self blood pressure measurement. Journal of hypertension, 23(4):697–701, 2005. ISSN 0263–6352. [DOI] [PubMed] [Google Scholar]
- Olsen Michael H, Angell Sonia Y, Asma Samira, Boutouyrie Pierre, Burger Dylan, Chirinos Julio A, Damasceno Albertino, Delles Christian, Gimenez-Roqueplo Anne-Paule, and Hering Dagmara. A call to action and a lifecourse strategy to address the global burden of raised blood pressure on current and future generations: the lancet commission on hypertension. The Lancet, 388(10060):2665–2712, 2016. ISSN 0140–6736. [DOI] [PubMed] [Google Scholar]
- Peter Lukáš, Noury Norbert, and Cerny M. A review of methods for non-invasive and continuous blood pressure monitoring: Pulse transit time method is promising? Irbm, 35(5):271–282, 2014. ISSN 1959–0318. [Google Scholar]
- Pickering Thomas G, Harshfield Gregory A, Kleinert Hollis D, Blank Seymour, and Laragh John H. Blood pressure during normal daily activities, sleep, and exercise: comparison of values in normal and hypertensive subjects. Jama, 247(7):992–996, 1982. ISSN 0098–7484. [PubMed] [Google Scholar]
- Pickering Thomas G, James Gary D, Boddie Charlene, Harshfield Gregory A, Blank Seymour, and Laragh John H. How common is white coat hypertension? Jama, 259(2): 225–228, 1988. ISSN 0098–7484. [PubMed] [Google Scholar]
- Shay Oliver Hao-Yuan and Dai Lillian Lei. System and method for biometric measurements, May 4 2017. US Patent App. 15/337,127.
- Snell Jake, Swersky Kevin, and Zemel Richard. Prototypical networks for few-shot learning. In Advances in neural information processing systems, pages 4077–4087. [Google Scholar]
- Stergiou George, Palatini Paolo, Asmar Roland, de la Sierra Alejandro, Myers Martin, Shennan Andrew, Wang Jiguang, O’brien Eoin, and Parati Gianfranco. Blood pressure measurement and hypertension diagnosis in the 2017 us guidelines: first things first. Hypertension, 71(6):963–965, 2018a. ISSN 0194–911X. [DOI] [PubMed] [Google Scholar]
- Stergiou George S, Alpert Bruce, Mieke Stephan, Asmar Roland, Atkins Neil, Eckert Siegfried, Frick Gerhard, Friedman Bruce, Graßl Thomas, and Ichikawa Tsutomu. A universal standard for the validation of blood pressure measuring devices: Association for the advancement of medical instrumentation/european society of hypertension/international organization for standardization (aami/esh/iso) collaboration statement. Hypertension, 71(3):368–374, 2018b. ISSN 0194–911X. [DOI] [PubMed] [Google Scholar]
- Taigman Yaniv, Polyak Adam, and Wolf Lior. Unsupervised cross-domain image generation. arXiv preprint arXiv:1611.02200, 2016.
- Tan Ben, Zhang Yu, Pan Sinno Jialin, and Yang Qiang. Distant domain transfer learning. In Thirty-First AAAI Conference on Artificial Intelligence. [Google Scholar]
- Thomas Simi Susan, Nathan Viswam, Zong Chengzhi, Soundarapandian Karthikeyan, Shi Xi-angrong, and Jafari Roozbeh. Biowatch: A noninvasive wrist-based blood pressure monitor that incorporates training techniques for posture and subject variability. IEEE journal of biomedical and health informatics, 20(5):1291–1300, 2016. ISSN 2168–2194. [DOI] [PubMed] [Google Scholar]
- Tzeng Eric, Hoffman Judy, Zhang Ning, Saenko Kate, and Darrell Trevor. Deep domain confusion: Maximizing for domain invariance. arXiv preprint arXiv:1412.3474, 2014.
- Wang Jindong, Chen Yiqiang, Hao Shuji, Feng Wenjie, and Shen Zhiqi. Balanced distribution adaptation for transfer learning. In 2017 IEEE International Conference on Data Mining (ICDM), pages 1129–1134. IEEE, a. ISBN 1538638355. [Google Scholar]
- Wang Jindong, Zheng Vincent W, Chen Yiqiang, and Huang Meiyu. Deep transfer learning for cross-domain activity recognition. In proceedings of the 3rd International Conference on Crowd Science and Engineering, pages 1–8, b. [Google Scholar]
- Wang Yang, Liu Zhiwen, and Ma Shaodong. Cuff-less blood pressure measurement from dual-channel photoplethysmographic signals via peripheral pulse transit time with singular spectrum analysis. Physiological measurement, 39(2):025010, 2018. ISSN 0967–3334. [DOI] [PubMed] [Google Scholar]
- Whelton Paul K, Carey Robert M, Aronow Wilbert S, Casey Donald E, Collins Karen J, Himmelfarb Cheryl Dennison, DePalma Sondra M, Gidding Samuel, Jamerson Kenneth A, and Jones Daniel W. 2017 acc/aha/aapa/abc/acpm/ags/apha/ash/aspc/nma/pcna guideline for the prevention, detection, evaluation, and management of high blood pressure in adults: a report of the american college of cardiology/american heart association task force on clinical practice guidelines. Journal of the American College of Cardiology, 71(19):e127–e248, 2018. ISSN 0735–1097. [DOI] [PubMed] [Google Scholar]
- Yao Yi and Doretto Gianfranco. Boosting for transfer learning with multiple sources. In 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages 1855–1862. IEEE. ISBN 1424469856. [Google Scholar]
- Yuan Ye, Chen Wuyang, Chen Tianlong, Yang Yang, Ren Zhou, Wang Zhangyang, and Hua Gang. Calibrated domain-invariant learning for highly generalizable large scale re-identification. In The IEEE Winter Conference on Applications of Computer Vision, pages 3589–3598. [Google Scholar]
- Zhao Zhongtang, Chen Yiqiang, Liu Junfa, Shen Zhiqi, and Liu Mingjie. Cross-people mobile-phone based activity recognition. In Twenty-second international joint conference on artificial intelligence. [Google Scholar]
- Zheng Ya-Li, Bryan P Yan Yuan-Ting Zhang, and Poon Carmen CY. An armband wearable device for overnight and cuff-less blood pressure measurement. IEEE transactions on biomedical engineering, 61(7):2179–2186, 2014. ISSN 0018–9294. [DOI] [PubMed] [Google Scholar]