(Su et al. 2018) [34] |
Proprietary data (84 subjects, 10 min each) |
PTT + deep learning (LSTM) |
Unknown |
RMSE |
3.73 |
2.43 |
(Mohammad Kachuee et al. 2017) [35] |
MIMIC II (1000 subjects) |
PTT + classical ML (AdaBoost) |
Yes |
MAE |
11.17 |
5.35 |
(Kurylyak, Lamonaca, and Grimaldi 2013b) [51] |
MIMIC (15,000 beats) |
Temporal PPG features + artificial neural network (ANN)) |
Unknown |
MAE |
3.8 |
2.21 |
(Gupta et al. 2022) [52] |
MIMIC I, II, and III (39 subjects, ? subjects, 510 subjects) |
PPG signal’s derivative contours + ML algorithms |
Yes |
MAE |
0.74, 1.69, 1.30 |
0.35, 0.77, 0.56 |
(Bernard, Msigwa, and Yun 2022) [36] |
MIMIC II (69 subjects) and proprietary data (23 subjects) |
five 1-D CNN, three Bi-directional LSTM networks |
No |
MAE |
1.38 |
0.95 |
(Leitner, Chiang, and Dey 2022) [37] |
MIMIC III (100 subjects) |
RCNN neural networks + personalization |
Yes |
MAE |
3.52 |
2.2 |
(Wang et al. 2022) [53] |
MIMIC II (348 records) |
Visibility graph + transfer learning |
No |
MAE |
6.17 |
3.66 |
(Schlesinger et al. 2020) [38] |
MIMIC II (304 subjects) |
CNN + Siamese Network |
No |
MAE |
5.95 |
3.41 |
Our proposed work (MLM-Transformer w/personalization) |
MIMIC III (1,732 subjects) |
3-layer Time Series Transformer + personalized fine-tuning |
Yes |
MAE |
2.41 |
1.31 |