Skip to main content
. 2020 May 27;11(3):387–398. doi: 10.1055/s-0040-1710525

Table 1. Overview of unique characteristics and differences between methods.

Author Year Goal Population location and size Feature set size “Gold standard” definition Model and performance metrics
Carrara et al 8 2015 Mortality prediction in septic shock patients MIMIC II (ICU)
30,000+ patients
30 variables Septic shock: 1991 SIRS criteria, ICD-9 code for septic shock, abnormal interval must exceed 5 h for each feature, SIRS 2 + , SIRS with low SBP despite adequate fluid resuscitation Multivariate linear regression with Shrinkage Techniques model
Mean square error (MSE): 0.03
Danner et al 9 2017 Assess the value of HR-to-systolic ratio in the accuracy of sepsis prediction after ED presentation Local (ED)
53,313 patients
9 vitals/variables Sepsis: Discharge diagnosis of sepsis, evaluated vitals, demographics, chief complaints Multivariate linear regression model
- Accuracy: 0.74
- HR to systolic ratio accounted for 69% of overall predictive ability
Capp et al 10 2015 Describe key patient characteristics present within 4 h of ED arrival that are associated with developing septic shock between 4 and 48 h of ED arrival Local (ED)
1,316 patients
5 risk factors Sepsis: manual chart review with SIRS 2 + , evidence of infection (excluded if gastrointestinal bleed)
Septic shock: SBP > 90 mm Hg despite appropriate fluid hydration of 30 mL/kg with presence of hypotension for at least 2 h after
Multivariable logistic regression model
Found risk factors associated with progression of sepsis to septic shock between 4 and 48 h of ED arrival:
- Female: 1.59 odds ratio (OR)
- Nonpersistent hypotension: 6.24 OR
- Lactate > 4 mmol/L: 5.30 OR
- Bandemia > 10%: 2.60 OR
- Past medical of coronary heart disease: 2.01 OR
Faisal et al 11 2018 To develop a logistic regression model to predict the risk of sepsis following emergency admission using the patient's first electronically recorded vital signs and blood test results and to validate this novel computer-aided risk of sepsis model, using data from another hospital Local (ED)
57,243 patients
12 vitals/variables Sepsis: ICD-10 codes without organ failure
Severe sepsis: ICD-10 codes with 1+ organ failure or septic shock
Logistic regression models
All area under the receiver operator curve (AUROC): 0.79
Sepsis AUROC: 0.70
Severe sepsis AUROC: 0.81
Ho et al 12 2012 Investigate how different imputation methods can overcome the handicap of missing information MIMIC II (ICU)
Sample size not stated
6 vitals Sepsis: ICD-9
Septic shock: examined clinical chart records
- Sepsis: Multivariate logistic regression models
- Septic shock: multivariate logistic regression, linear kernel SVM, and regression trees
H: Clinical history feature set
P: initial physiological state feature set
Sepsis AUROC (imputed mean and matrix factorization-based approaches)
All H: 0.791 (0.792)
Stepwise H: 0.790 (0.791)
All H ∪ P: 0.821 (0.822)
Stepwise H ∪ P: 0.823 (0.823)
Septic shock AUROC: 0.773–0.786
Langley et al 13 2013 Examine clinical features, plasma metabolome, and proteome of patients to predict patient survival of sepsis CAPSOD (ED)
1,152 individuals with suspected, community-acquired sepsis; Discovery set of 150 patients
4 vitals/variables Acute infection + 2+ SIRS Logistic regression (sepsis prediction) and SVM model (survival and death prediction)
Logistic regression AUROC: 0.847
Logistic regression accuracy: 0.851
*best stats occurred at enrollment
SVM AUROC: 0.740
SVM accuracy: 0.746
Sutherland et al 14 a 2011 Use gene expression biomarkers to prospectively distinguish patients with sepsis from those who experience systemic inflammation from healing of surgery Local (ICU)
85 patients
42 biomarkers Likely enter sepsis cohort if met ACCP/SCCM consensus statement and clinical suspicion of systemic infection
Confirmation performed retrospectively
Classifier: Recursive partitioning, LASSO, logistic regression. Individual genes examined via Bayes-adjusted linear model.
MT-PCR diagnostic classifier generated using a LogitBoost ML algorithm (tree-based)
PCR Accuracy: 92%
AUROC: 0.86–0.92
Gultepe et al 15 2014 Develop a decision support system to identify patients with hyperlactatemia and to predict mortality from sepsis using predicted lactate levels Local (ED)
741 patients
7 vitals/laboratories Sepsis: determined from EHR diagnosis and SIRS criteria SVM classifier
Accuracy: 0.73
AUROC: 0.73
Horng et al 16 2017 To demonstrate the incremental benefit of using free text data in addition to vital sign and demographic data to identify patients with suspected infection in the emergency department Local (ED)
198,833 control
32,103 cases
12 vitals/variables ED ICD-9-CM code Linear SVM and free text models
Bag of words AUROC: 0.86
Bag of words sensitivity: 0.78
Bag of words specificity: 0.79
Topic model AUROC: 0.85
Topic model sensitivity: 0.80
Topic model specificity: 0.75
Thottakkara et al 17 2016 To compare performance of risk prediction models for forecasting postoperative sepsis and acute kidney injury Local (in-patient)
50,318 patients
285 variables Forecast postop sepsis and acute kidney injury
AHRQ definition of “post-op sepsis” and organ failure associated with sepsis was identified by ICD-9-CM code for acute organ dysfunction
Comparison of models that used logistic regression, generalized additive models (GAM), naive Bayes, SVM
Naive Bayes performed the worst in the comparison; GAMs and SVMs had good performance; PCA feature extraction (reduced to 5 features) improved predictive performance for all models
Severe sepsis AUROC: 0.76–0.91
Vieira et al 18 2013 Proposed a modified binary particle swarm optimization method for feature selection to predict mortality in septic patients MEDAN (ICU)
382 patients
Model chooses custom number of features (2–7) MEDAN data set prelabeled patients for abdominal septic shock Support vector machine for mortality prediction
Modified binary particle swarm optimization (MBPSO): feature selection MBPSO
12 (28) features:
No-FS Accuracy: 72.6% (89%)
Accuracy: 76.5% (94.4%)
Ghosh et al 19 2017 Predict septic shock for ICU patients using noninvasive waveform measurements MIMIC II (ICU)
1,519 patients
3 vitals/laboratories Sepsis: ICD-9
Septic shock: examining clinical chart records
Coupled hidden Markov models (CHMM) with varying gap interval and observation window sizes
CHMM average: 0.85
Multichannel patterns (MCP)-CHMM average: 0.86
Peelen et al 20 2010 Develop a set of complex Markov models based on clinical data to extract meaningful clinical patterns and to provide prediction for sepsis and other diseases Local (ICU)
2,271 patients
6 variables Sever sepsis: SIRS 2+ within 24 h of ICU admission and 1+ dysfunctioning organ system (SOFA) 3 Markov models (amount of organ failure, type of organ failure, differences between development and persistence of organ failure)
ICU death the error rates were 17.7%, 18.1%, and 17.8% and the AUCs were 0.79, 0.79, and 0.80 for models I, II, and III
Stanculescu et al 21 2014 Demonstrate that by adding a higher-level discrete variable with semantics sepsis/nonsepsis, can detect changes in the physiological factors that signal the presence of sepsis Local (NICU)
24 neonates
Bradycardia, desaturation Laboratory result of blood culture for neonatal sepsis Hierarchical switching linear dynamical system (HSLDS)
Autoregressive (AR)-HMM AUROC: 0.72
HSLDS deep learning AUROC: 0.69
HSLDS known factors AUROC: 0.62
Stanculescu et al 22 2014 Detect and identify sepsis in neonates before a blood sample is drawn. Furthermore, they wanted to identify which physiological event would contribute most for detecting sepsis Local (NICU)
24 neonates
6 vitals/variables Positive cultures as pathogens: proven sepsis
Positive cultures as mixed growth/skin commensal: “suspected sepsis”
AR-HMM
AUROC: 0.74–0.75
AUROC with missing data: 0.72–0.73
AUROC with bradycardia and minibradycardia: 0.79–0.80
AUROC with desaturation: 0.76–0.78
AUROC with all states: 0.79–0.80
Gultepe et al 23 2012 Use a Bayesian network to detect sepsis early Local (ICU)
1,492 patients
BN1: 5 variables
BN2: 7 variables
“Sepsis occurrence” Bayesian network (BN) models
BN-1 (vitals) goodness of fit: 15.4
BN-2 (vitals + MAP) goodness of fit: 19.9
Found that lactate is a driver in both models and maybe an important feature for early sepsis detection
Nachimuthu and Haug 24 2012 Detect sepsis right after patients are admitted to the ED Local (ED)
3,100 patients
11 vitals/variables Clinician determined “sepsis” during retrospective chart review Dynamic Bayesian network
3 h after admission AUROC: 0.911
6 h after admission AUROC: 0.915
12 h after admission AUROC: 0.937
24 h after admission AUROC: 0.944
Calvert et al 25 2016 Detect and predict the onset of septic shock for alcohol-use disorder patients in the ICU MIMIC III (ICU)
1,394 patients
9 vitals/variables Septic shock: SIRS 2 + , ICD-9, organ dysfunction, SBP < 90 mm Hg for 1 h, total fluid replacement ≥ 1,200 mL or 20 mL/kg for 24 h InSight
Sensitivity: 0.93
Specificity: 0.91
Accuracy: 0.91
F1 score: 0.161
Calvert et al 26 2016 To develop high-performance early sepsis prediction technology for the “general patient population” MIMIC II (ICU)
29,083 patients
10 vitals/variables Sepsis: ICD-9 code, 1991 SIRS for 5 h InSight
Sensitivity: 0.90
Specificity: 0.81
AUROC: 0.92
Accuracy: 0.83
Desautels et al 27 2016 To validate InSight with the new Sep-3 definition and make predictions using minimal set of variables MIMIC III (ICU)
22,583 patients
8 vitals/laboratories Sepsis: Sep-3 definition, suspicion of infection equated with an order of culture laboratory draw and dose of antibiotics InSight
AUROC: 0.88
APR: 0.60
Mao et al 28 2018 Validate the InSight algorithm for detection and prediction of sepsis and septic shock MIMIC III (ICU)
Local (ED, general)
61,532 stays
6 vitals/laboratories Sepsis: ICD-9 + SIRS 2+ (995.91)
Severe sepsis: ICD-9 (955.92), organ dysfunction, SIRS 2+
Septic shock: ICD-9 (785.52), SBP < 90 mm Hg (at least 30 min), resuscitated with ≥ 20 mL/kg over 24 h, ≥ 1,200mL in total fluids
InSight
Detect sepsis AUROC: 0.92
Detect severe sepsis AUROC: 0.87
Detect 4 h before onset sepsis AUROC: 0.96
Detect 4 h before onset severe sepsis AUROC: 0.85
McCoy and Das 29 2017 Aimed to improve sepsis-related patient outcomes through a revised sepsis management approach Local (ICU)
407 patients
6 vitals/variables Severe sepsis: SIRS 2 + , qSOFA score Dascena
Sep-3 AUROC: 0.91
Sep-3 sensitivity: 0.83
Sep-3 specificity: 0.96
Severe sepsis AUROC: 0.96
Severe sepsis sensitivity: 0.90
Severe sepsis specificity: 0.85
Shimabukuro et al 30 a 2017 Randomized control trial to show lowered mortality and length of stay using a machine learning sepsis prediction algorithm Local (ICU)
75 controls
67 cases
7 vitals/laboratories Severe sepsis: “organ dysfunction caused by sepsis”
Random allocation sequence to put patients in groups
InSight
AUROC: 0.952
Sensitivity: 0.9
Specificity: 0.9
Average length of stay decreased from 13 to 10.3 d
In-hospital mortality decreased by 12.3%
Henry et al 32 2015 Create and test a score that predicts which patients will develop septic shock MIMIC II (ICU)
16,234 patients
54 features Suspicion of infection: ICD-9 or by presence of clinical note that mentioned sepsis or septic shock
Sepsis: suspicion + SIRS
Severe sepsis: sepsis + organ dysfunction
TREWScore (Cox proportional hazards model using the time until the onset of septic shock as the supervisory signal)
AUROC: 0.83
Specificity: 0.67
Sensitivity: 0.85
“Patients were identified a median of 28.2 h before shock onset”
Lukaszewski et al 31 a 2008 Detect and identify septic patients before displaying symptoms for ICU patients Local (ICU)
92 patients
7 cytokines Admission diagnosis upon ICU entry Neural networks using cytokine and chemokine data
Sensitivity: 0.91
Specificity: 0.80
Accuracy: 0.95
Nemati et al 35 2018 Aimed to develop and validate an artificial intelligence sepsis algorithm for early prediction of sepsis Local (ICU)
33,069 patients
65 variables Sepsis: Sepsis-3 Modified Weibull–Cox proportional hazards model
4 h in advance AUROC: 0.85
Pereira et al 33 2011 Examined different approaches to predicting septic shock with missing data MEDAN (ICU)
139 patients
2 sets of 12 and 28 “selected features” Septic shock: associated with abdominal causes (not clearly defined, data may be prelabeled) Zero-Order-Hold (ZOH) Fuzzy c-means clustering based on partial distance calculation strategy (FCM-PDS)
Performance improvements occur where up to 60% of the data are missing
ZOH-FCM-PDS 12 (28) feature AUROC: 0.899 (0.649); FCM-PDS 12 (28) feature AUROC: 0.786 (0.631)
Ribas et al 36 2011 Demonstrate that a SVM variant can provide automatic ranking of mortality predictor and have higher accuracy that current methods Local (ICU)
354 patients
4 vitals/laboratories Severe sepsis: organ dysfunction (SOFA) Relevance vector machine
AUROC: 0.80
Error rate: 0.24
Sensitivity: 0.66
Specificity: 0.80
Sawyer et al 38 a 2011 Evaluate if implementing an automated sepsis screening and alert system can facilitate in early interventions by identifying non-ICU patients at risk for developing sepsis Local (Non-ICU)
270 patients
9 vitals/variables Intervention group: real-time sepsis alert generated from Clinical Desktop Recursive partitioning regression tree analysis
Within 12 h of sepsis alert, 70.8% of patients in the intervention group received treatment versus 55.8% in control
Shashikumar et al 37 2017 Investigates the utility of high-resolution blood pressure and heart rate times series dynamics for the early prediction of sepsis Local (ICU)
242 patients
11 vitals/variables Sepsis: Seymour (Sep-3) at some point during ICU stay Elastic Net logistic classifier: 3 models: (1) entropy features, (2) EMR + sociodemographic-patient history features, (3) models 1 + 2
Model 1 AUROC (Accuracy): 0.67 (0.47)
Model 2 AUROC (Acc): 0.70 (0.50)
Model 3 AUROC (Acc): 0.78 (0.61)
Taylor et al 39 2016 Compare a machine learning approach to existing clinical decision rules to predict sepsis in-hospital mortality Local (ED)
4,676 patients
20 variables ICD-9 with AHRQ clinical classification software to obtain more exhaustive list of patients Random forest model
AUROC: 0.86

Abbreviations: AHRQ, Agency for Healthcare Research and Quality; ACCP/SCCM, American College of Chest Physicians/Society of Critical Care; ED, emergency department; EHR, electronic health record; HR, heart rate; ICD, International Classification of Disease; ICU, intensive care unit; MEDAN, Medical Data Warehousing and Analysis; MIMIC, Medical Information Mart for Intensive Care; ML, machine learning; MT-PCR, multiplex tandem-polymerase chain reaction; NICU, neonatal intensive care unit; PCA, principal component analysis; SBP, systolic blood pressure; SIRS, systemic inflammatory response syndrome; SOFA, sequential organ failure assessment; SVM, support vector machine.

a

Indicates prospective.