Abstract
Background
The greater the severity of illness of a patient, the more likely the patient will have a poor hospital outcome. However, hospital-wide severity of illness scores that are simple, widely available, and not diagnosis-specific are still needed. Laboratory tests could potentially be used as an alternative to estimate severity of illness.
Objective
To evaluate the ability of hospital laboratory tests, as measures of severity of illness, to predict in-hospital mortality among hospitalized patients, and therefore, their potential as an alternative method to severity of illness risk adjustment.
Designs and Patients
A retrospective cohort study among 38,367 adult non-trauma patients admitted to the University of Maryland Medical Center between November 2015 and November 2017 was performed. Laboratory tests (hemoglobin, platelet count, white blood cell count, urea nitrogen, creatinine, glucose, sodium, potassium, and total bicarbonate (HCO3)) were included when ordered within 24 h from the time of hospital admission. A multivariable logistic regression model to predict in-hospital mortality was constructed using a section of our cohort (n = 21,003).
Main Measures
Model performance was evaluated using the c-statistic and the Hosmer-Lemeshow (HL) test. In addition, a calibration belt was constructed to determine a confidence interval around the calibration curve with the purpose of identifying ranges of miscalibration.
Key Results
Patient age and all laboratory tests predicted mortality with good discrimination (c = 0.79). Patients with abnormal HCO3 levels or leukocyte counts at admission were twice as likely to die during their hospital stay as patients with normal results. A good model calibration and fit were observed (HL = 13.9, p = 0.18).
Conclusions
Admission laboratory tests are able to predict in-hospital mortality with good accuracy, providing an objective and widely accessible approach to severity of illness risk adjustment.
KEY WORDS: laboratory tests, severity of illness, mortality, prediction model
INTRODUCTION
Severity of illness refers to the degree of organ system derangement in a patient.1 The greater the severity of illness of a patient is, the more likely the patient will have a poor hospital outcome. Available severity of illness scoring systems can be divided into those that are specific for an organ or disease (e.g., Glasgow Coma Scale) and those that are applicable to specific patient populations (e.g., Acute Physiology and Chronic Health Evaluation II (APACHE) score).2 However, hospital-wide severity of illness scores that are simple, widely available, and not diagnosis-specific are still needed.
Laboratory tests could potentially be used as an alternative to estimate a hospital-wide severity of illness. Laboratory tests are routinely ordered upon patient hospital admission, are objective, and are easily accessible from electronic medical records. Leveraging the results of admission laboratory tests to generate a hospital-wide severity of illness score has the potential for a number of important applications. This study focused particularly on two important ones: (a) severity of illness confounding adjustment in epidemiological studies and (b) hospital-level pay-for-performance risk adjustment.
The Centers for Medicare and Medicaid Services (CMS) use risk adjustment to account for differences in beneficiary-level risk factors that can affect quality outcomes or medical costs when comparing hospitals. One of the main components used by CMS to account for patient severity is the Medicare Severity Diagnosis Related Group (MS-DRG).3 MS-DRGs classify each patient based on principal diagnosis, specific secondary diagnosis, procedures, gender, and discharge status.4 Each DRG is assigned a unique weight, which is later multiplied by the hospital’s payment rate per case.5 This approach to classifying illness severity is limited by its diagnosis-specific nature, which is highly dependent on accurate hospital coding of International Classification of Diseases (ICD) codes. Poor to fair ICD-10 coding accuracy and MS-DRG-assignment accuracy (61% and 71% respectively in 2017) have been consistently reported by the Annual National ICD-10 Coding Contests.6 Furthermore, MS-DRG is a proprietary coding system; therefore, this system is not always affordable or accessible to every hospital or researcher. Use of admission laboratory tests to predict severity of illness can circumvent many of these issues.
The objective of this study was to evaluate the ability of hospital admission laboratory tests to predict in-hospital mortality among hospitalized patients. If accurate, evaluation of laboratory tests could provide an alternative method for severity of illness risk adjustment.
METHODS
Study Cohort and Database
We conducted a retrospective cohort study among all non-trauma adult patients admitted to the University of Maryland Medical Center (UMMC), a 755-bed tertiary care hospital located in Baltimore, MA, between November 7, 2015, and October 31, 2017. Patients with multiple hospital admissions during this time period were included in the dataset. Institutional Review Board approval was obtained with a waiver of informed consent. The cohort was created using the hospital’s central data repository, which is a relational database containing administrative, pharmacy, clinical, and laboratory data obtained from electronic medical records. Validation of a random sample of 50 patients and their laboratory variables was performed by chart review and was greater than 99% accurate. Additionally, general validation of this data repository has previously demonstrated positive and negative predictive values greater than 99%.7–10
We chose to analyze laboratory tests that are commonly ordered early during hospital admission (only laboratory tests ordered in at least 80% of patients upon hospital admission at UMMC) and that have been associated with severity of illness (imminent death, in-hospital mortality, adverse events, etc.).11–18 Laboratory values selected included hemoglobin, platelet count, leucocyte count, blood urea nitrogen (BUN), serum creatinine, glucose, sodium, potassium, and total bicarbonate (HCO3). As our purpose was to assess severity of illness at admission, laboratory tests were only included if they were ordered within 24 h of admission. In cases of multiple tests obtained during this 24 h period, only the first laboratory result was selected.
Statistical Analyses
The association of each predictor with the outcome was assessed graphically. When this association was found to be linear, laboratory test results were analyzed as continuous variables. However, in the absence of linearity, variables were categorized as abnormally low, normal, and abnormally high based on standard clinical ranges reported by the UMCC’s Clinical Laboratory. Furthermore, continuous test results are difficult to interpret as clinically relevant unit changes are not consistent across laboratory tests (e.g., a one-unit increase in serum creatinine has a different clinical significance from a one-unit increase in serum glucose). Therefore, we developed another model for the outcome using only categorized laboratory tests (abnormal vs. normal) in order to improve clinical interpretation.
The dataset was randomly split into one development dataset (60% of the original dataset) and one validation dataset (40% of the original dataset). All laboratory tests significantly associated (p < 0.05) with the outcome in the bivariate analysis were considered for inclusion into the multivariable model. As patient’s age is associated with in-hospital mortality 19 and standardized ranges of laboratory values are commonly gender-specific,20 these two variables were also considered for entry into the multivariable model. Additionally, interaction terms between gender and specific laboratory tests were evaluated. Stepwise selection was used to construct multivariable logistic regression models using generalized estimating equations (GEE) to account for the correlation of multiple outcomes from the same patient. If a variable’s p value was < 0.15 when adjusted for the rest of the variables in the model, the variable was kept in the final multivariable model.
Model performance was tested using the validation dataset as follows. For all the records in the validation set, outcomes were predicted using the predictive equation derived from the development subset. The c-statistic evaluated the models’ discriminatory power (i.e., their ability to discriminate between those with the outcomes and those without).21 A c-statistic of 0.50 indicates that the model is no better than flipping a coin in predicting the outcome. Values ≥ 0.70 are considered good, while values ≥ 0.80 are considered excellent.21 The Hosmer-Lemeshow test was used to evaluate the model’s calibration (i.e., how closely the predictions of our model match the observed outcomes in our study population).21 The Hosmer-Lemeshow test with a non-significant p value demonstrates that there is no evidence of a poor model fit.22, 23 Additionally, a GiViTi calibration belt was constructed to determine a confidence interval around the calibration curve.24, 25 The calibration plot contains the bisector, which represents the line of perfect calibration between the predicted and the observed proportions. The calibration belt around the bisector represents the 80% and 95% confidence level calibration of the model. If the belt does not contain the bisector, it indicates a significant lack of fit. The model is either under- or overestimating the outcome.24, 25
RESULTS
Our development dataset had 26,857 admitted patients with a 3.6% mortality rate. Our validation dataset had 11,510 admitted patients with a 3.3% mortality rate. All analyzed laboratory tests were significantly associated with in-hospital mortality in the bivariate analysis and all laboratory tests remained in the first multivariable model (Table 1). The patients missing any of the laboratory tests included in the final multivariate model were automatically excluded from the final model. Twenty-two percent of the observations in the development dataset were excluded in the final model due to missing tests, i.e., not all patients having all the lab tests. The patients excluded were not different than the ones included in the model with respect to mortality. The model’s discriminatory power was good (c = 0.79 (95% CI 0.76–0.81)). After adding patient’s age and gender to the multivariable model, only age and the laboratory tests remained, leading to a slightly increased c-statistic (c = 0.81 (95% CI 0.78–0.83)) (Table 1). Additionally, as some included laboratory tests have gender-specific clinical cutoffs, we considered the interaction terms between gender and these laboratory tests; however, none of them remained in the final model.
Table 1.
Variables† | Bivariate OR (95%CI) |
Multivariable OR (95%CI) |
---|---|---|
Hemoglobin (g/dL) | 0.85 (0.82–0.88) | 0.93 (0.90–0.97) |
Leukocyte count (K/μL) | 1.02 (1.01–1.02) | 1.01 (1.01–1.02) |
Glucose (mg/dL) | 1.00 (1.00–1.00) | 1.00 (1.00–1.00)‡ |
Platelets | ||
Abnormally low (< 153,000 cells/μL) | 3.03 (2.65–3.48) | 2.25 (1.93–2.62) |
Abnormally high (> 367,000 cells/μL) | 1.04 (0.79–1.38) | 0.76 (0.55–1.05) |
Urea nitrogen (mg/dL) | 1.02 (1.02–1.03) | 1.01 (1.00–1.01) |
Serum creatinine | ||
Abnormally low (< 0.52 mg/dL) | 0.99 (0.72–1.35) | 1.02 (0.73–1.43) |
Abnormally high (> 1.04 mg/dL) | 3.84 (3.34–4.41) | 1.78 (1.48–2.14) |
Sodium | ||
Abnormally low (< 136 mmol/L) | 2.18 (1.88–2.52) | 1.58 (1.34–1.87) |
Abnormally high (> 147 mmol/L) | 2.79 (2.21–3.51) | 2.04 (1.58–2.64) |
Potassium | ||
Abnormally low (< 3.5 mmol/L) | 1.59 (1.31–1.92) | 1.39 (1.09–1.79) |
Abnormally high (> 5.2 mmol/L) | 3.46 (2.80–4.28) | 0.87 (0.64–1.17) |
Bicarbonate | ||
Abnormally low (< 21 mmol/L) | 4.40 (3.81–5.08) | 2.45 (2.05–2.93) |
Abnormally high (> 30 mmol/L) | 1.58 (1.24–2.00) | 1.43 (1.11–1.86) |
Age (years) | 1.03 (1.03–1.04) | 1.02 (1.02–1.03) |
*Analysis Was Evaluated Using the Dataset Used for Model Development. The Final Sample Size Included in the Final Multivariate Model is 21,003.
†Laboratory results were analyzed as continuous variables when their association with the outcome was linear. In the absence of linearity with the outcome, variables were categorized as abnormally low (levels below the reference ranges), normal (levels within the reference ranges), and abnormally high (levels above the reference ranges) based on standardized clinical ranges reported by the clinical laboratory. The normal category was defined as the reference category when categorical data were used
‡Glucose association was significantly associated (p < 0.0001) with mortality (OR = 1.003 (1.002–1.003))
In our second multivariate model where all the laboratory variables were categorized, having any of these abnormal tests at admission was associated on average with a 26 to 115% increase in the odds of dying during the hospital stay (Table 2). Patients with abnormal levels of HCO3 or leucocyte counts at admission were twice as likely to die during their hospital stay as patients with normal levels. However, due to the categorization of continuous variables, this model had a minimal decrease in discriminatory power (c = 0.79 (95%CI 0.77–0.81)). The Hosmer-Lemeshow test demonstrated a good model fit (X2 = 13.85, p = 0.18) (Fig. 1a). The calibration belt also demonstrated good calibration, as the belt consistently includes the bisector. Nevertheless, due to the small sample size at medium and high ranges of the predicted probabilities, the calibration belt is wider among probabilities above 0.40 (Fig. 1b).
Table 2.
Variables* | Multivariable OR (95%CI) |
---|---|
Hemoglobin | |
Abnormal | 1.26 (1.06–1.50) |
Females: < 11.9 or > 15.7 g/dL | |
Males: < 12.6 or > 17.4 g/dL | |
Leukocytes count | |
Abnormal (< 4, 500 or > 11, 000 cells/μL) | 2.01 (1.74–2.34) |
Glucose | |
Abnormal (< 70 or > 99 mg/dL) | 1.52 (1.74–2.34) |
Platelets | |
Abnormal (< 153,000 or > 367,000 cells/μL) | 1.80 (1.55–2.08) |
Urea nitrogen | |
Abnormal | 1.46 (1.23–1.73) |
Females: < 6 or > 17 mg/dL | |
Males: < 8 or > 20 mg/dL | |
Serum creatinine | |
Abnormal (< 0.52 or > 1.04 mg/dL) | 1.78 (1.51–2.11) |
Sodium | |
Abnormal (< 136 or > 147 mmol/L) | 1.62 (1.40–1.88) |
Potassium | |
Abnormal (< 3.5 or > 5.2 mmol/L) | 1.49 (1.26–1.75) |
Bicarbonate | |
Abnormal (< 21 or > 30 or mmol/L) | 2.15 (1.84–2.50) |
Age (years) | 1.03 (1.02–1.03) |
*Analysis was evaluated using the dataset used for model development. The final sample size included in the final multivariate model is 21,003. Laboratory variables were categorized as either abnormal (levels outside the reference ranges) or normal (levels within the reference ranges). The normal category was considered the reference category. The US traditional unit system is mostly used in our laboratory; therefore, this is the unit system reported in this manuscript. We detailed here the necessary SI unit conversions: hemoglobin (F, < 119 or > 157 g/L; M, < 126 or > 174 g/L); leukocytes (< 4.5 or > 11 × 109 cells/L); glucose (< 3.9 or > 5.5 mmol/L); platelets (< 153 or > 367 × 109 cells/L); urea nitrogen (F, < 2.1 or > 6.1 mmol/L; M, < 2.6 or > 7.1 mmol/L); serum creatinine (< 46.0 or > 91.9 μmol/L)
DISCUSSION
Our results demonstrate that admission laboratory tests can be used to predict severity of illness among hospitalized patients, suggesting their use as an alternative to the currently used CMS-DRGs. An admission laboratory–based method is more accessible, user-friendly, and more objective. Additionally, as it is not diagnosis-specific, it does not depend on the accuracy of ICD-10 coding in hospitals, which continues to pose challenges.
Previous investigators have evaluated the use of routine laboratory data to predict in-hospital mortality. In 1995, Pine and colleagues evaluated patients admitted with acute myocardial infarction, congestive heart failure, or pneumonia (n = 5966) to several acute care hospitals in Missouri to develop a predictive model for hospital mortality. These investigators included laboratory tests (blood chemistry, hematology, and arterial blood gases) and administrative data (age, sex, admission source, principal and secondary diagnoses based on ICD-9-CM codes) in their predictive model (c = 0.86).18 In 2005, Prytherch et al. used 1 year of adult hospital discharges from general medicine floors (n = 9497) to predict in-hospital mortality from routine laboratory tests (urea, sodium, potassium, albumin, hemoglobin, white blood cell count, creatinine), age, gender, and mode of admission (elective or emergency) (average c = 0.77).16 Limitations with both of these studies include small sample sizes and restriction to specific medical floors or services.
Froom et al. also developed a predictive model for hospital mortality among 10,308 patients hospitalized in internal medicine wards in a regional hospital over a 1-year period. Froom’s final model included age and admission laboratory tests (albumin, alkaline phosphatase, aspartate aminotransferase, urea, glucose, lactate dehydrogenase, neutrophil count, total leukocyte count) (c = 0.89).15 The investigators validated their model using data from the subsequent year; however, the investigators did not evaluate the model’s calibration, which is essential to assess the utility of a predictive model. Assadollahi et al. used a case-control design (550 cases and 1100 controls) to develop a simple clinical scoring system predicting in-hospital patient mortality at admission using seven laboratory variables (age, urea, hemoglobin, white blood cell count, platelet count, sodium, glucose) (c = 0.87).13 Although they selected both medical and surgical patients, they did restrict their study population to only patients admitted through the emergency room. They observed a lower score calibration in the highest score strata, likely as there were only small numbers of patients in this group, a similar phenomenon to what was observed in our model. To our knowledge, this study is the first to predict hospital-wide patient severity of illness using admission laboratory values.
We evaluated a large cohort of patients across several years. However, future validation of our results across other hospitals is necessary. Also, we were unable to include other laboratory tests that are potentially associated with mortality (i.e., albumin, alkaline phosphate, alanine aminotransferases, aspartate aminotransferases, coagulation tests, lactate dehydrogenase) as they were not ordered within 24 h of admission for the majority (> 80%) of patients during our study period. This limitation likely exists in all hospital datasets; furthermore, our model incorporates commonly ordered laboratory tests and is therefore applicable to most hospitalized patients and less likely to be affected by missing data in real-world use. Additionally, our outcome was a rare event in the study population, which may explain our lower model calibration particularly among those patients with higher mortality risk as the sample size used to inform this prediction was limited. Our database is also limited by the availability of MS-DRG codes for the totality of our cohort. As the majority of national hospitals, MS-DRG codes are only available for Medicare and Medicaid patients, as hospitals are required by CMS to report these codes only for this particular group of patients. Although it is beyond the scope of this manuscript, future studies could compare our approach with the CMS-DRG system. Future research using admission laboratory tests could include additional statistical techniques such as machine learning and the use of prediction rules/tools built into electronic medical records.
In conclusion, our findings suggest admission laboratory tests can predict severity of illness in a simple, widely available, and diagnosis non-specific manner assisting both clinicians with patient prognosis and healthcare researchers with adjusting for severity of illness in clinical studies.
Funders
K24AI079040 from the National Institutes of Health (NIH) (Dr. A.D. Harris)
Compliance with Ethical Standards
University of Maryland Baltimore Institutional Review Board approved this study (HP-00040458).
Conflict of Interest
The authors declare that they do not have a conflict of interest.
Footnotes
Prior Presentation
Poster presentation at Society of Epidemiologic Research (SER), June 18–22, 2018
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Vincent JL, Bruzzi de Carvahlo F. Severity of illness. Semin Respir Crit Care Med. 2010;31:31–38. doi: 10.1055/s-0029-1246287. [DOI] [PubMed] [Google Scholar]
- 2.Vincent JL, Moreno R. Clinical review: Scoring systems in the critically ill. Crit Care. 2010;14(2):207. doi: 10.1186/cc8204. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Risk adjustment. Fact Sheet [Internet]. [Accessed: June 6, 2019 ]. Available from: https://www.cms.gov/Medicare/Medicare-Fee-for-Service-Payment/PhysicianFeedbackProgram/Downloads/2015-RiskAdj-FactSheet.pdf
- 4.Defining the Medicare Severity Diagnosis Related Groups (MS-DRGs), Version 34.0 [Internet]. [Accessed: June 6, 2019 ]. Available from: https://www.cms.gov/ICD10Manual/version34-fullcode-cms/fullcode_cms/Defining_the_Medicare_Severity_Diagnosis_Related_Groups_(MS-DRGs)_PBL-038.pdf
- 5.Office of the Inspector General. Medicare Hospital Prospective Payment System. How DRG rates are calculated and updated. 2011.
- 6.2nd Annual ICD-10 Coding Contest Results [Internet]. [Accessed: June 6, 2019]. Available from:http://journal.ahima.org/2017/10/06/2nd-annual-icd-10-coding-contest-results-sponsored/.
- 7.Pepin CS, Thom KA, Sorkin JD, et al. Risk factors for central-line-associated bloodstream infections: a focus on comorbid conditions. Infect Control Hosp Epidemiol. 2015;36(4):479–81. doi: 10.1017/ice.2014.81. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Harris AD, Johnson JK, Thom KA, et al. Risk factors for development of intestinal colonization with imipenem-resistant Pseudomonas aeruginosa in the intensive care unit setting. Infect Control Hosp Epidemiol. 2011;32(7):719–22. doi: 10.1086/660763. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Harris AD, Jackson SS, Robinson G, Pineles L, Leekha S, Thom KA, et al. Pseudomonas aeruginosa Colonization in the Intensive Care Unit: Prevalence, Risk Factors, and Clinical Outcomes. Infect Control Hosp Epidemiol. 2016;37(5):544–8. doi: 10.1017/ice.2015.346. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Harris AD, Fleming B, Bromberg JS, Rock P, Nkonge G, Emerick M, et al. Surgical site infection after renal transplantation. Infect Control Hosp Epidemiol. 2015;36(4):417–23. doi: 10.1017/ice.2014.77. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.ten Boekel E, Vroonhof K, Huisman A, van Kampen C, de Kieviet W. Clinical laboratory findings associated with in-hospital mortality. Clin Chim Acta. 2006;372(1–2):1–13. doi: 10.1016/j.cca.2006.03.024. [DOI] [PubMed] [Google Scholar]
- 12.Kaufman M, Bebee B, Bailey J, Robbins R, Hart GK, Bellomo R. Laboratory tests to identify patients at risk of early major adverse events: a prospective pilot study. Intern Med J. 2014;44(10):1005–12. doi: 10.1111/imj.12509. [DOI] [PubMed] [Google Scholar]
- 13.Asadollahi K, Hastings IM, Gill GV, Beeching NJ. Prediction of hospital mortality from admission laboratory data and patient age: a simple model. Emerg Med Australas. 2011;23(3):354–63. doi: 10.1111/j.1742-6723.2011.01410.x. [DOI] [PubMed] [Google Scholar]
- 14.Asadollahi K, Hastings IM, Beeching NJ, Gill GV. Laboratory risk factors for hospital mortality in acutely admitted patients. QJM. 2007;100(8):501–7. doi: 10.1093/qjmed/hcm055. [DOI] [PubMed] [Google Scholar]
- 15.Froom P, Shimoni Z. Prediction of hospital mortality rates by admission laboratory tests. Clin Chem. 2006;52(2):325–8. doi: 10.1373/clinchem.2005.059030. [DOI] [PubMed] [Google Scholar]
- 16.Prytherch DR, Sirl JS, Schmidt P, Featherstone PI, Weaver PC, Smith GB. The use of routine laboratory data to predict in-hospital death in medical admissions. Resuscitation. 2005;66(2):203–7. doi: 10.1016/j.resuscitation.2005.02.011. [DOI] [PubMed] [Google Scholar]
- 17.Loekito E, Bailey J, Bellomo R, et al. Common laboratory tests predict imminent death in ward patients. Resuscitation. 2013;84(3):280–5. doi: 10.1016/j.resuscitation.2012.07.025. [DOI] [PubMed] [Google Scholar]
- 18.Pine M, Jones B, Lou YB. Laboratory values improve predictions of hospital mortality. Int J Qual Health Care. 1998;10(6):491–501. doi: 10.1093/intqhc/10.6.491. [DOI] [PubMed] [Google Scholar]
- 19.Cereda E, Klersy C, Hiesmayr M, et al. Body mass index, age and in-hospital mortality: The NutritionDay multinational survey. Clin Nutr. 2017;36(3):839–47. doi: 10.1016/j.clnu.2016.05.001. [DOI] [PubMed] [Google Scholar]
- 20.Lab values interpretation resources [Internet]. [Accessed: June 6, 2019]. Available from:https://c.ymcdn.com/sites/www.acutept.org/resource/resmgr/imported/labvalues.pdf
- 21.Grant SW, Collins GS, Nashef SAM. Statistical primer: developing and validating a risk prediction model. Eur J Cardiothorac Surg. 2018;0:1–6. doi: 10.1093/ejcts/ezy180. [DOI] [PubMed] [Google Scholar]
- 22.Hosmer-Lemeshow goodness of fit test [Internet]. [Accessed: June 6, 2019 ]. Available from: https://www.sealedenvelope.com/stata/hl/
- 23.Measures of Fit for Logistic Regression [Internet]. [Accessed: June 6, 2019 ]. Available from: https://support.sas.com/resources/papers/proceedings14/1485-2014.pdf.
- 24.givitiR package: assessing the calibration of binary outcome models with the GiViTI calibration belt [Internet]. [Accessed: June 6, 2019 ]. Available from: https://cran.r-project.org/web/packages/givitiR/vignettes/givitiR.html.
- 25.Finazzi S, Poole D, Luciani D, Cogo PE, Bertolini G. Calibration belt for quality for care assessment based on dichotomous outcomes. PLoS ONE. 2011;6(2):e16110. doi: 10.1371/journal.pone.0016110. [DOI] [PMC free article] [PubMed] [Google Scholar]