PURPOSE
Machine learning (ML) algorithms that incorporate routinely collected patient-reported outcomes (PROs) alongside electronic health record (EHR) variables may improve prediction of short-term mortality and facilitate earlier supportive and palliative care for patients with cancer.
METHODS
We trained and validated two-phase ML algorithms that incorporated standard PRO assessments alongside approximately 200 routinely collected EHR variables, among patients with medical oncology encounters at a tertiary academic oncology and a community oncology practice.
RESULTS
Among 12,350 patients, 5,870 (47.5%) completed PRO assessments. Compared with EHR- and PRO-only algorithms, the EHR + PRO model improved predictive performance in both tertiary oncology (EHR + PRO v EHR v PRO: area under the curve [AUC] 0.86 [0.85-0.87] v 0.82 [0.81-0.83] v 0.74 [0.74-0.74]) and community oncology (area under the curve 0.89 [0.88-0.90] v 0.86 [0.85-0.88] v 0.77 [0.76-0.79]) practices.
CONCLUSION
Routinely collected PROs contain added prognostic information not captured by an EHR-based ML mortality risk algorithm. Augmenting an EHR-based algorithm with PROs resulted in a more accurate and clinically relevant model, which can facilitate earlier and targeted supportive care for patients with cancer.
INTRODUCTION
For patients with cancer, early supportive care interventions, including serious illness conversations and palliative care, are evidence-based practices that improve quality of life and goal-concordant care.1-4 However, timely identification of patients who may benefit from early supportive care is challenging: Oncology clinicians are often unable to identify patients at risk of 6-month mortality and overestimate life expectancy for up to 70% of their patients.5-8 Interventions on the basis of electronic health record (EHR)–based machine learning (ML) prognostic algorithms increase serious illness conversations and palliative care referral and could lead to more goal-concordant cancer care for patients with cancer.9-19 However, such ML algorithms usually rely on structured EHR data, including laboratories, demographics, and diagnosis codes, which provide limited insight into patient symptoms or functional status.20
CONTEXT
Key Objectives
To train and compare algorithms on the basis of electronic health record (EHR) data alone, patient-reported outcome (PRO) data alone, and EHR plus PRO data, to estimate 6-month risk of mortality among patients seen in routine oncology practice.
Knowledge Generated
Compared with EHR- and PRO-only algorithms, the EHR + PRO model improved predictive performance in both tertiary oncology (EHR + PRO v EHR v PRO: area under the curve 0.86 [0.85-0.87] v 0.82 [0.81-0.83] v 0.74 [0.74-0.74]) and community oncology (area under the curve 0.89 [0.88-0.90] v 0.86 [0.85-0.88] v 0.77 [0.76-0.79]) settings. Performance was superior across patient- and clinician-specific risk threshold preferences and did not result in increased false-positives.
Relevance
Incorporating routinely collected PROs into automated EHR-based mortality prediction algorithms significantly improves performance and may aid in targeting supportive care interventions in oncology.
Patient-reported outcomes (PROs), which are independently associated with mortality,21 may augment such ML algorithms. Routine PRO assessment is now more common and may allow oncology clinicians to better identify patients with high symptom burden or declining functional status.22-26 However, the role of PROs in risk stratification remains unexplored. Incorporating PROs may improve traditional risk stratification tools used for supportive and end-of-life care planning.
In this study, we trained and compared algorithms on the basis of EHR data alone, PRO data alone, and EHR plus PRO data, to estimate 6-month risk of mortality among patients seen in either a large tertiary academic practice, or a community-based general oncology clinic. We hypothesized that adverse PROs would be independently associated with 6-month mortality, and that integrating routinely collected PROs into EHR-based ML algorithms would improve predictive performance compared with ML algorithms on the basis of EHR or PRO data alone in both oncology settings.
METHODS
Data Source
We derived our cohort from patients receiving care at the University of Pennsylvania Health System (UPHS) who were listed in Clarity, an EPIC reporting database, which contains individual electronic medical records for patients containing demographic, comorbidity, and laboratory data. Our study followed the transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD; Data Supplement) checklist for prediction model development and validation.27 We obtained approval and waiver of informed consent from the University of Pennsylvania institutional review board before conducting this study.
Study Population
Our cohort consisted of patients age 18 years or older who had outpatient medical oncology encounters at the Perelman School of Advanced Medicine (PCAM), a large tertiary practice with disease-specific oncology clinics, and Pennsylvania Hospital (PAH), a community oncology practice, between July 1, 2019, and January 1, 2020. We chose patients in these medical oncology clinics because (1) there has been routine collection of PROs in these clinics since mid-2019, (2) an EHR-based ML algorithm has been prospectively validated and implemented in these clinics as part of an initiative to increase serious illness conversations,28,29 and (3) a tertiary academic practice and a community oncology site are representative of the majority of oncology care settings. Details of PRO collection can be found in the Data Supplement. Patients were not required to have received cancer-directed treatment to be included in this study. We excluded patients who had benign hematology or genetics encounters, < 2 encounters during the study period, or no laboratory or comorbidity data within 6 months of the encounter. Our final cohort consisted of 12,350 patients (8,555 at PCAM and 3,795 at PAH); Appendix Figure A1 represents our cohort selection. In all statistical analysis and modeling, we used the first hematology/oncology encounter in the study period for each patient as the index encounter for statistical modeling. We chose not to incorporate PRO data from subsequent encounters because we found that trends in PROs were not meaningfully associated with mortality (Appendix Fig A2).
Features
Features included EHR and PRO data. Our EHR data set included three broad classes of features: (1) demographic variables, (2) comorbidities30; and (3) laboratory data. Our final feature set consisted of approximately 200 variables from the EHR (Appendix Table A1). PRO features were derived from the PRO version of The Common Terminology Criteria for Adverse Events (PRO-CTCAE)31 and the Patient-Reported Outcomes Measurement Information System (PROMIS) Global v.1.232 scales (Appendix Table A2). Further details on features are available in the Data Supplement.
Outcome
The primary outcome was death within 180 days of the index encounter at an oncology practice. We chose 180-day mortality because it is a common indicator of short-term mortality and is often used as a criterion for hospice referral.16 Date of death was derived from the first date of death recorded in either the EHR (Clarity database) or the Social Security Administration Death Master File, matched to UPHS patients by social security number and date of birth.33
Training and Validation Set Split
In the PCAM cohort, the study population was randomly split into a training cohort (70%), in which the mortality risk algorithms were derived, and a validation cohort (30%), in which the algorithms were applied and tested. Patients could not appear in both the training and validation sets. In the PAH cohort, splitting the data set into a training and testing set was not feasible because of the much lower number of cases.
Algorithm Development
To develop an algorithm on the basis of EHR variables alone (EHR algorithm), we fitted a logistic regression model with the adaptive LASSO algorithm to ensure consistent variable selection.34 To develop an algorithm on the basis of PROs alone (PRO algorithm), we fit a logistic regression model where all of the PROs are included as covariates, with observed 180-day mortality as the outcome. To develop an algorithm that includes both EHR and PRO variables (EHR + PRO algorithm), we applied a two-phase method to fit the prediction algorithm and estimate the area under the receiver operating characteristic curve (AUC) and area under the precision-recall curve (AUPRC) that makes full use of all available EHR (N = 8,555 PCAM; N = 3,795 PAH) and PRO (N = 4,677 PCAM; N = 1,193 PAH) data.35-37 Rationale and approaches for deriving the EHR, PRO, and EHR + PROs model are provided in the Data Supplement.
Statistical Analysis
We used descriptive statistics to compare the characteristics of the study population, stratified by whether PROs were collected. All algorithm analyses were performed separately for the PCAM and PAH cohorts using Rstudio software. We first explored correlations among individual PRO features using the aggregated PCAM and PAH data. We then fit logistic regression models with 180-day mortality as the outcome and each PRO as the only covariate. We also fit two-variable logistic regression models that assessed the association between each PRO and mortality, adjusted for the continuous 180-day mortality risk from the EHR algorithm. These exploratory analyses informed independent associations between PROs and mortality and the potential of PROs to augment ML performance.
Then the performance of the three different algorithms (EHR, PRO, and EHR + PRO) was assessed by calculating AUC and AUPRC, our primary performance metrics. True-positive rate (TPR) and false-positive rate at a previously specified 10% risk threshold29 were secondary performance metrics. The 95% CIs for each performance metric were derived using bootstrapping, where each of the two cohorts was repeatedly sampled with replacement to generate 1,000 data sets of the same size. Performance for the EHR model was evaluated for all individuals in the testing set for PCAM (n = 2,566) and in the entire cohort for PAH (n = 3,795). Performance of the PRO and EHR + PRO models were evaluated only for those who had complete PRO data (n = 1,387 for PCAM and n = 1,193 for PAH). Because estimation of the performance metrics for the EHR + PRO algorithm corrected for the potential nonrepresentativeness of the subset of individuals with complete PRO data, the EHR + PRO results are therefore representative of the complete test set. As a sensitivity analysis, we obtained predictive accuracy metrics for all models from the test set for PCAM using only the subset of individuals with PRO data available (n = 1,387). Finally, we conducted a decision curve analysis (see the Data Supplement for methodology) to assess the clinical impact of using each model to identify high-risk patients for the purpose of directing earlier supportive care.38-40
RESULTS
Cohort Description
The study cohort consisted of 8,555 patients who had 50,590 encounters from the tertiary oncology practice and 3,795 patients who had 32,805 encounters from the community oncology practice (median encounters per patient 4, interquartile range 2-7 during study period). The median age was 63.6 years (interquartile range 52.6-72.1 years), 6,989 (56.6%) were female, 8,250 (66.8%) were non-Hispanic White, 2,945 (23.8%) were non-Hispanic Black, 296 (2.4%) were Hispanic, and 6,207 (50.3%) had Medicare insurance. Among patients in the tertiary and community oncology practice cohort, 485 (5.7%) and 122 (3.2%), respectively, died during the 180-day follow-up period. 4,677 (54.7%) patients at the tertiary oncology practice and 1,193 (31.4%) patients at the community oncology practice had completed PRO assessments. Compared with patients who did not complete PRO assessments, patients who completed PRO assessments were more likely to be White (4,289 [73.1%] v 3,961 [61.1%]; P < .001) and have managed care insurance (1,962 [33.4%] v 1,811 [27.9%]; Table 1).
TABLE 1.
Correlation Among PRO Variables
In the combined tertiary and community oncology practice cohorts, decreased performance status was strongly correlated with fatigue (r = 0.69), decreased appetite (r = 0.5), and poorer quality of life (r = 0.58); fatigue was also strongly correlated with poorer quality of life (r = 0.6) and decreased appetite (r = 0.51; Fig 1). Increased anxiety was strongly correlated with increased sadness (r = 0.72). The correlation for all other PRO variable pairs was weak or moderate (r < 0.5). These results were consistent in practice-specific subset analyses (Appendix Figs A3 and A4).
PRO Associations With Observed Mortality
In unadjusted analyses, worse patient-reported performance status (odds ratio [OR], 2.13; 95% CI, 1.90 to 2.39), quality of life (OR, 1.97; 95% CI, 1.72 to 2.23), decreased appetite (OR, 1.89; 95% CI, 1.69 to 2.12), and fatigue (OR, 1.79; 95% CI, 1.61 to 2.00) had the strongest associations with observed mortality (Fig 2). After adjusting for EHR mortality risk, associations between adverse PROs and observed mortality remained significant for performance status, quality of life, fatigue, shortness of breath, anxiety, sadness, constipation, decreased appetite, and nausea (range of adjusted ORs 1.18-1.53; Appendix Table A3). We observed a similar pattern with community oncology practice data, although fewer associations were statistically significant (Appendix Table A4). Adverse PROs were also associated with higher EHR mortality risk for all PROs except rash.
Algorithm Performance
The final EHR + PRO model included the logit of the predicted probabilities from the EHR model, performance status, quality of life, numbness and tingling, and nausea (Appendix Fig A5). For the tertiary oncology practice data, the AUC of the EHR + PRO algorithm (0.86; 95% CI, 0.85 to 0.87) was significantly higher than that of the EHR (0.82; 95% CI, 0.81 to 0.83) and PRO (0.74; 95% CI, 0.73 to 0.75) algorithms (Fig 3A). The AUPRC of the EHR + PRO algorithm (0.36; 95% CI, 0.33 to 0.40) was significantly higher than that of the EHR (0.30; 95% CI, 0.27 to 0.32) and PRO (0.18; 95% CI, 0.17 to 0.20) algorithms (Fig 3B). The TPR of the EHR + PRO algorithm (0.67; 95% CI, 0.64 to 0.71) was significantly higher than that of the EHR (0.61; 95% CI, 0.58 to 0.64) and PRO (0.41; 95% CI, 0.39 to 0.44) algorithms (Fig 3C). There was no difference in false-positive rates among the EHR + PRO (0.12; 95% CI, 0.11 to 0.14), EHR (0.14; 95% CI, 0.13 to 0.15) and PRO (0.11; 95% CI, 0.10 to 0.12) algorithms (Fig 3D). The results were similar in the community oncology practice cohort (Figs 3A-D). In the sensitivity analysis among patients with only PRO data, the EHR + PRO model had consistently higher performance than the PRO model (Appendix Table A5).
Decision Curve Analysis
In both the tertiary and community oncology practice data sets, the decision curve for the EHR + PRO algorithm dominated the decision curves for the EHR and PRO algorithms, indicating that the EHR + PRO algorithm achieves greater clinical utility than the EHR and PRO algorithms regardless of risk preferences (Fig 4). At the clinically relevant mortality risk threshold of 10%, the standardized net benefit was higher for the EHR + PRO algorithm compared with the EHR or PRO algorithms in the tertiary oncology (0.42 [0.37 to 0.47] for EHR + PRO v 0.32 [0.27 to 0.38] for EHR v 0.19 [0.14 to 0.24] for PRO) and community oncology (0.43 [0.37 to 0.48] for EHR + PRO v 0.33 [0.27 to 0.37] for EHR v 0.22 [0.14 to 0.29] for PRO) cohorts.
DISCUSSION
Among patients with cancer treated at both a tertiary cancer center and community oncology practice, ML algorithms on the basis of combined structured EHR and PRO data outperformed algorithms on the basis of EHR or PRO data alone in predicting short-term mortality. Adverse PROs had strong associations with 180-day mortality. Decision curve analysis revealed that the EHR + PRO algorithm was consistently superior when considering patient-specific threshold preferences. Collectively, these findings suggest that routinely collected patient-reported symptoms, quality of life, and performance status have considerable independent prognostic value over and above structured EHR data and augment ML models on the basis of EHR data alone.
Accurate identification of patients at risk of short-term mortality is important in oncology, given guidelines around early palliative care, advance care planning, and serious illness communication for high-risk patients with cancer.41,42 EHR-based ML algorithms linked to automated clinician alerts increase rates of serious illness conversations and palliative care consultation among patients with cancer, with good acceptability among oncologists and no impact on conversation quality.29,43-45 Underidentification of high-risk patients is a barrier, as TPRs are generally below 50% in EHR algorithms. Incorporating PROs could improve accuracy, aid oncologist clinicians' risk assessments, and prompt clinician discussions about goals and end-of-life preferences. On the basis of our results, in a hypothetical population of 1,000 patients, integrating routinely collected PROs with EHR data would correctly identify an additional 60-80 patients at high risk for short-term mortality, compared with using EHR data alone. Routinely collected PROs add value to existing supportive care triggers in outpatient oncology.
Although routine PRO collection is recommended by consensus guidelines for clinical symptom management and toxicity monitoring during clinical trials,24,46 use of routinely collected PROs as part of risk stratification, including prognostic risk stratification, is rare in practice. Prior retrospective studies have found that adverse quality of life and symptoms such as depression, fatigue, and pain are independently associated with poorer survival.21,47,48 However, few studies have demonstrated the independent prognostic value of PROs in contemporary machine learning algorithms. Our study suggests that PROs are only modestly correlated with EHR-predicted mortality risk, and there is likely additional independent prognostic value of PROs that would be of benefit in ML algorithms. Although natural language processing for clinician notes is another potential option to elicit symptoms, there is significant discordance between actual patient-reported symptoms and clinicians' documentation in the EHR.49,50 Relying on routinely collected PROs is likely a better way to capture symptoms to maximally improve performance of predictive algorithms.
A strength of our two-phase methodology is its flexible approach, using PRO data when available and EHR data for all patients. This differs from traditional complete-case analyses, which may not use representative populations, or imputation-based approaches, which would perform poorly in a setting with a high missingness of PROs. Other advantages of this two-phase methodology are detailed in the Data Supplement.
There are several potential limitations to this study. First, although we trained EHR + PRO algorithms across academic and community oncology cohorts, validation across other institutions, including those with greater Hispanic representation, would be valuable. However, the EHR features used in our models are all commonly available in structured data fields in all health system EHRs, and the PRO features we used were based on standard instruments using Likert scale values. We did not derive any features from unstructured data, and thus, we would not expect semantic differences in coding between different systems. Nevertheless, our approach should be externally validated as other issues of data quality, including completeness of features and heterogeneity in coding practices between institutions, are well known and should be accounted for.51,52 Second, we did not validate a specific model in an external institutional cohort, but rather used a two-phase approach to test similar algorithms in two unique practices. This approach is justified because the purpose of our study was not to validate a specific algorithm, but rather to validate the conclusion that integrating PROs into routine prognostic algorithms improves risk prediction. Third, there is no gold standard reference for mortality prediction, and it is unclear how our EHR + PRO model compares with other published mortality prediction tools. However, we used the same features used in a validated EHR algorithm that is in routine use in medical oncology practices within our cancer center to prompt serious illness conversations.28 Fourth, although we expect that our institutional registry and Social Security death data captured most deaths, we were unable to use more robust death data including National Death Index and obituary data. Fifth, although algorithm performance in our PCAM sample is reported on a typical holdout test set, we were unable to use a train/test split using the PAH data because of the much smaller number of cases in that data set.
In conclusion, among 12,350 patients with cancer seen in tertiary and community oncology practices, ML algorithms to predict short-term mortality that integrated routinely collected patient-reported outcomes with electronic health record features significantly improved predictive performance, compared with algorithms on the basis of EHR or PRO data alone. Our findings suggest that PROs can significantly improve performance of predictive algorithms in oncology.
ACKNOWLEDGMENT
The authors thank Jay Fein for assistance with manuscript preparation.
APPENDIX
TABLE A1.
TABLE A2.
TABLE A3.
TABLE A4.
TABLE A5.
Ravi B. Parikh
Stock and Other Ownership Interests: Merck, Google, GNS Healthcare, Onc.AI
Consulting or Advisory Role: GNS Healthcare, Cancer Study Group, Onc.AI, Thyme Care, Humana, NanOlogy, Merck
Research Funding: Humana
Patents, Royalties, Other Intellectual Property: Technology to integrate patient-reported outcomes into electronic health record algorithms
Travel, Accommodations, Expenses: The Oncology Institute of Hope and Innovation
Jill S. Hasler
Patents, Royalties, Other Intellectual Property: Patent currently pending for machine learning systems using electronic health record data and patient-reported outcomes
William Ferrell
Research Funding: Humana
Peter E. Gabriel
Travel, Accommodations, Expenses: Varian Medical Systems
Justin E. Bekelman
Stock and Other Ownership Interests: Reimagine Care
Honoraria: National Comprehensive Cancer Network
Consulting or Advisory Role: UnitedHealthcare, Reimagine Care
No other potential conflicts of interest were reported.
DISCLAIMER
The funders had no role in the design and conduct of the study; collection, management, analysis, and interpretation of the data; preparation, review, or approval of the manuscript; and decision to submit the manuscript for publication.
SUPPORT
Supported by the National Cancer Institute (R35197461) C.L., National Cancer Institute (K08CA263541) R.B.P., National Cancer Institute (P30-014089) C.L., NHLBI (R01-HL138306) J.C., National Cancer Institute (R01- CA236468) Jinbo Chen, NIGMS (R01-GM140463) J.C., and National Center for Advancing Translational Sciences (UL1-TR001878) J.C.
R.B.P. and J.S.H. contributed equally to this work as co-first authors.
C.L., J.E.B., and J.C. contributed equally to this work as co-senior authors.
DATA SHARING STATEMENT
The data that support the findings of this study are available on request.
All statistical analysis was performed in R version 3.6.0. The validation for the EHR algorithm has been previous published (https://jamanetwork.com/journals/jamaoncology/article-abstract/2770698), and source code is available at https://github.com/pennsignals/eol-onc.
AUTHOR CONTRIBUTIONS
Conception and design: Ravi B. Parikh, Peter E. Gabriel, Caryn Lerman, Justin E. Bekelman, Jinbo Chen
Financial support: Ravi B. Parikh, Caryn Lerman, Justin E. Bekelman, Jinbo Chen
Administrative support: Ravi B. Parikh, William Ferrell
Provision of study materials or patients: Ravi B. Parikh, Jinbo Chen
Collection and assembly of data: Ravi B. Parikh, Yichen Zhang, Corey Chivers, William Ferrell, Peter E. Gabriel, Jinbo Chen
Data analysis and interpretation: Ravi B. Parikh, Jill S. Hasler, Yichen Zhang, Manqing Liu, Corey Chivers, Caryn Lerman, Justin E. Bekelman, Jinbo Chen
Manuscript writing: All authors
Final approval of manuscript: All authors
Accountable for all aspects of the work: All authors
AUTHORS' DISCLOSURES OF POTENTIAL CONFLICTS OF INTEREST
The following represents disclosure information provided by authors of this manuscript. All relationships are considered compensated unless otherwise noted. Relationships are self-held unless noted. I = Immediate Family Member, Inst = My Institution. Relationships may not relate to the subject matter of this manuscript. For more information about ASCO's conflict of interest policy, please refer to www.asco.org/rwc or ascopubs.org/cci/author-center.
Open Payments is a public database containing information reported by companies about payments made to US-licensed physicians (Open Payments).
Ravi B. Parikh
Stock and Other Ownership Interests: Merck, Google, GNS Healthcare, Onc.AI
Consulting or Advisory Role: GNS Healthcare, Cancer Study Group, Onc.AI, Thyme Care, Humana, NanOlogy, Merck
Research Funding: Humana
Patents, Royalties, Other Intellectual Property: Technology to integrate patient-reported outcomes into electronic health record algorithms
Travel, Accommodations, Expenses: The Oncology Institute of Hope and Innovation
Jill S. Hasler
Patents, Royalties, Other Intellectual Property: Patent currently pending for machine learning systems using electronic health record data and patient-reported outcomes
William Ferrell
Research Funding: Humana
Peter E. Gabriel
Travel, Accommodations, Expenses: Varian Medical Systems
Justin E. Bekelman
Stock and Other Ownership Interests: Reimagine Care
Honoraria: National Comprehensive Cancer Network
Consulting or Advisory Role: UnitedHealthcare, Reimagine Care
No other potential conflicts of interest were reported.
REFERENCES
- 1. Bernacki R, Paladino J, Neville BA, et al. Effect of the serious illness care program in outpatient oncology: A cluster randomized clinical trial. JAMA Intern Med. 2019;179:751–759. doi: 10.1001/jamainternmed.2019.0077. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Paladino J, Bernacki R, Neville BA, et al. Evaluating an intervention to improve communication between oncology clinicians and patients with life-limiting cancer: A cluster randomized clinical trial of the serious illness care program. JAMA Oncol. 2019;5:801–809. doi: 10.1001/jamaoncol.2019.0292. [DOI] [PubMed] [Google Scholar]
- 3. Temel JS, Greer JA, Muzikansky A, et al. Early palliative care for patients with metastatic non–small-cell lung cancer. N Engl J Med. 2010;363:733–742. doi: 10.1056/NEJMoa1000678. [DOI] [PubMed] [Google Scholar]
- 4. Wright AA. Associations between end-of-life discussions, patient mental health, medical care near death, and caregiver bereavement adjustment. JAMA. 2008;300:1665–1673. doi: 10.1001/jama.300.14.1665. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Krishnan M, Temel J, Wright A, et al. Predicting life expectancy in patients with advanced incurable cancer: A review. J Support Oncol. 2013;11:68–74. doi: 10.12788/j.suponc.0004. [DOI] [PubMed] [Google Scholar]
- 6. Chow E, Harth T, Hruby G, et al. How accurate are physicians’ clinical predictions of survival and the available prognostic tools in estimating survival times in terminally III cancer patients? A systematic review. Clin Oncol. 2001;13:209–218. doi: 10.1053/clon.2001.9256. [DOI] [PubMed] [Google Scholar]
- 7. White N, Reid F, Harris A, et al. A systematic review of predictions of survival in palliative care: How accurate are clinicians and who are the experts? PLoS One. 2016;11:e0161407. doi: 10.1371/journal.pone.0161407. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Christakis NA. Extent and determinants of error in doctors’ prognoses in terminally ill patients: Prospective cohort study commentary: Why do doctors overestimate? Commentary: Prognoses should be based on proved indices not intuition. BMJ. 2000;320:469–473. doi: 10.1136/bmj.320.7233.469. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Tang ST, Chen CH, Wen FH, et al. Accurate prognostic awareness facilitates, whereas better quality of life and more anxiety symptoms hinder end-of-life care discussions: A longitudinal survey study in terminally ill cancer patients’ last six months of life. J Pain Symptom Manage. 2018;55:1068–1076. doi: 10.1016/j.jpainsymman.2017.12.485. [DOI] [PubMed] [Google Scholar]
- 10. Nipp RD, Greer JA, El-Jawahri A, et al. Coping and prognostic awareness in patients with advanced cancer. J Clin Oncol. 2017;35:2551–2557. doi: 10.1200/JCO.2016.71.3404. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Lundquist G, Rasmussen BH, Axelsson B. Information of imminent death or not: Does it make a difference? J Clin Oncol. 2011;29:3927–3931. doi: 10.1200/JCO.2011.34.6247. [DOI] [PubMed] [Google Scholar]
- 12. Zhang B, Wright AA, Huskamp HA, et al. Health care costs in the last week of life: Associations with end-of-life conversations. Arch Intern Med. 2009;169:480–488. doi: 10.1001/archinternmed.2008.587. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. El-Jawahri A, Traeger L, Park ER, et al. Associations among prognostic understanding, quality of life, and mood in patients with advanced cancer: Prognosis, QoL, and Mood in Advanced CA. Cancer. 2014;120:278–285. doi: 10.1002/cncr.28369. [DOI] [PubMed] [Google Scholar]
- 14. Finlay E, Casarett D. Making difficult discussions easier: Using prognosis to facilitate transitions to hospice. CA Cancer J Clin. 2009;59:250–263. doi: 10.3322/caac.20022. [DOI] [PubMed] [Google Scholar]
- 15. Parikh RB, Manz C, Chivers C, et al. Machine learning approaches to predict 6-month mortality among patients with cancer. JAMA Netw Open. 2019;2:e1915997. doi: 10.1001/jamanetworkopen.2019.15997. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Elfiky AA, Pany MJ, Parikh RB, Obermeyer Z. Development and application of a machine learning approach to assess short-term mortality risk among patients with cancer starting chemotherapy. JAMA Netw Open. 2018;1:e180926. doi: 10.1001/jamanetworkopen.2018.0926. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Bertsimas D, Dunn J, Pawlowski C, et al. Applied informatics decision support tool for mortality predictions in patients with cancer. JCO Clin Cancer Inform. 2018;2:1–11. doi: 10.1200/CCI.18.00003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Gensheimer MF, Henry AS, Wood DJ, et al. Automated survival prediction in metastatic cancer patients using high-dimensional electronic medical record data. J Natl Cancer Inst. 2019;111:568–574. doi: 10.1093/jnci/djy178. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Gensheimer MF, Aggarwal S, Benson KR, et al. Automated model versus treating physician for predicting survival time of patients with metastatic cancer. J Am Med Inform Assoc. 2021;28:1108–1116. doi: 10.1093/jamia/ocaa290. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Baum LVM, Friedman D. The uncertain science of predicting death. JAMA Netw Open. 2020;3:e201736. doi: 10.1001/jamanetworkopen.2020.1736. [DOI] [PubMed] [Google Scholar]
- 21. Stukenborg GJ, Blackhall LJ, Harrison JH, et al. Longitudinal patterns of cancer patient reported outcomes in end of life care predict survival. Support Care Cancer. 2016;24:2217–2224. doi: 10.1007/s00520-015-3024-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Bubis LD, Davis L, Mahar A, et al. Symptom burden in the first year after cancer diagnosis: An analysis of patient-reported outcomes. J Clin Oncol. 2018;36:1103–1111. doi: 10.1200/JCO.2017.76.0876. [DOI] [PubMed] [Google Scholar]
- 23. Basch E, Deal AM, Kris MG, et al. Symptom monitoring with patient-reported outcomes during routine cancer treatment: A randomized controlled trial. J Clin Oncol. 2016;34:557–565. doi: 10.1200/JCO.2015.63.0830. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Basch E, Barbera L, Kerrigan CL, Velikova G. Implementation of patient-reported outcomes in routine medical care. Am Soc Clin Oncol Ed Book. 2018;38:122–134. doi: 10.1200/EDBK_200383. [DOI] [PubMed] [Google Scholar]
- 25. Yang LY, Manhas DS, Howard AF, Olson RA. Patient-reported outcome use in oncology: A systematic review of the impact on patient-clinician communication. Support Care Cancer. 2018;26:41–60. doi: 10.1007/s00520-017-3865-7. [DOI] [PubMed] [Google Scholar]
- 26. Basch E, Deal AM, Dueck AC, et al. Overall survival results of a trial assessing patient-reported outcomes for symptom monitoring during routine cancer treatment. JAMA. 2017;318:197–198. doi: 10.1001/jama.2017.7156. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Collins GS, Reitsma JB, Altman DG, et al. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): The TRIPOD statement. BMJ. 2015;350:g7594. doi: 10.1136/bmj.g7594. [DOI] [PubMed] [Google Scholar]
- 28. Manz CR, Chen J, Liu M, et al. Validation of a machine learning algorithm to predict 180-day mortality for outpatients with cancer. JAMA Oncol. 2020;6:1723–1730. doi: 10.1001/jamaoncol.2020.4331. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Manz CR, Parikh RB, Small DS, et al. Effect of integrating machine learning mortality estimates with behavioral nudges to clinicians on serious illness conversations among patients with cancer: A stepped-wedge cluster randomized clinical trial. JAMA Oncol. 2020;6:e204759. doi: 10.1001/jamaoncol.2020.4759. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data. Med Care. 1998;36:8–27. doi: 10.1097/00005650-199801000-00004. [DOI] [PubMed] [Google Scholar]
- 31.Patient-Reported Outcomes version of the Common Terminology Criteria for Adverse Events (PRO-CTCAE) https://healthcaredelivery.cancer.gov/pro-ctcae/ [Google Scholar]
- 32. Hays RD, Bjorner JB, Revicki DA, et al. Development of physical and mental health summary scores from the patient-reported outcomes measurement information system (PROMIS) global items. Qual Life Res. 2009;18:873–880. doi: 10.1007/s11136-009-9496-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.NTIS . Limited Access Death Master File Download. https://dmf.ntis.gov/ [Google Scholar]
- 34.Friedman J, Hastie T, Tibshirani R, et al. glmnet: Lasso and Elastic-Net Regularized Generalized Linear Models. 2021. https://dx.doi.org/10.18637/jss.v033.i01 [Google Scholar]
- 35. Breslow NE, Cain KC. Logistic regression for two-stage case-control data. Biometrika. 1988;75:11–20. [Google Scholar]
- 36. Cain KC, Breslow NE. Logistic regression analysis and efficient design for two-stage studies. Am J Epidemiol. 1988;128:1198–1206. doi: 10.1093/oxfordjournals.aje.a115074. [DOI] [PubMed] [Google Scholar]
- 37. Scott AJ, Wild CJ. Fitting logistic models under case-control or choice based sampling. J R Stat Soc Series B Stat Methodol. 1986;48:170–182. [Google Scholar]
- 38. Vickers AJ, Elkin EB. Decision curve analysis: A novel method for evaluating prediction models. Med Decis Making. 2006;26:565–574. doi: 10.1177/0272989X06295361. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Fitzgerald M, Saville BR, Lewis RJ. Decision curve analysis. JAMA. 2015;313:409–410. doi: 10.1001/jama.2015.37. [DOI] [PubMed] [Google Scholar]
- 40. Kerr KF, Brown MD, Zhu K, Janes H. Assessing the clinical impact of risk prediction models with decision curves: Guidance for correct interpretation and appropriate use. J Clin Oncol. 2016;34:2534–2540. doi: 10.1200/JCO.2015.65.5654. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Ferrell BR, Temel JS, Temin S, et al. Integration of palliative care into standard oncology care: American Society of Clinical Oncology clinical practice guideline update. J Clin Oncol. 2017;35:96–112. doi: 10.1200/JCO.2016.70.1474. [DOI] [PubMed] [Google Scholar]
- 42. Temel JS, Greer JA, El-Jawahri A, et al. Effects of early integrated palliative care in patients with lung and GI cancer: A randomized clinical trial. J Clin Oncol. 2017;35:834–841. doi: 10.1200/JCO.2016.70.5046. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Courtright KR, Chivers C, Becker M, et al. Electronic health record mortality prediction model for targeted palliative care among hospitalized medical patients: A pilot quasi-experimental study. J Gen Intern Med. 2019;34:1841–1847. doi: 10.1007/s11606-019-05169-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Parikh RB, Manz CR, Nelson MN, et al. Clinician perspectives on machine learning prognostic algorithms in the routine care of patients with cancer: A qualitative study. Support Care Cancer. 2022;30:4363–4372. doi: 10.1007/s00520-021-06774-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Li E, Manz C, Liu M, et al. Oncologist phenotypes and associations with response to a machine learning-based intervention to increase advance care planning: Secondary analysis of a randomized clinical trial. PLoS One. 2022;17:e0267012. doi: 10.1371/journal.pone.0267012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Basch E, Snyder C, McNiff K, et al. Patient-reported outcome performance measures in oncology. JCO Oncol Pract. 2014;10:209–211. doi: 10.1200/JOP.2014.001423. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47. Saint-Maurice PF, Troiano RP, Bassett DR, et al. Association of daily step count and step intensity with mortality among US adults. JAMA. 2020;323:1151–1160. doi: 10.1001/jama.2020.1382. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Seventer EE, Fish MG, Fosbenner K, et al. Associations of baseline patient-reported outcomes with treatment outcomes in advanced gastrointestinal cancer. Cancer. 2021;127:619–627. doi: 10.1002/cncr.33315. [DOI] [PubMed] [Google Scholar]
- 49. Pakhomov SV, Jacobsen SJ, Chute CG, Roger VL. Agreement between patient-reported symptoms and their documentation in the medical record. Am J Manag Care. 2008;14:530–539. [PMC free article] [PubMed] [Google Scholar]
- 50. Yeung AR, Pugh SL, Klopp AH, et al. Improvement in patient-reported outcomes with intensity-modulated radiotherapy (RT) compared with standard RT: A report from the NRG Oncology RTOG 1203 study. J Clin Oncol. 2020;38:1685–1692. doi: 10.1200/JCO.19.02381. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51. Fu S, Wen A, Schaeferle GM, et al. Assessment of data quality variability across two EHR systems through a case study of post-surgical complications. AMIA Annu Symp Proc AMIA Symp. 2022;2022:196–205. [PMC free article] [PubMed] [Google Scholar]
- 52. Horsky J, Drucker EA, Ramelson HZ. Accuracy and completeness of clinical coding using ICD-10 for ambulatory visits. AMIA Annu Symp Proc AMIA Symp. 2017;2017:912–920. [PMC free article] [PubMed] [Google Scholar]
- 53.Chivers, C: pennsignals/eol-onc. https://github.com/pennsignals/eol-onc [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data that support the findings of this study are available on request.
All statistical analysis was performed in R version 3.6.0. The validation for the EHR algorithm has been previous published (https://jamanetwork.com/journals/jamaoncology/article-abstract/2770698), and source code is available at https://github.com/pennsignals/eol-onc.