Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2022 Jan 1.
Published in final edited form as: J Pain Symptom Manage. 2020 Aug 25;61(1):136–142.e2. doi: 10.1016/j.jpainsymman.2020.08.024

Identifying Goals of Care Conversations in the Electronic Health Record Using Natural Language Processing and Machine Learning

Robert Y Lee 1,2, Lyndia C Brumback 1,3, William B Lober 1,4,5, James Sibley 1,4,5, Elizabeth L Nielsen 1,2, Patsy D Treece 1,2,4, Erin K Kross 1,2, Elizabeth T Loggers 1,8,9, James A Fausto 1,7, Charlotta Lindvall 10, Ruth A Engelberg 1,2, J Randall Curtis 1,2,4,6
PMCID: PMC7769906  NIHMSID: NIHMS1623210  PMID: 32858164

Abstract

Context

Goals-of-care discussions are an important quality metric in palliative care. However, goals-of-care discussions are often documented as free text in diverse locations. It is difficult to identify these discussions in the electronic health record (EHR) efficiently.

Objectives

To develop, train, and test an automated approach to identifying goals-of-care discussions in the EHR, using natural language processing (NLP) and machine learning (ML).

Methods

From the electronic health records of an academic health system, we collected a purposive sample of 3,183 EHR notes (1,435 inpatient notes and 1,748 outpatient notes) from 1,426 patients with serious illness over 2008–2016, and manually reviewed each note for documentation of goals-of-care discussions. Separately, we developed a program to identify notes containing documentation of goals-of-care discussions using NLP and supervised ML. We estimated the performance characteristics of the NLP/ML program across 100 pairs of randomly partitioned training and test sets. We repeated these methods for inpatient-only and outpatient-only subsets.

Results

Of 3,183 notes, 689 contained documentation of goals-of-care discussions. The mean sensitivity of the NLP/ML program was 82.3% (SD 3.2%), and the mean specificity was 97.4% (SD 0.7%). NLP/ML results had a median positive likelihood ratio of 32.2 (IQR 27.5–39.2) and a median negative likelihood ratio of 0.18 (IQR 0.16–0.20). Performance was better in inpatient-only samples than outpatient-only samples.

Conclusion

Using NLP and ML techniques, we developed a novel approach to identifying goals-of-care discussions in the EHR. NLP and ML represent a potential approach toward measuring goals-of-care discussions as a research outcome and quality metric.

Keywords: natural language processing, machine learning, goals of care, electronic health record, quality improvement, medical informatics

Introduction

Communication about goals of care is an important aspect of palliative and end-of-life care, and has been associated with improved patient and family outcomes as well as reduced intensity of treatment near the end of life (16). However, despite the importance of such communication, discussion and documentation of goals of care remains inadequate in many settings (1, 711). Wide-ranging deficiencies in timely communication and documentation of goals of care have been reported in both inpatient and outpatient settings (2, 1113). Even when goals-of-care discussions result in completion of an advance care planning document, lack of standardization in how such documentation is stored in the electronic health record (EHR) poses a barrier to communicating patients’ goals across the continuum of care (14). To promote the implementation of goals-of-care discussions for all patients with serious illness, it is important to be able to readily identify the timing and content of goals-of-care discussions (1518). However, an efficient means of identifying goals-of-care discussions in the EHR for patients with serious illness does not yet exist.

Natural language processing (NLP) methods enable computers to process and analyze textual data written in human languages (19). Although early NLP approaches were constrained to using rule-based methods developed iteratively by humans to classify textual data into understandable categories, the advent of machine learning (ML) methods that rely on automated computational techniques to classify data have allowed NLP to extract and measure more complex constructs (19). Inductive supervised machine learning is a family of ML techniques that utilizes statistical methods to derive rules that best explain the classification of training data—i.e., data independently labelled to reflect outcomes. These rules may subsequently be used to predict the classification of new data (20).

NLP and ML have been used to automatically extract a wide variety of data from unstructured clinical records (21, 22). While many clinical NLP efforts have been limited to rule-based measures of diagnoses, symptoms, and clinical events (2328), more recent efforts have incorporated ML techniques to infer diagnoses of cerebral aneurysms (29), measure patient-reported symptoms in patients with cancer (30), and enhance predictive modeling of mortality and long ICU stays among critically ill patients (31). In palliative care, rule-based NLP algorithms have been used to measure quality indicators and goals-of-care discussions for patients undergoing palliative surgery (32, 33), outpatients with COPD (34), and patients who died of cancer (35, 36); and, deep learning algorithms have been used to measure the documentation of patient preferences and communication among patients admitted to the intensive care unit (37, 38). However, additional work is needed to be able to accurately and efficiently identify documentation of goals-of-care discussions in the EHR across the continuum of serious illness. In this report, we describe the development and preliminary performance characteristics of a novel approach using NLP/ML to identify goals-of-care discussions for outpatients and inpatients with serious illness.

Methods

Data Sources

We assembled a retrospective sample containing clinical notes gathered from inpatients and outpatients with serious illness at the two teaching hospitals and affiliated clinics of UW Medicine, a large academic health system. Notes were collected from multiple sources: (a) all notes from participants in two randomized controlled trials (RCTs) of palliative care interventions, which had been previously abstracted for documentation of goals-of-care discussions (39, 40); (b) a convenience sample of notes written by palliative care specialists who had designated each note as containing, or not containing, goals-of-care discussions for the purpose of clinical service metrics; and, (c) clinical notes from a random sample of patients with serious illness (41, 42) (Table 1). We purposively selected these sources of EHR notes to enrich our data for notes containing goals-of-care discussions, which constitute a small proportion of randomly sampled EHR notes. All notes were extracted from the UW Medicine clinical EHR platforms (Cerner Millennium, Cerner Corp., North Kansas City, MO; and, EpicCare, Epic Systems Corp., Madison, WI) as plain text. The study was conducted as part of a quality improvement initiative to measure the prevalence of documented goals-of-care discussions for patients with serious illness at UW Medicine.

Table 1.

Composition of sample dataset.

Data source Determination of goals-of-care discussions Notes with goals-of-care discussion, n Notes without goals-of-care discussion, n Total notes, n
Outpatient clinical notes written by non-palliative providers for 537 patients with serious illness enrolled in RCT of communication-priming intervention, 2012–2016 (39) Abstraction of any discussions of advance care planning, referral to hospice, or referral to palliative care 212 1,070 1,282
Inpatient clinical notes for 168 ICU patients enrolled in RCT of communication facilitators, 2008–2013 (40) Abstraction of textual documentation of goals-of-care discussions 53 a 53
Clinical notes from 421 patients evaluated by palliative care specialists, 2012–2017 b Goals-of-care discussions indicated in structured quality-improvement metrics; confirmed by investigator review of subset of 139 notes 424 31 455
Clinical notes from random sample of 300 patients with serious illness who receive care through UW Medicine, 2014–2016 c Abstraction of textual documentation of goals-of-care discussions 0 d 1,393 1,393
Total 689 2,494 3,183
a

Goals-of-care-negative notes for this cohort were not recorded during this trial.

b

Contains 446 inpatient notes and 9 outpatient notes.

c

Criteria: age ≥ 18 years, cared for at UW Medicine between May 1, 2014 and August 1, 2016; with one or more of the nine Dartmouth Atlas chronic health conditions by ICD coding (41, 42); and receiving regular care through UW Medicine as defined by at least one non-surgical inpatient visit at a UW Medicine facility in the preceding two years, or two outpatient visits at the same UW Medicine facility in the preceding 32 months with at least one visit in the preceding two years. Contains 936 inpatient notes, and 457 outpatient notes.

d

Two-phase abstraction of this dataset identified no goals-of-care discussions.

Outcome Assessment and Labelling

Toward the goal of measuring conversations about goals of care, we adopted an operational definition of goals-of-care documentation as a provider note including any of the following: (1) advance care planning activities (discussion of values, goals, and preferences in consideration of future care), (2) completion of advance directives or Physician Order for Life-Sustaining Treatment (POLST) forms, or (3) referral to hospice or subspecialty palliative care services (Supplemental Appendix 1). (Subspecialty palliative care consultations in our system are most frequently requested for elucidation of goals-of-care). For the 1,335 notes from RCTs (39, 40), we evaluated the abstraction protocols from the original trials and mapped previously coded elements to this study’s definition of goals-of-care documentation, with input from the principal abstractor in both trials (PDT). For the 455 palliative care specialist notes, note authors designated each note as containing or not containing documented goals-of-care discussions at the time of note entry using a structured data collection instrument attached to each note. For the 1,393 notes collected from a random sample of patients with serious illness, two research coordinators abstracted each note using a similar definition of goals-of-care discussions, which identified no goals-of-care discussions. Because notes from the sample of palliative care specialist notes and from the random sample of patients with serious illness were not initially abstracted using the same protocols as the RCTs, two study investigators (PDT, RYL) re-abstracted 239 notes sampled from both corpuses using a study-specific protocol adapted from one of the RCTs (39) (Supplemental Appendix 1), finding excellent agreement between these notes’ initial designations and subsequent re-abstraction (94% agreement, kappa 0.87).

Natural Language Processing and Machine Learning

Following preprocessing of notes to remove non-textual templated content, we randomly partitioned the dataset into 100 pairs of training sets and test sets. Each training set consisted of a random sampling of 80% of the notes in the dataset, and corresponded to a test set consisting of the remaining 20% of notes in the dataset. We used NLP to tokenize each note into unigrams (i.e. tokens of 1 word length), excluding common stop words and negation terms. For each training set, we then trained an inductive supervised ML classifier, using the scikit-learn regularized logistic regression classifier with an L2 penalty and the default penalty parameter of 1, with a training feature set consisting of the count of each NLP unigram in the training set (43). Subsequently, we applied each trained ML classifier to its corresponding test set to compute the predicted probability of goals-of-care documentation for each note in the test set. A predicted probability of greater than (or less than) 0.5 was used to classify test notes as having (or not having) goals-of-care documentation for the purpose of estimating the sensitivity, specificity, and likelihood ratios of the NLP/ML classifier as compared to the clinician-abstracted gold-standard. We also summarized the performance of the NLP/ML classifier by the area under the receiver operating curve (AUC). This process was repeated for each of the 100 randomly partitioned sample sets. We reported summary statistics of performance metrics observed across all 100 random partitions.

To explore NLP/ML performance characteristics in the inpatient and outpatient settings, we separated the sample set into inpatient-only and outpatient-only sample sets, and then proceeded to train and test the NLP/ML program within these subsets using the same approach.

Software

All programming was performed in Python v3.6 (Python Software Foundation, python.org). Natural language processing was performed using the Natural Language Toolkit v3.2.4 (NLTK Project, nltk.org) (44), and machine learning was implemented using the scikit-learn library v0.19.0 (scikit-learn developers, scikit-learn.org) (43).

Results

The composition of the sample set is shown in Table 1. Overall, we identified 689 goals-of-care-positive notes across all four data sources, with 265 identified from prior RCTs, and 424 identified by note authors and confirmed by manual review. The remaining 2,494 notes in the dataset, including all of the 1,393 notes collected from a random sample of 300 seriously ill patients over a 27-month period, did not contain goals-of-care discussions. The sample set contained 1,435 inpatient notes, of which 468 contained goals-of-care discussions; and 1,748 outpatient notes, of which 221 contained goals-of-care discussions (Table 2).

Table 2.

Performance characteristics of NLP/ML by note type.

Note types Sensitivity Specificity LR+ LR− AUC
(note counts) Mean (SD) Mean (SD) Median (IQR) Median (IQR) Mean (SD)
All notes 82.3% (3.2%) 97.4% (0.7%) 32.2 (27.5–39.2) 0.18 (0.16–0.20) 0.943 (0.015)
Inpatient notes (468 GOC+, 967 GOC−) 97.1% (1.8%) 99.4% (0.6%) 194.7 (175.7–∞) a 0.03 (0.02–0.04) 0.996 (0.005)
Outpatient notes (221 GOC+, 1527 GOC−) 52.0% (6.7%) 96.8% (1.2%) 17.7 (12.0–22.5) 0.51 (0.45–0.54) 0.826 (0.040)
a

In the analysis of inpatient notes, 47 of 100 random splits resulted in 100% specific test-set performance, leading to an LR+ estimate of infinity.

Over 100 random partitions of the sample set, the observed sensitivity of NLP/ML ranged from 72–93%, with a mean sensitivity of 82.3% (SD 3.2%); and, the observed specificity of NLP/ML ranged from 95–99%, with a mean specificity of 97.4% (SD 0.7%). NLP/ML classifiers had a median positive likelihood ratio (LR+) of 32.2 (IQR 27.5–39.2, range 17.9–81.0), and a median negative likelihood ratio (LR–) of 0.18 (IQR 0.16–0.20, range 0.07–0.30). The AUC ranged from 0.91–0.98, with a mean AUC of 0.94 (SD 0.01). NLP/ML classifiers specific to inpatient records had better test performance compared to NLP/ML classifiers that were specific to outpatient records (mean sensitivity 97.1% vs. 52.0%; mean specificity 99.4% vs. 96.8%) (Table 2).

Discussion

By applying NLP/ML to a sample of inpatient and outpatient EHR notes, we were able to develop and train an automated program that identified documented goals-of-care discussions. While NLP techniques have been efficacious in identifying a variety of distinct clinical entities (2330), identifying goals-of-care discussions—a broad construct most often represented by descriptive free-text rather than precise signifiers—poses unique challenges to machine approaches (45). Most published studies that report the use of NLP to identify goals-of-care discussions are rule-based and focus on specific diseases or care contexts (30, 3236); two studies specifically examining ICU patients’ records have used ML to identify goals-of-care discussions (37, 38). Our work demonstrates the feasibility of NLP/ML as a potential approach for identifying goals-of-care discussions in the EHR across a broad range of clinical contexts. Additionally, our findings contribute to the growing application of NLP/ML for extracting increasingly complex constructs from the EHR (4648).

Stakeholders in palliative care research have given high priority to metrics that identify discussions between patients, families and providers about goals of care and advance care planning (15, 16, 49). Although there has been progress in measuring advance directive documentation—a metric rife with its own challenges (14, 34)—doing so ultimately falls short of measuring documented goals-of-care conversations, which are often held outside the context of advance directives (911, 15). Our approach of using NLP/ML to identify documentation of goals-of-care discussions brings the palliative care research community closer to the goal of quantifying and characterizing goals-of-care discussions across diverse clinical settings. Potential applications of this approach include the identification of target populations for palliative care interventions who have not had documented goals-of-care discussions, measuring outcomes for interventions designed to increase goals-of-care discussions, providing data to guide and evaluate quality improvement programs to improve communication, and developing clinician-facing EHR tools to assist clinicians in locating prior goals-of-care documentation (14). Each of these potential applications may prioritize different performance characteristics to maximize utility. For example, programs that aim to locate goals-of-care notes for point-of-care clinicians might prioritize high positive predictive value to avoid wasting clinicians’ time, while programs to identify potential study participants without prior goals-of-care discussions might prioritize high negative predictive value to avoid excluding eligible patients.

We encountered challenges in deciding what constituted documentation of a goals-of-care discussion. We elected to adopt a broad-ranging definition that encompassed many common elements of advance care planning encountered during prior clinical trials of patients with serious illness (39, 40). We also elected to exclude stand-alone documentation of code status from our definition, as these data often do not reflect the conversations we aimed to measure. Because the appropriate definition of a goals-of-care discussion differs over the research question being asked, we propose that future approaches adopt multidimensional definitions that specify different elements and types of goals-of-care documentation in order to maximize the adaptability of the end product.

Notably, in the entirety of the most generalized corpus of notes sampled—consisting of 1,393 notes authored over a 27-month period for a random sample of 300 patients with serious illness—we found no notes containing goals-of-care discussions (Table 1). This contrasted with the prevalence of goals-of-care discussions among patients enrolled in randomized trials of communication interventions (39, 40), and suggests that HER-documented goals-of-care discussions are quite rare even among patients with serious illness. This finding highlights both a clinical opportunity for improvement, and the utility of an automated approach to identifying goals-of-care documentation.

Our study has several limitations. First, our data were purposively sampled to enrich for goals-of-care documentation, and are not representative of EHR notes from a random sample of patients with serious illness. Although sensitivity, specificity, likelihood ratios, and AUC are often assumed to be inherent characteristics of the test, our observed performance is likely to be better than expected real-world performance in the general EHR. Second, because the dataset was purposively sampled, we are unable to report prevalence-dependent performance characteristics such as positive or negative predictive values. Owing to the low prevalence of goals-of-care discussions in the EHR, even algorithms with very high positive likelihood ratios may not achieve satisfactory positive predictive value. Third, our algorithm may have selected predictors unlikely to be associated with goals-of-care documentation in other samples or settings, leading to overly optimistic estimates of performance. Because our training data did not specifically identify which words or phrases of a given note reflected goals-of-care documentation, the algorithm may have selected predictors whose association with the outcome was specific to our data. Additionally, because our training data incorporated notes from four unique data sources, the algorithm may have selected predictors (e.g. predictors of RCT participation) whose association with goals-of-care documentation is mediated by the data source. Future studies should consider employing word- or phrase-level annotation of training data (e.g. identifying the specific portions of a note containing goals-of-care documentation), constructing ontologies to describe goals-of-care discussions, and exploring novel approaches to address biases in training data that differ from those present in real-world data (50). Fourth, because our sample was limited to notes from a single academic health system, our algorithm may not generalize to other health systems. Although this challenge is shared by many clinical applications of NLP/ML, ongoing advances in novel computational linguistic techniques are likely to facilitate generalization of NLP/ML to settings beyond the source of training data (51), which holds great promise for this application of NLP/ML. Fifth, our NLP/ML program demonstrated only modest performance in the analysis of outpatient records, suggesting that there is room for improvement. We hypothesize that goals-of-care discussions in the inpatient setting are more likely to reflect “in-the-moment” treatment decisions (52) occurring during medical crises or near the end of life; documentation of these discussions may be lengthier than outpatient advance care planning discussions, and may be easier for NLP/ML to identify. Despite these limitations, we believe our findings provide an important proof of concept for the use of NLP/ML in identifying goals-of-care discussions and provide direction for future studies to improve this approach.

Conclusion

Using natural language processing and machine learning techniques, we developed and evaluated a novel approach to identifying EHR documentation of goals-of-care discussions with performance characteristics that suggest proof of concept as well as room for improvement. Further study is needed to refine and validate this approach, particularly for identifying outpatient goals-of-care discussions. NLP and ML represent a novel and exciting approach toward measuring goals-of-care discussions as a research outcome and as a quality metric for palliative care.

Supplementary Material

1

KEY MESSAGE.

In this pilot study, natural language processing and machine learning achieved 82% sensitivity and 97% specificity for identifying documented goals-of-care discussions in medical records of patients with serious illness. The results suggest the potential of automated approaches to measuring goals-of-care discussions as a research outcome and quality metric.

DISCLOSURES AND ACKNOWLEDGEMENTS

This work was funded by the Cambia Health Foundation and by UW Medicine. Additionally, Dr. Lee was supported by an F32 award (HL142211) and a K12 award in implementation science (HL137940) from the National Heart, Lung, and Blood Institute. The funding sources had no role in study design; in the collection, analysis and interpretation of data; in the writing of the report; or in the decision to submit the article for publication.

Footnotes

Conflicts of interest: There are no conflicts of interest from any of the authors.

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

REFERENCES

  • 1.Wright AA, Zhang B, Ray A, et al. Associations between end-of-life discussions, patient mental health, medical care near death, and caregiver bereavement adjustment. JAMA 2008;300:1665–73. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Bernacki RE, Block SD, American College of Physicians High Value Care Task F. Communication about serious illness care goals: a review and synthesis of best practices. JAMA Intern Med 2014;174:1994–2003. [DOI] [PubMed] [Google Scholar]
  • 3.Brinkman-Stoppelenburg A, Rietjens JA, van der Heide A. The effects of advance care planning on end-of-life care: a systematic review. Palliat Med 2014;28:1000–25. [DOI] [PubMed] [Google Scholar]
  • 4.Mack JW, Cronin A, Taback N, et al. End-of-life care discussions among patients with advanced cancer: a cohort study. Ann Intern Med 2012;156:204–10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Detering KM, Hancock AD, Reade MC, Silvester W. The impact of advance care planning on end of life care in elderly patients: randomised controlled trial. BMJ 2010;340:c1345. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Institute of Medicine. Dying in America: Improving Quality and Honoring Individual Preferences Near the End of Life, Washington (DC): National Academies Press, 2015. [PubMed] [Google Scholar]
  • 7.Teno JM, Gruneir A, Schwartz Z, Nanda A, Wetle T. Association between advance directives and quality of end-of-life care: a national study. J Am Geriatr Soc 2007;55:189–94. [DOI] [PubMed] [Google Scholar]
  • 8.Silveira MJ, Kim SY, Langa KM. Advance directives and outcomes of surrogate decision making before death. N Engl J Med 2010;362:1211–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Fakhri S, Engelberg RA, Downey L, et al. Factors Affecting Patients’ Preferences for and Actual Discussions About End-of-Life Care. J Pain Symptom Manage 2016;52:386–94. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Heyland DK, Dodek P, You JJ, et al. Validation of quality indicators for end-of-life communication: results of a multicentre survey. CMAJ 2017;189:E980–E989. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Heyland DK, Barwich D, Pichora D, et al. Failure to engage hospitalized elderly patients and their families in advance care planning. JAMA Intern Med 2013;173:778–87. [DOI] [PubMed] [Google Scholar]
  • 12.Davison SN. End-of-life care preferences and needs: perceptions of patients with chronic kidney disease. Clin J Am Soc Nephrol 2010;5:195–204. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Yadav KN, Gabler NB, Cooney E, et al. Approximately One In Three US Adults Completes Any Type Of Advance Directive For End-Of-Life Care. Health Aff (Millwood) 2017;36:1244–1251. [DOI] [PubMed] [Google Scholar]
  • 14.Wilson CJ, Newman J, Tapper S, et al. Multiple locations of advance care planning documentation in an electronic health record: are they easy to find? J Palliat Med 2013;16:1089–94. [DOI] [PubMed] [Google Scholar]
  • 15.Curtis JR, Sathitratanacheewin S, Starks H, et al. Using Electronic Health Records for Quality Measurement and Accountability in Care of the Seriously Ill: Opportunities and Challenges. J Palliat Med 2018;21:S52–S60. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Tulsky JA, Beach MC, Butow PN, et al. A Research Agenda for Communication Between Health Care Professionals and Patients Living With Serious Illness. JAMA Intern Med 2017;177:1361–1366. [DOI] [PubMed] [Google Scholar]
  • 17.National Quality Forum. Palliative Care and End-of-Life Care—A Consensus Report. In: National Voluntary Consensus Standards, April 2012. [Google Scholar]
  • 18.Halpern SD. Goal-Concordant Care - Searching for the Holy Grail. N Engl J Med 2019;381:1603–1606. [DOI] [PubMed] [Google Scholar]
  • 19.Nadkarni PM, Ohno-Machado L, Chapman WW. Natural language processing: an introduction. J Am Med Inform Assoc 2011;18:544–51. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Kotsiantis SB. Supervised Machine Learning: A Review of Classification Techniques In: Maglogiannis I, Karpouzis K, Wallace BA, Soldatos J, eds. Emerging Artificial Intelligence Applications in Computer Engineering: Real Word AI Systems with Applications in EHealth, HCI, Information Retrieval and Pervasive Technologies, IOS Press, 2007:3–24. [Google Scholar]
  • 21.Jiang F, Jiang Y, Zhi H, et al. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol 2017;2:230–243. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Wang Y, Wang L, Rastegar-Mojarad M, et al. Clinical information extraction applications: A literature review. J Biomed Inform 2018;77:34–49. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Fiszman M, Chapman WW, Aronsky D, Evans RS, Haug PJ. Automatic detection of acute bacterial pneumonia from chest X-ray reports. J Am Med Inform Assoc 2000;7:593–604. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Murff HJ, FitzHenry F, Matheny ME, et al. Automated identification of postoperative complications within an electronic medical record using natural language processing. JAMA 2011;306:848–55. [DOI] [PubMed] [Google Scholar]
  • 25.Heintzelman NH, Taylor RJ, Simonsen L, et al. Longitudinal analysis of pain in patients with metastatic prostate cancer using natural language processing of medical record text. J Am Med Inform Assoc 2013;20:898–905. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Imler TD, Morea J, Kahi C, et al. Multi-center colonoscopy quality measurement utilizing natural language processing. Am J Gastroenterol 2015;110:543–52. [DOI] [PubMed] [Google Scholar]
  • 27.Lee JK, Jensen CD, Levin TR, et al. Accurate Identification of Colonoscopy Quality and Polyp Findings Using Natural Language Processing. J Clin Gastroenterol 2019;53:e25–e30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Nayor J, Borges LF, Goryachev S, Gainer VS, Saltzman JR. Natural Language Processing Accurately Calculates Adenoma and Sessile Serrated Polyp Detection Rates. Dig Dis Sci 2018;63:1794–1800. [DOI] [PubMed] [Google Scholar]
  • 29.Castro VM, Dligach D, Finan S, et al. Large-scale identification of patients with cerebral aneurysms using natural language processing. Neurology 2017;88:164–168. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Forsyth AW, Barzilay R, Hughes KS, et al. Machine Learning Methods to Extract Documentation of Breast Cancer Symptoms From Electronic Health Records. J Pain Symptom Manage 2018;55:1492–1499. [DOI] [PubMed] [Google Scholar]
  • 31.Weissman GE, Hubbard RA, Ungar LH, et al. Inclusion of Unstructured Clinical Text Improves Early Prediction of Death or Prolonged ICU Stay. Crit Care Med 2018;46:1125–1132. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Lilley EJ, Lindvall C, Lillemoe KD, et al. Measuring Processes of Care in Palliative Surgery: A Novel Approach Using Natural Language Processing. Ann Surg 2018;267:823–825. [DOI] [PubMed] [Google Scholar]
  • 33.Lindvall C, Lilley EJ, Zupanc SN, et al. Natural Language Processing to Assess End-of-Life Quality Indicators in Cancer Patients Receiving Palliative Surgery. J Palliat Med 2019;22:183–187. [DOI] [PubMed] [Google Scholar]
  • 34.Stephens AR, Wiener RS, Ieong MH. Comparison of Methods To Identify Advance Care Planning in Patients with Severe Chronic Obstructive Pulmonary Disease Exacerbation. J Palliat Med 2018;21:284–289. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Poort H, Zupanc SN, Leiter RE, Wright AA, Lindvall C. Documentation of Palliative and End-of-Life Care Process Measures Among Young Adults Who Died of Cancer: A Natural Language Processing Approach. J Adolesc Young Adult Oncol 2020;9:100–104. [DOI] [PubMed] [Google Scholar]
  • 36.Brizzi K, Zupanc SN, Udelsman BV, et al. Natural Language Processing to Assess Palliative Care and End-of-Life Process Measures in Patients With Breast Cancer With Leptomeningeal Disease. Am J Hosp Palliat Care 2020;37:371–376. [DOI] [PubMed] [Google Scholar]
  • 37.Chan A, Chien I, Moseley E, et al. Deep learning algorithms to identify documentation of serious illness conversations during intensive care unit admissions. Palliat Med 2019;33:187–196. [DOI] [PubMed] [Google Scholar]
  • 38.Udelsman BV, Moseley ET, Sudore RL, Keating NL, Lindvall C. Deep Natural Language Processing Identifies Variation in Care Preference Documentation. J Pain Symptom Manage 2020;59:1186–1194. [DOI] [PubMed] [Google Scholar]
  • 39.Curtis JR, Downey L, Back AL, et al. Effect of a Patient and Clinician Communication-Priming Intervention on Patient-Reported Goals-of-Care Discussions Between Patients With Serious Illness and Clinicians: A Randomized Clinical Trial. JAMA Intern Med 2018;178:930–940. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Curtis JR, Treece PD, Nielsen EL, et al. Randomized Trial of Communication Facilitators to Reduce Family Distress and Intensity of End-of-Life Care. Am J Respir Crit Care Med 2016;193:154–62. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Goodman DC, Esty AR, Fisher ES, Chang C-H. Trends and Variation in End-of-Life Care for Medicare Beneficiaries with Severe Chronic Illness: A Report of the Dartmouth Atlas Project In: Bronner KK, ed. The Dartmouth Atlas of Health Care, The Dartmouth Institute for Health Policy and Clinical Practice, 2011. [PubMed] [Google Scholar]
  • 42.Dartmouth Institute for Health Policy and Clinical Practice. Crosswalk File of ICD9 Diagnosis Codes to Risk Group Assessment. 2015. Available from: http://archive.dartmouthatlas.org/downloads/methods/Chronic_Disease_Codes.pdf. Accessed Aug 24, 2016. [Google Scholar]
  • 43.Pedregosa F, Varoquaux G, Gramfort A, et al. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 2011;12:2825–2830. [Google Scholar]
  • 44.Bird S, Klein E, Loper E. Natural language processing with Python, 1st ed. Beijing; Cambridge Mass.: O’Reilly, 2009. [Google Scholar]
  • 45.Uzuner O, South BR, Shen S, DuVall SL. 2010 i2b2/VA challenge on concepts, assertions, and relations in clinical text. J Am Med Inform Assoc 2011;18:552–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Gundlapalli AV, Carter ME, Palmer M, et al. Using natural language processing on the free text of clinical documents to screen for evidence of homelessness among US veterans. AMIA Annu Symp Proc 2013;2013:537–46. [PMC free article] [PubMed] [Google Scholar]
  • 47.Gundlapalli AV, Carter ME, Divita G, et al. Extracting Concepts Related to Homelessness from the Free Text of VA Electronic Medical Records. AMIA Annu Symp Proc 2014;2014:589–98. [PMC free article] [PubMed] [Google Scholar]
  • 48.Bejan CA, Angiolillo J, Conway D, et al. Mining 100 million notes to find homelessness and adverse childhood experiences: 2 case studies of rare and severe social determinants of health in electronic health records. J Am Med Inform Assoc 2018;25:61–71. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Sanders JJ, Curtis JR, Tulsky JA. Achieving Goal-Concordant Care: A Conceptual Model and Approach to Measuring Serious Illness Communication and Its Impact. J Palliat Med 2018;21:S17–S27. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Howell KP, Barnes MR, Curtis JR, et al. Controlling for confounding variables: accounting for dataset bias in classifying patient-provider interactions. In: Proceedings of the W3PHIAI International Workshop on Health Intelligence, New York, NY: 2020. [Google Scholar]
  • 51.Bender EM, Daumé H III, Ettinger A, Rao S. In: Proceedings of the First Workshop on Building Linguistically Generalizable NLP systems (BLGNLP), Empirical Methods in Natural Language Processing (EMNLP 2017), Copenhagen, Denmark: Association for Computational Linguistics, 2017. [Google Scholar]
  • 52.Sudore RL, Fried TR. Redefining the “planning” in advance care planning: preparing for end-of-life decision making. Ann Intern Med 2010;153:256–61. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

1

RESOURCES