Skip to main content
Lippincott Open Access logoLink to Lippincott Open Access
. 2024 Nov 25;8:e2400100. doi: 10.1200/CCI.24.00100

Automated Electronic Health Record Data Extraction and Curation Using ExtractEHR

Tamara P Miller 1,2, Kelly D Getz 3, Edward Krause 4, Yun Gun Jo 4, Sandhya Charapala 4, M Monica Gramatages 5,6, Karen Rabin 5,6, Michael E Scheurer 5,6, Jennifer J Wilkes 7, Brian T Fisher 3,8, Richard Aplenc 3,9,10,
PMCID: PMC11608624  PMID: 39586036

Abstract

PURPOSE

Although the potential transformative effect of electronic health record (EHR) data on clinical research in adult patient populations has been very extensively discussed, the effect on pediatric oncology research has been limited. Multiple factors contribute to this more limited effect, including the paucity of pediatric cancer cases in commercial EHR-derived cancer data sets and phenotypic case identification challenges in pediatric federated EHR data.

METHODS

The ExtractEHR software package was initially developed as a tool to improve clinical trial adverse event reporting but has expanded its use cases to include the development of multisite EHR data sets and the support of cancer cohorts. ExtractEHR enables customized, automated data extraction from the EHR that, when implemented across multiple hospitals, can create pediatric cancer EHR data sets to address a very wide range of research questions in pediatric oncology. After ExtractEHR data acquisition, EHR data can be cleaned and graded using CleanEHR and GradeEHR, companion software packages.

RESULTS

ExtractEHR has been installed at four leading pediatric institutions: Children's Healthcare of Atlanta, Children's Hospital of Philadelphia, Texas Children's Hospital, and Seattle Children's Hospital.

CONCLUSION

ExtractEHR has supported multiple use cases, including five clinical epidemiology studies, multicenter clinical trials, and cancer cohort assembly. Work is ongoing to develop Fast Health care Interoperability Resources ExtractEHR and implement other sustainability and scalability enhancements.


ExtractEHR is a low-cost software package to extract and curate electronic health record data in multiple use cases.

INTRODUCTION

Many publications have described both the potential and limitations of the electronic health record (EHR) in facilitating clinical research.1-4 This transformative potential arises from the extensive, digitally accessible data describing patients' medical experience with previously unavailable granularity. The well-described limitations stem in part from the typically siloed data structures that hinder broad data access.5 Many investigators and commercial entities seek to break down these silos to create data sets that span institutions and enable practice-changing data analyses.

CONTEXT

  • Key Objective

  • To describe development and use cases of ExtractEHR, a software package that was developed to extract and process data from the electronic health record (EHR).

  • Knowledge Generated

  • The process of implementing and using ExtractEHR for customized, automated data extraction from the EHR across multiple hospitals is described. Use cases demonstrate the range of opportunities for which ExtractEHR can be beneficial.

  • Relevance

  • A software tool that can be used to extract data from EHR (here used in pediatric oncology)—obviously very useful.

Typically these efforts focus on adult health conditions with large numbers of patients and potential revenue for commercial entities. The Flatiron Health database is perhaps best known and has been used for multiple studies in cancer cohorts.6-9 Although extensive, Flatiron does not currently include sufficient numbers of pediatric cancer patients, and access to the data requires Flatiron approval. The Center for International Blood and Marrow Transplant (CIBMTR) is piloting an automated Data Transformation Initiative to extract EHR data, including demographics, diagnoses and dates, labs, therapies, responses, relapses, and transplant data for pediatric and adult patients across sites receiving stem-cell and cellular therapies. However, this is specific to CIBMTR forms and limited to patients receiving hematopoietic stem-cell transplants, and thus may not be generalizable to broader populations. Finally, REDCap Clinical Data Interoperability Services (CDIS) enables transfer of EHR data directly into REDCap for research purposes.10 Although user-friendly, the CDIS does not currently include radiology results or clinician notes, does not incorporate data cleaning or synthesis procedures, and does not directly support federating EHR data across institutions.11

In pediatrics, EHR data federation has been led by PEDSNet, which federates data from multiple free-standing pediatric institutions and has supported multiple clinical research studies.12 This work has included creation of a computable phenotype identifying patients with acute leukemia and lymphoma.13 However, this computed phenotype does not distinguish between acute leukemia subtypes, an important potential limitation. Other pediatric groups have developed algorithms to identify local patient cohorts and to extract targeted critical clinical data, such as the cumulative anthracycline exposure.14 However, these efforts have been limited to single institutions and do not generate the larger scale, multi-institutional data sets needed to address complex clinical epidemiology questions.

In response to the need for a comprehensive data set inclusive of patients from multiple hospitals that can be used to answer clinical pediatric oncology research questions, we developed a software tool called ExtractEHR. We aim to describe the procedures used to develop and implement ExtractEHR, provide use cases across a range of clinical research applications, and discuss future potential applications.

METHODS

Overview

ExtractEHR is a software package written in R that extracts various structured, semi-structured, and unstructured data components from EHR systems. Data components include discrete chart elements, such as laboratory results or flowsheet data, and free-text documents, such as clinician notes or radiology reports. The specific components used for a given project are dictated by the project's protocol. During implementation, ExtractEHR is programmed to extract data from the EHR database for the patients and boundary dates needed for the specific use case. In addition to comprehensive EHR data extraction, we have developed two modules for postextraction EHR data cleaning and processing. CleanEHR cleans raw data for downstream analytic use. GradeEHR computes grades per Common Terminology Criteria for Adverse Events (CTCAE) for cleaned laboratory results and other data analyses. Figure 1 provides a schematic representation of the ExtractEHR, CleanEHR, and GradeEHR inputs, outputs, and processes, whereas Figure 2 provides an overview of ExtractEHR onboarding and setup.

FIG 1.

FIG 1.

EHR processing toolkit. EHR, electronic health record.

FIG 2.

FIG 2.

ExtractEHR setup. EHR, electronic health record; SQL, Structured Query Language.

Initial Setup

Implementation of ExtractEHR begins with an installation agreement between the site and the ExtractEHR Team. ExtractEHR is free but is not open-source due to the proprietary nature of the connections to the EHR vendors. ExtractEHR code is stored in a private GitHub repository with access given during initial implementation. ExtractEHR setup requires users to populate a configuration file with credentials for the source system connections and other settings that guide the data extraction process. Each institution must also complete a localized mapping process, in which specific data are defined and added to the template Structured Query Language (SQL) files included with the tool. Although institutions using the same brand of EHR system share a common data model, the data labels assigned to data elements used by ExtractEHR may differ between sites. Mapping ensures that data elements will be properly extracted and has been accomplished at Epic, Cerner, and All Scripts sites.

Mapping can be comprehensive across a data element type (eg, all laboratory tests) or tailored to include targeted elements (specific laboratory tests for a given use case). Pan-extraction (eg, all laboratory tests) substantially minimizes the mapping work required at ExtractEHR setup. However, it shifts the identification of individual components downstream in the analytic processing pipeline. In either mapping strategy, ExtractEHR may be used repeatedly for different use cases at the institution, unless other EHR components are desired or the EHR system undergoes a significant update with changes to the underlying data structure. Table 1 presents extracted data categories.

TABLE 1.

Data Element Categories Captured by ExtractEHR

Data Element Categories Type of Data LEARN AdOPT Cooperative Group Trials SEER CCSS
Patient demographics Structured X X X
Encounters: Inpatient, day hospital, emergency room, and outpatient Structured X X X
Flowsheet data (vital signs) Structured X X—only height and weight
Medications ordered, including outpatient prescription orders Structured X X X
Medications administered Structured X X X
Laboratory test results, including microbiology test results Structured X X X X
Radiology result reports Semi-structured X X X
Pathology result reports Semi-structured X X X
Procedures Structured X X X
Clinician notes Unstructured X X—only includes oncology clinician notes X
Genomic data, if available Variable by site X X

Abbreviations: AdOPT, Outcomes of Human Adenovirus (HAdV) Infection and Disease in Pediatric Allogeneic Hematopoietic Cell Transplant Recipients: A Multicenter Observational Cohort; CCSS, Childhood Cancer Survivorship Study; EHR, electronic health record; LEARN, Leukemia Electronic Abstraction of Records Network; SEER, Surveillance, Epidemiology, and End Results; Trials, Cooperative Oncology Group Clinical Trials.

Operation and Output

Figures 3 and 4 display the process for operating ExtractEHR. Once the ExtractEHR initial setup is completed and the patients (identified by medical record number [MRN] or patient name and date of birth) and date boundaries (eg, date of birth or start/end dates of therapy courses) are identified, ExtractEHR extracts the data elements for that project. When the program is launched, ExtractEHR first loads the configuration file and establishes a connection to the EHR database. With the initial settings in place, the patient identifiers and data boundaries are loaded into a tabular data object called a dataframe and then formatted for insertion into SQL queries. For each component of the extraction, the tool loads a query template and, using SQL parametrization, inserts the patient IDs, filter criteria, and date boundaries to create a complete query statement. The query is transmitted to the database server, executed against the EHR database, and the results are transmitted back to ExtractEHR and stored in a dataframe. The resulting dataframe is then formatted for output, which differs on the basis of the EHR component and may include unit conversions, organization of free-text notes and comments fields, and trimming records on the basis of individual patient date boundaries. Depending on the use case, the ExtractEHR code can also remove MRNs and replace them with study IDs. For example, for a clinical trial using ExtractEHR to provide data, the MRNs would be removed and trial study ID would be appended. The formatted dataframe is then exported to a CSV file located in the output directory included with the tool or may be imported into REDCap, and the process can be repeated for additional queries.

FIG 3.

FIG 3.

ExtractEHR process overview. EHR, electronic health record.

FIG 4.

FIG 4.

ExtractEHR query process overview. EHR, electronic health record.

POSTEXTRACTION PROCESSING

We have developed two postextraction modules for data cleaning and processing. Using a series of rules defined by the ExtractEHR clinical team, CleanEHR removes false or duplicate results from the extracted data set, standardizes data element formatting, and creates pre- and postcleaning summary metrics of each file for quality control purposes.15 Similar to ExtractEHR, CleanEHR requires a configuration file, which contains variables for site and project names, filepaths, and optional database connection parameters, allowing cleaned data sets to be written to CSV or uploaded into a database. CleanEHR also includes a set of file cleaning parameters, which define additional settings, such as data processing workflows and exclusion values. CleanEHR identifies suspected false positives, such as electrolyte or hematologic laboratory results that change from being abnormal to normal in less than an hour.15 False-positive results from improper specimen collection or handling, such as a hemolyzed potassium result (identified in the comment field for the laboratory result), are removed.15 Cleaning metric reports are generated, which include raw frequency distributions (including procedure names, component names, sources, and reference units), lab result values, missing results, references values, summary statistics, and cleaned frequency distributions.

The second postextraction module is GradeEHR that grades each laboratory result per CTCAE v5 definitions. Like CleanEHR, GradeEHR runs centrally on data sets produced by CleanEHR. GradeEHR requires a configuration file (providing the site name, filepaths, and optional source and target database connection information), as well as a set of grading parameters that define the variable names, ranges, and grades, and the CleanEHR output data set. For each laboratory AE, GradeEHR loads the appropriate laboratory result data set, assigns a CTCAE grade to each record, and calculates whether the AE has normalized within a specified time interval. The graded data sets can be exported to CSV or loaded into a database. Currently, GradeEHR is capable of grading 25 laboratory AEs, with plans for additional laboratory AEs to be added. GradeEHR also generates a grading metrics report, which includes time difference frequency distributions, AE grade frequency distributions, and cross-tabulations of AE grades.

ACCURACY

The accuracy of the ExtractEHR and postextraction CleanEHR and GradeEHR modules was tested in pediatric patients with AML at a single site.15 ExtractEHR had a sensitivity of over 98% for all laboratory AEs, whereas manual AE reports ascertained by clinical research associates had a sensitivity of 0%-21.1%.15 AE reporting errors included false-negative and false-positive results and incorrect grading, both ameliorated by the software packages.15

POSTEXTRACTION DATA TRANSFER

Data management of extracted but unprocessed data is determined by use case requirements. Site data may be transferred directly to a third party or a central data coordinating center (DCC). In all use cases, site-specific extracted data are stored locally at each site in the extracted format. For example, laboratory data obtained by mapped data queries are stored as individual specific laboratory data files. Laboratory data obtained from pan-extracted, unmapped processes are stored in a single file containing multiple laboratory results within the file. Currently, CleanEHR and GradeEHR are only run at the DCC. Thus, use cases requiring cleaned or graded data require unprocessed data to be at least temporarily resident at the DCC. Cleaned and graded data may be retained centrally or passed on to a third-party recipient. Site-specific cleaned and graded laboratory data are always provided to the individual site.

RESULTS

Leukemia Electronic Abstraction of Records Network

ExtractEHR has been implemented at four large pediatric hospitals that comprise the Leukemia Electronic Abstraction of Records Network (LEARN) collaborative network. The LEARN cohort includes patients age 0-21 years treated for de novo AML or acute lymphoblastic leukemia (ALL) who received at least one complete course of chemotherapy at one of the included sites. LEARN currently includes Children's Healthcare of Atlanta (CHOA), Children's Hospital of Philadelphia (CHOP), Texas Children's Hospital, and Seattle Children's Hospital with several other sites being onboarded. Institutional Review Board (IRB) approval for LEARN research was obtained at all sites with waiver of informed consent.

LEARN has produced five peer-reviewed publications to date, demonstrating the utility of collecting granular data from the EHR for answering clinical research questions.16-20 A recent Lancet Haematology publication described rates of laboratory AEs in children with ALL and AML and delineated that commonly occurring laboratory AEs occurred more frequently when identified using ExtractEHR than had previously been reported in the data collected on Children's Oncology Group (COG) clinical trials.17 Strikingly, the most common laboratory abnormalities in both ALL and AML had log-fold differences in frequency between data identified in the ExtractEHR-enabled LEARN cohort and manually reported COG trial data. Additionally, the ExtractEHR-enabled LEARN data included patients regardless of clinical trial enrollment and thus constitute comprehensive, real-world data. In a second publication, ExtractEHR-ascertained data were used to examine treatment-associated hepatotoxicity in a diverse sample of children with ALL.20 Other recent laboratory data–based publications include the description of lymphocyte count recovery after AML chemotherapy, acute kidney injury after ALL and AML chemotherapy, and differences in severity of illness at presentation across racial/ethnic groups.16,18,19

ExtractEHR can also ascertain complex phenotypes that may be poorly defined by discrete data EHR data elements, that is, diagnosis codes. For example, ExtractEHR-ascertained data have been used to identify typhlitis, a serious and potentially life-threatening complication in leukemia treatment, in a LEARN subcohort.21 This paper reported an automated algorithm combining discrete EHR elements and natural language–processed free-text clinician notes to identify typhlitis. The algorithm outperformed manual abstraction with a higher sensitivity and a log-fold reduction in the number of chemotherapy courses requiring manual review.22

Outcomes of Human Adenovirus Infection and Disease in Pediatric Allogeneic Hematopoietic Cell Transplant Recipients: A Multicenter Observational Cohort (AdOPT)

AdOPT is a mixed prospective and retrospective multicenter observational cohort study investigating Human Adenovirus infection and disease in pediatric allogeneic stem-cell transplant recipients. ExtractEHR is used at three of the 10 participating sites to obtain laboratory data; the other seven sites abstract laboratory data manually. The raw laboratory data obtained via ExtractEHR are transferred to the central data analysis team for import into the study's REDCap database. Sites using ExtractEHR report shorter chart abstraction times than those using manual laboratory data collection. Since the study population includes patients receiving an allogeneic stem-cell transplant for various conditions, the AdOPT study demonstrates the broad applicability of ExtractEHR. IRB approval for the AdOPT study was obtained at all sites with informed consent obtained for study participants.

Cooperative Oncology Group Clinical Trials

In the current COG Phase III clinical trial for high-risk neuroblastoma, ANBL1531 (ClinicalTrials.gov identifier: NCT03126916), ExtractEHR is being used at CHOP and CHOA to identify and grade laboratory AEs. This study subaim will identify the highest grade of specified laboratory AEs by reporting period and assess the number of AEs detected and the amount of time required for each AE to be reported both before and after ExtractEHR implementation. In addition, EHR ascertainment of laboratory test results is being evaluated on a joint trial between the COG Pediatric Early Phase Clinical Trials Network and the Pediatric Brain Tumor Consortium.23,24 In the recently completed study (ClinicalTrials.gov identifier: NCT05020951), ExtractEHR was used to extract and electronically transfer raw laboratory data from five participating sites to the cloud-based Medidata Rave Electronic Data Capture system.25 IRB approval and informed consent were obtained at all sites participating in ANBL1651. IRB approval with waiver of informed consent was obtained at one site for the NCT05020951 trial with IRB determination at all other sites that the trial was not human subjects research.

SEER Cancer Registry Pilot

The Surveillance, Epidemiology, and End Results (SEER) program of the National Cancer Institute includes 22 registries with collection of patient demographics, cancer information, and mortality and survival data.26 Cancer registries obtain data through provider reporting, which is required by law, but is labor-intensive and can be incomplete, inaccurate, and inconsistent.27 An ongoing pilot project has used ExtractEHR to ascertain and transfer demographics, clinical notes, laboratory results, medications ordered and administered, pathology reports, procedures, radiology reports, visits (clinic and inpatient), and vital signs (height and weight) from CHOA to the Georgia SEER Registry. Additional pilot testing in the Washington State, Texas, and New York SEER registries was recently completed. SEER pilot data are also shared with the National Childhood Cancer Registry (NCCR). IRB approval and consent were not obtained as this work was determined to be part of federally mandated cancer surveillance.

Childhood Cancer Survivor Study

The Childhood Cancer Survivor Study (CCSS) contains data on approximately 30,000 childhood cancer survivors and has enabled many high-impact publications that have defined the care of childhood cancer survivors. CCSS relies on manual data abstraction of approximately 200 data elements, including chemotherapy exposure. ExtractEHR is being tested at five sites to evaluate the feasibility of automating capture of elements in the medical record abstraction form, particularly chemotherapy. The protocol supporting this work received IRB approval with waiver of informed consent.

DISCUSSION

The successful implementation of ExtractEHR across a range of use cases demonstrates the broad potential applicability of ExtractEHR in pediatric cancer research, specifically the ability to facilitate development of real-world patient cohorts and to support clinical data registries. In these contexts, ExtractEHR may have particular value as a cost-effective tool for inclusive case identification regardless of clinical trial enrollment status. This capacity should address many of the well-described biases in research cohorts derived from clinical trials.28

As with any tool, ExtractEHR has important limitations that vary in impact across use cases. The primary limitation is the requirement for installation behind a local institutional firewall. Successful installation requires close collaboration between the ExtractEHR team and the local site team that typically comprises a clinical champion and a technical analyst. Support from institutional Information Systems is crucial, and the clinical champion's role in facilitating this support should not be underestimated. The requirement for local mapping, which can be labor-intensive, is another limitation. Although pan-extraction reduces local mapping effort, it requires additional downstream data processing before data analyses. In addition, depending on the number of patients included, pan-extraction may create very large data files. Finally, some institutions may require informed consent for certain EHR components, such as genomic data, and data sharing may require deidentification, which is challenging with unstructured EHR components. Work is ongoing to address this limitation by incorporating postextraction code to deidentify data elements or identify specific parts of each EHR component that do not include protected health information but retain clinical value. Use of existing packages such as Philter may also facilitate overcoming these challenges.29

Despite these limitations, ExtractEHR has several advantages. First, once installed, the package can be used for multiple, repeated use cases that can span varying, unrelated patient cohorts. These use cases can focus on any pediatric disease wherever the source population can be readily identified by a local site investigator. As an example, ExtractEHR has been used to extract data for children with chronic kidney disease at a single institution and is in process at other sites that are part of a multicenter collaboration. Second, once extracted, the data can be used not only for the defined use case but also for other local, secondary data analyses. This is particularly applicable when the mapped laboratory extraction approach is used, as that provides individual laboratory data files to the investigator. Furthermore, while ExtractEHR requires mapping at each institution, ExtractEHR does not require data transformation to a specific data format and rather extracts data as the institution has determined they should be stored. This provides an alternative to other existing data extraction methods that require use of a common data model or data standard. Finally, as an academically licensed software product, the only incurred costs are local site investigator and staff time.

As next steps, we are developing a Fast Health care Interoperability Resources (FHIR) version of ExtractEHR to determine whether local mapping requirements may be decreased or eliminated in specific use cases. The FHIR standard for information exchange simplifies EHR data sharing across institutions30 and has been widely adopted throughout the health care industry. The development of an FHIR ExtractEHR tool and broader ExtractEHR dissemination and use are crucial components of a sustainability strategy for ExtractEHR.

ACKNOWLEDGMENT

The authors would like to thank Ms Evanette Burrows for her foundational work in developing ExtractEHR, our publication co-authors, and ExtractEHR advocates, particularly Drs Malcolm Smith and Lynne Penberthy.

DISCLAIMER

The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

SUPPORT

Supported in whole or in part with Federal funds from the Childhood Cancer Data Initiative (CCDI), part of the National Cancer Institute under grant/Contract No. U24CA055727-29S3 and D2023-CHOP-1.

AUTHOR CONTRIBUTIONS

Conception and design: Tamara P. Miller, Edward Krause, Yun Gun Jo, Brian T. Fisher, Richard Aplenc

Financial support: Tamara P. Miller, Richard Aplenc

Administrative support: Tamara P. Miller, Richard Aplenc

Provision of study materials or patients: Tamara P. Miller, M. Monica Gramatages, Michael E. Scheurer, Jennifer J. Wilkes, Richard Aplenc

Collection and assembly of data: Tamara P. Miller, Edward Krause, Yun Gun Jo, Sandhya Charapala, Michael E. Scheurer, Jennifer J. Wilkes, Richard Aplenc

Data analysis and interpretation: Tamara P. Miller, Kelly D. Getz, M. Monica Gramatages, Karen Rabin, Michael E. Scheurer, Jennifer J. Wilkes, Brian T. Fisher, Richard Aplenc

Manuscript writing: All authors

Final approval of manuscript: All authors

Accountable for all aspects of the work: All authors

AUTHORS' DISCLOSURES OF POTENTIAL CONFLICTS OF INTEREST

The following represents disclosure information provided by authors of this manuscript. All relationships are considered compensated unless otherwise noted. Relationships are self-held unless noted. I = Immediate Family Member, Inst = My Institution. Relationships may not relate to the subject matter of this manuscript. For more information about ASCO's conflict of interest policy, please refer to www.asco.org/rwc or ascopubs.org/cci/author-center.

Open Payments is a public database containing information reported by companies about payments made to US-licensed physicians (Open Payments).

Tamara P. Miller

Stock and Other Ownership Interests: Gilead Sciences, Thermo Fisher Scientific, AbbVie, United Health Group

Kelly D. Getz

Research Funding: Harvard Pilgrim Healthcare

M. Monica Gramatges

Open Payments Link: https://openpaymentsdata.cms.gov/physician/1175834/

Brian T. Fisher

Consulting or Advisory Role: Astellas Pharma

Research Funding: Pfizer (Inst), Merck (Inst), Allovir (Inst)

Richard Aplenc

Expert Testimony: Vorys, Sater, Seymour, and Pease LLP

No other potential conflicts of interest were reported.

REFERENCES

  • 1.Casey JA, Schwartz BS, Stewart WF, et al. : Using electronic health records for population health research: A review of methods and applications. Annu Rev Public Health 37:61-81, 2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Cowie MR, Blomster JI, Curtis LH, et al. : Electronic health records to facilitate clinical research. Clin Res Cardiol 106:1-9, 2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Griffiths R, Herbert L, Akbari A, et al. : A methodology to facilitate critical care research using multiple linked electronic, clinical and administrative health records at population scale. Int J Popul Data Sci 7:1724, 2022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Mahmoudi E, Kamdar N, Kim N, et al. : Use of electronic medical records in development and validation of risk prediction models of hospital readmission: Systematic review. BMJ 369:m958, 2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Sauer CM, Chen LC, Hyland SL, et al. : Leveraging electronic health records for data science: Common pitfalls and how to avoid them. Lancet Digit Health 4:e893-e898, 2022 [DOI] [PubMed] [Google Scholar]
  • 6.Flatiron Health : Realizing the Full Potential of Real-World Evidence in Oncology. 2023 [Google Scholar]
  • 7.Ma X, Long L, Moon S, et al. : Comparison of population characteristics in real-world clinical oncology databases in the US: Flatiron health, SEER, and NPCR. medRxiv 2020. 2020.03.16.20037143 [Google Scholar]
  • 8.Liu R, Rizzo S, Waliany S, et al. : Systematic pan-cancer analysis of mutation-treatment interactions using large real-world clinicogenomics data. Nat Med 28:1656-1661, 2022 [DOI] [PubMed] [Google Scholar]
  • 9.Liu R, Rizzo S, Whipple S, et al. : Evaluating eligibility criteria of oncology trials using real-world data and AI. Nature 592:629-633, 2021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.REDCap : Clinical Data Interoperability Services (CDIS)
  • 11.Cheng AC, Duda SN, Taylor R, et al. : REDCap on FHIR: Clinical data interoperability Services. J Biomed Inform 121:103871, 2021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Forrest CB, Margolis PA, Bailey LC, et al. : PEDSnet: A National Pediatric Learning Health System. J Am Med Inform Assoc 21:602-606, 2014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Phillips CA, Razzaghi H, Aglio T, et al. : Development and evaluation of a computable phenotype to identify pediatric patients with leukemia and lymphoma treated with chemotherapy using electronic health record data. Pediatr Blood Cancer 66:e27876, 2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Noyd DH, Berkman A, Howell C, et al. : Leveraging clinical informatics tools to extract cumulative anthracycline exposure, measure cardiovascular Outcomes, and assess guideline adherence for children with cancer. JCO Clin Cancer Inform 10.1200/CCI.21.00099 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Miller TP, Li Y, Getz KD, et al. : Using electronic medical record data to report laboratory adverse events. Br J Haematol 177:283-286, 2017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Hsiao W, Li Y, Getz K, et al. : Acute and chronic kidney injury during therapy for pediatric acute leukemia: A report from the Leukemia Electronic Abstraction of Records Network (LEARN). Pediatr Blood Cancer 70:e30696, 2023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Miller TP, Getz KD, Li Y, et al. : Rates of laboratory adverse events by course in paediatric leukaemia ascertained with automated electronic health record extraction: A retrospective cohort study from the Children's Oncology Group. Lancet Haematol 9:e678-e688, 2022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Molina JC, Li Y, Otto WR, et al. : Absolute lymphocyte count recovery following initial acute myelogenous leukemia therapy: Implications for adoptive cell therapy. Pediatr Blood Cancer 70:e30062, 2023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Winestone LE, Getz KD, Li Y, et al. : Racial and ethnic disparities in acuity of presentation among children with newly diagnosed acute leukemia. Pediatr Blood Cancer 71:e30726, 2024 [DOI] [PubMed] [Google Scholar]
  • 20.Yi JS, Chambers TM, Getz KD, et al. : A report from the Leukemia Electronic Abstraction of Records Network on risk of hepatotoxicity during pediatric acute lymphoblastic leukemia treatment. Haematologica 107:1185-1188, 2022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Altınel E, Yarali N, Isık P, et al. : Typhlitis in acute childhood leukemia. Med Princ Pract 21:36-39, 2012 [DOI] [PubMed] [Google Scholar]
  • 22.Miller TP, Li Y, Masino AJ, et al. : Automated ascertainment of typhlitis from the electronic health record. JCO Clin Cancer Inform 10.1200/CCI.22.00081 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.National Cancer Institute : Pediatric Early Phase Clinical Trials Network (PEP-CTN). 2021 [Google Scholar]
  • 24.Pediatric Brain Tumor Consortium : About PBTC. 2023 [Google Scholar]
  • 25.Medidata : Rave EDC (Electronic Data Capture). 2023 [Google Scholar]
  • 26.National Cancer Institute : Overview of the SEER Program
  • 27.Park HS, Lloyd S, Decker RH, et al. : Limitations and biases of the surveillance, epidemiology, and end results database. Curr Probl Cancer 36:216-224, 2012 [DOI] [PubMed] [Google Scholar]
  • 28.Winestone LE, Getz KD, Rao P, et al. : Disparities in pediatric acute myeloid leukemia (AML) clinical trial enrollment. Leuk Lymphoma 60:2190-2198, 2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Norgeot B, Muenzen K, Peterson TA, et al. : Protected health information filter (Philter): Accurately and securely de-identifying free-text clinical notes. NPJ Digit Med 3:57, 2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.HL7 : Getting Started with FHIR

Articles from JCO Clinical Cancer Informatics are provided here courtesy of Wolters Kluwer Health

RESOURCES