Skip to main content
The BMJ logoLink to The BMJ
. 2021 Jan 12;372:m4856. doi: 10.1136/bmj.m4856

STaRT-RWE: structured template for planning and reporting on the implementation of real world evidence studies

Shirley V Wang 1,, Simone Pinheiro 2, Wei Hua 2, Peter Arlett 3 ,4, Yoshiaki Uyama 5, Jesse A Berlin 6, Dorothee B Bartels 7, Kristijan H Kahler 9, Lily G Bessette 1, Sebastian Schneeweiss 1
PMCID: PMC8489282  PMID: 33436424

Abstract

In alignment with the International Council of Harmonization’s strategic goals, a public-private consortium has developed a structured template for planning and reporting on the implementation of real world evidence (RWE) studies of the safety and effectiveness of treatments. The template serves as a guiding tool for designing and conducting reproducible RWE studies; set clear expectations for transparent communication of RWE methods; reduce misinterpretation of prose that lacks specificity; allow reviewers to quickly orient and find key information; and facilitate reproducibility, validity assessment, and evidence synthesis. The template is intended for use with studies of the effectiveness and safety of medical products and is compatible with multiple study designs, data sources, reporting guidelines, checklists, and bias assessment tools.


Real world evidence (RWE) generated from sources of real world data via the application of principled database epidemiology increasingly informs important decisions about the clinical effectiveness of medical products and interventions.1 2 3 4 5 Unlike clinical trials, which can leverage the power of randomisation, or non-randomised studies with prospective data collection for a specific research purpose, most RWE studies make secondary use of electronic data collected as part of routine healthcare processes (eg, administrative claims and electronic health records). Generating high quality evidence when analysing data not collected for research purposes requires decision making about many complex design and analytical parameters to handle temporality, measurement, confounding, and other potential sources of bias. Compared with trials and non-experimental studies that prospectively collect data for a research question, RWE studies have greater variability in design and analysis options. Owing to the current lack of structure in study reporting, assessment of RWE studies often requires substantial resources within regulatory and other organisations.

Despite recommendations from the International Committee of Medical Journal Editors that the methods sections of research publications should provide enough detail so that others with access to the data would be able to reproduce the results,6 attempts to replicate results from database studies have been hampered by a lack of clarity in reporting on critical study implementation details.7 8 9 10 11 Many organisations recognise this problem and have created guidelines and checklists for research reporting.12 13 14 15 16 17 18 19 20 21 22 Existing guidelines and checklists already have a strong consensus regarding what main elements are important to report. However, these guidelines are general in order to cover a broad base—which leaves room for ambiguity, assumptions, and misinterpretation when planning and implementing RWE studies.7 8

The multidisciplinary, multidatabase, and collaborative nature of RWE study design and conduct would be improved by clearer communication of critically important details. This need is particularly relevant for common protocol studies involving collaboration between multiple groups, where different interpretation by the groups executing a protocol can substantially influence results.23 Unambiguous documentation of a research team’s intended study implementation parameters can increase the effectiveness of communication within the multidisciplinary study team as well as between research teams and decision makers reviewing their studies.

Ambiguity in communication of key design and analytical details as well as data sources and their origins makes it difficult for reviewers to assess potential for bias or evaluate the robustness of study findings, and could limit the use of RWE studies in regulatory and other healthcare decision making.24 Recognising the high variability in quality and completeness of communication about RWE study implementation,7 8 an increasing number of stakeholders have moved towards routine pre-registration of RWE studies with fully specified protocols to support regulatory and coverage decisions.25 The International Council for Harmonization has set short term strategic goals of harmonising the structure and format of protocols and reporting documents in regulatory submissions to increase the acceptability of RWE studies.

Regulators are increasingly calling for high levels of transparency as an integral part of the science of RWE. Although the need for transparent and reproducible evidence was recognised earlier, the coronavirus disease 2019 pandemic has brought the field to an inflection point. Currently, timely high quality evidence is urgently needed to inform decision making. Such evidence can be obtained through rigorous analysis of routinely collected healthcare data. However, an influx of high profile studies with methodological issues has negatively affected the credibility of evidence derived from analysis of non-randomised healthcare databases. In order to have a foundation from which to evaluate and distinguish useful, rigorously designed studies from studies with validity problems, regulators and other decision makers have called for researchers to conduct and report RWE studies using standards agreed on by professional societies dedicated to RWE. Based on best scientific practice,26 this tool aims to resolve the identified gap in transparency on RWE study methods.

Summary points.

  • Compared with clinical trials and non-experimental studies that prospectively collect data, studies that use routinely collected electronic healthcare data have a greater variability in design and analysis options

  • Existing guidelines and checklists have a strong consensus regarding what main elements are important to report, but they can lead to ambiguity, assumptions, and misinterpretation when planning and implementing RWE studies

  • An increasing number of stakeholders have moved towards routine registration of RWE studies with fully specified study implementation protocols to support regulatory and coverage decisions

  • Through a public-private collaboration with broad and international stakeholder input, a structured template for planning and reporting on RWE study implementation (STaRT-RWE) has been developed

  • STaRT-RWE is intended to serve as a didactic tool for designing and conducting good RWE studies; set clear expectations for communication of RWE methods; reduce misinterpretation of prose that lacks specificity; allow reviewers to quickly find key information; and facilitate reproducibility, validity assessment, and evidence synthesis

  • The template has been endorsed by the International Society of Pharmacoepidemiology (ISPE) and the Transparency Initiative led by the International Society of Pharmacoeconomics and Outcomes Research in partnership with ISPE, Duke Margolis Health Policy, and the National Pharmaceutical Council

Structured template and reporting tool for real world evidence (STaRT-RWE)

While the simplicity of a checklist is ideal for summarising areas to report, it leaves room for misinterpretation and ambiguity about important details of study implementation. We complement the checklist approach by developing a study implementation template where methods related items from existing checklists correspond to the main headings in structured tables where critical details are communicated. Design details are visually summarised in a figure based on a framework for graphical depiction of longitudinal study design.27 The template is intended to support both research planning and communication of methods. As such, it should be completed when conceptualising and designing a study, updated with version changes during implementation, and then shared with the final study results at the time of submission to enable review and replication. The template tables can be referenced if they were previously used in study registration materials or publicly posted. Alternatively, they can be published as supplemental material. Used in this way, the template will increase the rigor of study conduct as well as support understanding of methods and interpretation of results by decision makers.

The study implementation template is designed to fulfil several aims: serve as a guiding tool for designing and conducting reproducible RWE studies; set clear expectations for transparent communication of RWE methods; reduce misinterpretation of prose that lacks specificity; allow reviewers to quickly orient and find key information; and facilitate reproducibility, validity assessment, and evidence synthesis. This multipurpose tool is particularly relevant for hypothesis evaluating studies on treatment effectiveness and safety28 that are intended to influence a regulatory or coverage decision, but is also compatible with RWE studies of various types (eg, exploratory, descriptive, prediction).

Development of a template for planning and reporting on RWE study implementation

The detailed study parameters requested in the structured template on study implementation are from a consensus document developed by a joint task force between the International Society for Pharmacoepidemiology (ISPE) and International Society for Pharmacoeconomics and Outcomes Research (ISPOR).15 The consensus document includes a detailed catalogue of specific parameters that represent key scientific and operational decisions made when implementing database studies to facilitate reproduction, facilitate replication, as well as evaluate validity. Many of these parameters are incorporated into existing reporting checklists at a high summary level, 14 19 as areas which researchers should provide detail on. However, the type and nature of detail have been explicitly delineated in the STaRT-RWE template for study implementation.

We cross referenced the catalogue in the ISPE-ISPOR consensus document against the design and analysis sections of existing reporting checklists for non-experimental studies such as RECORD-PE14 and ENCePP19 as well as against published quality assessment tools29 30 31 to ensure that key elements for assessing validity were covered (fig 1). We then created a series of structured tables detailing these study implementation parameters, accompanied by graphical depiction of longitudinal study design.27 The STaRT-RWE template tries to strike a balance between modularised response options while retaining flexibility. Prose is limited in the structured template tables, therefore the structured tables detailing study parameters should be accompanied by prose based background and rationale for scientific decisions, as well as interpretation of results and discussion in manuscripts or reports. Unlike reporting checklists, which have a more limited set of items that are requested somewhere in a paper, STaRT-RWE helps investigators walk through their intended scientific decisions in organised, unambiguous detail when planning their study. This implementation plan can then made available in online appendices when reporting on the completed study.

Fig 1.

Fig 1

Development of template for planning and reporting on real world evidence (RWE) study implementation.15 ISPE=International Society of Pharmacoepidemiology; ISPOR=International Society of Pharmacoeconomics and Outcomes Research; Transparency Initiative=multistakeholder group led by ISPOR in partnership with ISPE, Duke Margolis Health Policy, and National Pharmaceutical Council

STaRT-RWE template tables and figure

In close alignment with the detailed elements identified in the joint ISPE-ISPOR task force consensus document, the template includes several tables that are detailed below (appendix 1). We have included the template tables with example entries (highlighted in yellow) to show how the tables could be completed for a simple, comparative cohort study of a time-to-event outcome in pharmacoepidemiology. In practice, the template table entries would depend on the design and analysis parameters selected for the study. Detailed instructions integrated in the header text of the template tables can be deleted after the table is populated.

Administrative information

Administrative information is summarised in the first three tables of the template (appendix 1), including a table of contents, the study title, primary and secondary PICOT objectives (population, intervention, comparator, outcome, and time horizon), other administrative details as appropriate, and a version history documenting not only what changed, but also why it was changed, and when the change was made—documenting the history of the planned analyses.

Study design

The study design diagram (fig 2) provides a concise summary of how the analytical cohort was created. It is read from top to bottom, showing the sequence of actions taken to create the cohort. The vertical arrow denotes day 0, the point at which the study entry defining criterion is met. The size and location of the horizontal bars visually show the temporality of assessment windows for inclusion or exclusion criteria, covariates, and follow-up relative to day 0, while number ranges within each bar use standard mathematical notation to explicitly denote the temporality. The diagram can also have footnotes providing high level details about what is defined within each of the assessment windows. Details were published earlier this year, along with Microsoft PowerPoint templates.27

Fig 2.

Fig 2

Example study design diagram comparing use of macrolide versus fluoroquinolone on Clostridioides Difficile

Study population

The study population parameter table (appendix table 3) provides more detail than the design diagram, starting with metadata about the data source, any data linkages, and name or version of software used to create the analytical cohort. In subsection A of appendix table 3, the data source name and version are identified, as well as any sampling criteria applied (eg, the data cut only includes patients with a diagnosis of diabetes). If data linkage was involved, the table has a space to provide a citation or refer to an appendix with description of the linkage (eg, how it was done, performance characteristics).

Remaining template sections

The remaining subsections of this table detail, in a structured way, the operational definitions for the index date for cohort entry, exposures, inclusion or exclusion criteria, and order of application, covariates, outcomes, and follow-up time (appendix table 3B-H). These subsections summarise what is measured, timing of measurement, care setting (eg, inpatient, outpatient, emergency department), type of codes used to define the measure (eg, drug, diagnosis, procedure or laboratory codes), as well as sources for the algorithms (eg, publication, clinician review). For algorithms based on diagnosis codes, the template includes a section for researchers to define whether codes are required to be in the primary position (indicating that the diagnosis is the main reason for the encounter). The actual clinical codes used to define each measure are specified in appendices to the template tables; these appendices are structured and machine readable.

A subsection (appendix table 3B) focuses on detailing the day 0 defining criteria for entry to the study population. This subsection includes fields to specify how many times patients can enter the cohort, the type of entry, washout windows for incident entry if relevant, and what defines incident entry (eg, new users of a drug or newly diagnosed with a particular condition). The subsections on inclusion and exclusion criteria (appendix table 3C-D) have dropdown fields for specifying whether the criteria are applied before or after selection of the date for entry to the analytic study population. For covariates (appendix table 3E), how the variable is operationalised is also defined (eg, continuous, binary, categorical), with the option to provide more detail about transformations. The template includes an optional section on empirically defined covariates (appendix table 3F), where the source code or reference for the algorithm used for data driven selection of covariates are provided (eg, high dimensional propensity score algorithm32 33). Throughout the table, fields indicate whether the study parameter was prespecified and whether it was varied in sensitivity analyses.

In the outcome subsection (appendix table 3G), the primary outcome is specified along with measurement characteristics. Outcome measurement performance could come from published algorithm validation papers, from direct validation of outcomes within the analytical cohort, or researchers could explicitly note that there is no information about the accuracy of outcome ascertainment. The type of outcome and related washout windows to define incidence if relevant, care settings, and diagnosis positions are also specified in this section.

In the follow-up subsection (appendix table 3H), investigators define when follow-up begins relative to cohort entry and how it ends, which can guide investigators to consider each option and make an active choice. It also makes clear for the reviewer what are and are not used as censoring criteria. If censoring on discontinuation is selected, the table includes sections to define duration of exposure, including a grace period to allow for non-adherence, to define whether or how an algorithm is applied to handle early refills, or build in an induction window for the hypothesised biologic window of effect of exposure or carry over effects for the effect of a drug beyond the end of exposure.

Additional tables

Analysis specifications are provided in a separate table (appendix table 4), including fields for the hypothesis being tested and study population for the primary analysis and relevant subgroup analyses, followed by the same information for secondary analyses in subsequent sections. The software packages can be reported here, the models that are fit, the type of confounding adjustment and variables included, with specification of some parameters such as matching ratio and caliper, formulas for weights, trimming, and truncation rules. This table also includes fields to specify how missing data are handled in the analysis and subgroup analyses (eg, excluding patients with missing or unknown sex and multiple imputation for missing total cholesterol values).

A separate table for sensitivity analyses (appendix table 5) also exists, where investigators can specify which parameters are being changed, provide rationale for why they are being varied and detail what investigators expect to learn from this sensitivity analysis compared to the primary analysis.

Additional study population table shells that suggest the presentation of findings are provided (appendix tables 6 and 7), including an attrition table showing the counts as inclusion and exclusion criteria are applied to the source data and a power and sample size calculation table for feasibility counts, if relevant. The exact contents of this attrition table can vary depending on the type of calculation. A glossary of terms helps avoid misinterpretations, which over time will support standardisation of the terminology used by investigators and reviewers (appendix table 8). A list of abbreviations is also included (appendix table 9).

The template tables should always be accompanied by appendices that contain specific clinical code lists (eg, International Classification of Diseases-Clinical Modification, 9th revision; Current Procedural Terminology; READ codes) used to define study entry criteria or exposure, inclusion or exclusion, confounders, outcomes, and structured to be machine read or writeable. These tables can also be accompanied, if relevant, by additional appendices that detail decisions made during the conversion process from source data to a common data model, which provide more information about data linkages or other data processing steps, as well as appendices that contain or provide links to code used to create and analyse the study population. The focus group findings are summarised in appendix 2, arranged by main themes regarding the value, challenges, and usability for international RWE stakeholders.

STaRT-RWE example library from published studies

We developed a library of STaRT-RWE examples featuring common use cases involving different study designs, sources of data, and more complex algorithms to define key study parameters. These four case studies, based on published studies, aim to increase usability and facilitate adoption by providing tangible examples of how the template would be populated for different study designs and data sources. These examples are easily modifiable to fit similar use cases and include a comparative effectiveness cohort study, a predictive outcome modelling study using data from linked claims and electronic health records, a study of drug treatment safety in pregnancy, and a study using a self-controlled design (appendix 3). An example of structured appendices containing code lists that are machine readable or writable to accompany the STaRT-RWE study implementation template tables is provided (appendix 3).

Discussion

A key hurdle for RWE studies on the use, effectiveness, and safety of medical products to be considered by decision makers is the perceived ambiguity and lack of detail on complex design and analysis choices. Through a public-private collaboration with broad, international stakeholder input, we have developed a structured template for planning and reporting on RWE study implementation. The template is intended to aid study design, conduct, and review in a way that adds value for researchers, sponsors, reviewers, and decision makers. STaRT-RWE is not a reporting checklist but can complement existing checklists by providing guidance and a common structure to help research teams be clear and comprehensive when planning and communicating critical details of intended scientific decisions. By using tabular and visual formats, the template (unlike a checklist) minimises ambiguous prose and potential for misinterpretation.

STaRT-RWE is compatible with multiple study designs and RWD sources. It accords with an earlier recommendation by a joint ISPE-ISPOR consensus document on specific parameters that are necessary to make RWE research reproducible15 as well as published bias assessment tools.29 30 31 The study implementation template aims to be used in conjunction with unstructured prose based introductions and discussions detailing background and rationale, investigator’s interpretation of results, and noted strengths and limitations of the study. On completion of a study, the STaRT-RWE tables and figure detailing the study methods should be accompanied by appropriate tables and figures to communicate the results.

The planning and reporting of clinical trials are standardised. STaRT-RWE parallels the trial setting by providing a structured way to communicate complex RWE study design and analysis choices. The goal is not to standardise how people do the science of RWE, but rather to provide a framework for researchers to work through the details of design and analysis with a common understanding of the decisions that are made.

Clinical research studies are currently expected to provide a CONSORT diagram to describe how the study population was derived. We hope that RWE studies will include a “figure 1” for the study design diagram and an “appendix 1” for structured study parameter tables to clearly communicate the details of how the population, exposure, outcomes, and covariates being studied were derived from longitudinal streams of healthcare data not collected for research.

The structured template we have developed is operationally focused, tabular, and visual, minimising prose and reducing the potential for misinterpretation. The template’s study parameter tables and figure are designed to be flexible enough to incorporate into existing processes for protocol development and reporting. We expect that after the initial learning curve and refinements, use of the template would become routine with practice. Furthermore, software solutions can be developed to read input or output study tables, design diagrams, and appendices with a common structure.3 8 34 35 36 37 Given the trend towards registration of hypothesis confirming RWE studies,28 STaRT-RWE or a subset of core elements could be integrated with RWE study registration sites to simplify the registration process and improve standardised searches. However, many studies aim to raise hypotheses that need to be evaluated rather than confirm them. Registration might be less imperative in such settings. Furthermore, we recognise that other ways of communicating and implementing shared protocols already exist. Investigators who are unable to use the template should not be discouraged from publishing and reporting their observations.

Sharing data and code allows computational reproduction. STaRT-RWE facilitates replicability and validity assessment. Sharing of data and code should be encouraged when possible to allow computational reproduction. However, for research using routinely collected electronic healthcare data, public sharing of source data or data derivatives are typically not permissible. Furthermore, providing code does not substitute for clear communication of the intended study implementation parameters. A concern raised by stakeholders is that a narrow focus on computational reproducibility could encourage investigative teams to be less careful in communication of methods. The detailed logic flow contained in the template will provide more insight into the intended scientific design parameters than can be found in code by most readers. Additionally, code to create analytical study populations from source data are not readily transferable between data sources that have different database schemes whereas the information in the template is transferable. By providing the detailed study recipe that most reviewers would not be able to parse from code, the template facilitates validity assessment, independent reproducibility in the same data source, and replication in different data sources.

Limitations

The STaRT-RWE template has a few limitations. Firstly, while the template is designed to be flexible, the structure imposed in the tables might be an awkward fit for some use cases. Depending on the study, only a subset of tables in STaRT-RWE might be relevant. Secondly, use of the study implementation template to guide the planning of a study and provide clarity about scientific decisions does not guarantee that those decisions will result in unbiased findings. However, appropriate interpretation of study findings will be greatly facilitated if reviewers see unambiguous information on how those findings were derived, and what strategies were used to mitigate potential biases.

Thirdly, the template focuses primarily on study implementation decisions. Although the template includes a section about the data sources, the fields in the template do not capture all of the information needed to assess whether the data are fit-for-purpose. The fitness of data sources should be clearly documented, including a description of the available data fields and completeness of data capture (eg, closed v open system, inpatient v outpatient, primary v specialty care). They should also be accompanied by clear data provenance documenting transformations performed on the data streams used to create the research database. However, this level of metadata about the data source was not included in the STaRT-RWE template because we thought it would be more efficient for documentation to be created, maintained, and made publicly available as a citable resource for each RWD source, rather than having the detailed data source documentation repeated for every RWE study that uses the data source.

Dissemination plan

To harmonise protocol and reporting standards for regulatory submissions and coverage decisions,26 more work will be necessary, including dissemination of STaRT-RWE and discussion with regulators, health technology assessment agencies, and other relevant stakeholder organisations; development of training modules; use of the template for protocols posted in study registries such as EU-PASS or other governance processes; coordination with study registration sites to enable electronic uploading of the completed template when pre-registering a RWE study; and engagement of medical journal editors. Collection of feedback from broad stakeholder groups with practical experience of the template will be critical to inform updates. Multiple workstreams have been initiated to implement these next steps, support researcher training, and integrate STaRT-RWE with existing processes.25 26 38

Conclusion

STaRT-RWE is designed to reduce misinterpretation and create clear expectations regarding communication of how RWE is generated—expectations about which study design and analysis details should be reported as well as how and where that information should appear. The study implementation template is not a substitute for a well trained pharmacoepidemiologist or outcomes researcher, and unambiguous communication of study implementation is not necessarily equivalent to highly valid, rigorous study methodology. However, the availability of a baseline level of information about study methodology in a consistent structure would improve clarity regarding study implementation decisions between study investigators and decision makers. This information would increase healthcare decision makers’ ability to effectively evaluate RWE study validity.

Acknowledgments

We thank the following people for their participation in focus groups and invaluable feedback on STaRT-RWE: Sarah Priddy Alwardt (McKesson Specialty Health), Elande Baro (US Food and Drug Administration (FDA)), Pamela Bradt (Institute for Clinical and Economic Review), K Arnold Chan (National Taiwan University), Nam-Kyong Choi (Ewha Womans University), Noelle Cocoros (Department of Population Medicine, Harvard Medical School and Harvard Pilgrim Health Care Institute), Paul Coplan (Johnson and Johnson), Efe Eworuke (US Food and Drug Administration), Jesper Hallas (International Society for Pharmacoepidemiology (ISPE)), Christian Hampp (FDA), Laura Happe (Journal of Managed Care and Specialty Pharmacy), Katy Harrison (National Institute for Health and Care Excellence), Javier Jimenez (Sanofi), Clara Kim (FDA), Stefanie Kraus (FDA), Xavier Kurz (European Medicines Agency), Sinéad Langan (London School of Hygiene and Tropical Medicine), Nirosha Mahendraratnam Lederer (Duke-Margolis Center for Health Policy), Judith Maro (Sentinel Operations Center, Harvard Medical School, Harvard Pilgrim Health Care Institute), David Martin (FDA), Katherine McAllister (National Institute for Health and Care Excellence), David Mellor (Center for Open Science), Michael Nguyen (FDA), John Seeger (Optum Epidemiology), Richard Scott Swain (FDA), Peter Tugwell (Ottawa Hospital Research Institute, University of Ottawa), Richard Willke (Professional Society for Health Economics and Outcomes Research), Almut Gertrud Winterstein (University of Florida College of Pharmacy, Center for Drug Evaluation and Safety), and Xiaofeng Zhou (Pfizer). We also thank the following groups for their feedback on STaRT-RWE: ISPE executive board; steering group of the Transparency Initiative being led by the International Society of Pharmacoeconomics and Outcomes Research (ISPOR) in partnership with ISPE, Duke-Margolis Health Policy, and National Pharmaceutical Council; and RECORD committee co-chairs. The views expressed in this article are personal views of the authors and may not be understood or quoted as being made on behalf of or reflecting the position of the FDA, European Medicines Agency, Pharmaceuticals and Medical Devices Agency, or other organisations with which the authors are affiliated.

Web extra.

Extra material supplied by authors

Appendix 1: STaRT-RWE template tables and figure

wans059917.ww1.pdf (241.8KB, pdf)

Appendix 2: Summary of focus group feedback

wans059917.ww2.pdf (97.1KB, pdf)

Appendix 3: Summary of the library of STaRT-RWE examples

wans059917.ww3.pdf (123.2KB, pdf)

Contributors: SVW and SS conceived of the idea of the template and acquired funding. SVW and LGB managed project administration. All authors contributed to investigation and methodology, writing, review, and editing of the manuscript. SVW is guarantor. The corresponding author attests that all listed authors meet authorship criteria and that no others meeting the criteria have been omitted.

Funding: Development of the study implementation template was supported in two stages of grants to Brigham and Women’s Hospital. Regulatory collaborators did not receive funding from industry. The initial development of the template was supported by grants to Brigham and Women’s Hospital from Boehringer Ingelheim, Novartis, and Johnson and Johnson. The second phase of template development was supported by US FDA grant HHSF223201810136C to Brigham and Women’s Hospital. This second phase also included focus groups and preparation of a manuscript. The FDA sponsors participated in the design of the focus groups and interpretation of the data. Each of the study sponsors participated in the preparation of the manuscript.

Competing interests: All authors have completed the ICMJE uniform disclosure form at www.icmje.org/coi_disclosure.pdf and declare: support from Boehringer Ingelheim, Novartis, Johnson and Johnson, and the FDA for the submitted work; SVW received salary support from investigator-initiated grants to Brigham and Women’s Hospital from Boehringer Ingelheim, Johnson and Johnson, and Novartis Pharmaceuticals; SS is consultant to Aetion, a software manufacturer of which he also owns shares. He is principal investigator of investigator initiated grants to Brigham and Women’s Hospital from Boehringer Ingelheim unrelated to the topic of this study; JAB is employed by Johnson and Johnson; DBB is employed by UCB Pharmaceuticals; KHK is employed by Novartis Pharmaceuticals.

Ethical approval: This study was approved by the institutional review board of Brigham and Women’s Hospital.

Data sharing: The STaRT-RWE template and examples will be publicly available. Additional details of the focus groups beyond what is summarised in the appendices are available on request.

The corresponding author affirms that the manuscript is an honest, accurate, and transparent account of the study being reported; that no important aspects of the study have been omitted; and that any discrepancies from the study as planned have been explained.

Patient and public involvement: We sought feedback on the usability, value, and challenges of using the structured study implementation template via focus groups with international leaders involved in generating or reviewing RWE for decision making. Focus group participants included regulators, health technology assessors, industry, journal editors, payers, and leaders of large distributed data networks. After revising the template based on their suggestions, we sought feedback from ISPE and ISPOR membership and received endorsement from both international societies.

Dissemination to participants and related patient and public communities: We intend to disseminate STaRT-RWE to international stakeholder groups involved with generating or using real world evidence through publication, conference presentations, as well as multistakeholder efforts focused on increasing transparency of real world evidence studies. These efforts will involve collaboration and coordination with ISPE, ISPOR, the International Council of Harmonization workgroup on harmonizing technical protocols, Transparency Initiative, clinicaltrials.gov, Open Science Framework, European Medicines Agency’s Big Data Task Force, EU Electronic Register of Post-Authorisation Studies, European Network of Health Technology Assessment, the EQUATOR network, and others.

Provenance and peer review: Not commissioned; externally peer reviewed.

References

  • 1.34-21st Century Cures Act., Pub. L. No. Public Law No: 114-255.
  • 2.Eichler HG, Baird LG, Barker R, et al. From adaptive licensing to adaptive pathways: delivering a flexible life-span approach to bring new drugs to patients. Clin Pharmacol Ther 2015;97:234-46. 10.1002/cpt.59  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Ball R, Robb M, Anderson SA, Dal Pan G. The FDA’s sentinel initiative-a comprehensive approach to medical product surveillance. Clin Pharmacol Ther 2016;99:265-8. 10.1002/cpt.320  [DOI] [PubMed] [Google Scholar]
  • 4.Sun X, Tan J, Tang L, Guo JJ, Li X. Real world evidence: experience and lessons from China. BMJ 2018;360:j5262. 10.1136/bmj.j5262  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Makady A, Ham RT, de Boer A, Hillege H, Klungel O, Goettsch W, GetReal Workpackage 1 . Policies for use of real-world data in health technology assessment (HTA): a comparative study of six HTA agencies. Value Health 2017;20:520-32. 10.1016/j.jval.2016.12.003  [DOI] [PubMed] [Google Scholar]
  • 6.International Committee of Medical Journal Editors. Recommendations for the conduct, reporting, editing and publication of scholarly work in medical journals. 2021. http://www.icmje.org/recommendations/. [PubMed]
  • 7.Wang SV, Verpillat P, Rassen JA, Patrick A, Garry EM, Bartels DB. Transparency and reproducibility of observational cohort studies using large healthcare databases. Clin Pharmacol Ther 2016;99:325-32. 10.1002/cpt.329  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Reproducible Evidence: Practices to Enhance and Achieve Transparency (REPEAT) Initiative. https://www.repeatinitiative.org/.
  • 9.Matcho A, Ryan P, Fife D, Reich C. Fidelity assessment of a clinical practice research datalink conversion to the OMOP common data model. Drug Saf 2014;37:945-59. 10.1007/s40264-014-0214-3  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Lalmohamed A, van Staa TP, Vestergaard P, et al. Statins and risk of lower limb revision surgery: the influence of differences in study design using electronic health records from the United Kingdom and Denmark. Am J Epidemiol 2016;184:58-66. 10.1093/aje/kwv311  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.de Vries F, de Vries C, Cooper C, Leufkens B, van Staa T-P. Reanalysis of two studies with contrasting results on the association between statin use and fracture risk: the General Practice Research Database. Int J Epidemiol 2006;35:1301-8. 10.1093/ije/dyl147  [DOI] [PubMed] [Google Scholar]
  • 12.US Food and Drug Administration. Guidance for industry and FDA staff: best practices for conducting and reporting pharmacoepidemiologic safety studies using electronic healthcare data. 2013.
  • 13.Public Policy Committee, International Society of Pharmacoepidemiology . Guidelines for good pharmacoepidemiology practice (GPP). Pharmacoepidemiol Drug Saf 2016;25:2-10. 10.1002/pds.3891  [DOI] [PubMed] [Google Scholar]
  • 14.Langan SM, Schmidt SA, Wing K, et al. The reporting of studies conducted using observational routinely collected health data statement for pharmacoepidemiology (RECORD-PE). BMJ 2018;363:k3532. 10.1136/bmj.k3532  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Wang SV, Schneeweiss S, Berger ML, et al. joint ISPE-ISPOR Special Task Force on Real World Evidence in Health Care Decision Making . Reporting to improve reproducibility and facilitate validity assessment for healthcare database studies V1. 0. Pharmacoepidemiol Drug Saf 2017;26:1018-32. 10.1002/pds.4295  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Vandenbroucke JP, von Elm E, Altman DG, et al. STROBE Initiative . Strengthening the Reporting of Observational Studies in Epidemiology (STROBE): explanation and elaboration. PLoS Med 2007;4:e297. 10.1371/journal.pmed.0040297  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.European Medicines Agency. ENCePP guide on methodological standards in pharmacoepidemiology. 2014.
  • 18.European Medicines Agency. The ENCePP code of conduct for scientific independence and transparency in the conduct of pharmacoepidemiological and pharmacovigilance studies. 2018. http://www.encepp.eu/code_of_conduct/documents/ENCePPCodeofConduct.pdf.
  • 19.European Medicines Agency. ENCePP checklist for study protocols (revision 4). 2018. http://www.encepp.eu/standards_and_guidances/documents/ENCePPChecklistforStudyProtocols.doc.
  • 20.European Medicines Agency. Guideline on good pharmacovigilance practices (GVP) Module VIII—post-authorisation safety studies (revision 3). 2017.
  • 21.Berger ML, Martin BC, Husereau D, et al. A questionnaire to assess the relevance and credibility of observational studies to inform health care decision making: an ISPOR-AMCP-NPC Good Practice Task Force report. Value Health 2014;17:143-56. 10.1016/j.jval.2013.12.011  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.US Food and Drug Administration. Submitting documents using real-world data and real-world evidence to FDA for drugs and biologics guidance for industry. 2019. FDA-2019-D-1263.
  • 23.Klungel OH, Kurz X, de Groot MC, et al. Multi-centre, multi-database studies with common protocols: lessons learnt from the IMI PROTECT project. Pharmacoepidemiol Drug Saf 2016;25(Suppl 1):156-65. 10.1002/pds.3968  [DOI] [PubMed] [Google Scholar]
  • 24.Malone DC, Brown M, Hurwitz JT, Peters L, Graff JS. Real-world evidence: useful in the real world of US payer decision making? How? When? And what studies? Value Health 2018;21:326-33. 10.1016/j.jval.2017.08.3013  [DOI] [PubMed] [Google Scholar]
  • 25.Real-World Evidence Transparency Initiative. A partnership between ISPOR, the International Society for Pharmacoepidemiology, the Duke-Margolis Center for Health Policy, and the National Pharmaceutical Council. 2020. https://www.ispor.org/strategic-initiatives/real-world-evidence/real-world-evidence-transparency-initiative.
  • 26.International Council for Harmonisation of Technical Requirements for Pharmaceuticals for Human Use. ICH reflection paper: strategic approach to international harmonization of technical scientific requirements for pharmacoepidemiological studies submitted to regulatory agencies to advance more effective utilization of real-world data 2019. 2019. https://admin.ich.org/sites/default/files/2019-08/ICH_ReflectionPaper_Pharmacoepidemiology_2019_0605.pdf.
  • 27.Schneeweiss S, Rassen JA, Brown JS, et al. Graphical depiction of longitudinal study designs in health care databases. Ann Intern Med 2019;170:398-406. 10.7326/M18-3079  [DOI] [PubMed] [Google Scholar]
  • 28.Berger ML, Sox H, Willke RJ, et al. Good practices for real-world data studies of treatment and/or comparative effectiveness: recommendations from the joint ISPOR-ISPE Special Task Force on real-world evidence in health care decision making. Pharmacoepidemiol Drug Saf 2017;26:1033-9. 10.1002/pds.4297  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Sterne JA, Hernán MA, Reeves BC, et al. ROBINS-I: a tool for assessing risk of bias in non-randomised studies of interventions. BMJ 2016;355:i4919. 10.1136/bmj.i4919  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Dreyer NA, Bryant A, Velentgas P. The GRACE checklist: a validated assessment tool for high quality observational studies of comparative effectiveness. J Manag Care Spec Pharm 2016;22:1107-13. 10.18553/jmcp.2016.22.10.1107  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.US Food and Drug Administration. Guidance for industry and FDA staff: best practices for conducting and reporting pharmacoepidemiologic safety studies using electronic healthcare data. 2013. https://www.fda.gov/regulatory-information/search-fda-guidance-documents/best-practices-conducting-and-reporting-pharmacoepidemiologic-safety-studies-using-electronic.
  • 32.Schneeweiss S, Rassen JA, Glynn RJ, Avorn J, Mogun H, Brookhart MA. High-dimensional propensity score adjustment in studies of treatment effects using health care claims data. Epidemiology 2009;20:512-22. 10.1097/EDE.0b013e3181a663cc  [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Bross ID. Spurious effects from an extraneous variable. J Chronic Dis 1966;19:637-47. 10.1016/0021-9681(66)90062-2  [DOI] [PubMed] [Google Scholar]
  • 34.Franklin JM, Schneeweiss S. When and how can real world data analyses substitute for randomized controlled trials? Clin Pharmacol Ther 2017;102:924-33. 10.1002/cpt.857  [DOI] [PubMed] [Google Scholar]
  • 35.RCT DUPLICATE Project . 2020. https://www.rctduplicate.org/fda-demonstration-project.html.
  • 36.Schuemie MJ, Gini R, Coloma PM, et al. Replication of the OMOP experiment in Europe: evaluating methods for risk identification in electronic health record databases. Drug Saf 2013;36(Suppl 1):S159-69. 10.1007/s40264-013-0109-8  [DOI] [PubMed] [Google Scholar]
  • 37.Sentinel. Routine querying system 2020. 2020. https://www.sentinelinitiative.org/methods-data-tools/routine-querying-tools/routine-querying-system.
  • 38.International Society of Pharmacoepidemiology Real World Evidence Task Force . 2020. https://www.pharmacoepi.org/strategic-initiatives/rwe-task-force/.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Appendix 1: STaRT-RWE template tables and figure

wans059917.ww1.pdf (241.8KB, pdf)

Appendix 2: Summary of focus group feedback

wans059917.ww2.pdf (97.1KB, pdf)

Appendix 3: Summary of the library of STaRT-RWE examples

wans059917.ww3.pdf (123.2KB, pdf)

Articles from The BMJ are provided here courtesy of BMJ Publishing Group

RESOURCES