Abstract
Background
Real-world evidence (RWE) refers to observational health care data beyond clinical trial data. It holds the promise of transforming health care as a new form of evidence to support decision makers in making decisions when developing and regulating medicines. As the importance of RWE is recognized by industry and regulatory bodies, teaching RWE becomes an important matter to evaluate and refine in order to develop future researchers and stakeholders who can better integrate RWE into the routine development of medicine.
Objective
The aim of this review is to understand how RWE is currently being taught. From this landscape study, the insufficiencies of the current education of RWE can be identified and subsequently inform future education policies around RWE and its subfacets.
Methods
We will search MEDLINE, EMBASE, PsycINFO, Healthcare Management Information Consortium, Cochrane, and Web of Science for published studies using a combination of keywords and subject headings related to RWE and education. In addition, a Google search to identify grey literature will be conducted. Two authors will independently screen the titles and abstracts identified from the search and accept or reject the studies according to the study inclusion criteria; any discrepancies will be discussed and resolved. The quality of the included literature will be assessed using the Critical Appraisal Skills Programme systematic review checklist.
Results
Data from eligible publications will be abstracted into a predesigned form in order to better understand the current state of education of RWE and inform future RWE education directions and policies.
Conclusions
The subsequent systematic review will be published in a peer-reviewed journal.
International Registered Report Identifier (IRRID)
PRR1-10.2196/16941
Keywords: real-world evidence, teaching, public health
Introduction
Real-world evidence (RWE) is a subset of evidence-based medicine that refers to health care information gathered through means outside of the typical clinical research settings. Evidence-based medicine (EBM) is the use of best available clinical evidence from systematic research combined with clinical expertise to deliver the best possible clinical care to patients [1,2]. Since its advent in the 1990s, EBM has been shown to be the cornerstone of the medical profession, raising awareness of using reliable, published evidence to aid decision making in medicine [3,4]. Sources of RWE data can come from electronic health records (EHRs), health surveys, claims and billing data, product and disease registries, mobile health apps, and personal smart devices. Collectively, the real RWE base can generate invaluable insights and findings on diseases, products, and patient populations [5,6].
Regulatory bodies around the world are gaining interest in RWE. Notably, the US Food and Drug Administration has recently announced a $100 million project to build a modern system that will gather RWE from approximately 10 million individuals [7], and the European Union has funded over 170 initiatives related to RWE, of which 65 received over 734 million euros of public funding [8]. According to a report from McKinsey and Company [9], pharmaceutical companies have recognized the impact of RWE and are actively applying it to safety, postmarket, and end-to-end product development to facilitate research and development and commercial and safety decisions.
The anticipated importance of data collection and analysis of RWE in the future of clinical trials and development of medicines is evident. In order to meet the future demands of RWE researchers in terms of realizing the true potential of RWE in transforming health care, it is important to educate stakeholders such as researchers, clinicians, and policy makers in RWE in an appropriate way.
EBM has long been taught to medical professionals, including clinicians and nurses [10-12], and due to its importance in clinical care, teaching EBM in medical school has been investigated in various studies. Smith et al [13] conducted a controlled trial to look into the effectiveness of EBM courses for residents. Another recent study conducted by Nasr et al [14] evaluated four EBM workshops taught to residents-in-training and postgraduates in medical school. Interestingly, Slawson et al [15] suggested the importance of information management in the teaching of EBM back in 2005, and with the rapid development of big data, better computer processors, and the maturation of machine learning in recent years, information management is more important than ever in EBM. However, it is unknown how RWE is being taught and what the effects are.
This systematic review aims to answer the following research questions:
What are the current methods used to teach RWE, and what are the effects of those methods?
Who are the stakeholders teaching and learning about RWE?
Methods
Review Conduct
This systematic review will be conducted following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines [16]. The protocol methods will be reported following the 2015 PRISMA checklist [17]. The protocol will be registered with PROSPERO.
Eligibility Criteria
Inclusion Criteria
In accordance with PRISMA Protocols (PRISMA-P) checklist recommendation, the inclusion criteria for this protocol are in accordance with participants, interventions, comparators, and outcomes (PICO). Details of PICO to be included in the review are described in Textbox 1.
Participants, interventions, comparators, and outcomes for review inclusion criteria.
Population:
Participants with real-world evidence teaching
Intervention:
Any form of real-world evidence teaching
Comparator:
No real-world evidence teaching
Outcomes:
Learner-focused outcomes such as attitudes, cognitive changes, learner satisfaction
Study type:
Any study type (study type will not be subjected to any restrictions)
Must describe relevance to real-world evidence and its application in health care
Must describe a teaching method
English-language publications
Exclusion Criteria
Education methods that do not describe their relevance to RWE and its applications to health care will be excluded to limit the scope of the review to RWE-focused education and courses. Websites and articles describing RWE (although helpful for those actively seeking to learn more about RWE) will be excluded as there are no methods of teaching included. Studies not published in English are excluded due to the language barrier.
Search Strategy
The following databases will be searched: MEDLINE, EMBASE, PsycINFO, Healthcare Management Information Consortium, Cochrane, and Web of Science. In addition, a Google search for grey literature such as blog posts, opinion pieces, press releases, and online courses will be conducted. Online course platforms such as Coursera, edX, FutureLearn, and OpenClassrooms will be searched to identify relevant courses. Textbox 2 shows the search concept and keywords to be searched for this review. Search strings will be constructed using a combination of RWE-related and education-related keywords.
Concepts and keywords for search term development.
Real-world evidence:
Real world evidence, RWE, big data analytics, real world data, electronic health record*
Teaching:
Medical education, medical student*, medical curriculum, medical school*, health professionals
Study Selection
EndNote X8 software (Clarivate Analytics) will be used for the removal of duplicates. Textbox 1 describes the inclusion criteria of the review. Two independent reviewers will screen the titles and abstracts of papers to minimize the risk of bias and the risk of not including eligible papers due to oversight. Papers that are ineligible will be eliminated, and the full text of those that appear to meet the review’s eligibility criteria will be obtained and read in full to ensure eligibility. Any contradictions or discrepancies between the reviewers that arise will be discussed until consensus is reached. Valid studies will be assessed for their quality before any extraction of information.
Quality Assessment and Risk of Bias
Two reviewers will independently check each article to minimize bias using the Cochrane Collaboration’s risk of bias tool as described in the Cochrane Handbook for Systematic Review of Interventions [18]. All included articles will be judged for their quality based on the Critical Appraisal Skills Programme systematic review checklist [19] and data analysis.
Data Extraction
Eligible sources will subsequently be reviewed in detail, and key relevant challenges will be extracted, categorized, and recorded into a predesigned Excel spreadsheet (Microsoft Corp). A sample data abstraction form can be found in Figure 1.
Figure 1.
Sample data abstraction form.
Results
A sample search was conducted using PubMed, and the sample search string returned 943 results: (“real world evidence” OR “RWE” OR “big data analytic*” OR “real world data” OR “electronic health record*”) AND (“medical education” OR “medical student*” OR “medical curriculum” OR “medical school*”). The search string will be further fine-tuned in the review.
Discussion
Principal Findings
This study will offer a comprehensive overview of how RWE is taught to different stakeholders of the research and application of RWE in health care. However, traditional means of teaching, such as university lecturies, may not be published and hence may be underrepresented in this protocol.
Conclusions
This protocol will be executed in 2020 and published in a peer-reviewed journal in accordance with PRISMA guidelines. Any deviations in the execution shall be noted in the subsequent systematic review publication. The findings from this review will be used to inform the education strategy of RWE.
Acknowledgments
The authors are grateful to Imperial College London Charing Cross Library Manager and Liaison Librarian, Ms Rebecca Jones, for her expertise in database and search terms selection.
Abbreviations
- EBM
evidence-based medicine
- EHR
electronic health record
- PICO
participants, interventions, comparators, and outcomes
- PRISMA
Preferred Reporting Items for Systematic Reviews and Meta-Analyses
- PRISMA-P
Preferred Reporting Items for Systematic Reviews and Meta-Analyses Protocols
- RWE
real-world evidence
Footnotes
Authors' Contributions: The protocol was written by CL with revisions from MHvV, HC, EM.
Conflicts of Interest: None declared.
References
- 1.Sackett DL. Evidence-based medicine. Semin Perinatol. 1997 Feb;21(1):3–5. doi: 10.1016/s0146-0005(97)80013-4. [DOI] [PubMed] [Google Scholar]
- 2.Guyatt G. Evidence-based medicine. ACP J Club. 1991;114(Mar-Apr):A16. doi: 10.7326/ACPJC-1991-114-2-A16. http://www.acpjc.org/Content/114/2/issue/ACPJC-1991-114-2-A16.htm. [DOI] [Google Scholar]
- 3.Sheridan DJ, Julian DG. Achievements and limitations of evidence-based medicine. J Am Coll Cardiol. 2016 Jul;68(2):204–213. doi: 10.1016/j.jacc.2016.03.600. [DOI] [PubMed] [Google Scholar]
- 4.Djulbegovic B, Guyatt GH. Progress in evidence-based medicine: a quarter century on. Lancet. 2017 Jul;390(10092):415–423. doi: 10.1016/s0140-6736(16)31592-6. [DOI] [PubMed] [Google Scholar]
- 5.Garrison LP, Neumann PJ, Erickson P, Marshall D, Mullins CD. Using real-world data for coverage and payment decisions: the ISPOR Real-World Data Task Force report. Value Health. 2007;10(5):326–335. doi: 10.1111/j.1524-4733.2007.00186.x. https://linkinghub.elsevier.com/retrieve/pii/S1098-3015(10)60470-6. [DOI] [PubMed] [Google Scholar]
- 6.Sherman RE, Anderson SA, Dal Pan GJ, Gray GW, Gross T, Hunter NL, LaVange L, Marinac-Dabic D, Marks PW, Robb MA, Shuren J, Temple R, Woodcock J, Yue LQ, Califf RM. Real-world evidence—what is it and what can it tell us? N Engl J Med. 2016 Dec 08;375(23):2293–2297. doi: 10.1056/NEJMsb1609216. [DOI] [PubMed] [Google Scholar]
- 7.Glimcher PW. PharmExec.com. 2018. Nov 02, [2019-12-10]. FDA's $100 million RWE project: making it all it can be http://www.pharmexec.com/fdas-100-million-real-world-evidence-project-making-it-all-it-can-be.
- 8.Plueschke K, McGettigan P, Pacurariu A, Kurz X, Cave A. EU-funded initiatives for real world evidence: descriptive analysis of their characteristics and relevance for regulatory decision-making. BMJ Open. 2018 Jun 14;8(6):e021864. doi: 10.1136/bmjopen-2018-021864. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Cavlan O, Chilukuri S, Evers M, Westra A. McKinsey and Company. 2018. May, [2018-11-13]. Real-world evidence: from activity to impact https://www.mckinsey.com/industries/pharmaceuticals-and-medical-products/our-insights/real-world-evidence-from-activity-to-impact-in-healthcare-decision-making.
- 10.Mackey A, Bassendowski S. The history of evidence-based practice in nursing education and practice. J Prof Nurs. 2017;33(1):51–55. doi: 10.1016/j.profnurs.2016.05.009. [DOI] [PubMed] [Google Scholar]
- 11.Del Mar C, Glasziou P, Mayer D. Teaching evidence-based medicine. BMJ. 2004 Oct 28;329(7473):989–990. doi: 10.1136/bmj.329.7473.989. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Braithwaite RS. Evidence-based medicine: clinicians are taught to say it but not taught to think it. BMJ Evid Based Med. 2019 Oct;24(5):165–167. doi: 10.1136/bmjebm-2018-110970. http://ebm.bmj.com/cgi/pmidlookup?view=long&pmid=30275103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Smith CA, Ganschow PS, Reilly BM, Evans AT, McNutt RA, Osei A, Saquib M, Surabhi S, Yadav S. Teaching residents evidence-based medicine skills: a controlled trial of effectiveness and assessment of durability. J Gen Intern Med. 2000 Oct;15(10):710–715. doi: 10.1046/j.1525-1497.2000.91026.x. https://onlinelibrary.wiley.com/resolve/openurl?genre=article&sid=nlm:pubmed&issn=0884-8734&date=2000&volume=15&issue=10&spage=710. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Nasr J, Falatko J, Halalau A. The impact of critical appraisal workshops on residents' evidence based medicine skills and knowledge. Adv Med Educ Pract. 2018;9:267–272. doi: 10.2147/AMEP.S155676. doi: 10.2147/AMEP.S155676. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Slawson DC, Shaughnessy AF. Teaching evidence-based medicine: should we be teaching information management instead? Acad Med. 2005 Jul;80(7):685–689. doi: 10.1097/00001888-200507000-00014. [DOI] [PubMed] [Google Scholar]
- 16.Liberati A, Altman DG, Tetzlaff J, Mulrow C, Gøtzsche PC, Ioannidis JPA, Clarke M, Devereaux PJ, Kleijnen J, Moher D. The PRISMA statement for reporting systematic reviews and meta-analyses of studies that evaluate health care interventions: explanation and elaboration. PLoS Med. 2009 Jul 21;6(7):e1000100. doi: 10.1371/journal.pmed.1000100. http://dx.plos.org/10.1371/journal.pmed.1000100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Moher D, Liberati A, Tetzlaff J, Altman DG. Preferred Reporting Items for Systematic Reviews and Meta-Analyses: the PRISMA statement. PLoS Med. 2009 Jul 21;6(7):e1000097. doi: 10.1371/journal.pmed.1000097. http://dx.plos.org/10.1371/journal.pmed.1000097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Higgins JPT, Altman DG, Gøtzsche PC, Jüni P, Moher D, Oxman AD, Savovic J, Schulz KF, Weeks L, Sterne JAC, Cochrane Bias Methods Group. Cochrane Statistical Methods Group The Cochrane Collaboration's tool for assessing risk of bias in randomised trials. BMJ. 2011 Oct 18;343:d5928. doi: 10.1136/bmj.d5928. http://europepmc.org/abstract/MED/22008217. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.CASP checklist: 10 questions to help you make sense of a systematic review. Oxford: Critical Appraisal Skills Programme, part of Oxford Centre for Triple Value Healthcare Ltd; 2018. [2019-12-10]. https://casp-uk.net/wp-content/uploads/2018/01/CASP-Systematic-Review-Checklist_2018.pdf. [Google Scholar]