Skip to main content
PLOS Digital Health logoLink to PLOS Digital Health
. 2023 Nov 27;2(11):e0000255. doi: 10.1371/journal.pdig.0000255

An evidence-based approach to artificial intelligence education for medical students: A systematic review

Nikola Pupic 1, Aryan Ghaffari-zadeh 1,*, Ricky Hu 2, Rohit Singla 1, Kathryn Darras 3, Anna Karwowska 4,5, Bruce B Forster 3
Editor: Valentina Lichtner6
PMCID: PMC10681314  PMID: 38011214

Abstract

The exponential growth of artificial intelligence (AI) in the last two decades has been recognized by many as an opportunity to improve the quality of patient care. However, medical education systems have been slow to adapt to the age of AI, resulting in a paucity of AI-specific education in medical schools. The purpose of this systematic review is to evaluate the current evidence-based recommendations for the inclusion of an AI education curriculum in undergraduate medicine. Six databases were searched from inception to April 23, 2022 for cross sectional and cohort studies of fair quality or higher on the Newcastle-Ottawa scale, systematic, scoping, and integrative reviews, randomized controlled trials, and Delphi studies about AI education in undergraduate medical programs. The search yielded 991 results, of which 27 met all the criteria and seven more were included using reference mining. Despite the limitations of a high degree of heterogeneity among the study types and a lack of follow-up studies evaluating the impacts of current AI strategies, a thematic analysis of the key AI principles identified six themes needed for a successful implementation of AI in medical school curricula. These themes include ethics, theory and application, communication, collaboration, quality improvement, and perception and attitude. The themes of ethics, theory and application, and communication were further divided into subthemes, including patient-centric and data-centric ethics; knowledge for practice and knowledge for communication; and communication for clinical decision-making, communication for implementation, and communication for knowledge dissemination. Based on the survey studies, medical professionals and students, who generally have a low baseline knowledge of AI, have been strong supporters of adding formal AI education into medical curricula, suggesting more research needs to be done to push this agenda forward.

Author summary

Artificial intelligence (AI) has grown exponentially in the last two decades, presenting itself as a significant opportunity to improve the quality of patient care. Despite these advancements, medical programs have been slow to adapt and incorporate AI, which has resulted in a lack of AI-specific teaching in medical schools. Our work evaluates the reasons for why AI should be included in medical school programs to prepare learners for a future where they will have to interact with AI. Based on our analysis, we identified six themes that would be crucial to teach in order to successfully implement AI in medical schools and graduate AI-competent physicians, including: ethics, theory and application, communication, collaboration, quality improvement, and perception and attitude. The general atmosphere in the medical community agrees with our stance, with strong support among medical professionals and students for the inclusion of formal AI education in medical curricula. This highlights the importance of this topic and the need for further research to advance our AI agenda.

1. Introduction

Artificial intelligence (AI) applications in medicine have undergone exponential growth in the past two decades [1]. AI, a general term, implies “the use of a computer to model intelligent behaviour with minimal human intervention” [2]. For instance, in radiology, computer-aided tools can help in the detection of pneumonia or assist in liver and tumor segmentation. These programs aim to assist physicians in the detection and classification of disease, resulting in higher accuracy, reduced variability, or faster results. Despite its rapid emergence, medical training programs have yet to adapt and include AI education as a standard core component [3]. It is now reasonable to assume that graduating medical students should have a basic understanding of AI and how it can improve patient care.

Numerous experts have advocated for the incorporation of AI training and literacy into medical curricula [36]. Paranjape et al. and Wartman and Combs believe that clinical practice is changing from the information age to the AI age [3,4]. This shift promotes abstraction, which would require physicians to employ AI to manage the wealth of information without memorising it, allowing them to prioritize important tasks. However, abstraction requires education to contextualize AI’s capabilities. McCoy et al. state that graduating physicians do not need to understand the complexities of AI algorithms, but they must know when they are applicable and the strengths and weaknesses of the data output.6 These authors also differentiate between the knowledge that all physicians need for everyday practice versus what some physicians need to drive innovation, highlighting that a curricular reform should address the former and extracurricular programs the latter [6]. Support for a curricular reform has been echoed through the medical world, with Banerjee et al. and Teng et al., reporting that 81% of physicians and 63·36% of medical students, respectively, support fundamental AI literacy training [7,8]. Scheetz et al. reported that although physicians are aware of AI as a concept, only 5·5% report their knowledge as excellent, further suggesting the need for a curricular reform [9].

Several groups have investigated implementing ad-hoc extracurricular programs to improve AI literacy, such as the workshops led by Hu et al. and the Artificial Intelligence Curriculum for Residents (AI-RADS) program at Dartmouth [10,11]. These programs, and others, are mostly extracurricular and vary significantly in what is included in their proposed curricula.

The implementation of extracurricular training programs signals that AI training for physicians is at an early stage. Due to issues like topic selection, there is no unified medical curriculum. Despite the support for including AI education, financial barriers, resistance of educators to change curricula, lack of staff with teaching experience in AI, and limited digitalization are some of the main barriers for implementing AI into the medical curriculum [3,10,12,13]. The purpose of this systematic review is to identify and aggregate the current evidence-based recommendations for the development and implementation of an AI curriculum in undergraduate medical education (UGME). We hope that exposing medical students to an AI curriculum in the future will result in an increased integration of AI into their future practice and improved competency and confidence when using AI in medicine.

2. Methods overview

This systematic review followed PRISMA (Preferred Reporting Items for Systematic Review and Meta-analysis) guidelines [14]. This systematic review also used the Newcastle-Ottawa scale (NOS) and the Risk of Bias in Systematic Reviews (ROBIS) tools to assess bias [15,16].

Search strategy

MEDLINE, EMBASE, CINAHL, ERIC, NCBI, and Web of Science were searched from database inception to April 23, 2022, for articles addressing AI education in undergraduate medical education. To guide our search, the search terms of “medical education”, “artificial intelligence”, “medical curriculum”, and “medical program” were used combined with Boolean operators “AND”, “OR”, and “ADJACENT”.

Assessment of study eligibility

The inclusion and exclusion criteria for this study were determined a priori. Criteria for inclusion were: studies about undergraduate medical students, cross sectional and cohort studies assessed as fair quality or higher by NOS, systematic and scoping reviews, randomized controlled trials (RCTs), and Delphi studies.15 Based on the NOS, fair quality evidence includes 2 stars in the selection domain AND 1 or 2 stars in the comparability domain AND 2 or 3 stars in the outcome/exposure domain [15]. Exclusion criteria were: non-English studies, post-graduate medical education studies, papers targeting allied healthcare workers, original research that focuses on applications of AI and not AI education, conference abstracts, case reports, narrative studies, studies assessed as poor quality using the NOS. Based on the NOS, poor quality evidence includes 0 or 1 star in the selection domain OR 0 stars in the comparability domain OR 0 or 1 stars in the outcome/exposure domain [15].

Study screening

Abstract and full-text screening were performed independently and in duplicate by two reviewers (NP and AG) using the Covidence platform (Veritas Health Innovation Ltd., Melbourne, Australia). Any conflicts that arose when consensus was not reached during the screening steps were resolved by a third reviewer (RS).

Data extraction

Data extraction was performed in Excel (Microsoft, Washington, USA). The studies were subdivided into two categories of: survey-based studies and non-survey studies.

For all studies, the following data fields were extracted manually: title, author, country, year of publication, study type, population, sample size, AI principles focused on, any previous exposure to AI, years of experience working with AI, male to female ratio, education level of respondents, factors that may influence respondents’ answers (i.e., anonymized responses, closed-ended questions, open-ended questions), and the study’s level of evidence. Specific to survey studies, additional information extracted included: survey validation status, number of respondents, types of questions asked, survey structure, outcomes evaluated, AI principles focused on, AI exposure of respondents, framework for questions (structured or unstructured), sampling method, and prior exposure to AI. Finally, for non-survey studies, the additional fields extracted were type of validation, sampling type, sensitivity, specificity, key results reported, outcomes evaluated, effect size measurements, statistical significance, clinical significance, and type of intervention.

Data extraction was checked for accuracy and completion by a third reviewer (RS).

Additional references

The reference lists of all included articles were subsequently reviewed for additional relevant articles. Eight additional articles were included.

Thematic analysis

Using the extracted data, two reviewers (NP and AG) created a MindMap using Mural (Mind Mapping Software, Buenos Aires, Argentina) to group various themes that were identified throughout the extraction process. Each group contained information supporting its status using evidence from the extracted data.

Quality assessment and statistical analysis

Quality assessment was performed by two reviewers (NP and AG) using the NOS criteria for cross sectional and cohort studies and the ROBIS criteria for reviews [15,16]. The NOS evaluates studies based on the selection of the study groups, the comparability of groups, and the ascertainment of either the exposure or outcome of interest [15]. A meta-analysis of the included articles was not performed due to the significant heterogeneity in the study types included. However, the article data was categorized based on the frequency of keywords related to the themes identified in the thematic analysis.

3. Results

The original search yielded 991 studies after duplicates were removed. A total of 34 studies were included for analysis after title, abstract, full-text screen, and reference mining (Fig 1). Of the 34 total included studies, 23 were evaluated with the NOS criteria and 11 with the ROBIS criteria. One study met the NOS criteria for good [17] and 22 studies met the NOS criteria for fair [7,8,1836,37]. Seven studies met the ROBIS criteria for low risk [12,13,38,39,40,41,42], one study was of unclear risk [43], and three studies were of high risk [4446]. No RCTs were published within our search criteria. Table 1 provides a summary of the studies included. S1 and S2 Tables in the supplemental documents show an overview of the NOS scoring for cross-sectional and cohort studies and ROBIS scoring for reviews, respectively. The studies were separated into two categories: survey (n = 23) and non-survey (n = 12). The mean sample size of the survey studies was 578 (range 62–1459). The mean respondent years of experience working with AI was only reported in three studies [7,21,28]. The mean male to female ratio of respondents was 54% female and 46% male. A thematic analysis of the included studies revealed six themes: ethics, theory and application, communication, teamwork, quality improvement, and perception and attitude. Fig 2 lists the percentage of total studies supporting each theme. This section also discusses each topic and applicable sub-themes with supporting evidence.

Fig 1. PRISMA diagram depicting the screening process undertaken.

Fig 1

Table 1. A summary of the included studies in the review.

The studies evaluated the current state of AI education in undergraduate medical curricula, as well as the perception, attitude, and knowledge of undergraduate medical students and physicians regarding AI, the current state of AI education in medicine, and the impact of AI on the future of medicine. ML = machine learning, DL = deep learning, EHR = electronic health records.

Author(s) Study Type Year Published Country Sample Size Male to Female Ratio Education Level Outcomes Evaluated/Questions Asked Key Results Reported NOS Score ROBIS Score
Banerjee et al. [7] Survey Study 2021 UK 210 47% female Trainee doctors Training in AI and interaction with AI systems 92% reported insufficient training in AI in their current training, 81% supported more formal training in AI, 62% believed AI would reduce clinical workload 5 N/a
Teng et al. [8] Survey Study 2022 Canada 2167 62·53% female Medical students and physicians Demographics, understanding of AI, attitude towards the impact of AI, priorities in AI literacy and education 63·36% believed that gaining basic literacy in AI should be part of their curriculum, 29·44% preferred either a multiple-workshop series, 16·11% preferred a one-day course. 5 N/a
Lee et al. [12] Scoping review 2021 Canada N/A N/A N/A Identify key themes and gaps on how to train and prepare students for using AI in clinical practice Curriculum delivery should include experiential learning, could also include modules and small group sessions. There are barriers to implementation such as faculty resistance, lack of AI accreditation and licensing, lack of AI core competencies, lack of faculty expertise on AI, and lack of awareness regarding how AI will impact the future of healthcare. N/a Low Risk
Charow et al. [13] Scoping Review 2021 Canada N/A N/A N/A Assess current and past AI education programs to inform curricular content, delivery, and effectiveness There are 3 main barriers to AI adaption: regulatory, economic, and organizational culture. 10 studies present 13 unique programs teaching AI and many are not specifically designed for undergraduate medical education. The cognitive aspect of AI education includes things needed to know, psychomotor involves skills needed to adapt and master, and the affective component considers that attitude needed to develop and incorporate into practice. N/a Low Risk
Karaca et al. [17] Intervention Study 2021 Turkey 897 N/A Undergraduate medical students Cognition, ability, ethics, vision—explain 50·9% of the cumulative variance The Medical Artificial Intelligence Readiness Scale for Medical Students (MAIRS-MS) could be used as an effective screening tool for evaluation and monitoring of medical students’ readiness on the topic of AI. 6 N/a
Gong et al. [18] Survey Study 2018 Canada 322 N/A Medical students Impact of AI in radiology in terms of training, workload, and employment 29·3% of respondents agreed AI would replace radiologists in foreseeable future, 67·7% agreed AI would reduce the demand for radiologists, 48·6% agreed AI caused anxiety when considering the radiology specialty 5 N/a
dos Santos et al. [19] Survey Study 2019 Germany 263 63·8% female Medical students Awareness of AI in the context of radiology, current uses of AI in medicine 52% were aware of AI in radiology, 68% were unaware of the AI technologies involved, 71% agreed on the need for AI to be included in medical training 4 N/a
Sit et al. [20] Survey Study 2020 UK 484 N/A Medical students AI role in in healthcare future, AI impacting specialty choice, current understanding in AI comfort with nomenclature associated with AI, benefits of AI teaching 89% believed that teaching in AI would be beneficial for their careers, 78% agreed that students should receive training in AI as part of their medical degree 6 N/a
Park et al. [21] Survey Study 2020 USA 156 N/A Medical students Medical students’ perception towards AI, primary source of information regarding AI in medicine, AI impact on their enthusiasm for choosing a specialty 75% agreed that AI would have a significant role in the future of medicine, 66% agreed that diagnostic radiology would be the specialty most greatly affected, 44% reported that AI made them less enthusiastic about radiology 5 N/a
Reeder and Lee [22] Survey Study 2021 USA 463 43·2% female Medical students Impact of AI on radiology ranking, Opinions on radiology and AI, Exposure to radiology and AI, Methods for AI education 40% of students expressed a concern towards choosing radiology due to AI, 51% of students predicted a decrease in radiology job opportunities due to AI 6 N/a
Alelyani et al. [23] Survey Study 2021 Saudi Arabia 714 54·6% female Radiologists, technicians, and radiological science students AI awareness, AI practices, AI validation, AI outcomes 81·9% believe that Artificial intelligence must be included in the curriculum and training of medicine and health sciences colleges. 5 N/a
Auloge et al. [24] Survey Study 2020 France 1459 65% female Medical students The future of radiology, awareness regarding interventional radiology, the implementation of AI in radiology. 65% believe AI is not a threat to radiology 5 N/a
Doganer et al. [25] Cross-Sectional Survey Study 2021 Turkey 550 66·2% female Undergraduate students in health sciences (not limited to Medicine) Use of AI, effects of AI in the future of medicine, effects of AI on business life in the future Medical students think that artificial intelligence will increase unemployment and will have a negative sociological impact. 4 N/a
Ejaz et al. [26] Survey Study and Focus Group 2022 UK 128 56% female Medical students AI knowledge, approach to learning AI, applications of AI in clinical medicine, patient safety 86% were interested in exploring interdisciplinary learning with engineering, 92% expressed that AI-related teaching needs to be incorporated into the core medical curriculum 5 N/a
Gillissen et al. [27] Survey and Discussion Groups 2022 Switzerland 1053 74% female Medical students Digitization of patient information, digitization of doctor-patient interaction, demographics, and learning Students within a Case Based Learning (CBL) curriculum believe that AI solutions result in better diagnosis 5 N/a
Blease et al. [28] Cross-Sectional Survey Study 2022 Ireland 252 62·6% female Medical students Familiarity with AI, exposure to AI education 62·4% of students stated there was 0 hours of training in AI. 48·8% of students somewhat agreed, 18·6% moderately agreed, and 11·6% strongly agreed that AI should be a part of their medical program training. 4 N/a
van Hoek et al. [29] Survey Study 2019 Switzerland 170 40% female Doctors, surgeons, and students AI exposure, how AI should be learned, opinions on AI and radiology AI significantly lowered students’ preference for ranking radiology and it was significantly associated with a lower understanding of radiology. Curricular integration was the preferred method of students for teaching AI. 6 N/a
Wood et al. [30] Survey Study 2021 USA 161 33% female Medical students and clinical faculty Participant’s background, AI awareness, and AI applications in medicine. Students were more interested in AI in patient care training (28% vs 14%), while faculty were more interested in AI in teaching training (16% vs 2%). 5 N/a
Blease et al. [31] Survey Study and Open Commentary 2019 UK 66 42·4% female General practitioners Participants were asked to provide any comments on the survey topics seen in study 29. Eight themes were identified, including empathy and communication, clinical reasoning, patient-centredness, improved efficiency, administrative roles, understaffing, acceptability of AI, and ethics of innovation. 5 N/a
Blease et al. [32] Cross-Sectional Survey Study 2018 UK 720 44·9% female General practitioners Could AI replace general practitioners in analyzing patient data, reaching a diagnosis, and providing empathetic care The majority of GPs thought it was unlikely for technology to replace physicians in diagnosing patients (68%), creating personalized treatment plans (61%), and providing empathetic (94%). 5 N/a
Blease et al. [33] Delphi Method Study 2020 USA 29 Round 1: 25% female
Round 2 and 3: 31% female
Leading health Informaticians Forecasts of the impact of AI/ML on patient care, access to care, and the long-future for primary care physicians Experts anticipated that by 2029 workplace changes within healthcare would require increased AI/ML training for medical students. 4 N/a
Blacketer et al. [34] Cross-Sectional Study and Survey Study 2021 Australia 245 N/A Undergraduate medical students ML Knowledge Students performed poorly on questions related to study design and knowledge questions. However, they performed well on interpreting conclusions and statistical significance of ML research. 5 N/a
Kansal et al. [35] Cross-Sectional Study 2019 India 212 40·6% female Medical students and doctors Association between knowledge of AI, gender, medical experience, etc. as well as knowledge and interest in AI 74·4% of participants were not knowledgeable about AI and medical students were more interested in learning about AI than physicians (69·3% vs. 51·6%). Female students were significantly less knowledgeable in AI than male students but they were significantly more interested in learning about AI in medicine. 6 N/a
Giunti et al. [36] Cross-Sectional Descriptive Study 2019 Finland N/A N/A Medical students N/A 30% of schools within 28 members of the European Union offered a type of course within information technology and 64·4% of them made the course mandatory. 5 N/a
Harshana Liyanage et al. [37] Delphi Method Study 2019 Germany 20 experts N/A Panelists with previous exposure to AI Identify stakeholders’ perceptions, issues, and challenges surrounding AI in primary care Primary care community needs to be proactive in guiding the ethical development of AI applications. There should be a formal process to develop an ethics committee that could assess the ethical processing of data in AI applications 5 N/a
Yang et al. [38] Scoping Review 2022 Canada N/A N/A N/A Examines stakeholders’ perspectives on the use of AI in radiology Seven themes of AI in radiology were identified, including predicted impact of AI on radiology, potential replacement of radiologists, trust in AI, knowledge of AI, education on AI, economic considerations, and medicolegal implications. N/a Low Risk
Sapci and Sapci [39] Integrative Review 2020 USA N/A N/A N/A Examine and evaluate the current state of AI training in medicine 10/26 papers evaluated AI education in medicine. AI in healthcare is still an emerging field that does not have much high quality evidence. Most studies were either case reports or opinion pieces N/a Low Risk
Eui-Ryoung Han et al. [40] Integrative Review 2019 Korea N/A N/A N/A Identify and evaluate the themes that need to be implemented in AI curricula for medical education Mainly focuses on medical education trends fostering a connection with advanced technologies amongst physicians considering integrative themes in education such as humanistic approach to patient safety, early experience and longitudinal integration, and student-driven learning with advanced technology N/a Low Risk
Lorainne Tudor Car et al. [41] Review 2021 Singapore N/A N/A N/A Analyze the current digital health education for medical students and identify curricular changes that need to be improved Does not focus on AI and what encompasses digital health is discussed. Courses focusing on digital health within undergraduate medical curriculum were heterogeneous in content and teaching time. N/a Low Risk
Grunhut et al. [42] Integrative review 2021 USA N/A N/A N/A Evaluate the attitudes of medical students towards AI and how to approach implementing AI into the medical curriculum In UME programs, there should be a focus on medical students developing the skill to create validated information for AI systems and learn about the capabilities of AI. There is an overall lack of implementation of AI within medical education even though there is a general consensus as to a need for AI principles being a part of medical curriculum. N/a Low Risk
Chan and Zary [43] Integrative Review 2019 UAE N/A N/A N/A Review current use of AI in medical education, identify challenges of implementing AI in medical education The primary use of AI in medical education was for learning support. To better integrate Ai into the medical profession, AI should be integrated into the medical school curriculum. N/a Unclear Risk
Maksut Senbekov et al. [44] Review 2020 Kazakhstan N/A N/A N/A Discuss and analyze recent changes in digital health topics including AI and medical education AI can be used for clinical decision making and care management as well as proactive detection to forecast hospital mortality. There has not been much effort in the past for revising current curricula. AI-based training in medical education can complement and enrich the curriculum so students know how to apply AI tools to clinical problems. N/a High Risk
Mark P. Khurana et al. [45] Delphi Method Study and Scoping Review 2022 Denmark 18 experts N/A Digital health experts Knowledge, skills, and attitude Attitude towards digital health and basic understanding of digital health are significantly more important than practical skills within digital health. N/a High Risk
Santomartino et al. [46] Systematic Review 2022 USA N/A N/A N/A Evaluate and summarizes the attitudes of medical stakeholders’ toward the role and impact of AI on radiology Medical students and radiologists favored the inclusion of implementing AI solutions in medicine with an overall optimism about the integration of AI within radiology. N/a High Risk

Fig 2. Bar graph depicting the percentage of total studies that support each of the six themes.

Fig 2

Medical AI ethics divides into patient-centric and data-centric ethics. Patient-centric ethics refers to the safe and effective implementation of AI to build awareness of patient inclusion, rights, and equity while addressing bias [13,45]. For example, ensuring that data collection and algorithm design are made equitable by collecting representative training sets to minimize bias [13,37]. Data-centric ethics refers to the legal aspects of safe data handling to preserve patient privacy and to prevent data from being compromised [13]. It requires algorithmic validation, using techniques such as federated learning or subgroup analysis, to ensure they are controlled for bias [37,45], and that the intellectual property over the algorithms is protected by copyright [13,46].

Two sub-themes were identified within theory and application: knowledge for practice and knowledge for development. Knowledge for practice refers to first establishing a strong foundation of basic statistics that would help facilitate learning the basic AI specific knowledge needed to effectively use AI tools in a clinical environment [12,13,17,36,39,42,44,45]. Overall, this includes learning the AI terminology [13,37,45], strengths and limitations of AI [12,13,16,37,42], risks of AI [12,13,17,37,39,42], controlling for bias [42], critical appraisal AI literature and tools [12,13,34,37,39,45,46], and shared decision-making [12,13,37,41,45]. In terms of implementation, developing the ability to visualize how the role and workflow of a physician may change with AI during UGME [13,33], and learning implementation techniques that would help with the AI transition, may improve the future integration of AI in healthcare.

Knowledge for development refers to what physicians should know to better contextualize AI in the clinical environment, communicate to data scientists and allied healthcare professionals, and actively contribute to the evaluation, design, and integration of new AI tools [12,13,44]. Eleven studies suggest that physicians should understand the role of data science in AI [12,13,16,37,39,41,42,45]. In particular, knowledge related to data stewardship [12,13,37,41], data preprocessing and acquisition [13], standardization [13], data analytics [13,17,39], health data infrastructure [45], programming [39,41], and big data [13,45], may help physicians appreciate the obstacles faced by data scientists and improve the multidisciplinary collaboration [13,39]. Other important skills include knowledge related to machine learning (ML) [12,13], deep learning (DL) [13], natural language processing (NLP) [13], model development [13], and AI tool design and development [13,39].

Three key types of communication were identified: communication for clinical decision-making [13], communication for implementation [13,37], and communication for knowledge dissemination [12,13,37,41,42]. Furthermore, four key relationships for communication were identified: data scientist to physician [13], physician to physician/allied healthcare professional [12,13,42], physician to patient [12,13,37,41,42], and physician to payer [13,37].

Communication for clinical decision-making focuses on ameliorating a physician’s own ability to understand how AI tools are designed, including open communication about the data collection, preprocessing, and acquisition [13], as well as what factors the AI tool uses to make decisions [13]. The goal of this communication is to equip physicians with the tools needed to successfully use AI in a clinical setting [12,13,16,40,44,45]. The key relationships for communication for practice are data scientist to physician and physician to physician/allied healthcare professional.

Communication for implementation focuses on understanding how physician roles and workflows will change when AI gets implemented [13,33]. It also focuses on the economic effects of implementing AI tools and how to communicate with stakeholders [13,37]. The key relationships involved will be physician to physician/allied healthcare professional and physician to payer.

Communication for knowledge dissemination focuses on sharing knowledge to improve the understanding of AI tools with respect to their role in care, as well as the strengths, limitations, and risks of AI [12,13,17,37,39,42]. It also includes the need to show empathy and patience when communicating with and educating others about AI [13,33,40]. The key relationships involved will be physician to physician/allied healthcare professional and physician to patient.

The implementation of AI in medicine requires a multidisciplinary approach, integrating perspectives from developers and clinicians [13]. However, regulatory, economic, and organizational fragmentation between clinical and data science domains inhibits the adoption of AI [13]. Regulatory challenges include integrating AI into clinical settings and workflow, data sharing between institutions, and the translation and interpretation of AI models in a health- care setting [3,13,43]. From an economic perspective, learning about the various economic considerations with respect to AI is crucial to having an understanding of the business and clinical aspects of AI implementation [13]. From an organizational culture perspective, including legal strategies and governance in programs can build awareness regarding the ethics, inclusion, equity, patent rights, and confidentiality when using AI tools [3,5,13].

Quality improvement was reflected through the need for critical appraisal of AI literature and tools [12,13,34,37,39], physician involvement in AI tool design/development [13,39], the ability to identify and control for bias [45], communication with data scientists and engineers [13], and communication with patients [12,13,37,41,42]. Quality improvement was also reflected by the need for evaluating and iterating on any AI curriculum designed for medical schools on a regular basis to further improve the learning experience [12].

The perceptions and attitudes of medical students and professionals reflected some hesitancy surrounding the impacts of AI on medicine, such as unemployment [25], but ultimately showed overwhelming support in favour of AI [79,26]. For example, Ejaz et al. report that despite 43% of participants expressing worry regarding the impact of AI on medicine, 92% expressed that AI-related teaching needs to be incorporated into the core medical curriculum [26]. The same authors also report that 86% of students were interested in more interdisciplinary learning within the computer science and biomedical engineering fields [26].

4. Discussion

This is the first study that examines the evidence-based recommendations for implementation of AI educational pro- grams in the global undergraduate medical curriculum. Through systematically evaluating various aspects of AI, we were able to look at the current trends in multiple domains of AI curricula including ethical, technical, and economic points of view.

Nine studies discuss the need for a fundamental understanding of AI ethics to improve the adoption of AI in medicine and ensure its safe use [12,13,16,37,39,45]. Charow et al. support the need for AI-specific ethics to be taught in medical schools to address ethical, legal, and data governance issues [13]. Charow et al. also suggest that this knowledge gap can be bridged by establishing a regulatory body that is tasked with creating systematic guidelines for AI ethics [13]. This regulatory body would also monitor data processing, regulate data sharing between institutions, and oversee the general implementation of AI in healthcare [13]. Lee et al. support this sentiment, suggesting the need for the development of a standardized list of core AI competencies and the implementation of an AI curriculum to medical schools [12]. With a regulatory body and a standardized curriculum, medical schools could ensure that future practitioners are able to recognize and identify ethical and legal issues that may arise such as patient confidentiality [13], equity [13,45], inclusion [13], and patient rights [13]. Furthermore, by providing medical students with an opportunity to learn about AI ethics before they enter their clinical experiences, it could not only better prepare them to work with AI, but also better communicate to patients about AI [12,40].

We also found that establishing a strong theoretical base of AI for practice, including its strengths and limitations, is what we should strive for when educating future learners. This would equip them with the skills needed to critically appraise AI literature and tools before implementing them [12,13,34,37,39], interpret model output [13,37,42], and communicate AI findings to others [12,13,42]. Due to the inability for contextualization of AI algorithms, physicians will still need to regularly conduct contextual analyses when handling the care of patients to supplement the AI. Although important, knowledge for development is likely more suited for select individuals who are more passionate for AI and would like to play a larger role in its advocacy and implementation.

Unfortunately, studies did not always clarify specific issues within each area. Many studies advocated for fundamental statistics knowledge but did not define what is entailed. AI is advanced predictive statistics, therefore medical students could consider it part of their biostatistics coursework. Studies suggest that ML, DL, and NLP should be included in the early years of training in undergraduate medical education in order to increase student exposure to AI [12,13]. Post-graduate medical education can then develop these talents further. Communication for clinical decision-making and communication for knowledge dissemination are both patient-facing skills that would be used on a daily basis in clinics. Therefore, focusing on these skills to bridge the gap between a learner’s knowledge of the theory and application of AI and providing care should be a primary objective taught to medical students. This would allow physicians to explain the strengths and limitations of AI in a way that patients can understand.

AI-interested physicians can also focus on implementation communication, using the skills they developed during UGME and elsewhere to better connect AI with healthcare. Predicting workflow changes, economic implications, and stakeholder communication would be more logistical. Interdisciplinary engagement with data scientists would improve AI infrastructure by solving regulatory, economic, and organisational obstacles and applying quality improvement techniques for constant monitoring. This role’s knowledge may vary by country and healthcare system (i.e., public vs private payer healthcare system). Although there is hesitancy from students surrounding AI implementation, most can see the benefits outweighing the risks, leading to a positive attitude toward learning about AI. As medical school curricula are already saturated with content [47], it may be difficult to move forward without gaining the support of these stakeholders. In addition, there needs to be capacity-building to increase the pool of qualified instructors as those already teaching is overwhelmed [48]. Physicians and patients should play a role in leading AI curriculum development, in addition to data scientists and educators, to ensure the learning outcomes are designed to be clinically useful and patient-centred.

This study had several limitations. The heterogeneity between studies included in this review made it difficult to draw strongly supported conclusions. A lack of follow-up studies evaluating the impacts of current strategies addressing AI education made it difficult to assess their effectiveness. Without these studies, we were also unable to conduct a meta-analysis to strengthen our findings. Lastly, this study focused specifically on AI education in undergraduate medicine, rather than on the continuum of medical education.

Future steps include defining a list of curricular elements that have been validated both by existing literature and domain experts to create a standardized AI curriculum that medical programs can integrate with their current content. We also believe that more cohort studies need to be conducted in order to better evaluate the long-term outcome of AI teachings in undergraduate medicine.

Our study shows that although there is a high degree of heterogeneity among programs and research, the themes of ethics, theory and application, communication, collaboration, quality improvement, and perception and attitudes are recurring. The feedback and attitude of medical trainees towards AI implementation in medical curriculum makes it evident that a systematic approach towards AI education in medicine is warranted.

Supporting information

S1 Table. A table displaying the NOS criteria and scoring for all included survey studies.

The NOS criteria is comprised of three categories: selection, comparability, and outcome.

(DOCX)

S2 Table. A table outlining the ROBIS criteria and scoring for all included review studies, with Phase 2 addressing process concerns and Phase 3 assessing bias risk.

(DOCX)

S1 PRISMA Checklist. PRISMA 2020 Main Checklist.

(PDF)

Data Availability

All relevant data are within the manuscript and its Supporting information files.

Funding Statement

The authors received no specific funding for this work.

References

  • 1.Briganti G, Le Moine O. Artificial intelligence in medicine: today and tomorrow. Front Med 2020; 7: 27. doi: 10.3389/fmed.2020.00027 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Hamet P, Tremblay J. Artificial intelligence in medicine. Metabolism 2017; 69: S36–40. doi: 10.1016/j.metabol.2017.01.011 [DOI] [PubMed] [Google Scholar]
  • 3.Paranjape K, Schinkel M, Panday RN, Car J, Nanayakkara P. Introducing artificial intelligence training in medical education. JMIR Med Educ 2019; 5: e16048. doi: 10.2196/16048 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Wartman SA, Combs CD. Medical education must move from the information age to the age of artificial intelligence. Acad Med 2018; 93: 1107–9. doi: 10.1097/ACM.0000000000002044 [DOI] [PubMed] [Google Scholar]
  • 5.Masters K. Artificial intelligence in medical education. Med Teach 2019; 41: 976–80. doi: 10.1080/0142159X.2019.1595557 [DOI] [PubMed] [Google Scholar]
  • 6.McCoy LG, Nagaraj S, Morgado F, Harish V, Das S, Celi LA. What do medical students actually need to know about artificial intelligence?. NPJ Digit Med 2020; 3: 1–3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Banerjee M, Chiew D, Patel KT, et al. The impact of artificial intelligence on clinical education: perceptions of postgraduate trainee doctors in London (UK) and recommendations for trainers. BMC Med Educ 2021; 21: 429. doi: 10.1186/s12909-021-02870-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Teng M, Singla R, Yau O, et al. Health Care Students’ Perspectives on Artificial Intelligence: Countrywide Survey in Canada. JMIR Med Educ 2022; 8: e33390. doi: 10.2196/33390 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Scheetz J, Rothschild P, McGuinness M, et al. A survey of clinicians on the use of artificial intelligence in ophthalmology, dermatology, radiology and radiation oncology. Sci Rep 2021; 11: 5193. doi: 10.1038/s41598-021-84698-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Hu R, Fan KY, Pandey P, et al. Insights from teaching artificial intelligence to medical students in Canada. Commun Med 2022; 2: 63. doi: 10.1038/s43856-022-00125-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Lindqwister AL, Hassanpour S, Lewis PJ, Sin JM. AI-RADS: an artificial intelligence curriculum for residents. Acad Radiol 2021; 28: 1810–6. doi: 10.1016/j.acra.2020.09.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Lee J, Wu AS, Li D, Kulasegaram KM. Artificial intelligence in undergraduate medical education: a scoping review. Acad Med 2021; 96: S62–70. doi: 10.1097/ACM.0000000000004291 [DOI] [PubMed] [Google Scholar]
  • 13.Charow R, Jeyakumar T, Younus S, et al. Artificial intelligence education programs for health care professionals: Scoping review. JMIR Med Educ 2021; 7: e31043. doi: 10.2196/31043 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Shamseer L, Moher D, Clarke M, et al. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015: elaboration and explanation. BMJ 2015; 349: g7647. doi: 10.1136/bmj.g7647 [DOI] [PubMed] [Google Scholar]
  • 15.Wells GA, Shea B, O’Connell D, et al. The Newcastle-Ottawa Scale (NOS) for assessing the quality of nonrandomised studies in meta-analyses. 2019. http://www.ohri.ca/programs/clinical_epidemiology/oxford.asp (accessed Nov, 1, 2022).
  • 16.Whiting P, Savović J, Higgins JP, et al. ROBIS: a new tool to assess risk of bias in systematic reviews was developed. J Clin Epi 2016; 69: 225–34. doi: 10.1016/j.jclinepi.2015.06.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Karaca O, Çalışkan SA, Demir K. Medical artificial intelligence readiness scale for medical students (MAIRS-MS)–development, validity and reliability study. BMC Med Educ 2021; 21: 1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Gong B, Nugent JP, Guest W, et al. Influence of artificial intelligence on Canadian medical students’ preference for radiology specialty: A National survey study. Acad Radiol 2019; 26: 566–77. doi: 10.1016/j.acra.2018.10.007 [DOI] [PubMed] [Google Scholar]
  • 19.Pinto Dos Santos D, Giese D, Brodehl S, et al. Medical students’ attitude towards artificial intelligence: a multicentre survey. Eur Radiol 2019; 29: 1640–6. doi: 10.1007/s00330-018-5601-1 [DOI] [PubMed] [Google Scholar]
  • 20.Sit C, Srinivasan R, Amlani A, et al. Attitudes and perceptions of UK medical students towards artificial intelligence and radiology: a multicentre survey. Insights Imaging 2020; 11: 1–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Park CJ, Paul HY, Siegel EL. Medical student perspectives on the impact of artificial intelligence on the practice of medicine. Curr Probl Diagn Radiol 2021; 50: 614–9. doi: 10.1067/j.cpradiol.2020.06.011 [DOI] [PubMed] [Google Scholar]
  • 22.Reeder K, Lee H. Impact of artificial intelligence on US medical students’ choice of radiology. Clin Imaging 2022; 81: 67–71. doi: 10.1016/j.clinimag.2021.09.018 [DOI] [PubMed] [Google Scholar]
  • 23.Alelyani M, Alamri S, Alqahtani MS, et al. Radiology community attitude in Saudi Arabia about the applications of artificial intelligence in radiology. Healthcare 2021; 9: 834. doi: 10.3390/healthcare9070834 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Auloge P, Garnon J, Robinson JM, et al. Interventional radiology and artificial intelligence in radiology: Is it time to enhance the vision of our medical students?. Insights Imaging 2020; 11: 1–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Doganer A. The approaches and expectations of the health sciences students towards artificial intelligence. Karya J Health Sci 2021; 2: 5–11. [Google Scholar]
  • 26.Ejaz H, McGrath H, Wong BL, Guise A, Vercauteren T, Shapey J. Artificial intelligence and medical education: A global mixed-methods study of medical students’ perspectives. Digit Health 2022; 8: 1–11. doi: 10.1177/20552076221089099 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Gillissen A, Kochanek T, Zupanic M, Ehlers J. Medical Students’ Perceptions towards Digitization and Artificial Intelligence: A Mixed-Methods Study. Healthcare 2022; 10: 723. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Blease C, Kharko A, Bernstein M, et al. Machine learning in medical education: a survey of the experiences and opinions of medical students in Ireland. BMJ Health Care Inform 2022; 29: e100480. doi: 10.1136/bmjhci-2021-100480 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.van Hoek J, Huber A, Leichtle A, et al. A survey on the future of radiology among radiologists, medical students and surgeons: students and surgeons tend to be more skeptical about artificial intelligence and radiologists may fear that other disciplines take over. Eur Radiol 2019; 121: 108742. doi: 10.1016/j.ejrad.2019.108742 [DOI] [PubMed] [Google Scholar]
  • 30.Wood EA, Ange BL, Miller DD. Are we ready to integrate artificial intelligence literacy into medical school curriculum: students and faculty survey. J Med Educ Curric Dev 2021; 8: 1–5. doi: 10.1177/23821205211024078 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Blease C, Kaptchuk TJ, Bernstein MH, Mandl KD, Halamka JD, DesRoches CM. Artificial intelligence and the future of primary care: exploratory qualitative study of UK general practitioners’ views. J Med Internet Res 2019; 21: e12802. doi: 10.2196/12802 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Blease C, Bernstein MH, Gaab J, et al. Computerization and the future of primary care: a survey of general practitioners in the UK. PLoS One 2018; 13: e0207418. doi: 10.1371/journal.pone.0207418 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Blease C, Kharko A, Locher C, DesRoches CM, Mandl KD. US primary care in 2029: A Delphi survey on the impact of machine learning. PLoS One 2020; 15: e0239947. doi: 10.1371/journal.pone.0239947 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Blacketer C, Parnis R, Franke KB, et al. Medical student knowledge and critical appraisal of machine learning: a multicentre international cross-sectional study. Intern Med J 2021; 51: 1539–42. doi: 10.1111/imj.15479 [DOI] [PubMed] [Google Scholar]
  • 35.Kansal R, Bawa A, Bansal A, et al. Differences in Knowledge and Perspectives on the Usage of Artificial Intelligence Among Doctors and Medical Students of a Developing Country: A Cross-Sectional Study. Cureus 2022; 14: e21434. doi: 10.7759/cureus.21434 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Giunti G, Guisado-Fernandez E, Belani H, Lacalle-Remigio JR. Mapping the access of future doctors to health information technologies training in the European Union: cross-sectional descriptive study. J Med Internet Res 2019; 21: e14086. doi: 10.2196/14086 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Liyanage H, Liaw ST, Jonnagaddala J, et al. Artificial intelligence in primary health care: perceptions, issues, and challenges. Yearb Med Inform 2019; 28: 41–6. doi: 10.1055/s-0039-1677901 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Yang L, Ene IC, Arabi Belaghi R, Koff D, Stein N, Santaguida PL. Stakeholders’ perspectives on the future of artificial intelligence in radiology: a scoping review. Eur Radiol 2022; 32: 1477–1495. doi: 10.1007/s00330-021-08214-z [DOI] [PubMed] [Google Scholar]
  • 39.Sapci AH, Sapci HA. Artificial intelligence education and tools for medical and health informatics students: systematic review. JMIR Med Educ 2020; 6: e19285. doi: 10.2196/19285 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Han ER, Yeo S, Kim MJ, Lee YH, Park KH, Roh H. Medical education trends for future physicians in the era of advanced technology and artificial intelligence: an integrative review. BMC Med Educ 2019; 19: 460. doi: 10.1186/s12909-019-1891-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Car LT, Kyaw BM, Panday RS, et al. Digital health training programs for medical students: scoping review. JMIR Med Educ 2021; 7: e28275. doi: 10.2196/28275 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Grunhut J, Wyatt AT, Marques O. Educating future physicians in artificial intelligence (AI): An integrative review and proposed changes. J Med Educ Curric Dev 2021; 8: 1–8. doi: 10.1177/23821205211036836 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Chan KS, Zary N. Applications and challenges of implementing artificial intelligence in medical education: integrative review. JMIR Med Educ 2019; 5: e13930. doi: 10.2196/13930 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Senbekov M, Saliev T, Bukeyeva Z, et al. The recent progress and applications of digital technologies in healthcare: a review. Int J Telemed Appl 2020; 2020: 8830200. doi: 10.1155/2020/8830200 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Khurana MP, Raaschou-Pedersen DE, Kurtzhals J, Bardram JE, Ostrowski SR, Bundgaard JS. Digital health competencies in medical school education: a scoping review and Delphi method study. BMC Med Educ 2022; 22: 129. doi: 10.1186/s12909-022-03163-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Santomartino SM, Paul HY. Systematic Review of Radiologist and Medical Student Attitudes on the Role and Impact of AI in Radiology. Acad Radiol 2022; 29: 1748–56. doi: 10.1016/j.acra.2021.12.032 [DOI] [PubMed] [Google Scholar]
  • 47.IsHak W, Nikravesh R, Lederer S, Perry R, Ogunyemi D, Bernstein C. Burnout in medical students: a systematic review. Clin Teach 2013; 10: 242–5. doi: 10.1111/tct.12014 [DOI] [PubMed] [Google Scholar]
  • 48.Dandar V, Grigsby RK, Bunton S. Burnout among US medical school faculty. AAMC Anal Brief 2019; 19: 1–3. https://www.aamc.org/media/9921/download?attachment (Accessed Dec 15, 2022).
PLOS Digit Health. doi: 10.1371/journal.pdig.0000255.r001

Decision Letter 0

Valentina Lichtner, Piyush Mathur

18 May 2023

PDIG-D-23-00115

Artificial Intelligence Education for Medical Students: A Systematic Review

PLOS Digital Health

Dear Dr. Ghaffari-zadeh,

Thank you for submitting your manuscript to PLOS Digital Health. After careful consideration, we feel that it has merit but does not fully meet PLOS Digital Health's publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please specifically address the comments from Reviewer 2.

Please submit your revised manuscript within 60 days Jul 17 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at digitalhealth@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pdig/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

* A rebuttal letter that responds to each point raised by the editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

* A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

* An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

We look forward to receiving your revised manuscript.

Kind regards,

Valentina Lichtner

Academic Editor

PLOS Digital Health

Journal Requirements:

1. Please send a completed 'Competing Interests' statement, including any COIs declared by your co-authors. If you have no competing interests to declare, please state "The authors have declared that no competing interests exist". Otherwise please declare all competing interests beginning with twhe statement "I have read the journal's policy and the authors of this manuscript have the following competing interests:"

2. Please provide separate figure files in .tif or .eps format only and remove any figures embedded in your manuscript file. Please also ensure that all files are under our size limit of 10MB.

For more information about figure files please see our guidelines:

https://journals.plos.org/digitalhealth/s/figures

https://journals.plos.org/digitalhealth/s/figures#loc-file-requirements

3. We ask that a manuscript source file is provided at Revision. Please upload your manuscript file as a .doc, .docx, .rtf or .tex.

4. We have noticed that you have uploaded Supporting Information files, but you have not included a list of legends. Please add a full list of legends for your Supporting Information files after the references list.

Additional Editor Comments (if provided):

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Does this manuscript meet PLOS Digital Health’s publication criteria? Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe methodologically and ethically rigorous research with conclusions that are appropriately drawn based on the data presented.

Reviewer #1: Yes

Reviewer #2: No

Reviewer #3: Yes

--------------------

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

Reviewer #3: Yes

--------------------

3. Have the authors made all data underlying the findings in their manuscript fully available (please refer to the Data Availability Statement at the start of the manuscript PDF file)?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception. The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

--------------------

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS Digital Health does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

--------------------

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This systematic review attempts to identify emerging themes in incorporating artificial intelligence literacy in medical education. The authors do a nice job of synthesizing their findings into relevant, albeit broad, learning objectives for undergraduate medical curricula.

The study title is too broad and fails to highlight the important goal of this study which was to identify evidence-based recommendations. It is highly recommended to focus the title and align it with the study's objective.

The methodology is well explained and executed. My only concern is the exclusion of the randomized control trial study design from the review. The rationale and decision to not include RCTs must be clearly explained in the methodology. For example, interventional studies that have looked at the efficacy of utilizing AI for instruction would provide further evidence to contextualize the authors' recommendations and identify more nuanced strengths and weaknesses.

The discussion is well balanced and adequately states potential challenges in enriching an already saturated curriculum. The article can possibly benefit from a discussion on the role of medical licensing bodies such as the LCME or the Royal College for Physicians and Surgeons of Canada in defining new core competencies on AI or Digital Health Literacy.

Overall, this is an insightful study that identifies key themes to be addressed in an AI-aware medical curriculum.

Reviewer #2: This paper aimed to identify evidence-based recommendations for developing and implementing an AI curriculum in undergraduate medical education through reviewing articles. The authors claimed six themes for the successful implementation of AI in medical school curricula: ethics, theory and application, communication, collaboration, quality improvement, and perception and attitude. Furthermore, they suggested future research about adding formal AI education into medical curricula.

This paper is very timely and novel. The AI education issue is exciting to global medical educators because of ChagGPT. However, the article needs rigorous efforts for good original content and a valid research methodology.

1) Research gaps

The authors needed to present research gaps from previous review literature sufficiently. This paper is a systematic review, so the authors should show how many review articles have already existed and which research agenda has been explored in these previous review articles. When the authors present current review aims and methodologies, readers can understand the context of the previous literature and the necessity of this research.

2) Data and analyses

The data and analyses did not fully support the authors’ claims that the themes are evidence-based. First, the themes were mainly from review articles, but the authors did not evaluate their quality and evidence level. Instead, they only assessed the quality of the cohort and cross-sectional studies based on the Newcastle-Ottawa quality assessment scale.

Second, the purpose of the previous article did not aim to implement a successful program. Therefore, I don’t think the authors thematized the findings from the included reports. Instead, I guess they thematized the discussion contents from the previous articles. I don't think the discussion contents can be evidence-based.

Therefore, I suggest the authors collect and evaluate articles implementing and evaluating AI programs to support the claims that recommendations are evidence-based. In addition, they should analyze the outcomes of articles related to the research questions.

3) Additional work to improve the paper

Rather than a systematic review, this paper seems more likely to be an integrative or scoping review. The authors wanted to explore implementation strategies using thematic analysis. This aim matches well with the integrative review.

Or the authors could narrow the focus on survey or review articles to improve the research design.

There are three more things. One is that the results used non-included articles from 50 to 65. I think the authors should remove them. Another is the discussion. AI in healthcare seems unrelated to the research questions in this paper. I hope to see more detailed and profound talks about developing and implementing AI education from the results. The other is Figure 2. As a reader, I would like to see the quality level in Table 1 instead of Figure 2.

4) Research protocol

The authors did not present a registered protocol. So, it would be good for the authors to register the research protocol before starting the study.

5) Guidelines and methodology

This paper needs to describe a more detailed methodology for reproduction. The authors tried to follow the PRISMA guideline, but more was needed. PRISMA 2020 for abstracts and AMSTAR 2 checklist will be helpful.

First, the abstract should include more detailed information, such as specific inclusion and exclusion criteria, information sources, methods to assess the risk of biases in studies, and limitations of the evidence.

Second, key research questions did not specify PICO and others, such as time, setting, or study design.

Third, inclusion criteria only described article types, with no exclusion criteria. I wanted to know whether the authors include non-English articles, experimental studies, or single-institution studies. Besides, I asked whether the authors excluded low-quality articles.

Fourth, two or more researchers should select articles independently. The authors described two researchers conducting study screening, data extraction, and thematic analysis. However, they should have commented on how many researchers perform study selection and quality assessment.

Fifth, I wondered why the authors chose NOS because NOS can evaluate only cohort and case-control studies. Many other evaluation tools exist, such as AMSTAR, MERSQI, or BEME. This paper included various articles, so I suggest the authors use more appropriate tools for surveys and review articles.

6) Organization of the manuscript

The manuscript was understandable, but needed to check for omissions and errors. In addition, for readers who are not familiar with article evaluation tools, NOS only is difficult to understand.

Furthermore, the authors said that they separated the previous studies into survey and interventional. Interventional studies were mainly reports and reviews, which confused me because interventional studies usually mean experimental studies. Therefore, I suggest the authors change the term ‘interventional’ and explain the definition of it.

Reviewer #3: Pupic et al. did a systematic review on the AI-specific education in medical schools. They had collected more than 900 original paper and scoped down into 30 ones fit the study criteria. In conclusion they found six core themes should be suggested implant into the future AI related medical education.

Minor points:

1. It is also mentioned some field of medical education developed more dependent on current state of art technique of AI, while other filed less developed. So, it is suggested to draw a bar chart to demonstrate the diversity of the current AI application situation in various medial field or specialty. And the distribution of the 30 selected paper should also be more elucidated on each research field.

2. The specific indicators of quality assessment and statistical analysis should be clearer. The current presentation format of Table 1 and Figure 2 and 3 are not strait forward and difficult to link with the result part.

In all, the topic is meaningful and the research method was thoroughly with high efficiency and plentiful of data. The current manuscript should be accepted after the minor revisions.

--------------------

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

Do you want your identity to be public for this peer review? If you choose “no”, your identity will remain anonymous but your review may still be made public.

For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: HyeRin Roh

Reviewer #3: Yes: Dawei Yang

--------------------

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLOS Digit Health. doi: 10.1371/journal.pdig.0000255.r003

Decision Letter 1

Valentina Lichtner, Piyush Mathur

14 Sep 2023

An Evidence-Based Approach to Artificial Intelligence Education for Medical Students: A Systematic Review

PDIG-D-23-00115R1

Dear Mr Ghaffari-zadeh,

We are pleased to inform you that your manuscript 'An Evidence-Based Approach to Artificial Intelligence Education for Medical Students: A Systematic Review' has been provisionally accepted for publication in PLOS Digital Health.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow-up email from a member of our team. 

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they'll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact digitalhealth@plos.org.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Digital Health.

Best regards,

Valentina Lichtner

Academic Editor

PLOS Digital Health

***********************************************************

Reviewer Comments (if any, and for reference):

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. A table displaying the NOS criteria and scoring for all included survey studies.

    The NOS criteria is comprised of three categories: selection, comparability, and outcome.

    (DOCX)

    S2 Table. A table outlining the ROBIS criteria and scoring for all included review studies, with Phase 2 addressing process concerns and Phase 3 assessing bias risk.

    (DOCX)

    S1 PRISMA Checklist. PRISMA 2020 Main Checklist.

    (PDF)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting information files.


    Articles from PLOS Digital Health are provided here courtesy of PLOS

    RESOURCES