Abstract
Background
The widening gap between innovations in the medical field and the dissemination of such information to doctors may affect the quality of care. Offline computer-based digital education (OCDE) may be a potential solution to overcoming the geographical, financial, and temporal obstacles faced by doctors.
Objective
The objectives of this systematic review were to evaluate the effectiveness of OCDE compared with face-to-face learning, no intervention, or other types of digital learning for improving medical doctors’ knowledge, cognitive skills, and patient-related outcomes. Secondary objectives were to assess the cost-effectiveness (CE) of OCDE and any adverse effects.
Methods
We searched major bibliographic databases from 1990 to August 2017 to identify relevant articles and followed the Cochrane methodology for systematic reviews of intervention.
Results
Overall, 27 randomized controlled trials (RCTs), 1 cluster RCT (cRCT), and 1 quasi-RCT were included in this review. The total number of participants was 1690 in addition to the cRCT, which included 24 practices. Due to the heterogeneity of the participants, interventions, and outcomes, meta-analysis was not feasible, and the results were presented as narrative summary.
Compared with face-to-face learning, the effect of OCDE on knowledge gain is uncertain (ratio of the means [RM] range 0.95-1.17; 8 studies, 495 participants; very low grade of evidence). From the same comparison, the effect of OCDE on cognitive skill gain is uncertain (RM range 0.1-0.9; 8 studies, 375 participants; very low grade of evidence). OCDE may have little or no effect on patients’ outcome compared with face-to-face education (2 studies, 62 participants; low grade of evidence).
Compared with no intervention, OCDE may improve knowledge gain (RM range 1.36-0.98; 4 studies, 401 participants; low grade of evidence). From the same comparison, the effect of OCDE on cognitive skill gain is uncertain (RM range 1.1-1.15; 4 trials, 495 participants; very low grade of evidence). One cRCT, involving 24 practices, investigated patients’ outcome in this comparison and showed no difference between the 2 groups with low-grade evidence.
Compared with text-based learning, the effect of OCDE on cognitive skills gain is uncertain (RM range 0.91-1.46; 3 trials with 4 interventions; 68 participants; very low-grade evidence). No study in this comparison investigated knowledge gain or patients’ outcomes.
One study assessed the CE and showed that OCDE was cost-effective when compared with face-to-face learning if the cost is less than or equal to Can $200. No trial evaluated the adverse effect of OCDE.
Conclusions
The effect of OCDE compared with other methods of education on medical doctors’ knowledge and cognitive skill gain is uncertain. OCDE may improve doctors’ knowledge compared with no intervention but its effect on doctors’ cognitive skills is uncertain. OCDE may have little or no effect in improving patients’ outcome.
Keywords: systematic review; medical education, digital education
Introduction
Background
Faced with the rapid innovations in medicine, structured postgraduate residency programs, in addition to continuing professional development (CPD) and continuing medical education, (CME), have been developed to advance and update the skills and knowledge of medical doctors and other health professions [1]. Evidence suggested that these programs are effective in improving the diagnostic and therapeutic competencies of health care professionals and patient-related outcomes [2-6]. Although face-to-face learning is the dominant method of teaching in these programs, it is being increasingly supplemented or replaced with digital learning using both Web-based and offline options [7]. Digital education involves the delivery of educational material through Information and Communication Technology using a wide variety of pedagogical designs and formats [8,9]. Digital learning is a plausible low-cost platform that provides convenient access to educational materials with flexibility in terms of pace, place, and time [10]. It has many advantages over traditional learning, such as outreach, flexibility, and adaptability, as it has the potential to reach a large number of learners regardless of the physical distance; in addition, it caters for the pace and time of the individual participant while reducing the overhead costs of the learning process [11-13]. These features of digital learning could prove it to be a plausible solution to the constraints faced by medical education in low- and middle-income countries (LMICs) where digital learning has been used effectively to improve existing health services [12] and the diagnostic and therapeutic competencies of the health care providers [2].
Despite the potential of digital learning to be the leading learning method in sciences, including medical education, some disadvantages have been reported for this method of learning. Lack of interaction with other learners and the instructor, because of flexibility in the timing of learning, may lead to social isolation and discouragement of team work [11,14].
Although digital learning caters for students’ learning pace, it tends to deindividualize the instructor and fails to respond to the learners’ individual needs [15]. Although digital learning is associated with reduced cost for the learner, the potentially large cost of designing some courses, such as those that involve virtual reality and simulation, cannot be ignored [15].
Unlike other learning methods, digitally-based courses must be designed carefully to meet the instructional objectives as the absence of an instructor for explanation makes such courses more susceptible to design flaws [15,16]. For effective digital learning in medical education, many essential design characteristics have been suggested, such as effective communication among the learners and validation and assessment of knowledge gained in addition to use of real-world scenarios [17].
Currently, there are many different technological platforms for digital learning, including Web-based learning activities, computer-based instruction, and mobile learning. The availability of many platforms paved the way for many learning opportunities in medical education, such as massive open Web-based courses, serious gaming and gamification, and virtual patient-based learning. The focus of this review is offline computer-based digital education (OCDE).
This type of digital education does not require internet or local area network connection, and the learning material is typically kept in either magnetic storage such as floppy discs or optical storage such as CD, digital versatile disc, flash memory, multimedia cards, and external hard discs, which facilitates the delivery of various educational material such as text, images, audio, and video material [18-20]. OCDE has many advantages, especially in settings where internet access is absent or limited. It has many of the advantages over other digital learning platforms as it provides a solution for those learners faced with geographical, financial, and temporal barriers to face-to-face education without compromising on the teaching and learning process or the intended outcomes [18,19,21].
Objectives
The objective of this systematic review was to evaluate the effectiveness of OCDE compared with face-to-face learning, no intervention, text-based learning, or other type of digital learning for improving medical doctors’ knowledge, cognitive skills, and patient-related outcomes. The secondary objectives were to assess the cost and cost-effectiveness (CE) of OCDE and adverse effects of the interventions.
Methods
Eligibility Criteria
Types of Studies
We included randomized controlled trials (RCTs), cluster RCTs (cRCTs), and quasi-randomized trials that compared OCDE (personal computer or laptop) methods for medical doctors and dentists with face-to-face learning, written information, no intervention, or other OCDE. We excluded crossover trials because of the high likelihood of a carry-over effect [22]. We included trials reported in conference proceedings and abstracts when information could be obtained by contacting the authors. For a detailed description of the methodology, please refer to the study by Car et al [23].
Types of Participants
We included studies in which participants (learners) were medical doctors and dentists who were enrolled in postgraduate medical education programs. Studies with mixed participant groups, such as doctors and nurses, in which results for medical doctors could not be obtained separately, were excluded.
Types of Interventions
We included studies in which OCDE interventions were used to deliver educational content. Residency training programs and CME- and CPD-based programs that involved the use of OCDE interventions were included [24,25].
OCDE intervention refers to the use of personal computers or laptops that have assisted in the delivery of standalone multimedia materials without the need for internet or local area network connections [24,25].
We only considered studies that made the following intervention comparisons:
Offline-based intervention versus traditional face-to-face learning.
Offline-based intervention versus no intervention.
Offline-based digital learning versus written text-based learning.
Offline-based digital learning versus another method of digital learning.
Primary Outcomes
We included studies that reported at least one of the following primary or secondary outcomes
Learner’s knowledge: defined as the learners’ factual or conceptual understanding.
Learner’s cognitive skills: defined as skills used in the process of acquiring knowledge, for example, skills learned for reading an x-ray film or learning the steps of performing a procedure.
Patients’ outcomes: defined as the direct observation of the application of knowledge and skills on the patients where the outcome of the trial is patients’ physical, mental, and psychological condition, such as the clinical effect of optimizing medication regimens. In addition, we considered studies that aimed at improving health services, such as improving existing screening programs, as patients’ outcomes.
Secondary Outcomes
CE of digital learning
Any adverse outcome
Types of Outcome Measures
These outcomes were assessed using any validated or non-validated instrument to measure the difference in pre and posttest scores. These assessments were either subjective (eg, self-reported) or objective (eg, questionnaire). When several posttest results were available, data were recorded as to when those tests were conducted and the difference between the pretest and the first posttest was used for the analysis.
Search Methods
We searched major bibliographic databases from 1990, when the virtual learning environments began and schools started delivering Web-based courses, till August 2017 to identify all relevant articles. We searched in English but included papers published in any language.
Electronic Searches
We searched the following databases: Medical Literature Analysis and Retrieval System Online (MEDLINE, via Ovid), EMBASE (via Ovid), Web of Science, Educational Resource Information Centre (via Ovid), Cochrane Central Register of Controlled Trials, Cochrane Library, PsycINFO (Ovid), Cumulative Index to Nursing and Allied Health Literature (via EBSCO), and ProQuest Dissertation and Theses Database. The MEDLINE search strategy was adapted to search other databases (Multimedia Appendix 1).
Searching Other Resources
We searched the reference lists of all included studies and relevant systematic reviews. We also searched the International Clinical Trials Registry Platform Search Portal and Current Controlled Trials metaRegister of Controlled Trials to identify the unpublished trials and contacted relevant investigators for further information.
Data Collection and Analysis
Study Selection
Two reviewers independently screened the titles and abstracts and identified studies potentially meeting the inclusion criteria. The full-text versions were retrieved and read in full. Finally, 2 review authors independently assessed the articles against the eligibility criteria. Any disagreements were resolved through discussion between the 2 authors. If no agreement was reached, a third author acted as an arbiter. Two reviewers verified the final list of included studies.
Data Extraction and Management
Two reviewers independently extracted and managed the data for each of the included studies and used a structured data recording form. In addition to the usual information on the study design and participants’ demographics, we extracted data on relevant fields such as the country where the trial was conducted, funding source, and duration of intervention. Disagreements between the review authors were resolved by discussion. A third review author acted as an arbiter in case disagreements were not resolved.
Dealing with Missing Data
Whenever possible, we attempted to obtain missing data from the original authors.
Assessment of Risk of Bias in Included Studies
Two reviewers independently assessed the risk of bias of each of the included studies using the Cochrane Collaboration’s risk of bias tool [22]. Studies were assessed for the risk of bias in the following domains: random sequence generation, allocation concealment, blinding of participants or personnel, blinding of outcome assessors, completeness of outcome data (attrition bias), selective outcome reporting (reporting bias), and other sources of bias including baseline imbalance and contamination. For cRCTs, we assessed and reported the risk of bias associated with an additional domain: selective recruitment of cluster participants, baseline imbalance, attrition of clusters, and not accounting for cluster effect in analysis [26].
We judged the risk of bias for each study to be of 1 of 3 levels: high, low, or unclear risk of bias. We scored each study for risk of bias as follows: low if all key domains were scored as low risk or if 1 domain is scored as unclear. We scored the trial as unclear if 2 key domains were scored as unclear risk and high if more than 2 key domains were scored unclear risk or 1 domain scored high risk, adapted from the study by Davey et al [27].
Reporting bias was assessed qualitatively on the basis of the characteristics of included studies. Due to the heterogeneity of the trials (in terms of populations, interventions, comparator groups, and outcomes), data pooling was not feasible.
We used Evers checklist [28] to evaluate risk of bias in articles that examined the CE of OCDE.
Data Synthesis
We reported post intervention values for the outcomes of intervention and control groups and the effect size as reported by the authors (P value). In addition, we calculated the ratio of the means (RM) [4] whenever feasible. As the heterogeneity of populations, outcomes, and comparisons precluded meta-analysis, we provided a narrative summary of the results.
Summary of Findings Tables
For main comparisons, 2 authors used the Grading of Recommendations Assessment, Development and Evaluation (GRADE) criteria independently of one another to assess the quality of evidence [29]. We considered the following limitations: risk of bias, inconsistency of results, indirectness of the evidence, imprecision, or publication bias, and subsequently downgraded the quality of evidence where appropriate [30].
Results
Overview
The study selection process is shown in the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) flow diagram (Figure 1).
The initial search yielded 21,849 records. After screening the titles and abstracts, we obtained the full-text reports for 197 records and assessed them for inclusion in the review. Of these, we excluded 168 studies that did not meet the eligibility criteria (Multimedia Appendix 2). The remaining 29 articles were included in this review.
We contacted authors of abstracts for further information and data on their trials [31-37], we received a response from 2 authors, Rae et al [37], where the data were subsequently included in the review, and Ukabiala et al [36], which was subsequently excluded. We tried to contact the authors of the cRCT [38] for data and information to assess the risk of bias and to calculate the intracluster correlation coefficients (ICC); however, the contact information was void.
Included Studies
A total of 24 of the 29 included trials were parallel RCTs, each included 2 arms and 1 trial had 4 arms, 3 of which were included in this review, resulting in 2 interventions [37]. In addition, 1 trial had 3 arms [39], 1 was a factorial RCT [40], 1 was a cRCT [38], and 1 was a quasi-RCT [41]. All trials were published in peer-reviewed journals except for 1 trial where unpublished data were obtained from the authors [37]. Overall, 24 studies (83%, 24/29) were conducted in high-income countries [37-60], and the remaining 5 studies (17%, 5/29) were conducted in upper middle-income countries [61-65].
Two studies investigated OCDE in dentistry [39,45], whereas the rest investigated it in medicine, including 6 in surgery or anesthesia [37,47,48,50,51,54], 7 in internal medicine or family medicine [40,44,49,53,57,60,65], 5 in pediatrics [38,41,46,56,58], 3 in psychiatry [52,61,63], 1 in obstetrics and gynecology [42], and 1 in radiology [43]. In the 4 remaining trials, the subject of intervention was evidence-based medicine [55] or advanced life support [59,62,64]. We included 1 cRCT in this review [38]. We obtained data at the participants’ level to estimate the outcome effect; however, the effect of clustering was not adjusted for and sufficient information was not available to perform reanalysis to account for ICC (Multimedia Appendix 3).
Participants’ Characteristics
The total number of participants included across all trials was 1690 in addition to 1 cRCT [38], which was conducted in 24 medical practices, but the number of participants was not specified. The study with the largest number of participants included 88 dentists in the control group and 86 in the intervention group [45], and the smallest study included 5 participants in the control and 6 participants in the intervention group [58]. Most of the participants were trainees. In addition, 4 studies were conducted among 377 medical interns [55,62-64], 16 trials among a total of 686 postgraduate residents [37,42,44,46-50,52,54,58-61,65], and 1 study included 49 residents and faculty members [43], of the remaining studies, 7 were performed on 578 practicing doctors [39-41,45,53,56,57] (Multimedia Appendix 3).
Intervention Characteristics
Overall, 15 studies compared OCDE with classroom or face-to-face learning [41,43,47-51,53-55,60-64], 3 studies with 4 interventions [37,46,58] compared OCDE with text-based resources, and 9 studies [38-40,42,44,45,52,56,59] compared OCDE with no intervention. The other 2 trials compared OCDE with another OCDE [57,65]. In addition, 23 trials reported the duration of the exposure to the intervention [38,40-42,46-51,53-65], which ranged between 2.5 min [58] and 12 months [38,59] (Multimedia Appendix 3).
All the studies used OCDE that was delivered by either personal computers or laptops. In addition, 19 studies used software- or computer-based programs delivered via a variety of sources such as CD-ROM and stored in the computer [38-41,44,46,47,50,52,53,55-57,59,60,62-65], 6 studies used video recording [42,48,49,51,58,61], 3 studies [43,45,54] investigated the use of multimedia, and 1 study used both computer-based program and video recording [37].
Primary Outcomes
As meta-analysis was not feasible, we presented the results in a narrative summary format (Multimedia Appendix 4) and Summary of Findings (Multimedia Appendix 3).
Doctors’ Knowledge
Among the 29 studies that compared OCDE with other interventions or to no intervention, knowledge was assessed in 13 studies (44%) [41-45,47,49,55,59,60,63-65]. Knowledge gain was assessed by multiple choice questions (MCQs) in 5 studies [42,47,59,63,64] and by test of true or false questions in 2 studies [44,60], none of the tests were validated. Either non-validated or low-internal validity open-ended questions were used in 3 studies to assess knowledge [43,49,65]: 1 study used a validated questionnaire comprising MCQs and structured questions [55], 1 used non-validated Likert scale questions [45], and 1 used invalidated anonymous scoring by the authors compared with a gold standard prepared by 3 neurologists [41] (Multimedia Appendix 3).
Doctors’ Cognitive Skills
Overall, 16 RCTs with 17 interventions [37,39,43,45-48, 50,52-54,56,58,61,62,65] assessed cognitive skills as an outcome. A total of 3 studies [48,50,54] used the validated anesthetist nontechnical skills scoring system to assess doctors’ cognitive skills. One study [62] used the objective structured clinical examination (OSCE) tool for evaluation of skills. Four studies [52,53,58,65] used thematic analysis to assess skills gain, including thinking process [58], task completion rate [65], number of empathetic statement responses to patients [53], and agreement between expert and participant in mental health capacity assessment [52].
Cognitive skills gain was assessed by many tools including calculating the accuracy of decision making within and between the study groups [39], questionnaires [45,56], and a 36-item checklist [37]. Multiple assessment tools were used in 2 studies to assess cognitive skills, Ottolini et al [46] used 2 tools, thematic analysis and a questionnaire, whereas Esfahani et al [61] used both the Jefferson empathy scale and OSCE. A questionnaire with open-ended questions was used in 1 study [43], and an MCQ test was used in another [47] (Multimedia Appendix 3).
Patients’ Outcomes
A total of 4 studies [38,40,53,57] assessed patient outcomes. Bonevski et al [57] examined doctors’ improved screening and detection of the patients’ risk behaviors using a self-reported survey. Millard et al [40] evaluated the improvement in dementia diagnosis, following a computer-generated audit of the participants’ practices. Tulsky et al [53] used a telephone survey to evaluate the patients’ trust in their oncologists and the oncologists’ perceived empathy and knowledge of the patients, following CD-ROM–based education. Lavigne et al [38] evaluated patients’ outcomes by measuring the improvement in the children’s attention-deficit/hyperactivity disorder (ADHD) symptoms using the Beck Anxiety Inventory, Swanson, Nolan, and Pehlam-IV Rating Scale, and the ADHD Rating Scales-IV completed by teachers and parents following the use of computer software education in medication dose titration (Multimedia Appendix 3).
Secondary Outcomes
The CE of OCDE computer-based digital learning was examined in 1 study [51]. The study used data from an RCT included in this review [48]. It compared the cost and effectiveness of self-debriefing versus instructor debriefing using net benefit regression. The CE estimate was reported as the incremental net benefit, and the uncertainty was presented using a CE acceptability curve. The study concluded that digital learning was cost-effective if the intervention cost was less than or equal to Can $200 in the 2012 rate.
We did not find any RCT that compared the adverse effects of OCDE with other interventions.
Risk of Bias in Included Studies
The assessment of risk of bias is described in detail in Multimedia Appendix 4 and shown on Figures 2 and 3. A total of 24 of the 29 included trials (83%) were judged to be at a high risk of bias [37-41,43-48,50-53,56-64]. In addition, 2 trials were judged to be of unclear risk of bias [42,49], whereas only 3 trials were judged to be at low risk of bias [54,55,65]. The assessment of the methodological quality of economic evaluation using the Evers checklist is presented in Multimedia Appendix 5.
Effects of the Interventions
The studies were divided into 4 comparisons, which evaluate the impact of OCDE compared with face-to-face learning, with no intervention, OCDE with text-based learning, and with another OCDE method.
Offline Computer-Based Digital Education Compared With Face-To-Face Learning
The characteristics of the studies are presented in (Multimedia Appendix 3); GRADE of evidence is presented in Summary of Findings (Multimedia Appendix 6).
Knowledge Gain
Overall, 4 studies [43,49,55,64] showed no significant difference in posttest knowledge scores between digital learning and face-to-face learning (ratio of the mean [RM] ranges from 1.0 to 1.13). In addition, 3 studies [41,60,63] showed the OCDE group to have significantly higher scores than the face-to-face group. However, the difference among the posttest mean scores of the participants was modest, as indicated by the small RM (RM=1.1, RM=1.17, and RM=1.13, respectively). One study [47] showed face-to-face participants to score higher than OCDE group (RM=0.95). The grade of evidence for this outcome is very low because of the high risk of bias in the included studies, the heterogeneity of participants and interventions, and the indirectness of evidence; therefore, it is uncertain whether there is difference in knowledge gain between OCDE and face-to-face learning.
Cognitive Skills Gain
Overall, 4 RCTs showed no significant difference between OCDE and face-to-face learning in posttest mean scores of skills gain [43,48,50,54] (RM range 0.94-1.0). In 2 studies [47,61], the mean posttest scores of face-to-face participants were significantly higher than those of the OCDE intervention (RM 0.91 and 0.95, respectively). In addition, 2 studies [62,53] showed the mean posttest score for the participants of the OCDE to be significantly higher than that of the face-to-face controls. The grade of evidence for this outcome is very low because of the high risk of bias in the included studies, the heterogeneity of participants and interventions, and the indirectness of evidence; therefore, it is uncertain whether there is a difference in cognitive skill gain between OCDE and face-to-face learning.
Patients’ Outcomes
Patients’ outcomes were examined in 2 trials [40,53]. Patients’ outcomes significantly improved in the OCDE group. The grade of evidence for this outcome is low because of the high risk of bias and heterogeneity of participants and interventions; therefore, OCDE may improve patients’ outcome compared with face-to-face learning.
Offline Computer-Based Digital Education Compared With No Intervention
The characteristics of the studies are presented in (Multimedia Appendix 3); GRADE of evidence is presented in Summary of Findings (Multimedia Appendix 6).
Knowledge Gain
Four RCTs [42,44,45,59] investigated the effect of OCDE compared with no intervention on knowledge gain. Of them, 3 trials [42,44,45] showed that OCDE was significantly more effective than no intervention with modest effect (RM range 1.11-1.36). The fourth trial [59] showed no significant difference in the posttest knowledge scores between intervention and control (RM=0.98). The grade of evidence for this outcome is low because of the high risk of bias and heterogeneity of participants and interventions; therefore, digital learning may improve knowledge gain compared with no intervention.
Overall, 3 trials [39,45,52] showed that OCDE had similar effect to no intervention in cognitive skills gain (RM=1.01), whereas Gordon et al [56] showed that participants in OCDE had significantly higher posttest scores compared with control (RM=1.25). The grade of evidence for this outcome is very low because of the high risk of bias, heterogeneity of participants and interventions, and indirectness of evidence; therefore, there is uncertainty about the effectiveness of OCDE compared with no intervention in cognitive skill gain.
Patients’ Outcomes
Only 1 cRCT, at high risk of bias, investigated the effect of OCDE compared with no intervention on patients’ outcome [38]. The trial was conducted among doctors of 24 pediatric practices where the number of doctors was not specified. It showed similar effectiveness in the treatment of patients with ADHD in intervention and control groups. The grade of evidence for this outcome was low because of the high risk of bias and the fact that evidence was drawn from a single study; therefore, OCDE may have an equal effect as no intervention in patients’ outcomes.
Offline Computer-Based Digital Education Compared With Text-Based Learning: Cognitive Skills Gain
The characteristics of the studies are presented in Multimedia Appendix 3; GRADE of evidence is presented in Summary of Findings Table (Multimedia Appendix 6).
Overall, 3 RCTs with 4 interventions [37,46,58] investigated the effect of OCDE compared with text-based learning on cognitive skills gain. In the 3 comparisons [37,46,58], OCDE was significantly more effective than text (RM range 1.14-1.46). In the fourth comparison [37], there was no difference in the posttest scores between the intervention and control groups (RM=0.91). The grade of evidence is very low because of the high risk of bias, small number of participants, and indirectness of evidence; therefore, there is uncertainty about the effect of OCDE compared with text-based learning in cognitive skills gain.
Offline Computer-Based Digital Education Compared With Other Digital Learning
The characteristics of the studies are presented in Multimedia Appendix 3.
Knowledge Gain
Only 1 trial at low risk of bias investigated knowledge gain [65]. It showed no difference in the effects of 2 methods of digital learning (RM=0.98). The body of evidence is considered low grade as the evidence is driven from a single study with a small number of participants.
Cognitive Skills Gain
Only 1 RCT, at low risk of bias, in this comparison investigated cognitive skill gain [65]. It showed equal effects from 2 methods of OCDE (RM=0.98). The body of evidence is considered low grade as the evidence is driven from a single study with a small number of participants.
Patients’ Outcomes
One RCT compared offline computer-based CME with feedback with the same CME without feedback in improving screening behavior as patients’ outcome [57]. The RCT showed better patients’ outcomes for CME with audit compared with the same program without audit. The body of evidence is considered very low grade as the evidence is driven from a single study at high risk of bias with a small number of participants.
Discussion
Principal Findings
This systematic review showed that the effectiveness of OCDE compared with other methods of education, on medical doctors’ knowledge and cognitive skill gain, is uncertain. OCDE may improve doctors’ knowledge compared with no intervention, but its effect on doctors’ cognitive skills is uncertain. OCDE may have little or no effect in improving patients’ outcomes.
The evidence for this review is driven from 29 RCTs, which covered a wide range of offline digital learning interventions in a variety of clinical and nonclinical medical disciplines. The studies investigated multiple outcomes of the intervention in 1690 doctors and dentists; therefore, they provide a considerable body of evidence. However, heterogeneity of participants, interventions, and methods of assessment of outcomes, in addition to the poor methodological quality of the trials, resulted in uncertainty about the effectiveness of OCDE compared with other instruction methods.
The quality of evidence for all outcomes was rated as low or very low (for different outcomes) on the GRADE scale because of the poor methodological quality of the included studies, as 24 out of the 29 included studies were judged to be at high risk of bias in addition to the marked clinical heterogeneity of the body of evidence.
It is worth noting that in all trials that compared OCDE with other types of learning, the measured outcome was the participants’ improved knowledge or skills (contents) rather than the methods of learning. This surrogate outcome (indirectness of evidence) is valid for the evaluation of the methods of learning as long as the assumption that the participants in the intervention and control groups had equal baseline knowledge with respect to the contents of the interventions is valid. Nevertheless, bias can be introduced if the participants have different levels of knowledge about the contents (eg, same content was taught in medical school, participants at different level of training) and no pre-intervention test was performed or was completed with an invalidated tool, which is the case in most of the trials included in this review. Furthermore, the body of evidence in this review has been drawn from small individual studies, as 12 of the included studies had less than 50 participants.
The external validity of the interventions in this review has been compromised by the recruitment of volunteers, which might have resulted in the selection of participants who were more computer literate, and therefore overestimated the effect of offline OCDE by excluding participants who did not know how to use the technology or were unwilling to do so. We believe variation in computer literacy and cultural differences may influence the generalization of our results to LMICs. Furthermore, most of the included studies were experimental trials conducted in ideal university hospital settings rather than implemented in programs in the field; therefore, the true applicability of OCDE could not be examined by this review.
The results of this review are inconsistent with previous evidence about the effectiveness of digital learning for health care professionals in improving knowledge and skills gain. A systematic review of 15 RCTs on the effects of digital learning (both on the Web and offline) showed digital learning to outperform or have equal effects as face-to-face learning in knowledge gain and practice improvement [10]. Similar effects of Web-based continuing education compared with face-to-face learning for medical doctors were found by Wulto et al [66] in their systematic review, which included 16 RCTs. Another systematic review [67], which examined the effectiveness of computer-based programs on the dentists’ performance, time spent, and attitude toward the programs, showed that in all the 12 included studies except 1, the computer-based programs were either better or similar to face-to-face learning in knowledge gain and that dentists had a positive attitude toward the program. More recent systematic reviews that investigated the effectiveness of a specific type of digital learning on the knowledge and skills gain of health professionals have shown similar results [68,69]. However, our conclusion of the effect of OCDE has been based on grading the evidence base of the effectiveness of OCDE, which we believe gives a more accurate evaluation of the effectiveness of the intervention on the desired outcomes. To that end, our conclusion agrees with a recently published Cochrane systematic review [70], which considered grading of the body of evidence an integral part of its conclusion.
Strengths and Limitations
To complete this review, we followed the robust methodology outlined by the Cochrane collaboration for searching, assessing, and reporting of the body of evidence for the effectiveness of OCDE in improving medical doctors’ knowledge and cognitive skills.
The review comprehensively evaluated the OCDE for medical doctors and dentists. The participants of the included studies are representative of the target population of medical doctors and dentists in training and non-training posts. Furthermore, the interventions accommodated participants from almost all clinical disciplines in addition to 4 fields of general knowledge and skills, including evidence-based medicine, research methodology, advanced life support, and cardiopulmonary resuscitation.
However, the heterogeneity of the participants, interventions, and outcomes precluded meta-analysis and subgroup analysis, which would have improved our certainty about the effectiveness of the intervention.
Implications for Practice and Research
The uncertainty associated with the effectiveness of OCDE for medical doctors’ education calls for limited-scale implementation of OCDE in the context of experimental settings and research.
Research in digital education should be employed to investigate effectiveness in updating medical doctors’ knowledge and skills, considering the patient as the center of care and the improvement of patients’ health as the main outcome, especially in LMICs. Future trials should follow a robust methodology, focusing on avoiding major biases by employing valid methods for randomization and allocation concealment, in addition to the use of validated tests to assess the outcomes. As most of the participants are not blinded to the intervention in these types of studies, there is high risk of attrition bias for any outcome that relied on active participation and follow-up (eg, demonstrating skills or taking a knowledge test). Such bias can be reduced by securing the anonymity of the participants, for example, replacing their names with numbers or letters.
The indirectness of evidence will continue to downgrade the evidence base of the effectiveness of OCDE unless validated pre and posttests are used to evaluate the outcomes in addition to attentive selection of trial participants who have no previous knowledge about the subject of the education (eg, new imaging technique).
Furthermore, we believe that the development of a common taxonomy for digital learning will facilitate easier comparison among studies and therefore better the quality of evidence.
Evaluating the CE of the various methods of digital learning is an important field for future research, considering the need of such interventions in LMICs.
Conclusions
The effectiveness of OCDE when compared with other methods of education, on medical doctors’ knowledge and cognitive skill gain, is uncertain. OCDE may improve doctors’ knowledge when compared with no intervention, but its effect on doctors’ cognitive skills is uncertain. OCDE may have little or no effect in improving patients’ outcome.
Acknowledgments
This review was conducted in collaboration with the Health Workforce Department at the World Health Organization. The authors would also like to thank Mr Carl Gornitzki, Ms GunBrit Knutssön, and Mr Klas Moberg from University Library, Karolinska Institute for developing and undertaking our search strategy. We gratefully acknowledge King Saud University, Deanship of Scientific Research, Research Chair for Evidence-Based Health Care and Knowledge Translation for funding the reviewers from King Saud University and the funding from Nanyang Technological University and National Healthcare Group Polyclinics, Singapore for this project. We gratefully acknowledge funding from the Lee Kong Chian School of Medicine, Nanyang Technological University Singapore, Singapore, e-learning for health professionals education grant.
Abbreviations
- ADHD
attention-deficit/hyperactivity disorder
- CE
cost-effectiveness
- CME
continuing medical education
- CPD
continuing professional development
- cRCT
cluster randomized controlled trial
- GRADE
Grading of Recommendations Assessment, Development and Evaluation
- LMICs
low- and middle-income countries
- MCQs
multiple choice questions
- MEDLINE
Medical Literature Analysis and Retrieval System Online
- OCDE
offline computer-based digital education
- OSCE
objective structured clinical examination
- RCT
randomized controlled trial
- RM
ratio of the means
- ICC
intracluster correlation coefficients
Search Strategy.
Excluded studies.
Characteristics of the included studies.
Risk of bias summary.
Evers checklist for economic evaluation.
Summary of findings tables.
Footnotes
Authors' Contributions: HW is the main review author and has been involved with all aspects of the review. HW, MS, PP, and LTC wrote the protocol and commented on the draft of the review. HW, SE, YA, AF, AJ, NZ, MT, KB, and ARS screened eligible studies, conducted the quality assessment of eligible studies, and extracted data from the included studies. HW, SE, YA, AF, AJ, MT, KB, and ARS assessed the risk of bias of the included studies. HW and ARS conducted the data synthesis and performed the GRADE data synthesis. HW, SE, YA, MT, KB, AF, and PP wrote the draft of the review.
Conflicts of Interest: None declared.
References
- 1.Davis N, Davis D, Bloch R. Continuing medical education: AMEE Education Guide No 35. Med Teach. 2008;30(7):652–66. doi: 10.1080/01421590802108323.902180057 [DOI] [PubMed] [Google Scholar]
- 2.Paixão MP, Miot HA, Wen CL. Tele-education on leprosy: evaluation of an educational strategy. Telemed J E Health. 2009;15(6):552–9. doi: 10.1089/tmj.2008.0137. [DOI] [PubMed] [Google Scholar]
- 3.Lawton S, Wimpenny P. Continuing professional development: a review. Nurs Stand. 2003;17(24):41–4. doi: 10.7748/ns.17.24.41.s52. [DOI] [PubMed] [Google Scholar]
- 4.Marinopoulos SS, Dorman T, Ratanawongsa N, Wilson LM, Ashar BH, Magaziner JL, Miller RG, Thomas PA, Prokopowicz GP, Qayyum R, Bass EB. Effectiveness of continuing medical education. Evid Rep Technol Assess (Full Rep) 2007 Jan;(149):1–69. [PMC free article] [PubMed] [Google Scholar]
- 5.O'Neil KM, Addrizzo-Harris DJ, American College of Chest Physicians Health and Science Policy Committee Continuing medical education effect on physician knowledge application and psychomotor skills: effectiveness of continuing medical education: American College of Chest Physicians Evidence–Based Educational Guidelines. Chest. 2009 Mar;135(3 Suppl):37S–41S. doi: 10.1378/chest.08-2516.S0012-3692(09)60172-0 [DOI] [PubMed] [Google Scholar]
- 6.Thepwongsa I, Kirby CN, Schattner P, Piterman L. Online continuing medical education (CME) for GPs: does it work? A systematic review. Aust Fam Physician. 2014 Oct;43(10):717–21. http://www.racgp.org.au/afp/2014/october/online-continuing-medical-education-(cme)-for-gps-does-it-work-a-systematic-review/ [PubMed] [Google Scholar]
- 7.Masic I. E-learning as new method of medical education. Acta Inform Med. 2008;16(2):102–17. doi: 10.5455/aim.2008.16.102-117. http://europepmc.org/abstract/MED/24109154 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Cook DA, Levinson AJ, Garside S, Dupras DM, Erwin PJ, Montori VM. Instructional design variations in internet–based learning for health professions education: a systematic review and meta-analysis. Acad Med. 2010 May;85(5):909–22. doi: 10.1097/ACM.0b013e3181d6c319.00001888-201005000-00042 [DOI] [PubMed] [Google Scholar]
- 9.Sinclair PM, Kable A, Levett-Jones T, Booth D. The effectiveness of Internet–based e-learning on clinician behaviour and patient outcomes: a systematic review. Int J Nurs Stud. 2016 May;57:70–81. doi: 10.1016/j.ijnurstu.2016.01.011.S0020-7489(16)00012-2 [DOI] [PubMed] [Google Scholar]
- 10.Lam-Antoniades M, Ratnapalan S, Tait G. Electronic continuing education in the health professions: an update on evidence from RCTs. J Contin Educ Health Prof. 2009;29(1):44–51. doi: 10.1002/chp.20005. [DOI] [PubMed] [Google Scholar]
- 11.Bouhnik D, Marcus T. Interaction in distance-learning courses. J Am Soc Inf Sci. 2006 Feb 1;57(3):299–305. doi: 10.1002/asi.20277. doi: 10.1002/asi.20277. [DOI] [Google Scholar]
- 12.Andreazzi DB, Rossi F, Wen CL. Interactive tele-education applied to a distant clinical microbiology specialization university course. Telemed J E Health. 2011 Sep;17(7):524–9. doi: 10.1089/tmj.2011.0010. [DOI] [PubMed] [Google Scholar]
- 13.Sezer B. Faculty of medicine students' attitudes towards electronic learning and their opinion for an example of distance learning application. Comput Human Behav. 2016 Feb;:932–9. doi: 10.1016/j.chb.2015.10.018. doi: 10.1016/j.chb.2015.10.018. [DOI] [Google Scholar]
- 14.Cantoni V, Cellario M, Porta M. Perspectives and challenges in e-learning: towards natural interaction paradigms. J Vis Lang Comput. 2004 Oct;:333–45. doi: 10.1016/j.jvlc.2003.10.002. doi: 10.1016/j.jvlc.2003.10.002. [DOI] [Google Scholar]
- 15.Cook DA. Web-based learning: pros, cons and controversies. Clin Med (Lond) 2007;7(1):37–42. doi: 10.7861/clinmedicine.7-1-37. https://www.tau.ac.il/medicine/cme/pituach/240210/4.pdf . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Alur P, Fatima K, Joseph R. Medical teaching websites: do they reflect the learning paradigm? Med Teach. 2002 Jul;24(4):422–4. doi: 10.1080/01421590220145815. [DOI] [PubMed] [Google Scholar]
- 17.Carroll C, Booth A, Papaioannou D, Sutton A, Wong R. UK health-care professionals' experience of on-line learning techniques: a systematic review of qualitative data. J Contin Educ Health Prof. 2009;29(4):235–41. doi: 10.1002/chp.20041. [DOI] [PubMed] [Google Scholar]
- 18.Greenhalgh T. Computer assisted learning in undergraduate medical education. Br Med J. 2001 Jan 6;322(7277):40–4. doi: 10.1136/bmj.322.7277.40. http://europepmc.org/abstract/MED/11141156 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Triola MM, Huwendiek S, Levinson AJ, Cook DA. New directions in e-learning research in health professions education: report of two symposia. Med Teach. 2012;34(1):e15–20. doi: 10.3109/0142159X.2012.638010. [DOI] [PubMed] [Google Scholar]
- 20.Balslev T, de Grave W, Muijtjens AM, Eika B, Scherpbier AJ. The development of shared cognition in paediatric residents analysing a patient video versus a paper patient case. Adv Health Sci Educ Theory Pract. 2009 Oct;14(4):557–65. doi: 10.1007/s10459-008-9138-y. [DOI] [PubMed] [Google Scholar]
- 21.Cook DA, McDonald FS. E-learning: is there anything special about the. Perspect Biol Med. 2008;51(1):5–21. doi: 10.1353/pbm.2008.0007.S1529879507100053 [DOI] [PubMed] [Google Scholar]
- 22.Higgins J, Green S. Cochrane Handbook for Systematic Reviews of Interventions Version 5. United States: 2011. http://handbook-5-1.cochrane.org/ [Google Scholar]
- 23.Car J, Carlstedt-Duke J, Tudor CL, Posadzki P, Whiting P, Zary N, Atun R, Majeed A, Campbell J. Digital education for health professions: methods for overarching evidence syntheses J Med Internet Res. Journal of medical internet research. 2019:-. doi: 10.2196/preprints.12913. (forthcoming)(forthcoming) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Ward JP, Gordon J, Field MJ, Lehmann HP. Communication and information technology in medical education. Lancet. 2001 Mar 10;357(9258):792–6. doi: 10.1016/S0140-6736(00)04173-8.S0140-6736(00)04173-8 [DOI] [PubMed] [Google Scholar]
- 25.Rasmussen K, Belisario JM, Wark PA, Molina JA, Loong SL, Cotic Z, Papachristou N, Riboli-Sasco E, Tudor CL, Musulanov EM, Kunz H, Zhang Y, George PP, Heng BH, Wheeler EL, al Shorbaji N, Svab I, Atun R, Majeed A, Car J. Offline eLearning for undergraduates in health professions: a systematic review of the impact on knowledge, skills, attitudes and satisfaction. J Glob Health. 2014 Jun;4(1):010405. doi: 10.7189/jogh.04.010405. http://europepmc.org/abstract/MED/24976964 .jogh-04-010405 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Puffer S, Torgerson D, Watson J. Evidence for risk of bias in cluster randomised trials: review of recent trials published in three general medical journals. Br Med J. 2003 Oct 4;327(7418):785–9. doi: 10.1136/bmj.327.7418.785. http://europepmc.org/abstract/MED/14525877 .327/7418/785 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Davey P, Brown E, Charani E, Fenelon L, Gould IM, Holmes A, Ramsay CR, Wiffen PJ, Wilcox M. Interventions to improve antibiotic prescribing practices for hospital inpatients. Cochrane Database Syst Rev. 2013;4:CD003543. doi: 10.1002/14651858.CD003543.pub3. [DOI] [PubMed] [Google Scholar]
- 28.Evers S, Goossens M, de Vet H, van Tulder M, Ament A. Criteria list for assessment of methodological quality of economic evaluations: Consensus on Health Economic Criteria. Int J Technol Assess Health Care. 2005;21(2):240–5. doi: 10.1017/S0266462305050324. [DOI] [PubMed] [Google Scholar]
- 29.GRADE Working Group. 2014. [2019-02-09]. http://www.gradeworkinggroup.org/
- 30.Schünemann H, Oxman A, Higgins J, Vist G, Glasziou P, Guyatt G. Cochrane Handbook for Systematic Reviews of Interventions. United States: 2011. [2019-02-09]. Presenting results and Summary of findings' tables https://handbook-5-1.cochrane.org/chapter_11/11_10_chapter_information.htm . [Google Scholar]
- 31.Bharati A, Datta A, Gupta S, Chun-Lema P, Ryan J. Resident education in ultrasonography: assessment of identification of basic cardiac anatomy and pathology after a multimedia tutorial. Ann Emerg Med S. 2012;60(4):S41. doi: 10.1016/j.annemergmed.2012.06.089. [DOI] [Google Scholar]
- 32.Juyal M, Ware A, Houk J, Mina R. Electronic learning module enhances rheumatology education. 2013 ACR/ARHP Annual Meeting; October 25-30, 2013; San Diego, California, US. 2013. https://acrabstracts.org/abstract/electronic-learning-module-enhances-rheumatology-education/ [Google Scholar]
- 33.Khan T, Gupta N, Wani SR. Comparison of classroom didactic teaching (DT) and computer based self-learning (Sl) in recognizing narrow band imaging (NBI) patterns for diminutive polyp (DP) histology characterization. Gastrointest Endosc. 2015;81(5):AB325. doi: 10.1016/j.gie.2015.03.1449. [DOI] [Google Scholar]
- 34.Lichtman AS, Goff B, Schreuder HW, Lenihan JJ, Lefebvre G, Parker W. A validation study: does use of an interactive computer based laparoscopic hysterectomy trainer expand cognitive surgical knowledge of ob/gyn trainees beyond that of traditional educational tools? J Minim Invasive Gynecol. 2014;21(6):S87. doi: 10.1016/j.jmig.2014.08.304. [DOI] [Google Scholar]
- 35.Muhammad A, Nutland S, Unsworth J. Electronic vs standard learning in hospital trust induction for junior doctors - a randomized trial. Br J Surg. 2010;97(S2):49–50. https://onlinelibrary.wiley.com/doi/epdf/10.1002/bjs.7050 . [Google Scholar]
- 36.Springer. 2015. Apr, Abstracts from the 38th Annual Meeting of the Society of General Internal Medicine https://link.springer.com/content/pdf/10.1007%2Fs11606-015-3271-0.pdf.
- 37.Rae A, Khatib M, Sarker S, Bello F. The Effect of a Computer Based Open Surgery Simulation of an Inguinal Hernia Repair on the Results of Cognitive Task Analysis Performance of Surgical Trainees: An Educational Trial. ASGBI International Surgical Congress 2015; April 22-24, 2015; Manchester, England. 2015. [Google Scholar]
- 38.Lavigne JV, Dulcan MK, LeBailly SA, Binns HJ, Cummins TK, Jha P. Computer-assisted management of attention-deficit/hyperactivity disorder. Pediatrics. 2011 Jul;128(1):e46–53. doi: 10.1542/peds.2010-2684. http://europepmc.org/abstract/MED/21669891 .peds.2010-2684 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Kay EJ, Silkstone B, Worthington HV. Evaluation of computer aided learning in developing clinical decision-making skills. Br Dent J. 2001 May 26;190(10):554–7. doi: 10.1038/sj.bdj.4801033a. doi: 10.1038/sj.bdj.4801033a. [DOI] [PubMed] [Google Scholar]
- 40.Millard FB, Thistlehwaite J, Spagnolo C, Kennedy LR, Baune BT. Dementia diagnosis: a pilot randomised controlled trial of education and IT audit to assess change in GP dementia documentation. Aust J Prim Health. 2008 Jan 10;14(3):141–9. doi: 10.1071/PY08046. doi: 10.1071/PY08046. [DOI] [Google Scholar]
- 41.Farrar M, Connolly AM, Lawson J, Burgess A, Lonergan A, Bye AM. Teaching doctors how to diagnose paroxysmal events: a comparison of two educational methods. Med Educ. 2008 Sep;42(9):909–14. doi: 10.1111/j.1365-2923.2008.03146.x.MED3146 [DOI] [PubMed] [Google Scholar]
- 42.Chao TT, Wendel GD, McIntire DD, Corton MM. Effectiveness of an instructional DVD on third- and fourth-degree laceration repair for obstetrics and gynecology postgraduate trainees. Int J Gynaecol Obstet. 2010 Apr;109(1):16–9. doi: 10.1016/j.ijgo.2009.10.016.S0020-7292(09)00613-4 [DOI] [PubMed] [Google Scholar]
- 43.D'Alessandro MP, Galvin JR, Erkonen WE, Albanese MA, Michaelsen VE, Huntley JS, McBurney RM, Easley G. The instructional effectiveness of a radiology multimedia textbook (HyperLung) versus a standard lecture. Invest Radiol. 1993 Jul;28(7):643–8. doi: 10.1097/00004424-199307000-00017. [DOI] [PubMed] [Google Scholar]
- 44.Garrett TJ, Selnow G, Dobkin JF, Healton C. Computer-assisted instruction in aids infection control for physicians. Teach Learn Med. 2009:2015–218. doi: 10.1080/10401339009539463. doi: 10.1080/10401339009539463. [DOI] [Google Scholar]
- 45.Hsieh NK, Herzig K, Gansky SA, Danley D, Gerbert B. Changing dentists' knowledge, attitudes and behavior regarding domestic violence through an interactive multimedia tutorial. J Am Dent Assoc. 2006 May;137(5):596–603. doi: 10.14219/jada.archive.2006.0254.S0002-8177(14)64859-5 [DOI] [PubMed] [Google Scholar]
- 46.Ottolini MC, Greenberg L. Development and evaluation of a CD-ROM computer program to teach residents telephone management. Pediatrics. 1998 Mar;101(3):E2. doi: 10.1542/peds.101.3.e2. [DOI] [PubMed] [Google Scholar]
- 47.Platz E, Liteplo A, Hurwitz S, Hwang J. Are live instructors replaceable? Computer vs classroom lectures for EFAST training. J Emerg Med. 2011 May;40(5):534–8. doi: 10.1016/j.jemermed.2009.08.030.S0736-4679(09)00755-0 [DOI] [PubMed] [Google Scholar]
- 48.Boet S. Self debriefing versus instructor debriefing: a prospective randomized trial. Can J Anesth/J Can Anesth. 2010 Nov 13;57(S1):S26–S26. doi: 10.1007/s12630-010-9415-0. https://www.cas.ca/2010/abstracts/2.%20SESA%20and%20Health/802764.pdf . [DOI] [Google Scholar]
- 49.Garcia-Rodriguez JA, Donnon T. Using comprehensive video-module instruction as an alternative approach for teaching IUD insertion. Fam Med. 2016 Jan;48(1):15–20. http://www.stfm.org/FamilyMedicine/Vol48Issue1/GarciaRodriguez15 . [PubMed] [Google Scholar]
- 50.Hards A, Davies S, Salman A, Erik-Soussi M, Balki M. Management of simulated maternal cardiac arrest by residents: didactic teaching versus electronic learning. Can J Anaesth. 2012 Sep;59(9):852–60. doi: 10.1007/s12630-012-9752-2. [DOI] [PubMed] [Google Scholar]
- 51.Isaranuwatchai W, Alam F, Hoch J, Boet S. A cost-effectiveness analysis of self-debriefing versus instructor debriefing for simulated crises in perioperative medicine in Canada. J Educ Eval Health Prof. 2017;13:44. doi: 10.3352/jeehp.2016.13.44. doi: 10.3352/jeehp.2016.13.44.jeehp.2016.13.44 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Naglie G, Silberfeld M, O'Rourke K, Fried B, Corber W, Bombardier C, Detsky A. A randomized trial of a decisional aid for mental capacity assessments. J Clin Epidemiol. 1993 Mar;46(3):221–30. doi: 10.1016/0895-4356(93)90069-D.0895-4356(93)90069-D [DOI] [PubMed] [Google Scholar]
- 53.Tulsky JA, Arnold RM, Alexander SC, Olsen MK, Jeffreys AS, Rodriguez KL, Skinner CS, Farrell D, Abernethy AP, Pollak KI. Enhancing communication between oncologists and patients with a computer-based training program: a randomized trial. Ann Intern Med. 2011 Nov 1;155(9):593–601. doi: 10.7326/0003-4819-155-9-201111010-00007. http://europepmc.org/abstract/MED/22041948 .155/9/593 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Welke TM, LeBlanc VR, Savoldelli GL, Joo HS, Chandra DB, Crabtree NA, Naik VN. Personalized oral debriefing versus standardized multimedia instruction after patient crisis simulation. Anesth Analg. 2009 Jul;109(1):183–9. doi: 10.1213/ane.0b013e3181a324ab.109/1/183 [DOI] [PubMed] [Google Scholar]
- 55.Davis J, Chryssafidou E, Zamora J, Davies D, Khan K, Coomarasamy A. Computer-based teaching is as good as face to face lecture-based teaching of evidence based medicine: a randomised controlled trial. BMC Med Educ. 2007 Jul 20;7:23. doi: 10.1186/1472-6920-7-23. https://bmcmededuc.biomedcentral.com/articles/10.1186/1472-6920-7-23 .1472-6920-7-23 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Gordon M, Chandratilake M, Baker P. Improved junior paediatric prescribing skills after a short e-learning intervention: a randomised controlled trial. Arch Dis Child. 2011 Dec;96(12):1191–4. doi: 10.1136/archdischild-2011-300577.archdischild-2011-300577 [DOI] [PubMed] [Google Scholar]
- 57.Bonevski B, Sanson-Fisher RW, Campbell E, Carruthers A, Reid AL, Ireland M. Randomized controlled trial of a computer strategy to increase general practitioner preventive care. Prev Med. 1999 Dec;29(6 Pt 1):478–86. doi: 10.1006/pmed.1999.0567.S0091-7435(99)90567-4 [DOI] [PubMed] [Google Scholar]
- 58.Balslev T, de GW, Muijtjens AM, Scherpbier AJ. Comparison of text and video cases in a postgraduate problem-based learning format. Med Educ. 2005 Nov;39(11):1086–92. doi: 10.1111/j.1365-2929.2005.02314.x.MED2314 [DOI] [PubMed] [Google Scholar]
- 59.Jensen ML, Mondrup F, Lippert F, Ringsted C. Using e-learning for maintenance of ALS competence. Resuscitation. 2009 Aug;80(8):903–8. doi: 10.1016/j.resuscitation.2009.06.005.S0300-9572(09)00316-5 [DOI] [PubMed] [Google Scholar]
- 60.Pelayo-Alvarez M, Albert-Ros X, Gil-Latorre F, Gutierrez-Sigler D. Feasibility analysis of a personalized training plan for learning research methodology. Med Educ. 2000 Feb;34(2):139–45. doi: 10.1046/j.1365-2923.2000.00383.x.med383 [DOI] [PubMed] [Google Scholar]
- 61.Esfahani MN, Behzadipour M, Nadoushan AJ, Shariat SV. A pilot randomized controlled trial on the effectiveness of inclusion of a distant learning component into empathy training. Med J Islam Repub Iran. 2014;28:65. http://europepmc.org/abstract/MED/25405130 . [PMC free article] [PubMed] [Google Scholar]
- 62.Farahmand S, Jalili E, Arbab M, Sedaghat M, Shirazi M, Keshmiri F, Azizpour A, Valadkhani S, Bagheri-Hariri S. Distance learning can be as effective as traditional learning for medical students in the initial assessment of trauma patients. Acta Med Iran. 2016 Sep;54(9):600–4. http://acta.tums.ac.ir/index.php/acta/article/view/5034 . [PubMed] [Google Scholar]
- 63.Farokhi RM, Zarifsanaiey N, Haghighi F, Mehrabi M. E-learning or in-person approaches in continuous medical education: a comparative study. IIOAB J. 2016;7(2):472–7. https://www.iioab.org/articles/IIOABJ_7.S2_472-476.pdf . [Google Scholar]
- 64.Khoshbaten M, Soleimanpour H, Ala A, Vahdati SS, Ebrahimian K, Safari S, Golzari SE, Ranjbarzadeh FS, Esfanjani RM. Which form of medical training is the best in improving interns' knowledge related to advanced cardiac life support drugs pharmacology? An educational analytical intervention study between electronic learning and lecture-based education. Anesth Pain Med. 2014 Feb;4(1):e15546. doi: 10.5812/aapm.15546. http://europepmc.org/abstract/MED/24719802 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Davids MR, Chikte UM, Halperin ML. Effect of improving the usability of an e-learning resource: a randomized trial. Adv Physiol Educ. 2014 Jun;38(2):155–60. doi: 10.1152/advan.00119.2013. http://www.physiology.org/doi/full/10.1152/advan.00119.2013?url_ver=Z39.88-2003&rfr_id=ori:rid:crossref.org&rfr_dat=cr_pub%3dpubmed .38/2/155 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Wutoh R, Boren SA, Balas EA. eLearning: a review of internet-based continuing medical education. J Contin Educ Health Prof. 2004;24(1):20–30. doi: 10.1002/chp.1340240105. [DOI] [PubMed] [Google Scholar]
- 67.Rosenberg H, Grad HA, Matear DW. The effectiveness of computer-aided, self-instructional programs in dental education: a systematic review of the literature. J Dent Educ. 2003 May;67(5):524–32. http://www.jdentaled.org/cgi/pmidlookup?view=long&pmid=12809187 . [PubMed] [Google Scholar]
- 68.Ilgen JS, Sherbino J, Cook DA. Technology-enhanced simulation in emergency medicine: a systematic review and meta-analysis. Acad Emerg Med. 2013 Feb;20(2):117–27. doi: 10.1111/acem.12076. doi: 10.1111/acem.12076. [DOI] [PubMed] [Google Scholar]
- 69.Mundell WC, Kennedy CC, Szostek JH, Cook DA. Simulation technology for resuscitation training: a systematic review and meta-analysis. Resuscitation. 2013 Sep;84(9):1174–83. doi: 10.1016/j.resuscitation.2013.04.016.S0300-9572(13)00246-3 [DOI] [PubMed] [Google Scholar]
- 70.Vaona A, Banzi R, Kwag KH, Rigon G, Cereda D, Pecoraro V, Tramacere I, Moja L. E-learning for health professionals. Cochrane Database Syst Rev. 2018 Jan 21;1:CD011736. doi: 10.1002/14651858.CD011736.pub2. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Search Strategy.
Excluded studies.
Characteristics of the included studies.
Risk of bias summary.
Evers checklist for economic evaluation.
Summary of findings tables.