Abstract
Background
The use of flipped classroom approach has become increasingly popular in health professions education. However, no meta-analysis has been published that specifically examines the effect of flipped classroom versus traditional classroom on student learning. This study examined the findings of comparative articles through a meta-analysis in order to summarize the overall effects of teaching with the flipped classroom approach. We focused specifically on a set of flipped classroom studies in which pre-recorded videos were provided before face-to-face class meetings. These comparative articles focused on health care professionals including medical students, residents, doctors, nurses, or learners in other health care professions and disciplines (e.g., dental, pharmacy, environmental or occupational health).
Method
Using predefined study eligibility criteria, seven electronic databases were searched in mid-April 2017 for relevant articles. Methodological quality was graded using the Medical Education Research Study Quality Instrument (MERSQI). Effect sizes, heterogeneity estimates, analysis of possible moderators, and publication bias were computed using the Comprehensive Meta-Analysis software.
Results
A meta-analysis of 28 eligible comparative studies (between-subject design) showed an overall significant effect in favor of flipped classrooms over traditional classrooms for health professions education (standardized mean difference, SMD = 0.33, 95% confidence interval, CI = 0.21–0.46, p < 0.001), with no evidence of publication bias. In addition, the flipped classroom approach was more effective when instructors used quizzes at the start of each in-class session. More respondents reported they preferred flipped to traditional classrooms.
Conclusions
Current evidence suggests that the flipped classroom approach in health professions education yields a significant improvement in student learning compared with traditional teaching methods.
Keywords: Flipped classroom, Flipped learning, Health professions education, Meta-analysis
Background
Many classes in higher education institutes now employ blended learning; whereby students learn in part at a supervised face-to-face location on campus, and in part through the Internet with some elements of student choice over place and pace [1].Of the many different models of blended learning in practice, the use of flipped classroom approach has become increasingly widespread [2–4].
Initially popularized in the United States [5], flipped classrooms replace teacher-led in-class instructions with individual homework or group activities [6]. Recently, the flipped classroom approach has made inroads into health professions education, and has even been touted “a new paradigm” in medical education [7]. Various health professions have adopted this instructional approach into their curricula [8]. A recent review of learner perceptions of flipped classrooms in health professions education [8] found an overwhelming positive response from students who attended flipped courses. More specifically, students expressed high levels of satisfaction with pre-class video lectures because the videos can be accessed at any time and as often as they desire. Students also highly regarded the use of small group discussion-based activities in flipped classroom face-to-face sessions because these sessions help increase their motivation to learn, enhance their level of engagement, and interest in the subject matter [8].
But does using the flipped classroom approach in health professions education really improve student learning? It is important to note that positive student perception toward flipped classrooms does not necessarily imply that this instructional approach will significantly improve student learning [8]. For example, flipping the evidence-based medicine (EBM) course by Ilic et al. [9] did not improve scores on the Berlin objective assessment of EBM competencies compared to its traditional counterpart, despite students reporting a positive perception of the flipped course.
Up to now the effectiveness of flipped classroom approach compared with traditional learning has not been established. Although several literature review studies have been conducted in health care professionals such as nursing [10, 11] and medical education [8, 12], no meta-analysis has been published that specifically examines the effect of flipped classroom versus traditional classroom on student learning. By traditional classroom, we refer to the approach of having students come to class during which teachers use a range of pedagogical strategies (e.g., lecture, case discussion, student presentation), and then students complete most of their homework after school [13, 14]. The uncertainty about the effectiveness of flipped classroom approach over traditional instruction provided the impetus for the current study. We opted to contrast flipped classrooms with traditional classrooms because the latter is still widely used in health professions education [15].
Conceptual framework
For the purposes of this study, we adopted the conceptual framework of flipped classroom approach by educause [16], one of the leading associations that focus on instructional technology in higher education, as a pedagogical strategy in which the typical lecture and homework elements of a course are reversed. In a typical traditional classroom, students listen to lectures in class and complete most of their homework after class. In a flipped classroom, students listen or watch pre-recorded lectures before class and perform active learning activities such as exercises, projects, or discussions [16]. Problem-based learning may be one of the activities used in flipped classroom [17].
Although pre-recorded lectures could be a podcast or other audio format, the use of videos has become so ubiquitous that the flipped classroom approach has come to be identified with pre-recorded videos [16, 18, 23]. Therefore, in the present review, we focused specifically on flipped classroom studies in which pre-recorded videos were provided (rather than live lectures, or intelligent tutoring system without a video or instructor) prior to face-to-face class meetings.
It is important to stress that this definition excludes the sole use of pre-class reading materials as a form of flipped learning. Text-based materials cannot “closely mimic what students in a traditional setting would experience” [19] since it does not involve instructors’ explanation and elaboration of contents [20]. Muir and Geiger [21] reported that “a book doesn’t really walk through the steps on how to do something.” In contrast, the use of video lectures enables the instructors to elaborate the course contents as in a traditional lecture [20, 21].
Methods
Data sources and search strategies
This meta-analysis and review were carried out according to the PRISMA (Preferred Reporting Items for Systematic reviews and Meta-Analysis) guidelines [22]. Relevant online databases were searched from January 2012 through March 2017. January 2012 onwards was chosen because 2012 was the year of the first publication of an application of flipped classroom approach to health professions student teaching [23]. Altogether, seven electronic databases were searched in mid-April 2017, including Academic Search Complete, PubMed, PsycINFO, CINAHL Plus, TOC Premier, British Nursing Index, and ERIC. To capture a broader range of potentially eligible articles, we employed the following search terms with Boolean operators: “(flip* or invert*) and (class* or learn* or instruction* or course*) and (medic* or nurs* or pharmac* or physiotherap* or dental or dentist* or chiropract*)”. The asterisk was used as a wild card to include most of the common expressions of the flipped classroom approach (e.g., flipped learning, flipped class, flipping the classroom). The search term was entered as a string and searched in each of the seven databases.
Eligibility criteria
To examine the possible effects of flipped classroom on student achievement, a meta-analysis was done on eligible articles. The eligibility criteria are as follows:
Inclusion
The studies must report at least one comparison of a flipped classroom condition versus a traditional classroom condition (i.e., between-subject design) focusing on health care professionals including medical students, residents, doctors, nurses, or learners in other health care professions and disciplines (e.g., dental, pharmacy, environmental or occupational health). The flipped classroom implementation must consist of both pre-class and in-class activities.
The pre-class flipped classroom activities must at least include the use of instructor-recorded classroom lectures, PowerPoints with instructor talking head, YouTube videos, Khan Academy videos, TED (technology, entertainment, design) video talks, screencast, or PowerPoints with instructor’s voice over.
In the present review, we include only flipped classroom studies that have a face-to-face meeting because face-to-face class meeting is typically used in many flipped classroom implementations [18, 24]. Comparing online courses (without face-to-face meeting) with flipped classrooms is outside the scope of our review.
The traditional classroom involves students coming to class during which teachers typically give a lecture [25], and might use other presentation strategies (e.g., case discussion, group-work, student presentation).
Only comparative studies such as randomized controlled trial, quasi-experiments, and historical cohort controlled research designs were included for review.
Articles must measure student learning on similar course topics using some form of objective-based assessments such as post-tests or exams. These assessment instruments must be similar or identical.
No geographical restrictions were imposed, however the articles must be written in English and published in peer-reviewed journals. Searching for peer-reviewed publications is a useful criterion for selecting studies of sufficient quality [26].
Exclusion
a) Published studies were excluded if their datasets or results were incomplete – such as if they lacked sufficient information to calculate effect sizes (e.g., sample sizes), or if effort to obtain data from corresponding authors was not successful.
Study selection and data extraction
The title and abstract of the studies from initial the search process were screened in order to derive a preliminary set of full articles for potential final review. This was followed by a reading of the preliminary full articles by the authors individually to confirm the relevance of the studies before a final decision (through discussion) was made to confirm the studies to be included in the final review. To extract the data, we looked for information including authors of the study, publication year, location in which the study was conducted, subject topic, participant sample, study design such as quasi-experiments, randomized control trials, or historical controls, and details of the flipped classroom implementation such as the types of pre-class and in-class activities used. The percent agreement between the coders concerning the data extraction was high – 95%. To reach consensus, the discrepancies between the extracted data of the two researchers were reviewed, discussed, and resolved prior to data entry and analysis.
For all the eligible studies, one effect size was calculated for each study to meet the assumption of the independence of the effect sizes based on independent samples of students. In cases where articles reported multiple assessments of a single course subject, we selected the assessment that was most summative, as suggested by Freeman et al. [27]. For example, we chose final exam over other assessments (e.g., mid-term exam, weekly quiz). We also chose assessment that is recognized as a widely-used inventory [27] – for example, the Objective Structured Clinical Examination over an instructor-written examination. In cases, where a single study had multiple outcomes from different course subjects from the same set of students, we computed a single combined effect size using the formulas proposed by Borenstein et al. [28]. In doing so, we assumed that the correlation between the different outcomes within a comparison was 1 as suggested by Freeman et al. [27] since the same students were sampled for each outcome. This is a conservative measure as the actual correlation between outcomes is likely lower than 1 [27].
Meta analyses
We computed effect sizes using the Comprehensive Meta-Analysis Version 3 software (Biostat, Inc., Englewood Cliffs, NJ, USA). All reported p values are two-tailed, unless otherwise reported. To compare the effect sizes, we used a random effects model, or random effects analysis because conditions that could affect student achievements differed among studies in the analysis, including the frequency of lessons flipped, student population, and course level. We computed effect sizes using standardized mean differences (SMDs) from the means and standard deviations of student achievement data (e.g., exam scores, post-test scores). If standard errors were used in the previous empirical studies but not the standard deviations, we used the following formula [29] to calculate the standard deviations:
If the means and standard deviations were not reported in the previous empirical studies, the standardized mean difference was estimated using a variety of sources, including t-tests (or formulas, see Borenstein et al. [28]; Lipsey & Wilson [30]). The presence of heterogeneity (i.e. the degree of inconsistency in the studies’ results) was detected by the I2 test. Publication bias (or otherwise known as file-drawer problem) occurs when researchers publish only favorable results [63]. Therefore, in order to determine whether the present review suffers from publication bias, we conducted the following standard tests used to analyze publication bias including: (a) assessing the funnel plot, (b) computing Begg and Mazumdar rank correlation, (c) calculating Egger’s regression, (d) computing Duval and Tweedie’s trim and fill, and (e) calculating the classic fail-safe N test. Currently, sufficient information was available for all these tests.
We also conducted subgroup analyses across six major categorical moderator variables in order to identify the possible source of variation among the effect sizes and the differences among the subgroups. These moderator variables were: (a) student initial equivalence, (b) instructor equivalence, (c) research design, (d) types of students; (e) pre-class component of flipped classroom, and (f) in-class component of flipped classroom.
To determine student initial equivalence, we examined whether the study design was based on the following categories: (a) comparative studies where authors provided no data or no statistical control on student equivalence in terms of initial academic performance, or where authors merely claimed students were equivalent but did not provide any relevant statistical evidence (e.g., t-test results); or (b) comparative studies where data indicated no statistical difference on a pretest that directly relate to the topic, or on a metric of academic performance (e.g., college GPA). To evaluate instructor equivalence, we checked whether the study involved: (a) identical instructor for the flipped-and-non-flipped classes; or (b) different instructors for different classes; or (c) no data provided.
In addition, we stratified the analysis according to research design (randomized control, quasi-experiment, or historical control), types of students (e.g., medicine, pharmacy, public health, etc.), pre-class component (availability of pre-class assessment/exercise or not, availability of readings/notes or not), and in-class component (use of quiz at start of lesson or not).
Results
The literature search process is depicted in Fig. 1. Our initial search yielded a total of 2129 articles, which was reduced to 1655 after duplicates were removed. However, after reviewing their titles and abstracts, many articles (n = 1611) were found to be irrelevant, particularly those that did not report empirical research, or were not related to health professions education. For example, many irrelevant search outcomes came from articles that did not collect and analyze any form of data (i.e., non-empirical), and from basic science disciplines (e.g., inverted structural plasticity, invertebrates, and tissue classification). A total of 44 full text articles were read and assessed for eligibility. Out of these 44 articles, six were removed because it did not fulfill our criteria of a flipped classroom approach, eight were removed due to insufficient data on student performance or descriptions of the learning activities, and three were removed because the flipped course and traditional course offering were not comparable. Twenty-eight articles were included in the final analysis.
Study demographics
Table 1 descriptively summarizes the main elements of the 28 studies included in the meta-analysis. Most of the studies were based on historical control designs. Only four studies were quasi-experiments. The remaining four were randomized controlled trials.
Table 1.
Study | Country of origin | Subject matter | Sample | Study design | Duration of intervention |
---|---|---|---|---|---|
Bakr et al. (2016) [31] | Australia | Dental anatomy | 1st year dental students | Historical control | 13 weeks |
Bossaer et al. (2016) [32] | USA | Pharmacy | 3rd year pharmacy students | Historical control | 1 semester (actual duration not specified) |
Cheng et al. (2016) [33] | China | Histology | 1st year medicine students | Quasi-experiment | 1 semester (actual duration not specified) |
Cotta et al. (2016) [13] | USA | Pharmaceutical calculations | 1st year pharmacy students | Historical control | 10 weeks |
Galway et al. (2014) [34] | Canada | Environmental and occupational health | 1st or 2nd year graduate public health students | Historical control | 13 week |
Gillispie (2016) [35] | USA | Obstetrics and gynecology | 3rd/4th year medical students | Historical control | 8 weeksa |
Giuliano and Moser (2016) [36] | USA | Drug literature | 1st year pharmacy students | Historical control | 1 semester (actual duration not specified) |
Harrington et al. (2015) [37] | USA | Nursing | Baccalaureate students | Randomized | 4 months |
Hsu et al. (2016) [38] | Taiwan | General medicine | Postgraduate year medical students | Historical control | 12 months |
Ilic et al. (2015) [9] | Australia | Evidence-based medicine | 3rd year medical students | Randomized | 10 two-hour teaching sessions |
Kiviniemi (2014) [39] | USA | Public health | Graduate students | Historical control | 1 semester (actual duration not specified) |
Koo et al. (2016) [40] | USA | Pharmacotherapy | 2nd year doctor of pharmacy students | Historical control | 8 sessions |
Liebert et al. (2016) [41] | USA | Surgery | 3rd year medical student | Historical control | 8 weeks |
Lin et al. (2017) [14] | China | Ophthalmology | Medical undergraduates | Randomized | 1 semester (actual duration not specified) |
McLaughlin et al. (2013) [42] | USA | Pharmaceutics | 1st year pharmacy students | Historical control | 13 weeks |
McLaughlin et al. (2014) [43] | USA | Pharmaceutics | 1st year professional students | Historical control | 13 weeks |
Morton and Colbert-Getz (2016) [44] | USA | Anatomy | 1st year medical students | Historical control | 17 weeks |
Munson and Pierce (2015) [45] | USA | Pharmacogenomics | Mixture of students with bachelor degree and students without first degree | Historical control | 15 weeks |
O’Connor et al. (2016) [15] | USA | Radiology | 3rd or 4th year medical students | Quasi-experiment | 2 weeks |
Pierce and Fox (2012) [6] | USA | Pharmacotherapy | Pharmacy students | Historical control | 8 weeks |
Porcaro et al. (2016) [46] | Australia | Hematology | Mixture of undergraduate and postgraduate students in medical science | Historical control | 1 semester (actual duration not specified) |
Prescott et al. (2016) [47] | USA | Patient assessment | 1st year pharmacy students | Historical control | 2 semesters (actual duration not specified) |
Rui et al. (2017) [48] | China | Medical diagnostics | Junior-year medical students | Randomized | 3 weeks |
Sajid et al. (2016) [49] | Saudi Arabia | Hematology | 3rd year medical students | Historical control | 5 lectures |
Street et al. (2015) [17] | USA | Physiology | 1st year medical students | Quasi-experiment | 6 weeks |
Tune et al. (2013) [50] | USA | Physiology | 1st year graduate medical students | Quasi-experiment | 1 semester (actual duration not specified) |
Whillier and Lystad (2015) [51] | Australia | Neuroanatomy | 2nd year undergraduate chiropractic students | Historical control | 5 weeks |
Wong et al. (2014) [52] | USA | Pharmacy | 1st year pharmacy students | Historical control | 1 week |
anote: Gillispie (2016) [35] - 8 weeks information obtained from the website: https://medicine-program.uq.edu.au/current-students/placements/core-rotations
Methodological quality was graded using the Medical Education Research Study Quality Instrument (MERSQI) [53]. We summarize the study quality in Table 2. The mean MERSQI score was 12.5 on an 18-point scale.
Table 2.
Scale item (max. points) | Subscale (points if present) | No. (%) present |
---|---|---|
Study design (max. 3) | Non-randomized 2-group (2) | 24 (86) |
Randomized 2-group (3) | 4 (14) | |
Sampling: no. of institutions (max. 1.5) | 1(0.5) | 26 (93) |
2 (1) | 1 (3.5) | |
3 (1.5) | 1 (3.5) | |
aSampling: response rate (max. 1.5) | < 50% or not reported (0.5) | 5 (17) |
50%–74% (1) (1) | 1 (3.5) | |
≥75% (1.5) | 22 (79) | |
Type of data: outcome assessment (max. 3) | Objective (3) | 28 (100) |
Validity evidence (max. 3) | Content (1) | 28 (100) |
Data analysis: appropriate (max. 1) | Appropriate (1) | 28 (100) |
Data analysis: sophistication (max. 2) | Beyond descriptive analysis (2) | 28 (100) |
Highest outcome type (max. 3) | Knowledge, skills (1.5) | 28 (100) |
aNote: Sampling (response rate) refers to the proportion of students enrolled who completed the flipped classroom approach
Meta-analysis
A meta-analysis of 28 eligible comparative studies involving 2295 subjects exposed to flipped classroom and 2420 subjects exposed to traditional classroom showed an overall significant effect in favor of the flipped classroom approach for health professions education (SMD = 0.33, 95% CI 0.21–0.46, p < 0.001) as shown in Fig. 2. A significant Q statistic (p < 0.001) indicated the presence of heterogeneity (I2 = 75.6%).
Moderator analyses
In order to explore the possible causes for the significant heterogeneity, we conducted several moderator analyses using the random-effects mode. The results of the analyses are summarized in Table 2.
When we analyzed the data based on whether the studies controlled for student or instructor equivalence, we found no evidence of heterogeneity between studies that reported initial student equivalence, studies that employed randomized student assignment, and studies that did not provide any such data (Q = 0.43, df = 2, p = 0.81).
Analyzing variation with respect to instructor equivalence also suggested no evidence of heterogeneity (Q = 4.72, df = 2, p = 0.09). Thus, the overall effect size for student performance data appears to be robust to varying methodological rigor of published studies (e.g., poorly controlled studies with different instructors, or with no data provided on student or instructor equivalence).
Heterogeneity analyses also indicated no significant variation when comparing (a) studies with different research design such as historical control, quasi-experiment, or randomized experiment (Q = 0.52, df = 2, p = 0.77); (b) studies with different types of students such as medicine, pharmacy (Q = 2.44, df = 5, p = 0.78); (c) studies that employed pre-class assessment/exercise, or not (Q = 2.67, df = 1, p = 0.10); or (d) studies that provided pre-class readings/notes, or not (Q = 0.11, df = 1, p = 0.74).
However, a heterogeneity analysis in Table 3 indicated that the effect size is significantly higher when instructor(s) employed quizzes at the start of an in-class session to assess students’ learning of the pre-class video contents as opposed to instructor(s) who did not (Q = 5.34, df = 1, p = 0.02).
Table 3.
Moderator | 95% CI | ||||
---|---|---|---|---|---|
n | SMD | SE | LL | UL | |
For student equivalence | |||||
No data provided | 13 | 0.31 | 0.10 | 0.11 | 0.50 |
No statistical difference (pretest, or other scores) | 11 | 0.39 | 0.10 | 0.18 | 0.59 |
Randomized assignment | 4 | 0.28 | 0.18 | −0.08 | 0.63 |
For instructor equivalence | |||||
No data | 17 | 0.23 | 0.08 | 0.07 | 0.39 |
Different instructors | 3 | 0.39 | 0.19 | 0.01 | 0.76 |
Identical instructors | 8 | 0.55 | 0.12 | 0.31 | 0.79 |
Research design | |||||
Historical control | 20 | 0.32 | 0.08 | 0.17 | 0.48 |
Quasi-experiment | 4 | 0.45 | 0.18 | 0.09 | 0.81 |
Random | 4 | 0.28 | 0.18 | −0.08 | 0.63 |
Types of students | |||||
Medicine | 13 | 0.26 | 0.10 | 0.08 | 0.45 |
Pharmacy | 10 | 0.45 | 0.10 | 0.24 | 0.65 |
Public health | 2 | 0.40 | 0.29 | −0.18 | 0.98 |
Nursing | 1 | 0.08 | 0.36 | −0.62 | 0.79 |
Dental | 1 | 0.19 | 0.30 | −0.39 | 0.78 |
Chiropractic | 1 | 0.29 | 0.39 | −0.47 | 1.04 |
Availability of pre-class assessment/ exercise? | |||||
No | 17 | 0.25 | 0.08 | 0.09 | 0.41 |
Yes | 11 | 0.46 | 0.10 | 0.26 | 0.65 |
Availability of pre-class readings/ notes? | |||||
No | 18 | 0.35 | 0.08 | 0.19 | 0.51 |
Yes | 10 | 0.30 | 0.11 | 0.08 | 0.53 |
Availability of quiz at start of in-class? | |||||
No | 20 | 0.26 | 0.07 | 0.12 | 0.38 |
Yes | 8 | 0.56 | 0.11 | 0.34 | 0.78* |
n number of studies, SMD standardized mean difference, SE standard error, 95% CI 95% confidence interval, LL lower limit, UL upper limit, *p < 0.05
Publication bias
Visual inspection of Fig. 3 suggested no presence of publication bias. This is supported by two statistical analyses: Begg and Mazumdar rank correction (Kendall’s Tau with continuity correction) = 0.08, one-tailed p = 0.27; and Egger’s regression intercept 0.79, one-tailed p = 0.24. Computation of Duval and Tweedie’s trim and fill method using the random effects model revealed no studies were trimmed using the random effects model. We also conducted a classic fail-safe N test to determine the number of null effect studies needed to raise the p value associated with the mean effect above an arbitrary alpha level (α = 0.05). Results showed that 747 additional missing studies with zero mean effect size would be required to make the overall effect statistically insignificant. There would therefore have to be an unreasonably large number of undetected studies with zero effect to bring the effect sizes reported in this paper to values that might be statistically insignificant. Based on the visual inspection of funnel plot, statistical analyses, and class fail-safe N, we believe that the overall mean effect size is not inflated by publication bias.
Analysis of learners’ preference
In the course of the review, we found five articles explicitly compared student preference for flipped or traditional classroom via survey questionnaires (Table 4). Overall, among the studies listed in Table 4, preference for flipped classroom was reported by an average of 70% of total respondents.
Table 4.
Study | Prefer flipped classroom n (%) | Prefer traditional classroom n (%) | Total respondents |
---|---|---|---|
Cotta et al. (2016) [13] | 87 (73) | 33 (27) | 120 |
Galway et al. (2014) [34] | 9 (82) | 2 (18) | 11 |
Giuliano and Moser (2016) [36] | 49 (60) | 33 (40) | 82 |
aKiviniemi (2014) [39] | 33 (89) | 4 (11) | 37 |
Porcaro et al. (2016) [46] | 50 (71) | 20 (29) | 70 |
Total | 228 (71) | 92 (29) | 320 |
aThe remaining 3 participants in Kiviniemi [57] indicated a preference for a combination of flipped and traditional classrooms
Discussion
Overall, the data reported in this study indicate that more students favored the flipped classroom approach over traditional classroom. In addition, the flipped classroom approach was more effective than traditional classroom in increasing student learning performance. One explanation for the more positive student perception, as well as the greater effect of flipped classroom over traditional classroom, is that having unrestricted access to pre-recorded video lectures before class enables students to learn anywhere and at any time, at their own pace [36]. Students can also watch the videos multiple times to better understand a particular topic [13, 36]. Another explanation is the availability of more in-class active learning time to help increase students’ understanding of the subject material. Many of the in-class activities such as small-group discussion promoted students’ interactions with their peers [34]. Instructors also felt they had greater opportunity to provide more feedback during in-class sessions [34]. There were also greater opportunities for students to apply their knowledge in flipped classes [13, 34].
Further analyses suggest that the increase in performance holds across studies of different research designs or methodological quality. Experiments where students were randomly assigned to flipped classrooms produced results that were indistinguishable from quasi-experiments and historical controls. Analyzing variation with respect to controls over student or instructor equivalence also produced no evidence of heterogeneity. In addition, the availability of online assessment/exercises, or readings before face-to-face class did not appear to moderate performance gains, as no heterogeneity was detected between the subgroups.
However, we found that the use of quizzes at the start of a face-to-face class would make flipped classroom more effective. The quizzes consisted of specific questions that were developed by the instructor beforehand and were used to assess student learning of the pre-class learning materials. One explanation for this finding is that quizzes at the beginning of class helped students recall the knowledge learned prior to the class. Prior knowledge has long been considered an important factor influencing learning [54, 55]. Stimulating the recall of prior knowledge helps learners to make better sense of new information by connecting it to something they already know. In addition, the retrieval of information from memory makes the path to that information in memory stronger; this consequently enables the information to be more easily retrieved by the learner on the next occasion [56].
Having quizzes at the beginning of class also allows an instructor to identify students’ possible misconceptions of the pre-class materials. Students’ misconceptions can prevent further learning if not addressed. Based on student performance, instructors can provide remedial action if necessary such as reviewing the pre-class video lectures or making adjustments to the in-class teaching plans to specifically address the students’ misconceptions. The use of quizzes at the beginning of an in-class lesson can also serve as a strong motivator for students to watch the pre-class video lectures [34, 50]. This finding thus implies that instructors use quizzes as a regular part of the in-class activities to assess students’ mastery of the pre-class learning materials.
Students who preferred a traditional classroom reported that watching video lectures took a lot of additional time [13]. In a traditional class, students learn about the subject matter through a teacher-led lecture format during class time [13]; however in a flipped class students are now required to watch the video lectures before class. Students were unhappy being asked to do work at home that was traditionally done in a face-to-face class format, and considered watching the pre-class videos as burdensome in terms of time [13]. Studies from several non-health professions education flipped classrooms also supported this finding. For example, half of the students who would unwilling to take another flipped class cited the additional time required to complete the pre-class work as a reason [57]. This finding thus implies that instructors who wish to employ flipped classroom should first promote students’ understanding of this new instructional approach by explaining the rationale, and potential benefits of flipped classroom [20]. In addition, instructors may consider limiting total length of all combined video segments to about 20 min. Support for this comes from several non health professions education-related flipped classroom studies [58, 59] which reported that most students spent up to 20–25 min on viewing pre-class video lectures.
Strengths and limitations
Our meta-analysis has several strengths. The flipped classroom approach has grown rapidly and is now widely used in health professions education. To our knowledge, this is the first meta-analysis to summarize the evidence to date concerning the effectiveness of flipped classroom on student learning compared with traditional classroom. Evaluating the effectiveness of flipped classroom is therefore both timely and important for instructors and learners. We also intentionally kept our literature search very recent (up to April 15, 2017) and broad in terms of subjects from various health professions using multiple academic online databases.
However, there are some limitations that should be considered. First, this review focused on flipped classroom studies in which pre-class videos were provided, and class attendance was mandatory. To broaden the scope of review, future reviews can examine other flipped learning studies that do not restrain the instructors’ use of technological tools. Future reviews can also compare the use of online course without face-to-face meeting and flipped classes with face-to-face meeting. Second, the flipped classroom designs in the reviewed studies were not always clearly reported. For example, the specific types of video lecture used were not described. Video lectures can include many different styles including recorded classroom lecture, Khan-style freehand writing video, PowerPoint presentation with instructor talking head, PowerPoint presentation with more than one people in conversation, among others [60–62]. Different video styles may affect student learning. We also could not identify information related to the actual time allocated to different instructional activities (e.g., small-group activities), and the actual details of small-group learning activities. Specifically, with regard to the small-group learning activities, there is no consensus about what various activities actually entail in practice [27]. As a result, we could not really differentiate the activities when the authors merely stated the use of small-group discussion, in-class collaboration, or group problem solving without providing specific details on the actual tasks involved in the group activities because group discussion, collaboration, problem-solving all involved discussions; and problem-solving can also be a form of collaboration. The absence of all this information prevented us from conducting further moderator analyses to discern additional key factors that could affect flipped classroom effectiveness. Additionally, the results are limited because no study included long-term follow-ups to assess learning retention. Also, the overall effect size for randomized controlled trials is usually bigger than for cohort studies.
Conclusions
Current evidence suggests that the flipped classroom approach in health professions education overall yields a statistically significant improvement in learner performance compared with traditional teaching methods. In addition, the flipped classroom would be more effective when instructors use quizzes at the start of each in-class session. Future research can be conducted to examine the possible effect of specific types of teaching method or presentation on student learning. Future research should also examine the possible impact of video styles. Despite the increasing popularity of using video-recorded lectures, we still understand little about how different video styles may impact student learning. Longitudinal studies should also be conducted to examine whether the flipped classroom approach can foster learning retention over a long period of time.
Acknowledgements
Not applicable.
Funding
No funding was received.
Availability of data and materials
The data used in this study can be found in previously published manuscripts.
Abbreviations
- CI
Confidence interval
- EBM
Evidence-based medicine
- LL
Lower limit
- MERSQI
Medical education research study quality instrument
- PRISMA
Preferred reporting items for systematic reviews and meta-analysis
- SE
Standard error
- SMD
Standardized mean differences
- TED
Technology, entertainment, design
- UL
Upper limit
Authors’ contributions
Both authors contributed to the design of the study. KFH performed the statistical analysis and wrote the first draft. Both authors carried out data collection and data coding. CKL helped to revise the manuscript. Both authors read and approved the final manuscript.
Ethics approval and consent to participate
Ethical approval and consent to participate are not applicable for this study.
Consent for publication
Consent for publication is not applicable for this work.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Khe Foon HEW, Phone: (852) 2859-1903, Email: kfhew@hku.hk.
Chung Kwan LO, Email: cklohku@hku.hk.
References
- 1.Horn MB, Staker HC. The rise of K-12 blended learning. San Mateo: Innosight Institute, Inc.; Retrieved from https://www.christenseninstitute.org/wp-content/uploads/2013/04/The-rise-of-K-12-blended-learning.pdf. 2011.
- 2.Giannakos MN, Krogstie J, Chrisochoides N. Proceedings of the computer science education research conference. 2014. Reviewing the flipped classroom research: reflections for computer science education; pp. 23–29. [Google Scholar]
- 3.Karabulut-Ilgu A, Jaramillo Cherrez N, Jahren CT. A systematic review of research on the flipped learning method in engineering education. Br J Educ Technol. 2017; [cited 2017 Mar 29]. Available from: http://onlinelibrary.wiley.com/doi/10.1111/bjet.12548/full
- 4.O’Flaherty J, Phillips C. The use of flipped classrooms in higher education: a scoping review. Internet High Educ. 2015;25:85–95. doi: 10.1016/j.iheduc.2015.02.002. [DOI] [Google Scholar]
- 5.Bergmann J, Sams A. Remixing chemistry class. Learn Lead Technol. 2008;36(4):24–27. [Google Scholar]
- 6.Pierce R, Fox J. Vodcasts and active-learning exercises in a “flipped classroom” model of a renal pharmacotherapy module. Am J Pharm Educ. 2012;76(10):196. doi: 10.5688/ajpe7610196. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Mehta NB, Hull AL, Young JB, Stoller JK. Just imagine: new paradigms for medical education. Acad Med. 2013;88(10):1418–1423. doi: 10.1097/ACM.0b013e3182a36a07. [DOI] [PubMed] [Google Scholar]
- 8.Ramnanan CJ, Pound LD. Advances in medical education and practice: student perceptions of the flipped classroom. Adv Med Educ Pract. 2017;8:63–73. doi: 10.2147/AMEP.S109037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Ilic D, Nordin RB, Glasziou P, Tilson JK, Villanueva EA. Randomised controlled trial of a blended learning evaluation intervention for teaching evidence-based medicine. BMC Med Educ. 2015;15:39. doi: 10.1186/s12909-015-0321-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Betihavas V, Bridgman H, Kornhaber R, Cross M. (2016). The evidence for ‘flipping out’: a systematic review of the flipped classroom in nursing education. Nurse Educ Today. 2016;38:15–21. doi: 10.1016/j.nedt.2015.12.010. [DOI] [PubMed] [Google Scholar]
- 11.Presti CR. The flipped learning approach in nursing education: a literature review. J Nurs Educ. 2016;55(5):252–257. doi: 10.3928/01484834-20160414-03. [DOI] [PubMed] [Google Scholar]
- 12.Chen F, Lui AM, Martinelli SM. A systematic review of the effectiveness of flipped classrooms in medical education. Med Educ. 2017;51(6):585–597. doi: 10.1111/medu.13272. [DOI] [PubMed] [Google Scholar]
- 13.Cotta KI, Shah S, Almgren MM, Macías-Moriarity LZ, Mody V. Effectiveness of flipped classroom instructional model in teaching pharmaceutical calculations. Curr Pharm Teach Learn. 2016;8(5):646–653. doi: 10.1016/j.cptl.2016.06.011. [DOI] [Google Scholar]
- 14.Lin Y, Zhu Y, Chen C, Wang W, Chen T, Li T, Li Y, Liu B, Lian Y, Lu L, Zou Y, Liu Y. Facing the challenges in ophthalmology clerkship teaching: is flipped classroom the answer? PLoS One. 2017;12(4):e0174829. doi: 10.1371/journal.pone.0174829. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.O’Connor EE, Fried J, McNulty N, Shah P, Hogg JP, Lewis P, Zeffiro T, Agarwal V, Reddy S. Flipping radiology education right side up. Acad Radiol. 2016;23(7):810–822. doi: 10.1016/j.acra.2016.02.011. [DOI] [PubMed] [Google Scholar]
- 16.Educause. 7 things you should know about flipped classrooms. [cited 2017 Mar 29]. Available from https://www.rit.edu/academicaffairs/tls/sites/rit.edu.academicaffairs.tls/files/directory/ELI7081-1.pdf. 2012.
- 17.Street SE, Gilliland KO, McNeil C, Royal K. The flipped classroom improved medical student performance and satisfaction in a pre-clinical physiology course. Med Sci Educ. 2015;25(1):35–43. doi: 10.1007/s40670-014-0092-4. [DOI] [Google Scholar]
- 18.Bishop JL, Verleger MA. Proceedings of the 120th ASEE National Conference and exposition, Atlanta, GA (paper ID 6219) 2013. The flipped classroom: a survey of the research. [Google Scholar]
- 19.Love B, Hodge A, Corritore C, Ernst DC. Inquiry-based learning and the flipped classroom model. Primus. 2015;25(8):745–762. doi: 10.1080/10511970.2015.1046005. [DOI] [Google Scholar]
- 20.Lo CK, Hew KFA. Critical review of flipped classroom challenges in K-12 education: possible solutions and recommendations for future research. Res Pract Technol Enhanc Learn. 2017;12(1):4. doi: 10.1186/s41039-016-0044-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Muir T, Geiger V. The affordances of using a flipped classroom approach in the teaching of mathematics: a case study of a grade 10 mathematics class. Math Educ Res J. 2016;28(1):149–171. doi: 10.1007/s13394-015-0165-8. [DOI] [Google Scholar]
- 22.Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. Ann Intern Med. 2009;151(4):264–269. doi: 10.7326/0003-4819-151-4-200908180-00135. [DOI] [PubMed] [Google Scholar]
- 23.Prober CG, Heath C. Lecture halls without lectures – a proposal for medical education. N Engl J Med. 2012;366(18):1657–1659. doi: 10.1056/NEJMp1202451. [DOI] [PubMed] [Google Scholar]
- 24.Abeysekera L, Dawson P. Motivation and cognitive load in the flipped classroom: definition, rationale and a call for research. High Educ Res Dev. 2015;34(1):1–14. doi: 10.1080/07294360.2014.934336. [DOI] [Google Scholar]
- 25.Williams DE. The future of medical education: flipping the classroom and education technology. Ochsner J. 2016;16(1):14–15. [PMC free article] [PubMed] [Google Scholar]
- 26.Korpershoek H, Harms T, de Boer H, van Kuijk M, Doolaard S. A meta-analysis of the effects of classroom management strategies and classroom management programs on students’ academic, behavioral, emotional, and motivational outcomes. Rev Educ Res. 2016;86(3):643–680. doi: 10.3102/0034654315626799. [DOI] [Google Scholar]
- 27.Freeman S, Eddy SL, McDonough M, Smith MK, Okoroafor N, Jordt H, Wenderoth MP. Active learning increases student performance in science, engineering, and mathematics. Proc Natl Acad Sci U S A. 2014;111(23):8410–8415. doi: 10.1073/pnas.1319030111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Borenstein M, Hedges LV, Higgins JPT, Rothstein HR. Introduction to meta-analysis. Chichester, UK: Wiley; 2009. [Google Scholar]
- 29.Altman DG, Bland JM. Standard deviations and standard errors. BMJ. 2005;331(7521):903. doi: 10.1136/bmj.331.7521.903. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Lipsey MW, Wilson DB. Practical meta-analysis. Thousand Oaks, CA: Sage; 2001. [Google Scholar]
- 31.Bakr MM, Massey WL, Massa HM. Flipping a dental anatomy course: a retrospective study over four years. Educ Res Int. 2016;2016:7097398. [Google Scholar]
- 32.Bossaer JB, Panus P, Stewart DW, Hagemeier NE, George J. Student performance in a pharmacotherapy oncology module before and after flipping the classroom. Am J Pharm Educ. 2016;80(2):31. doi: 10.5688/ajpe80231. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Cheng X, Lee KKH, Chang EY, Yang X. The “flipped classroom” approach: stimulating positive learning attitudes and improving mastery of histology among medical students. Anat Sci Educ. 2016; 10.1002/ase.1664. [DOI] [PubMed]
- 34.Galway LP, Corbett KK, Takaro TK, Tairyan K, Frank EA. Novel integration of online and flipped classroom instructional models in public health higher education. BMC Med Educ. 2014;14(1):181. doi: 10.1186/1472-6920-14-181. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Gillispie V. Using the flipped classroom to bridge the gap to generation Y. Ochsner J. 2016;16(1):32–36. [PMC free article] [PubMed] [Google Scholar]
- 36.Giuliano CA, Moser LR. Evaluation of a flipped drug literature evaluation course. Am J Pharm Educ. 2016;80(4):66. doi: 10.5688/ajpe80466. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Harrington SA, Bosch MV, Schoofs N, Beel-Bates C, Anderson K. Quantitative outcomes for nursing students in a flipped classroom. Nurs Educ Perspect. 2015;36(3):179–181. doi: 10.5480/13-1255. [DOI] [Google Scholar]
- 38.Hsu SD, Chen CJ, Chang WK, An HYJ. Investigation of the outcomes of PGY students’ cognition of and persistent behavior in learning through the intervention of the flipped classroom in Taiwan. PLoS One. 2016;11(12):e0167598. doi: 10.1371/journal.pone.0167598. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Kiviniemi MT. Effects of a blended learning approach on student outcomes in a graduate-level public health course. BMC Med Educ. 2014;14:181. doi: 10.1186/1472-6920-14-47. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Koo CL, Demps EL, Farris C, Bowman JD, Panahi L, Boyle P. Impact of flipped classroom design on student performance and perceptions in a pharmacotherapy course. Am J Pharm Educ. 2016;80(2):33. doi: 10.5688/ajpe80233. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Liebert CA, Lin DT, Mazer LM, Bereknyei S, Lau JN. Effectiveness of the surgery core clerkship flipped classroom: a prospective cohort trial. Am J Surg. 2016;211(2):451–457. doi: 10.1016/j.amjsurg.2015.10.004. [DOI] [PubMed] [Google Scholar]
- 42.McLaughlin JE, Griffin LM, Esserman DA, Davidson CA, Glatt DM, Roth MT, Gharkholonarehe N, Mumper RJ. Pharmacy student engagement, performance, and perception in a flipped satellite classroom. Am J Pharm Educ. 2013;77(9):196. doi: 10.5688/ajpe779196. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.McLaughlin JE, Roth MT, Glatt DM, Gharkholonarehe N, Davidson CA, Griffin LM, Esserman DA, Mumper RJ. The flipped classroom: a course redesign to foster learning and engagement in a health professions school. Acad Med. 2014;89(2):236–243. doi: 10.1097/ACM.0000000000000086. [DOI] [PubMed] [Google Scholar]
- 44.Morton DA, Colbert-Getz JM. Measuring the impact of the flipped anatomy classroom: the importance of categorizing an assessment by Bloom’s taxonomy. Anat Sci Educ. 2016;10(2):170–175. doi: 10.1002/ase.1635. [DOI] [PubMed] [Google Scholar]
- 45.Munson A, Pierce R. Flipping content to improve student examination performance in a pharmacogenomics course. Am J Pharm Educ. 2015;79(7):103. doi: 10.5688/ajpe797103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Porcaro PA, Jackson DE, McLaughlin PM, O’Malley CJ. Curriculum design of a flipped classroom to enhance haematology learning. J Sci Educ Technol. 2016;25(3):345–357. doi: 10.1007/s10956-015-9599-8. [DOI] [Google Scholar]
- 47.Prescott Jr WA, Woodruff A, Prescott GM, Albanese N, Bernhardi C, Doloresco F. Introduction and assessment of a blended-learning model to teach patient assessment in a doctor of pharmacy program. Am J Pharm Educ. 2016;80(10):176. doi: 10.5688/ajpe8010176. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Rui Z, Lian-rui X, Rong-zheng Y, Jing Z, Xue-hong W, Chuan Z. Friend or foe? Flipped classroom for undergraduate electrocardiogram learning: a randomized controlled study. BMC Med Educ. 2017;17:53. doi: 10.1186/s12909-017-0881-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Sajid MR, Laheji AF, Abothenain F, Salam Y, AlJayar D, Obeidat A. Can blended learning and the flipped classroom improve student learning and satisfaction in Saudi Arabia? Int J Med Educ. 2016;7:281–285. doi: 10.5116/ijme.57a7.83d4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Tune JD, Sturek M, Basile DP. Flipped classroom model improves graduate student performance in cardiovascular, respiratory, and renal physiology. AJP Adv Physiol Educ. 2013;37(4):316–320. doi: 10.1152/advan.00091.2013. [DOI] [PubMed] [Google Scholar]
- 51.Whillier S, Lystad RP. No differences in grades or level of satisfaction in a flipped classroom for neuroanatomy. J Chiropr Educ. 2015;29(2):127–133. doi: 10.7899/JCE-14-28. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Wong TH, Ip EJ, Lopes I, Rajagopalan V. Pharmacy students’ performance and perceptions in a flipped teaching pilot on cardiac arrhythmias. Am J Pharm Educ. 2014;78(10):185. doi: 10.5688/ajpe7810185. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Reed DA, Cook DA, Beckman TJ, Levine RB, Kern DE, Wright SM. Association between funding and quality of published medical education research. JAMA. 2007;298:1002–1009. doi: 10.1001/jama.298.9.1002. [DOI] [PubMed] [Google Scholar]
- 54.Hailikari T, Katajavuori N, Lindblom-Ylanne S. The relevance of prior knowledge in learning and instructional design. Am J Pharm Educ. 2008;72(5):113. doi: 10.5688/aj7205113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Merrill MD. First principles of instruction. Educ Technol Res Dev. 2002;50(3):43–59. doi: 10.1007/BF02505024. [DOI] [Google Scholar]
- 56.Dirkx KJH, Kester L, Kirschner PA. The testing effect for learning principles and procedures from texts. J Educ Res. 2014;107:357–364. doi: 10.1080/00220671.2013.823370. [DOI] [Google Scholar]
- 57.Murphy J, Chang JM, Suaray K. Student performance and attitudes in a collaborative and flipped linear algebra course. Int J Math Educ Sci Technol. 2016;47(5):653–673. doi: 10.1080/0020739X.2015.1102979. [DOI] [Google Scholar]
- 58.McGivney-Burelle J, Xue F. Flipping calculus. Primus. 2013;23(5):477–486. doi: 10.1080/10511970.2012.757571. [DOI] [Google Scholar]
- 59.Vazquez JJ, Chiang EP. Flipping out! A case study on how to flip the principles of economics classroom. Int Adv Econ Res. 2015;21:379–390. doi: 10.1007/s11294-015-9549-5. [DOI] [Google Scholar]
- 60.Chen CM, Wu CH. Effects of different video lecture types on sustained attention, emotion, cognitive load, and learning performance. Comput Educ. 2015;80:108–121. doi: 10.1016/j.compedu.2014.08.015. [DOI] [Google Scholar]
- 61.Chorianopoulos K, Giannakos MN. Proceedings of the 11th European conference on interactive TV and video. 2013. Usability design for video lectures; pp. 163–164. [Google Scholar]
- 62.Guo PJ, Kim J, Rubin R. Proceedings of the first ACM conference on learning@ scale conference. 2014. How video production affects student engagement: an empirical study of MOOC videos; pp. 41–50. [Google Scholar]
- 63.Peplow M. Social Sciences suffer from severe publication bias. Nature. Retrieved on August 25, 2017 from http://www.nature.com/news/social-sciences-suffer-from-severe-publication-bias-1.15787.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data used in this study can be found in previously published manuscripts.