Skip to main content
Academic Pathology logoLink to Academic Pathology
. 2021 Dec 13;8:23742895211061822. doi: 10.1177/23742895211061822

Outcomes of Remote Pathology Instruction in Student Performance and Course Evaluation

Tahyna Hernandez 1, Robert Fallar 2, Alexandros D Polydorides 1,
PMCID: PMC8679015  PMID: 34926798

Abstract

The coronavirus disease 2019 (COVID-19) pandemic has disrupted undergraduate medical education, including preclinical class-based courses, by requiring social distancing and essentially eliminating in-person teaching. The aim of this study was to compare student performance and satisfaction before and after implementation of remote instruction in a first-year introductory pathology course. Assessments (3 quizzes, 1 practical exam, and 1 final) were compared between courses given before (January 2020) and during (January 2021) the COVID-19 pandemic, in terms of mean scores, degree of difficulty, and item discrimination, both overall and across different question types. Students’ evaluations of the course (Likert scale-based) were also compared between the 2 years. Significantly higher mean scores were observed during remote instruction (compared to the prior, in-person year) on verbatim-repeated questions (94.9 ± 8.8 vs 89.4 ± 12.2; P = .002) and on questions incorporating a gross specimen image (88.4 ± 7.5 vs 84.4 ± 10.3; P = .007). The percentage of questions that were determined to be moderate/hard in degree of difficulty and good/very good in item discrimination remained similar between the 2 time periods. In the practical examination, students performed significantly better during remote instruction on questions without specimen images (96.5 ± 7.0 vs 91.2 ± 15.2; P = .004). Finally, course evaluation metrics improved, with students giving a higher mean rating value in each measured end point of course quality during the year of remote instruction. In conclusion, student performance and course satisfaction generally improved with remote instruction, suggesting that the changes implemented, and their consequences, should perhaps inform future curriculum improvements.

Keywords: medical education, COVID-19, assessment scores, online/virtual teaching, curriculum

Introduction

The coronavirus disease 2019 (COVID-19) pandemic has severely disrupted both health care and higher education throughout the world, and as a result, medical education, being at the nexus between the two, has been profoundly impacted. 1,2 Social distancing measures, complete lockdowns and the overwhelming burden on health care utilization due to the disease, have led to major disruptions in the content, conduct, and evaluation of undergraduate medical courses. 3 In order to ensure continuity in training, most academic medical centers in the United States made the strategic decision to transition to an almost completely remote learning model, including online instruction and assessment of preclinical courses. Thus, the pandemic has served as a catalyst in a paradigm shift that has significantly accelerated recent trends in medical teaching, such as reducing the number and size of large lectures, providing opportunities for individualized instruction, and shifting to electronic educational media. 4,5 However, the acute chronology of these changes has also brought to the forefront existing risks and difficulties in terms of mental health and the social and economic well-being of students, with likely long-term implications. 6,7

The discipline of anatomic pathology, with its emphasis on the histomorphologic appearance of cells and tissues and the schematic representation of pathophysiology pathways in human disease is perhaps uniquely positioned for a seamless transition to remote learning entirely based on digital images. 8,9 In fact, even before the pandemic, efforts had been well underway to implement updated technology modalities, such as livestreaming, social media, and virtual microscopy, in pathology training. 10 -12 Nevertheless, there are obstacles during the employment of these learning techniques, including, but not limited to, the need for appropriate hardware and compatible devices, and the requirement for adequate connectivity and storage capacity. 13,14 Besides technical issues, distance learning faces additional challenges in the realms of engagement/commitment, structure/organization, and communication/mental well-being. 15,16

To that end, data are needed that determine the efficacy and evaluate the outcomes of purposefully and completely online medical teaching programs, but these are generally lacking. The aim of this study was to measure the effect of remote instruction on students, both in terms of their capacity for learning as well as their perception of course quality. Given the context of reacting to COVID-19, we reviewed assessment performance and course satisfaction among undergraduate medical students remotely participating in a first-year introductory pathology course and compared them in detail to their classmate cohort from the immediately preceding year, taking the same course entirely in-person. Based on our previous item-level analysis, we specifically examined student performance among various types of assessment questions and also determined item discrimination and degree of difficulty for these question categories. 17 In addition, we evaluated student performance on a team-based practical examination which had been transformed from a hands-on activity to an online conference-room format. Finally, we compared the students’ post-course evaluation of its different components regarding learning objectives, instruction methods and materials, and content quality before and after the pandemic-induced transition in mode of instruction.

Materials and Methods

The study was approved by our Institutional Review Board (#18-00479) and conducted according to the World Medical Association Declaration of Helsinki. Deidentified scores were retrospectively collected for the academic year immediately before COVID-19 (2019-2020) and the year during the pandemic (2020-2021) from the web-based course management platform Blackboard Learn (Blackboard Inc.). Student attempts were recorded for all assessments and average responses (percent correct) were calculated for each question in each of the 2 events (years). Each question was characterized across a number of categorical variables as previously described. 17 Briefly, questions were categorized in terms of: answer type (multiple choice or matching), presence of clinical vignette (if so, whether simple or complex), presence of specimen image (gross or microscopic), depth of information (simple recall or interpretation), knowledge density (first or second order), Bloom level (1-3, in increasing complexity), and, for the final exam, subject familiarity (new or repeated subject and, if so, slightly modified or verbatim). Item difficulty was graded according to published guidelines: hard if <30%, moderate (desired range) if 30% to 80%, and easy if >80%. 18 Item discrimination was calculated based on the point biserial index and graded as follows: poor if below 0.2, fair if 0.2 to 0.29, good if 0.3 to 0.39, and very good if 0.4 to 0.7. 19 Student evaluations of various course parameters were obtained using a 5-level Likert scale. Mean student scores and rating values across the various question characteristics and evaluation parameters, respectively, were compared between the 2 time periods using unpaired t test with a two-tailed P value. The Bonferroni correction for multiple univariate comparisons was applied by dividing each P value with the number of hypotheses (ie, variables) tested. Percentage of questions in terms of degree of difficulty (hard and moderate together vs easy) and item discrimination (very good and good together vs fair and poor together) were compared using Pearson χ2 tests. All statistical analyses were carried out using Statistical Package for the Social Sciences software (SPSS; Build 1.0.0.1327; copyright 2019, IBM) with P < .05 considered significant.

Course Design

General Pathology is a 4-week introductory first-year medical school course at our institution, which normally consists of 32 hours of instruction including 17 hours of lectures and 15 hours of lab-based interaction. Lectures are created in Microsoft PowerPoint (Microsoft Corporation), presented to the entire class, recorded using Echo360 (Echo360 Inc), and are available on Blackboard for future access and archiving. Labs are organized in small groups of 36 students each. Each lab includes review of gross organ specimens under the guidance of pathology residents (the small group of 36 students is divided into 4 groups of 9 students each for this part of the lab). Subsequently, students participate in computer-based exercises in teams of 3, under the direction and with the support of pathology faculty instructors. These patient vignette-based lab exercises present clinical scenarios and digital images and ask the students to answer questions reinforcing concepts introduced in the preceding lecture(s). Students complete the exercise through real-time (wiki) collaboration within their teams on Microsoft Word-based documents, available through Google Docs software (Google LLC).

During the COVID-19 pandemic, instruction at the medical school switched from in-person for the academic year 2019-2020 to remote teaching in 2020-2021. In fact, this first-year class of medical school students had not been physically on campus and had not interacted in person with one another at all for the Fall semester. The General Pathology course also switched from completely in-person instruction in January 2020 to entirely remote teaching in January 2021. For that purpose, lectures were transformed from being physically given in a large auditorium to digitally delivered (same topic by the same instructor) over video conferencing using the screen-sharing tool on Zoom software (version 5.4.7; Zoom Video Communications, Inc) and were subsequently available as Blackboard-linked Echo360 video recordings. At the same time, lab exercises transitioned to virtual events also held via Zoom video conference. Residents displayed gross specimens using a high-definition desktop document camera (Lumens PS760; Lumens Digital Optics, Inc; Figure 1A). Students subsequently completed the same lab exercises within their teams, using the “breakout room” functionality on Zoom. Instructors and assisting residents “floated” between breakout rooms to answer questions and the entire group reconvened within the main Zoom meeting room at the end of each exercise for final summation, group discussion, and concept consolidation. This implementation and execution of remote instruction was delegated to individual course directors, with substantial administrative and logistic support from the medical school’s Office of Curricular Affairs and with the expressed involvement of the Office of Student Affairs, the latter regarding issues of student safety and overall well-being.

Figure 1.

Figure 1.

Remote instruction using Zoom video conferencing. A, Photograph of the desktop camera used to capture video of the specimen on the stage and project the feed onto a Zoom meeting with students. In the photograph background, there are trays with additional specimens to be shown. B, Screen shot of a question from the practical examination that required review of gross specimens. A video recording of the pathology resident (inset) presenting specimen findings was embedded within the assessment question. The students had the ability to expand the video to fill the entire screen or click a link for the video to open in a separate window.

Course assessments of individual students comprise a total of 125 questions which were identical between the 2 academic years 2019-2020 and 2021-2022. Questions are distributed among 3 weekly quizzes and 1 final exam (quiz one: 15 questions, quizzes two and three: 20 questions each, and final: 70 questions). All assessment examinations are closed-book and are administered individually, remotely, and online, with each student completing an honor code attestation. Some of these assessment questions include digital photographs of gross and/or microscopic specimens that are similar or identical to images previously included in lectures or lab exercises. However, neither these images included in assessment questions nor the images that students used to study from (ie, from lectures and lab exercises) were any different between the course taught in-person (2019-2020) and the course given remotely (2020-2021).

In addition, the last lab session of the course takes the form of a practical examination, whereby students answer 38 different questions collectively as a team-based exercise, in their usual 3-member lab group teams. These practical exam questions are based on unknown surgical and autopsy patient cases. In prior years, these cases were presented to the students in person and they had the opportunity to physically examine gross specimens, before attempting to answer assessment questions on the associated online quiz. During the pandemic, this entire practical assessment was moved online, with pathology residents using the camera set-up described above to present cases in video vignettes that were then embedded within the online quiz questions (Figure 1B).

Results

The number of students enrolled in the course during in-person instruction (January 2020) was 140 and during remote instruction (January 2021) was 139, for a total of 17,500 and 17,375 graded attempts, respectively, which were scored as correct or incorrect (Table 1). Overall, the mean percentage of correct answers was not significantly different (86.1 ± 13.3 vs 85.9 ± 16.5). However, when looking at questions with different characteristics, students had significantly higher mean scores during remote instruction (2021), compared to in-person teaching (2020) on questions repeated verbatim in the final exam from prior quizzes (94.9 ± 8.8 vs 89.4 ± 12.2; P = .002 with Bonferroni correction) and, surprisingly, on questions that contained a gross specimen image (88.4 ± 7.5 vs 84.4 ± 10.3; P = .007, corrected).

Table 1.

Student Performance in Assessment Questions During In-Person and Remote Instruction.*

Mean Score (%) ± SD
In-person (2020) Remote (2021) P value
(n = 140) (n = 139)
All questions (n = 125) 86.1 ± 13.3 85.9 ± 16.5 .91
Question type
 Multiple choice (n = 103) 84.9 ± 14.0 84.3 ± 17.5 .75
 Matching answer (n = 22) 92.1 ± 6.5 93.0 ± 7.0 .27
Clinical vignette
 Absent (n = 48) 87.1 ± 11.8 89.2 ± 10.7 .12
 Present (n = 77) 85.5 ± 14.2 83.8 ± 19.0 .40
  Simple (n = 72) 86.2 ± 12.6 84.2 ± 18.4 .29
  Complex (n = 5) 75.7 ± 30.1 77.3 ± 28.6 .65
Specimen image
 Absent (n = 94) 86.8 ± 12.4 85.9 ± 17.0 .61
 Present (n = 31) 84.2 ± 15.9 85.7 ± 15.2 .42
  Gross (n = 5) 84.4 ± 10.3 88.4 ± 7.5 .0003
  Microscopic (n = 22) 83.7 ± 18.1 85.3 ± 17.4 .45
  Both (n = 4) 86.6 ± 7.9 84.0 ± 9.5 .01
Information depth
 Simple recall (n = 120) 86.8 ± 12.1 86.5 ± 15.8 .86
 Interpretation (n = 5) 69.3 ± 27.7 70.2 ± 26.1 .78
Knowledge density
 First order (n = 76) 86.6 ± 12.5 85.7 ± 17.8 .63
 Second order (n = 49) 85.3 ± 14.5 86.1 ± 14.4 .64
Bloom taxonomy
 Level 1 (n = 76) 86.6 ± 12.5 85.7 ± 17.8 .63
 Level 2 (n = 44) 87.2 ± 11.4 87.9 ± 11.7 .61
 Level 3 (n = 5) 69.3 ± 27.7 70.2 ± 26.1 .78
Subject familiarity
 New information (n = 40) 84.2 ± 11.9 81.8 ± 19.4 .21
 Repeated subject (n = 30) 92.5 ± 7.7 90.9 ± 13.2 .22
  Similarity (n = 26) 92.9 ± 7.0 90.3 ± 13.8 .048
  Verbatim (n = 4) 89.4 ± 12.2 94.9 ± 8.8 .0001

Abbreviation: SD, standard deviation

* Statistically significant P values (ie, < .05) are bold. Applying the Bonferroni correction for 23 univariate tests leaves only presence of gross image and verbatim repetition of subject in the final as significantly different between the 2 sets (P values of .007 and .002, respectively).

As expected, given that the questions did not change year-over-year, the percentages of questions with moderate or high (hard) degrees of difficulty compared to those that were considered easy were similar between the 2 years, regardless of question type or characteristic (Table 2). Similarly, capacity for item discrimination (good/very good vs fair/poor) was analogously distributed during in-person and remote instruction, across all question subtypes (Table 3).

Table 2.

Degree of Difficulty in Assessment Questions During In-Person and Remote Instruction.*

Easy Moderate-Hard Easy Moderate-Hard
All questions 93 (74.4%) 32 (25.6%) 98 (78.4%) 27 (21.6%)
Question type
 Multiple choice 72 (69.9%) 31 (30.1%) 78 (75.7%) 25 (24.3%)
 Matching answer 21 (95.5%) 1 (4.5%) 20 (90.9%) 2 (9.1%)
Clinical vignette
 Absent 39 (81.3%) 9 (18.8%) 40 (83.3%) 8 (16.7%)
 Present 54 (70.1%) 23 (29.9%) 58 (75.3%) 19 (24.7%)
  Simple 51 (70.8%) 21 (29.2%) 55 (76.4%) 17 (23.6%)
  Complex 3 (60.0%) 2 (40.0%) 3 (60.0%) 2 (40.0%)
Specimen image
 Absent 71 (75.5%) 23 (24.5%) 74 (78.7%) 20 (21.3%)
 Present 22 (71.0%) 9 (29.0%) 24 (77.4%) 7 (22.6%)
  Gross 3 (60.0%) 2 (40.0%) 5 (100.0%) 0 (0%)
  Microscopic 16 (72.7%) 6 (27.3%) 17 (77.3%) 5 (22.7%)
  Both 3 (75.0%) 1 (25.0%) 2 (50.0%) 2 (50.0%)
Information depth
 Simple recall 91 (75.8%) 29 (24.2%) 97 (80.8%) 23 (19.2%)
 Interpretation 2 (40.0%) 3 (60.0%) 1 (20.0%) 4 (80.0%)
Knowledge density
 First order 58 (76.3%) 18 (23.7%) 62 (81.6%) 14 (18.4%)
 Second order 35 (71.4%) 14 (28.6%) 36 (73.5%) 13 (26.5%)
Bloom taxonomy
 Level 1 58 (76.3%) 18 (23.7%) 62 (81.6%) 14 (18.4%)
 Level 2 33 (75.0%) 11 (25.0%) 35 (79.5%) 9 (20.5%)
 Level 3 2 (40.0%) 3 (60.0%) 1 (20.0%) 4 (80.0%)
Subject familiarity
 New information 28 (70.0%) 12 (30.0%) 29 (72.5%) 11 (27.5%)
 Repeated subject 26 (86.7%) 4 (13.3%) 28 (93.3%) 2 (6.7%)
  Similarity 23 (88.5%) 3 (11.5%) 24 (92.3%) 2 (7.7%)
  Verbatim 3 (75.0%) 1 (25.0%) 4 (100.0%) 0 (0%)

* There are no significant differences between the 2 years (ie, all are P > .05).

Table 3.

Item Discrimination of Assessment Questions During In-Person and Remote Instruction.*

In-person (2020) Remote (2021)
Good-Very Good Fair-Poor Good-Very Good Fair-Poor
All questions 40 (32.0%) 85 (68.0%) 44 (35.2%) 81 (64.8%)
Question type
 Multiple choice 27 (26.2%) 76 (73.8%) 32 (31.1%) 71 (68.9)
 Matching 13 (59.1%) 9 (40.9%) 12 (54.6%) 10 (45.4%)
Clinical vignette
 Absent 16 (33.3%) 32 (66.7%) 12 (25.0%) 36 (75.0%)
 Present 24 (31.2%) 53 (68.9%) 32 (41.6%) 45 (58.4%)
  Simple 22 (30.5%) 50 (69.5%) 30 (41.6%) 42 (58.4%)
  Complex 2 (40.0%) 3 (60.0%) 2 (40.0%) 3 (60.0%)
Specimen image
 Absent 26 (27.6%) 68 (72.3%) 32 (34.0%) 62 (66.0%)
 Present 14 (45.2%) 17 (54.8%) 12 (38.7%) 19 (61.3%)
  Gross 3 (60.0%) 2 (40.0%) 2 (40.0%) 3 (60.0%)
  Microscopic 9 (40.9%) 13 (59.1%) 8 (36.4%) 14 (63.6%)
  Both 2 (50.0%) 2 (50.0%) 2 (50.0%) 2 (50.0%)
Information depth
 Simple recall 37 (30.8%) 83 (69.2%) 42 (35.0%) 78 (65.0%)
 Interpretation 3 (60.0%) 2 (40.0%) 2 (40.0%) 3 (60.0%)
Knowledge density
 First order 21 (27.7%) 55 (72.4%) 23 (30.3%) 53 (69.7%)
 Second order 19 (38.7%) 30 (61.2%) 21 (42.9%) 28 (57.1%)
Bloom taxonomy
 Level 1 21 (27.7%) 55 (72.4%) 23 (30.3%) 53 (69.7%)
 Level 2 16 (36.4%) 28 (63.6%) 19 (43.2%) 25 (56.8%)
 Level 3 3 (60.0%) 2 (40.0%) 2 (40.0%) 3 (60.0%)
Subject familiarity
 New information 10 (25.0%) 30 (75.0%) 11 (27.5%) 29 (72.5%)
 Repeated subject 6 (20.0%) 24 (80.0%) 7 (23.3%) 23 (76.7%)
  Similarity 5 (19.2%) 21 (80.8%) 7 (26.9%) 19 (73.1%)
  Verbatim 1 (25.0%) 3 (75.0%) 0 (0%) 4 (100%)

* There are no significant differences between the 2 years (ie, all are P > .05).

In terms of the 38 different questions in the separate practical examination, students performed better during the year of remote instruction on matching type questions, those with simple clinical vignettes, questions testing first order knowledge, and questions assigned a Bloom taxonomy level one, and performed significantly worse on questions with an image of a gross specimen, compared to students during the prior year with in-person teaching (Table 4). However, these differences did not survive the Bonferroni correction for multiple univariate comparisons. In contrast, the improved performance of students during COVID-19 on questions lacking a specimen image with a mean score of 96.5 ± 7.0 (compared to 91.2 ± 15.2 the year before) remained statistically significant (P = .004, with Bonferroni correction).

Table 4.

Student Performance in the Practical Exam During In-Person and Remote Instruction.*

Mean score (%) ± SD
In-person (2020) Remote (2021) P value
(n = 140) (n = 139)
All questions (n = 38) 90.7 ± 15.2 91.8 ± 10.2 .48
Question type
 Multiple choice (n = 37) 90.6 ± 15.2 91.7 ± 10.3 .48
 Matching (n = 1) 93.9 ± 15.7 97.1 ± 7.1 .03
Clinical vignette
 Absent (n = 8) 99.0 ± 3.8 99.6 ± 0.9 .07
 Present (n = 30) 88.4 ± 18.3 89.8 ± 12.7 .46
 Simple (n = 15) 93.0 ± 13.5 95.7 ± 7.4 .04
 Complex (n = 15) 83.9 ± 23.1 83.8 ± 18.0 .97
Specimen image
 Absent (n = 13) 91.2 ± 15.2 96.5 ± 7.0 .0002
 Present (n = 25) 90.4 ± 15.3 89.4 ± 11.9 .54
  Gross (n = 9) 81.9 ± 21.4 75.4 ± 25.3 .02
  Microscopic (n = 3) 81.9 ± 31.1 86.9 ± 16.6 .10
  Both (n = 13) 98.2 ± 7.3 99.7 ± 1.5 .21
Information depth
 Simple recall (n = 36) 91.4 ± 14.3 92.5 ± 9.4 .45
 Interpretation (n = 2) 77.2 ± 32.4 80.4 ± 24.9 .36
Knowledge density
 First order (n = 14) 97.6 ± 9.2 99.5 ± 1.9 .02
 Second order (n = 24) 86.6 ± 18.8 87.4 ± 15.0 .77
Bloom taxonomy
 Level 1 (n = 14) 97.6 ± 9.2 99.5 ± 1.9 .02
 Level 2 (n = 22) 87.5 ± 17.5 88.0 ± 14.2 .79
 Level 3 (n = 2) 77.2 ± 32.4 80.4 ± 24.9 .36

Abbreviation: SD, standard deviation.

* Statistically significant P values (ie, < .05) are in bold. Applying the Bonferroni correction for 19 univariate tests leaves only the absence of specimen image as significantly different between the 2 sets (P value of .004).

Finally, students’ responses during the post-course evaluation were recorded and compared between the course given remotely and the course with in-person instruction the prior year. Each year, approximately half the students in the class are randomly asked to anonymously and confidentially provide ratings based on a 5-level Likert scale (from 1-strongly disagree/poor to 5-strongly agree/excellent) on the quality of various course parameters, including organization, learning objectives, content, teaching, and instruction methods (Table 5). There was improvement in the mean rating values assigned by students during remote teaching (compared to those during in-person instruction) in every single parameter measured, however none reached statistical significance.

Table 5.

Comparison of Students’ Course Evaluations Following In-Person and Remote Instruction.*

Mean rating value (#) ± SD
In-person (2020) Remote (2021)
(n = 67) (n = 64)
Overall evaluation
 Course quality 4.69 ± 0.61 4.70 ± 0.55
 Faculty teaching 4.67 ± 0.59 4.80 ± 0.48
 Content organization 4.63 ± 0.69 4.73 ± 0.51
Learning objectives§
 Clearly stated 4.65 ± 0.57 4.75 ± 0.47
 Followed in content 4.65 ± 0.57 4.76 ± 0.43
Content and assessments§
 Illustrations of clinical relevance 4.75 ± 0.47 4.83 ± 0.38
 Manageable workload 4.70 ± 0.49 4.73 ± 0.54
 Instruction methods facilitated  learning 4.58 ± 0.65 4.72 ± 0.52
 Fair assessment of concepts 4.69 ± 0.58 4.75 ± 0.44

Abbreviation: SD, standard deviation.

*P values for univariate comparison across each row were not statistically significant (ie, > .05).

 Rated as follows: 1 = poor, 2 = fair, 3 = good, 4 = very good, 5 = excellent.

§Rated as follows: 1 = strongly disagree, 2 = disagree, 3 = neutral, 4 = agree, 5 = strongly agree.

Discussion

The results of this study show that transition to remote instruction is not associated with a degradation in student performance and in fact may be accompanied by improvement, particularly on assessment questions with certain characteristics, such as those being repeated from previous exams or those that contain an image of a gross specimen and, in the practical examination, those that lacked a specimen image. However, assessment questions retained their degree of difficulty and capacity for item discrimination overall, compared to the prior year with in-person teaching. Importantly, student satisfaction with course quality improved over the previous year, as evidenced in evaluation ratings.

The rapid and comprehensive disruption caused by the COVID-19 pandemic compelled academic institutions worldwide to reevaluate and adapt their organizational models. In medical education, and specifically pathology teaching, a varying array of solutions were implemented, from mixed models incorporating flipped classrooms with limited, unidirectional, and timed in-person specimen examination to fully virtual laboratory classes using online video-conferencing tools. 20 -22 These changes reflected similar adjustments imposed on pathology resident training and upper-class student electives and mostly concerned the replacement of instruction centered on a physical microscope with the use of virtually accessed digitized slides. 23 -27 Among individual medical schools, the decision of how to modify instruction depended on factors such as content complexity, technical feasibility, time limitations, and infrastructure capacity. 28 In our institution, the decision involved an immediate transition to completely remote instruction for all preclinical courses and was partly influenced by the fact that our school’s location was at the time an epicenter of the pandemic.

This precipitous and near universal adoption of remote learning modalities also necessitates a framework for their evaluation, both for quality assurance as well as for the development of standards and best practice indicators as they have been identified in the areas of organizational capacity, educational effectiveness, and human resources. 29 We were especially interested in examining how online instructional methods affected student involvement in and understanding of the learning material as well as their perception of course quality. To measure these effects, we examined students’ performance on the course learning assessments and their subsequent evaluation of course content and delivery. While it may be difficult to compare student performance across different years, the large number of scored attempts and the fact that course content and assessments have remained relatively constant in recent years hopefully ameliorated some of the cyclical variation. In addition, we had previously observed that year-over-year divergence has been minimal in this cohort. 17 Finally, with the course being in the middle of an academic year with completely remote instruction, the students were used to the video conferencing format and thus fairly capable in its usage and comfortable with its features.

Student assessment scores in this cohort were remarkably similar to prior years, with the transition to remote instruction leading to significant increases in the mean percentage of correct answers in only 2 types of questions, those with a gross specimen image and those repeated verbatim in the final exam from prior quizzes. We hypothesize that improved performance in both of these question categories reveals an effort on the part of the students to concentrate more and invest time in their studying. Having been warned that images would be included in the assessments and that question topics may be repeated in the final exam, they may have focused their individual learning on preparing for these exact eventualities, with the resultant better scores. The partial lockdown still in effect at the time probably minimized extraneous distractions. Supporting this explanation, a recent study of a medical histopathology course found that, while student scores on team quizzes during the lab sessions declined significantly during COVID-19 compared to prior years, their individual performance on weekly quizzes and module examinations (more involved assessments occurring at a later time) was unaffected. 21 Other authors have also found increased medical student academic performance with the transition to e-learning necessitated by COVID-19, however they were not based in the United States and examined overall test scores between semesters without a detailed breakdown by course and question type. 30 Furthermore, degree of difficulty and item discrimination for these questions in our course did not differ from prior years, implying that educational goals and program objectives, vis-à-vis the function of student assessments, were still being met. 17

At the same time, the lack of in-person instruction may have been an incentive for students to specifically emphasize learning themes that may have been affected by remote instruction, such as gross pathologic appearance of disease processes, hence improving their understanding and scores on such topics. Consistent with this hypothesis, a large cross-sectional survey in the United Kingdom found a significant increase in the amount of time spent by medical students on electronic learning platforms, something that has been echoed by others. 31,32 We posit that the lab practical, being an exercise given in real time and for which students didn’t really prepare (since it was supposed to be open-book and taken collectively in teams of 3), did not pose as much of a challenge or an impetus for them to study gross images ahead of time, thus leading to lower performance in this assessment (Table 4). In addition, perhaps because of this lower performance on questions with gross images in the lab practical, students were compelled to prepare and study these types of images for the upcoming final exam, leading to a higher score in such questions during the final, summative assessment (Table 1).

An intriguing alternative explanation is that due to the remote set-up used in this course, pathology residents were displaying via camera only one gross specimen at a time (as opposed to students congregating around a table with all the organs of the day’s session simultaneously exposed). This may have allowed students to concentrate on one specimen at a time and follow the resident’s guidance in describing its physical characteristics (color, size, shape, texture, consistency, etc). On the other hand, students performed significantly better in the practical examination during COVID-19, but only on questions without a specimen image, suggesting that there are still barriers to remote instruction in terms of adequately portraying perhaps more subtle findings on gross organs.

We also found that students were overall very pleased with the course delivered remotely, which scored higher on their evaluations compared to the previous year. Even though it is a different set of students responding to these evaluation surveys and we did not specifically address remote instruction and online learning, general trends can be surmised from their narrative answers that accompanied these questionnaires (data not shown). Students appreciated the interactive, albeit virtual format of the course: they responded positively to the lab setups with team-based learning and the opportunity to interact with residents in small groups. They also welcomed the chance to work in groups privately as afforded to them in the breakout rooms, before reconvening in the main online meeting room for group discussion. From their comments, it was evident that students valued the organized, targeted, and effective teaching provided, and rated the course very highly despite the disruption caused by COVID-19. As is also being reported by others, while online teaching is favorably received, most students report that they still prefer the physical interaction of in-person instruction. 22,25,31,33 -36

Generally, reported strengths of remote learning include increased participation, adaptability, and flexibility of both content and schedule, organized and personalized delivery geared toward individualized goals, and a positive education environment that is useful and constructive. 28,31,35 Nevertheless, significant challenges to remote learning remain, including infrastructure and technical issues such as internet connectivity (speed, bandwidth, interruptions in service) and physical space requirements that may lead to confidentiality issues and family or other distractions. 28,31 For example, students have reported that online educational tools tend to decrease peer-to-peer teaching, obstruct the learning experience, and increase anxiety, outcomes that may disproportionally affect some groups of learners more than others. 21,37 In addition, the teaching of practical skills, which are essential in medical education, particularly in the clinical years, may not be as amenable to remote instruction. 38 Similarly, in the preclinical years, remote instruction faces an almost insurmountable deficit in teaching dissection and grossing. 35 Finally, impaired or reduced social connections during virtual interactions may lead to diminished observation of students’ performance in these skills and hinder proper and adequate feedback, particularly in terms of narrative assessment. 21,28,35 Going forward, with the pandemic’s grip easing, it would be important to observe and evaluate student performance and attitudes as we return to in-person instruction and this study would form the basis for continued analysis.

In conclusion, while the necessary transition to completely remote instruction has mostly proven to be a success, in terms of outcomes such as students’ assessment performance and course satisfaction, the long-term effects of this paradigm shift in medical (and all) education will need to keep being monitored and studied. Crucially, student satisfaction with course quality improved over the previous year as evidenced in evaluation ratings, suggesting that the impact of COVID-19 may not manifest itself in ways previously feared, but also instructing future attempts in curriculum development. It is difficult to imagine a future where online teaching is not a major component of medical learning but the well-being of both students and instructors, particularly in these interactions with reduced physical contact and less structured frameworks will need to remain in the forefront. 31,39,40 This is undoubtedly an opportunity for additional innovation in the delivery and assessment of medical education and, in this context, perhaps a hybrid model will be best suited to combine the efficacy and usefulness of remote instruction with the benefits and community of in-person engagement.

Acknowledgments

Authors would like to thank Megan Sacco, Kira Schmidt, and Jeff Cavalli for administrative and technical support during the transition to remote instruction.

Footnotes

Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Funding: The author(s) received no financial support for the research, authorship, and/or publication of this article.

ORCID iD: Alexandros D. Polydorides Inline graphic https://orcid.org/0000-0001-8604-0756

References

  • 1. Dedeilia A, Sotiropoulos MG, Hanrahan JG, Janga D, Dedeilias P, Sideris M. Medical and surgical education challenges and innovations in the COVID-19 era: a systematic review. In Vivo. 2020;34:1603–1611. doi:10.21873/invivo.11950 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Sahi PK, Mishra D, Singh T. Medical education amid the COVID-19 pandemic. Indian Pediatr. 2020;57:652–657. doi:10.1007/s13312-020-1894-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Franchi T. The impact of the Covid-19 pandemic on current anatomy education and future careers: a student’s perspective. Anat Sci Educ. 2020;13:312–315. doi:10.1002/ase.1966 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Alsafi Z, Rajabali H. Medical education in the midst of the coronavirus (COVID-19) pandemic-one year later. Int J Surg. 2021;106000. doi:10.1016/j.ijsu.2021.106000 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Rose S. Medical student education in the time of COVID-19. JAMA. 2020;323:2131–2132. doi:10.1001/jama.2020.5227 [DOI] [PubMed] [Google Scholar]
  • 6. Gill D, Whitehead C, Wondimagegn D. Challenges to medical education at a time of physical distancing. Lancet. 2020;396:77–79. doi:10.1016/S0140-6736(20)31368-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Sharma D, Bhaskar S. Addressing the Covid-19 burden on medical education and training: the role of telemedicine and tele-education during and beyond the pandemic. Front Public Health. 2020;8:589669. doi:10.3389/fpubh.2020.589669 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Koch LK, Chang OH, Dintzis SM. Medical education in pathology: general concepts and strategies for implementation. Arch Pathol Lab Med. 2021;145:1081–1088. doi:10.5858/arpa.2020-0463-RA [DOI] [PubMed] [Google Scholar]
  • 9. Mukhopadhyay S, Booth AL, Calkins SM, et al. Leveraging technology for remote learning in the era of COVID-19 and social distancing. Arch Pathol Lab Med. 2020;144:1027–1036. doi:10.5858/arpa.2020-0201-ED [DOI] [PubMed] [Google Scholar]
  • 10. Cima L, Mannan R, Madrigal E, Barbareschi M. Towards a “Net” generation of pathologists: the pathCast online remote learning platform. Pathologica. 2020;112:160–171. doi:10.32074/1591-951X-210 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Madrigal E, Mannan R. pathCast: an interactive medical education curriculum that leverages livestreaming on Facebook and YouTube. Acad Med. 2020;95:744–750. doi:10.1097/ACM.0000000000003148 [DOI] [PubMed] [Google Scholar]
  • 12. Master SR, Neide B, Kollu S, Mulchandani S, Harris RM, Obstfeld AE. Facilitating test status information communication in the time of COVID-19. Acad Pathol. 2021;8:23742895211006822. doi:10.1177/23742895211006822 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Hojilla C, Armstrong S, Pun C, et al. A holistic approach to pathology education during the coronavirus disease 2019 (COVID-19) pandemic. Arch Pathol Lab Med. 2021;145:923–924. doi:10.5858/arpa.2020-0668-LE [DOI] [PubMed] [Google Scholar]
  • 14. Zalat MM, Hamed MS, Bolbol SA. The experiences, challenges, and acceptance of e-learning as a tool for teaching during the COVID-19 pandemic among university medical staff. PLoS One. 2021;16:e0248758. doi:10.1371/journal.pone.0248758 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Kwon R, Zhang ML, VandenBussche CJ. Considerations for remote learning in pathology during COVID-19 social distancing. Cancer Cytopathol. 2020;128:642–647. doi:10.1002/cncy.22289 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Reinhart A, Malzkorn B, Doing C, Beyer I, Junger J, Bosse HM. Undergraduate medical education amid COVID-19: a qualitative analysis of enablers and barriers to acquiring competencies in distant learning using focus groups. Med Educ Online. 2021;26:1940765. doi:10.1080/10872981.2021.1940765 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Hernandez T, Magid MS, Polydorides AD. Assessment question characteristics predict medical student performance in general pathology. Arch Pathol Lab Med. 2021;145:1280-1288. doi:10.5858/arpa.2020-0624-OA [DOI] [PubMed] [Google Scholar]
  • 18. McDonald ME. Interpreting test results. In: The Nurse Educator’s Guide to Assessing Learning Outcomes. 4th ed. Jones & Bartlett Learning; 2018. Chapter 11. [Google Scholar]
  • 19. McGahee TW, Ball J. How to read and really use an item analysis. Nurse Educ. 2009;34:166–171. doi:10.1097/NNE.0b013e3181aaba94 [DOI] [PubMed] [Google Scholar]
  • 20. Bryant RJ, Wilcox R, Zhang B. Engage in exploration: pathology gross laboratory in the COVID-era. Acad Pathol. 2021;8:23742895211002843. doi:10.1177/23742895211002843 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Sharma R, King TS, Hanson ER, Fiebelkorn K. Medical histopathology laboratories: remote teaching in response to COVID-19 pandemic. Acad Pathol. 2021:8. doi:10.1177/2374289521998049 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Singh K, Srivastav S, Bhardwaj A, Dixit A, Misra S. Medical education during the COVID-19 pandemic: a single institution experience. Indian Pediatr. 2020;57:678–679. doi:10.1007/s13312-020-1899-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Chin KE, Kwon D, Gan Q, et al. Transition from a standard to a hybrid on-site and remote anatomic pathology training model during the coronavirus disease 2019 (COVID-19) pandemic. Arch Pathol Lab Med. 2021;145:22–31. doi:10.5858/arpa.2020-0467-SA [DOI] [PubMed] [Google Scholar]
  • 24. Fu L, Swete M, Selgrade D, et al. Virtual pathology elective provides uninterrupted medical education and impactful pathology education during the COVID-19 pandemic. Acad Pathol. 2021;8:23742895211010275. doi:10.1177/23742895211010275 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Guiter GE, Sapia S, Wright AI, Hutchins GGA, Arayssi T. Development of a remote online collaborative medical school pathology curriculum with clinical correlations, across several international sites, through the Covid-19 pandemic. Med Sci Educ. 2021;31:1–8. doi:10.1007/s40670-021-01212-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Parker EU, Chang O, Koch L. Remote anatomic pathology medical student education in Washington State. Am J Clin Pathol. 2020;154:585–591. doi:10.1093/ajcp/aqaa154 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. White MJ, Birkness JE, Salimian KJ, et al. Continuing undergraduate pathology medical education in the coronavirus disease 2019 (COVID-19) global pandemic: the Johns Hopkins virtual surgical pathology clinical elective. Arch Pathol Lab Med. 2021;145:814–820. doi:10.5858/arpa.2020-0652-SA [DOI] [PubMed] [Google Scholar]
  • 28. Hayat AA, Keshavarzi MH, Zare S, et al. Challenges and opportunities from the COVID-19 pandemic in medical education: a qualitative study. BMC Med Educ. 2021;21:247. doi:10.1186/s12909-021-02682-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Wasfy NF, Abouzeid E, Nasser AA, et al. A guide for evaluation of online learning in medical education: a qualitative reflective analysis. BMC Med Educ. 2021;21:339. doi:10.1186/s12909-021-02752-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Tashkandi E. E-Learning for undergraduate medical students. Adv Med Educ Pract. 2021;12:665–674. doi:10.2147/AMEP.S314509 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. Dost S, Hossain A, Shehab M, Abdelwahed A, Al-Nusair L. Perceptions of medical students towards online teaching during the COVID-19 pandemic: a national cross-sectional survey of 2721 UK medical students. BMJ Open. 2020;10:e042378. doi:10.1136/bmjopen-2020-042378 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Radhamani R, Kumar D, Nizar N, Achuthan K, Nair B, Diwakar S. What virtual laboratory usage tells us about laboratory skill education pre- and post-COVID-19: Focus on usage, behavior, intention and adoption. Educ Inf Technol (Dordr). 2021:1–19. doi:10.1007/s10639-021-10583-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Franklin G, Martin C, Ruszaj M, et al. How the COVID-19 pandemic impacted medical education during the last year of medical school: a class survey. Life (Basel). 2021;11:294. doi:10.3390/life11040294 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Villanueva EW, Meissner H, Walters RW. Medical student perceptions of the learning environment, quality of life, and the school of medicine’s response to the COVID-19 pandemic: a single institution perspective. Med Sci Educ. 2021;31:1–10. doi:10.1007/s40670-021-01223-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Tanaka KS, Ramachandran R. Perceptions of a remote learning pathology elective for advanced clinical medical students. Acad Pathol. 2021;8:23742895211006846. doi:10.1177/23742895211006846 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Vatier C, Carrie A, Renaud MC, Simon-Tillaux N, Hertig A, Jeru I. Lessons from the impact of COVID-19 on medical educational continuity and practices. Adv Physiol Educ. 2021;45:390–398. doi: 10.1152/advan.00243.2020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Wilcha RJ. Effectiveness of virtual medical teaching during the COVID-19 crisis: systematic review. JMIR Med Educ. 2020;6:e20963. doi:10.2196/20963 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Enoch TR, Williams RC. Why face-to-face medical education will prevail despite the world’s swift acclimatisation to virtual learning [published online June 30, 2021]. Postgrad Med J. 2021. doi:10.1136/postgradmedj-2021-140426 [DOI] [PubMed] [Google Scholar]
  • 39. Leong JMC, Lam WL, Tan SZ, Ng CY. Changing face of medical education during a pandemic: tragedy or opportunity [published online June 30, 2021]? Postgrad Med J. 2021. doi:10.1136/postgradmedj-2021-140330 [DOI] [PubMed] [Google Scholar]
  • 40. Gaur U, Majumder MAA, Sa B, Sarkar S, Williams A, Singh K. Challenges and opportunities of preclinical medical education: COVID-19 crisis and beyond. SN Compr Clin Med. 2020:1–6. doi:10.1007/s42399-020-00528-1 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Academic Pathology are provided here courtesy of Elsevier

RESOURCES