Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2022 Jan 27;17(1):e0263182. doi: 10.1371/journal.pone.0263182

Predicting the effectiveness of the online clinical clerkship curriculum: Development of a multivariate prediction model and validation study

Naoto Kuroda 1,2,*, Anna Suzuki 3, Kai Ozawa 4, Nobuhiro Nagai 5, Yurika Okuyama 6, Kana Koshiishi 7, Masafumi Yamada 5, Makoto Kikukawa 8
Editor: Feroze Kaliyadan9
PMCID: PMC8794117  PMID: 35085367

Abstract

Given scientific and technological advancements, expectations of online medical education are increasing. However, there is no way to predict the effectiveness of online clinical clerkship curricula. To develop a prediction model, we conducted cross-sectional national surveys in Japan. Social media surveys were conducted among medical students in Japan during the periods May–June 2020 and February–March 2021. We used the former for the derivation dataset and the latter for the validation dataset. We asked students questions in three areas: 1) opportunities to learn from each educational approach (lectures, medical quizzes, assignments, oral presentations, observation of physicians’ practice, clinical skills practice, participation in interprofessional meetings, and interactive discussions with physicians) in online clinical clerkships compared to face-to-face, 2) frequency of technical problems on online platforms, and 3) satisfaction and motivation as outcome measurements. We developed a scoring system based on a multivariate prediction model for satisfaction and motivation in a cross-sectional study of 1,671 medical students during the period May–June 2020. We externally validated this scoring with a cross-sectional study of 106 medical students during February–March 2021 and assessed its predictive performance. The final prediction models in the derivation dataset included eight variables (frequency of lectures, medical quizzes, oral presentations, observation of physicians’ practice, clinical skills practice, participation in interprofessional meetings, interactive discussions with physicians, and technical problems). We applied the prediction models created using the derivation dataset to a validation dataset. The prediction performance values, based on the area under the receiver operating characteristic curve, were 0.69 for satisfaction (sensitivity, 0.50; specificity, 0.89) and 0.75 for motivation (sensitivity, 0.71; specificity, 0.85). We developed a prediction model for the effectiveness of the online clinical clerkship curriculum, based on students’ satisfaction and motivation. Our model will accurately predict and improve the online clinical clerkship curriculum effectiveness.

Introduction

Given scientific and technological advancements, the medical education community’s expectations of online education are rising [13]. In addition, coronavirus disease 2019 (COVID-19) has necessitated the use of the online platform in medical education [48]. Globally, online clinical clerkship was proposed to replace face-to-face clinical clerkship during the COVID-19 pandemic (mid-COVID-19) [9]. Various opinions and practical innovations have been reported in relation to making online clinical clerkship a viable alternative to face-to-face clerkship [1012]. However, there are currently no scores that can predict the effectiveness of online clinical clerkship curricula. Developing such predictive scores would make online clinical clerkship more effective.

We aimed to develop and externally validate a prognostic score that represents students’ satisfaction and motivation, and can predict the effectiveness of the online clinical clerkship curriculum.

Material and methods

Survey design and data collection

This study followed the principles of the Declaration of Helsinki and was approved by Kyushu University’s ethics committee. To achieve our objectives, we conducted cross-sectional surveys. We developed the survey questions with a focus on content validity through an iterative process within the team, which included a medical educator. In addition, we administered the survey to a pilot group of 30 medical students. During a semi-structured interview after the survey was administered, these students were asked to provide feedback on the clarity of formulation of the items and on the format of the questionnaire.

The study consisted of two parts. First, we collected the survey responses for the derivation dataset. Through derivation analysis, we developed a scoring model for the online curriculum to predict medical students’ satisfaction and motivation. We then collected the survey responses for the validation dataset. Developed using the derivation dataset, we assessed our predictive model’s prediction performance via validation analysis.

The study participants were undergraduate medical school students in Japan. The survey that yielded the derivation dataset was conducted from May 29, 2020 to June 14, 2020. This period was chosen because the first declaration of a state of emergency due to COVID-19 in Japan (April 7–May 25, 2020) led to the widespread use of online clinical clerkship. We conducted the survey that yielded the validation dataset from February 26 to March 11, 2021. This period was chosen because the second declaration of a state of emergency due to COVID-19 in Japan (January 7–March 21, 2021) also led to the widespread use of online clinical clerkship.

Both surveys have identical content. The questionnaire was presented along with an informative letter summarizing the research purpose, as well as the informed consent form, written briefly and clearly in Japanese to avoid misinterpretation. The questionnaire took 3–5 min to complete.

The inclusion criteria were as follows: students (1) who switched from face-to-face to online clinical clerkship mid-COVID-19 and (2) completed all the survey questions. We included fifth- and sixth-year medical students in the derivation dataset and fourth- and fifth-year students in the validation study. This is because in Japan, medical students graduate in March, so sixth-year students often do not participate in clinical clerkships between January and March. The Japanese undergraduate medical education curriculum is six years period. It usually consists of four/three and half years of preclinical education and two/two and half years of clinical clerkship. Clinical clerkship has medical students participate as members of a medical team in a variety of clinical settings. Medical students can acquire clinical competencies through clinical clerkship.

Survey items

The questionnaire asked respondents to share information about their demographic information. If participants were medical students who switched from face-to-face to online clinical clerkship, we inquired as to the effectiveness of online compared to face-to-face clinical clerkship and, how frequently they had various experiences while completing their online clinical clerkship.

Exposure factors and other factors

We examined the extent of students’ exposure to factors associated with the effectiveness of online clinical clerkship. We decided which factors should be included in our survey based on previous literature and recommendations for medical education. The factors incorporated into this study included lecture duration [1315] and lecture frequency per week [13, 16, 17]; opportunities to take quizzes [18, 19], submit assignments [20, 21], give oral presentations [22, 23], observe physicians’ practice [10, 24], practice clinical skills [10, 24], participate in interprofessional meetings [2426], and interact with physicians [24, 27]; and the frequency of technical Internet-related problems [28]. For details regarding the options for each answer, see S1 Method

Measuring outcomes

Based on Kirkpatrick’s assessment model for assessing the usefulness of medical education and medical curricula [26, 2931], we used two outcome measures of the effectiveness of online clinical clerkship: satisfaction level (Level 1 in Kirkpatrick’s assessment model) and motivation level (Level 2a in Kirkpatrick’s assessment model). In questions about satisfaction and motivation, the answers were based on a 5-point Likert scale, ranging from 1 = the level/amount was much higher during face-to-face clerkship than it was during online clinical clerkship and 5 = the level was much higher during online clinical clerkship than it was during face-to-face clerkship.

Total number of respondents and the number of respondents included in the study (Fig 1)

Fig 1. A flowchart showing the selection of suitable participants for inclusion in this study.

Fig 1

To compile the derivation dataset, we collected responses from 2,640 Japanese medical students. Among the 2,640 respondents, 2,594 (98.3%) consented to participate in this study. Respondents with answers missing from their profile were excluded from the analysis (n = 127). In Japan, fifth- and sixth-year students are the equivalent years of practice, so we also excluded responses from students in years below the fifth year (n = 19). The next step was to categorize the responses according to the type of education the students received mid-COVID-19. Sixteen respondents were excluded for inappropriate responses. Of the 2,432 remaining respondents, 1,758 reported that they had experienced a face-to-face clerkship in a hospital pre-COVID-19 and then switched to an online clinical clerkship due to the COVID-19 crisis. Of the 1,758 respondents, those with missing responses regarding the content of the education to which they were exposed were excluded (n = 87). Finally, 1,671 responses were used for derivation analysis in this study.

For the derivation dataset, we collected responses from 2,640 Japanese medical students. Among the 2,640 respondents, 2,594 (98.3%) consented to participate in this study. The sample size was calculated based on a 99% confidence interval (CI) and a 5% margin of error. The student population enrolled in their fifth and sixth years at Japanese medical universities totaled 18,195. The required sample size was 642, and in this study, the total number of respondents was 2,594, which is 4-fold larger than required. Respondents with missing answers were excluded from the study. Of the remaining respondents, we included those who indicated having experienced a face-to-face clerkship in a hospital prior to the first emergency declaration and then subsequently switching to an online clinical clerkship. Finally, 1,671 responses were used in the derivation analysis (Fig 1).

For the validation dataset, we collected responses from 132 Japanese medical students. Among the 132 respondents, 130 (98.5%) consented to participate in this study. Of the remaining respondents, we included those who indicated having experienced a face-to-face clerkship in a hospital prior to the second emergency declaration and then subsequently switching to an online clinical clerkship. Finally, 106 responses were used for the validation analysis (Fig 1).

To compile the validation dataset, we collected responses from 132 Japanese medical students. Among the 132 respondents, 130 (98.5%) consented to participate in this study. Respondents with answers missing from their profile were excluded from the analysis (n = 0). In Japan, fourth- and fifth-year students are the equivalent years of practice, so we also excluded responses from students in other years (n = 0). The next step was to categorize the responses according to the type of education the students received mid-COVID-19. No respondents were excluded due to an inappropriate response. Of the 130 remaining respondents, 107 reported that they had experienced a face-to-face clerkship in a hospital pre-COVID-19 and then switched to an online clinical clerkship due to the COVID-19 crisis. Of the 107 respondents, those with missing responses regarding the content of the education to which they were exposed were excluded (n = 1). Finally, 106 responses were used for validation analysis in this study.

Statistical analysis

We used SPSS v27 (IBM, Armonk, NY, USA). Statistical significance was set at a two-sided p-value of 0.05.

Using the derivation dataset, we performed a multivariate logistic regression analysis to identify the factors associated with each outcome measurement of the effectiveness of online clinical clerkship, per the outcomes defined in the Measuring Outcomes section. We categorized the outcomes into the following binary: equivalent/better than face-to-face clerkship or not. The predictive variables included (a) lecture duration, (b) lecture frequency, (c) quizzes, (d) assignment submission, (e) student oral presentations, (f) observation of practice, (g) clinical skills practice, (h) participation in interprofessional meetings, (i) interactive discussions with physicians, and (j) technical Internet-related problems. We categorized the answers to (a)–(h) into three categorical variables: more opportunities in face-to-face clerkship, same opportunities in online as in face-to-face clinical clerkship, and more opportunities in online clinical clerkship. We categorized (j) (technical Internet-related problems) into three variables: not at all, a little, and a lot. In this study, we did not incorporate variables related to medical students’ profiles because doing so would not have been in keeping with the study’s aim to develop a prediction model for the effectiveness of the online clinical clerkship curriculum.

Based on multivariate logistic regression analysis, we only incorporated associated factors (p < 0.1) into the final prediction model. Using multivariate logistic regression analysis, the final model determined point values for each associated variable and developed an integer-based estimation system for satisfaction and motivation, respectively.

By applying this model to the validation dataset, we assessed the prognostic instrument’s predictive accuracy, with discrimination. Discrimination (i.e., the degree to which a model differentiates between participants answering equivalent/better satisfaction/motivation in face-to-face clerkship) was calculated using concordance (c-) statistics, ranging from 0.5 (no discrimination) to 1.0 (perfect discrimination).

Results

Table 1 presents the participants’ demographic information and the proportion of respondents that selected each answer to every survey question.

Table 1. Summary of respondents’ answers in each survey.

Participants Derivation dataset (n = 1,671) Validation dataset (n = 106)
Participants’ profiles
Sex, No. (%) Male: 979 (58.6), Female: 692 (41.4) Male: 60 (56.6), Female: 46 (43.4)
Year in medical school, No. (%) 5th: 707 (42.3), 6th: 964 (57.7) 4th: 51 (48.1), 5th: 55 (51.9)
Factors in online clinical clerkship  
Lecture durationa No. (%) 1: 287 (17.2), 2: 54 (3.2), 3: 139 (8.3), 4: 250 (15.0), 5: 406 (24.3), 6: 329 (19.7), 7: 118(7.1), 8: 88 (5.3) 1: 10 (9.4), 2: 2 (1.9), 3: 4 (3.8), 4: 21 (19.8), 5: 37 (34.9), 6: 26 (24.5), 7: 2 (1.9), 8: 4 (3.8)
Lecture frequency per weekb No. (%) 1: 287 (17.2), 2: 141 (8.4), 3: 166 (9.9), 4: 138 (8.3), 5: 181 (10.8), 6: 139 (8.3), 7: 357 (21.4), 8: 61 (3.7), 9: 50 (3.0), 10: 28 (1.7), 11: 12 (0.7), 12: 63 (3.8), 13: 48 (2.9) 1: 10 (9.4), 2: 4 (3.8), 3: 16 (15.1), 4: 21 (19.8), 5: 24 (22.6), 6: 9 (8.5), 7: 16 (15.1), 8: 2 (1.9), 9: 1 (0.9), 10: 1 (0.9), 11: 0 (0), 12: 2 (1.9), 13: 0 (0)
Opportunity to take quizzesc No. (%) 1: 357 (21.4), 2: 240 (14.4), 3: 753 (45.1), 4: 206 (12.3), 5: 115 (6.9) 1: 16 (15.1), 2: 17 (16.0), 3: 62 (58.5), 4: 10 (9.4), 5: 1 (0.9)
Opportunity to submit assignmentsc No. (%) 1: 212 (12.7), 2: 171 (10.2), 3: 473 (28.3), 4: 382 (22.9), 5: 433 (25.9) 1: 6 (5.7), 2: 6 (5.7), 3: 35 (33.0), 4: 36 (34.0), 5: 23 (21.7)
Opportunity to give oral presentationsc No. (%) 1: 700 (41.9), 2: 370 (22.1), 3: 400 (23.9), 4: 119 (7.1), 5: 82 (4.9) 1: 35 (33.0), 2: 33 (31.1), 3: 27 (25.5), 4: 9 (8.5), 5: 2 (1.9)
Opportunity to observe physicians’ practicec No. (%) 1: 1,286 (77.0), 2: 238 (14.2), 3: 87 (5.2), 4: 25 (1.5), 5: 35 (2.1) 1: 74 (69.8), 2: 22 (20.8), 3: 7 (6.6), 4: 2 (1.9), 5: 1 (0.9)
Opportunity to practice clinical skillsc No. (%) 1: 1,250 (74.8), 2: 263 (15.7), 3: 124 (7.4), 4: 14 (0.8), 5: 20 (1.2) 1: 74 (69.8), 2: 19 (17.9), 3: 12 (11.3), 4: 0 (0), 5: 1 (0.9)
Opportunity to participate in interprofessional meetingsc No. (%) 1: 1,202 (71.9), 2: 244 (14.6), 3: 152 (9.1), 4: 37 (2.2), 5: 36 (2.2) 1: 72 (67.9), 2: 20 (18.9), 3: 10 (9.4), 4: 3 (2.8), 5: 1 (0.9)
Opportunity to interact with physiciansc No. (%) 1: 798 (47.8), 2: 382 (22.9), 3: 303 (18.1), 4: 108 (6.5), 5: 80 (4.8) 1: 42 (39.6), 2: 36 (34.0), 3: 23 (21.7), 4: 5 (4.7), 5: 0 (0)
Frequency of technical Internet-related problemsd No. (%) 1: 575 (34.4), 2: 416 (24.9), 3: 296 (17.7), 4: 297 (17.8), 5: 87 (5.2) 1: 23 (21.7), 2: 27 (25.5), 3: 20 (18.9), 4: 29 (27.4), 5: 7 (6.6)
Outcome measurement
Satisfaction levele No. (%) 1: 517 (30.9), 2: 570 (34.1), 3: 348 (20.8), 4: 138 (8.3), 5: 98 (5.9) 1: 47 (44.3), 2: 35 (33.0), 3: 15 (14.2), 4: 8 (7.5), 5: 1 (0.9)
Motivation levele No. (%) 1: 543 (32.5), 2: 525 (31.4), 3: 381 (22.8), 4: 108 (6.5), 5: 114 (6.8) 1: 51 (48.1), 2: 38 (35.8), 3: 12 (11.3), 4: 4 (3.8), 5: 1 (0.9)

a: Eight choices, arranged in 15-min increments: 1) not at all, 2) ≤15 min per lecture, 3) ≥15 min to ≤30 min per lecture, 4) ≥30 min to ≤45 min per lecture, 5) ≥45 min to ≤60 min per lecture, 6) ≥60 min to ≤75 min per lecture, 7) ≥75 min to ≤90 min per lecture, and 8) ≥90 min per lecture.

b: 13 choices: 1) not at all, 2) less than once a week, 3) about once a week, 4) about twice a week, 5) about three times a week, 6) about four times a week, 7) about five times a week, 8) about six times a week, 9) about seven times a week, 10) about eight times a week, 11) about nine times a week, 12) about ten times a week, and 13) over ten times a week.

c: 5-point Likert scale, ranging from 1 = there were many more opportunities during face-to-face clinical clerkship than there were during online clinical clerkship to 5 = there were many more opportunities during online clinical clerkship than there were during face-to-face clinical clerkship.

d: 5-point Likert scale, ranging from 1 = not at all to 5 = very frequently.

e: 5-point Likert scale, ranging from 1 = the level/amount was much higher during face-to-face clinical clerkship than it was during online clinical clerkship to 5 = the level/amount was much higher during online clinical clerkship than it was during face-to-face clinical clerkship.

Identifying the factors associated with satisfaction/motivation level regarding online clinical clerkship

S1 Table shows the results of the multivariate logistic regression analysis for students’ satisfaction with online clinical clerkship. The positive factors associated with satisfaction were lecture frequency (estimate = 0.054; p = .010), quizzes (estimate = 0.291; p = .001), student oral presentations (estimate = 0.175; p = .046), observation of practice (estimate = 0.584; p = .001), clinical skills practice (estimate = 0.594; p = .002), participation in interprofessional meetings (estimate = 0.421; p = .003), and interactive discussion with physicians (estimate = 0.494; p < .001). The negative associated factor was technical problems (estimate = -0.155, p = .043).

S2 Table shows the results of the multivariate logistic regression analysis for students’ motivation with regard to online clinical clerkship. The positive factors associated with motivation were quizzes (estimate = 0.264; p = .002), student oral presentations (estimate = 0.313; p < .001), observation of practice (estimate = 0.427; p = .010), clinical skills practice (estimate = 0.657; p = .001), participation in interprofessional meetings (estimate = 0.497; p = .001), and interactive discussion with physicians (estimate = 0.338; p < .001). The negative associated factor was technical Internet-related problems (estimate = -0.270; p < .001).

Final prediction model

Having identified the associated variables via multivariate logistic regression analysis, we again performed multivariate logistic regression analysis, this time incorporating only the associated variables identified in the first analysis. In the prediction model for satisfaction, eight predictors remained in the final multivariable model after simplification: lecture frequency (estimate = 0.049; p = .007), quizzes (estimate = 0.321; p < .001), oral presentations (estimate = 0.192; p = .027), observation (estimate = 0.583; p = .001), practice (estimate = 0.586; p = .002), interprofessional meetings (estimate = 0.416; p = .004), interactive discussion (estimate = 0.492; p < .001), and technical problems (estimate = -0.158; p = .038) (S3 Table). In the prediction model for motivation, seven predictors remained in the final multivariable model after simplification: quizzes (estimate = 0.270; p = .001), oral presentations (estimate = 0.331; p < .001), observation (estimate = 0.422; p = .010), practice (estimate = 0.651; p = .001), interprofessional meetings (estimate = 0.497; p = .001)), interactive discussion (estimate = 0.358; p < .001), and technical problems (estimate = -0.248; p = .001) (S4 Table). Based on the results of this final multivariate logistic regression analysis, we assigned point values to these variables and developed an integer-based estimation system for satisfaction and motivation, respectively (Table 2).

Table 2. Scores for each variable.

Satisfaction Motivation
Estimate Estimate/0.06 Opportunities compared to face-to-face clerkship Estimate Estimate/0.06 Opportunities compared to face-to-face clerkship
Less Same More Score Less Same More Score
Lecture frequency 0.049 0.82 a /12 - - - - - -
Quizzes 0.321 5.35 0 5 10 /10 0.270 4.50 0 5 10 /10
Oral presentations 0.192 3.20 0 3 6 /6 0.331 5.52 0 6 12 /12
Observation 0.583 9.72 0 10 20 /20 0.422 7.03 0 7 14 /14
Practice 0.586 9.77 0 10 20 /20 0.651 10.85 0 11 22 /22
Interprofessional meetings 0.416 6.93 0 7 14 /14 0.497 8.28 0 8 16 /16
Interactive discussion 0.492 8.20 0 8 16 /16 0.358 5.97 0 6 12 /12
Technical problems -0.158 -2.63 0 (Not at all) -3 (A little) -6 (A lot) /-6 -0.248 -3.97 0 (Not at all) -4 (A little) -8 (A lot) /-8
Sum /98 /86
Cut-off value 22 17

a: Not at all: 0, less than once a week: 1, once a week: 2, twice a week: 3, three times a week: 4, four times a week: 5, five times a week: 6, six times a week: 7, seven times a week: 8, eight times a week: 9, nine times a week: 10, ten times a week: 11, over ten times a week: 12.

Validation analysis to assess the scoring model’s prediction performance

We determined our predictive instrument’s discrimination by applying it to the validation dataset. Based on these scores, we created receiver operating characteristic (ROC) curves to predict whether students’ satisfaction/motivation was equivalent/better than in face-to-face clerkship, as shown in Fig 2.

Fig 2. The prediction models’ accuracy for satisfaction and motivation.

Fig 2

We applied the scoring prediction model developed using the derivation dataset to the validation dataset. Based on the model’s prediction performance in the validation dataset, we created receiver operating characteristic (ROC) curves for satisfaction and motivation. (A) A given ROC curve delineates predictive accuracy for satisfaction. (B) A given ROC curve delineates predictive accuracy for motivation. AUC: area under the curve.

The c-statistics, which indicate the area under the curve to predict satisfaction and motivation, were 0.69 · (95% CI: 0.55–0.83, p = .005) and 0.75 · (95% CI: 0.60–0.90, p = .001), respectively. When the cut-off value for the satisfaction scores was set to 22, sensitivity was 0.50, and specificity was 0.89. When the cut-off value for the motivation scores was set to 17, sensitivity was 0.71 and specificity was 0.85.

Discussion

We developed a practical instrument to predict the effectiveness of online clinical clerkship based on medical students’ satisfaction and motivation. Although an instrument to predict the online clinical clerkship curriculum’s effectiveness has not yet been developed, such an instrument is needed, especially given the increase in demand for online clinical clerkship due to the COVID-19 pandemic.

As the results show, our model’s prediction performance was relatively high (satisfaction: 0.69, motivation: 0.75). Satisfaction and motivation are based on individual self-assessment, which means that different people can rate their satisfaction with and motivation for the same curriculum differently. Individual factors (such as sex and year in medical school) are reported to be related to satisfaction and motivation [30, 32]. Nevertheless, we did not incorporate individual components into the score in our prediction model. This was because our goal was to develop a scoring model that predicts the effectiveness of online clinical clerkship curricula. In short, our model’s prediction performance was high in spite of not incorporating individual demographic information.

Using our scoring system, educators will be able to both predict the effectiveness of and improve their online clinical clerkships. Our instrument will be of great help to educators in building a better online clinical clerkship curriculum. In our study, the best cutoff values for motivation and satisfaction were 22 and 17, respectively (Table 2). Educators can use these scores as guides to improve online clinical clerkship by enhancing the educational approaches in their curriculum, with the goal of exceeding the scores we obtained. For example, to maintain medical students’ satisfaction and motivation during online clinical clerkships in a manner that is comparable to face-to-face clerkship, educators might consider providing the students with the same opportunities to observe physicians’ practice and to practice their clinical skills as face-to-face clinical clerkship (at only these two variables, score for satisfaction:20, and score for motivation:18) (Table 2).

This study has several strengths. The assembled population is one of the largest study sample sizes for predicting the effectiveness of online clinical clerkship. Our instruments were developed using big data from national-level cross-sectional studies involving a total of 1,671 medical students in the derivation dataset and 106 students in the validation dataset. Another strength is that our scoring system is a practical instrument that can be applied to the online clinical clerkship curriculum. The predictors are well defined and easily measured, so that students’ online opportunities to experience each educational approach can be evaluated as equivalent to, or more or less frequent than in face-to-face clerkship. We also made technical troubles simple to evaluate in three frequency categories: never, occasionally, and often.

Our study has some limitations. First, the cross-sectional study retrospectively measures subjects’ exposure and outcome simultaneously. Thus, recall bias was unavoidable. Second, our prediction model is based on medical students’ satisfaction and motivation, which are self-assessed outcomes that are easily measured using a questionnaire. However, knowledge and skill acquisition, which are Level 2b in Kirkpatrick’s assessment model, are more objective outcomes [26, 2931]. Predictive instruments based on such outcomes, as opposed to satisfaction and/or motivation, might be more useful. Therefore, more research might be required to show the correlation with exam scores or patient outcomes before using this study as a tool to modify curricula. The third limitation is that qualitative factors on the part of the faculty were not considered in the current model because the current study was a quantitative survey. For example, the content and quality of lectures, the faculty’s skills of online facilitation might also affect the students’ satisfaction and motivation. Forth, in this study, we only included medical students in Japan. Future research should consider the applicability of these results to other countries. Online clinical clerkship would greatly depend on health care resources, Internet infrastructure in countries/communities, and whether students are familiar with the devices required to access online education [33, 34]. This indicates that our model would be difficult to apply in countries with limited healthcare resources and/or a lack of Internet infrastructure. Another limitation is that these data are based on online clinical clerkship during the COVID-19 pandemic. The impact of COVID-19 varies across countries and communities. Furthermore, online clinical clerkship post-COVID-19 pandemic may be of a different nature (e.g., a combination of face-to-face and online). We believe that our prediction model can be applied in other countries’ online clinical clerkship curriculum or post-COVID-19 pandemic. However, we need to evaluate our model’s validity in other countries and post-COVID-19 pandemic in future studies.

Conclusion

We successfully developed a scoring model to predict the effectiveness of the online clinical clerkship curriculum, based on students’ satisfaction and motivation. Our scoring model will accurately predict and improve the online clinical clerkship curriculum’s effectiveness.

Supporting information

S1 Method. Detailed description of the answer options for each survey question.

(DOCX)

S1 Table. Multivariate logistic regression analysis to identify the factors associated with maintaining medical students’ satisfaction with online clerkship (Level 1 in Kirkpatrick’s assessment model).

(DOCX)

S2 Table. Multivariate logistic regression analysis to identify the factors associated with maintaining medical students’ motivation during online clerkship (Level 2a in Kirkpatrick’s assessment model).

(DOCX)

S3 Table. The final model using multivariate logistic regression analysis to predict medical students’ satisfaction with online clerkship (Level 1 in Kirkpatrick’s assessment model).

(DOCX)

S4 Table. The final model using multivariate logistic regression analysis to predict medical students’ motivation during online clerkship (Level 2a in Kirkpatrick’s assessment model).

(DOCX)

S1 Dataset

(XLSX)

S2 Dataset

(XLSX)

Acknowledgments

The authors are grateful to Dr. Eishi Asano at Children’s Hospital of Michigan, Detroit Medical Center, Wayne State University for the advice on statistical analysis. The authors are also grateful to all participants of this survey-based study.

Data Availability

All relevant data are within the paper and its Supporting Information files.

Funding Statement

M.K. was supported by JSPS KAKENHI (grant number 17H04097). URL: https://www.jsps.go.jp/english/index.html This funder did not any role in this study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Chen BY, Kern DE, Kearns RM, et al. From modules to MOOCs: application of the six-step approach to online curriculum development for medical education. Acad Med. 2019;94(5):678–685. doi: 10.1097/ACM.0000000000002580 [DOI] [PubMed] [Google Scholar]
  • 2.Pickering JD, Henningsohn L, DeRuiter MC, et al. Twelve tips for developing and delivering a massive open online course in medical education. Med Teach. 2017;39(7):691–696. doi: 10.1080/0142159X.2017.1322189 [DOI] [PubMed] [Google Scholar]
  • 3.Pei L, Wu H. Does online learning work better than offline learning in undergraduate medical education? A systematic review and meta-analysis. Med Educ Online. 2019;24(1):1666538. doi: 10.1080/10872981.2019.1666538 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Daniel M, Gordon M, Patricio M, et al. An update on developments in medical education in response to the COVID-19 pandemic: a BEME scoping review: BEME Guide No. 64 published online January 26, 2021. Med Teach. 2021;43(3):253–271. doi: 10.1080/0142159X.2020.1864310 [DOI] [PubMed] [Google Scholar]
  • 5.Gibbs T. The Covid-19 pandemic: provoking thought and encouraging change. Med Teach. 2020;42(7):738–740. doi: 10.1080/0142159X.2020.1775967 [DOI] [PubMed] [Google Scholar]
  • 6.Gordon M, Patricio M, Horne L, et al. Developments in medical education in response to the COVID-19 pandemic: a rapid BEME systematic review: BEME Guide No. 63. Med Teach. 2020;42(11):1202–1215. doi: 10.1080/0142159X.2020.1807484 [DOI] [PubMed] [Google Scholar]
  • 7.Tolsgaard MG, Cleland J, Wilkinson T, Ellaway RH. How we make choices and sacrifices in medical education during the COVID-19 pandemic. Med Teach. 2020;42(7):741–743. doi: 10.1080/0142159X.2020.1767769 [DOI] [PubMed] [Google Scholar]
  • 8.Yang DY, Cheng SY, Wang SZ, et al. Preparedness of medical education in China: lessons from the COVID-19 outbreak. Med Teach. 2020;42(7):787–790. doi: 10.1080/0142159X.2020.1770713 [DOI] [PubMed] [Google Scholar]
  • 9.Rose S. Medical student education in the time of COVID-19. JAMA. 2020;323(21):2131–2132. doi: 10.1001/jama.2020.5227 [DOI] [PubMed] [Google Scholar]
  • 10.Chao TN, Frost AS, Brody RM, et al. Creation of an interactive virtual surgical rotation for undergraduate medical education during the COVID-19 pandemic. J Surg Educ. 2021;78(1):346–350. doi: 10.1016/j.jsurg.2020.06.039 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Co M, Chu KM. Distant surgical teaching during COVID-19—a pilot study on final year medical students. Surg Pract. 2020;24:105–109. doi: 10.1111/1744-1633.12436 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Jack MM, Gattozzi DA, Camarata PJ, et al. Live-streaming surgery for medical student education—educational solutions in neurosurgery during the COVID-19 pandemic. J Surg Educ. 2021;78(1):99–103. doi: 10.1016/j.jsurg.2020.07.005 [DOI] [PubMed] [Google Scholar]
  • 13.Mansouri M, Lockyer J. A meta-analysis of continuing medical education effectiveness. J Contin Educ Health Prof. 2007;27(1):6–15. doi: 10.1002/chp.88 [DOI] [PubMed] [Google Scholar]
  • 14.Mustafa T, Farooq Z, Asad Z, et al. Lectures in medical educaton: what students think? J Ayub Med Coll Abbottabad. 2014;26(1):21–25. [PubMed] [Google Scholar]
  • 15.Hashmi NR, Daud S, Manzoor I. Medical education: views and recommendations by final year MBBS students of a private medical college in Lahore. J Coll Phys Surg Pak. 2010;20(2):93–97. doi: 02.2010/JCPSP.9397 [PubMed] [Google Scholar]
  • 16.Najmi RS. Lecture as a mode of instruction in undergraduate medical education. J Pak Med Assoc. 1999;49(2):30–33. [PubMed] [Google Scholar]
  • 17.Manzoor I, Mumtaz A, Habib M, et al. Lectures in medical education: students’ views. J Ayub Med Coll Abbottabad. 2011;23(4):118–121. [PubMed] [Google Scholar]
  • 18.Karpicke JD, Roediger HL, III. The critical importance of retrieval for learning. Science. 2008;319(5865):966–968. doi: 10.1126/science.1152408 [DOI] [PubMed] [Google Scholar]
  • 19.Larsen DP, Butler AC, Roediger HL III. Repeated testing improves long-term retention relative to repeated study: a randomised controlled trial. Med Educ. 2009;43(12):1174–1181. doi: 10.1111/j.1365-2923.2009.03518.x [DOI] [PubMed] [Google Scholar]
  • 20.Epp S. The value of reflective journaling in undergraduate nursing education: a literature review. Int J Nurs Stud. 2008;45(9):1379–1388. doi: 10.1016/j.ijnurstu.2008.01.006 [DOI] [PubMed] [Google Scholar]
  • 21.August E, Burke K, Fleischer C, et al. Writing assignments in epidemiology courses: how many and how good? Public Health Rep. 2019;134(4):441–446. doi: 10.1177/0033354919849942 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Plack MM, Driscoll M, Marquez M, et al. Peer-facilitated virtual action learning: reflecting on critical incidents during a pediatric clerkship. Acad Pediatr. 2010;10(2):146–152. doi: 10.1016/j.acap.2009.11.007 [DOI] [PubMed] [Google Scholar]
  • 23.Singh K, Bharatha A, Sa B, et al. Teaching anatomy using an active and engaging learning strategy. BMC Med Educ. 2019;19(1):149. doi: 10.1186/s12909-019-1590-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Reed T, Pirotte M, McHugh M, et al. Simulation-based mastery learning improves medical student performance and retention of core clinical skills. Simul Healthc. 2016;11(3):173–180. doi: 10.1097/SIH.0000000000000154 [DOI] [PubMed] [Google Scholar]
  • 25.Mann KV, Mcfetridge-Durdle J, Martin-Misener R, et al. Interprofessional education for students of the health professions: the “Seamless Care” model. J Interprof Care. 2009;23(3):224–233. doi: 10.1080/13561820802697735 [DOI] [PubMed] [Google Scholar]
  • 26.Peiris-John R, Selak V, Robb G, et al. The state of quality improvement teaching in medical schools: a systematic review. J Surg Educ. 2020;77(4):889–904. doi: 10.1016/j.jsurg.2020.01.003 [DOI] [PubMed] [Google Scholar]
  • 27.Alsoufi A, Alsuyihili A, Msherghi A, et al. Impact of the COVID-19 pandemic on medical education: medical students’ knowledge, attitudes, and practices regarding electronic learning. PLOS ONE. 2020;15(11):e0242905. doi: 10.1371/journal.pone.0242905 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Dost S, Hossain A, Shehab M, et al. Perceptions of medical students towards online teaching during the COVID-19 pandemic: a national cross-sectional survey of 2721 UK medical students. BMJ Open. 2020;10(11):e042378. doi: 10.1136/bmjopen-2020-042378 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Kirkpatrick DL. Evaluation of training. In:, Craig RL, Bittel LR, eds. Training and Development Handbook. New York, NY: McGraw-Hill; 1967:87112. [Google Scholar]
  • 30.Colthart I, Bagnall G, Evans A, et al. The effectiveness of self-assessment on the identification of learner needs, learner activity, and impact on clinical practice: BEME Guide no. 10. Med Teach. 2008;30(2):124–145. doi: 10.1080/01421590701881699 [DOI] [PubMed] [Google Scholar]
  • 31.Wong BM, Etchells EE, Kuper A, et al. Teaching quality improvement and patient safety to trainees: a systematic review. Acad Med. 2010;85(9):1425–1439. doi: 10.1097/ACM.0b013e3181e2d0c6 [DOI] [PubMed] [Google Scholar]
  • 32.Çalışkan F, Mıdık Ö, Baykan Z, et al. The knowledge level and perceptions toward COVID-19 among Turkish final year medical students. Postgrad Med. 2020;132(8):764–772. doi: 10.1080/00325481.2020.1795486 [DOI] [PubMed] [Google Scholar]
  • 33.Walsh S, De Villiers MR, Golakai VK. Introducing an e-learning solution for medical education in Liberia. Ann Glob Health. 2018;84(1):190–197. doi: 10.29024/aogh.21 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Barteit S, Neuhann F, Bärnighausen T, et al. Perspectives of nonphysician clinical students and medical lecturers on tablet-based health care practice support for medical education in Zambia, Africa: qualitative study. JMIR MHealth UHealth. 2019;7(1):e12637. doi: 10.2196/12637 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Feroze Kaliyadan

5 Jan 2022

PONE-D-21-32687Predicting the effectiveness of the online clinical clerkship curriculum: Development of a multivariate prediction model and validation studyPLOS ONE

Dear Dr. Kuroda,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Feb 19 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Feroze Kaliyadan, M.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

Additional Editor Comments:

The concept of developing a predictive model for evaluating effectiveness of the online clinical clerkship curriculum is important, interesting and very relevant.

You will need to add more details on the questionnaire validation process.

A brief summary of the curricular model followed (especially for year 5 and 6 )normally needs to be added to put things in context

Limitations need to be elaborated, would have been especially good if faculty perspective was also incorporated

Finally, the discussion needs to elaborate on how far the results can be extrapolated to other regions

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: I Don't Know

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Article is recommended for submission since it meets all the criteria of the journal. The tool for validation of online clerkship is something that is a requisite today. The statistical works and the language of the paper are acceptable. The reference styling is acceptable.

Reviewer #2: The study is very relevant considering the pandemic and could be very useful to Medical Colleges all over the world to evaluate and improve on their curricula.

A potential source of the results not reflecting actual improvement in the curricula, is the fact, that in the current scenario online learning is more of a necessity rather than an option. The student satisfaction being high could be reflecting their greater comfort rather than better learning outcomes as the study is only assessing student satisfaction and has no information regarding correlation with exam scores or patient outcomes. More research might be required in these avenues before using this study as a tool to modify curricula.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Jan 27;17(1):e0263182. doi: 10.1371/journal.pone.0263182.r002

Author response to Decision Letter 0


12 Jan 2022

Reviewer #1: Article is recommended for submission since it meets all the criteria of the journal. The tool for validation of online clerkship is something that is a requisite today. The statistical works and the language of the paper are acceptable. The reference styling is acceptable.

→Thank you for your favorable comments.

Reviewer #2: The study is very relevant considering the pandemic and could be very useful to Medical Colleges all over the world to evaluate and improve on their curricula.

A potential source of the results not reflecting actual improvement in the curricula, is the fact, that in the current scenario online learning is more of a necessity rather than an option. The student satisfaction being high could be reflecting their greater comfort rather than better learning outcomes as the study is only assessing student satisfaction and has no information regarding correlation with exam scores or patient outcomes. More research might be required in these avenues before using this study as a tool to modify curricula.

→Thank you for your comments. As you pointed out, the clear limitation of this study is that we did not investigate the correlation with exam scores or patient’s outcomes. We added this limitation as follows:

Therefore, more research might be required to show the correlation with exam scores or patient outcomes before using this study as a tool to modify curricula.

Attachment

Submitted filename: rebuttal_letter_nk5.docx

Decision Letter 1

Feroze Kaliyadan

14 Jan 2022

Predicting the effectiveness of the online clinical clerkship curriculum: Development of a multivariate prediction model and validation study

PONE-D-21-32687R1

Dear Dr. Kuroda,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Feroze Kaliyadan, M.D.

Academic Editor

PLOS ONE

Acceptance letter

Feroze Kaliyadan

19 Jan 2022

PONE-D-21-32687R1

Predicting the effectiveness of the online clinical clerkship curriculum: Development of a multivariate prediction model and validation study

Dear Dr. Kuroda:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Feroze Kaliyadan

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Method. Detailed description of the answer options for each survey question.

    (DOCX)

    S1 Table. Multivariate logistic regression analysis to identify the factors associated with maintaining medical students’ satisfaction with online clerkship (Level 1 in Kirkpatrick’s assessment model).

    (DOCX)

    S2 Table. Multivariate logistic regression analysis to identify the factors associated with maintaining medical students’ motivation during online clerkship (Level 2a in Kirkpatrick’s assessment model).

    (DOCX)

    S3 Table. The final model using multivariate logistic regression analysis to predict medical students’ satisfaction with online clerkship (Level 1 in Kirkpatrick’s assessment model).

    (DOCX)

    S4 Table. The final model using multivariate logistic regression analysis to predict medical students’ motivation during online clerkship (Level 2a in Kirkpatrick’s assessment model).

    (DOCX)

    S1 Dataset

    (XLSX)

    S2 Dataset

    (XLSX)

    Attachment

    Submitted filename: rebuttal_letter_nk5.docx

    Data Availability Statement

    All relevant data are within the paper and its Supporting Information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES